Searched refs:preferred_pset_id (Results 1 – 2 of 2) sorted by relevance
| /xnu-12377.1.9/tests/sched/ |
| H A D | rt_migration.c | 102 for (int preferred_pset_id = 0; preferred_pset_id < topo.num_psets; preferred_pset_id++) { variable 103 set_tg_sched_bucket_preferred_pset(tg, TH_BUCKET_FIXPRI, preferred_pset_id); 104 sched_policy_push_metadata("preferred_pset_id", preferred_pset_id); 107 choose_pset_for_thread_expect(thread, preferred_pset_id); 114 for (int i = 0; i < topo.psets[preferred_pset_id].num_cpus; i++) { 115 int cpu_id = pset_id_to_cpu_id(preferred_pset_id) + i; 120 T_QUIET; T_EXPECT_NE(chosen, preferred_pset_id, "chose an unloaded cluster"); 121 …T_QUIET; T_EXPECT_EQ(topo.psets[chosen].cpu_type, topo.psets[preferred_pset_id].cpu_type, "chose a… 124 cpu_set_thread_current(pset_id_to_cpu_id(preferred_pset_id), later_thread); 126 T_QUIET; T_EXPECT_EQ(chosen, preferred_pset_id, "preempting later-deadline thread"); [all …]
|
| H A D | edge_migration.c | 277 for (int preferred_pset_id = 0; preferred_pset_id < topo.num_psets; preferred_pset_id++) { variable 278 set_tg_sched_bucket_preferred_pset(tg, sched_bucket, preferred_pset_id); 279 sched_policy_push_metadata("preferred_pset_id", preferred_pset_id); 295 …bool should_rebalance = (topo.psets[evaluate_pset].cpu_type == topo.psets[preferred_pset_id].cpu_t… 296 (topo.psets[running_on_pset_id].cpu_type != topo.psets[preferred_pset_id].cpu_type); 305 if ((topo.psets[p].cpu_type == topo.psets[preferred_pset_id].cpu_type) && 315 if ((topo.psets[p].cpu_type == topo.psets[preferred_pset_id].cpu_type) && 335 (topo.psets[preferred_pset_id].cpu_type != TEST_CPU_TYPE_PERFORMANCE))) { 338 …rebalance_steal = (topo.psets[evaluate_pset].cpu_type == topo.psets[preferred_pset_id].cpu_type) && 339 (topo.psets[enqueued_pset].cpu_type != topo.psets[preferred_pset_id].cpu_type); [all …]
|