Searched refs:TH_BUCKET_SHARE_BG (Results 1 – 10 of 10) sorted by relevance
| /xnu-12377.41.6/tests/sched/ |
| H A D | clutch_runqueue.c | 26 …t_thread_t lowest_bound = create_thread(TH_BUCKET_SHARE_BG, low_tg, root_bucket_to_highest_pri[TH_… 30 …thread_t lowest_unbound = create_thread(TH_BUCKET_SHARE_BG, low_tg, root_bucket_to_highest_pri[TH_… 103 for (int bucket = TH_BUCKET_SHARE_BG; bucket >= TH_BUCKET_SHARE_FG; bucket--) { 104 if (bucket < TH_BUCKET_SHARE_BG) { 122 for (int bucket = TH_BUCKET_SHARE_BG; bucket >= TH_BUCKET_SHARE_FG; bucket--) { 123 if (bucket < TH_BUCKET_SHARE_BG) { 306 for (int bucket = TH_BUCKET_SHARE_BG; bucket >= TH_BUCKET_SHARE_FG; bucket--) { 307 if (bucket < TH_BUCKET_SHARE_BG) { 317 …root_bucket_arg = (bucket < TH_BUCKET_SHARE_BG ? SELECTION_OPENED_WARP_WINDOW : SELECTION_WAS_EDF)… 328 root_bucket_arg = (bucket < TH_BUCKET_SHARE_BG ? 0 : SELECTION_WAS_EDF) | CTS_VERSION; [all …]
|
| H A D | edge_runqueue.c | 81 …threads[3] = create_thread(TH_BUCKET_SHARE_BG, tg, root_bucket_to_highest_pri[TH_BUCKET_SHARE_BG]);
|
| H A D | edge_migration.c | 67 …test_thread_t background = create_thread(TH_BUCKET_SHARE_BG, tg, root_bucket_to_highest_pri[TH_BUC…
|
| /xnu-12377.41.6/osfmk/kern/ |
| H A D | sched_average.c | 195 load_now[TH_BUCKET_SHARE_BG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_BG], relaxed); in compute_sched_load() 206 load_now[TH_BUCKET_SHARE_BG], load_now[TH_BUCKET_SHARE_UT], 0); in compute_sched_load() 216 for (uint32_t i = TH_BUCKET_SHARE_FG; i <= TH_BUCKET_SHARE_BG; i++) { in compute_sched_load() 245 if (load_now[TH_BUCKET_SHARE_BG] > 0) { in compute_sched_load() 259 for (uint32_t i = TH_BUCKET_SHARE_FG; i <= TH_BUCKET_SHARE_BG; i++) { in compute_sched_load() 282 …SCALE(sched_load[TH_BUCKET_SHARE_UT]), SCHED_LOAD_EWMA_UNSCALE(sched_load[TH_BUCKET_SHARE_BG]), 0); in compute_sched_load() 292 for (uint32_t i = TH_BUCKET_SHARE_FG; i <= TH_BUCKET_SHARE_BG; i++) { in compute_averages()
|
| H A D | priority.c | 750 bucket <= TH_BUCKET_SHARE_BG); in sched_incr_bucket() 759 bucket <= TH_BUCKET_SHARE_BG); in sched_decr_bucket() 770 bucket <= TH_BUCKET_SHARE_BG); in sched_add_bucket() 779 bucket <= TH_BUCKET_SHARE_BG); in sched_sub_bucket() 863 new_bucket = TH_BUCKET_SHARE_BG; in sched_update_thread_bucket() 903 new_bucket = TH_BUCKET_SHARE_BG; in sched_smt_update_thread_bucket()
|
| H A D | sched.h | 230 TH_BUCKET_SHARE_BG, /* Timeshare thread between MAXPRI_THROTTLE and MINPRI */ enumerator
|
| H A D | sched_clutch.c | 367 bucket = TH_BUCKET_SHARE_BG; in sched_convert_pri_to_bucket() 1557 …case TH_BUCKET_SHARE_BG: os_atomic_inc(&root_clutch->scr_cumulative_run_count[TH_BUCKET_SHARE_BG]… in sched_edge_cluster_cumulative_count_incr() 1572 …case TH_BUCKET_SHARE_BG: os_atomic_dec(&root_clutch->scr_cumulative_run_count[TH_BUCKET_SHARE_BG]… in sched_edge_cluster_cumulative_count_decr()
|
| H A D | thread_group.c | 1308 case TH_BUCKET_SHARE_BG: in sched_bucket_to_perfcontrol_class()
|
| H A D | sched_prim.c | 3969 if ((thread != THREAD_NULL) && thread->th_sched_bucket == TH_BUCKET_SHARE_BG) { in sched_timeshare_initial_quantum_size() 6694 sched_pri_shifts[TH_BUCKET_SHARE_FG], sched_pri_shifts[TH_BUCKET_SHARE_BG], in sched_timeshare_maintenance_continue()
|
| /xnu-12377.41.6/tests/sched/sched_test_harness/ |
| H A D | sched_clutch_harness_impl.c | 164 thread_t idle_thread = clutch_impl_create_thread(TH_BUCKET_SHARE_BG, not_real_idle_tg, IDLEPRI); in clutch_impl_init_topology()
|