Searched refs:TH_BUCKET_SHARE_BG (Results 1 – 9 of 9) sorted by relevance
| /xnu-11417.140.69/tests/sched/ |
| H A D | clutch_runqueue.c | 26 …t_thread_t lowest_bound = create_thread(TH_BUCKET_SHARE_BG, low_tg, root_bucket_to_highest_pri[TH_… 30 …thread_t lowest_unbound = create_thread(TH_BUCKET_SHARE_BG, low_tg, root_bucket_to_highest_pri[TH_… 103 for (int bucket = TH_BUCKET_SHARE_BG; bucket >= TH_BUCKET_SHARE_FG; bucket--) { 104 if (bucket < TH_BUCKET_SHARE_BG) { 122 for (int bucket = TH_BUCKET_SHARE_BG; bucket >= TH_BUCKET_SHARE_FG; bucket--) { 123 if (bucket < TH_BUCKET_SHARE_BG) { 306 for (int bucket = TH_BUCKET_SHARE_BG; bucket >= TH_BUCKET_SHARE_FG; bucket--) { 307 if (bucket < TH_BUCKET_SHARE_BG) { 317 …root_bucket_arg = (bucket < TH_BUCKET_SHARE_BG ? SELECTION_OPENED_WARP_WINDOW : SELECTION_WAS_EDF)… 328 root_bucket_arg = (bucket < TH_BUCKET_SHARE_BG ? 0 : SELECTION_WAS_EDF) | CTS_VERSION; [all …]
|
| H A D | edge_runqueue.c | 81 …threads[3] = create_thread(TH_BUCKET_SHARE_BG, tg, root_bucket_to_highest_pri[TH_BUCKET_SHARE_BG]);
|
| H A D | edge_migration.c | 67 …test_thread_t background = create_thread(TH_BUCKET_SHARE_BG, tg, root_bucket_to_highest_pri[TH_BUC…
|
| /xnu-11417.140.69/osfmk/kern/ |
| H A D | sched_average.c | 195 load_now[TH_BUCKET_SHARE_BG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_BG], relaxed); in compute_sched_load() 206 load_now[TH_BUCKET_SHARE_BG], load_now[TH_BUCKET_SHARE_UT], 0); in compute_sched_load() 216 for (uint32_t i = TH_BUCKET_SHARE_FG; i <= TH_BUCKET_SHARE_BG; i++) { in compute_sched_load() 245 if (load_now[TH_BUCKET_SHARE_BG] > 0) { in compute_sched_load() 259 for (uint32_t i = TH_BUCKET_SHARE_FG; i <= TH_BUCKET_SHARE_BG; i++) { in compute_sched_load() 282 …SCALE(sched_load[TH_BUCKET_SHARE_UT]), SCHED_LOAD_EWMA_UNSCALE(sched_load[TH_BUCKET_SHARE_BG]), 0); in compute_sched_load() 292 for (uint32_t i = TH_BUCKET_SHARE_FG; i <= TH_BUCKET_SHARE_BG; i++) { in compute_averages()
|
| H A D | priority.c | 750 bucket <= TH_BUCKET_SHARE_BG); in sched_incr_bucket() 759 bucket <= TH_BUCKET_SHARE_BG); in sched_decr_bucket() 770 bucket <= TH_BUCKET_SHARE_BG); in sched_add_bucket() 779 bucket <= TH_BUCKET_SHARE_BG); in sched_sub_bucket() 863 new_bucket = TH_BUCKET_SHARE_BG; in sched_update_thread_bucket() 903 new_bucket = TH_BUCKET_SHARE_BG; in sched_smt_update_thread_bucket()
|
| H A D | sched.h | 225 TH_BUCKET_SHARE_BG, /* Timeshare thread between MAXPRI_THROTTLE and MINPRI */ enumerator
|
| H A D | sched_clutch.c | 347 bucket = TH_BUCKET_SHARE_BG; in sched_convert_pri_to_bucket() 1532 …case TH_BUCKET_SHARE_BG: os_atomic_inc(&root_clutch->scr_cumulative_run_count[TH_BUCKET_SHARE_BG]… in sched_edge_cluster_cumulative_count_incr() 1547 …case TH_BUCKET_SHARE_BG: os_atomic_dec(&root_clutch->scr_cumulative_run_count[TH_BUCKET_SHARE_BG]… in sched_edge_cluster_cumulative_count_decr()
|
| H A D | thread_group.c | 1300 …[TH_BUCKET_SHARE_BG] = (overrides[PERFCONTROL_CLASS_BACKGROUND] != SCHED_PERFCONTROL_PREFERRED_CLU… in sched_perfcontrol_thread_group_preferred_clusters_set()
|
| H A D | sched_prim.c | 4291 if ((thread != THREAD_NULL) && thread->th_sched_bucket == TH_BUCKET_SHARE_BG) { in sched_timeshare_initial_quantum_size() 7542 sched_pri_shifts[TH_BUCKET_SHARE_FG], sched_pri_shifts[TH_BUCKET_SHARE_BG], in sched_timeshare_maintenance_continue()
|