Searched refs:TH_BUCKET_SHARE_DF (Results 1 – 9 of 9) sorted by relevance
| /xnu-12377.81.4/tests/sched/ |
| H A D | edge_runqueue.c | 20 …thread_t thread = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BUCKET_SHARE… 26 ret = tracepoint_expect(CLUTCH_THREAD_SELECT, i, 0, TH_BUCKET_SHARE_DF, bound_arg); 76 …threads[1] = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BUCKET_SHARE_DF]); 94 …tie_break_threads[k * 2] = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BUC… 100 …_break_threads[k * 2 + 1] = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BU… 116 …_t low_iscore_bound = create_thread(TH_BUCKET_SHARE_DF, low_iscore_tg, root_bucket_to_highest_pri[… 118 …t high_iscore_bound = create_thread(TH_BUCKET_SHARE_DF, high_iscore_tg, root_bucket_to_highest_pri…
|
| H A D | clutch_runqueue.c | 41 …test_thread_t bound = create_thread(TH_BUCKET_SHARE_DF, low_tg, root_bucket_to_highest_pri[TH_BUCK… 43 …ead_t higherpri_unbound = create_thread(TH_BUCKET_SHARE_DF, low_tg, root_bucket_to_highest_pri[TH_… 44 …tive_higherpri_unbound = create_thread(TH_BUCKET_SHARE_DF, high_tg, root_bucket_to_highest_pri[TH_… 45 …ctive_lowerpri_unbound = create_thread(TH_BUCKET_SHARE_DF, high_tg, root_bucket_to_highest_pri[TH_… 411 …_thread_t bound_thread = create_thread(TH_BUCKET_SHARE_DF, same_tg, root_bucket_to_highest_pri[TH_… 417 ret = tracepoint_expect(CLUTCH_THREAD_SELECT, 10, 0, TH_BUCKET_SHARE_DF, root_bucket_arg); 432 …st_thread_t def_thread = create_thread(TH_BUCKET_SHARE_DF, same_tg, root_bucket_to_highest_pri[TH_… 437 increment_mock_time_us(clutch_root_bucket_wcel_us[TH_BUCKET_SHARE_DF] + 1); 451 ret = tracepoint_expect(CLUTCH_THREAD_SELECT, 0, 0, TH_BUCKET_SHARE_DF, root_bucket_arg); 456 increment_mock_time_us(clutch_root_bucket_wcel_us[TH_BUCKET_SHARE_DF] + 1); [all …]
|
| H A D | edge_migration.c | 24 …threads[i] = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BUCKET_SHARE_DF]); 26 set_pset_load_avg(i, TH_BUCKET_SHARE_DF, low_load); 32 set_pset_load_avg(j, TH_BUCKET_SHARE_DF, high_load); 35 set_pset_load_avg(j, TH_BUCKET_SHARE_DF, low_load); 47 set_pset_load_avg(replacement_pset, TH_BUCKET_SHARE_DF, idle_load); 55 set_pset_load_avg(replacement_pset, TH_BUCKET_SHARE_DF, low_load); 67 …test_thread_t yielder = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BUCKET… 121 …test_thread_t starts_p = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BUCKE… 122 …test_thread_t starts_e = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BUCKE… 123 …t_thread_t other_p_thread = create_thread(TH_BUCKET_SHARE_DF, tg, root_bucket_to_highest_pri[TH_BU… [all …]
|
| /xnu-12377.81.4/osfmk/kern/ |
| H A D | sched_average.c | 193 load_now[TH_BUCKET_SHARE_DF] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_DF], relaxed); in compute_sched_load() 205 load_now[TH_BUCKET_FIXPRI], (load_now[TH_BUCKET_SHARE_FG] + load_now[TH_BUCKET_SHARE_DF]), in compute_sched_load() 225 if (load_now[TH_BUCKET_SHARE_DF] <= (ncpus * 2)) { in compute_sched_load() 226 load_now[TH_BUCKET_SHARE_FG] += load_now[TH_BUCKET_SHARE_DF]; in compute_sched_load() 236 load_now[TH_BUCKET_SHARE_DF] += load_now[TH_BUCKET_SHARE_UT]; in compute_sched_load() 239 load_now[TH_BUCKET_SHARE_DF] += ncpus; in compute_sched_load() 247 load_now[TH_BUCKET_SHARE_DF] += 1; in compute_sched_load() 281 …A_UNSCALE(sched_load[TH_BUCKET_SHARE_FG]), SCHED_LOAD_EWMA_UNSCALE(sched_load[TH_BUCKET_SHARE_DF]), in compute_sched_load()
|
| H A D | sched.h | 228 …TH_BUCKET_SHARE_DF, /* Timeshare thread between BASEPRI_DEFAULT and BASEPRI_UT… enumerator
|
| H A D | priority.c | 859 new_bucket = TH_BUCKET_SHARE_DF; in sched_update_thread_bucket() 899 new_bucket = TH_BUCKET_SHARE_DF; in sched_smt_update_thread_bucket()
|
| H A D | sched_clutch.c | 355 bucket = TH_BUCKET_SHARE_DF; in sched_convert_pri_to_bucket() 1607 …case TH_BUCKET_SHARE_DF: os_atomic_inc(&root_clutch->scr_cumulative_run_count[TH_BUCKET_SHARE_DF]… in sched_edge_cluster_cumulative_count_incr() 1622 …case TH_BUCKET_SHARE_DF: os_atomic_dec(&root_clutch->scr_cumulative_run_count[TH_BUCKET_SHARE_DF]… in sched_edge_cluster_cumulative_count_decr()
|
| H A D | thread_group.c | 1304 case TH_BUCKET_SHARE_DF: in sched_bucket_to_perfcontrol_class()
|
| H A D | sched_prim.c | 6709 sched_pri_shifts[TH_BUCKET_SHARE_UT], sched_pri_shifts[TH_BUCKET_SHARE_DF], 0); in sched_timeshare_maintenance_continue()
|