Searched refs:TH_BUCKET_RUN (Results 1 – 6 of 6) sorted by relevance
| /xnu-12377.81.4/osfmk/kern/ |
| H A D | sched_average.c | 190 load_now[TH_BUCKET_RUN] = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in compute_sched_load() 197 assert(load_now[TH_BUCKET_RUN] >= 0); in compute_sched_load() 200 uint32_t nthreads = load_now[TH_BUCKET_RUN]; in compute_sched_load() 288 uint32_t nthreads = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed) - 1; in compute_averages()
|
| H A D | priority.c | 791 uint32_t new_count = os_atomic_inc(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in sched_run_incr() 805 uint32_t new_count = os_atomic_dec(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in sched_run_decr() 822 uint32_t new_count = os_atomic_add(&sched_run_buckets[TH_BUCKET_RUN], run_weight, relaxed); in sched_smt_run_incr() 838 uint32_t new_count = os_atomic_sub(&sched_run_buckets[TH_BUCKET_RUN], run_weight, relaxed); in sched_smt_run_decr() 847 sched_bucket_t new_bucket = TH_BUCKET_RUN; in sched_update_thread_bucket() 887 sched_bucket_t new_bucket = TH_BUCKET_RUN; in sched_smt_update_thread_bucket()
|
| H A D | sched.h | 231 TH_BUCKET_RUN, /* All runnable threads */ enumerator 232 TH_BUCKET_SCHED_MAX = TH_BUCKET_RUN, /* Maximum schedulable buckets */
|
| H A D | sched_clutch.c | 348 sched_bucket_t bucket = TH_BUCKET_RUN; in sched_convert_pri_to_bucket() 2064 assert(bucket != TH_BUCKET_RUN); in sched_clutch_run_bucket_incr() 2092 assert(bucket != TH_BUCKET_RUN); in sched_clutch_run_bucket_decr() 3112 assert(bucket != TH_BUCKET_RUN); in sched_clutch_thread_pri_shift() 3691 uint32_t new_count = os_atomic_inc(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in sched_clutch_run_incr() 3700 uint32_t new_count = os_atomic_dec(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in sched_clutch_run_decr() 3736 if (old_bucket != TH_BUCKET_RUN) { in sched_clutch_update_thread_bucket()
|
| H A D | sched_prim.c | 886 new_run_count = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in thread_unblock() 965 sched_run_buckets[TH_BUCKET_RUN], 0); in thread_unblock() 3261 sampled_sched_run_count = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in pset_cancel_deferred_dispatch() 3419 sched_run_buckets[TH_BUCKET_RUN], 0); in thread_dispatch() 3583 sched_run_buckets[TH_BUCKET_RUN], 0); in thread_dispatch() 3764 …if ((os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed) == 1) && !(self->state & TH_IDLE))… in thread_dispatch() 5931 if (thread->th_sched_bucket == TH_BUCKET_RUN) { in set_sched_pri()
|
| H A D | thread.c | 227 .th_sched_bucket = TH_BUCKET_RUN,
|