Home
last modified time | relevance | path

Searched refs:rt_runq_count (Results 1 – 7 of 7) sorted by relevance

/xnu-12377.81.4/osfmk/kern/
H A Dsched_rt.c367 if (rt_runq_count(pset) < lowest_count) { in sched_rtlocal_choose_processor_smt()
371 lowest_count = rt_runq_count(pset); in sched_rtlocal_choose_processor_smt()
621 if (rt_runq_count(pset) > 0) { in sched_rt_choose_thread()
656 while (rt_runq_count(pset) > 0) { in sched_rt_queue_shutdown()
796 return rt_runq_count(pset) > bit_count(avail_map); in rt_pset_has_stealable_threads()
877 int rt_rq_count = rt_runq_count(pset); in rt_pset_needs_a_followup_IPI()
1012 if (bit_count(non_realtime_map) > rt_runq_count(pset)) { in pset_choose_processor_for_realtime_thread_smt()
1025 if (bit_count(non_realtime_map) > rt_runq_count(pset)) { in pset_choose_processor_for_realtime_thread_smt()
1061 if (bit_count(non_realtime_map) > rt_runq_count(pset)) { in pset_choose_processor_for_realtime_thread()
1320 rt_runq_count(processor_set_t pset) in rt_runq_count() function
H A Dsched_amp_common.c93 …int load = ((bit_count(pset->cpu_state_map[PROCESSOR_RUNNING]) + non_rt_load + rt_runq_count(pset)… in sched_amp_update_pset_load_average()
99 …(bit_count(pset->cpu_state_map[PROCESSOR_RUNNING]) + pset->pset_runq.count + rt_runq_count(pset))); in sched_amp_update_pset_load_average()
H A Dsched_rt.h173 int rt_runq_count(processor_set_t);
H A Dsched_prim.c682 return !SCHED(processor_queue_empty)(processor) || rt_runq_count(processor->processor_set) > 0; in sched_thread_should_yield()
2219 …return (pset->pset_runq.count > 0) && ((pset->pset_runq.count + rt_runq_count(pset)) > bit_count(a… in pset_has_stealable_threads()
2437 } else if (rt_runq_count(pset)) { in thread_select()
2567 …} else if ((rt_runq_count(pset) > 0) && (rt_deadline_add(rt_runq_earliest_deadline(pset), rt_deadl… in thread_select()
2613 if ((rt_runq_count(pset) == 0) && in thread_select()
2727 …if (sched_ok_to_run_realtime_thread(pset, processor, true) && (spill_pending || rt_runq_count(pset… in thread_select()
2838 …cessor) || (sched_ok_to_run_realtime_thread(pset, processor, true) && (rt_runq_count(pset) > 0))) { in thread_select()
4187 int existing_backups = bit_count(pset->pending_AST_URGENT_cpu_mask) - rt_runq_count(pset); in realtime_setrun()
5711 if (rt_runq_count(pset) > 0) { in csw_check_locked()
6391 if (rt_runq_count(pset)) { in processor_idle()
[all …]
H A Dsched_amp.c312 if (!sched_amp_processor_queue_empty(processor) || (rt_runq_count(processor->processor_set) > 0)) { in sched_amp_thread_should_yield()
H A Dsched_clutch.c4274 …if (!sched_edge_processor_queue_empty(processor) || (rt_runq_count(processor->processor_set) > 0))… in sched_edge_thread_should_yield()
4449 …ge_cluster_cumulative_count(&pset->pset_clutch_root, sched_bucket) + rt_runq_count(pset) + runnin… in sched_edge_update_pset_load_average()
/xnu-12377.81.4/tests/sched/sched_test_harness/
H A Dsched_edge_harness.c271 if (rt_runq_count(pset) > 0) { in impl_cpu_dequeue_thread_compare_current()