Searched refs:rt_runq_count (Results 1 – 7 of 7) sorted by relevance
367 if (rt_runq_count(pset) < lowest_count) { in sched_rtlocal_choose_processor_smt()371 lowest_count = rt_runq_count(pset); in sched_rtlocal_choose_processor_smt()621 if (rt_runq_count(pset) > 0) { in sched_rt_choose_thread()656 while (rt_runq_count(pset) > 0) { in sched_rt_queue_shutdown()796 return rt_runq_count(pset) > bit_count(avail_map); in rt_pset_has_stealable_threads()877 int rt_rq_count = rt_runq_count(pset); in rt_pset_needs_a_followup_IPI()1012 if (bit_count(non_realtime_map) > rt_runq_count(pset)) { in pset_choose_processor_for_realtime_thread_smt()1025 if (bit_count(non_realtime_map) > rt_runq_count(pset)) { in pset_choose_processor_for_realtime_thread_smt()1061 if (bit_count(non_realtime_map) > rt_runq_count(pset)) { in pset_choose_processor_for_realtime_thread()1320 rt_runq_count(processor_set_t pset) in rt_runq_count() function
93 …int load = ((bit_count(pset->cpu_state_map[PROCESSOR_RUNNING]) + non_rt_load + rt_runq_count(pset)… in sched_amp_update_pset_load_average()99 …(bit_count(pset->cpu_state_map[PROCESSOR_RUNNING]) + pset->pset_runq.count + rt_runq_count(pset))); in sched_amp_update_pset_load_average()
173 int rt_runq_count(processor_set_t);
682 return !SCHED(processor_queue_empty)(processor) || rt_runq_count(processor->processor_set) > 0; in sched_thread_should_yield()2219 …return (pset->pset_runq.count > 0) && ((pset->pset_runq.count + rt_runq_count(pset)) > bit_count(a… in pset_has_stealable_threads()2437 } else if (rt_runq_count(pset)) { in thread_select()2567 …} else if ((rt_runq_count(pset) > 0) && (rt_deadline_add(rt_runq_earliest_deadline(pset), rt_deadl… in thread_select()2613 if ((rt_runq_count(pset) == 0) && in thread_select()2727 …if (sched_ok_to_run_realtime_thread(pset, processor, true) && (spill_pending || rt_runq_count(pset… in thread_select()2838 …cessor) || (sched_ok_to_run_realtime_thread(pset, processor, true) && (rt_runq_count(pset) > 0))) { in thread_select()4187 int existing_backups = bit_count(pset->pending_AST_URGENT_cpu_mask) - rt_runq_count(pset); in realtime_setrun()5711 if (rt_runq_count(pset) > 0) { in csw_check_locked()6391 if (rt_runq_count(pset)) { in processor_idle()[all …]
312 if (!sched_amp_processor_queue_empty(processor) || (rt_runq_count(processor->processor_set) > 0)) { in sched_amp_thread_should_yield()
4274 …if (!sched_edge_processor_queue_empty(processor) || (rt_runq_count(processor->processor_set) > 0))… in sched_edge_thread_should_yield()4449 …ge_cluster_cumulative_count(&pset->pset_clutch_root, sched_bucket) + rt_runq_count(pset) + runnin… in sched_edge_update_pset_load_average()
271 if (rt_runq_count(pset) > 0) { in impl_cpu_dequeue_thread_compare_current()