Searched refs:earliest_deadline (Results 1 – 3 of 3) sorted by relevance
148 return os_atomic_load_wide(&SCHED(rt_runq)(pset)->earliest_deadline, relaxed); in rt_runq_earliest_deadline()176 uint64_t earliest_deadline = RT_DEADLINE_NONE; in check_rt_runq_consistency() local214 if (rt_runq->pri_earliest_deadline < earliest_deadline) { in check_rt_runq_consistency()215 earliest_deadline = rt_runq->pri_earliest_deadline; in check_rt_runq_consistency()221 assert(os_atomic_load_wide(&rt_run_queue->earliest_deadline, relaxed) == earliest_deadline); in check_rt_runq_consistency()568 os_atomic_init(&pset->rt_runq.earliest_deadline, RT_DEADLINE_NONE); in pset_rt_init()2204 …er_psets_have_earlier_rt_threads_pending(processor_set_t stealing_pset, uint64_t earliest_deadline) in other_psets_have_earlier_rt_threads_pending() argument2213 …dline_add(nset->stealable_rt_threads_earliest_deadline, rt_deadline_epsilon) < earliest_deadline) { in other_psets_have_earlier_rt_threads_pending()2229 uint64_t earliest_deadline = rt_runq_earliest_deadline(starting_pset); in choose_next_rt_processor_for_IPI() local2234 if (earliest_deadline < rt_constraint_ll + ctime) { in choose_next_rt_processor_for_IPI()[all …]
259 _Atomic uint64_t earliest_deadline; /* earliest deadline */ member
104 …ad_t sched_rtlocal_steal_thread(processor_set_t stealing_pset, uint64_t earliest_deadline);861 thread_t (*rt_steal_thread)(processor_set_t pset, uint64_t earliest_deadline);