Searched refs:SCHED (Results 1 – 15 of 15) sorted by relevance
| /xnu-11215.81.4/osfmk/kern/ |
| H A D | syscall_subr.c | 110 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_continue() 127 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch() 147 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_pri_continue() 164 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch_pri() 306 bool should_yield = SCHED(thread_should_yield)(current_processor(), current_thread()); in thread_switch() 607 if (!SCHED(thread_should_yield)(myprocessor, self)) { in thread_yield_internal()
|
| H A D | sched_prim.c | 147 return os_atomic_load(&SCHED(rt_runq)(pset)->count, relaxed); in rt_runq_count() 153 return os_atomic_load_wide(&SCHED(rt_runq)(pset)->earliest_deadline, relaxed); in rt_runq_earliest_deadline() 160 rt_queue_t rt_run_queue = SCHED(rt_runq)(pset); in rt_runq_priority() 244 return os_atomic_load(&SCHED(rt_runq)(pset)->constraint, relaxed) <= rt_constraint_threshold; in rt_runq_is_low_latency() 566 kprintf("Scheduler: Default of %s\n", SCHED(sched_name)); in sched_init() 583 strlcpy(sched_string, SCHED(sched_name), sizeof(sched_string)); in sched_init() 589 SCHED(init)(); in sched_init() 590 SCHED(rt_init)(&pset0); in sched_init() 593 SCHED(pset_init)(&pset0); in sched_init() 594 SCHED(processor_init)(master_processor); in sched_init() [all …]
|
| H A D | priority.c | 201 if (SCHED(can_update_priority)(thread)) { in thread_quantum_expire() 202 SCHED(update_priority)(thread); in thread_quantum_expire() 204 SCHED(lightweight_update_priority)(thread); in thread_quantum_expire() 208 SCHED(quantum_expire)(thread); in thread_quantum_expire() 327 SCHED(update_thread_bucket)(thread); in sched_set_thread_base_priority() 392 priority = (int16_t)SCHED(compute_timeshare_priority)(thread); in thread_recompute_sched_pri() 969 SCHED(update_thread_bucket)(thread); in sched_set_thread_mode()
|
| H A D | sched_prim.h | 841 #define SCHED(f) (sched_edge_dispatch.f) macro 844 #define SCHED(f) (sched_amp_dispatch.f) macro 851 #define SCHED(f) (sched_clutch_dispatch.f) macro 854 #define SCHED(f) (sched_dualq_dispatch.f) macro
|
| H A D | processor.c | 309 SCHED(processor_init)(processor); in processor_init() 578 if (SCHED(multiple_psets_enabled) == FALSE) { in pset_create() 696 SCHED(pset_init)(pset); in pset_init() 697 SCHED(rt_init)(pset); in pset_init()
|
| H A D | host.c | 259 quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); in host_info() 1052 out->ps_runq_count_sum = SCHED(processor_runq_stats_count_sum)(processor); in get_sched_statistics() 1067 out->ps_runq_count_sum = SCHED(rt_runq_count_sum)(); in get_sched_statistics()
|
| H A D | startup.c | 960 SCHED(run_count_incr)(thread); in load_context()
|
| H A D | thread.c | 1513 new_thread->sched_mode = SCHED(initial_thread_sched_mode)(parent_task); in thread_create_internal() 1986 if (SCHED(can_update_priority)(thread)) { in retrieve_thread_basic_info() 1987 SCHED(update_priority)(thread); in retrieve_thread_basic_info() 2167 quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); in thread_info_internal() 3653 return (uint64_t) (SCHED(initial_quantum_size)(thread) / 2); in thread_workq_quantum_size()
|
| H A D | thread_group.c | 1330 SCHED(thread_group_recommendation_change)(tg, new_recommendation); in sched_perfcontrol_thread_group_recommend()
|
| H A D | sched_clutch.c | 1240 if (SCHED(priority_is_urgent)(thread->sched_pri)) { in sched_clutch_root_urgency_inc() 1260 if (SCHED(priority_is_urgent)(thread->sched_pri)) { in sched_clutch_root_urgency_dec()
|
| H A D | thread_policy.c | 1266 sched_mode_t newmode = SCHED(initial_thread_sched_mode)(get_threadtask(thread)); in thread_policy_reset()
|
| H A D | task.c | 5616 quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); in task_info()
|
| /xnu-11215.81.4/tests/sched/sched_test_harness/shadow_headers/ |
| H A D | sched_prim.c | 139 if (SCHED(priority_is_urgent)(rq->highq)) { in run_queue_dequeue() 182 if (SCHED(priority_is_urgent)(thread->sched_pri)) { in run_queue_enqueue() 208 if (SCHED(priority_is_urgent)(thread->sched_pri)) { in run_queue_remove()
|
| /xnu-11215.81.4/tests/sched/sched_test_harness/shadow_headers/kern/ |
| H A D | sched_prim.h | 93 #define SCHED(f) (sched_clutch_dispatch.f) macro
|
| /xnu-11215.81.4/doc/scheduler/ |
| H A D | sched_clutch_edge.md | 248 …SCHED(steal_thread)` scheduler callout is invoked when the processor does not find any thread for … 262 …turn THREAD_NULL for the steal callout and perform rebalancing as part of SCHED(processor_balance)… 269 If `SCHED(steal_thread)` did not return a thread for the processor, it indicates that the processor…
|