Searched refs:SCHED (Results 1 – 17 of 17) sorted by relevance
| /xnu-8020.101.4/osfmk/kern/ |
| H A D | syscall_subr.c | 110 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_continue() 127 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch() 147 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_pri_continue() 164 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch_pri() 306 bool should_yield = SCHED(thread_should_yield)(current_processor(), current_thread()); in thread_switch() 606 if (!SCHED(thread_should_yield)(myprocessor, self)) { in thread_yield_internal()
|
| H A D | sched_prim.c | 160 return os_atomic_load(&SCHED(rt_runq)(pset)->count, relaxed); in rt_runq_count() 166 return os_atomic_load_wide(&SCHED(rt_runq)(pset)->earliest_deadline, relaxed); in rt_runq_earliest_deadline() 173 rt_queue_t rt_run_queue = SCHED(rt_runq)(pset); in rt_runq_priority() 257 return os_atomic_load(&SCHED(rt_runq)(pset)->constraint, relaxed) <= rt_constraint_threshold; in rt_runq_is_low_latency() 433 kprintf("Scheduler: Default of %s\n", SCHED(sched_name)); in sched_init() 450 strlcpy(sched_string, SCHED(sched_name), sizeof(sched_string)); in sched_init() 454 SCHED(init)(); in sched_init() 455 SCHED(rt_init)(&pset0); in sched_init() 458 SCHED(pset_init)(&pset0); in sched_init() 459 SCHED(processor_init)(master_processor); in sched_init() [all …]
|
| H A D | priority.c | 182 if (SCHED(can_update_priority)(thread)) { in thread_quantum_expire() 183 SCHED(update_priority)(thread); in thread_quantum_expire() 185 SCHED(lightweight_update_priority)(thread); in thread_quantum_expire() 189 SCHED(quantum_expire)(thread); in thread_quantum_expire() 320 SCHED(update_thread_bucket)(thread); in sched_set_thread_base_priority() 385 priority = (int16_t)SCHED(compute_timeshare_priority)(thread); in thread_recompute_sched_pri() 956 SCHED(update_thread_bucket)(thread); in sched_set_thread_mode()
|
| H A D | sched_multiq.c | 403 if (!SCHED(sched_groups_enabled)) { in sched_group_create() 427 if (!SCHED(sched_groups_enabled)) { in sched_group_destroy() 636 if (SCHED(priority_is_urgent)(rq->highq)) { in entry_queue_dequeue_entry() 679 if (SCHED(priority_is_urgent)(sched_pri)) { in entry_queue_enqueue_entry() 710 if (SCHED(priority_is_urgent)(sched_pri)) { in entry_queue_remove_entry() 769 if (SCHED(priority_is_urgent)(rq->highq)) { in group_run_queue_dequeue_thread() 815 if (SCHED(priority_is_urgent)(thread_pri)) { in group_run_queue_enqueue_thread() 844 if (SCHED(priority_is_urgent)(thread_pri)) { in group_run_queue_remove_thread()
|
| H A D | sched_prim.h | 691 #define SCHED(f) (sched_edge_dispatch.f) macro 694 #define SCHED(f) (sched_amp_dispatch.f) macro 701 #define SCHED(f) (sched_clutch_dispatch.f) macro 704 #define SCHED(f) (sched_dualq_dispatch.f) macro
|
| H A D | sched_traditional.c | 369 if (SCHED(priority_is_urgent)(pri)) { in sched_traditional_choose_thread_from_runq() 552 if (SCHED(priority_is_urgent)(pri)) { in sched_traditional_processor_queue_shutdown() 675 if (SCHED(priority_is_urgent)(pri)) { in sched_traditional_steal_processor_thread()
|
| H A D | machine.c | 150 SCHED(pset_made_schedulable)(processor, pset, false); in processor_up() 337 SCHED(processor_queue_shutdown)(processor); in processor_doshutdown() 339 SCHED(rt_queue_shutdown)(processor); in processor_doshutdown()
|
| H A D | processor.c | 254 SCHED(processor_init)(processor); in processor_init() 500 if (SCHED(multiple_psets_enabled) == FALSE) { in pset_create() 605 SCHED(pset_init)(pset); in pset_init() 606 SCHED(rt_init)(pset); in pset_init()
|
| H A D | host.c | 257 quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); in host_info() 1017 out->ps_runq_count_sum = SCHED(processor_runq_stats_count_sum)(processor); in get_sched_statistics() 1032 out->ps_runq_count_sum = SCHED(rt_runq_count_sum)(); in get_sched_statistics()
|
| H A D | thread.c | 1405 new_thread->sched_mode = SCHED(initial_thread_sched_mode)(parent_task); in thread_create_internal() 1842 if (SCHED(can_update_priority)(thread)) { in retrieve_thread_basic_info() 1843 SCHED(update_priority)(thread); in retrieve_thread_basic_info() 2023 quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); in thread_info_internal() 3465 return (uint64_t) (SCHED(initial_quantum_size)(thread) / 2); in thread_workq_quantum_size()
|
| H A D | startup.c | 947 SCHED(run_count_incr)(thread); in load_context()
|
| H A D | thread_group.c | 1366 SCHED(thread_group_recommendation_change)(tg, new_recommendation); in sched_perfcontrol_thread_group_recommend()
|
| H A D | sched_clutch.c | 947 if (SCHED(priority_is_urgent)(thread->sched_pri)) { in sched_clutch_root_urgency_inc() 967 if (SCHED(priority_is_urgent)(thread->sched_pri)) { in sched_clutch_root_urgency_dec()
|
| H A D | thread_policy.c | 1214 sched_mode_t newmode = SCHED(initial_thread_sched_mode)(get_threadtask(thread)); in thread_policy_reset()
|
| H A D | task.c | 5235 quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); in task_info()
|
| /xnu-8020.101.4/doc/ |
| H A D | sched_clutch_edge.md | 249 …SCHED(steal_thread)` scheduler callout is invoked when the processor does not find any thread for … 263 …turn THREAD_NULL for the steal callout and perform rebalancing as part of SCHED(processor_balance)… 270 If `SCHED(steal_thread)` did not return a thread for the processor, it indicates that the processor…
|
| /xnu-8020.101.4/osfmk/arm64/ |
| H A D | machine_routines.c | 645 if (SCHED(cpu_init_completed) != NULL) { in ml_cpu_init_completed() 646 SCHED(cpu_init_completed)(); in ml_cpu_init_completed()
|