| /xnu-11215.81.4/osfmk/kern/ |
| H A D | ast.c | 98 thread_preempted_in_kernel(ast_t urgent_reason) in thread_preempted_in_kernel() 143 ast_t urgent_reason = ast_consume(AST_PREEMPTION); in ast_taken_kernel() 177 ast_t urgent_reason = ast_consume(AST_PREEMPTION); in ast_taken_user() 211 ast_t reasons = ast_consume(AST_PER_THREAD | AST_KPERF | AST_DTRACE); in ast_taken_user() 289 ast_t telemetry_reasons = reasons & AST_TELEMETRY_ALL; in ast_taken_user() 326 ast_t preemption_reasons = ast_consume(AST_PREEMPTION); in ast_taken_user() 390 ast_on(ast_t reasons) in ast_on() 392 ast_t *pending_ast = ast_pending(); in ast_on() 402 ast_off(ast_t reasons) in ast_off() 404 ast_t *pending_ast = ast_pending(); in ast_off() [all …]
|
| H A D | ast.h | 109 __options_decl(ast_t, uint32_t, { 160 extern ast_t *ast_pending(void); 163 extern void ast_on(ast_t reasons); 166 extern void ast_off(ast_t reasons); 169 extern ast_t ast_consume(ast_t reasons); 172 extern ast_t ast_peek(ast_t reasons);
|
| H A D | sched.h | 314 extern ast_t csw_check( 317 ast_t check_reason); 322 extern ast_t update_pending_nonurgent_preemption(processor_t processor, ast_t reason);
|
| H A D | sfi.h | 61 ast_t sfi_thread_needs_ast(thread_t thread, sfi_class_id_t *out_class /* optional */); 62 ast_t sfi_processor_needs_ast(processor_t processor);
|
| H A D | sched_dualq.c | 65 static ast_t 90 sched_dualq_choose_thread(processor_t processor, int priority, __unused thread_t prev, ast_t reason… 99 sched_dualq_thread_avoid_processor(processor_t processor, thread_t thread, __unused ast_t reason); 219 __unused ast_t reason) in sched_dualq_choose_thread() 309 static ast_t 552 sched_dualq_thread_avoid_processor(processor_t processor, thread_t thread, __unused ast_t reason) in sched_dualq_thread_avoid_processor()
|
| H A D | sched_amp.c | 66 static ast_t 91 sched_amp_choose_thread(processor_t processor, int priority, __unused thread_t prev, ast_t reason); 103 sched_amp_thread_avoid_processor(processor_t processor, thread_t thread, __unused ast_t reason); 236 __unused ast_t reason) in sched_amp_choose_thread() 316 static ast_t 638 sched_amp_thread_avoid_processor(processor_t processor, thread_t thread, __unused ast_t reason) in sched_amp_thread_avoid_processor()
|
| H A D | sched_prim.h | 264 ast_t reason); 876 ast_t reason); 941 ast_t (*processor_csw_check)(processor_t processor); 997 bool (*thread_avoid_processor)(processor_t processor, thread_t thread, ast_t reason);
|
| H A D | sched_amp_common.c | 145 ast_t preempt = AST_NONE; in pset_signal_spill() 179 ast_t new_preempt = update_pending_nonurgent_preemption(processor, preempt); in pset_signal_spill()
|
| H A D | syscall_subr.c | 635 ast_t *pending_ast = ast_pending(); in thread_yield_to_preemption() 636 ast_t ast = AST_NONE; in thread_yield_to_preemption()
|
| H A D | sfi.c | 873 ast_t 900 ast_t 1044 ast_t sfi_ast; in sfi_reevaluate()
|
| H A D | sched_prim.c | 482 static ast_t 487 ast_t check_reason); 2645 ast_t *reason) in thread_select() 3163 ast_t reason) in thread_invoke() 3802 ast_t reason = thread->reason; in thread_dispatch() 3926 ast_t new_ast; in thread_dispatch() 4034 ast_t reason) in thread_block_reason() 4124 ast_t reason = AST_NONE; in thread_run() 4859 ast_t preempt = AST_NONE; in realtime_setrun() 5110 ast_t preempt = AST_NONE; in processor_setrun() [all …]
|
| H A D | priority.c | 102 ast_t preempt; in thread_quantum_expire() 229 ast_t check_reason = AST_QUANTUM; in thread_quantum_expire()
|
| H A D | thread_act.c | 1367 ast_t ast) in act_set_ast() 1401 ast_t ast) in act_set_ast_async()
|
| H A D | sched_clutch.c | 3049 static ast_t 3081 sched_clutch_choose_thread(processor_t processor, int priority, thread_t prev_thread, ast_t reason); 3323 __unused ast_t reason) in sched_clutch_choose_thread() 3355 static ast_t 3699 sched_edge_thread_avoid_processor(processor_t processor, thread_t thread, ast_t reason); 4406 sched_edge_thread_avoid_processor(processor_t processor, thread_t thread, ast_t reason) in sched_edge_thread_avoid_processor() 5090 ast_t new_preempt = update_pending_nonurgent_preemption(src_processor, AST_PREEMPT); in sched_edge_migrate_thread_group_running_threads()
|
| H A D | thread.h | 469 ast_t reason; /* why we blocked */ 798 os_atomic(ast_t) ast;
|
| H A D | lock_mtx.c | 690 ast_t *const astp = ast_pending(); in lck_mtx_lock_adaptive_spin()
|
| /xnu-11215.81.4/tests/sched/sched_test_harness/shadow_headers/kern/ |
| H A D | sched_prim.h | 113 ast_t reason); 146 ast_t (*processor_csw_check)(processor_t processor); 178 bool (*thread_avoid_processor)(processor_t processor, thread_t thread, ast_t reason);
|
| /xnu-11215.81.4/tests/sched/sched_test_harness/shadow_headers/ |
| H A D | misc_needed_defines.h | 25 typedef uint32_t ast_t; typedef
|
| /xnu-11215.81.4/osfmk/i386/ |
| H A D | cpu.c | 211 ast_t *
|
| H A D | cpu_data.h | 185 ast_t cpu_pending_ast;
|
| /xnu-11215.81.4/osfmk/arm/ |
| H A D | cpu_data_internal.h | 150 ast_t cpu_pending_ast;
|
| H A D | cpu_common.c | 768 ast_t *
|
| /xnu-11215.81.4/tests/sched/sched_test_harness/ |
| H A D | sched_clutch_harness.c | 217 ast_t preempt_ast = sched_clutch_processor_csw_check(&cpu0); in impl_processor_csw_check()
|
| /xnu-11215.81.4/osfmk/arm64/sptm/pmap/ |
| H A D | pmap.h | 695 return !!(*((volatile ast_t*)ast_pending()) & AST_URGENT); in _pmap_pending_preemption_real()
|
| /xnu-11215.81.4/osfmk/arm/pmap/ |
| H A D | pmap.h | 784 return !!(*((volatile ast_t*)ast_pending()) & AST_URGENT); in _pmap_pending_preemption_real()
|