| /xnu-11417.101.15/tests/sched/sched_test_harness/ |
| H A D | sched_edge_harness.c | 65 return SCHED(ipi_policy)(dst, thread, (dst->active_thread == NULL), event); in sched_ipi_action() 305 assert(cpus[cpu_id]->active_thread != NULL); in impl_cpu_dequeue_thread_compare_current() 306 return sched_clutch_choose_thread(cpus[cpu_id], MINPRI, cpus[cpu_id]->active_thread, 0); in impl_cpu_dequeue_thread_compare_current() 313 assert(cpus[cpu_id]->active_thread != NULL); in impl_processor_csw_check() 400 assert(cpus[cpu_id]->active_thread != NULL); in impl_thread_should_yield() 401 return sched_edge_thread_should_yield(cpus[cpu_id], cpus[cpu_id]->active_thread); in impl_thread_should_yield()
|
| H A D | sched_clutch_harness.c | 71 assert(cpus[cpu_id]->active_thread != NULL); in impl_cpu_dequeue_thread_compare_current() 72 return sched_clutch_choose_thread(cpus[cpu_id], MINPRI, cpus[cpu_id]->active_thread, 0); in impl_cpu_dequeue_thread_compare_current()
|
| H A D | sched_clutch_harness_impl.c | 125 cpus[c]->active_thread = NULL; in clutch_impl_init_topology() 222 cpus[cpu_id]->active_thread = thread; in clutch_impl_cpu_set_thread_current() 233 cpus[cpu_id]->active_thread = NULL; in clutch_impl_cpu_clear_thread_current()
|
| /xnu-11417.101.15/osfmk/kdp/ml/arm/ |
| H A D | kdp_vm.c | 333 if ((cpudatap == NULL) || (processor->active_thread == NULL)) { in kern_collectth_state() 339 void *kpcb = processor->active_thread->machine.kpcb; in kern_collectth_state() 350 vm_offset_t kstackptr = (vm_offset_t) processor->active_thread->machine.kstackptr; in kern_collectth_state() 361 vm_offset_t kstackptr = (vm_offset_t) processor->active_thread->machine.kstackptr; in kern_collectth_state()
|
| /xnu-11417.101.15/tools/lldbmacros/ |
| H A D | process.py | 1692 active_thread = current_processor.active_thread 1693 if unsigned(active_thread) != 0: 1694 task_val = active_thread.t_tro.tro_task 1699 print(GetThreadSummary(active_thread, O=O)) 1701 print(GetThreadBackTrace(active_thread, prefix="\t")) 1713 active_thread = current_processor.active_thread 1714 if unsigned(active_thread) != 0 : 1715 task_val = active_thread.t_tro.tro_task 1720 print(GetThreadSummary(active_thread, O=O)) 1766 active_thread = processor.active_thread [all …]
|
| H A D | scheduler.py | 357 active_thread = current_processor.active_thread 360 if unsigned(active_thread) != 0 : 361 task_val = active_thread.t_tro.tro_task 362 proc_val = active_thread.t_tro.tro_proc 364 thread_id = active_thread.thread_id 379 …time_since_dispatch_us, time_since_debugger_us, t=active_thread, thread_id=thread_id, proc_name=pr… 382 …: 0x{t:<16x} 0x{thread_id:<8x} {proc_name:s}".format(last_dispatch, t=active_thread, thread_id=thr… 624 if (processor.active_thread != 0) : 626 print("\t" + GetThreadSummary(processor.active_thread))
|
| H A D | recount.py | 485 active_thread = unsigned(pr.active_thread) 486 if active_thread != 0: 488 thread_plan, [active_thread], indent=True, O=O)
|
| /xnu-11417.101.15/osfmk/kern/ |
| H A D | thread_act.c | 1130 processor->active_thread == thread) { in thread_set_apc_ast_locked() 1382 processor->active_thread == thread) { in act_set_ast() 1491 processor->active_thread == thread) { in act_set_ast_reset_pcs()
|
| H A D | sched_dualq.c | 328 assert(processor->active_thread != NULL); in sched_dualq_processor_csw_check()
|
| H A D | sched_clutch.c | 2869 highest_thread = processor->active_thread; in sched_clutch_hierarchy_thread_highest() 2894 … == SCHED_CLUTCH_TRAVERSE_REMOVE_CONSIDER_CURRENT && highest_thread != processor->active_thread)) { in sched_clutch_hierarchy_thread_highest() 2895 …mode != SCHED_CLUTCH_TRAVERSE_REMOVE_HIERARCHY_ONLY || highest_thread != processor->active_thread); in sched_clutch_hierarchy_thread_highest() 3275 thread_t prev_thread = has_prev_thread ? processor->active_thread : NULL; in sched_clutch_processor_highest_thread() 3325 assert(prev_thread == NULL || prev_thread == processor->active_thread); in sched_clutch_choose_thread() 3358 assert(processor->active_thread != NULL); in sched_clutch_processor_csw_check() 3360 if (runqueue_thread != processor->active_thread) { in sched_clutch_processor_csw_check()
|
| H A D | startup.c | 1025 processor->active_thread = thread; in load_context()
|
| H A D | processor.h | 406 struct thread *active_thread; /* thread running on processor */ member
|
| H A D | recount.c | 1072 thread_t thread = processor->active_thread; in recount_transition() 1130 thread_t thread = processor->active_thread; in recount_leave_interrupt()
|
| H A D | sched_amp.c | 325 assert(processor->active_thread != NULL); in sched_amp_processor_csw_check()
|
| H A D | sfi.c | 1102 processor->active_thread == thread) { in sfi_reevaluate()
|
| H A D | sched_prim.c | 3335 processor->active_thread = thread; in thread_invoke() 3481 processor->active_thread = thread; in thread_invoke() 5262 …if ((preempt = csw_check_locked(processor->active_thread, processor, pset, AST_NONE)) != AST_NONE)… in processor_setrun() 6477 assert(thread == processor->active_thread); in csw_check() 6676 thread_t thread = processor->active_thread; in ast_check() 6889 processor->active_thread == thread) { in set_sched_pri() 7891 last_processor->active_thread == thread) { in thread_set_eager_preempt()
|
| H A D | machine.c | 501 assert(processor->active_thread == current_thread()); in processor_offline_intstack()
|
| H A D | processor.c | 315 processor->active_thread = processor->startup_thread = processor->idle_thread = THREAD_NULL; in processor_init()
|
| H A D | kern_stackshot.c | 4589 last_processor->active_thread == thread); in classify_thread()
|
| H A D | task.c | 10598 processor->active_thread == thread) { in task_set_ast_mach_exception()
|
| /xnu-11417.101.15/osfmk/kperf/ |
| H A D | thread_samplers.c | 70 if ((last_processor != PROCESSOR_NULL) && (thread == last_processor->active_thread)) { in kperf_thread_info_runmode_legacy()
|
| /xnu-11417.101.15/osfmk/i386/ |
| H A D | i386_init.c | 1103 if (current_processor()->active_thread != THREAD_NULL) { in do_init_slave()
|