| /xnu-12377.1.9/osfmk/kern/ |
| H A D | kpc_thread.c | 107 if (!current_thread()->kpc_buf) { in kpc_set_thread_counting() 108 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_set_thread_counting() 109 act_set_kperf(current_thread()); in kpc_set_thread_counting() 141 if (!current_thread()->kpc_buf) { in kpc_update_thread_counters() 142 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_update_thread_counters() 143 act_set_kperf(current_thread()); in kpc_update_thread_counters() 156 thread_t thread = current_thread(); in kpc_get_curthread_counters() 172 kpc_update_thread_counters( current_thread()); in kpc_get_curthread_counters()
|
| H A D | kern_apfs_reflock.c | 161 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_get_ref() 193 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref() 205 … panic("kern_apfs_reflock: too many refs for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref() 229 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_put_ref() 246 … panic("kern_apfs_reflock: over releasing reflock %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref() 268 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref() 308 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_lock() 323 …nic("Trying to lock a reflock owned by the same thread %p, reflock %p", current_thread(), reflock); in kern_apfs_reflock_try_lock() 334 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_lock() 366 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_wait_for_unlock() [all …]
|
| H A D | syscall_subr.c | 112 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_continue() 129 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch() 145 thread_depress_abort(current_thread()); in swtch_pri_continue() 149 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_pri_continue() 166 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch_pri() 181 thread_t self = current_thread(); in thread_switch_continue() 204 thread_t self = current_thread(); in thread_switch() 312 bool should_yield = SCHED(thread_should_yield)(current_processor(), current_thread()); in thread_switch() 358 thread_t self = current_thread(); in thread_handoff_internal() 443 thread_t self = current_thread(); in thread_depress_abstime() [all …]
|
| H A D | thread_test_context.c | 93 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests() 105 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests() 112 if (current_thread()->th_test_ctx != &ctx) { in thread_test_context_tests() 124 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests() 199 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
|
| H A D | thread_test_context.h | 103 return current_thread()->th_test_ctx; in thread_get_test_context() 152 thread_t thread = current_thread(); in thread_set_test_context() 195 thread_test_context_t *thread_ctx = current_thread()->th_test_ctx; in thread_cleanup_test_context() 199 current_thread()->th_test_ctx = NULL; in thread_cleanup_test_context()
|
| H A D | thread_act.c | 283 if (thread != current_thread() && result == KERN_SUCCESS) { in thread_terminate_internal() 307 if (task == kernel_task && thread != current_thread()) { in thread_terminate() 359 assert(thread == current_thread()); in thread_terminate_immovable() 437 if (thread != current_thread() && result == KERN_SUCCESS) { in thread_suspend() 635 if (thread != current_thread()) { in thread_get_state_internal() 740 if (thread != current_thread()) { in thread_set_state_internal() 833 from_thread = current_thread(); in thread_convert_thread_state() 835 to_thread = current_thread(); in thread_convert_thread_state() 884 if (thread != current_thread()) { in thread_state_initialize() 902 thread_t self = current_thread(); in thread_dup() [all …]
|
| H A D | lock_ticket.c | 137 thread_t self = current_thread(); in __lck_ticket_owned_panic() 146 thread_t self = current_thread(); in __lck_ticket_not_owned_panic() 171 if ((l)->lck_ticket_owner != current_thread()->ctid) { \ 205 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_init_locked() 634 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock() 674 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_to() 698 thread_t cthread = current_thread(); in lck_ticket_lock() 707 thread_t cthread = current_thread(); in lck_ticket_lock_nopreempt() 722 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_try_internal() 757 thread_t cthread = current_thread(); in lck_ticket_lock_try_internal() [all …]
|
| H A D | locks.c | 238 irq_time = current_thread()->machine.int_time_mt; in hw_spin_should_keep_spinning() 695 thread_t thread = current_thread(); 709 thread_t thread = current_thread(); 726 thread_t thread = current_thread(); 742 thread_t thread = current_thread(); 768 thread_t thread = current_thread(); 781 thread_t thread = current_thread(); 792 current_thread(), lock); in __hw_lock_unlock_unowned_panic() 805 if (HW_LOCK_STATE_TO_THREAD(lock->lock_data) != current_thread() && in hw_lock_unlock_internal() 837 thread = current_thread(); in hw_lock_assert() [all …]
|
| H A D | thread.c | 432 thread_t thread = current_thread(); in thread_corpse_continue() 462 thread_t thread = current_thread(); in thread_terminate_self() 625 task_deliver_crash_notification(task, current_thread(), etype, subcode); in thread_terminate_self() 1186 assert(th_remove != current_thread()); in thread_terminate_crashed_threads() 1317 if (__improbable(current_thread() == &init_thread)) { in thread_create_internal() 2312 if (host_priv == NULL || thread != current_thread()) { in thread_wire_internal() 2353 return current_thread() == pgo_iothread_external_state.pgo_iothread; in is_external_pageout_thread() 2359 return current_thread()->options & TH_OPT_VMPRIV ? TRUE : FALSE; in is_vm_privileged() 2367 if (current_thread()->options & TH_OPT_VMPRIV) { in set_vm_privilege() 2374 current_thread()->options |= TH_OPT_VMPRIV; in set_vm_privilege() [all …]
|
| H A D | exclaves.c | 885 thread_t thread = current_thread(); in exclaves_allocate_ipc_buffer() 913 thread_t thread = current_thread(); in exclaves_free_ipc_buffer() 943 assert(thread == current_thread()); in exclaves_thread_terminate() 967 thread_t thread = current_thread(); in exclaves_get_ipc_buffer() 1021 thread_t thread = current_thread(); in exclaves_boot_early() 1147 arm_sme_saved_state_t *sme_state = machine_thread_get_sme_state(current_thread()); in exclaves_save_matrix_state() 1161 arm_sme_saved_state_t *sme_state = machine_thread_get_sme_state(current_thread()); in exclaves_restore_matrix_state() 1190 thread_t thread = current_thread(); in exclaves_enter() 1586 while ((os_atomic_load(¤t_thread()->th_exclaves_inspection_state, in exclaves_scheduler_request() 1663 __assert_only ctid_t ctid = thread_get_ctid(current_thread()); in handle_response_yield() [all …]
|
| H A D | lock_rw.c | 898 __assert_only thread_t self = current_thread(); 975 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockWrite); 1051 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockWrite); 1150 thread_t thread = current_thread(); in lck_rw_lock_exclusive_check_contended() 1186 thread_t thread = current_thread(); 1295 __assert_only thread_t self = current_thread(); 1370 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockRead); 1424 thread_t thread = current_thread(); 1573 thread_t thread = current_thread(); in lck_rw_lock_shared_to_exclusive_failure() 1673 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockUpgrade); in lck_rw_lock_shared_to_exclusive_success() [all …]
|
| /xnu-12377.1.9/osfmk/tests/ |
| H A D | kernel_tests.c | 925 clear_wait(current_thread(), THREAD_AWAKENED); in wait_threads() 948 thread_t thread = current_thread(); in thread_lock_unlock_kernel_primitive() 1118 if (info->threads[i] != current_thread()) { in wait_for_waiters() 1157 if (os_atomic_load(&info->threads[i], acquire) == current_thread()) { in exclude_current_waiter() 1491 uint my_pri = current_thread()->sched_pri; in thread_inheritor_like_mutex() 1493 T_LOG("Started thread pri %d %p", my_pri, current_thread()); in thread_inheritor_like_mutex() 1504 info->thread_inheritor = current_thread(); in thread_inheritor_like_mutex() 1516 …T_ASSERT(info->thread_inheritor == current_thread(), "thread_inheritor is %p", info->thread_inheri… in thread_inheritor_like_mutex() 1524 …T_ASSERT(info->thread_inheritor != current_thread(), "thread_inheritor is %p", info->thread_inheri… in thread_inheritor_like_mutex() 1530 assert(current_thread()->kern_promotion_schedpri == 0); in thread_inheritor_like_mutex() [all …]
|
| H A D | pmap_tests.c | 161 &args, thread_kern_get_pri(current_thread()), &disconnect_thread); in test_pmap_enter_disconnect() 223 &args, thread_kern_get_pri(current_thread()), &remove_thread); in test_pmap_compress_remove() 295 clear_wait(current_thread(), THREAD_AWAKENED); in pmap_nest_thread() 542 pmap_t pmap = current_thread()->map->pmap; in test_pmap_call_overhead() 789 &args, thread_kern_get_pri(current_thread()), &remove_enter_thread); in test_pmap_huge_pv_list() 797 thread_kern_get_pri(current_thread()), &fast_fault_thread); in test_pmap_huge_pv_list() 805 thread_kern_get_pri(current_thread()), &cache_attr_thread); in test_pmap_huge_pv_list() 1017 thread_kern_get_pri(current_thread()) - 1, in test_surt() 1024 clear_wait(current_thread(), THREAD_AWAKENED); in test_surt() 1042 thread_kern_get_pri(current_thread()) - 1, in test_surt() [all …]
|
| /xnu-12377.1.9/osfmk/arm/ |
| H A D | preemption_disable.c | 84 if (current_thread()->machine.CpuDatap->cpu_pending_ast & AST_URGENT) { in kernel_preempt_check() 120 thread_t thread = current_thread(); in _disable_preemption() 153 thread_t thread = current_thread(); in _disable_preemption_without_measurements() 191 thread_t thread = current_thread(); in _enable_preemption() 230 return get_preemption_level_for_thread(current_thread()); in get_preemption_level() 328 thread_t thread = current_thread(); in _prepare_preemption_disable_measurement() 399 _enable_preemption_write_count(current_thread(), 0); in _collect_preemption_disable_measurement()
|
| H A D | machdep_call.c | 65 return machine_thread_set_tsd_base(current_thread(), self); in thread_set_cthread_self() 74 assert( self == current_thread()->machine.cthread_self); in thread_get_cthread_self()
|
| /xnu-12377.1.9/osfmk/i386/ |
| H A D | locks_i386_opt.c | 151 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock() 200 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock() 261 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock_spin_always() 341 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock_spin_always() 429 if (__improbable(lock->lck_mtx_owner != current_thread()->ctid)) { in lck_mtx_unlock()
|
| H A D | locks_i386.c | 422 thread = current_thread(); in lck_spin_assert() 486 lowner, current_thread(), lsti->owner_cpu, in usimple_lock_acquire_timeout_panic() 816 l->debug.lock_thread == (void *) current_thread()) { in usld_lock_pre() 859 l->debug.lock_thread = (void *)current_thread(); in usld_lock_post() 894 if (l->debug.lock_thread != (void *) current_thread()) { in usld_unlock() 964 l->debug.lock_thread = (void *) current_thread(); in usld_lock_try_post() 1115 thread = current_thread(); in lck_mtx_unlock_slow() 1201 assert(current_thread()->turnstile != NULL); in lck_mtx_unlock_wakeup_tail() 1236 turnstile_update_inheritor(ts, current_thread(), in lck_mtx_lock_acquire_inline() 1245 assert(current_thread()->turnstile != NULL); in lck_mtx_lock_acquire_inline() [all …]
|
| H A D | bsd_i386_native.c | 133 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self() 149 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self64() 199 pcb = THREAD_TO_PCB(current_thread()); in thread_set_user_ldt()
|
| /xnu-12377.1.9/osfmk/arm64/ |
| H A D | copyio.c | 304 vm_map_t map = current_thread()->map; in copyin() 336 vm_map_t map = current_thread()->map; in copyin_atomic32() 353 vm_map_t map = current_thread()->map; in copyin_atomic32_wait_if_equals() 369 vm_map_t map = current_thread()->map; in copyin_atomic64() 385 vm_map_t map = current_thread()->map; in copyout_atomic32() 401 vm_map_t map = current_thread()->map; in copyout_atomic64() 417 vm_map_t map = current_thread()->map; in copyinstr() 445 vm_map_t map = current_thread()->map; in copyout() 472 vm_map_t map = current_thread()->map; in copyoutstr_prevalidate()
|
| H A D | bsd_arm64.c | 150 if (thread == current_thread()) { in thread_setsinglestep() 196 struct uthread *ut = get_bsdthread_info(current_thread()); in mach_syscall() 209 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall() 286 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall() 298 prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name; in mach_syscall()
|
| /xnu-12377.1.9/bsd/kern/ |
| H A D | tty_tty.c | 99 context.vc_thread = current_thread(); in cttyopen() 170 context.vc_thread = current_thread(); in cttyread() 190 context.vc_thread = current_thread(); in cttywrite() 225 context.vc_thread = current_thread(); in cttyioctl() 241 context.vc_thread = current_thread(); in cttyselect()
|
| H A D | sys_work_interval.c | 81 kret = kern_work_interval_create(current_thread(), &create_args); in work_interval_ctl() 105 kern_work_interval_destroy(current_thread(), create_args.wica_id); in work_interval_ctl() 210 kret = kern_work_interval_destroy(current_thread(), uap->work_interval_id); in work_interval_ctl() 245 kret = kern_work_interval_notify(current_thread(), &kwi_args); in work_interval_ctl() 260 kret = kern_work_interval_join(current_thread(), in work_interval_ctl()
|
| /xnu-12377.1.9/bsd/dev/dtrace/ |
| H A D | profile_prvd.c | 156 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_fire() 162 pal_register_cache_state(current_thread(), VALID); in profile_fire() 164 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_fire() 182 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_fire() 194 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_fire() 215 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_tick() 221 pal_register_cache_state(current_thread(), VALID); in profile_tick() 223 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_tick() 241 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_tick() 250 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_tick()
|
| /xnu-12377.1.9/osfmk/vm/ |
| H A D | vm_dyld_pager.c | 169 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in _delta_ptr_within_page_nospec() 208 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in _delta_ptr_within_page32_nospec() 391 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64() 409 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64() 456 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageChain32() 473 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageChain32() 537 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage32() 593 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in signPointer() 627 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64() 651 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64() [all …]
|
| /xnu-12377.1.9/bsd/vm/ |
| H A D | vm_unix.c | 1011 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1042 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1054 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1070 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1086 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1101 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np() 1125 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_copyin() 1176 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup() 1214 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup() 1230 (void *)VM_KERNEL_ADDRPERM(current_thread()))); in shared_region_map_and_slide_setup() [all …]
|