Home
last modified time | relevance | path

Searched refs:current_thread (Results 1 – 25 of 274) sorted by relevance

1234567891011

/xnu-12377.1.9/osfmk/kern/
H A Dkpc_thread.c107 if (!current_thread()->kpc_buf) { in kpc_set_thread_counting()
108 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_set_thread_counting()
109 act_set_kperf(current_thread()); in kpc_set_thread_counting()
141 if (!current_thread()->kpc_buf) { in kpc_update_thread_counters()
142 current_thread()->kperf_ast |= T_KPC_ALLOC; in kpc_update_thread_counters()
143 act_set_kperf(current_thread()); in kpc_update_thread_counters()
156 thread_t thread = current_thread(); in kpc_get_curthread_counters()
172 kpc_update_thread_counters( current_thread()); in kpc_get_curthread_counters()
H A Dkern_apfs_reflock.c161 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_get_ref()
193 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref()
205 … panic("kern_apfs_reflock: too many refs for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_get_ref()
229 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_put_ref()
246 … panic("kern_apfs_reflock: over releasing reflock %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref()
268 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_put_ref()
308 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_try_lock()
323 …nic("Trying to lock a reflock owned by the same thread %p, reflock %p", current_thread(), reflock); in kern_apfs_reflock_try_lock()
334 … panic("kern_apfs_reflock: too many waiters for %p thread %p", reflock, current_thread()); in kern_apfs_reflock_try_lock()
366 ctid_t my_ctid = thread_get_ctid(current_thread()); in kern_apfs_reflock_wait_for_unlock()
[all …]
H A Dsyscall_subr.c112 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_continue()
129 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch()
145 thread_depress_abort(current_thread()); in swtch_pri_continue()
149 result = SCHED(thread_should_yield)(myprocessor, current_thread()); in swtch_pri_continue()
166 if (!SCHED(thread_should_yield)(myprocessor, current_thread())) { in swtch_pri()
181 thread_t self = current_thread(); in thread_switch_continue()
204 thread_t self = current_thread(); in thread_switch()
312 bool should_yield = SCHED(thread_should_yield)(current_processor(), current_thread()); in thread_switch()
358 thread_t self = current_thread(); in thread_handoff_internal()
443 thread_t self = current_thread(); in thread_depress_abstime()
[all …]
H A Dthread_test_context.c93 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
105 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
112 if (current_thread()->th_test_ctx != &ctx) { in thread_test_context_tests()
124 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
199 if (current_thread()->th_test_ctx != NULL) { in thread_test_context_tests()
H A Dthread_test_context.h103 return current_thread()->th_test_ctx; in thread_get_test_context()
152 thread_t thread = current_thread(); in thread_set_test_context()
195 thread_test_context_t *thread_ctx = current_thread()->th_test_ctx; in thread_cleanup_test_context()
199 current_thread()->th_test_ctx = NULL; in thread_cleanup_test_context()
H A Dthread_act.c283 if (thread != current_thread() && result == KERN_SUCCESS) { in thread_terminate_internal()
307 if (task == kernel_task && thread != current_thread()) { in thread_terminate()
359 assert(thread == current_thread()); in thread_terminate_immovable()
437 if (thread != current_thread() && result == KERN_SUCCESS) { in thread_suspend()
635 if (thread != current_thread()) { in thread_get_state_internal()
740 if (thread != current_thread()) { in thread_set_state_internal()
833 from_thread = current_thread(); in thread_convert_thread_state()
835 to_thread = current_thread(); in thread_convert_thread_state()
884 if (thread != current_thread()) { in thread_state_initialize()
902 thread_t self = current_thread(); in thread_dup()
[all …]
H A Dlock_ticket.c137 thread_t self = current_thread(); in __lck_ticket_owned_panic()
146 thread_t self = current_thread(); in __lck_ticket_not_owned_panic()
171 if ((l)->lck_ticket_owner != current_thread()->ctid) { \
205 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_init_locked()
634 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock()
674 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_to()
698 thread_t cthread = current_thread(); in lck_ticket_lock()
707 thread_t cthread = current_thread(); in lck_ticket_lock_nopreempt()
722 lock_disable_preemption_for_thread(current_thread()); in hw_lck_ticket_lock_try_internal()
757 thread_t cthread = current_thread(); in lck_ticket_lock_try_internal()
[all …]
H A Dlocks.c238 irq_time = current_thread()->machine.int_time_mt; in hw_spin_should_keep_spinning()
695 thread_t thread = current_thread();
709 thread_t thread = current_thread();
726 thread_t thread = current_thread();
742 thread_t thread = current_thread();
768 thread_t thread = current_thread();
781 thread_t thread = current_thread();
792 current_thread(), lock); in __hw_lock_unlock_unowned_panic()
805 if (HW_LOCK_STATE_TO_THREAD(lock->lock_data) != current_thread() && in hw_lock_unlock_internal()
837 thread = current_thread(); in hw_lock_assert()
[all …]
H A Dthread.c432 thread_t thread = current_thread(); in thread_corpse_continue()
462 thread_t thread = current_thread(); in thread_terminate_self()
625 task_deliver_crash_notification(task, current_thread(), etype, subcode); in thread_terminate_self()
1186 assert(th_remove != current_thread()); in thread_terminate_crashed_threads()
1317 if (__improbable(current_thread() == &init_thread)) { in thread_create_internal()
2312 if (host_priv == NULL || thread != current_thread()) { in thread_wire_internal()
2353 return current_thread() == pgo_iothread_external_state.pgo_iothread; in is_external_pageout_thread()
2359 return current_thread()->options & TH_OPT_VMPRIV ? TRUE : FALSE; in is_vm_privileged()
2367 if (current_thread()->options & TH_OPT_VMPRIV) { in set_vm_privilege()
2374 current_thread()->options |= TH_OPT_VMPRIV; in set_vm_privilege()
[all …]
H A Dexclaves.c885 thread_t thread = current_thread(); in exclaves_allocate_ipc_buffer()
913 thread_t thread = current_thread(); in exclaves_free_ipc_buffer()
943 assert(thread == current_thread()); in exclaves_thread_terminate()
967 thread_t thread = current_thread(); in exclaves_get_ipc_buffer()
1021 thread_t thread = current_thread(); in exclaves_boot_early()
1147 arm_sme_saved_state_t *sme_state = machine_thread_get_sme_state(current_thread()); in exclaves_save_matrix_state()
1161 arm_sme_saved_state_t *sme_state = machine_thread_get_sme_state(current_thread()); in exclaves_restore_matrix_state()
1190 thread_t thread = current_thread(); in exclaves_enter()
1586 while ((os_atomic_load(&current_thread()->th_exclaves_inspection_state, in exclaves_scheduler_request()
1663 __assert_only ctid_t ctid = thread_get_ctid(current_thread()); in handle_response_yield()
[all …]
H A Dlock_rw.c898 __assert_only thread_t self = current_thread();
975 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockWrite);
1051 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockWrite);
1150 thread_t thread = current_thread(); in lck_rw_lock_exclusive_check_contended()
1186 thread_t thread = current_thread();
1295 __assert_only thread_t self = current_thread();
1370 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockRead);
1424 thread_t thread = current_thread();
1573 thread_t thread = current_thread(); in lck_rw_lock_shared_to_exclusive_failure()
1673 thread_set_pending_block_hint(current_thread(), kThreadWaitKernelRWLockUpgrade); in lck_rw_lock_shared_to_exclusive_success()
[all …]
/xnu-12377.1.9/osfmk/tests/
H A Dkernel_tests.c925 clear_wait(current_thread(), THREAD_AWAKENED); in wait_threads()
948 thread_t thread = current_thread(); in thread_lock_unlock_kernel_primitive()
1118 if (info->threads[i] != current_thread()) { in wait_for_waiters()
1157 if (os_atomic_load(&info->threads[i], acquire) == current_thread()) { in exclude_current_waiter()
1491 uint my_pri = current_thread()->sched_pri; in thread_inheritor_like_mutex()
1493 T_LOG("Started thread pri %d %p", my_pri, current_thread()); in thread_inheritor_like_mutex()
1504 info->thread_inheritor = current_thread(); in thread_inheritor_like_mutex()
1516 …T_ASSERT(info->thread_inheritor == current_thread(), "thread_inheritor is %p", info->thread_inheri… in thread_inheritor_like_mutex()
1524 …T_ASSERT(info->thread_inheritor != current_thread(), "thread_inheritor is %p", info->thread_inheri… in thread_inheritor_like_mutex()
1530 assert(current_thread()->kern_promotion_schedpri == 0); in thread_inheritor_like_mutex()
[all …]
H A Dpmap_tests.c161 &args, thread_kern_get_pri(current_thread()), &disconnect_thread); in test_pmap_enter_disconnect()
223 &args, thread_kern_get_pri(current_thread()), &remove_thread); in test_pmap_compress_remove()
295 clear_wait(current_thread(), THREAD_AWAKENED); in pmap_nest_thread()
542 pmap_t pmap = current_thread()->map->pmap; in test_pmap_call_overhead()
789 &args, thread_kern_get_pri(current_thread()), &remove_enter_thread); in test_pmap_huge_pv_list()
797 thread_kern_get_pri(current_thread()), &fast_fault_thread); in test_pmap_huge_pv_list()
805 thread_kern_get_pri(current_thread()), &cache_attr_thread); in test_pmap_huge_pv_list()
1017 thread_kern_get_pri(current_thread()) - 1, in test_surt()
1024 clear_wait(current_thread(), THREAD_AWAKENED); in test_surt()
1042 thread_kern_get_pri(current_thread()) - 1, in test_surt()
[all …]
/xnu-12377.1.9/osfmk/arm/
H A Dpreemption_disable.c84 if (current_thread()->machine.CpuDatap->cpu_pending_ast & AST_URGENT) { in kernel_preempt_check()
120 thread_t thread = current_thread(); in _disable_preemption()
153 thread_t thread = current_thread(); in _disable_preemption_without_measurements()
191 thread_t thread = current_thread(); in _enable_preemption()
230 return get_preemption_level_for_thread(current_thread()); in get_preemption_level()
328 thread_t thread = current_thread(); in _prepare_preemption_disable_measurement()
399 _enable_preemption_write_count(current_thread(), 0); in _collect_preemption_disable_measurement()
H A Dmachdep_call.c65 return machine_thread_set_tsd_base(current_thread(), self); in thread_set_cthread_self()
74 assert( self == current_thread()->machine.cthread_self); in thread_get_cthread_self()
/xnu-12377.1.9/osfmk/i386/
H A Dlocks_i386_opt.c151 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock()
200 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock()
261 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_lock_spin_always()
341 ordered_store_mtx_owner(lock, current_thread()->ctid); in lck_mtx_try_lock_spin_always()
429 if (__improbable(lock->lck_mtx_owner != current_thread()->ctid)) { in lck_mtx_unlock()
H A Dlocks_i386.c422 thread = current_thread(); in lck_spin_assert()
486 lowner, current_thread(), lsti->owner_cpu, in usimple_lock_acquire_timeout_panic()
816 l->debug.lock_thread == (void *) current_thread()) { in usld_lock_pre()
859 l->debug.lock_thread = (void *)current_thread(); in usld_lock_post()
894 if (l->debug.lock_thread != (void *) current_thread()) { in usld_unlock()
964 l->debug.lock_thread = (void *) current_thread(); in usld_lock_try_post()
1115 thread = current_thread(); in lck_mtx_unlock_slow()
1201 assert(current_thread()->turnstile != NULL); in lck_mtx_unlock_wakeup_tail()
1236 turnstile_update_inheritor(ts, current_thread(), in lck_mtx_lock_acquire_inline()
1245 assert(current_thread()->turnstile != NULL); in lck_mtx_lock_acquire_inline()
[all …]
H A Dbsd_i386_native.c133 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self()
149 machine_thread_set_tsd_base(current_thread(), self); in thread_fast_set_cthread_self64()
199 pcb = THREAD_TO_PCB(current_thread()); in thread_set_user_ldt()
/xnu-12377.1.9/osfmk/arm64/
H A Dcopyio.c304 vm_map_t map = current_thread()->map; in copyin()
336 vm_map_t map = current_thread()->map; in copyin_atomic32()
353 vm_map_t map = current_thread()->map; in copyin_atomic32_wait_if_equals()
369 vm_map_t map = current_thread()->map; in copyin_atomic64()
385 vm_map_t map = current_thread()->map; in copyout_atomic32()
401 vm_map_t map = current_thread()->map; in copyout_atomic64()
417 vm_map_t map = current_thread()->map; in copyinstr()
445 vm_map_t map = current_thread()->map; in copyout()
472 vm_map_t map = current_thread()->map; in copyoutstr_prevalidate()
H A Dbsd_arm64.c150 if (thread == current_thread()) { in thread_setsinglestep()
196 struct uthread *ut = get_bsdthread_info(current_thread()); in mach_syscall()
209 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall()
286 proc_pid(current_proc()), thread_tid(current_thread())); in mach_syscall()
298 prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name; in mach_syscall()
/xnu-12377.1.9/bsd/kern/
H A Dtty_tty.c99 context.vc_thread = current_thread(); in cttyopen()
170 context.vc_thread = current_thread(); in cttyread()
190 context.vc_thread = current_thread(); in cttywrite()
225 context.vc_thread = current_thread(); in cttyioctl()
241 context.vc_thread = current_thread(); in cttyselect()
H A Dsys_work_interval.c81 kret = kern_work_interval_create(current_thread(), &create_args); in work_interval_ctl()
105 kern_work_interval_destroy(current_thread(), create_args.wica_id); in work_interval_ctl()
210 kret = kern_work_interval_destroy(current_thread(), uap->work_interval_id); in work_interval_ctl()
245 kret = kern_work_interval_notify(current_thread(), &kwi_args); in work_interval_ctl()
260 kret = kern_work_interval_join(current_thread(), in work_interval_ctl()
/xnu-12377.1.9/bsd/dev/dtrace/
H A Dprofile_prvd.c156 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_fire()
162 pal_register_cache_state(current_thread(), VALID); in profile_fire()
164 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_fire()
182 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_fire()
194 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_fire()
215 x86_saved_state_t *kern_regs = find_kern_regs(current_thread()); in profile_tick()
221 pal_register_cache_state(current_thread(), VALID); in profile_tick()
223 x86_saved_state_t *tagged_regs = (x86_saved_state_t *)find_user_regs(current_thread()); in profile_tick()
241 arm_saved_state_t *arm_kern_regs = (arm_saved_state_t *) find_kern_regs(current_thread()); in profile_tick()
250 arm_saved_state_t *arm_user_regs = (arm_saved_state_t *)find_user_regs(current_thread()); in profile_tick()
/xnu-12377.1.9/osfmk/vm/
H A Dvm_dyld_pager.c169 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in _delta_ptr_within_page_nospec()
208 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in _delta_ptr_within_page32_nospec()
391 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64()
409 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage64()
456 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageChain32()
473 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageChain32()
537 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPage32()
593 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in signPointer()
627 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64()
651 …ktriage_record(thread_tid(current_thread()), KDBG_TRIAGE_EVENTID(KDBG_TRIAGE_SUBSYS_DYLD_PAGER, KD… in fixupPageAuth64()
[all …]
/xnu-12377.1.9/bsd/vm/
H A Dvm_unix.c1011 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1042 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1054 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1070 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1086 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1101 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_check_np()
1125 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_copyin()
1176 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup()
1214 (void *)VM_KERNEL_ADDRPERM(current_thread()), in shared_region_map_and_slide_setup()
1230 (void *)VM_KERNEL_ADDRPERM(current_thread()))); in shared_region_map_and_slide_setup()
[all …]

1234567891011