| /xnu-12377.41.6/osfmk/ipc/ |
| H A D | ipc_eventlink.c | 758 CAST_EVENT64_T(wait_eventlink), in ipc_eventlink_signal_wait_internal() 876 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked() 895 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked() 1134 if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[0])) { in kdp_eventlink_find_owner() 1141 } else if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[1])) { in kdp_eventlink_find_owner()
|
| H A D | ipc_mqueue.h | 121 #define IPC_MQUEUE_FULL CAST_EVENT64_T(&ipc_mqueue_full)
|
| /xnu-12377.41.6/osfmk/kern/ |
| H A D | epoch_sync.c | 467 CAST_EVENT64_T(sync), interruptible, TIMEOUT_WAIT_FOREVER); in esync_wait() 547 kr = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake() 555 kr = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake() 563 kr = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake() 573 kr = waitq_wakeup64_thread(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake()
|
| H A D | thread_call.c | 1492 kr = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_wake() 1509 CAST_EVENT64_T(&thread_call_daemon_awake), in thread_call_wake() 1657 waitq_wakeup64_all(&group->waiters_waitq, CAST_EVENT64_T(call), in thread_call_finish() 1862 wres = waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_INTERRUPTIBLE, 0); in thread_call_thread() 1874 …waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_UNINT, 0); /* Interrupted me… in thread_call_thread() 1937 waitq_assert_wait64(&daemon_waitq, CAST_EVENT64_T(&thread_call_daemon_awake), THREAD_UNINT, 0); in thread_call_daemon_continue() 2164 res = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_dealloc_timer() 2227 …wait_result_t res = waitq_assert_wait64(&group->waiters_waitq, CAST_EVENT64_T(call), THREAD_UNINT,… in thread_call_wait_once_locked() 2310 CAST_EVENT64_T(call), THREAD_UNINT, 0); in thread_call_wait_locked()
|
| H A D | kern_types.h | 81 #define CAST_EVENT64_T(a_ptr) ((event64_t)((uintptr_t)(a_ptr))) macro
|
| H A D | locks.c | 1305 CAST_EVENT64_T(event), result, flags); in wakeup_with_inheritor_and_turnstile() 1310 CAST_EVENT64_T(event), result, flags); in wakeup_with_inheritor_and_turnstile() 1322 ret = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(event), in wakeup_with_inheritor_and_turnstile() 1328 ret = waitq_wakeup64_thread(&ts->ts_waitq, CAST_EVENT64_T(event), in wakeup_with_inheritor_and_turnstile() 1398 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(event), interruptible, deadline); 1737 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(cond), interruptible, deadline); 2112 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_open_turnstile() 2183 hp_thread = waitq_wakeup64_identify(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_handoff_turnstile() 2371 waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), interruptible, deadline); in gate_wait_turnstile()
|
| H A D | sfi.c | 429 CAST_EVENT64_T(sfi_class_id), in sfi_timer_per_class_on() 1017 CAST_EVENT64_T(class_id), in sfi_ast() 1084 CAST_EVENT64_T(current_class_id), in sfi_reevaluate()
|
| H A D | waitq.h | 399 #define global_eventq(event) _global_eventq(CAST_EVENT64_T(event))
|
| H A D | mpsc_queue.c | 182 assert_wait_queue(dq), CAST_EVENT64_T(dq), in _mpsc_daemon_queue_init_with_thread()
|
| H A D | smr.c | 460 return CAST_EVENT64_T(&smrw->sect_queue); in __smrw_oncore_event() 466 return CAST_EVENT64_T(&smrw->whead); in __smrw_drain_event() 834 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(pcpu), in __smr_leave_stalled() 940 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(pcpu), in __smr_wait_for_stalled()
|
| H A D | sync_sema.c | 59 #define SEMAPHORE_EVENT CAST_EVENT64_T(&semaphore_event)
|
| H A D | sched_prim.c | 1147 return waitq_assert_wait64(waitq, CAST_EVENT64_T(event), interruptible, TIMEOUT_WAIT_FOREVER); in assert_wait() 1190 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout() 1238 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout_with_leeway() 1272 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline() 1307 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline_with_leeway() 1721 count = waitq_wakeup64_nthreads(wq, CAST_EVENT64_T(event), result, in thread_wakeup_nthreads_prim() 1764 return waitq_wakeup64_thread(wq, CAST_EVENT64_T(event), thread, THREAD_AWAKENED); in thread_wakeup_thread()
|
| H A D | lock_mtx.c | 105 #define LCK_MTX_EVENT(lck) CAST_EVENT64_T(&(lck)->lck_mtx.data)
|
| H A D | turnstile.c | 3578 CAST_EVENT64_T(test_prim), wait_type, in tstile_test_prim_lock() 3684 CAST_EVENT64_T(test_prim), in tstile_test_prim_unlock()
|
| H A D | thread.c | 1744 assert_wait_queue(event), CAST_EVENT64_T(event), in thread_create_waiting_internal()
|
| H A D | task.c | 911 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_clear_return_wait() 990 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_wait_to_return()
|
| /xnu-12377.41.6/bsd/kern/ |
| H A D | sys_ulock.c | 755 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in sys_ulock_wait2() 1038 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1045 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1055 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1058 waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
|
| H A D | kern_aio.c | 1482 waitq_wakeup64_one(&queue->aioq_waitq, CAST_EVENT64_T(queue), in aio_try_enqueue_work_locked() 1838 waitq_assert_wait64(&queue->aioq_waitq, CAST_EVENT64_T(queue), THREAD_UNINT, 0); in aio_get_some_work()
|
| H A D | kern_event.c | 546 return CAST_EVENT64_T(kn); in knote_filt_wev64()
|
| /xnu-12377.41.6/osfmk/i386/ |
| H A D | locks.h | 150 #define LCK_MTX_EVENT(lck) CAST_EVENT64_T(&(lck)->lck_mtx_owner)
|