| /xnu-8792.41.9/osfmk/ipc/ |
| H A D | ipc_eventlink.c | 758 CAST_EVENT64_T(wait_eventlink), in ipc_eventlink_signal_wait_internal() 876 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked() 895 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked() 1134 if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[0])) { in kdp_eventlink_find_owner() 1141 } else if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[1])) { in kdp_eventlink_find_owner()
|
| H A D | ipc_mqueue.h | 131 #define IPC_MQUEUE_FULL CAST_EVENT64_T(&ipc_mqueue_full)
|
| /xnu-8792.41.9/bsd/kern/ |
| H A D | sys_ulock.c | 714 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in sys_ulock_wait2() 997 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1004 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1014 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1017 waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
|
| H A D | kern_aio.c | 1350 waitq_wakeup64_one(&queue->aioq_waitq, CAST_EVENT64_T(queue), in aio_try_enqueue_work_locked() 1647 waitq_assert_wait64(&queue->aioq_waitq, CAST_EVENT64_T(queue), THREAD_UNINT, 0); in aio_get_some_work()
|
| H A D | kern_event.c | 511 return CAST_EVENT64_T(kn); in knote_filt_wev64()
|
| /xnu-8792.41.9/osfmk/kern/ |
| H A D | thread_call.c | 1491 kr = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_wake() 1508 CAST_EVENT64_T(&thread_call_daemon_awake), in thread_call_wake() 1656 waitq_wakeup64_all(&group->waiters_waitq, CAST_EVENT64_T(call), in thread_call_finish() 1861 wres = waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_INTERRUPTIBLE, 0); in thread_call_thread() 1873 …waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_UNINT, 0); /* Interrupted me… in thread_call_thread() 1936 waitq_assert_wait64(&daemon_waitq, CAST_EVENT64_T(&thread_call_daemon_awake), THREAD_UNINT, 0); in thread_call_daemon_continue() 2150 res = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_dealloc_timer() 2213 …wait_result_t res = waitq_assert_wait64(&group->waiters_waitq, CAST_EVENT64_T(call), THREAD_UNINT,… in thread_call_wait_once_locked() 2296 CAST_EVENT64_T(call), THREAD_UNINT, 0); in thread_call_wait_locked()
|
| H A D | kern_types.h | 81 #define CAST_EVENT64_T(a_ptr) ((event64_t)((uintptr_t)(a_ptr))) macro
|
| H A D | sfi.c | 427 CAST_EVENT64_T(sfi_class_id), in sfi_timer_per_class_on() 1018 CAST_EVENT64_T(class_id), in sfi_ast() 1085 CAST_EVENT64_T(current_class_id), in sfi_reevaluate()
|
| H A D | locks.c | 1238 CAST_EVENT64_T(event), result, flags); in wakeup_with_inheritor_and_turnstile() 1257 ret = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(event), in wakeup_with_inheritor_and_turnstile() 1325 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(event), interruptible, deadline); 1758 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(cond), interruptible, deadline); 2133 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_open_turnstile() 2204 hp_thread = waitq_wakeup64_identify(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_handoff_turnstile() 2392 waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), interruptible, deadline); in gate_wait_turnstile()
|
| H A D | sched_prim.c | 1176 return waitq_assert_wait64(waitq, CAST_EVENT64_T(event), interruptible, TIMEOUT_WAIT_FOREVER); in assert_wait() 1219 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout() 1267 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout_with_leeway() 1301 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline() 1336 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline_with_leeway() 1714 return waitq_wakeup64_one(wq, CAST_EVENT64_T(event), result, WAITQ_WAKEUP_DEFAULT); in thread_wakeup_prim() 1716 return waitq_wakeup64_all(wq, CAST_EVENT64_T(event), result, WAITQ_WAKEUP_DEFAULT); in thread_wakeup_prim() 1738 return waitq_wakeup64_thread(wq, CAST_EVENT64_T(event), thread, THREAD_AWAKENED); in thread_wakeup_thread() 1757 return waitq_wakeup64_one(wq, CAST_EVENT64_T(event), THREAD_AWAKENED, priority); in thread_wakeup_one_with_pri() 1777 return waitq_wakeup64_identify(wq, CAST_EVENT64_T(event), THREAD_AWAKENED, priority); in thread_wakeup_identify()
|
| H A D | sync_sema.c | 61 #define SEMAPHORE_EVENT CAST_EVENT64_T(&semaphore_event)
|
| H A D | thread_act.c | 138 wait_result = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in thread_start_in_assert_wait()
|
| H A D | lock_mtx.c | 143 #define LCK_MTX_EVENT(lck) CAST_EVENT64_T(&(lck)->lck_mtx.data)
|
| H A D | turnstile.c | 3607 CAST_EVENT64_T(test_prim), wait_type, in tstile_test_prim_lock() 3713 CAST_EVENT64_T(test_prim), in tstile_test_prim_unlock()
|
| H A D | task.c | 761 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_clear_return_wait() 791 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_wait_to_return()
|
| /xnu-8792.41.9/osfmk/i386/ |
| H A D | locks.h | 160 #define LCK_MTX_EVENT(lck) CAST_EVENT64_T(&(lck)->lck_mtx_owner)
|