| /xnu-8019.80.24/osfmk/ipc/ |
| H A D | ipc_eventlink.c | 761 CAST_EVENT64_T(wait_eventlink), in ipc_eventlink_signal_wait_internal() 878 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked() 898 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked() 1140 if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[0])) { in kdp_eventlink_find_owner() 1147 } else if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[1])) { in kdp_eventlink_find_owner()
|
| H A D | ipc_mqueue.h | 131 #define IPC_MQUEUE_FULL CAST_EVENT64_T(&ipc_mqueue_full)
|
| /xnu-8019.80.24/osfmk/kern/ |
| H A D | thread_call.c | 1434 kr = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_wake() 1450 waitq_wakeup64_all(&daemon_waitq, CAST_EVENT64_T(&thread_call_daemon_awake), in thread_call_wake() 1581 waitq_wakeup64_all(&group->waiters_waitq, CAST_EVENT64_T(call), in thread_call_finish() 1772 wres = waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_INTERRUPTIBLE, 0); in thread_call_thread() 1784 …waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_UNINT, 0); /* Interrupted me… in thread_call_thread() 1847 waitq_assert_wait64(&daemon_waitq, CAST_EVENT64_T(&thread_call_daemon_awake), THREAD_UNINT, 0); in thread_call_daemon_continue() 2057 res = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_dealloc_timer() 2120 …wait_result_t res = waitq_assert_wait64(&group->waiters_waitq, CAST_EVENT64_T(call), THREAD_UNINT,… in thread_call_wait_once_locked() 2200 CAST_EVENT64_T(call), THREAD_UNINT, 0); in thread_call_wait_locked()
|
| H A D | kern_types.h | 70 #define CAST_EVENT64_T(a_ptr) ((event64_t)((uintptr_t)(a_ptr))) macro
|
| H A D | locks.c | 1481 …waitq_assert_wait64(&turnstile->ts_waitq, CAST_EVENT64_T(LCK_MTX_EVENT(mutex)), THREAD_UNINT | THR… in lck_mtx_lock_wait() 1591 …did_wake = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(LCK_MTX_EVENT(mutex)), THREAD_AWAKENED… in lck_mtx_unlock_wakeup() 1593 …did_wake = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(LCK_MTX_EVENT(mutex)), THREAD_AWAKENED… in lck_mtx_unlock_wakeup() 1808 wokeup = waitq_wakeup64_identify(&ts->ts_waitq, CAST_EVENT64_T(event), result, priority); in wakeup_with_inheritor_and_turnstile_type() 1827 ret = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(event), result, WAITQ_ALL_PRIORITIES); in wakeup_with_inheritor_and_turnstile_type() 1894 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(event), interruptible, deadline); 2433 …waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), THREAD_AWAKENED, WAITQ_ALL_PRI… in gate_open_turnstile() 2504 …hp_thread = waitq_wakeup64_identify(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), THREAD_AWAKEN… in gate_handoff_turnstile() 2691 waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), interruptible, deadline); in gate_wait_turnstile()
|
| H A D | sfi.c | 420 CAST_EVENT64_T(sfi_class_id), in sfi_timer_per_class_on() 1013 CAST_EVENT64_T(class_id), in sfi_ast() 1080 CAST_EVENT64_T(current_class_id), in sfi_reevaluate()
|
| H A D | sched_prim.c | 1161 return waitq_assert_wait64(waitq, CAST_EVENT64_T(event), interruptible, TIMEOUT_WAIT_FOREVER); in assert_wait() 1204 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout() 1252 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout_with_leeway() 1286 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline() 1321 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline_with_leeway() 1671 return waitq_wakeup64_one(wq, CAST_EVENT64_T(event), result, WAITQ_ALL_PRIORITIES); in thread_wakeup_prim() 1673 return waitq_wakeup64_all(wq, CAST_EVENT64_T(event), result, WAITQ_ALL_PRIORITIES); in thread_wakeup_prim() 1695 return waitq_wakeup64_thread(wq, CAST_EVENT64_T(event), thread, THREAD_AWAKENED); in thread_wakeup_thread() 1714 return waitq_wakeup64_one(wq, CAST_EVENT64_T(event), THREAD_AWAKENED, priority); in thread_wakeup_one_with_pri() 1734 return waitq_wakeup64_identify(wq, CAST_EVENT64_T(event), THREAD_AWAKENED, priority); in thread_wakeup_identify()
|
| H A D | sync_sema.c | 61 #define SEMAPHORE_EVENT CAST_EVENT64_T(&semaphore_event)
|
| H A D | thread_act.c | 137 wait_result = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in thread_start_in_assert_wait()
|
| H A D | turnstile.c | 3459 CAST_EVENT64_T(test_prim), wait_type, in tstile_test_prim_lock() 3557 CAST_EVENT64_T(test_prim), in tstile_test_prim_unlock()
|
| H A D | task.c | 714 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_clear_return_wait() 746 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_wait_to_return()
|
| /xnu-8019.80.24/bsd/kern/ |
| H A D | sys_ulock.c | 708 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wait2() 977 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 988 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 997 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1000 waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
|
| H A D | kern_aio.c | 1350 waitq_wakeup64_one(&queue->aioq_waitq, CAST_EVENT64_T(queue), in aio_try_enqueue_work_locked() 1647 waitq_assert_wait64(&queue->aioq_waitq, CAST_EVENT64_T(queue), THREAD_UNINT, 0); in aio_get_some_work()
|
| H A D | kern_event.c | 510 return CAST_EVENT64_T(kn); in knote_filt_wev64()
|
| /xnu-8019.80.24/osfmk/i386/ |
| H A D | locks_i386.c | 1316 …did_wake = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(LCK_MTX_EVENT(mutex)), THREAD_AWAKENED… in lck_mtx_unlock_wakeup_tail() 1318 …did_wake = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(LCK_MTX_EVENT(mutex)), THREAD_AWAKENED… in lck_mtx_unlock_wakeup_tail() 2415 …waitq_assert_wait64(&turnstile->ts_waitq, CAST_EVENT64_T(LCK_MTX_EVENT(mutex)), THREAD_UNINT | THR… in lck_mtx_lock_wait_x86()
|