Home
last modified time | relevance | path

Searched refs:CAST_EVENT64_T (Results 1 – 19 of 19) sorted by relevance

/xnu-11215.61.5/osfmk/ipc/
H A Dipc_eventlink.c758 CAST_EVENT64_T(wait_eventlink), in ipc_eventlink_signal_wait_internal()
876 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked()
895 CAST_EVENT64_T(signal_eventlink), in ipc_eventlink_signal_internal_locked()
1134 if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[0])) { in kdp_eventlink_find_owner()
1141 } else if (event == CAST_EVENT64_T(&ipc_eventlink_base->elb_eventlink[1])) { in kdp_eventlink_find_owner()
H A Dipc_mqueue.h131 #define IPC_MQUEUE_FULL CAST_EVENT64_T(&ipc_mqueue_full)
/xnu-11215.61.5/osfmk/kern/
H A Depoch_sync.c469 CAST_EVENT64_T(sync), interruptible, TIMEOUT_WAIT_FOREVER); in esync_wait()
549 kr = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake()
557 kr = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake()
565 kr = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake()
575 kr = waitq_wakeup64_thread(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake()
H A Dthread_call.c1492 kr = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_wake()
1509 CAST_EVENT64_T(&thread_call_daemon_awake), in thread_call_wake()
1657 waitq_wakeup64_all(&group->waiters_waitq, CAST_EVENT64_T(call), in thread_call_finish()
1862 wres = waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_INTERRUPTIBLE, 0); in thread_call_thread()
1874 …waitq_assert_wait64(&group->idle_waitq, CAST_EVENT64_T(group), THREAD_UNINT, 0); /* Interrupted me… in thread_call_thread()
1937 waitq_assert_wait64(&daemon_waitq, CAST_EVENT64_T(&thread_call_daemon_awake), THREAD_UNINT, 0); in thread_call_daemon_continue()
2151 res = waitq_wakeup64_one(&group->idle_waitq, CAST_EVENT64_T(group), in thread_call_dealloc_timer()
2214 …wait_result_t res = waitq_assert_wait64(&group->waiters_waitq, CAST_EVENT64_T(call), THREAD_UNINT,… in thread_call_wait_once_locked()
2297 CAST_EVENT64_T(call), THREAD_UNINT, 0); in thread_call_wait_locked()
H A Dkern_types.h81 #define CAST_EVENT64_T(a_ptr) ((event64_t)((uintptr_t)(a_ptr))) macro
H A Dlocks.c1305 CAST_EVENT64_T(event), result, flags); in wakeup_with_inheritor_and_turnstile()
1310 CAST_EVENT64_T(event), result, flags); in wakeup_with_inheritor_and_turnstile()
1322 ret = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(event), in wakeup_with_inheritor_and_turnstile()
1328 ret = waitq_wakeup64_thread(&ts->ts_waitq, CAST_EVENT64_T(event), in wakeup_with_inheritor_and_turnstile()
1398 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(event), interruptible, deadline);
1737 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(cond), interruptible, deadline);
2112 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_open_turnstile()
2183 hp_thread = waitq_wakeup64_identify(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_handoff_turnstile()
2371 waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), interruptible, deadline); in gate_wait_turnstile()
H A Dsfi.c428 CAST_EVENT64_T(sfi_class_id), in sfi_timer_per_class_on()
1015 CAST_EVENT64_T(class_id), in sfi_ast()
1082 CAST_EVENT64_T(current_class_id), in sfi_reevaluate()
H A Dmpsc_queue.c182 assert_wait_queue(dq), CAST_EVENT64_T(dq), in _mpsc_daemon_queue_init_with_thread()
H A Dsmr.c460 return CAST_EVENT64_T(&smrw->sect_queue); in __smrw_oncore_event()
466 return CAST_EVENT64_T(&smrw->whead); in __smrw_drain_event()
834 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(pcpu), in __smr_leave_stalled()
940 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(pcpu), in __smr_wait_for_stalled()
H A Dsched_prim.c1329 return waitq_assert_wait64(waitq, CAST_EVENT64_T(event), interruptible, TIMEOUT_WAIT_FOREVER); in assert_wait()
1372 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout()
1420 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_timeout_with_leeway()
1454 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline()
1489 wresult = waitq_assert_wait64_locked(waitq, CAST_EVENT64_T(event), in assert_wait_deadline_with_leeway()
1904 return waitq_wakeup64_one(wq, CAST_EVENT64_T(event), result, WAITQ_WAKEUP_DEFAULT); in thread_wakeup_prim()
1906 return waitq_wakeup64_all(wq, CAST_EVENT64_T(event), result, WAITQ_WAKEUP_DEFAULT); in thread_wakeup_prim()
1928 return waitq_wakeup64_thread(wq, CAST_EVENT64_T(event), thread, THREAD_AWAKENED); in thread_wakeup_thread()
1947 return waitq_wakeup64_one(wq, CAST_EVENT64_T(event), THREAD_AWAKENED, priority); in thread_wakeup_one_with_pri()
1967 return waitq_wakeup64_identify(wq, CAST_EVENT64_T(event), THREAD_AWAKENED, priority); in thread_wakeup_identify()
H A Dsync_sema.c61 #define SEMAPHORE_EVENT CAST_EVENT64_T(&semaphore_event)
H A Dlock_mtx.c105 #define LCK_MTX_EVENT(lck) CAST_EVENT64_T(&(lck)->lck_mtx.data)
H A Dturnstile.c3564 CAST_EVENT64_T(test_prim), wait_type, in tstile_test_prim_lock()
3670 CAST_EVENT64_T(test_prim), in tstile_test_prim_unlock()
H A Dthread.c1742 assert_wait_queue(event), CAST_EVENT64_T(event), in thread_create_waiting_internal()
H A Dtask.c916 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_clear_return_wait()
947 CAST_EVENT64_T(task_get_return_wait_event(task)), in task_wait_to_return()
/xnu-11215.61.5/bsd/kern/
H A Dsys_ulock.c745 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in sys_ulock_wait2()
1028 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
1035 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
1045 CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
1048 waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
H A Dkern_aio.c1349 waitq_wakeup64_one(&queue->aioq_waitq, CAST_EVENT64_T(queue), in aio_try_enqueue_work_locked()
1646 waitq_assert_wait64(&queue->aioq_waitq, CAST_EVENT64_T(queue), THREAD_UNINT, 0); in aio_get_some_work()
H A Dkern_event.c545 return CAST_EVENT64_T(kn); in knote_filt_wev64()
/xnu-11215.61.5/osfmk/i386/
H A Dlocks.h150 #define LCK_MTX_EVENT(lck) CAST_EVENT64_T(&(lck)->lck_mtx_owner)