| /xnu-10063.141.1/osfmk/kern/ |
| H A D | turnstile.c | 827 waitq_init(&turnstile->ts_waitq, WQT_TURNSTILE, in turnstile_alloc() 841 waitq_unlock(&turnstile->ts_waitq); in turnstile_alloc() 933 waitq_deinit(&turnstile->ts_waitq); in turnstile_destroy() 1345 assert(waitq_held(&turnstile->ts_waitq)); in turnstile_update_inheritor_locked() 1572 waitq_lock(&turnstile->ts_waitq); in turnstile_update_inheritor() 1576 waitq_unlock(&turnstile->ts_waitq); in turnstile_update_inheritor() 1605 &dst_turnstile->ts_waitq.waitq_prio_queue, in turnstile_need_thread_promotion_update() 1671 &dst_turnstile->ts_waitq.waitq_prio_queue, in turnstile_update_thread_promotion_locked() 1684 &dst_turnstile->ts_waitq.waitq_prio_queue, in turnstile_update_thread_promotion_locked() 2270 waitq_lock(&dst_turnstile->ts_waitq); in turnstile_update_turnstile_promotion() [all …]
|
| H A D | turnstile.h | 344 WAITQ_FLAGS(ts_waitq 348 struct waitq ts_waitq; /* waitq embedded in turnstile */ member 350 #define ts_inheritor ts_waitq.waitq_inheritor /* thread/turnstile inheriting the pri… 373 #define waitq_to_turnstile(waitq) __container_of(waitq, struct turnstile, ts_waitq)
|
| H A D | epoch_sync.c | 370 wait_result_t wr = waitq_assert_wait64(&ts->ts_waitq, in esync_wait() 437 kr = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake() 443 kr = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake() 449 kr = waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake() 457 kr = waitq_wakeup64_thread(&ts->ts_waitq, CAST_EVENT64_T(sync), in esync_wake()
|
| H A D | locks.c | 1304 wokeup = waitq_wakeup64_identify(&ts->ts_waitq, in wakeup_with_inheritor_and_turnstile() 1309 ret = waitq_wakeup64_one(&ts->ts_waitq, in wakeup_with_inheritor_and_turnstile() 1322 ret = waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(event), in wakeup_with_inheritor_and_turnstile() 1328 ret = waitq_wakeup64_thread(&ts->ts_waitq, CAST_EVENT64_T(event), in wakeup_with_inheritor_and_turnstile() 1398 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(event), interruptible, deadline); 1737 ret = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(cond), interruptible, deadline); 2112 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_open_turnstile() 2183 hp_thread = waitq_wakeup64_identify(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), in gate_handoff_turnstile() 2371 waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(GATE_EVENT(gate)), interruptible, deadline); in gate_wait_turnstile()
|
| H A D | lock_mtx.c | 571 waitq_assert_wait64(&ts->ts_waitq, LCK_MTX_EVENT(lck), in lck_mtx_lock_wait() 641 did_wake = waitq_wakeup64_one(&ts->ts_waitq, LCK_MTX_EVENT(lck), in lck_mtx_unlock_wakeup()
|
| H A D | waitq.c | 419 priority_queue_empty(&ts->ts_waitq.waitq_prio_queue); in waitq_empty() 522 return ts ? &ts->ts_waitq : NULL; in waitq_get_safeq()
|
| H A D | smr.c | 833 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(pcpu), in __smr_leave_stalled() 939 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(pcpu), in __smr_wait_for_stalled()
|
| H A D | task.c | 929 waitq_wakeup64_all(&turnstile->ts_waitq, in task_clear_return_wait() 960 waitq_assert_wait64(&turnstile->ts_waitq, in task_wait_to_return()
|
| /xnu-10063.141.1/bsd/kern/ |
| H A D | sys_ulock.c | 741 wr = waitq_assert_wait64(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in sys_ulock_wait2() 1023 kern_return_t kr = waitq_wakeup64_thread(&ts->ts_waitq, in ulock_wake() 1031 waitq_wakeup64_all(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake() 1040 new_owner = waitq_wakeup64_identify(&ts->ts_waitq, in ulock_wake() 1044 waitq_wakeup64_one(&ts->ts_waitq, CAST_EVENT64_T(ULOCK_TO_EVENT(ull)), in ulock_wake()
|
| H A D | kern_event.c | 2255 waitq_wakeup64_thread(&ts->ts_waitq, knote_filt_wev64(kn), in filt_wlupdate() 2573 waitq_assert_wait64(&ts->ts_waitq, knote_filt_wev64(kn), in filt_wlpost_register_wait()
|
| /xnu-10063.141.1/tools/lldbmacros/ |
| H A D | waitq.py | 48 if ts: return ts.ts_waitq.waitq_ts 100 for t in IterateSchedPriorityQueue(ts.ts_waitq.waitq_prio_queue,
|
| H A D | turnstile.py | 101 waitq = Waitq(addressof(turnstile.ts_waitq))
|
| /xnu-10063.141.1/osfmk/ipc/ |
| H A D | ipc_mqueue.c | 514 &send_turnstile->ts_waitq, in ipc_mqueue_send_locked() 637 if (waitq_wakeup64_one(&send_turnstile->ts_waitq, in ipc_mqueue_release_msgcount() 1442 waitq_wakeup64_all(&send_turnstile->ts_waitq, in ipc_mqueue_destroy_locked() 1519 waitq_wakeup64_one(&send_turnstile->ts_waitq, in ipc_mqueue_set_qlimit_locked()
|
| /xnu-10063.141.1/bsd/pthread/ |
| H A D | pthread_shims.c | 237 wr = waitq_assert_wait64_leeway(&ts->ts_waitq, (event64_t)kwq, in psynch_wait_prepare() 300 kr = waitq_wakeup64_thread(&ts->ts_waitq, (event64_t)kwq, th, in psynch_wait_wakeup()
|
| /xnu-10063.141.1/osfmk/i386/ |
| H A D | locks_i386.c | 1190 did_wake = waitq_wakeup64_one(&ts->ts_waitq, LCK_MTX_EVENT(mutex), in lck_mtx_unlock_wakeup_tail() 2189 …waitq_assert_wait64(&turnstile->ts_waitq, LCK_MTX_EVENT(mutex), THREAD_UNINT | THREAD_WAIT_NOREPOR… in lck_mtx_lock_wait_x86()
|