| /xnu-8792.41.9/osfmk/kern/ |
| H A D | cpu_quiesce.c | 193 struct cpu_quiesce *st = PERCPU_GET(cpu_quiesce); in cpu_quiescent_counter_join() 219 struct cpu_quiesce *st = PERCPU_GET(cpu_quiesce); in cpu_quiescent_counter_ast() 270 struct cpu_quiesce *st = PERCPU_GET(cpu_quiesce); in cpu_quiescent_counter_leave() 323 struct cpu_quiesce *st = PERCPU_GET(cpu_quiesce); in cpu_quiescent_counter_checkin() 373 struct cpu_quiesce *st = PERCPU_GET(cpu_quiesce); in cpu_quiescent_counter_assert_ast()
|
| H A D | iotrace.h | 78 nextidxp = PERCPU_GET(iotrace_next); in iotrace() 80 cur_iotrace_ring = *PERCPU_GET(iotrace_ring); in iotrace()
|
| H A D | stack.c | 240 cache = PERCPU_GET(stack_cache); in stack_free_stack() 275 cache = PERCPU_GET(stack_cache); in stack_alloc_try()
|
| H A D | recount.c | 422 struct recount_snap *this_snap = PERCPU_GET(_snaps_percpu); in recount_update_snap() 452 struct recount_snap *last = PERCPU_GET(_snaps_percpu); in recount_current_thread_usage() 492 struct recount_snap *last = PERCPU_GET(_snaps_percpu); in recount_current_thread_perf_level_usage() 553 struct recount_snap *last = PERCPU_GET(_snaps_percpu); in _time_since_last_snapshot() 768 struct recount_snap *last = PERCPU_GET(_snaps_percpu); in recount_switch_thread() 862 struct recount_snap *last = PERCPU_GET(_snaps_percpu); in recount_kernel_transition()
|
| H A D | percpu.h | 107 #define PERCPU_GET(name) \ macro
|
| H A D | debug.c | 191 return PERCPU_GET(debugger_state); in current_debugger_state() 2037 if (__probable(*PERCPU_GET(hv_entry_detected) || !awl_scratch_reg_supported)) { 2040 *PERCPU_GET(hv_entry_detected) = true; 2053 if (*PERCPU_GET(hv_entry_detected)) {
|
| H A D | smr.c | 701 slot = PERCPU_GET(smr_bucket); in smr_global_retire()
|
| H A D | lock_mtx.c | 465 mcs = PERCPU_GET(lck_mtx_mcs); in lck_mtx_ilk_lock_contended() 715 mcs = PERCPU_GET(lck_mtx_mcs); in lck_mtx_lock_adaptive_spin()
|
| H A D | locks.c | 271 PERCPU_GET(lck_spinlock_to_info)->owner_thread_orig = owner & ~0x7ul; in lck_spinlock_timeout_set_orig_owner() 282 PERCPU_GET(lck_spinlock_to_info)->owner_thread_orig = in lck_spinlock_timeout_set_orig_ctid() 292 lck_spinlock_to_info_t lsti = PERCPU_GET(lck_spinlock_to_info); in lck_spinlock_timeout_hit()
|
| H A D | sched_prim.h | 463 PERCPU_GET(sched_stats)->field++; \
|
| H A D | lock_ticket.c | 503 PERCPU_GET(lck_spinlock_to_info)->extra = arg.mt; in hw_lck_ticket_contended()
|
| H A D | waitq.c | 2020 id = (*PERCPU_GET(select_setid) += inc); in select_set_nextid() 2034 id = os_atomic_add(PERCPU_GET(select_setid), inc, relaxed); in select_set_nextid()
|
| /xnu-8792.41.9/san/memory/ |
| H A D | ubsan_minimal.c | 149 bool *in_handler = PERCPU_GET(ubsan_minimal_in_handler); in ubsan_minimal_stash_telemetry() 159 uintptr_t *cache_address = PERCPU_GET(ubsan_minimal_cache_address); in ubsan_minimal_stash_telemetry() 200 bool *in_handler = PERCPU_GET(ubsan_minimal_in_handler); in ubsan_minimal_flush_entries()
|
| /xnu-8792.41.9/osfmk/arm64/ |
| H A D | lock_ticket_pv.c | 108 lck_tktlock_pv_info_t ltpi = PERCPU_GET(lck_tktlock_pv_info); in hw_lck_ticket_lock_wait_pv()
|
| /xnu-8792.41.9/osfmk/i386/ |
| H A D | lock_ticket_pv.c | 129 lck_tktlock_pv_info_t ltpi = PERCPU_GET(lck_tktlock_pv_info); in hw_lck_ticket_lock_wait_pv()
|
| H A D | cpu_data.h | 563 nextidxp = PERCPU_GET(traptrace_next); in traptrace_start() 569 *PERCPU_GET(traptrace_ring), sizeof(traptrace_entry_t) * traptrace_entries_per_cpu); in traptrace_start()
|
| /xnu-8792.41.9/bsd/vfs/ |
| H A D | vfs_io_compression_stats.c | 276 lz4_encode_scratch_t *scratch_buf = *PERCPU_GET(per_cpu_scratch_buf); in iocs_compress_block() 277 uint8_t *dest_buf = *PERCPU_GET(per_cpu_compression_buf); in iocs_compress_block()
|
| /xnu-8792.41.9/libkern/os/ |
| H A D | log_queue.c | 543 log_queue_t lq = PERCPU_GET(oslog_queue); in log_queue_dispatch() 595 log_queue_t lq = PERCPU_GET(oslog_queue); in log_queue_add()
|
| /xnu-8792.41.9/tools/lldbmacros/ |
| H A D | recount.py | 448 snap = kern.PERCPU_GET('_snaps_percpu', cpu_id) 522 snap = kern.PERCPU_GET('_snaps_percpu', i)
|
| H A D | xnu.py | 1068 getattr(kern.PERCPU_GET(ring, 0)[0], field_arg) 1094 ring_slice = [(x, y, kern.PERCPU_GET(ring, x)[y]) for y in range(entries_per_cpu)]
|
| H A D | memory.py | 2510 mcs = kern.PERCPU_GET('lck_mtx_mcs', cpu) 2523 mcs = addressof(kern.PERCPU_GET('lck_mtx_mcs', idx - 1)) 2540 mcs = addressof(kern.PERCPU_GET('lck_mtx_mcs', idx - 1))
|
| /xnu-8792.41.9/osfmk/prng/ |
| H A D | entropy.c | 228 entropy_cpu_data_t *e = PERCPU_GET(entropy_cpu_data); in entropy_collect()
|
| /xnu-8792.41.9/osfmk/arm/ |
| H A D | locks_arm.c | 309 uint64_t _Atomic * const max_duration = PERCPU_GET(preemption_disable_max_mt); in _collect_preemption_disable_measurement()
|
| H A D | cpu_common.c | 702 return PERCPU_GET(processor); in current_processor()
|
| /xnu-8792.41.9/tools/lldbmacros/core/ |
| H A D | kernelcore.py | 440 def PERCPU_GET(self, name, cpu): member in KernelTarget
|