| /xnu-10002.61.3/osfmk/arm/ |
| H A D | preemption_disable.c | 264 struct _preemption_disable_pcpu *pcpu = PERCPU_GET(_preemption_disable_pcpu_data); in _preemption_disable_snap_start() 297 struct _preemption_disable_pcpu *pcpu = PERCPU_GET(_preemption_disable_pcpu_data); in _preemption_disable_snap_end() 460 struct _preemption_disable_pcpu *pcpu = PERCPU_GET(_preemption_disable_pcpu_data); in abandon_preemption_disable_measurement() 476 struct _preemption_disable_pcpu *pcpu = PERCPU_GET(_preemption_disable_pcpu_data); in _do_disable_preemption_without_measurements()
|
| H A D | cpu_common.c | 702 return PERCPU_GET(processor); in current_processor()
|
| /xnu-10002.61.3/osfmk/kern/ |
| H A D | iotrace.h | 78 nextidxp = PERCPU_GET(iotrace_next); in iotrace() 80 cur_iotrace_ring = *PERCPU_GET(iotrace_ring); in iotrace()
|
| H A D | stack.c | 240 cache = PERCPU_GET(stack_cache); in stack_free_stack() 275 cache = PERCPU_GET(stack_cache); in stack_alloc_try()
|
| H A D | percpu.h | 108 #define PERCPU_GET(name) \ macro
|
| H A D | smr.c | 699 smrw = PERCPU_GET(smr_worker); in smr_enter_sleepable() 739 __smr_wake_oncore_sleepers(PERCPU_GET(smr_worker)); in smr_ack_ipi() 745 struct smr_worker *smrw = PERCPU_GET(smr_worker); in smr_mark_active_trackers_stalled() 868 smrw = PERCPU_GET(smr_worker); in smr_leave_sleepable() 873 __smr_wake_oncore_sleepers(PERCPU_GET(smr_worker)); in smr_leave_sleepable() 1455 smrw = PERCPU_GET(smr_worker); in smr_call() 1563 __smr_cpu_lazy_up_if_needed(PERCPU_GET(smr_worker)); in smr_barrier() 2167 struct smr_worker *smrw = PERCPU_GET(smr_worker); in smr_cpu_tick()
|
| H A D | locks_internal.h | 257 return PERCPU_GET(lck_mcs); in lck_mcs_get_current()
|
| H A D | debug.c | 203 return PERCPU_GET(debugger_state); in current_debugger_state() 2097 if (__probable(*PERCPU_GET(hv_entry_detected) || !awl_scratch_reg_supported)) { 2100 *PERCPU_GET(hv_entry_detected) = true; 2113 if (*PERCPU_GET(hv_entry_detected)) {
|
| H A D | telemetry.c | 1406 bool *in_handler = PERCPU_GET(brk_telemetry_in_handler); in telemetry_stash_ca_event() 1415 uintptr_t *cache_address = PERCPU_GET(brk_telemetry_cache_address); in telemetry_stash_ca_event() 1478 bool *in_handler = PERCPU_GET(brk_telemetry_in_handler); in telemetry_flush_ca_events()
|
| H A D | sched_prim.h | 483 PERCPU_GET(sched_stats)->field++; \
|
| H A D | locks.c | 273 PERCPU_GET(lck_spinlock_to_info)->owner_thread_orig = owner & ~0x7ul; in lck_spinlock_timeout_set_orig_owner() 284 PERCPU_GET(lck_spinlock_to_info)->owner_thread_orig = in lck_spinlock_timeout_set_orig_ctid() 294 lck_spinlock_to_info_t lsti = PERCPU_GET(lck_spinlock_to_info); in lck_spinlock_timeout_hit()
|
| H A D | lock_ticket.c | 503 PERCPU_GET(lck_spinlock_to_info)->extra = arg.mt; in hw_lck_ticket_contended()
|
| H A D | waitq.c | 2122 id = (*PERCPU_GET(select_setid) += inc); in select_set_nextid() 2136 id = os_atomic_add(PERCPU_GET(select_setid), inc, relaxed); in select_set_nextid()
|
| /xnu-10002.61.3/osfmk/arm/pmap/ |
| H A D | pmap_misc.c | 46 struct _preemption_disable_pcpu *pcpu = PERCPU_GET(_preemption_disable_pcpu_data); in pmap_abandon_measurement()
|
| /xnu-10002.61.3/osfmk/arm64/ |
| H A D | lock_ticket_pv.c | 108 lck_tktlock_pv_info_t ltpi = PERCPU_GET(lck_tktlock_pv_info); in hw_lck_ticket_lock_wait_pv()
|
| /xnu-10002.61.3/osfmk/i386/ |
| H A D | lock_ticket_pv.c | 129 lck_tktlock_pv_info_t ltpi = PERCPU_GET(lck_tktlock_pv_info); in hw_lck_ticket_lock_wait_pv()
|
| H A D | cpu_data.h | 570 nextidxp = PERCPU_GET(traptrace_next); in traptrace_start() 576 *PERCPU_GET(traptrace_ring), sizeof(traptrace_entry_t) * traptrace_entries_per_cpu); in traptrace_start()
|
| /xnu-10002.61.3/bsd/vfs/ |
| H A D | vfs_io_compression_stats.c | 276 lz4_encode_scratch_t *scratch_buf = *PERCPU_GET(per_cpu_scratch_buf); in iocs_compress_block() 277 uint8_t *dest_buf = *PERCPU_GET(per_cpu_compression_buf); in iocs_compress_block()
|
| /xnu-10002.61.3/libkern/os/ |
| H A D | log_queue.c | 543 log_queue_t lq = PERCPU_GET(oslog_queue); in log_queue_dispatch() 595 log_queue_t lq = PERCPU_GET(oslog_queue); in log_queue_add()
|
| /xnu-10002.61.3/osfmk/prng/ |
| H A D | entropy.c | 223 entropy_cpu_data_t *e = PERCPU_GET(entropy_cpu_data); in entropy_collect()
|
| /xnu-10002.61.3/san/memory/ |
| H A D | kasan-classic.c | 637 kasan_quarantine_t q = PERCPU_GET(kasan_quarantine); in kasan_quarantine()
|
| /xnu-10002.61.3/tools/lldbmacros/core/ |
| H A D | kernelcore.py | 350 def PERCPU_GET(self, name, cpu): member in KernelTarget
|
| /xnu-10002.61.3/tools/lldbmacros/ |
| H A D | xnu.py | 1301 getattr(kern.PERCPU_GET(ring, 0)[0], field_arg) 1327 ring_slice = [(x, y, kern.PERCPU_GET(ring, x)[y]) for y in range(entries_per_cpu)]
|
| H A D | memory.py | 2374 mcs = kern.PERCPU_GET('lck_mcs', cpu).mcs_mtx 2387 mcs = addressof(kern.PERCPU_GET('lck_mcs', idx & 0x3fff).mcs_mtx) 2404 mcs = addressof(kern.PERCPU_GET('lck_mcs', idx & 0x3fff).mcs_mtx)
|
| /xnu-10002.61.3/osfmk/vm/ |
| H A D | vm_resident.c | 3163 if ((mem = *PERCPU_GET(free_pages))) { in vm_page_grab_options() 3302 if ((mem = *PERCPU_GET(free_pages))) { in vm_page_grab_options() 3317 color = *PERCPU_GET(start_color); in vm_page_grab_options()
|