| /xnu-11417.121.6/osfmk/i386/ |
| H A D | cpu_threads.h | 40 #define cpu_to_core_lapic(cpu) (cpu_to_lapic[cpu] & ~CPU_THREAD_MASK) argument 41 #define cpu_to_core_cpu(cpu) (lapic_to_cpu[cpu_to_core_lapic(cpu)]) argument 42 #define cpu_to_logical_cpu(cpu) (cpu_to_lapic[cpu] & CPU_THREAD_MASK) argument 43 #define cpu_is_core_cpu(cpu) (cpu_to_logical_cpu(cpu) == 0) argument 45 #define _cpu_to_lcpu(cpu) (&cpu_datap(cpu)->lcpu) argument 46 #define _cpu_to_core(cpu) (_cpu_to_lcpu(cpu)->core) argument 47 #define _cpu_to_package(cpu) (_cpu_to_core(cpu)->package) argument 49 #define cpu_to_lcpu(cpu) ((cpu_datap(cpu) != NULL) ? _cpu_to_lcpu(cpu) : NULL) argument 50 #define cpu_to_core(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->core : NUL… argument 51 #define cpu_to_die(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->die : NULL) argument [all …]
|
| H A D | mp.c | 109 void i386_cpu_IPI(int cpu); 116 static boolean_t cpu_signal_pending(int cpu, mp_event_t event); 156 volatile uint16_t cpu; member 158 .cpu = KDP_XCPU_NONE 187 void i386_send_NMI(int cpu); 674 cpu_interrupt(int cpu) in cpu_interrupt() argument 679 && pmCPUExitIdle(cpu_datap(cpu))) { in cpu_interrupt() 680 i386_cpu_IPI(cpu); in cpu_interrupt() 684 KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_SCHED, MACH_REMOTE_AST), cpu, did_IPI, 0, 0, 0); in cpu_interrupt() 691 cpu_NMI_interrupt(int cpu) in cpu_NMI_interrupt() argument [all …]
|
| H A D | pmCPU.h | 98 void (*resyncDeadlines)(int cpu); 100 x86_lcpu_t *(*GetLCPU)(int cpu); 101 x86_core_t *(*GetCore)(int cpu); 102 x86_die_t *(*GetDie)(int cpu); 103 x86_pkg_t *(*GetPackage)(int cpu); 110 boolean_t (*GetHibernate)(int cpu); 114 void (*pmSendIPI)(int cpu); 117 uint32_t (*timerQueueMigrate)(int cpu); 133 uint64_t pmCPUGetDeadline(struct cpu_data *cpu); 134 uint64_t pmCPUSetDeadline(struct cpu_data *cpu, uint64_t deadline); [all …]
|
| H A D | mp_native.c | 39 void i386_cpu_IPI(int cpu); 43 void i386_send_NMI(int cpu); 89 i386_send_NMI(int cpu) in i386_send_NMI() argument 94 i386_cpu_IPI(cpu); in i386_send_NMI() 96 lapic_send_nmi(cpu); in i386_send_NMI() 114 i386_cpu_IPI(int cpu) in i386_cpu_IPI() argument 117 if (cpu_datap(cpu)->cpu_signals & 6) { /* (BRINGUP) */ in i386_cpu_IPI() 118 …i386_cpu_IPI: sending enter debugger signal (%08X) to cpu %d\n", cpu_datap(cpu)->cpu_signals, cpu); in i386_cpu_IPI() 122 lapic_send_ipi(cpu, LAPIC_VECTOR(INTERPROCESSOR)); in i386_cpu_IPI()
|
| H A D | lapic.c | 68 lapic_cpu_map(int apic_id, int cpu) in lapic_cpu_map() argument 71 assert(cpu < MAX_CPUS); in lapic_cpu_map() 72 cpu_to_lapic[cpu] = apic_id; in lapic_cpu_map() 73 lapic_to_cpu[apic_id] = cpu; in lapic_cpu_map() 84 ml_get_apicid(uint32_t cpu) in ml_get_apicid() argument 86 if (cpu >= (uint32_t)MAX_CPUS) { in ml_get_apicid() 90 return (uint32_t)cpu_to_lapic[cpu]; in ml_get_apicid()
|
| H A D | mp.h | 93 extern void cpu_interrupt(int cpu); 163 cpu_to_cpumask(cpu_t cpu) in cpu_to_cpumask() argument 165 return (cpu < MAX_CPUS) ? (1ULL << cpu) : 0; in cpu_to_cpumask() 173 extern void mp_cpus_call_cpu_init(int cpu); 222 extern void cpu_PM_interrupt(int cpu); 229 int cpu; member 252 eventp->cpu = _cpu; \ 262 #define DBGLOG_CPU_INIT(cpu) { \ argument 263 cpu_signal_event_log_t **sig_logpp = &cpu_signal[cpu]; \ 264 cpu_signal_event_log_t **hdl_logpp = &cpu_handle[cpu]; \ [all …]
|
| H A D | cpu_threads.c | 337 x86_lcpu_init(int cpu) in x86_lcpu_init() argument 343 cpup = cpu_datap(cpu); in x86_lcpu_init() 347 lcpu->cpu = cpup; in x86_lcpu_init() 354 lcpu->cpu_num = cpu; in x86_lcpu_init() 355 lcpu->lnum = cpu; in x86_lcpu_init() 364 x86_core_alloc(int cpu) in x86_core_alloc() argument 369 cpup = cpu_datap(cpu); in x86_core_alloc() 404 x86_package_find(int cpu) in x86_package_find() argument 410 cpup = cpu_datap(cpu); in x86_package_find() 426 x86_die_find(int cpu) in x86_die_find() argument [all …]
|
| H A D | cpu.c | 119 int cpu) in cpu_start() argument 123 if (cpu == cpu_number()) { in cpu_start() 133 ret = intel_startCPU_fast(cpu); in cpu_start() 139 ret = intel_startCPU(cpu); in cpu_start() 143 panic("cpu_start(%d) failed: %d\n", cpu, ret); in cpu_start() 149 int cpu) in cpu_exit_wait() argument 151 cpu_data_t *cdp = cpu_datap(cpu); in cpu_exit_wait() 172 panic("cpu_exit_wait(%d) timeout", cpu); in cpu_exit_wait() 206 int cpu) in cpu_to_processor() argument 208 return cpu_datap(cpu)->cpu_processor; in cpu_to_processor()
|
| H A D | pmCPU.c | 314 pmGetLogicalCPU(int cpu) in pmGetLogicalCPU() argument 316 return cpu_to_lcpu(cpu); in pmGetLogicalCPU() 328 pmGetCore(int cpu) in pmGetCore() argument 330 return cpu_to_core(cpu); in pmGetCore() 342 pmGetDie(int cpu) in pmGetDie() argument 344 return cpu_to_die(cpu); in pmGetDie() 356 pmGetPackage(int cpu) in pmGetPackage() argument 358 return cpu_to_package(cpu); in pmGetPackage() 386 pmCPUGetDeadline(cpu_data_t *cpu) in pmCPUGetDeadline() argument 393 deadline = (*pmDispatch->GetDeadline)(&cpu->lcpu); in pmCPUGetDeadline() [all …]
|
| H A D | hpet.c | 109 hpet_request(uint32_t cpu) in hpet_request() argument 126 if (cpu >= real_ncpus) { in hpet_request() 128 lcpu = cpu_to_lcpu(cpu); in hpet_request() 142 rc = (*hpet_req)(ml_get_apicid(cpu), hpet_arg, &hpetReq); in hpet_request() 148 lcpu = cpu_to_lcpu(cpu); in hpet_request() 322 ml_hpet_cfg(uint32_t cpu, uint32_t hpetVect) in ml_hpet_cfg() argument 331 if (cpu > 1) { in ml_hpet_cfg() 332 panic("ml_hpet_cfg: invalid cpu = %d", cpu); in ml_hpet_cfg() 335 lcpu = cpu_to_lcpu(cpu); in ml_hpet_cfg() 349 hpetVaddr = (uint64_t *)(((uintptr_t)&(((hpetReg_t *)hpetArea)->TIM1_CONF)) + (cpu << 5)); in ml_hpet_cfg() [all …]
|
| /xnu-11417.121.6/osfmk/i386/vmx/ |
| H A D | vmx_cpu.c | 169 vmx_cpu_t *cpu = ¤t_cpu_datap()->cpu_vmx; in vmx_on() local 174 cpu_number(), cpu->specs.vmx_on); in vmx_on() 176 assert(cpu->specs.vmx_present); in vmx_on() 178 if (NULL == cpu->vmxon_region) { in vmx_on() 181 vmxon_region_paddr = vmx_paddr(cpu->vmxon_region); in vmx_on() 186 if (FALSE == cpu->specs.vmx_on) { in vmx_on() 187 assert(vmx_is_cr0_valid(&cpu->specs)); in vmx_on() 188 assert(vmx_is_cr4_valid(&cpu->specs)); in vmx_on() 196 cpu->specs.vmx_on = TRUE; in vmx_on() 199 cpu_number(), cpu->specs.vmx_on); in vmx_on() [all …]
|
| /xnu-11417.121.6/iokit/Kernel/arm/ |
| H A D | AppleARMSMP.cpp | 122 IOService *cpu = IOService::waitForMatchingService(matching, UINT64_MAX); in register_aic_handlers() local 125 OSArray *irqs = (OSArray *) cpu->getProperty(gIOInterruptSpecifiersKey); in register_aic_handlers() 134 if (cpu->registerInterrupt(0, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers() 135 cpu->enableInterrupt(0) != kIOReturnSuccess || in register_aic_handlers() 136 cpu->registerInterrupt(2, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers() 137 cpu->enableInterrupt(2) != kIOReturnSuccess) { in register_aic_handlers() 149 …if (cpu->registerInterrupt(1, NULL, (IOInterruptAction)(void (*)(void))pmi_handler, NULL) != kIORe… in register_aic_handlers() 150 cpu->enableInterrupt(1) != kIOReturnSuccess) { in register_aic_handlers() 179 for (unsigned int cpu = 0; cpu < topology_info->num_cpus; cpu++) { in cpu_boot_thread() local 180 const ml_topology_cpu *cpu_info = &topology_info->cpus[cpu]; in cpu_boot_thread() [all …]
|
| /xnu-11417.121.6/osfmk/kern/ |
| H A D | kpc_thread.c | 125 cpu_data_t *cpu = NULL; in kpc_update_thread_counters() local 127 cpu = current_cpu_datap(); in kpc_update_thread_counters() 131 NULL, cpu->cpu_kpc_buf[1] ); in kpc_update_thread_counters() 136 thread->kpc_buf[i] += cpu->cpu_kpc_buf[1][i] - cpu->cpu_kpc_buf[0][i]; in kpc_update_thread_counters() 147 tmp = cpu->cpu_kpc_buf[1]; in kpc_update_thread_counters() 148 cpu->cpu_kpc_buf[1] = cpu->cpu_kpc_buf[0]; in kpc_update_thread_counters() 149 cpu->cpu_kpc_buf[0] = tmp; in kpc_update_thread_counters()
|
| H A D | monotonic.h | 89 void mt_cpu_idle(cpu_data_t *cpu); 90 void mt_cpu_run(cpu_data_t *cpu); 95 void mt_cpu_down(cpu_data_t *cpu); 96 void mt_cpu_up(cpu_data_t *cpu); 111 uint64_t mt_cpu_update_count(cpu_data_t *cpu, unsigned int ctr);
|
| H A D | kpc.h | 66 #define FIXED_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr)]) argument 68 #define CONFIGURABLE_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr) + kpc_fixed_c… argument 72 #define FIXED_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr)]) argument 74 #define CONFIGURABLE_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr) + kpc_fixed_c… argument
|
| /xnu-11417.121.6/osfmk/arm/ |
| H A D | model_dep.c | 385 for (int cpu = 0; cpu <= max_cpu; cpu++) { in panic_display_last_pc_lr() local 386 cpu_data_t *current_cpu_datap = cpu_datap(cpu); in panic_display_last_pc_lr() 397 …g_append_noflush("CORE %u is the one that panicked. Check the full backtrace for details.\n", cpu); in panic_display_last_pc_lr() 401 paniclog_append_noflush("CORE %u: PC=0x%016llx, LR=0x%016llx, FP=0x%016llx\n", cpu, in panic_display_last_pc_lr() 991 int cpu, xnu_saved_registers_t *regp) in DebuggerPatchupThreadState() argument 997 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerPatchupThreadState() 1042 int cpu; in DebuggerXCallEnter() local 1081 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter() 1082 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerXCallEnter() 1127 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter() [all …]
|
| /xnu-11417.121.6/libkern/os/ |
| H A D | refcnt.c | 764 zpercpu_foreach_cpu(cpu) { in zpercpu_foreach_cpu() argument 765 os_atomic_init(zpercpu_get_cpu(rc, cpu), in zpercpu_foreach_cpu() 766 OS_PCPU_REF_LIVE + (cpu ? 0 : OS_PCPU_REF_INC)); in zpercpu_foreach_cpu() 787 for (int cpu = zpercpu_count(); cpu-- > 1;) { variable 788 n |= os_atomic_load_wide(zpercpu_get_cpu(rc, cpu), relaxed); 816 int cpu; in __os_pcpu_ref_delta() local 818 cpu = cpu_number(); in __os_pcpu_ref_delta() 819 rcp = zpercpu_get_cpu(rc, cpu); in __os_pcpu_ref_delta() 826 *cpup = cpu; in __os_pcpu_ref_delta() 836 __os_pcpu_ref_retain_slow(__os_pcpu_ref_t rc, int cpu, uint64_t v) in __os_pcpu_ref_retain_slow() argument [all …]
|
| /xnu-11417.121.6/SETUP/config/ |
| H A D | doconf | 178 set cpu=`echo $MACHINE | tr A-Z a-z` 183 …ster_file (${MASTER}.${cpu}.${SOC_CONFIG}.${PLATFORM} ${MASTER}.${cpu}.${SOC_CONFIG} ${MASTER}.${c… 199 echo "[ generating $SYSID from $MASTER_DIR/MASTER{,.$cpu}{,.local} ]" 206 …unifdef -t -DPLATFORM_${PLATFORM} -DCPU_$cpu -DSOC_CONFIG_${SOC_CONFIG} -DSYS_${SYS} $DEF_AVP $DEF… 277 echo "${prog}: ${$SYSID}: no such configuration in $MASTER_DIR/MASTER{,.$cpu}"
|
| /xnu-11417.121.6/osfmk/arm64/ |
| H A D | machine_routines.c | 772 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_up_update_counts() local 774 os_atomic_inc(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_up_update_counts() 873 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_down_update_counts() local 875 os_atomic_dec(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_down_update_counts() 1181 ml_topology_cpu_t *cpu = &topology_info.cpus[topology_info.num_cpus]; in ml_parse_cpu_topology() local 1183 cpu->cpu_id = topology_info.num_cpus++; in ml_parse_cpu_topology() 1184 assert(cpu->cpu_id < MAX_CPUS); in ml_parse_cpu_topology() 1185 topology_info.max_cpu_id = MAX(topology_info.max_cpu_id, cpu->cpu_id); in ml_parse_cpu_topology() 1187 cpu->die_id = (int)ml_readprop(child, "die-id", 0); in ml_parse_cpu_topology() 1188 topology_info.max_die_id = MAX(topology_info.max_die_id, cpu->die_id); in ml_parse_cpu_topology() [all …]
|
| H A D | monotonic_arm64.c | 287 core_idle(__unused cpu_data_t *cpu) in core_idle() argument 289 assert(cpu != NULL); in core_idle() 545 unsigned int cpu = cpu_number(); in _broadcast_block_trampoline() local 546 unsigned int cluster = topo->cpus[cpu].cluster_id; in _broadcast_block_trampoline() 547 if (topo->clusters[cluster].first_cpu_id == cpu) { in _broadcast_block_trampoline() 548 cb(topo->cpus[cpu].cluster_id); in _broadcast_block_trampoline() 1328 mt_cpu_idle(cpu_data_t *cpu) in mt_cpu_idle() argument 1330 core_idle(cpu); in mt_cpu_idle() 1334 mt_cpu_run(cpu_data_t *cpu) in mt_cpu_run() argument 1338 assert(cpu != NULL); in mt_cpu_run() [all …]
|
| /xnu-11417.121.6/bsd/kern/ |
| H A D | kdebug_common.c | 257 struct kd_buffer *kd_data_page, int cpu) in allocate_storage_unit() argument 268 kdbp = &kd_data_page->kdb_info[cpu]; in allocate_storage_unit() 393 release_storage_unit(struct kd_control *kd_ctrl_page, struct kd_buffer *kd_data_page, int cpu, uint… in release_storage_unit() argument 399 kdbp = &kd_data_page->kdb_info[cpu]; in release_storage_unit() 466 int cpu; in kernel_debug_write() local 477 if (kd_rec.cpu == -1) { in kernel_debug_write() 478 cpu = cpu_number(); in kernel_debug_write() 480 cpu = kd_rec.cpu; in kernel_debug_write() 483 kdbp = &kd_data_page->kdb_info[cpu]; in kernel_debug_write() 518 if (allocate_storage_unit(kd_ctrl_page, kd_data_page, cpu) == false) { in kernel_debug_write() [all …]
|
| /xnu-11417.121.6/osfmk/x86_64/ |
| H A D | monotonic_x86_64.c | 208 core_down(cpu_data_t *cpu) in core_down() argument 214 struct mt_cpu *mtc = &cpu->cpu_monotonic; in core_down() 222 core_up(cpu_data_t *cpu) in core_up() argument 232 mtc = &cpu->cpu_monotonic; in core_up() 242 mt_cpu_down(cpu_data_t *cpu) in mt_cpu_down() argument 244 core_down(cpu); in mt_cpu_down() 248 mt_cpu_up(cpu_data_t *cpu) in mt_cpu_up() argument 252 core_up(cpu); in mt_cpu_up() 261 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis() local 262 npmis += cpu->cpu_monotonic.mtc_npmis; in mt_count_pmis()
|
| /xnu-11417.121.6/pexpert/arm/ |
| H A D | pe_identify_machine.c | 38 DTEntry cpus, cpu; in pe_identify_machine() local 63 while (kSuccess == SecureDTIterateEntries(&iter, &cpu)) { in pe_identify_machine() 64 if ((kSuccess != SecureDTGetProperty(cpu, "state", &value, &size)) || in pe_identify_machine() 70 if (SecureDTGetProperty(cpu, "timebase-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 85 if (SecureDTGetProperty(cpu, "bus-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 102 if (SecureDTGetProperty(cpu, "memory-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 113 if (SecureDTGetProperty(cpu, "peripheral-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 124 if (SecureDTGetProperty(cpu, "fixed-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 132 if (SecureDTGetProperty(cpu, "clock-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
|
| /xnu-11417.121.6/tools/lldbmacros/kmemory/ |
| H A D | kmem.py | 224 def PERCPU_BASE(self, cpu): argument 268 def PERCPU_BASE(self, cpu): argument 269 cpu_data = self.arm64_CpuDataEntries.chkGetChildAtIndex(cpu) 304 def PERCPU_BASE(self, cpu): argument 305 cpu_data = self.intel_cpu_data.chkGetChildAtIndex(cpu) 328 def __getitem__(self, cpu): argument 329 if cpu in self.kmem.pcpus: 331 addr = sbv.GetLoadAddress() + self.kmem.PERCPU_BASE(cpu) 350 (cpu, sbv.chkCreateValueFromAddress(name, addr + kmem.PERCPU_BASE(cpu), ty)) 351 for cpu in kmem.pcpus
|
| /xnu-11417.121.6/tests/cpu_counters/ |
| H A D | kpc_tests.c | 411 struct cpu { struct 447 struct cpu *cpus = calloc(mch.ncpus, sizeof(*cpus)); 481 struct cpu *cpu = &cpus[tp->cpuid]; 483 if (cpu->prev_count != 0) { 484 uint64_t delta = count - cpu->prev_count; 486 if (skid > cpu->max_skid) { 487 cpu->max_skid = skid; 490 cpu->prev_count = count; 497 cpu->pmi_timeslices[(unsigned int)slice] += 1; 513 struct cpu *cpu = &cpus[tp->cpuid]; variable [all …]
|