| /xnu-10063.101.15/osfmk/i386/ |
| H A D | cpu_threads.h | 40 #define cpu_to_core_lapic(cpu) (cpu_to_lapic[cpu] & ~CPU_THREAD_MASK) argument 41 #define cpu_to_core_cpu(cpu) (lapic_to_cpu[cpu_to_core_lapic(cpu)]) argument 42 #define cpu_to_logical_cpu(cpu) (cpu_to_lapic[cpu] & CPU_THREAD_MASK) argument 43 #define cpu_is_core_cpu(cpu) (cpu_to_logical_cpu(cpu) == 0) argument 45 #define _cpu_to_lcpu(cpu) (&cpu_datap(cpu)->lcpu) argument 46 #define _cpu_to_core(cpu) (_cpu_to_lcpu(cpu)->core) argument 47 #define _cpu_to_package(cpu) (_cpu_to_core(cpu)->package) argument 49 #define cpu_to_lcpu(cpu) ((cpu_datap(cpu) != NULL) ? _cpu_to_lcpu(cpu) : NULL) argument 50 #define cpu_to_core(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->core : NUL… argument 51 #define cpu_to_die(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->die : NULL) argument [all …]
|
| H A D | mp.c | 108 void i386_cpu_IPI(int cpu); 115 static boolean_t cpu_signal_pending(int cpu, mp_event_t event); 155 volatile uint16_t cpu; member 157 .cpu = KDP_XCPU_NONE 186 void i386_send_NMI(int cpu); 673 cpu_interrupt(int cpu) in cpu_interrupt() argument 678 && pmCPUExitIdle(cpu_datap(cpu))) { in cpu_interrupt() 679 i386_cpu_IPI(cpu); in cpu_interrupt() 683 KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_SCHED, MACH_REMOTE_AST), cpu, did_IPI, 0, 0, 0); in cpu_interrupt() 690 cpu_NMI_interrupt(int cpu) in cpu_NMI_interrupt() argument [all …]
|
| H A D | pmCPU.h | 98 void (*resyncDeadlines)(int cpu); 100 x86_lcpu_t *(*GetLCPU)(int cpu); 101 x86_core_t *(*GetCore)(int cpu); 102 x86_die_t *(*GetDie)(int cpu); 103 x86_pkg_t *(*GetPackage)(int cpu); 110 boolean_t (*GetHibernate)(int cpu); 114 void (*pmSendIPI)(int cpu); 117 uint32_t (*timerQueueMigrate)(int cpu); 133 uint64_t pmCPUGetDeadline(struct cpu_data *cpu); 134 uint64_t pmCPUSetDeadline(struct cpu_data *cpu, uint64_t deadline); [all …]
|
| H A D | mp_native.c | 39 void i386_cpu_IPI(int cpu); 43 void i386_send_NMI(int cpu); 89 i386_send_NMI(int cpu) in i386_send_NMI() argument 94 i386_cpu_IPI(cpu); in i386_send_NMI() 96 lapic_send_nmi(cpu); in i386_send_NMI() 114 i386_cpu_IPI(int cpu) in i386_cpu_IPI() argument 117 if (cpu_datap(cpu)->cpu_signals & 6) { /* (BRINGUP) */ in i386_cpu_IPI() 118 …i386_cpu_IPI: sending enter debugger signal (%08X) to cpu %d\n", cpu_datap(cpu)->cpu_signals, cpu); in i386_cpu_IPI() 122 lapic_send_ipi(cpu, LAPIC_VECTOR(INTERPROCESSOR)); in i386_cpu_IPI()
|
| H A D | lapic.c | 68 lapic_cpu_map(int apic_id, int cpu) in lapic_cpu_map() argument 71 assert(cpu < MAX_CPUS); in lapic_cpu_map() 72 cpu_to_lapic[cpu] = apic_id; in lapic_cpu_map() 73 lapic_to_cpu[apic_id] = cpu; in lapic_cpu_map() 84 ml_get_apicid(uint32_t cpu) in ml_get_apicid() argument 86 if (cpu >= (uint32_t)MAX_CPUS) { in ml_get_apicid() 90 return (uint32_t)cpu_to_lapic[cpu]; in ml_get_apicid()
|
| H A D | mp.h | 90 extern void cpu_interrupt(int cpu); 160 cpu_to_cpumask(cpu_t cpu) in cpu_to_cpumask() argument 162 return (cpu < MAX_CPUS) ? (1ULL << cpu) : 0; in cpu_to_cpumask() 170 extern void mp_cpus_call_cpu_init(int cpu); 219 extern void cpu_PM_interrupt(int cpu); 226 int cpu; member 249 eventp->cpu = _cpu; \ 259 #define DBGLOG_CPU_INIT(cpu) { \ argument 260 cpu_signal_event_log_t **sig_logpp = &cpu_signal[cpu]; \ 261 cpu_signal_event_log_t **hdl_logpp = &cpu_handle[cpu]; \ [all …]
|
| H A D | cpu_threads.c | 337 x86_lcpu_init(int cpu) in x86_lcpu_init() argument 343 cpup = cpu_datap(cpu); in x86_lcpu_init() 347 lcpu->cpu = cpup; in x86_lcpu_init() 354 lcpu->cpu_num = cpu; in x86_lcpu_init() 355 lcpu->lnum = cpu; in x86_lcpu_init() 364 x86_core_alloc(int cpu) in x86_core_alloc() argument 369 cpup = cpu_datap(cpu); in x86_core_alloc() 404 x86_package_find(int cpu) in x86_package_find() argument 410 cpup = cpu_datap(cpu); in x86_package_find() 426 x86_die_find(int cpu) in x86_die_find() argument [all …]
|
| H A D | cpu.c | 119 int cpu) in cpu_start() argument 123 if (cpu == cpu_number()) { in cpu_start() 133 ret = intel_startCPU_fast(cpu); in cpu_start() 139 ret = intel_startCPU(cpu); in cpu_start() 143 kprintf("cpu: cpu_start(%d) returning failure!\n", cpu); in cpu_start() 151 int cpu) in cpu_exit_wait() argument 153 cpu_data_t *cdp = cpu_datap(cpu); in cpu_exit_wait() 174 panic("cpu_exit_wait(%d) timeout", cpu); in cpu_exit_wait() 208 int cpu) in cpu_to_processor() argument 210 return cpu_datap(cpu)->cpu_processor; in cpu_to_processor()
|
| H A D | pmCPU.c | 314 pmGetLogicalCPU(int cpu) in pmGetLogicalCPU() argument 316 return cpu_to_lcpu(cpu); in pmGetLogicalCPU() 328 pmGetCore(int cpu) in pmGetCore() argument 330 return cpu_to_core(cpu); in pmGetCore() 342 pmGetDie(int cpu) in pmGetDie() argument 344 return cpu_to_die(cpu); in pmGetDie() 356 pmGetPackage(int cpu) in pmGetPackage() argument 358 return cpu_to_package(cpu); in pmGetPackage() 386 pmCPUGetDeadline(cpu_data_t *cpu) in pmCPUGetDeadline() argument 393 deadline = (*pmDispatch->GetDeadline)(&cpu->lcpu); in pmCPUGetDeadline() [all …]
|
| H A D | hpet.c | 109 hpet_request(uint32_t cpu) in hpet_request() argument 126 if (cpu >= real_ncpus) { in hpet_request() 128 lcpu = cpu_to_lcpu(cpu); in hpet_request() 142 rc = (*hpet_req)(ml_get_apicid(cpu), hpet_arg, &hpetReq); in hpet_request() 148 lcpu = cpu_to_lcpu(cpu); in hpet_request() 322 ml_hpet_cfg(uint32_t cpu, uint32_t hpetVect) in ml_hpet_cfg() argument 331 if (cpu > 1) { in ml_hpet_cfg() 332 panic("ml_hpet_cfg: invalid cpu = %d", cpu); in ml_hpet_cfg() 335 lcpu = cpu_to_lcpu(cpu); in ml_hpet_cfg() 349 hpetVaddr = (uint64_t *)(((uintptr_t)&(((hpetReg_t *)hpetArea)->TIM1_CONF)) + (cpu << 5)); in ml_hpet_cfg() [all …]
|
| /xnu-10063.101.15/osfmk/i386/vmx/ |
| H A D | vmx_cpu.c | 169 vmx_cpu_t *cpu = ¤t_cpu_datap()->cpu_vmx; in vmx_on() local 174 cpu_number(), cpu->specs.vmx_on); in vmx_on() 176 assert(cpu->specs.vmx_present); in vmx_on() 178 if (NULL == cpu->vmxon_region) { in vmx_on() 181 vmxon_region_paddr = vmx_paddr(cpu->vmxon_region); in vmx_on() 186 if (FALSE == cpu->specs.vmx_on) { in vmx_on() 187 assert(vmx_is_cr0_valid(&cpu->specs)); in vmx_on() 188 assert(vmx_is_cr4_valid(&cpu->specs)); in vmx_on() 196 cpu->specs.vmx_on = TRUE; in vmx_on() 199 cpu_number(), cpu->specs.vmx_on); in vmx_on() [all …]
|
| /xnu-10063.101.15/iokit/Kernel/arm/ |
| H A D | AppleARMSMP.cpp | 123 IOService *cpu = IOService::waitForMatchingService(matching, UINT64_MAX); in register_aic_handlers() local 126 OSArray *irqs = (OSArray *) cpu->getProperty(gIOInterruptSpecifiersKey); in register_aic_handlers() 135 if (cpu->registerInterrupt(0, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers() 136 cpu->enableInterrupt(0) != kIOReturnSuccess || in register_aic_handlers() 137 cpu->registerInterrupt(2, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers() 138 cpu->enableInterrupt(2) != kIOReturnSuccess) { in register_aic_handlers() 150 …if (cpu->registerInterrupt(1, NULL, (IOInterruptAction)(void (*)(void))pmi_handler, NULL) != kIORe… in register_aic_handlers() 151 cpu->enableInterrupt(1) != kIOReturnSuccess) { in register_aic_handlers() 203 for (unsigned int cpu = 0; cpu < topology_info->num_cpus; cpu++) { in cpu_boot_thread() local 204 const ml_topology_cpu *cpu_info = &topology_info->cpus[cpu]; in cpu_boot_thread() [all …]
|
| /xnu-10063.101.15/osfmk/kern/ |
| H A D | kpc_thread.c | 125 cpu_data_t *cpu = NULL; in kpc_update_thread_counters() local 127 cpu = current_cpu_datap(); in kpc_update_thread_counters() 131 NULL, cpu->cpu_kpc_buf[1] ); in kpc_update_thread_counters() 136 thread->kpc_buf[i] += cpu->cpu_kpc_buf[1][i] - cpu->cpu_kpc_buf[0][i]; in kpc_update_thread_counters() 147 tmp = cpu->cpu_kpc_buf[1]; in kpc_update_thread_counters() 148 cpu->cpu_kpc_buf[1] = cpu->cpu_kpc_buf[0]; in kpc_update_thread_counters() 149 cpu->cpu_kpc_buf[0] = tmp; in kpc_update_thread_counters()
|
| H A D | monotonic.h | 89 void mt_cpu_idle(cpu_data_t *cpu); 90 void mt_cpu_run(cpu_data_t *cpu); 95 void mt_cpu_down(cpu_data_t *cpu); 96 void mt_cpu_up(cpu_data_t *cpu); 111 uint64_t mt_cpu_update_count(cpu_data_t *cpu, unsigned int ctr);
|
| H A D | kpc.h | 66 #define FIXED_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr)]) argument 68 #define CONFIGURABLE_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr) + kpc_fixed_c… argument 72 #define FIXED_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr)]) argument 74 #define CONFIGURABLE_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr) + kpc_fixed_c… argument
|
| /xnu-10063.101.15/osfmk/arm/ |
| H A D | model_dep.c | 382 for (int cpu = 0; cpu <= max_cpu; cpu++) { in panic_display_last_pc_lr() local 383 cpu_data_t *current_cpu_datap = cpu_datap(cpu); in panic_display_last_pc_lr() 394 …g_append_noflush("CORE %u is the one that panicked. Check the full backtrace for details.\n", cpu); in panic_display_last_pc_lr() 398 paniclog_append_noflush("CORE %u: PC=0x%016llx, LR=0x%016llx, FP=0x%016llx\n", cpu, in panic_display_last_pc_lr() 957 int cpu; in DebuggerXCallEnter() local 994 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter() 995 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerXCallEnter() 1032 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter() 1033 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerXCallEnter() 1041 if (processor_array[cpu]->state <= PROCESSOR_PENDING_OFFLINE) { in DebuggerXCallEnter() [all …]
|
| H A D | cpu_common.c | 80 cpu_datap(int cpu) in cpu_datap() argument 82 assert(cpu <= ml_get_max_cpu_number()); in cpu_datap() 83 return CpuDataEntries[cpu].cpu_data_vaddr; in cpu_datap() 257 int cpu; in cpu_broadcast_xcall_internal() local 274 for (cpu = 0; cpu < max_cpu; cpu++) { in cpu_broadcast_xcall_internal() 275 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in cpu_broadcast_xcall_internal() 643 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in cpu_exit_wait() local 644 assert(cpu && cpu->cpu_IMPL_regs); in cpu_exit_wait() 645 volatile uint64_t *cpu_sts = (void *)(cpu->cpu_IMPL_regs + CPU_PIO_CPU_STS_OFFSET); in cpu_exit_wait() 686 cpu_can_exit(__unused int cpu) in cpu_can_exit() argument [all …]
|
| /xnu-10063.101.15/SETUP/config/ |
| H A D | doconf | 171 set cpu=`echo $MACHINE | tr A-Z a-z` 176 …ster_file (${MASTER}.${cpu}.${SOC_CONFIG}.${PLATFORM} ${MASTER}.${cpu}.${SOC_CONFIG} ${MASTER}.${c… 192 echo "[ generating $SYSID from $MASTER_DIR/MASTER{,.$cpu}{,.local} ]" 199 …unifdef -t -DPLATFORM_${PLATFORM} -DCPU_$cpu -DSOC_CONFIG_${SOC_CONFIG} -DSYS_${SYS} $DEF_AVP $DEF… 270 echo "${prog}: ${$SYSID}: no such configuration in $MASTER_DIR/MASTER{,.$cpu}"
|
| /xnu-10063.101.15/osfmk/arm64/ |
| H A D | machine_routines.c | 734 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_up_update_counts() local 736 os_atomic_inc(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_up_update_counts() 784 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_down_update_counts() local 786 os_atomic_dec(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_down_update_counts() 1058 ml_topology_cpu_t *cpu = &topology_info.cpus[topology_info.num_cpus]; in ml_parse_cpu_topology() local 1060 cpu->cpu_id = topology_info.num_cpus++; in ml_parse_cpu_topology() 1061 assert(cpu->cpu_id < MAX_CPUS); in ml_parse_cpu_topology() 1062 topology_info.max_cpu_id = MAX(topology_info.max_cpu_id, cpu->cpu_id); in ml_parse_cpu_topology() 1064 cpu->die_id = 0; in ml_parse_cpu_topology() 1067 cpu->phys_id = (uint32_t)ml_readprop(child, "reg", ML_READPROP_MANDATORY); in ml_parse_cpu_topology() [all …]
|
| H A D | monotonic_arm64.c | 275 core_idle(__unused cpu_data_t *cpu) in core_idle() argument 277 assert(cpu != NULL); in core_idle() 1236 mt_cpu_idle(cpu_data_t *cpu) in mt_cpu_idle() argument 1238 core_idle(cpu); in mt_cpu_idle() 1242 mt_cpu_run(cpu_data_t *cpu) in mt_cpu_run() argument 1246 assert(cpu != NULL); in mt_cpu_run() 1249 mtc = &cpu->cpu_monotonic; in mt_cpu_run() 1262 mt_cpu_down(cpu_data_t *cpu) in mt_cpu_down() argument 1264 mt_cpu_idle(cpu); in mt_cpu_down() 1268 mt_cpu_up(cpu_data_t *cpu) in mt_cpu_up() argument [all …]
|
| /xnu-10063.101.15/bsd/kern/ |
| H A D | kdebug_common.c | 256 struct kd_buffer *kd_data_page, int cpu) in allocate_storage_unit() argument 267 kdbp = &kd_data_page->kdb_info[cpu]; in allocate_storage_unit() 392 release_storage_unit(struct kd_control *kd_ctrl_page, struct kd_buffer *kd_data_page, int cpu, uint… in release_storage_unit() argument 398 kdbp = &kd_data_page->kdb_info[cpu]; in release_storage_unit() 465 int cpu; in kernel_debug_write() local 476 if (kd_rec.cpu == -1) { in kernel_debug_write() 477 cpu = cpu_number(); in kernel_debug_write() 479 cpu = kd_rec.cpu; in kernel_debug_write() 482 kdbp = &kd_data_page->kdb_info[cpu]; in kernel_debug_write() 517 if (allocate_storage_unit(kd_ctrl_page, kd_data_page, cpu) == false) { in kernel_debug_write() [all …]
|
| /xnu-10063.101.15/osfmk/x86_64/ |
| H A D | monotonic_x86_64.c | 208 core_down(cpu_data_t *cpu) in core_down() argument 214 struct mt_cpu *mtc = &cpu->cpu_monotonic; in core_down() 222 core_up(cpu_data_t *cpu) in core_up() argument 232 mtc = &cpu->cpu_monotonic; in core_up() 242 mt_cpu_down(cpu_data_t *cpu) in mt_cpu_down() argument 244 core_down(cpu); in mt_cpu_down() 248 mt_cpu_up(cpu_data_t *cpu) in mt_cpu_up() argument 252 core_up(cpu); in mt_cpu_up() 261 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis() local 262 npmis += cpu->cpu_monotonic.mtc_npmis; in mt_count_pmis()
|
| /xnu-10063.101.15/pexpert/arm/ |
| H A D | pe_identify_machine.c | 38 DTEntry cpus, cpu; in pe_identify_machine() local 63 while (kSuccess == SecureDTIterateEntries(&iter, &cpu)) { in pe_identify_machine() 64 if ((kSuccess != SecureDTGetProperty(cpu, "state", &value, &size)) || in pe_identify_machine() 70 if (SecureDTGetProperty(cpu, "timebase-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 85 if (SecureDTGetProperty(cpu, "bus-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 102 if (SecureDTGetProperty(cpu, "memory-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 113 if (SecureDTGetProperty(cpu, "peripheral-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 124 if (SecureDTGetProperty(cpu, "fixed-frequency", &value, &size) == kSuccess) { in pe_identify_machine() 132 if (SecureDTGetProperty(cpu, "clock-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
|
| /xnu-10063.101.15/tools/lldbmacros/kmemory/ |
| H A D | kmem.py | 219 def PERCPU_BASE(self, cpu): argument 263 def PERCPU_BASE(self, cpu): argument 264 cpu_data = self.arm64_CpuDataEntries.chkGetChildAtIndex(cpu) 299 def PERCPU_BASE(self, cpu): argument 300 cpu_data = self.intel_cpu_data.chkGetChildAtIndex(cpu) 323 def __getitem__(self, cpu): argument 324 if cpu in self.kmem.pcpus: 326 addr = sbv.GetLoadAddress() + self.kmem.PERCPU_BASE(cpu) 345 (cpu, sbv.chkCreateValueFromAddress(name, addr + kmem.PERCPU_BASE(cpu), ty)) 346 for cpu in kmem.pcpus
|
| /xnu-10063.101.15/tests/cpu_counters/ |
| H A D | kpc_tests.c | 411 struct cpu { struct 447 struct cpu *cpus = calloc(mch.ncpus, sizeof(*cpus)); 481 struct cpu *cpu = &cpus[tp->cpuid]; 483 if (cpu->prev_count != 0) { 484 uint64_t delta = count - cpu->prev_count; 486 if (skid > cpu->max_skid) { 487 cpu->max_skid = skid; 490 cpu->prev_count = count; 497 cpu->pmi_timeslices[(unsigned int)slice] += 1; 513 struct cpu *cpu = &cpus[tp->cpuid]; variable [all …]
|