Home
last modified time | relevance | path

Searched refs:cpu (Results 1 – 25 of 127) sorted by relevance

123456

/xnu-10063.101.15/osfmk/i386/
H A Dcpu_threads.h40 #define cpu_to_core_lapic(cpu) (cpu_to_lapic[cpu] & ~CPU_THREAD_MASK) argument
41 #define cpu_to_core_cpu(cpu) (lapic_to_cpu[cpu_to_core_lapic(cpu)]) argument
42 #define cpu_to_logical_cpu(cpu) (cpu_to_lapic[cpu] & CPU_THREAD_MASK) argument
43 #define cpu_is_core_cpu(cpu) (cpu_to_logical_cpu(cpu) == 0) argument
45 #define _cpu_to_lcpu(cpu) (&cpu_datap(cpu)->lcpu) argument
46 #define _cpu_to_core(cpu) (_cpu_to_lcpu(cpu)->core) argument
47 #define _cpu_to_package(cpu) (_cpu_to_core(cpu)->package) argument
49 #define cpu_to_lcpu(cpu) ((cpu_datap(cpu) != NULL) ? _cpu_to_lcpu(cpu) : NULL) argument
50 #define cpu_to_core(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->core : NUL… argument
51 #define cpu_to_die(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->die : NULL) argument
[all …]
H A Dmp.c108 void i386_cpu_IPI(int cpu);
115 static boolean_t cpu_signal_pending(int cpu, mp_event_t event);
155 volatile uint16_t cpu; member
157 .cpu = KDP_XCPU_NONE
186 void i386_send_NMI(int cpu);
673 cpu_interrupt(int cpu) in cpu_interrupt() argument
678 && pmCPUExitIdle(cpu_datap(cpu))) { in cpu_interrupt()
679 i386_cpu_IPI(cpu); in cpu_interrupt()
683 KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_SCHED, MACH_REMOTE_AST), cpu, did_IPI, 0, 0, 0); in cpu_interrupt()
690 cpu_NMI_interrupt(int cpu) in cpu_NMI_interrupt() argument
[all …]
H A DpmCPU.h98 void (*resyncDeadlines)(int cpu);
100 x86_lcpu_t *(*GetLCPU)(int cpu);
101 x86_core_t *(*GetCore)(int cpu);
102 x86_die_t *(*GetDie)(int cpu);
103 x86_pkg_t *(*GetPackage)(int cpu);
110 boolean_t (*GetHibernate)(int cpu);
114 void (*pmSendIPI)(int cpu);
117 uint32_t (*timerQueueMigrate)(int cpu);
133 uint64_t pmCPUGetDeadline(struct cpu_data *cpu);
134 uint64_t pmCPUSetDeadline(struct cpu_data *cpu, uint64_t deadline);
[all …]
H A Dmp_native.c39 void i386_cpu_IPI(int cpu);
43 void i386_send_NMI(int cpu);
89 i386_send_NMI(int cpu) in i386_send_NMI() argument
94 i386_cpu_IPI(cpu); in i386_send_NMI()
96 lapic_send_nmi(cpu); in i386_send_NMI()
114 i386_cpu_IPI(int cpu) in i386_cpu_IPI() argument
117 if (cpu_datap(cpu)->cpu_signals & 6) { /* (BRINGUP) */ in i386_cpu_IPI()
118 …i386_cpu_IPI: sending enter debugger signal (%08X) to cpu %d\n", cpu_datap(cpu)->cpu_signals, cpu); in i386_cpu_IPI()
122 lapic_send_ipi(cpu, LAPIC_VECTOR(INTERPROCESSOR)); in i386_cpu_IPI()
H A Dlapic.c68 lapic_cpu_map(int apic_id, int cpu) in lapic_cpu_map() argument
71 assert(cpu < MAX_CPUS); in lapic_cpu_map()
72 cpu_to_lapic[cpu] = apic_id; in lapic_cpu_map()
73 lapic_to_cpu[apic_id] = cpu; in lapic_cpu_map()
84 ml_get_apicid(uint32_t cpu) in ml_get_apicid() argument
86 if (cpu >= (uint32_t)MAX_CPUS) { in ml_get_apicid()
90 return (uint32_t)cpu_to_lapic[cpu]; in ml_get_apicid()
H A Dmp.h90 extern void cpu_interrupt(int cpu);
160 cpu_to_cpumask(cpu_t cpu) in cpu_to_cpumask() argument
162 return (cpu < MAX_CPUS) ? (1ULL << cpu) : 0; in cpu_to_cpumask()
170 extern void mp_cpus_call_cpu_init(int cpu);
219 extern void cpu_PM_interrupt(int cpu);
226 int cpu; member
249 eventp->cpu = _cpu; \
259 #define DBGLOG_CPU_INIT(cpu) { \ argument
260 cpu_signal_event_log_t **sig_logpp = &cpu_signal[cpu]; \
261 cpu_signal_event_log_t **hdl_logpp = &cpu_handle[cpu]; \
[all …]
H A Dcpu_threads.c337 x86_lcpu_init(int cpu) in x86_lcpu_init() argument
343 cpup = cpu_datap(cpu); in x86_lcpu_init()
347 lcpu->cpu = cpup; in x86_lcpu_init()
354 lcpu->cpu_num = cpu; in x86_lcpu_init()
355 lcpu->lnum = cpu; in x86_lcpu_init()
364 x86_core_alloc(int cpu) in x86_core_alloc() argument
369 cpup = cpu_datap(cpu); in x86_core_alloc()
404 x86_package_find(int cpu) in x86_package_find() argument
410 cpup = cpu_datap(cpu); in x86_package_find()
426 x86_die_find(int cpu) in x86_die_find() argument
[all …]
H A Dcpu.c119 int cpu) in cpu_start() argument
123 if (cpu == cpu_number()) { in cpu_start()
133 ret = intel_startCPU_fast(cpu); in cpu_start()
139 ret = intel_startCPU(cpu); in cpu_start()
143 kprintf("cpu: cpu_start(%d) returning failure!\n", cpu); in cpu_start()
151 int cpu) in cpu_exit_wait() argument
153 cpu_data_t *cdp = cpu_datap(cpu); in cpu_exit_wait()
174 panic("cpu_exit_wait(%d) timeout", cpu); in cpu_exit_wait()
208 int cpu) in cpu_to_processor() argument
210 return cpu_datap(cpu)->cpu_processor; in cpu_to_processor()
H A DpmCPU.c314 pmGetLogicalCPU(int cpu) in pmGetLogicalCPU() argument
316 return cpu_to_lcpu(cpu); in pmGetLogicalCPU()
328 pmGetCore(int cpu) in pmGetCore() argument
330 return cpu_to_core(cpu); in pmGetCore()
342 pmGetDie(int cpu) in pmGetDie() argument
344 return cpu_to_die(cpu); in pmGetDie()
356 pmGetPackage(int cpu) in pmGetPackage() argument
358 return cpu_to_package(cpu); in pmGetPackage()
386 pmCPUGetDeadline(cpu_data_t *cpu) in pmCPUGetDeadline() argument
393 deadline = (*pmDispatch->GetDeadline)(&cpu->lcpu); in pmCPUGetDeadline()
[all …]
H A Dhpet.c109 hpet_request(uint32_t cpu) in hpet_request() argument
126 if (cpu >= real_ncpus) { in hpet_request()
128 lcpu = cpu_to_lcpu(cpu); in hpet_request()
142 rc = (*hpet_req)(ml_get_apicid(cpu), hpet_arg, &hpetReq); in hpet_request()
148 lcpu = cpu_to_lcpu(cpu); in hpet_request()
322 ml_hpet_cfg(uint32_t cpu, uint32_t hpetVect) in ml_hpet_cfg() argument
331 if (cpu > 1) { in ml_hpet_cfg()
332 panic("ml_hpet_cfg: invalid cpu = %d", cpu); in ml_hpet_cfg()
335 lcpu = cpu_to_lcpu(cpu); in ml_hpet_cfg()
349 hpetVaddr = (uint64_t *)(((uintptr_t)&(((hpetReg_t *)hpetArea)->TIM1_CONF)) + (cpu << 5)); in ml_hpet_cfg()
[all …]
/xnu-10063.101.15/osfmk/i386/vmx/
H A Dvmx_cpu.c169 vmx_cpu_t *cpu = &current_cpu_datap()->cpu_vmx; in vmx_on() local
174 cpu_number(), cpu->specs.vmx_on); in vmx_on()
176 assert(cpu->specs.vmx_present); in vmx_on()
178 if (NULL == cpu->vmxon_region) { in vmx_on()
181 vmxon_region_paddr = vmx_paddr(cpu->vmxon_region); in vmx_on()
186 if (FALSE == cpu->specs.vmx_on) { in vmx_on()
187 assert(vmx_is_cr0_valid(&cpu->specs)); in vmx_on()
188 assert(vmx_is_cr4_valid(&cpu->specs)); in vmx_on()
196 cpu->specs.vmx_on = TRUE; in vmx_on()
199 cpu_number(), cpu->specs.vmx_on); in vmx_on()
[all …]
/xnu-10063.101.15/iokit/Kernel/arm/
H A DAppleARMSMP.cpp123 IOService *cpu = IOService::waitForMatchingService(matching, UINT64_MAX); in register_aic_handlers() local
126 OSArray *irqs = (OSArray *) cpu->getProperty(gIOInterruptSpecifiersKey); in register_aic_handlers()
135 if (cpu->registerInterrupt(0, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers()
136 cpu->enableInterrupt(0) != kIOReturnSuccess || in register_aic_handlers()
137 cpu->registerInterrupt(2, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers()
138 cpu->enableInterrupt(2) != kIOReturnSuccess) { in register_aic_handlers()
150 …if (cpu->registerInterrupt(1, NULL, (IOInterruptAction)(void (*)(void))pmi_handler, NULL) != kIORe… in register_aic_handlers()
151 cpu->enableInterrupt(1) != kIOReturnSuccess) { in register_aic_handlers()
203 for (unsigned int cpu = 0; cpu < topology_info->num_cpus; cpu++) { in cpu_boot_thread() local
204 const ml_topology_cpu *cpu_info = &topology_info->cpus[cpu]; in cpu_boot_thread()
[all …]
/xnu-10063.101.15/osfmk/kern/
H A Dkpc_thread.c125 cpu_data_t *cpu = NULL; in kpc_update_thread_counters() local
127 cpu = current_cpu_datap(); in kpc_update_thread_counters()
131 NULL, cpu->cpu_kpc_buf[1] ); in kpc_update_thread_counters()
136 thread->kpc_buf[i] += cpu->cpu_kpc_buf[1][i] - cpu->cpu_kpc_buf[0][i]; in kpc_update_thread_counters()
147 tmp = cpu->cpu_kpc_buf[1]; in kpc_update_thread_counters()
148 cpu->cpu_kpc_buf[1] = cpu->cpu_kpc_buf[0]; in kpc_update_thread_counters()
149 cpu->cpu_kpc_buf[0] = tmp; in kpc_update_thread_counters()
H A Dmonotonic.h89 void mt_cpu_idle(cpu_data_t *cpu);
90 void mt_cpu_run(cpu_data_t *cpu);
95 void mt_cpu_down(cpu_data_t *cpu);
96 void mt_cpu_up(cpu_data_t *cpu);
111 uint64_t mt_cpu_update_count(cpu_data_t *cpu, unsigned int ctr);
H A Dkpc.h66 #define FIXED_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr)]) argument
68 #define CONFIGURABLE_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr) + kpc_fixed_c… argument
72 #define FIXED_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr)]) argument
74 #define CONFIGURABLE_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr) + kpc_fixed_c… argument
/xnu-10063.101.15/osfmk/arm/
H A Dmodel_dep.c382 for (int cpu = 0; cpu <= max_cpu; cpu++) { in panic_display_last_pc_lr() local
383 cpu_data_t *current_cpu_datap = cpu_datap(cpu); in panic_display_last_pc_lr()
394 …g_append_noflush("CORE %u is the one that panicked. Check the full backtrace for details.\n", cpu); in panic_display_last_pc_lr()
398 paniclog_append_noflush("CORE %u: PC=0x%016llx, LR=0x%016llx, FP=0x%016llx\n", cpu, in panic_display_last_pc_lr()
957 int cpu; in DebuggerXCallEnter() local
994 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter()
995 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerXCallEnter()
1032 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter()
1033 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerXCallEnter()
1041 if (processor_array[cpu]->state <= PROCESSOR_PENDING_OFFLINE) { in DebuggerXCallEnter()
[all …]
H A Dcpu_common.c80 cpu_datap(int cpu) in cpu_datap() argument
82 assert(cpu <= ml_get_max_cpu_number()); in cpu_datap()
83 return CpuDataEntries[cpu].cpu_data_vaddr; in cpu_datap()
257 int cpu; in cpu_broadcast_xcall_internal() local
274 for (cpu = 0; cpu < max_cpu; cpu++) { in cpu_broadcast_xcall_internal()
275 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in cpu_broadcast_xcall_internal()
643 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in cpu_exit_wait() local
644 assert(cpu && cpu->cpu_IMPL_regs); in cpu_exit_wait()
645 volatile uint64_t *cpu_sts = (void *)(cpu->cpu_IMPL_regs + CPU_PIO_CPU_STS_OFFSET); in cpu_exit_wait()
686 cpu_can_exit(__unused int cpu) in cpu_can_exit() argument
[all …]
/xnu-10063.101.15/SETUP/config/
H A Ddoconf171 set cpu=`echo $MACHINE | tr A-Z a-z`
176 …ster_file (${MASTER}.${cpu}.${SOC_CONFIG}.${PLATFORM} ${MASTER}.${cpu}.${SOC_CONFIG} ${MASTER}.${c…
192 echo "[ generating $SYSID from $MASTER_DIR/MASTER{,.$cpu}{,.local} ]"
199 …unifdef -t -DPLATFORM_${PLATFORM} -DCPU_$cpu -DSOC_CONFIG_${SOC_CONFIG} -DSYS_${SYS} $DEF_AVP $DEF…
270 echo "${prog}: ${$SYSID}: no such configuration in $MASTER_DIR/MASTER{,.$cpu}"
/xnu-10063.101.15/osfmk/arm64/
H A Dmachine_routines.c734 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_up_update_counts() local
736 os_atomic_inc(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_up_update_counts()
784 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_down_update_counts() local
786 os_atomic_dec(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_down_update_counts()
1058 ml_topology_cpu_t *cpu = &topology_info.cpus[topology_info.num_cpus]; in ml_parse_cpu_topology() local
1060 cpu->cpu_id = topology_info.num_cpus++; in ml_parse_cpu_topology()
1061 assert(cpu->cpu_id < MAX_CPUS); in ml_parse_cpu_topology()
1062 topology_info.max_cpu_id = MAX(topology_info.max_cpu_id, cpu->cpu_id); in ml_parse_cpu_topology()
1064 cpu->die_id = 0; in ml_parse_cpu_topology()
1067 cpu->phys_id = (uint32_t)ml_readprop(child, "reg", ML_READPROP_MANDATORY); in ml_parse_cpu_topology()
[all …]
H A Dmonotonic_arm64.c275 core_idle(__unused cpu_data_t *cpu) in core_idle() argument
277 assert(cpu != NULL); in core_idle()
1236 mt_cpu_idle(cpu_data_t *cpu) in mt_cpu_idle() argument
1238 core_idle(cpu); in mt_cpu_idle()
1242 mt_cpu_run(cpu_data_t *cpu) in mt_cpu_run() argument
1246 assert(cpu != NULL); in mt_cpu_run()
1249 mtc = &cpu->cpu_monotonic; in mt_cpu_run()
1262 mt_cpu_down(cpu_data_t *cpu) in mt_cpu_down() argument
1264 mt_cpu_idle(cpu); in mt_cpu_down()
1268 mt_cpu_up(cpu_data_t *cpu) in mt_cpu_up() argument
[all …]
/xnu-10063.101.15/bsd/kern/
H A Dkdebug_common.c256 struct kd_buffer *kd_data_page, int cpu) in allocate_storage_unit() argument
267 kdbp = &kd_data_page->kdb_info[cpu]; in allocate_storage_unit()
392 release_storage_unit(struct kd_control *kd_ctrl_page, struct kd_buffer *kd_data_page, int cpu, uint… in release_storage_unit() argument
398 kdbp = &kd_data_page->kdb_info[cpu]; in release_storage_unit()
465 int cpu; in kernel_debug_write() local
476 if (kd_rec.cpu == -1) { in kernel_debug_write()
477 cpu = cpu_number(); in kernel_debug_write()
479 cpu = kd_rec.cpu; in kernel_debug_write()
482 kdbp = &kd_data_page->kdb_info[cpu]; in kernel_debug_write()
517 if (allocate_storage_unit(kd_ctrl_page, kd_data_page, cpu) == false) { in kernel_debug_write()
[all …]
/xnu-10063.101.15/osfmk/x86_64/
H A Dmonotonic_x86_64.c208 core_down(cpu_data_t *cpu) in core_down() argument
214 struct mt_cpu *mtc = &cpu->cpu_monotonic; in core_down()
222 core_up(cpu_data_t *cpu) in core_up() argument
232 mtc = &cpu->cpu_monotonic; in core_up()
242 mt_cpu_down(cpu_data_t *cpu) in mt_cpu_down() argument
244 core_down(cpu); in mt_cpu_down()
248 mt_cpu_up(cpu_data_t *cpu) in mt_cpu_up() argument
252 core_up(cpu); in mt_cpu_up()
261 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis() local
262 npmis += cpu->cpu_monotonic.mtc_npmis; in mt_count_pmis()
/xnu-10063.101.15/pexpert/arm/
H A Dpe_identify_machine.c38 DTEntry cpus, cpu; in pe_identify_machine() local
63 while (kSuccess == SecureDTIterateEntries(&iter, &cpu)) { in pe_identify_machine()
64 if ((kSuccess != SecureDTGetProperty(cpu, "state", &value, &size)) || in pe_identify_machine()
70 if (SecureDTGetProperty(cpu, "timebase-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
85 if (SecureDTGetProperty(cpu, "bus-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
102 if (SecureDTGetProperty(cpu, "memory-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
113 if (SecureDTGetProperty(cpu, "peripheral-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
124 if (SecureDTGetProperty(cpu, "fixed-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
132 if (SecureDTGetProperty(cpu, "clock-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
/xnu-10063.101.15/tools/lldbmacros/kmemory/
H A Dkmem.py219 def PERCPU_BASE(self, cpu): argument
263 def PERCPU_BASE(self, cpu): argument
264 cpu_data = self.arm64_CpuDataEntries.chkGetChildAtIndex(cpu)
299 def PERCPU_BASE(self, cpu): argument
300 cpu_data = self.intel_cpu_data.chkGetChildAtIndex(cpu)
323 def __getitem__(self, cpu): argument
324 if cpu in self.kmem.pcpus:
326 addr = sbv.GetLoadAddress() + self.kmem.PERCPU_BASE(cpu)
345 (cpu, sbv.chkCreateValueFromAddress(name, addr + kmem.PERCPU_BASE(cpu), ty))
346 for cpu in kmem.pcpus
/xnu-10063.101.15/tests/cpu_counters/
H A Dkpc_tests.c411 struct cpu { struct
447 struct cpu *cpus = calloc(mch.ncpus, sizeof(*cpus));
481 struct cpu *cpu = &cpus[tp->cpuid];
483 if (cpu->prev_count != 0) {
484 uint64_t delta = count - cpu->prev_count;
486 if (skid > cpu->max_skid) {
487 cpu->max_skid = skid;
490 cpu->prev_count = count;
497 cpu->pmi_timeslices[(unsigned int)slice] += 1;
513 struct cpu *cpu = &cpus[tp->cpuid]; variable
[all …]

123456