Home
last modified time | relevance | path

Searched refs:cpu (Results 1 – 25 of 132) sorted by relevance

123456

/xnu-11417.121.6/osfmk/i386/
H A Dcpu_threads.h40 #define cpu_to_core_lapic(cpu) (cpu_to_lapic[cpu] & ~CPU_THREAD_MASK) argument
41 #define cpu_to_core_cpu(cpu) (lapic_to_cpu[cpu_to_core_lapic(cpu)]) argument
42 #define cpu_to_logical_cpu(cpu) (cpu_to_lapic[cpu] & CPU_THREAD_MASK) argument
43 #define cpu_is_core_cpu(cpu) (cpu_to_logical_cpu(cpu) == 0) argument
45 #define _cpu_to_lcpu(cpu) (&cpu_datap(cpu)->lcpu) argument
46 #define _cpu_to_core(cpu) (_cpu_to_lcpu(cpu)->core) argument
47 #define _cpu_to_package(cpu) (_cpu_to_core(cpu)->package) argument
49 #define cpu_to_lcpu(cpu) ((cpu_datap(cpu) != NULL) ? _cpu_to_lcpu(cpu) : NULL) argument
50 #define cpu_to_core(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->core : NUL… argument
51 #define cpu_to_die(cpu) ((cpu_to_lcpu(cpu) != NULL) ? _cpu_to_lcpu(cpu)->die : NULL) argument
[all …]
H A Dmp.c109 void i386_cpu_IPI(int cpu);
116 static boolean_t cpu_signal_pending(int cpu, mp_event_t event);
156 volatile uint16_t cpu; member
158 .cpu = KDP_XCPU_NONE
187 void i386_send_NMI(int cpu);
674 cpu_interrupt(int cpu) in cpu_interrupt() argument
679 && pmCPUExitIdle(cpu_datap(cpu))) { in cpu_interrupt()
680 i386_cpu_IPI(cpu); in cpu_interrupt()
684 KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_SCHED, MACH_REMOTE_AST), cpu, did_IPI, 0, 0, 0); in cpu_interrupt()
691 cpu_NMI_interrupt(int cpu) in cpu_NMI_interrupt() argument
[all …]
H A DpmCPU.h98 void (*resyncDeadlines)(int cpu);
100 x86_lcpu_t *(*GetLCPU)(int cpu);
101 x86_core_t *(*GetCore)(int cpu);
102 x86_die_t *(*GetDie)(int cpu);
103 x86_pkg_t *(*GetPackage)(int cpu);
110 boolean_t (*GetHibernate)(int cpu);
114 void (*pmSendIPI)(int cpu);
117 uint32_t (*timerQueueMigrate)(int cpu);
133 uint64_t pmCPUGetDeadline(struct cpu_data *cpu);
134 uint64_t pmCPUSetDeadline(struct cpu_data *cpu, uint64_t deadline);
[all …]
H A Dmp_native.c39 void i386_cpu_IPI(int cpu);
43 void i386_send_NMI(int cpu);
89 i386_send_NMI(int cpu) in i386_send_NMI() argument
94 i386_cpu_IPI(cpu); in i386_send_NMI()
96 lapic_send_nmi(cpu); in i386_send_NMI()
114 i386_cpu_IPI(int cpu) in i386_cpu_IPI() argument
117 if (cpu_datap(cpu)->cpu_signals & 6) { /* (BRINGUP) */ in i386_cpu_IPI()
118 …i386_cpu_IPI: sending enter debugger signal (%08X) to cpu %d\n", cpu_datap(cpu)->cpu_signals, cpu); in i386_cpu_IPI()
122 lapic_send_ipi(cpu, LAPIC_VECTOR(INTERPROCESSOR)); in i386_cpu_IPI()
H A Dlapic.c68 lapic_cpu_map(int apic_id, int cpu) in lapic_cpu_map() argument
71 assert(cpu < MAX_CPUS); in lapic_cpu_map()
72 cpu_to_lapic[cpu] = apic_id; in lapic_cpu_map()
73 lapic_to_cpu[apic_id] = cpu; in lapic_cpu_map()
84 ml_get_apicid(uint32_t cpu) in ml_get_apicid() argument
86 if (cpu >= (uint32_t)MAX_CPUS) { in ml_get_apicid()
90 return (uint32_t)cpu_to_lapic[cpu]; in ml_get_apicid()
H A Dmp.h93 extern void cpu_interrupt(int cpu);
163 cpu_to_cpumask(cpu_t cpu) in cpu_to_cpumask() argument
165 return (cpu < MAX_CPUS) ? (1ULL << cpu) : 0; in cpu_to_cpumask()
173 extern void mp_cpus_call_cpu_init(int cpu);
222 extern void cpu_PM_interrupt(int cpu);
229 int cpu; member
252 eventp->cpu = _cpu; \
262 #define DBGLOG_CPU_INIT(cpu) { \ argument
263 cpu_signal_event_log_t **sig_logpp = &cpu_signal[cpu]; \
264 cpu_signal_event_log_t **hdl_logpp = &cpu_handle[cpu]; \
[all …]
H A Dcpu_threads.c337 x86_lcpu_init(int cpu) in x86_lcpu_init() argument
343 cpup = cpu_datap(cpu); in x86_lcpu_init()
347 lcpu->cpu = cpup; in x86_lcpu_init()
354 lcpu->cpu_num = cpu; in x86_lcpu_init()
355 lcpu->lnum = cpu; in x86_lcpu_init()
364 x86_core_alloc(int cpu) in x86_core_alloc() argument
369 cpup = cpu_datap(cpu); in x86_core_alloc()
404 x86_package_find(int cpu) in x86_package_find() argument
410 cpup = cpu_datap(cpu); in x86_package_find()
426 x86_die_find(int cpu) in x86_die_find() argument
[all …]
H A Dcpu.c119 int cpu) in cpu_start() argument
123 if (cpu == cpu_number()) { in cpu_start()
133 ret = intel_startCPU_fast(cpu); in cpu_start()
139 ret = intel_startCPU(cpu); in cpu_start()
143 panic("cpu_start(%d) failed: %d\n", cpu, ret); in cpu_start()
149 int cpu) in cpu_exit_wait() argument
151 cpu_data_t *cdp = cpu_datap(cpu); in cpu_exit_wait()
172 panic("cpu_exit_wait(%d) timeout", cpu); in cpu_exit_wait()
206 int cpu) in cpu_to_processor() argument
208 return cpu_datap(cpu)->cpu_processor; in cpu_to_processor()
H A DpmCPU.c314 pmGetLogicalCPU(int cpu) in pmGetLogicalCPU() argument
316 return cpu_to_lcpu(cpu); in pmGetLogicalCPU()
328 pmGetCore(int cpu) in pmGetCore() argument
330 return cpu_to_core(cpu); in pmGetCore()
342 pmGetDie(int cpu) in pmGetDie() argument
344 return cpu_to_die(cpu); in pmGetDie()
356 pmGetPackage(int cpu) in pmGetPackage() argument
358 return cpu_to_package(cpu); in pmGetPackage()
386 pmCPUGetDeadline(cpu_data_t *cpu) in pmCPUGetDeadline() argument
393 deadline = (*pmDispatch->GetDeadline)(&cpu->lcpu); in pmCPUGetDeadline()
[all …]
H A Dhpet.c109 hpet_request(uint32_t cpu) in hpet_request() argument
126 if (cpu >= real_ncpus) { in hpet_request()
128 lcpu = cpu_to_lcpu(cpu); in hpet_request()
142 rc = (*hpet_req)(ml_get_apicid(cpu), hpet_arg, &hpetReq); in hpet_request()
148 lcpu = cpu_to_lcpu(cpu); in hpet_request()
322 ml_hpet_cfg(uint32_t cpu, uint32_t hpetVect) in ml_hpet_cfg() argument
331 if (cpu > 1) { in ml_hpet_cfg()
332 panic("ml_hpet_cfg: invalid cpu = %d", cpu); in ml_hpet_cfg()
335 lcpu = cpu_to_lcpu(cpu); in ml_hpet_cfg()
349 hpetVaddr = (uint64_t *)(((uintptr_t)&(((hpetReg_t *)hpetArea)->TIM1_CONF)) + (cpu << 5)); in ml_hpet_cfg()
[all …]
/xnu-11417.121.6/osfmk/i386/vmx/
H A Dvmx_cpu.c169 vmx_cpu_t *cpu = &current_cpu_datap()->cpu_vmx; in vmx_on() local
174 cpu_number(), cpu->specs.vmx_on); in vmx_on()
176 assert(cpu->specs.vmx_present); in vmx_on()
178 if (NULL == cpu->vmxon_region) { in vmx_on()
181 vmxon_region_paddr = vmx_paddr(cpu->vmxon_region); in vmx_on()
186 if (FALSE == cpu->specs.vmx_on) { in vmx_on()
187 assert(vmx_is_cr0_valid(&cpu->specs)); in vmx_on()
188 assert(vmx_is_cr4_valid(&cpu->specs)); in vmx_on()
196 cpu->specs.vmx_on = TRUE; in vmx_on()
199 cpu_number(), cpu->specs.vmx_on); in vmx_on()
[all …]
/xnu-11417.121.6/iokit/Kernel/arm/
H A DAppleARMSMP.cpp122 IOService *cpu = IOService::waitForMatchingService(matching, UINT64_MAX); in register_aic_handlers() local
125 OSArray *irqs = (OSArray *) cpu->getProperty(gIOInterruptSpecifiersKey); in register_aic_handlers()
134 if (cpu->registerInterrupt(0, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers()
135 cpu->enableInterrupt(0) != kIOReturnSuccess || in register_aic_handlers()
136 cpu->registerInterrupt(2, NULL, (IOInterruptAction)ipi_handler, NULL) != kIOReturnSuccess || in register_aic_handlers()
137 cpu->enableInterrupt(2) != kIOReturnSuccess) { in register_aic_handlers()
149 …if (cpu->registerInterrupt(1, NULL, (IOInterruptAction)(void (*)(void))pmi_handler, NULL) != kIORe… in register_aic_handlers()
150 cpu->enableInterrupt(1) != kIOReturnSuccess) { in register_aic_handlers()
179 for (unsigned int cpu = 0; cpu < topology_info->num_cpus; cpu++) { in cpu_boot_thread() local
180 const ml_topology_cpu *cpu_info = &topology_info->cpus[cpu]; in cpu_boot_thread()
[all …]
/xnu-11417.121.6/osfmk/kern/
H A Dkpc_thread.c125 cpu_data_t *cpu = NULL; in kpc_update_thread_counters() local
127 cpu = current_cpu_datap(); in kpc_update_thread_counters()
131 NULL, cpu->cpu_kpc_buf[1] ); in kpc_update_thread_counters()
136 thread->kpc_buf[i] += cpu->cpu_kpc_buf[1][i] - cpu->cpu_kpc_buf[0][i]; in kpc_update_thread_counters()
147 tmp = cpu->cpu_kpc_buf[1]; in kpc_update_thread_counters()
148 cpu->cpu_kpc_buf[1] = cpu->cpu_kpc_buf[0]; in kpc_update_thread_counters()
149 cpu->cpu_kpc_buf[0] = tmp; in kpc_update_thread_counters()
H A Dmonotonic.h89 void mt_cpu_idle(cpu_data_t *cpu);
90 void mt_cpu_run(cpu_data_t *cpu);
95 void mt_cpu_down(cpu_data_t *cpu);
96 void mt_cpu_up(cpu_data_t *cpu);
111 uint64_t mt_cpu_update_count(cpu_data_t *cpu, unsigned int ctr);
H A Dkpc.h66 #define FIXED_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr)]) argument
68 #define CONFIGURABLE_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr) + kpc_fixed_c… argument
72 #define FIXED_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr)]) argument
74 #define CONFIGURABLE_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr) + kpc_fixed_c… argument
/xnu-11417.121.6/osfmk/arm/
H A Dmodel_dep.c385 for (int cpu = 0; cpu <= max_cpu; cpu++) { in panic_display_last_pc_lr() local
386 cpu_data_t *current_cpu_datap = cpu_datap(cpu); in panic_display_last_pc_lr()
397 …g_append_noflush("CORE %u is the one that panicked. Check the full backtrace for details.\n", cpu); in panic_display_last_pc_lr()
401 paniclog_append_noflush("CORE %u: PC=0x%016llx, LR=0x%016llx, FP=0x%016llx\n", cpu, in panic_display_last_pc_lr()
991 int cpu, xnu_saved_registers_t *regp) in DebuggerPatchupThreadState() argument
997 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerPatchupThreadState()
1042 int cpu; in DebuggerXCallEnter() local
1081 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter()
1082 target_cpu_datap = (cpu_data_t *)CpuDataEntries[cpu].cpu_data_vaddr; in DebuggerXCallEnter()
1127 for (cpu = 0; cpu <= max_cpu; cpu++) { in DebuggerXCallEnter()
[all …]
/xnu-11417.121.6/libkern/os/
H A Drefcnt.c764 zpercpu_foreach_cpu(cpu) { in zpercpu_foreach_cpu() argument
765 os_atomic_init(zpercpu_get_cpu(rc, cpu), in zpercpu_foreach_cpu()
766 OS_PCPU_REF_LIVE + (cpu ? 0 : OS_PCPU_REF_INC)); in zpercpu_foreach_cpu()
787 for (int cpu = zpercpu_count(); cpu-- > 1;) { variable
788 n |= os_atomic_load_wide(zpercpu_get_cpu(rc, cpu), relaxed);
816 int cpu; in __os_pcpu_ref_delta() local
818 cpu = cpu_number(); in __os_pcpu_ref_delta()
819 rcp = zpercpu_get_cpu(rc, cpu); in __os_pcpu_ref_delta()
826 *cpup = cpu; in __os_pcpu_ref_delta()
836 __os_pcpu_ref_retain_slow(__os_pcpu_ref_t rc, int cpu, uint64_t v) in __os_pcpu_ref_retain_slow() argument
[all …]
/xnu-11417.121.6/SETUP/config/
H A Ddoconf178 set cpu=`echo $MACHINE | tr A-Z a-z`
183 …ster_file (${MASTER}.${cpu}.${SOC_CONFIG}.${PLATFORM} ${MASTER}.${cpu}.${SOC_CONFIG} ${MASTER}.${c…
199 echo "[ generating $SYSID from $MASTER_DIR/MASTER{,.$cpu}{,.local} ]"
206 …unifdef -t -DPLATFORM_${PLATFORM} -DCPU_$cpu -DSOC_CONFIG_${SOC_CONFIG} -DSYS_${SYS} $DEF_AVP $DEF…
277 echo "${prog}: ${$SYSID}: no such configuration in $MASTER_DIR/MASTER{,.$cpu}"
/xnu-11417.121.6/osfmk/arm64/
H A Dmachine_routines.c772 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_up_update_counts() local
774 os_atomic_inc(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_up_update_counts()
873 ml_topology_cpu_t *cpu = &ml_get_topology_info()->cpus[cpu_id]; in ml_cpu_down_update_counts() local
875 os_atomic_dec(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_down_update_counts()
1181 ml_topology_cpu_t *cpu = &topology_info.cpus[topology_info.num_cpus]; in ml_parse_cpu_topology() local
1183 cpu->cpu_id = topology_info.num_cpus++; in ml_parse_cpu_topology()
1184 assert(cpu->cpu_id < MAX_CPUS); in ml_parse_cpu_topology()
1185 topology_info.max_cpu_id = MAX(topology_info.max_cpu_id, cpu->cpu_id); in ml_parse_cpu_topology()
1187 cpu->die_id = (int)ml_readprop(child, "die-id", 0); in ml_parse_cpu_topology()
1188 topology_info.max_die_id = MAX(topology_info.max_die_id, cpu->die_id); in ml_parse_cpu_topology()
[all …]
H A Dmonotonic_arm64.c287 core_idle(__unused cpu_data_t *cpu) in core_idle() argument
289 assert(cpu != NULL); in core_idle()
545 unsigned int cpu = cpu_number(); in _broadcast_block_trampoline() local
546 unsigned int cluster = topo->cpus[cpu].cluster_id; in _broadcast_block_trampoline()
547 if (topo->clusters[cluster].first_cpu_id == cpu) { in _broadcast_block_trampoline()
548 cb(topo->cpus[cpu].cluster_id); in _broadcast_block_trampoline()
1328 mt_cpu_idle(cpu_data_t *cpu) in mt_cpu_idle() argument
1330 core_idle(cpu); in mt_cpu_idle()
1334 mt_cpu_run(cpu_data_t *cpu) in mt_cpu_run() argument
1338 assert(cpu != NULL); in mt_cpu_run()
[all …]
/xnu-11417.121.6/bsd/kern/
H A Dkdebug_common.c257 struct kd_buffer *kd_data_page, int cpu) in allocate_storage_unit() argument
268 kdbp = &kd_data_page->kdb_info[cpu]; in allocate_storage_unit()
393 release_storage_unit(struct kd_control *kd_ctrl_page, struct kd_buffer *kd_data_page, int cpu, uint… in release_storage_unit() argument
399 kdbp = &kd_data_page->kdb_info[cpu]; in release_storage_unit()
466 int cpu; in kernel_debug_write() local
477 if (kd_rec.cpu == -1) { in kernel_debug_write()
478 cpu = cpu_number(); in kernel_debug_write()
480 cpu = kd_rec.cpu; in kernel_debug_write()
483 kdbp = &kd_data_page->kdb_info[cpu]; in kernel_debug_write()
518 if (allocate_storage_unit(kd_ctrl_page, kd_data_page, cpu) == false) { in kernel_debug_write()
[all …]
/xnu-11417.121.6/osfmk/x86_64/
H A Dmonotonic_x86_64.c208 core_down(cpu_data_t *cpu) in core_down() argument
214 struct mt_cpu *mtc = &cpu->cpu_monotonic; in core_down()
222 core_up(cpu_data_t *cpu) in core_up() argument
232 mtc = &cpu->cpu_monotonic; in core_up()
242 mt_cpu_down(cpu_data_t *cpu) in mt_cpu_down() argument
244 core_down(cpu); in mt_cpu_down()
248 mt_cpu_up(cpu_data_t *cpu) in mt_cpu_up() argument
252 core_up(cpu); in mt_cpu_up()
261 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis() local
262 npmis += cpu->cpu_monotonic.mtc_npmis; in mt_count_pmis()
/xnu-11417.121.6/pexpert/arm/
H A Dpe_identify_machine.c38 DTEntry cpus, cpu; in pe_identify_machine() local
63 while (kSuccess == SecureDTIterateEntries(&iter, &cpu)) { in pe_identify_machine()
64 if ((kSuccess != SecureDTGetProperty(cpu, "state", &value, &size)) || in pe_identify_machine()
70 if (SecureDTGetProperty(cpu, "timebase-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
85 if (SecureDTGetProperty(cpu, "bus-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
102 if (SecureDTGetProperty(cpu, "memory-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
113 if (SecureDTGetProperty(cpu, "peripheral-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
124 if (SecureDTGetProperty(cpu, "fixed-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
132 if (SecureDTGetProperty(cpu, "clock-frequency", &value, &size) == kSuccess) { in pe_identify_machine()
/xnu-11417.121.6/tools/lldbmacros/kmemory/
H A Dkmem.py224 def PERCPU_BASE(self, cpu): argument
268 def PERCPU_BASE(self, cpu): argument
269 cpu_data = self.arm64_CpuDataEntries.chkGetChildAtIndex(cpu)
304 def PERCPU_BASE(self, cpu): argument
305 cpu_data = self.intel_cpu_data.chkGetChildAtIndex(cpu)
328 def __getitem__(self, cpu): argument
329 if cpu in self.kmem.pcpus:
331 addr = sbv.GetLoadAddress() + self.kmem.PERCPU_BASE(cpu)
350 (cpu, sbv.chkCreateValueFromAddress(name, addr + kmem.PERCPU_BASE(cpu), ty))
351 for cpu in kmem.pcpus
/xnu-11417.121.6/tests/cpu_counters/
H A Dkpc_tests.c411 struct cpu { struct
447 struct cpu *cpus = calloc(mch.ncpus, sizeof(*cpus));
481 struct cpu *cpu = &cpus[tp->cpuid];
483 if (cpu->prev_count != 0) {
484 uint64_t delta = count - cpu->prev_count;
486 if (skid > cpu->max_skid) {
487 cpu->max_skid = skid;
490 cpu->prev_count = count;
497 cpu->pmi_timeslices[(unsigned int)slice] += 1;
513 struct cpu *cpu = &cpus[tp->cpuid]; variable
[all …]

123456