| /xnu-11215.81.4/osfmk/arm64/ |
| H A D | cpu.c | 321 cpu_data_t *cpu_data_ptr = getCpuDatap(); in cpu_sleep() local 323 cpu_data_ptr->cpu_active_thread = current_thread(); in cpu_sleep() 325 cpu_data_ptr->cpu_reset_handler = (uintptr_t) VM_KERNEL_STRIP_PTR(arm_init_cpu); in cpu_sleep() 327 cpu_data_ptr->cpu_reset_handler = (uintptr_t) start_cpu_paddr; in cpu_sleep() 329 os_atomic_or(&cpu_data_ptr->cpu_flags, SleepState, relaxed); in cpu_sleep() 331 if (cpu_data_ptr->cpu_user_debug != NULL) { in cpu_sleep() 337 mt_cpu_down(cpu_data_ptr); in cpu_sleep() 347 PE_cpu_machine_quiesce(cpu_data_ptr->cpu_id); in cpu_sleep() 349 bool deep_sleep = PE_cpu_down(cpu_data_ptr->cpu_id); in cpu_sleep() 350 cpu_data_ptr->cpu_sleep_token = ARM_CPU_ON_SLEEP_PATH; in cpu_sleep() [all …]
|
| H A D | machine_routines.c | 753 cpu_data_t *cpu_data_ptr = getCpuDatap(); in ml_cpu_up() local 755 assert(!bit_test(os_atomic_load(&ml_cpu_up_processors, relaxed), cpu_data_ptr->cpu_number)); in ml_cpu_up() 757 atomic_bit_set(&ml_cpu_up_processors, cpu_data_ptr->cpu_number, memory_order_relaxed); in ml_cpu_up() 819 cpu_data_t *cpu_data_ptr = getCpuDatap(); in ml_cpu_down() local 820 cpu_data_ptr->cpu_running = FALSE; in ml_cpu_down() 822 assert((cpu_data_ptr->cpu_signal & SIGPdisabled) == 0); in ml_cpu_down() 823 assert(bit_test(os_atomic_load(&ml_cpu_up_processors, relaxed), cpu_data_ptr->cpu_number)); in ml_cpu_down() 825 atomic_bit_clear(&ml_cpu_up_processors, cpu_data_ptr->cpu_number, memory_order_release); in ml_cpu_down() 827 if (cpu_data_ptr == &BootCpuData && ml_is_quiescing()) { in ml_cpu_down() 833 } else if (cpu_data_ptr != &BootCpuData || (support_bootcpu_shutdown && !ml_is_quiescing())) { in ml_cpu_down() [all …]
|
| H A D | pcb.c | 854 struct cpu_data * cpu_data_ptr; in arm_debug_set32() local 862 cpu_data_ptr = getCpuDatap(); in arm_debug_set32() 863 cpu_debug = cpu_data_ptr->cpu_user_debug; in arm_debug_set32() 871 cpu_data_ptr->cpu_user_debug = NULL; in arm_debug_set32() 875 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set32() 1056 struct cpu_data * cpu_data_ptr; in arm_debug_set64() local 1064 cpu_data_ptr = getCpuDatap(); in arm_debug_set64() 1065 cpu_debug = cpu_data_ptr->cpu_user_debug; in arm_debug_set64() 1073 cpu_data_ptr->cpu_user_debug = NULL; in arm_debug_set64() 1077 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set64()
|
| /xnu-11215.81.4/osfmk/arm/ |
| H A D | cpu_common.c | 120 cpu_data_t *cpu_data_ptr = CpuDataEntries[slot_num].cpu_data_vaddr; in cpu_info() local 130 cpu_stat->irq_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info() 131 cpu_stat->ipi_cnt = (uint32_t)cpu_data_ptr->cpu_stat.ipi_cnt; in cpu_info() 132 cpu_stat->timer_cnt = (uint32_t)cpu_data_ptr->cpu_stat.timer_cnt; in cpu_info() 133 cpu_stat->undef_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.undef_ex_cnt; in cpu_info() 134 cpu_stat->unaligned_cnt = (uint32_t)cpu_data_ptr->cpu_stat.unaligned_cnt; in cpu_info() 135 cpu_stat->vfp_cnt = (uint32_t)cpu_data_ptr->cpu_stat.vfp_cnt; in cpu_info() 137 cpu_stat->data_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.data_ex_cnt; in cpu_info() 138 cpu_stat->instr_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.instr_ex_cnt; in cpu_info() 152 cpu_stat->irq_ex_cnt = cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info() [all …]
|
| H A D | caches.c | 66 cpu_data_t *cpu_data_ptr = getCpuDatap(); in flush_dcache() local 85 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in flush_dcache() 86 …cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlushRegion, (unsigned int) paddr… in flush_dcache() 100 cpu_data_t *cpu_data_ptr = getCpuDatap(); in clean_dcache() local 119 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in clean_dcache() 120 …cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanRegion, (unsigned int) paddr, (un… in clean_dcache() 151 cpu_data_t *cpu_data_ptr = getCpuDatap(); in dcache_incoherent_io_flush64() local 155 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64() 156 cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlush, 0x0UL, 0x0UL); in dcache_incoherent_io_flush64() 182 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64() [all …]
|
| H A D | arm_timer.c | 71 cpu_data_t *cpu_data_ptr; in timer_intr() local 74 cpu_data_ptr = getCpuDatap(); in timer_intr() 75 mytimer = &cpu_data_ptr->rtclock_timer; /* Point to the event timer */ in timer_intr() 79 if ((cpu_data_ptr->idle_timer_deadline > 0) && (cpu_data_ptr->idle_timer_deadline <= abstime)) { in timer_intr() 80 cpu_data_ptr->idle_timer_deadline = 0x0ULL; in timer_intr() 84 cpu_data_ptr->idle_timer_notify(cpu_data_ptr->idle_timer_refcon, &new_idle_timeout_ticks); in timer_intr() 89 …clock_absolutetime_interval_to_deadline(new_idle_timeout_ticks, &cpu_data_ptr->idle_timer_deadline… in timer_intr() 108 processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr); in timer_intr() 115 cpu_data_ptr->rtcPop = EndOfAllTime; in timer_intr() 128 cpu_data_t *cpu_data_ptr; in timer_set_deadline() local [all …]
|
| H A D | arm_init.c | 688 cpu_data_t *cpu_data_ptr, in arm_init_cpu() argument 706 os_atomic_andnot(&cpu_data_ptr->cpu_flags, SleepState, relaxed); in arm_init_cpu() 709 machine_set_current_thread(cpu_data_ptr->cpu_active_thread); in arm_init_cpu() 717 …if ((cpu_data_ptr == &BootCpuData) && (gIOHibernateState == kIOHibernateStateWakingFromHibernate) … in arm_init_cpu() 730 if (get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread) != in arm_init_cpu() 733 get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread), in arm_init_cpu() 736 cpu_data_ptr->cpu_active_thread->machine.preemption_count--; in arm_init_cpu() 751 if ((cpus_defeatures & (0xF << 4 * cpu_data_ptr->cpu_number)) != 0) { in arm_init_cpu() 752 cpu_defeatures_set((cpus_defeatures >> 4 * cpu_data_ptr->cpu_number) & 0xF); in arm_init_cpu() 765 if (cpu_data_ptr == &BootCpuData && ml_is_quiescing()) { in arm_init_cpu() [all …]
|
| H A D | model_dep.c | 825 cpu_data_t * cpu_data_ptr = getCpuDatap(); in print_all_backtraces() local 827 assert(cpu_data_ptr->PAB_active == FALSE); in print_all_backtraces() 828 cpu_data_ptr->PAB_active = TRUE; in print_all_backtraces() 840 cpu_data_ptr->PAB_active = FALSE; in print_all_backtraces() 1047 cpu_data_t *cpu_data_ptr = getCpuDatap(); in DebuggerXCallEnter() local 1050 cpu_data_ptr->debugger_active++; in DebuggerXCallEnter() 1051 if (cpu_data_ptr->debugger_active != 1) { in DebuggerXCallEnter() 1084 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() 1130 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() 1190 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() [all …]
|
| H A D | cpu_data_internal.h | 313 extern void cpu_data_init(cpu_data_t *cpu_data_ptr); 314 extern void cpu_data_register(cpu_data_t *cpu_data_ptr);
|
| /xnu-11215.81.4/osfmk/i386/ |
| H A D | Diagnostics.c | 152 cpu_data_ptr[i]->cpu_hwIntCnt[j] = 0; in diagCall64() 177 …(void) copyout((char *) &cpu_data_ptr[i]->cpu_hwIntCnt, curpos + 8, 256 * sizeof(uint32_t)); /*… in diagCall64() 273 cest.caperf = cpu_data_ptr[i]->cpu_aperf; in diagCall64() 274 cest.cmperf = cpu_data_ptr[i]->cpu_mperf; in diagCall64() 275 cest.ccres[0] = cpu_data_ptr[i]->cpu_c3res; in diagCall64() 276 cest.ccres[1] = cpu_data_ptr[i]->cpu_c6res; in diagCall64() 277 cest.ccres[2] = cpu_data_ptr[i]->cpu_c7res; in diagCall64() 279 bcopy(&cpu_data_ptr[i]->cpu_rtimes[0], &cest.crtimes[0], sizeof(cest.crtimes)); in diagCall64() 280 bcopy(&cpu_data_ptr[i]->cpu_itimes[0], &cest.citimes[0], sizeof(cest.citimes)); in diagCall64() 282 cest.citime_total = cpu_data_ptr[i]->cpu_itime_total; in diagCall64() [all …]
|
| H A D | trap.c | 318 cpu_data_ptr[i]->cpu_max_observed_int_latency = in interrupt_reset_latency_stats() 319 cpu_data_ptr[i]->cpu_max_observed_int_latency_vector = 0; in interrupt_reset_latency_stats() 330 if (cur_max < cpu_data_ptr[i]->cpu_max_observed_int_latency) { in interrupt_populate_latency_stats() 331 cur_max = cpu_data_ptr[i]->cpu_max_observed_int_latency; in interrupt_populate_latency_stats() 337 …f(buf, bufsize, "0x%x 0x%x 0x%llx", tcpu, cpu_data_ptr[tcpu]->cpu_max_observed_int_latency_vector,… in interrupt_populate_latency_stats() 358 cpu_data_t *cdp = cpu_data_ptr[cnum]; in interrupt() 376 if (cpu_data_ptr[cnum]->lcpu.package->num_idle == topoParms.nLThreadsPerPackage) { in interrupt() 377 cpu_data_ptr[cnum]->cpu_hwIntpexits[interrupt_num]++; in interrupt()
|
| H A D | cpu_data.h | 342 extern cpu_data_t *__single cpu_data_ptr[MAX_CPUS]; 705 return cpu_data_ptr[cpu]; in cpu_datap() 718 return cpu_data_ptr[cpu]->cd_shadow; in cpu_shadowp()
|
| H A D | mp_desc.c | 185 cpu_data_t *cpu_data_ptr[MAX_CPUS] = {[0] = &scdatas[0] }; variable 669 cpu_data_ptr[cnum] = cdp; in cpu_data_alloc() 783 bcopy((void *) cpu_data_ptr[0], (void*) cdp, sizeof(cpu_data_t)); in cpu_data_realloc() 798 cpu_data_ptr[0] = cdp; in cpu_data_realloc()
|
| H A D | cpu_topology.c | 121 qsort((void *) &cpu_data_ptr[1], in cpu_topology_sort()
|
| H A D | locks_i386.c | 1998 if ((cpu_data_ptr[i] != NULL) && (cpu_data_ptr[i]->cpu_active_thread == owner_thread)) { in lck_mtx_lock_spinwait_x86()
|
| H A D | mp.c | 508 my_word = &cpu_data_ptr[my_cpu]->cpu_signals; in cpu_signal_handler() 514 cpu_data_ptr[my_cpu]->cpu_prior_signals = *my_word; in cpu_signal_handler()
|
| /xnu-11215.81.4/osfmk/arm64/sptm/ |
| H A D | arm_init_sptm.c | 983 cpu_data_t *cpu_data_ptr, in arm_init_cpu() argument 1012 os_atomic_andnot(&cpu_data_ptr->cpu_flags, SleepState, relaxed); in arm_init_cpu() 1015 machine_set_current_thread(cpu_data_ptr->cpu_active_thread); in arm_init_cpu() 1024 …if ((cpu_data_ptr == &BootCpuData) && (gIOHibernateState == kIOHibernateStateWakingFromHibernate) … in arm_init_cpu() 1037 if (get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread) != in arm_init_cpu() 1040 get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread), in arm_init_cpu() 1043 cpu_data_ptr->cpu_active_thread->machine.preemption_count--; in arm_init_cpu() 1057 if ((cpus_defeatures & (0xF << 4 * cpu_data_ptr->cpu_number)) != 0) { in arm_init_cpu() 1058 cpu_defeatures_set((cpus_defeatures >> 4 * cpu_data_ptr->cpu_number) & 0xF); in arm_init_cpu() 1071 if (cpu_data_ptr == &BootCpuData && ml_is_quiescing()) { in arm_init_cpu() [all …]
|
| /xnu-11215.81.4/tools/lldbmacros/ |
| H A D | misc.py | 29 while kern.globals.cpu_data_ptr[cpu]: 30 cd = kern.globals.cpu_data_ptr[cpu] 162 cpu_data = kern.globals.cpu_data_ptr[cpu_id]
|
| H A D | scheduler.py | 373 cpu_data = kern.globals.cpu_data_ptr[current_processor.cpu_id]
|
| H A D | process.py | 1889 cpu_data = kern.globals.cpu_data_ptr[processor.cpu_id]
|
| /xnu-11215.81.4/osfmk/x86_64/ |
| H A D | monotonic_x86_64.c | 261 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis()
|
| /xnu-11215.81.4/osfmk/arm/pmap/ |
| H A D | pmap.c | 823 static void pmap_switch_user_ttb(pmap_t pmap, pmap_cpu_data_t *cpu_data_ptr); 4584 pmap_cpu_data_t *cpu_data_ptr = pmap_get_cpu_data(); local 4586 os_atomic_store(&cpu_data_ptr->active_pmap, pmap, relaxed); 4613 pmap_t last_nested_pmap = cpu_data_ptr->cpu_nested_pmap; 4614 __unused const pt_attr_t *last_nested_pmap_attr = cpu_data_ptr->cpu_nested_pmap_attr; 4615 __unused vm_map_address_t last_nested_region_addr = cpu_data_ptr->cpu_nested_region_addr; 4616 __unused vm_map_offset_t last_nested_region_size = cpu_data_ptr->cpu_nested_region_size; 4626 assert(asid_index < (sizeof(cpu_data_ptr->cpu_sw_asids) / sizeof(*cpu_data_ptr->cpu_sw_asids))); 4630 uint8_t last_sw_asid = cpu_data_ptr->cpu_sw_asids[asid_index]; 4639 cpu_data_ptr->cpu_sw_asids[asid_index] = new_sw_asid; [all …]
|
| /xnu-11215.81.4/osfmk/kdp/ml/x86_64/ |
| H A D | kdp_machdep.c | 262 if (cpu_data_ptr[i] == NULL) { in kdp_machine_hostinfo()
|
| /xnu-11215.81.4/tools/lldbmacros/core/ |
| H A D | kernelcore.py | 337 return unsigned(self.globals.cpu_data_ptr[cpu].cpu_pcpu_base)
|