| /xnu-10063.101.15/osfmk/arm64/ |
| H A D | cpu.c | 317 cpu_data_t *cpu_data_ptr = getCpuDatap(); in cpu_sleep() local 319 cpu_data_ptr->cpu_active_thread = current_thread(); in cpu_sleep() 321 cpu_data_ptr->cpu_reset_handler = (uintptr_t) VM_KERNEL_STRIP_PTR(arm_init_cpu); in cpu_sleep() 323 cpu_data_ptr->cpu_reset_handler = (uintptr_t) start_cpu_paddr; in cpu_sleep() 325 cpu_data_ptr->cpu_flags |= SleepState; in cpu_sleep() 327 if (cpu_data_ptr->cpu_user_debug != NULL) { in cpu_sleep() 333 mt_cpu_down(cpu_data_ptr); in cpu_sleep() 343 PE_cpu_machine_quiesce(cpu_data_ptr->cpu_id); in cpu_sleep() 345 bool deep_sleep = PE_cpu_down(cpu_data_ptr->cpu_id); in cpu_sleep() 346 cpu_data_ptr->cpu_sleep_token = ARM_CPU_ON_SLEEP_PATH; in cpu_sleep() [all …]
|
| H A D | pcb.c | 603 struct cpu_data * cpu_data_ptr; in arm_debug_set32() local 611 cpu_data_ptr = getCpuDatap(); in arm_debug_set32() 612 cpu_debug = cpu_data_ptr->cpu_user_debug; in arm_debug_set32() 620 cpu_data_ptr->cpu_user_debug = NULL; in arm_debug_set32() 624 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set32() 805 struct cpu_data * cpu_data_ptr; in arm_debug_set64() local 813 cpu_data_ptr = getCpuDatap(); in arm_debug_set64() 814 cpu_debug = cpu_data_ptr->cpu_user_debug; in arm_debug_set64() 822 cpu_data_ptr->cpu_user_debug = NULL; in arm_debug_set64() 826 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set64()
|
| H A D | machine_routines.c | 764 cpu_data_t *cpu_data_ptr = getCpuDatap(); in ml_cpu_down() local 765 cpu_data_ptr->cpu_running = FALSE; in ml_cpu_down() 767 if (cpu_data_ptr != &BootCpuData) { in ml_cpu_down() 772 timer_queue_shutdown(&cpu_data_ptr->rtclock_timer.queue); in ml_cpu_down() 862 cpu_data_t *cpu_data_ptr; in ml_install_interrupt_handler() local 866 cpu_data_ptr = getCpuDatap(); in ml_install_interrupt_handler() 868 cpu_data_ptr->interrupt_nub = nub; in ml_install_interrupt_handler() 869 cpu_data_ptr->interrupt_source = source; in ml_install_interrupt_handler() 870 cpu_data_ptr->interrupt_target = target; in ml_install_interrupt_handler() 871 cpu_data_ptr->interrupt_handler = handler; in ml_install_interrupt_handler() [all …]
|
| /xnu-10063.101.15/osfmk/arm/ |
| H A D | cpu_common.c | 119 cpu_data_t *cpu_data_ptr = CpuDataEntries[slot_num].cpu_data_vaddr; in cpu_info() local 129 cpu_stat->irq_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info() 130 cpu_stat->ipi_cnt = (uint32_t)cpu_data_ptr->cpu_stat.ipi_cnt; in cpu_info() 131 cpu_stat->timer_cnt = (uint32_t)cpu_data_ptr->cpu_stat.timer_cnt; in cpu_info() 132 cpu_stat->undef_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.undef_ex_cnt; in cpu_info() 133 cpu_stat->unaligned_cnt = (uint32_t)cpu_data_ptr->cpu_stat.unaligned_cnt; in cpu_info() 134 cpu_stat->vfp_cnt = (uint32_t)cpu_data_ptr->cpu_stat.vfp_cnt; in cpu_info() 136 cpu_stat->data_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.data_ex_cnt; in cpu_info() 137 cpu_stat->instr_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.instr_ex_cnt; in cpu_info() 151 cpu_stat->irq_ex_cnt = cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info() [all …]
|
| H A D | caches.c | 66 cpu_data_t *cpu_data_ptr = getCpuDatap(); in flush_dcache() local 85 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in flush_dcache() 86 …cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlushRegion, (unsigned int) paddr… in flush_dcache() 100 cpu_data_t *cpu_data_ptr = getCpuDatap(); in clean_dcache() local 119 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in clean_dcache() 120 …cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanRegion, (unsigned int) paddr, (un… in clean_dcache() 151 cpu_data_t *cpu_data_ptr = getCpuDatap(); in dcache_incoherent_io_flush64() local 155 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64() 156 cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlush, 0x0UL, 0x0UL); in dcache_incoherent_io_flush64() 182 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64() [all …]
|
| H A D | arm_timer.c | 71 cpu_data_t *cpu_data_ptr; in timer_intr() local 74 cpu_data_ptr = getCpuDatap(); in timer_intr() 75 mytimer = &cpu_data_ptr->rtclock_timer; /* Point to the event timer */ in timer_intr() 79 if ((cpu_data_ptr->idle_timer_deadline > 0) && (cpu_data_ptr->idle_timer_deadline <= abstime)) { in timer_intr() 80 cpu_data_ptr->idle_timer_deadline = 0x0ULL; in timer_intr() 84 cpu_data_ptr->idle_timer_notify(cpu_data_ptr->idle_timer_refcon, &new_idle_timeout_ticks); in timer_intr() 89 …clock_absolutetime_interval_to_deadline(new_idle_timeout_ticks, &cpu_data_ptr->idle_timer_deadline… in timer_intr() 108 processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr); in timer_intr() 115 cpu_data_ptr->rtcPop = EndOfAllTime; in timer_intr() 128 cpu_data_t *cpu_data_ptr; in timer_set_deadline() local [all …]
|
| H A D | arm_init.c | 667 cpu_data_t *cpu_data_ptr) in arm_init_cpu() argument 678 cpu_data_ptr->cpu_flags &= ~SleepState; in arm_init_cpu() 681 machine_set_current_thread(cpu_data_ptr->cpu_active_thread); in arm_init_cpu() 688 …if ((cpu_data_ptr == &BootCpuData) && (gIOHibernateState == kIOHibernateStateWakingFromHibernate))… in arm_init_cpu() 701 if (get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread) != in arm_init_cpu() 704 get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread), in arm_init_cpu() 707 cpu_data_ptr->cpu_active_thread->machine.preemption_count--; in arm_init_cpu() 722 if ((cpus_defeatures & (0xF << 4 * cpu_data_ptr->cpu_number)) != 0) { in arm_init_cpu() 723 cpu_defeatures_set((cpus_defeatures >> 4 * cpu_data_ptr->cpu_number) & 0xF); in arm_init_cpu() 736 if (cpu_data_ptr == &BootCpuData) { in arm_init_cpu() [all …]
|
| H A D | model_dep.c | 783 cpu_data_t * cpu_data_ptr = getCpuDatap(); in print_all_backtraces() local 785 assert(cpu_data_ptr->PAB_active == FALSE); in print_all_backtraces() 786 cpu_data_ptr->PAB_active = TRUE; in print_all_backtraces() 798 cpu_data_ptr->PAB_active = FALSE; in print_all_backtraces() 960 cpu_data_t *cpu_data_ptr = getCpuDatap(); in DebuggerXCallEnter() local 963 cpu_data_ptr->debugger_active++; in DebuggerXCallEnter() 964 if (cpu_data_ptr->debugger_active != 1) { in DebuggerXCallEnter() 997 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() 1035 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() 1072 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() [all …]
|
| H A D | cpu_data_internal.h | 285 extern void cpu_data_init(cpu_data_t *cpu_data_ptr); 286 extern void cpu_data_free(cpu_data_t *cpu_data_ptr); 287 extern kern_return_t cpu_data_register(cpu_data_t *cpu_data_ptr);
|
| /xnu-10063.101.15/osfmk/i386/ |
| H A D | Diagnostics.c | 152 cpu_data_ptr[i]->cpu_hwIntCnt[j] = 0; in diagCall64() 177 …(void) copyout((char *) &cpu_data_ptr[i]->cpu_hwIntCnt, curpos + 8, 256 * sizeof(uint32_t)); /*… in diagCall64() 273 cest.caperf = cpu_data_ptr[i]->cpu_aperf; in diagCall64() 274 cest.cmperf = cpu_data_ptr[i]->cpu_mperf; in diagCall64() 275 cest.ccres[0] = cpu_data_ptr[i]->cpu_c3res; in diagCall64() 276 cest.ccres[1] = cpu_data_ptr[i]->cpu_c6res; in diagCall64() 277 cest.ccres[2] = cpu_data_ptr[i]->cpu_c7res; in diagCall64() 279 bcopy(&cpu_data_ptr[i]->cpu_rtimes[0], &cest.crtimes[0], sizeof(cest.crtimes)); in diagCall64() 280 bcopy(&cpu_data_ptr[i]->cpu_itimes[0], &cest.citimes[0], sizeof(cest.citimes)); in diagCall64() 282 cest.citime_total = cpu_data_ptr[i]->cpu_itime_total; in diagCall64() [all …]
|
| H A D | trap.c | 317 cpu_data_ptr[i]->cpu_max_observed_int_latency = in interrupt_reset_latency_stats() 318 cpu_data_ptr[i]->cpu_max_observed_int_latency_vector = 0; in interrupt_reset_latency_stats() 329 if (cur_max < cpu_data_ptr[i]->cpu_max_observed_int_latency) { in interrupt_populate_latency_stats() 330 cur_max = cpu_data_ptr[i]->cpu_max_observed_int_latency; in interrupt_populate_latency_stats() 336 …f(buf, bufsize, "0x%x 0x%x 0x%llx", tcpu, cpu_data_ptr[tcpu]->cpu_max_observed_int_latency_vector,… in interrupt_populate_latency_stats() 357 cpu_data_t *cdp = cpu_data_ptr[cnum]; in interrupt() 375 if (cpu_data_ptr[cnum]->lcpu.package->num_idle == topoParms.nLThreadsPerPackage) { in interrupt() 376 cpu_data_ptr[cnum]->cpu_hwIntpexits[interrupt_num]++; in interrupt()
|
| H A D | cpu_data.h | 342 extern cpu_data_t *__single cpu_data_ptr[MAX_CPUS]; 705 return cpu_data_ptr[cpu]; in cpu_datap() 718 return cpu_data_ptr[cpu]->cd_shadow; in cpu_shadowp()
|
| H A D | mp_desc.c | 185 cpu_data_t *cpu_data_ptr[MAX_CPUS] = {[0] = &scdatas[0] }; variable 669 cpu_data_ptr[cnum] = cdp; in cpu_data_alloc() 783 bcopy((void *) cpu_data_ptr[0], (void*) cdp, sizeof(cpu_data_t)); in cpu_data_realloc() 798 cpu_data_ptr[0] = cdp; in cpu_data_realloc()
|
| H A D | cpu_topology.c | 121 qsort((void *) &cpu_data_ptr[1], in cpu_topology_sort()
|
| H A D | locks_i386.c | 1998 if ((cpu_data_ptr[i] != NULL) && (cpu_data_ptr[i]->cpu_active_thread == owner_thread)) { in lck_mtx_lock_spinwait_x86()
|
| H A D | mp.c | 507 my_word = &cpu_data_ptr[my_cpu]->cpu_signals; in cpu_signal_handler() 513 cpu_data_ptr[my_cpu]->cpu_prior_signals = *my_word; in cpu_signal_handler()
|
| /xnu-10063.101.15/osfmk/arm64/sptm/ |
| H A D | arm_init_sptm.c | 917 cpu_data_t *cpu_data_ptr) in arm_init_cpu() argument 928 cpu_data_ptr->cpu_flags &= ~SleepState; in arm_init_cpu() 931 machine_set_current_thread(cpu_data_ptr->cpu_active_thread); in arm_init_cpu() 934 …if ((cpu_data_ptr == &BootCpuData) && (gIOHibernateState == kIOHibernateStateWakingFromHibernate))… in arm_init_cpu() 947 if (get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread) != in arm_init_cpu() 950 get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread), in arm_init_cpu() 953 cpu_data_ptr->cpu_active_thread->machine.preemption_count--; in arm_init_cpu() 967 if ((cpus_defeatures & (0xF << 4 * cpu_data_ptr->cpu_number)) != 0) { in arm_init_cpu() 968 cpu_defeatures_set((cpus_defeatures >> 4 * cpu_data_ptr->cpu_number) & 0xF); in arm_init_cpu() 981 if (cpu_data_ptr == &BootCpuData) { in arm_init_cpu() [all …]
|
| /xnu-10063.101.15/tools/lldbmacros/ |
| H A D | misc.py | 29 while kern.globals.cpu_data_ptr[cpu]: 30 cd = kern.globals.cpu_data_ptr[cpu] 162 cpu_data = kern.globals.cpu_data_ptr[cpu_id]
|
| H A D | scheduler.py | 370 cpu_data = kern.globals.cpu_data_ptr[current_processor.cpu_id]
|
| H A D | process.py | 1646 cpu_data = kern.globals.cpu_data_ptr[processor.cpu_id]
|
| /xnu-10063.101.15/osfmk/x86_64/ |
| H A D | monotonic_x86_64.c | 261 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis()
|
| /xnu-10063.101.15/osfmk/arm/pmap/ |
| H A D | pmap.c | 788 static void pmap_switch_user_ttb(pmap_t pmap, pmap_cpu_data_t *cpu_data_ptr); 4511 pmap_cpu_data_t *cpu_data_ptr = pmap_get_cpu_data(); local 4513 os_atomic_store(&cpu_data_ptr->active_pmap, pmap, relaxed); 4537 pmap_t last_nested_pmap = cpu_data_ptr->cpu_nested_pmap; 4538 __unused const pt_attr_t *last_nested_pmap_attr = cpu_data_ptr->cpu_nested_pmap_attr; 4539 __unused vm_map_address_t last_nested_region_addr = cpu_data_ptr->cpu_nested_region_addr; 4540 __unused vm_map_offset_t last_nested_region_size = cpu_data_ptr->cpu_nested_region_size; 4550 assert(asid_index < (sizeof(cpu_data_ptr->cpu_sw_asids) / sizeof(*cpu_data_ptr->cpu_sw_asids))); 4554 uint8_t last_sw_asid = cpu_data_ptr->cpu_sw_asids[asid_index]; 4563 cpu_data_ptr->cpu_sw_asids[asid_index] = new_sw_asid; [all …]
|
| /xnu-10063.101.15/osfmk/kdp/ml/x86_64/ |
| H A D | kdp_machdep.c | 262 if (cpu_data_ptr[i] == NULL) { in kdp_machine_hostinfo()
|
| /xnu-10063.101.15/tools/lldbmacros/core/ |
| H A D | kernelcore.py | 337 return unsigned(self.globals.cpu_data_ptr[cpu].cpu_pcpu_base)
|