| /xnu-8019.80.24/osfmk/arm/ |
| H A D | cpu.c | 99 cpu_data_t *cpu_data_ptr = getCpuDatap(); in cpu_sleep() local 100 cpu_data_ptr->cpu_active_thread = current_thread(); in cpu_sleep() 101 cpu_data_ptr->cpu_reset_handler = (vm_offset_t) start_cpu_paddr; in cpu_sleep() 102 cpu_data_ptr->cpu_flags |= SleepState; in cpu_sleep() 103 cpu_data_ptr->cpu_user_debug = NULL; in cpu_sleep() 107 PE_cpu_machine_quiesce(cpu_data_ptr->cpu_id); in cpu_sleep() 119 cpu_data_t *cpu_data_ptr = getCpuDatap(); in cpu_idle() local 122 if ((!idle_enable) || (cpu_data_ptr->cpu_signal & SIGPdisabled)) { in cpu_idle() 128 lastPop = cpu_data_ptr->rtcPop; in cpu_idle() 130 cpu_data_ptr->cpu_active_thread = current_thread(); in cpu_idle() [all …]
|
| H A D | cpu_common.c | 118 cpu_data_t *cpu_data_ptr = CpuDataEntries[slot_num].cpu_data_vaddr; in cpu_info() local 128 cpu_stat->irq_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info() 129 cpu_stat->ipi_cnt = (uint32_t)cpu_data_ptr->cpu_stat.ipi_cnt; in cpu_info() 130 cpu_stat->timer_cnt = (uint32_t)cpu_data_ptr->cpu_stat.timer_cnt; in cpu_info() 131 cpu_stat->undef_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.undef_ex_cnt; in cpu_info() 132 cpu_stat->unaligned_cnt = (uint32_t)cpu_data_ptr->cpu_stat.unaligned_cnt; in cpu_info() 133 cpu_stat->vfp_cnt = (uint32_t)cpu_data_ptr->cpu_stat.vfp_cnt; in cpu_info() 135 cpu_stat->data_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.data_ex_cnt; in cpu_info() 136 cpu_stat->instr_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.instr_ex_cnt; in cpu_info() 150 cpu_stat->irq_ex_cnt = cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info() [all …]
|
| H A D | arm_timer.c | 71 cpu_data_t *cpu_data_ptr; in timer_intr() local 74 cpu_data_ptr = getCpuDatap(); in timer_intr() 75 mytimer = &cpu_data_ptr->rtclock_timer; /* Point to the event timer */ in timer_intr() 79 if ((cpu_data_ptr->idle_timer_deadline > 0) && (cpu_data_ptr->idle_timer_deadline <= abstime)) { in timer_intr() 80 cpu_data_ptr->idle_timer_deadline = 0x0ULL; in timer_intr() 84 cpu_data_ptr->idle_timer_notify(cpu_data_ptr->idle_timer_refcon, &new_idle_timeout_ticks); in timer_intr() 89 …clock_absolutetime_interval_to_deadline(new_idle_timeout_ticks, &cpu_data_ptr->idle_timer_deadline… in timer_intr() 108 processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr); in timer_intr() 115 cpu_data_ptr->rtcPop = EndOfAllTime; in timer_intr() 128 cpu_data_t *cpu_data_ptr; in timer_set_deadline() local [all …]
|
| H A D | caches.c | 66 cpu_data_t *cpu_data_ptr = getCpuDatap(); in flush_dcache() local 85 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in flush_dcache() 86 …cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlushRegion, (unsigned int) paddr… in flush_dcache() 100 cpu_data_t *cpu_data_ptr = getCpuDatap(); in clean_dcache() local 119 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in clean_dcache() 120 …cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanRegion, (unsigned int) paddr, (un… in clean_dcache() 155 cpu_data_t *cpu_data_ptr = getCpuDatap(); in dcache_incoherent_io_flush64() local 162 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64() 163 cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlush, 0x0UL, 0x0UL); in dcache_incoherent_io_flush64() 190 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64() [all …]
|
| H A D | arm_init.c | 641 cpu_data_t *cpu_data_ptr) in arm_init_cpu() argument 652 cpu_data_ptr->cpu_flags &= ~SleepState; in arm_init_cpu() 654 cpu_data_ptr->cpu_CLW_active = 1; in arm_init_cpu() 658 machine_set_current_thread(cpu_data_ptr->cpu_active_thread); in arm_init_cpu() 661 …if ((cpu_data_ptr == &BootCpuData) && (gIOHibernateState == kIOHibernateStateWakingFromHibernate))… in arm_init_cpu() 674 if (cpu_data_ptr->cpu_active_thread->machine.preemption_count != expected_preemption_count) { in arm_init_cpu() 676 cpu_data_ptr->cpu_active_thread->machine.preemption_count, in arm_init_cpu() 679 cpu_data_ptr->cpu_active_thread->machine.preemption_count--; in arm_init_cpu() 699 if ((cpus_defeatures & (0xF << 4 * cpu_data_ptr->cpu_number)) != 0) { in arm_init_cpu() 700 cpu_defeatures_set((cpus_defeatures >> 4 * cpu_data_ptr->cpu_number) & 0xF); in arm_init_cpu() [all …]
|
| H A D | machine_routines.c | 221 cpu_data_t *cpu_data_ptr; in ml_cpu_down() local 240 cpu_data_ptr = getCpuDatap(); in ml_cpu_down() 241 cpu_data_ptr->cpu_running = FALSE; in ml_cpu_down() 354 cpu_data_t *cpu_data_ptr; in ml_install_interrupt_handler() local 358 cpu_data_ptr = getCpuDatap(); in ml_install_interrupt_handler() 360 cpu_data_ptr->interrupt_nub = nub; in ml_install_interrupt_handler() 361 cpu_data_ptr->interrupt_source = source; in ml_install_interrupt_handler() 362 cpu_data_ptr->interrupt_target = target; in ml_install_interrupt_handler() 363 cpu_data_ptr->interrupt_handler = handler; in ml_install_interrupt_handler() 364 cpu_data_ptr->interrupt_refCon = refCon; in ml_install_interrupt_handler() [all …]
|
| H A D | model_dep.c | 708 cpu_data_t * cpu_data_ptr = getCpuDatap(); in print_all_backtraces() local 710 assert(cpu_data_ptr->PAB_active == FALSE); in print_all_backtraces() 711 cpu_data_ptr->PAB_active = TRUE; in print_all_backtraces() 723 cpu_data_ptr->PAB_active = FALSE; in print_all_backtraces() 895 cpu_data_t *cpu_data_ptr = getCpuDatap(); in DebuggerXCallEnter() local 898 cpu_data_ptr->debugger_active++; in DebuggerXCallEnter() 899 if (cpu_data_ptr->debugger_active != 1) { in DebuggerXCallEnter() 939 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() 984 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter() 1009 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr) || in DebuggerXCallEnter() [all …]
|
| H A D | pcb.c | 357 struct cpu_data *cpu_data_ptr; in arm_debug_set() local 362 cpu_data_ptr = getCpuDatap(); in arm_debug_set() 365 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set() 369 uintptr_t debug_map = cpu_data_ptr->cpu_debug_interface_map; in arm_debug_set()
|
| H A D | cpu_data_internal.h | 330 extern void cpu_data_init(cpu_data_t *cpu_data_ptr); 331 extern void cpu_data_free(cpu_data_t *cpu_data_ptr); 332 extern kern_return_t cpu_data_register(cpu_data_t *cpu_data_ptr);
|
| H A D | locks_arm.c | 497 cpu_data_t *cpu_data_ptr; in kernel_preempt_check() local 510 cpu_data_ptr = os_atomic_load(&thread->machine.CpuDatap, compiler_acq_rel); in kernel_preempt_check() 511 if (__probable((cpu_data_ptr->cpu_pending_ast & AST_URGENT) == 0)) { in kernel_preempt_check() 524 cpu_data_ptr = os_atomic_load(&thread->machine.CpuDatap, compiler_acq_rel); in kernel_preempt_check() 1323 cpu_data_t *cpu_data_ptr = CpuDataEntries[i].cpu_data_vaddr; in lck_mtx_lock_contended_spinwait_arm() local 1324 if ((cpu_data_ptr != NULL) && (cpu_data_ptr->cpu_active_thread == owner)) { in lck_mtx_lock_contended_spinwait_arm()
|
| H A D | pmap.c | 772 static void pmap_switch_user_ttb(pmap_t pmap, pmap_cpu_data_t *cpu_data_ptr); 4463 pmap_cpu_data_t *cpu_data_ptr = pmap_get_cpu_data(); local 4466 os_atomic_store(&cpu_data_ptr->active_pmap, pmap, relaxed); 4480 pmap_t last_nested_pmap = cpu_data_ptr->cpu_nested_pmap; 4482 __unused const pt_attr_t *last_nested_pmap_attr = cpu_data_ptr->cpu_nested_pmap_attr; 4483 __unused vm_map_address_t last_nested_region_addr = cpu_data_ptr->cpu_nested_region_addr; 4484 __unused vm_map_offset_t last_nested_region_size = cpu_data_ptr->cpu_nested_region_size; 4494 assert(asid_index < (sizeof(cpu_data_ptr->cpu_sw_asids) / sizeof(*cpu_data_ptr->cpu_sw_asids))); 4498 uint8_t last_sw_asid = cpu_data_ptr->cpu_sw_asids[asid_index]; 4507 cpu_data_ptr->cpu_sw_asids[asid_index] = new_sw_asid; [all …]
|
| /xnu-8019.80.24/osfmk/arm64/ |
| H A D | cpu.c | 319 cpu_data_t *cpu_data_ptr = getCpuDatap(); in cpu_sleep() local 321 cpu_data_ptr->cpu_active_thread = current_thread(); in cpu_sleep() 322 cpu_data_ptr->cpu_reset_handler = (uintptr_t) start_cpu_paddr; in cpu_sleep() 323 cpu_data_ptr->cpu_flags |= SleepState; in cpu_sleep() 324 cpu_data_ptr->cpu_user_debug = NULL; in cpu_sleep() 329 mt_cpu_down(cpu_data_ptr); in cpu_sleep() 339 PE_cpu_machine_quiesce(cpu_data_ptr->cpu_id); in cpu_sleep() 341 bool deep_sleep = PE_cpu_down(cpu_data_ptr->cpu_id); in cpu_sleep() 342 cpu_data_ptr->cpu_sleep_token = ARM_CPU_ON_SLEEP_PATH; in cpu_sleep() 344 cpu_data_ptr->cpu_reset_handler = (uintptr_t)0; in cpu_sleep() [all …]
|
| H A D | machine_routines.c | 668 cpu_data_t *cpu_data_ptr; in ml_cpu_down() local 687 cpu_data_ptr = getCpuDatap(); in ml_cpu_down() 688 cpu_data_ptr->cpu_running = FALSE; in ml_cpu_down() 690 if (cpu_data_ptr != &BootCpuData) { in ml_cpu_down() 695 timer_queue_shutdown(&cpu_data_ptr->rtclock_timer.queue); in ml_cpu_down() 771 cpu_data_t *cpu_data_ptr; in ml_install_interrupt_handler() local 775 cpu_data_ptr = getCpuDatap(); in ml_install_interrupt_handler() 777 cpu_data_ptr->interrupt_nub = nub; in ml_install_interrupt_handler() 778 cpu_data_ptr->interrupt_source = source; in ml_install_interrupt_handler() 779 cpu_data_ptr->interrupt_target = target; in ml_install_interrupt_handler() [all …]
|
| H A D | pcb.c | 476 struct cpu_data * cpu_data_ptr; in arm_debug_set32() local 483 cpu_data_ptr = getCpuDatap(); in arm_debug_set32() 486 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set32() 668 struct cpu_data * cpu_data_ptr; in arm_debug_set64() local 675 cpu_data_ptr = getCpuDatap(); in arm_debug_set64() 678 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set64()
|
| /xnu-8019.80.24/osfmk/i386/ |
| H A D | Diagnostics.c | 155 cpu_data_ptr[i]->cpu_hwIntCnt[j] = 0; in diagCall64() 180 …(void) copyout((char *) &cpu_data_ptr[i]->cpu_hwIntCnt, curpos + 8, 256 * sizeof(uint32_t)); /*… in diagCall64() 275 cest.caperf = cpu_data_ptr[i]->cpu_aperf; in diagCall64() 276 cest.cmperf = cpu_data_ptr[i]->cpu_mperf; in diagCall64() 277 cest.ccres[0] = cpu_data_ptr[i]->cpu_c3res; in diagCall64() 278 cest.ccres[1] = cpu_data_ptr[i]->cpu_c6res; in diagCall64() 279 cest.ccres[2] = cpu_data_ptr[i]->cpu_c7res; in diagCall64() 281 bcopy(&cpu_data_ptr[i]->cpu_rtimes[0], &cest.crtimes[0], sizeof(cest.crtimes)); in diagCall64() 282 bcopy(&cpu_data_ptr[i]->cpu_itimes[0], &cest.citimes[0], sizeof(cest.citimes)); in diagCall64() 284 cest.citime_total = cpu_data_ptr[i]->cpu_itime_total; in diagCall64() [all …]
|
| H A D | trap.c | 315 cpu_data_ptr[i]->cpu_max_observed_int_latency = in interrupt_reset_latency_stats() 316 cpu_data_ptr[i]->cpu_max_observed_int_latency_vector = 0; in interrupt_reset_latency_stats() 327 if (cur_max < cpu_data_ptr[i]->cpu_max_observed_int_latency) { in interrupt_populate_latency_stats() 328 cur_max = cpu_data_ptr[i]->cpu_max_observed_int_latency; in interrupt_populate_latency_stats() 334 …f(buf, bufsize, "0x%x 0x%x 0x%llx", tcpu, cpu_data_ptr[tcpu]->cpu_max_observed_int_latency_vector,… in interrupt_populate_latency_stats() 355 cpu_data_t *cdp = cpu_data_ptr[cnum]; in interrupt() 373 if (cpu_data_ptr[cnum]->lcpu.package->num_idle == topoParms.nLThreadsPerPackage) { in interrupt() 374 cpu_data_ptr[cnum]->cpu_hwIntpexits[interrupt_num]++; in interrupt()
|
| H A D | mp_desc.c | 189 cpu_data_t *cpu_data_ptr[MAX_CPUS] = {[0] = &scdatas[0] }; variable 683 cpu_data_ptr[cnum] = cdp; in cpu_data_alloc() 799 bcopy((void *) cpu_data_ptr[0], (void*) cdp, sizeof(cpu_data_t)); in cpu_data_realloc() 814 cpu_data_ptr[0] = cdp; in cpu_data_realloc()
|
| H A D | cpu_data.h | 373 extern cpu_data_t *__single cpu_data_ptr[MAX_CPUS]; 760 return cpu_data_ptr[cpu]; in cpu_datap() 773 return cpu_data_ptr[cpu]->cd_shadow; in cpu_shadowp()
|
| H A D | cpu_topology.c | 122 qsort((void *) &cpu_data_ptr[1], in cpu_topology_sort()
|
| H A D | mp.c | 510 my_word = &cpu_data_ptr[my_cpu]->cpu_signals; in cpu_signal_handler() 516 cpu_data_ptr[my_cpu]->cpu_prior_signals = *my_word; in cpu_signal_handler()
|
| H A D | locks_i386.c | 2205 if ((cpu_data_ptr[i] != NULL) && (cpu_data_ptr[i]->cpu_active_thread == owner)) { in lck_mtx_lock_spinwait_x86()
|
| /xnu-8019.80.24/osfmk/x86_64/ |
| H A D | monotonic_x86_64.c | 266 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis()
|
| /xnu-8019.80.24/tools/lldbmacros/ |
| H A D | misc.py | 60 while kern.globals.cpu_data_ptr[cpu]: 61 cd = kern.globals.cpu_data_ptr[cpu] 166 cpu_data = kern.globals.cpu_data_ptr[cpu_id]
|
| /xnu-8019.80.24/osfmk/kdp/ml/x86_64/ |
| H A D | kdp_machdep.c | 261 if (cpu_data_ptr[i] == NULL) { in kdp_machine_hostinfo()
|
| /xnu-8019.80.24/tools/lldbmacros/core/ |
| H A D | kernelcore.py | 417 return unsigned(self.globals.cpu_data_ptr[cpu].cpu_pcpu_base)
|