Home
last modified time | relevance | path

Searched refs:cpu_data_ptr (Results 1 – 24 of 24) sorted by relevance

/xnu-11215.81.4/osfmk/arm64/
H A Dcpu.c321 cpu_data_t *cpu_data_ptr = getCpuDatap(); in cpu_sleep() local
323 cpu_data_ptr->cpu_active_thread = current_thread(); in cpu_sleep()
325 cpu_data_ptr->cpu_reset_handler = (uintptr_t) VM_KERNEL_STRIP_PTR(arm_init_cpu); in cpu_sleep()
327 cpu_data_ptr->cpu_reset_handler = (uintptr_t) start_cpu_paddr; in cpu_sleep()
329 os_atomic_or(&cpu_data_ptr->cpu_flags, SleepState, relaxed); in cpu_sleep()
331 if (cpu_data_ptr->cpu_user_debug != NULL) { in cpu_sleep()
337 mt_cpu_down(cpu_data_ptr); in cpu_sleep()
347 PE_cpu_machine_quiesce(cpu_data_ptr->cpu_id); in cpu_sleep()
349 bool deep_sleep = PE_cpu_down(cpu_data_ptr->cpu_id); in cpu_sleep()
350 cpu_data_ptr->cpu_sleep_token = ARM_CPU_ON_SLEEP_PATH; in cpu_sleep()
[all …]
H A Dmachine_routines.c753 cpu_data_t *cpu_data_ptr = getCpuDatap(); in ml_cpu_up() local
755 assert(!bit_test(os_atomic_load(&ml_cpu_up_processors, relaxed), cpu_data_ptr->cpu_number)); in ml_cpu_up()
757 atomic_bit_set(&ml_cpu_up_processors, cpu_data_ptr->cpu_number, memory_order_relaxed); in ml_cpu_up()
819 cpu_data_t *cpu_data_ptr = getCpuDatap(); in ml_cpu_down() local
820 cpu_data_ptr->cpu_running = FALSE; in ml_cpu_down()
822 assert((cpu_data_ptr->cpu_signal & SIGPdisabled) == 0); in ml_cpu_down()
823 assert(bit_test(os_atomic_load(&ml_cpu_up_processors, relaxed), cpu_data_ptr->cpu_number)); in ml_cpu_down()
825 atomic_bit_clear(&ml_cpu_up_processors, cpu_data_ptr->cpu_number, memory_order_release); in ml_cpu_down()
827 if (cpu_data_ptr == &BootCpuData && ml_is_quiescing()) { in ml_cpu_down()
833 } else if (cpu_data_ptr != &BootCpuData || (support_bootcpu_shutdown && !ml_is_quiescing())) { in ml_cpu_down()
[all …]
H A Dpcb.c854 struct cpu_data * cpu_data_ptr; in arm_debug_set32() local
862 cpu_data_ptr = getCpuDatap(); in arm_debug_set32()
863 cpu_debug = cpu_data_ptr->cpu_user_debug; in arm_debug_set32()
871 cpu_data_ptr->cpu_user_debug = NULL; in arm_debug_set32()
875 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set32()
1056 struct cpu_data * cpu_data_ptr; in arm_debug_set64() local
1064 cpu_data_ptr = getCpuDatap(); in arm_debug_set64()
1065 cpu_debug = cpu_data_ptr->cpu_user_debug; in arm_debug_set64()
1073 cpu_data_ptr->cpu_user_debug = NULL; in arm_debug_set64()
1077 cpu_data_ptr->cpu_user_debug = debug_state; in arm_debug_set64()
/xnu-11215.81.4/osfmk/arm/
H A Dcpu_common.c120 cpu_data_t *cpu_data_ptr = CpuDataEntries[slot_num].cpu_data_vaddr; in cpu_info() local
130 cpu_stat->irq_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info()
131 cpu_stat->ipi_cnt = (uint32_t)cpu_data_ptr->cpu_stat.ipi_cnt; in cpu_info()
132 cpu_stat->timer_cnt = (uint32_t)cpu_data_ptr->cpu_stat.timer_cnt; in cpu_info()
133 cpu_stat->undef_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.undef_ex_cnt; in cpu_info()
134 cpu_stat->unaligned_cnt = (uint32_t)cpu_data_ptr->cpu_stat.unaligned_cnt; in cpu_info()
135 cpu_stat->vfp_cnt = (uint32_t)cpu_data_ptr->cpu_stat.vfp_cnt; in cpu_info()
137 cpu_stat->data_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.data_ex_cnt; in cpu_info()
138 cpu_stat->instr_ex_cnt = (uint32_t)cpu_data_ptr->cpu_stat.instr_ex_cnt; in cpu_info()
152 cpu_stat->irq_ex_cnt = cpu_data_ptr->cpu_stat.irq_ex_cnt; in cpu_info()
[all …]
H A Dcaches.c66 cpu_data_t *cpu_data_ptr = getCpuDatap(); in flush_dcache() local
85 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in flush_dcache()
86cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlushRegion, (unsigned int) paddr… in flush_dcache()
100 cpu_data_t *cpu_data_ptr = getCpuDatap(); in clean_dcache() local
119 if (paddr && (cpu_data_ptr->cpu_cache_dispatch != NULL)) { in clean_dcache()
120cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanRegion, (unsigned int) paddr, (un… in clean_dcache()
151 cpu_data_t *cpu_data_ptr = getCpuDatap(); in dcache_incoherent_io_flush64() local
155 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64()
156 cpu_data_ptr->cpu_cache_dispatch(cpu_data_ptr->cpu_id, CacheCleanFlush, 0x0UL, 0x0UL); in dcache_incoherent_io_flush64()
182 if (cpu_data_ptr->cpu_cache_dispatch != NULL) { in dcache_incoherent_io_flush64()
[all …]
H A Darm_timer.c71 cpu_data_t *cpu_data_ptr; in timer_intr() local
74 cpu_data_ptr = getCpuDatap(); in timer_intr()
75 mytimer = &cpu_data_ptr->rtclock_timer; /* Point to the event timer */ in timer_intr()
79 if ((cpu_data_ptr->idle_timer_deadline > 0) && (cpu_data_ptr->idle_timer_deadline <= abstime)) { in timer_intr()
80 cpu_data_ptr->idle_timer_deadline = 0x0ULL; in timer_intr()
84 cpu_data_ptr->idle_timer_notify(cpu_data_ptr->idle_timer_refcon, &new_idle_timeout_ticks); in timer_intr()
89 …clock_absolutetime_interval_to_deadline(new_idle_timeout_ticks, &cpu_data_ptr->idle_timer_deadline… in timer_intr()
108 processor = PERCPU_GET_RELATIVE(processor, cpu_data, cpu_data_ptr); in timer_intr()
115 cpu_data_ptr->rtcPop = EndOfAllTime; in timer_intr()
128 cpu_data_t *cpu_data_ptr; in timer_set_deadline() local
[all …]
H A Darm_init.c688 cpu_data_t *cpu_data_ptr, in arm_init_cpu() argument
706 os_atomic_andnot(&cpu_data_ptr->cpu_flags, SleepState, relaxed); in arm_init_cpu()
709 machine_set_current_thread(cpu_data_ptr->cpu_active_thread); in arm_init_cpu()
717 …if ((cpu_data_ptr == &BootCpuData) && (gIOHibernateState == kIOHibernateStateWakingFromHibernate) … in arm_init_cpu()
730 if (get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread) != in arm_init_cpu()
733 get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread), in arm_init_cpu()
736 cpu_data_ptr->cpu_active_thread->machine.preemption_count--; in arm_init_cpu()
751 if ((cpus_defeatures & (0xF << 4 * cpu_data_ptr->cpu_number)) != 0) { in arm_init_cpu()
752 cpu_defeatures_set((cpus_defeatures >> 4 * cpu_data_ptr->cpu_number) & 0xF); in arm_init_cpu()
765 if (cpu_data_ptr == &BootCpuData && ml_is_quiescing()) { in arm_init_cpu()
[all …]
H A Dmodel_dep.c825 cpu_data_t * cpu_data_ptr = getCpuDatap(); in print_all_backtraces() local
827 assert(cpu_data_ptr->PAB_active == FALSE); in print_all_backtraces()
828 cpu_data_ptr->PAB_active = TRUE; in print_all_backtraces()
840 cpu_data_ptr->PAB_active = FALSE; in print_all_backtraces()
1047 cpu_data_t *cpu_data_ptr = getCpuDatap(); in DebuggerXCallEnter() local
1050 cpu_data_ptr->debugger_active++; in DebuggerXCallEnter()
1051 if (cpu_data_ptr->debugger_active != 1) { in DebuggerXCallEnter()
1084 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter()
1130 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter()
1190 if ((target_cpu_datap == NULL) || (target_cpu_datap == cpu_data_ptr)) { in DebuggerXCallEnter()
[all …]
H A Dcpu_data_internal.h313 extern void cpu_data_init(cpu_data_t *cpu_data_ptr);
314 extern void cpu_data_register(cpu_data_t *cpu_data_ptr);
/xnu-11215.81.4/osfmk/i386/
H A DDiagnostics.c152 cpu_data_ptr[i]->cpu_hwIntCnt[j] = 0; in diagCall64()
177 …(void) copyout((char *) &cpu_data_ptr[i]->cpu_hwIntCnt, curpos + 8, 256 * sizeof(uint32_t)); /*… in diagCall64()
273 cest.caperf = cpu_data_ptr[i]->cpu_aperf; in diagCall64()
274 cest.cmperf = cpu_data_ptr[i]->cpu_mperf; in diagCall64()
275 cest.ccres[0] = cpu_data_ptr[i]->cpu_c3res; in diagCall64()
276 cest.ccres[1] = cpu_data_ptr[i]->cpu_c6res; in diagCall64()
277 cest.ccres[2] = cpu_data_ptr[i]->cpu_c7res; in diagCall64()
279 bcopy(&cpu_data_ptr[i]->cpu_rtimes[0], &cest.crtimes[0], sizeof(cest.crtimes)); in diagCall64()
280 bcopy(&cpu_data_ptr[i]->cpu_itimes[0], &cest.citimes[0], sizeof(cest.citimes)); in diagCall64()
282 cest.citime_total = cpu_data_ptr[i]->cpu_itime_total; in diagCall64()
[all …]
H A Dtrap.c318 cpu_data_ptr[i]->cpu_max_observed_int_latency = in interrupt_reset_latency_stats()
319 cpu_data_ptr[i]->cpu_max_observed_int_latency_vector = 0; in interrupt_reset_latency_stats()
330 if (cur_max < cpu_data_ptr[i]->cpu_max_observed_int_latency) { in interrupt_populate_latency_stats()
331 cur_max = cpu_data_ptr[i]->cpu_max_observed_int_latency; in interrupt_populate_latency_stats()
337 …f(buf, bufsize, "0x%x 0x%x 0x%llx", tcpu, cpu_data_ptr[tcpu]->cpu_max_observed_int_latency_vector,… in interrupt_populate_latency_stats()
358 cpu_data_t *cdp = cpu_data_ptr[cnum]; in interrupt()
376 if (cpu_data_ptr[cnum]->lcpu.package->num_idle == topoParms.nLThreadsPerPackage) { in interrupt()
377 cpu_data_ptr[cnum]->cpu_hwIntpexits[interrupt_num]++; in interrupt()
H A Dcpu_data.h342 extern cpu_data_t *__single cpu_data_ptr[MAX_CPUS];
705 return cpu_data_ptr[cpu]; in cpu_datap()
718 return cpu_data_ptr[cpu]->cd_shadow; in cpu_shadowp()
H A Dmp_desc.c185 cpu_data_t *cpu_data_ptr[MAX_CPUS] = {[0] = &scdatas[0] }; variable
669 cpu_data_ptr[cnum] = cdp; in cpu_data_alloc()
783 bcopy((void *) cpu_data_ptr[0], (void*) cdp, sizeof(cpu_data_t)); in cpu_data_realloc()
798 cpu_data_ptr[0] = cdp; in cpu_data_realloc()
H A Dcpu_topology.c121 qsort((void *) &cpu_data_ptr[1], in cpu_topology_sort()
H A Dlocks_i386.c1998 if ((cpu_data_ptr[i] != NULL) && (cpu_data_ptr[i]->cpu_active_thread == owner_thread)) { in lck_mtx_lock_spinwait_x86()
H A Dmp.c508 my_word = &cpu_data_ptr[my_cpu]->cpu_signals; in cpu_signal_handler()
514 cpu_data_ptr[my_cpu]->cpu_prior_signals = *my_word; in cpu_signal_handler()
/xnu-11215.81.4/osfmk/arm64/sptm/
H A Darm_init_sptm.c983 cpu_data_t *cpu_data_ptr, in arm_init_cpu() argument
1012 os_atomic_andnot(&cpu_data_ptr->cpu_flags, SleepState, relaxed); in arm_init_cpu()
1015 machine_set_current_thread(cpu_data_ptr->cpu_active_thread); in arm_init_cpu()
1024 …if ((cpu_data_ptr == &BootCpuData) && (gIOHibernateState == kIOHibernateStateWakingFromHibernate) … in arm_init_cpu()
1037 if (get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread) != in arm_init_cpu()
1040 get_preemption_level_for_thread(cpu_data_ptr->cpu_active_thread), in arm_init_cpu()
1043 cpu_data_ptr->cpu_active_thread->machine.preemption_count--; in arm_init_cpu()
1057 if ((cpus_defeatures & (0xF << 4 * cpu_data_ptr->cpu_number)) != 0) { in arm_init_cpu()
1058 cpu_defeatures_set((cpus_defeatures >> 4 * cpu_data_ptr->cpu_number) & 0xF); in arm_init_cpu()
1071 if (cpu_data_ptr == &BootCpuData && ml_is_quiescing()) { in arm_init_cpu()
[all …]
/xnu-11215.81.4/tools/lldbmacros/
H A Dmisc.py29 while kern.globals.cpu_data_ptr[cpu]:
30 cd = kern.globals.cpu_data_ptr[cpu]
162 cpu_data = kern.globals.cpu_data_ptr[cpu_id]
H A Dscheduler.py373 cpu_data = kern.globals.cpu_data_ptr[current_processor.cpu_id]
H A Dprocess.py1889 cpu_data = kern.globals.cpu_data_ptr[processor.cpu_id]
/xnu-11215.81.4/osfmk/x86_64/
H A Dmonotonic_x86_64.c261 cpu_data_t *cpu = cpu_data_ptr[i]; in mt_count_pmis()
/xnu-11215.81.4/osfmk/arm/pmap/
H A Dpmap.c823 static void pmap_switch_user_ttb(pmap_t pmap, pmap_cpu_data_t *cpu_data_ptr);
4584 pmap_cpu_data_t *cpu_data_ptr = pmap_get_cpu_data(); local
4586 os_atomic_store(&cpu_data_ptr->active_pmap, pmap, relaxed);
4613 pmap_t last_nested_pmap = cpu_data_ptr->cpu_nested_pmap;
4614 __unused const pt_attr_t *last_nested_pmap_attr = cpu_data_ptr->cpu_nested_pmap_attr;
4615 __unused vm_map_address_t last_nested_region_addr = cpu_data_ptr->cpu_nested_region_addr;
4616 __unused vm_map_offset_t last_nested_region_size = cpu_data_ptr->cpu_nested_region_size;
4626 assert(asid_index < (sizeof(cpu_data_ptr->cpu_sw_asids) / sizeof(*cpu_data_ptr->cpu_sw_asids)));
4630 uint8_t last_sw_asid = cpu_data_ptr->cpu_sw_asids[asid_index];
4639 cpu_data_ptr->cpu_sw_asids[asid_index] = new_sw_asid;
[all …]
/xnu-11215.81.4/osfmk/kdp/ml/x86_64/
H A Dkdp_machdep.c262 if (cpu_data_ptr[i] == NULL) { in kdp_machine_hostinfo()
/xnu-11215.81.4/tools/lldbmacros/core/
H A Dkernelcore.py337 return unsigned(self.globals.cpu_data_ptr[cpu].cpu_pcpu_base)