| /xnu-12377.81.4/osfmk/x86_64/ |
| H A D | pmap_pcid.c | 100 if (cpu_datap(i)) { in pmap_pcid_configure() 101 cpu_datap(i)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 105 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 114 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 119 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = TRUE; in pmap_pcid_configure() 144 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = TRUE; in pmap_pcid_configure() 149 cpu_datap(ccpu)->cpu_pmap_pcid_coherentp = in pmap_pcid_configure() 150 cpu_datap(ccpu)->cpu_pmap_pcid_coherentp_kernel = in pmap_pcid_configure() 152 cpu_datap(ccpu)->cpu_pcid_data = &pcid_data[ccpu]; in pmap_pcid_configure() 153 cpu_datap(ccpu)->cpu_pcid_data->cpu_pcid_refcounts[0] = 1; in pmap_pcid_configure() [all …]
|
| H A D | pmap.c | 2777 bool gflushed = (cpu_datap(rcpu)->cpu_tlb_invalid_global_count != in pmap_tlbi_response() 2778 cpu_datap(lcpu)->cpu_tlb_gen_counts_global[rcpu]); in pmap_tlbi_response() 2788 bool lflushed = (cpu_datap(rcpu)->cpu_tlb_invalid_local_count != in pmap_tlbi_response() 2789 cpu_datap(lcpu)->cpu_tlb_gen_counts_local[rcpu]); in pmap_tlbi_response() 2797 if ((cpu_datap(rcpu)->cpu_tlb_invalid == 0) || in pmap_tlbi_response() 2838 cpu_datap(cpu)->cpu_tlb_invalid_global = 1; in pmap_flush() 2841 cpu_datap(cpu)->cpu_tlb_invalid_local = 1; in pmap_flush() 2843 cpu_datap(my_cpu)->cpu_tlb_gen_counts_global[cpu] = cpu_datap(cpu)->cpu_tlb_invalid_global_count; in pmap_flush() 2844 cpu_datap(my_cpu)->cpu_tlb_gen_counts_local[cpu] = cpu_datap(cpu)->cpu_tlb_invalid_local_count; in pmap_flush() 3024 … cpu_datap(my_cpu)->cpu_tlb_gen_counts_global[cpu] = cpu_datap(cpu)->cpu_tlb_invalid_global_count; in pmap_flush_tlbs() [all …]
|
| /xnu-12377.81.4/osfmk/i386/ |
| H A D | cpu_topology.c | 68 cpu_data_t *cpup = cpu_datap(i); in cpu_shadow_sort() 69 ptrdiff_t coff = cpup - cpu_datap(0); in cpu_shadow_sort() 95 assert(cpu_datap(0)->cpu_number == 0); in cpu_topology_sort() 110 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 129 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 139 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 177 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 246 TOPO_DBG("\tlcpu %d\n", cpu_datap(i)->cpu_number); in cpu_topology_start_cpu() 247 processor_boot(cpu_datap(i)->cpu_processor); in cpu_topology_start_cpu()
|
| H A D | cpu.c | 151 cpu_data_t *cdp = cpu_datap(cpu); in cpu_exit_wait() 208 return cpu_datap(cpu)->cpu_processor; in cpu_to_processor() 221 return cpu_datap(slot_num)->cpu_type; in slot_type() 228 return cpu_datap(slot_num)->cpu_subtype; in slot_subtype() 235 return cpu_datap(slot_num)->cpu_threadtype; in slot_threadtype() 262 cpu_data_t *cpup = cpu_datap(cpu_num); in processor_to_datastring()
|
| H A D | pmap.h | 654 cpu_datap(ccpu)->cpu_task_cr3 = pcr3; in set_dirbase() 657 cpu_datap(ccpu)->cpu_ucr3 = ucr3; in set_dirbase() 660 cpu_datap(ccpu)->cpu_task_map = cpu_shadowp(ccpu)->cpu_task_map = in set_dirbase() 671 boolean_t priorpagezero = cpu_datap(ccpu)->cpu_pagezero_mapped; in set_dirbase() 672 cpu_datap(ccpu)->cpu_pagezero_mapped = nopagezero; in set_dirbase() 684 } else if (get_cr3_base() != cpu_datap(ccpu)->cpu_kernel_cr3) { in set_dirbase() 685 set_cr3_raw(cpu_datap(ccpu)->cpu_kernel_cr3); in set_dirbase() 696 if (get_cr3_base() != cpu_datap(ccpu)->cpu_kernel_cr3) { in set_dirbase() 697 set_cr3_raw(cpu_datap(ccpu)->cpu_kernel_cr3); in set_dirbase() 813 …3_raw(), map->pmap->pm_cr3, cpu_datap(ccpu)->cpu_kernel_cr3, kernel_pmap->pm_cr3, cpu_datap(ccpu)-… [all …]
|
| H A D | mp.c | 268 if (cpu_datap(slot_num)->cpu_running) { in mp_wait_for_cpu_up() 314 slot_num, cpu_datap(slot_num)->cpu_running, 0, 0, 0); in intel_startCPU_fast() 320 if (cpu_datap(slot_num)->cpu_running) { in intel_startCPU_fast() 381 cpu_datap(psip->target_cpu)->cpu_running, 0, 0, 0); in start_cpu() 384 cpu_datap(psip->target_cpu)->cpu_running) { in start_cpu() 410 cpu_datap(psip->target_cpu)->tsc_sync_delta = tsc_delta; in start_cpu() 443 cpu_desc_init(cpu_datap(slot_num)); in intel_startCPU() 473 if (!cpu_datap(slot_num)->cpu_running) { in intel_startCPU() 679 && pmCPUExitIdle(cpu_datap(cpu))) { in cpu_interrupt() 711 cpu_datap(cpu)->cpu_NMI_acknowledged = FALSE; in NMI_cpus() [all …]
|
| H A D | machine_check.c | 198 if (cpu_datap(master_cpu)->cpu_mca_state == NULL) { in mca_cpu_alloc() 199 mca_cpu_alloc(cpu_datap(master_cpu)); in mca_cpu_alloc() 341 if (!cpu_datap(i)->cpu_mca_state->mca_is_saved) { in mca_dump() 361 mca_state_t *mcsp = cpu_datap(i)->cpu_mca_state; in mca_dump()
|
| H A D | mp_native.c | 117 if (cpu_datap(cpu)->cpu_signals & 6) { /* (BRINGUP) */ in i386_cpu_IPI() 118 …kprintf("i386_cpu_IPI: sending enter debugger signal (%08X) to cpu %d\n", cpu_datap(cpu)->cpu_sign… in i386_cpu_IPI()
|
| H A D | cpu_threads.h | 45 #define _cpu_to_lcpu(cpu) (&cpu_datap(cpu)->lcpu) 49 #define cpu_to_lcpu(cpu) ((cpu_datap(cpu) != NULL) ? _cpu_to_lcpu(cpu) : NULL)
|
| H A D | acpi.c | 159 cpu_datap(0)->cpu_hibernate = 0; in acpi_hibernate() 438 mt_cpu_down(cpu_datap(0)); in acpi_idle_kernel() 459 cpu_datap(0)->cpu_hibernate = 0; in acpi_idle_kernel() 510 mt_cpu_up(cpu_datap(0)); in acpi_idle_kernel()
|
| H A D | i386_timer.c | 350 queue = &cpu_datap(master_cpu)->rtclock_timer.queue; in timer_queue_assign() 380 cpu_data_t *target_cdp = cpu_datap(target_cpu); in timer_queue_migrate_cpu() 422 return &cpu_datap(cpu)->rtclock_timer.queue; in timer_queue_cpu()
|
| H A D | i386_init.c | 788 cpu_desc_init(cpu_datap(0)); in vstart() 790 cpu_desc_load(cpu_datap(0)); in vstart() 793 cpu_syscall_init(cpu_datap(0)); /* cpu_syscall_init() will be in vstart() 809 cpu_desc_load(cpu_datap(cpu)); in vstart() 826 cpu_datap(cpu)->cpu_int_stack_top); in vstart() 1013 mt_cpu_up(cpu_datap(0)); in i386_init()
|
| H A D | cpu_threads.c | 343 cpup = cpu_datap(cpu); in x86_lcpu_init() 369 cpup = cpu_datap(cpu); in x86_core_alloc() 410 cpup = cpu_datap(cpu); in x86_package_find() 433 cpup = cpu_datap(cpu); in x86_die_find() 461 cpup = cpu_datap(cpu); in x86_core_find() 515 cpup = cpu_datap(cpu); in x86_die_alloc() 556 cpup = cpu_datap(cpu); in x86_package_alloc() 835 cpup = cpu_datap(cpu); in cpu_thread_alloc()
|
| H A D | pmap_pcid.h | 65 volatile uint8_t *cptr = cpu_datap(ccpu)->cpu_pmap_pcid_coherentp; in pmap_pcid_validate_current()
|
| H A D | cpu_data.h | 703 cpu_datap(int cpu) in cpu_datap() function 711 return (cpu_datap(cpu) != NULL) && (cpu_datap(cpu)->cpu_running); in cpu_is_running()
|
| H A D | pmCPU.c | 739 return cpu_datap(chosenCPU)->cpu_processor; in machine_choose_processor() 899 return cpu_datap(cpu)->cpu_hibernate; in pmCPUGetHibernate() 905 return cpu_datap(lcpu)->cpu_processor; in pmLCPUtoProcessor()
|
| H A D | hibernate_i386.c | 250 cpu_datap(0)->cpu_hibernate = 1; in hibernate_processor_setup()
|
| /xnu-12377.81.4/osfmk/arm/ |
| H A D | arm_timer.c | 237 queue = &cpu_datap(master_cpu)->rtclock_timer.queue; in timer_queue_assign() 259 return &cpu_datap(cpu)->rtclock_timer.queue; in timer_queue_cpu() 265 cpu_signal(cpu_datap(cpu), SIGPxcall, (void *) fn, arg); in timer_call_cpu() 272 cpu_signal(cpu_datap(cpu), SIGPxcall, (void *) fn, arg); in timer_call_nosync_cpu()
|
| H A D | cpu_common.c | 81 cpu_datap(int cpu) in cpu_datap() function 726 cpu_data_t *cpu_data = cpu_datap(cpu); in cpu_to_processor() 801 return cpu_datap(slot_num)->cpu_type; in slot_type() 807 return cpu_datap(slot_num)->cpu_subtype; in slot_subtype() 813 return cpu_datap(slot_num)->cpu_threadtype; in slot_threadtype() 851 return (vm_address_t)cpu_datap(cpu) - __PERCPU_ADDR(cpu_data); in other_percpu_base()
|
| H A D | cpu_data_internal.h | 317 extern cpu_data_t *cpu_datap(int cpu);
|
| /xnu-12377.81.4/osfmk/kern/ |
| H A D | kpc.h | 66 #define FIXED_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr)]) 68 #define CONFIGURABLE_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr) + kpc_fixed_c… 72 #define FIXED_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr)]) 74 #define CONFIGURABLE_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr) + kpc_fixed_c…
|
| /xnu-12377.81.4/osfmk/arm64/ |
| H A D | dbgwrap.c | 73 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_cpu_is_halted() 84 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_wait_cpu_halted() 106 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_halt_cpu() 247 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_halt_cpu_with_state()
|
| H A D | hibernate_arm64.c | 295 cpu_datap(master_cpu)->cpu_hibernate = 1; in hibernate_processor_setup()
|
| /xnu-12377.81.4/osfmk/i386/vmx/ |
| H A D | vmx_cpu.c | 240 vmx_cpu_t *cpu = &cpu_datap(i)->cpu_vmx; in vmx_allocate_vmxon_regions() 261 vmx_cpu_t *cpu = &cpu_datap(i)->cpu_vmx; in vmx_free_vmxon_regions() 280 vmx_cpu_t *cpu = &cpu_datap(i)->cpu_vmx; in vmx_globally_available()
|
| /xnu-12377.81.4/san/coverage/ |
| H A D | kcov-denylist | 26 fun:cpu_datap
|