| /xnu-10063.141.1/osfmk/x86_64/ |
| H A D | pmap_pcid.c | 100 if (cpu_datap(i)) { in pmap_pcid_configure() 101 cpu_datap(i)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 105 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 114 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 119 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = TRUE; in pmap_pcid_configure() 144 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = TRUE; in pmap_pcid_configure() 149 cpu_datap(ccpu)->cpu_pmap_pcid_coherentp = in pmap_pcid_configure() 150 cpu_datap(ccpu)->cpu_pmap_pcid_coherentp_kernel = in pmap_pcid_configure() 152 cpu_datap(ccpu)->cpu_pcid_data = &pcid_data[ccpu]; in pmap_pcid_configure() 153 cpu_datap(ccpu)->cpu_pcid_data->cpu_pcid_refcounts[0] = 1; in pmap_pcid_configure() [all …]
|
| H A D | pmap.c | 2746 bool gflushed = (cpu_datap(rcpu)->cpu_tlb_invalid_global_count != in pmap_tlbi_response() 2747 cpu_datap(lcpu)->cpu_tlb_gen_counts_global[rcpu]); in pmap_tlbi_response() 2757 bool lflushed = (cpu_datap(rcpu)->cpu_tlb_invalid_local_count != in pmap_tlbi_response() 2758 cpu_datap(lcpu)->cpu_tlb_gen_counts_local[rcpu]); in pmap_tlbi_response() 2766 if ((cpu_datap(rcpu)->cpu_tlb_invalid == 0) || in pmap_tlbi_response() 2807 cpu_datap(cpu)->cpu_tlb_invalid_global = 1; in pmap_flush() 2810 cpu_datap(cpu)->cpu_tlb_invalid_local = 1; in pmap_flush() 2812 cpu_datap(my_cpu)->cpu_tlb_gen_counts_global[cpu] = cpu_datap(cpu)->cpu_tlb_invalid_global_count; in pmap_flush() 2813 cpu_datap(my_cpu)->cpu_tlb_gen_counts_local[cpu] = cpu_datap(cpu)->cpu_tlb_invalid_local_count; in pmap_flush() 2993 … cpu_datap(my_cpu)->cpu_tlb_gen_counts_global[cpu] = cpu_datap(cpu)->cpu_tlb_invalid_global_count; in pmap_flush_tlbs() [all …]
|
| /xnu-10063.141.1/osfmk/i386/ |
| H A D | cpu_topology.c | 68 cpu_data_t *cpup = cpu_datap(i); in cpu_shadow_sort() 69 ptrdiff_t coff = cpup - cpu_datap(0); in cpu_shadow_sort() 95 assert(cpu_datap(0)->cpu_number == 0); in cpu_topology_sort() 110 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 129 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 139 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 177 cpu_data_t *cpup = cpu_datap(i); in cpu_topology_sort() 244 TOPO_DBG("\tlcpu %d\n", cpu_datap(i)->cpu_number); in cpu_topology_start_cpu() 245 processor_start(cpu_datap(i)->cpu_processor); in cpu_topology_start_cpu()
|
| H A D | cpu.c | 153 cpu_data_t *cdp = cpu_datap(cpu); in cpu_exit_wait() 210 return cpu_datap(cpu)->cpu_processor; in cpu_to_processor() 223 return cpu_datap(slot_num)->cpu_type; in slot_type() 230 return cpu_datap(slot_num)->cpu_subtype; in slot_subtype() 237 return cpu_datap(slot_num)->cpu_threadtype; in slot_threadtype() 264 cpu_data_t *cpup = cpu_datap(cpu_num); in processor_to_datastring()
|
| H A D | pmap.h | 654 cpu_datap(ccpu)->cpu_task_cr3 = pcr3; in set_dirbase() 657 cpu_datap(ccpu)->cpu_ucr3 = ucr3; in set_dirbase() 660 cpu_datap(ccpu)->cpu_task_map = cpu_shadowp(ccpu)->cpu_task_map = in set_dirbase() 671 boolean_t priorpagezero = cpu_datap(ccpu)->cpu_pagezero_mapped; in set_dirbase() 672 cpu_datap(ccpu)->cpu_pagezero_mapped = nopagezero; in set_dirbase() 684 } else if (get_cr3_base() != cpu_datap(ccpu)->cpu_kernel_cr3) { in set_dirbase() 685 set_cr3_raw(cpu_datap(ccpu)->cpu_kernel_cr3); in set_dirbase() 696 if (get_cr3_base() != cpu_datap(ccpu)->cpu_kernel_cr3) { in set_dirbase() 697 set_cr3_raw(cpu_datap(ccpu)->cpu_kernel_cr3); in set_dirbase() 813 …3_raw(), map->pmap->pm_cr3, cpu_datap(ccpu)->cpu_kernel_cr3, kernel_pmap->pm_cr3, cpu_datap(ccpu)-… [all …]
|
| H A D | mp.c | 267 if (cpu_datap(slot_num)->cpu_running) { in mp_wait_for_cpu_up() 313 slot_num, cpu_datap(slot_num)->cpu_running, 0, 0, 0); in intel_startCPU_fast() 319 if (cpu_datap(slot_num)->cpu_running) { in intel_startCPU_fast() 380 cpu_datap(psip->target_cpu)->cpu_running, 0, 0, 0); in start_cpu() 383 cpu_datap(psip->target_cpu)->cpu_running) { in start_cpu() 409 cpu_datap(psip->target_cpu)->tsc_sync_delta = tsc_delta; in start_cpu() 442 cpu_desc_init(cpu_datap(slot_num)); in intel_startCPU() 472 if (!cpu_datap(slot_num)->cpu_running) { in intel_startCPU() 678 && pmCPUExitIdle(cpu_datap(cpu))) { in cpu_interrupt() 710 cpu_datap(cpu)->cpu_NMI_acknowledged = FALSE; in NMI_cpus() [all …]
|
| H A D | machine_check.c | 197 if (cpu_datap(master_cpu)->cpu_mca_state == NULL) { in mca_cpu_alloc() 198 mca_cpu_alloc(cpu_datap(master_cpu)); in mca_cpu_alloc() 340 if (!cpu_datap(i)->cpu_mca_state->mca_is_saved) { in mca_dump() 360 mca_state_t *mcsp = cpu_datap(i)->cpu_mca_state; in mca_dump()
|
| H A D | mp_native.c | 117 if (cpu_datap(cpu)->cpu_signals & 6) { /* (BRINGUP) */ in i386_cpu_IPI() 118 …kprintf("i386_cpu_IPI: sending enter debugger signal (%08X) to cpu %d\n", cpu_datap(cpu)->cpu_sign… in i386_cpu_IPI()
|
| H A D | cpu_threads.h | 45 #define _cpu_to_lcpu(cpu) (&cpu_datap(cpu)->lcpu) 49 #define cpu_to_lcpu(cpu) ((cpu_datap(cpu) != NULL) ? _cpu_to_lcpu(cpu) : NULL)
|
| H A D | acpi.c | 158 cpu_datap(0)->cpu_hibernate = 0; in acpi_hibernate() 437 mt_cpu_down(cpu_datap(0)); in acpi_idle_kernel() 458 cpu_datap(0)->cpu_hibernate = 0; in acpi_idle_kernel() 509 mt_cpu_up(cpu_datap(0)); in acpi_idle_kernel()
|
| H A D | i386_timer.c | 341 queue = &cpu_datap(master_cpu)->rtclock_timer.queue; in timer_queue_assign() 371 cpu_data_t *target_cdp = cpu_datap(target_cpu); in timer_queue_migrate_cpu() 413 return &cpu_datap(cpu)->rtclock_timer.queue; in timer_queue_cpu()
|
| H A D | i386_init.c | 786 cpu_desc_init(cpu_datap(0)); in vstart() 788 cpu_desc_load(cpu_datap(0)); in vstart() 791 cpu_syscall_init(cpu_datap(0)); /* cpu_syscall_init() will be in vstart() 807 cpu_desc_load(cpu_datap(cpu)); in vstart() 824 cpu_datap(cpu)->cpu_int_stack_top); in vstart() 1011 mt_cpu_up(cpu_datap(0)); in i386_init()
|
| H A D | cpu_threads.c | 343 cpup = cpu_datap(cpu); in x86_lcpu_init() 369 cpup = cpu_datap(cpu); in x86_core_alloc() 410 cpup = cpu_datap(cpu); in x86_package_find() 433 cpup = cpu_datap(cpu); in x86_die_find() 461 cpup = cpu_datap(cpu); in x86_core_find() 515 cpup = cpu_datap(cpu); in x86_die_alloc() 556 cpup = cpu_datap(cpu); in x86_package_alloc() 835 cpup = cpu_datap(cpu); in cpu_thread_alloc()
|
| H A D | pmap_pcid.h | 65 volatile uint8_t *cptr = cpu_datap(ccpu)->cpu_pmap_pcid_coherentp; in pmap_pcid_validate_current()
|
| H A D | cpu_data.h | 703 cpu_datap(int cpu) in cpu_datap() function 711 return (cpu_datap(cpu) != NULL) && (cpu_datap(cpu)->cpu_running); in cpu_is_running()
|
| H A D | pmCPU.c | 739 return cpu_datap(chosenCPU)->cpu_processor; in machine_choose_processor() 891 return cpu_datap(cpu)->cpu_hibernate; in pmCPUGetHibernate() 897 return cpu_datap(lcpu)->cpu_processor; in pmLCPUtoProcessor()
|
| H A D | hibernate_i386.c | 250 cpu_datap(0)->cpu_hibernate = 1; in hibernate_processor_setup()
|
| /xnu-10063.141.1/osfmk/arm/ |
| H A D | arm_timer.c | 222 queue = &cpu_datap(master_cpu)->rtclock_timer.queue; in timer_queue_assign() 244 return &cpu_datap(cpu)->rtclock_timer.queue; in timer_queue_cpu() 250 cpu_signal(cpu_datap(cpu), SIGPxcall, (void *) fn, arg); in timer_call_cpu() 257 cpu_signal(cpu_datap(cpu), SIGPxcall, (void *) fn, arg); in timer_call_nosync_cpu()
|
| H A D | cpu_common.c | 80 cpu_datap(int cpu) in cpu_datap() function 719 cpu_data_t *cpu_data = cpu_datap(cpu); in cpu_to_processor() 794 return cpu_datap(slot_num)->cpu_type; in slot_type() 800 return cpu_datap(slot_num)->cpu_subtype; in slot_subtype() 806 return cpu_datap(slot_num)->cpu_threadtype; in slot_threadtype() 842 return (char *)cpu_datap(cpu) - __PERCPU_ADDR(cpu_data); in other_percpu_base()
|
| H A D | cpu_data_internal.h | 282 extern cpu_data_t *cpu_datap(int cpu);
|
| /xnu-10063.141.1/osfmk/kern/ |
| H A D | kpc.h | 66 #define FIXED_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr)]) 68 #define CONFIGURABLE_RELOAD_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_reload[(ctr) + kpc_fixed_c… 72 #define FIXED_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr)]) 74 #define CONFIGURABLE_SHADOW_CPU(cpu, ctr) (cpu_datap(cpu)->cpu_kpc_shadow[(ctr) + kpc_fixed_c…
|
| /xnu-10063.141.1/osfmk/arm64/ |
| H A D | dbgwrap.c | 73 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_cpu_is_halted() 84 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_wait_cpu_halted() 106 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_halt_cpu() 230 cpu_data_t *cdp = cpu_datap(cpu_index); in ml_dbgwrap_halt_cpu_with_state()
|
| H A D | hibernate_arm64.c | 212 cpu_datap(master_cpu)->cpu_hibernate = 1; in hibernate_processor_setup()
|
| /xnu-10063.141.1/osfmk/i386/vmx/ |
| H A D | vmx_cpu.c | 240 vmx_cpu_t *cpu = &cpu_datap(i)->cpu_vmx; in vmx_allocate_vmxon_regions() 261 vmx_cpu_t *cpu = &cpu_datap(i)->cpu_vmx; in vmx_free_vmxon_regions() 280 vmx_cpu_t *cpu = &cpu_datap(i)->cpu_vmx; in vmx_globally_available()
|
| /xnu-10063.141.1/san/coverage/ |
| H A D | kcov-blacklist | 23 fun:cpu_datap
|