| /xnu-8020.121.3/osfmk/x86_64/ |
| H A D | pmap.c | 232 const pmap_t kernel_pmap = &kernel_pmap_store; variable 284 kr = pmap_enter(kernel_pmap, (vm_map_offset_t)virt, in pmap_map() 325 cdp->cpu_kernel_cr3 = kernel_pmap->pm_cr3; in pmap_cpu_init() 327 cdp->cpu_active_cr3 = kernel_pmap->pm_cr3; in pmap_cpu_init() 500 os_ref_init(&kernel_pmap->ref_count, NULL); in pmap_bootstrap() 502 kernel_pmap->nx_enabled = TRUE; in pmap_bootstrap() 504 kernel_pmap->pm_task_map = TASK_MAP_64BIT; in pmap_bootstrap() 505 kernel_pmap->pm_obj = (vm_object_t) NULL; in pmap_bootstrap() 506 kernel_pmap->pm_pml4 = IdlePML4; in pmap_bootstrap() 507 kernel_pmap->pm_upml4 = IdlePML4; in pmap_bootstrap() [all …]
|
| H A D | pmap_pcid.c | 151 &(kernel_pmap->pmap_pcid_coherency_vector[ccpu]); in pmap_pcid_configure() 281 active_pmap = kernel_pmap; in pcid_for_pmap_cpu_tuple() 326 pmap_assert(((tpmap == kernel_pmap) && new_pcid == 0) || in pmap_pcid_activate() 332 if (pml4[KERNEL_PML4_INDEX] != kernel_pmap->pm_pml4[KERNEL_PML4_INDEX]) { in pmap_pcid_activate() 342 new_pcid = kernel_pmap->pmap_pcid_cpus[ccpu]; in pmap_pcid_activate() 343 ncr3 = kernel_pmap->pm_cr3; in pmap_pcid_activate() 345 cpu_datap(ccpu)->cpu_kernel_pcid = kernel_pmap->pmap_pcid_cpus[ccpu]; in pmap_pcid_activate()
|
| H A D | copyio.c | 224 if (__improbable(((pmap != kernel_pmap) && (use_kernel_map == 0)) && in copyio() 231 if (__improbable(pmap == kernel_pmap)) { in copyio() 283 if ((pmap != kernel_pmap) && !use_kernel_map && in copyio() 388 if (!recursive_CopyIOActive && (get_cr3_raw() != kernel_pmap->pm_cr3)) { in copyio() 392 set_cr3_raw(kernel_pmap->pm_cr3); in copyio()
|
| /xnu-8020.121.3/osfmk/kdp/ml/i386/ |
| H A D | kdp_x86_common.c | 96 pmap_t src_pmap = kernel_pmap; in kdp_machine_vm_read() 127 if (!(cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst))) { in kdp_machine_vm_read() 189 if (!(cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst))) { in kdp_machine_phys_read() 241 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_vm_write() 245 if ((cur_phys_src = kdp_vtophys(kernel_pmap, cur_virt_src)) == 0) { in kdp_machine_vm_write() 301 if ((cur_phys_src = kdp_vtophys(kernel_pmap, cur_virt_src)) == 0) { in kdp_machine_phys_write() 427 debugger_ptep = pmap_pte(kernel_pmap, debugger_window_kva); in kdp_map_debug_pagetable_window() 430 pmap_expand(kernel_pmap, debugger_window_kva, PMAP_EXPAND_OPTIONS_NONE); in kdp_map_debug_pagetable_window() 431 debugger_ptep = pmap_pte(kernel_pmap, debugger_window_kva); in kdp_map_debug_pagetable_window() 443 kdp_jtag_coredump.kernel_pmap_pml4 = (uint64_t) kernel_pmap->pm_pml4; in kdp_jtag_coredump_init()
|
| /xnu-8020.121.3/osfmk/i386/ |
| H A D | pmap_x86_common.c | 345 if ((pmap != kernel_pmap) && not_in_kdp) { in pmap_find_pa() 400 if ((pmap == kernel_pmap) || in pmap_find_phys_nofault() 427 if ((pmap != kernel_pmap) && not_in_kdp) { in pmap_get_prot() 533 …ked: Missing PTE, pmap: %p, pn: 0x%x vaddr: 0x%llx kernel_pmap: %p", pmap, pn, vaddr, kernel_pmap); in pmap_update_cache_attributes_locked() 698 if (__improbable(set_NX && (pmap == kernel_pmap) && in pmap_enter_options() 772 assert(pmap != kernel_pmap); in pmap_enter_options() 837 if (pmap != kernel_pmap && !is_ept) { in pmap_enter_options() 974 if (pmap != kernel_pmap) { in pmap_enter_options() 1011 if (pmap != kernel_pmap) { in pmap_enter_options() 1086 if (kernel_pmap == pmap) { in pmap_enter_options() [all …]
|
| H A D | machine_routines.c | 172 if (pmap_get_prot(kernel_pmap, base + offset, &pageprot) == KERN_FAILURE) { in ml_static_verify_page_protections() 209 map_size = pmap_query_pagesize(kernel_pmap, vaddr_cur); in ml_static_mfree() 222 ppn = pmap_find_phys(kernel_pmap, vaddr_cur); in ml_static_mfree() 225 pmap_remove(kernel_pmap, vaddr_cur, vaddr_cur + map_size); in ml_static_mfree() 259 pmap_mark_range(kernel_pmap, vmaddr, size, NX, ro); in ml_static_protect()
|
| H A D | x86_hypercall.c | 130 kernel_pmap->pm_cr3, /* args[3]: Kernel CR3 */ in hvg_hcall_set_coredump_data()
|
| H A D | phys.c | 189 pa = ((pmap_paddr_t)pmap_find_phys(kernel_pmap, addr)) << INTEL_PGSHIFT; in kvtophys()
|
| H A D | pmap_internal.h | 400 pmap_assert(p != kernel_pmap); in pmap_pv_throttle() 802 if (pmap == kernel_pmap) { in pmap_classify_pagetable_corruption() 817 } else if ((pmap != kernel_pmap) && (!is_ept) && ((cpte & INTEL_PTE_USER) == 0)) { in pmap_classify_pagetable_corruption()
|
| H A D | Diagnostics.c | 338 rval = pmap_permissions_verify(kernel_pmap, kernel_map, 0, ~0ULL); in diagCall64()
|
| /xnu-8020.121.3/osfmk/kdp/ml/arm/ |
| H A D | kdp_vm.c | 66 if (!pmap || ((pmap != kernel_pmap) && ((va < pmap->min) || (va >= pmap->max)))) { in kdp_vtophys() 118 pmap = kernel_pmap; /* otherwise, use kernel's */ in kdp_machine_vm_read() 125 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_vm_read() 185 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_phys_read() 232 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_vm_write() 236 if ((cur_phys_src = kdp_vtophys(kernel_pmap, cur_virt_src)) == 0) { in kdp_machine_vm_write()
|
| /xnu-8020.121.3/osfmk/arm/ |
| H A D | pmap.c | 406 const pmap_t kernel_pmap = &kernel_pmap_store; variable 563 if (pmap == kernel_pmap) { in pte_set_wired() 730 if ((current != kernel_pmap) && in current_pmap() 1366 pt_entry_t * const ptep = pmap_pte(kernel_pmap, kva); in pmap_set_xprr_perm() 1370 native_pt_ops.flush_tlb_region_async(kva, PAGE_SIZE, kernel_pmap, true); in pmap_set_xprr_perm() 1448 tt_entry_t *ttep = pmap_tte(kernel_pmap, va); in pmap_set_range_xprr_perm() 1487 PMAP_UPDATE_TLBS(kernel_pmap, start, end, false, true); in pmap_set_range_xprr_perm() 1512 if (pmap != kernel_pmap) { in pmap_tt_ledger_credit() 1523 if (pmap != kernel_pmap) { in pmap_tt_ledger_debit() 1671 kr = pmap_enter(kernel_pmap, virt, (ppnum_t)atop(start), in pmap_map() [all …]
|
| H A D | machine_routines.c | 726 pmap_remove(kernel_pmap, addr, addr + sz); in ml_io_unmap() 847 ppn = pmap_find_phys(kernel_pmap, vaddr_cur); in ml_static_protect() 849 tt_entry_t *ttp = &kernel_pmap->tte[ttenum(vaddr_cur)]; in ml_static_protect() 905 ppn = pmap_find_phys(kernel_pmap, vaddr_cur); in ml_static_mfree()
|
| H A D | arm_vm_init.c | 369 pmap_init_pte_page(kernel_pmap, ppte, HIGH_EXC_VECTORS & ~ARM_TT_L1_PT_OFFMASK, 2, TRUE); in arm_vm_prot_init() 602 pmap_init_pte_page(kernel_pmap, ptp, va + off, 2, TRUE); in arm_vm_init()
|
| H A D | pcb.c | 174 thread->machine.kptw_ttb = ((unsigned int) kernel_pmap->ttep) | TTBR_SETUP; in machine_thread_create()
|
| H A D | model_dep.c | 591 print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, TRUE, filesetKC); in do_print_all_backtraces() 593 print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, FALSE, filesetKC); in do_print_all_backtraces()
|
| /xnu-8020.121.3/osfmk/arm64/ |
| H A D | platform_tests.c | 1552 ro_pn = pmap_find_phys(kernel_pmap, ro_test_va); in ctrr_test_cpu() 1553 nx_pn = pmap_find_phys(kernel_pmap, nx_test_va); in ctrr_test_cpu() 1559 prot = pmap_get_arm64_prot(kernel_pmap, ctrr_test_page); in ctrr_test_cpu() 1563 kr = pmap_enter(kernel_pmap, ctrr_test_page, ro_pn, in ctrr_test_cpu() 1569 prot = pmap_get_arm64_prot(kernel_pmap, ctrr_test_page); in ctrr_test_cpu() 1591 pmap_remove(kernel_pmap, ctrr_test_page, ctrr_test_page + PAGE_SIZE); in ctrr_test_cpu() 1595 kr = pmap_enter(kernel_pmap, ctrr_test_page, nx_pn, in ctrr_test_cpu() 1600 prot = pmap_get_arm64_prot(kernel_pmap, ctrr_test_page); in ctrr_test_cpu() 1624 pmap_remove(kernel_pmap, ctrr_test_page, ctrr_test_page + PAGE_SIZE); in ctrr_test_cpu()
|
| H A D | copyio.c | 114 return current_thread()->map->pmap == kernel_pmap; in is_kernel_to_kernel_copy()
|
| /xnu-8020.121.3/osfmk/i386/vmx/ |
| H A D | vmx_shims.c | 53 return ptoa_64(pmap_find_phys(kernel_pmap, (addr64_t)(uintptr_t)va)); in vmx_paddr()
|
| /xnu-8020.121.3/osfmk/vm/ |
| H A D | vm_kern.c | 366 assert(kernel_map && map->pmap == kernel_pmap); in kernel_memory_allocate() 651 PMAP_ENTER_CHECK(kernel_pmap, mem); in kernel_memory_populate_object_and_unlock() 665 pe_result = pmap_enter_options(kernel_pmap, in kernel_memory_populate_object_and_unlock() 675 pe_result = pmap_enter_options(kernel_pmap, in kernel_memory_populate_object_and_unlock() 783 pmap_protect(kernel_pmap, offset, offset + size, VM_PROT_NONE); in kernel_memory_depopulate() 1110 assert(map->pmap == kernel_pmap); in kmem_free() 1177 assert(parent->pmap == kernel_pmap); in kmem_suballoc()
|
| H A D | pmap.h | 702 extern const pmap_t kernel_pmap; /* The kernel's map */ 704 extern pmap_t kernel_pmap; /* The kernel's map */ 706 #define pmap_kernel() (kernel_pmap)
|
| H A D | vm_swapfile_pager.c | 387 retval = pmap_enter(kernel_pmap, in swapfile_pager_data_request() 409 pmap_remove(kernel_pmap, in swapfile_pager_data_request()
|
| /xnu-8020.121.3/osfmk/kdp/ |
| H A D | kdp_core.c | 299 pmap_find_phys(kernel_pmap, vaddr) : in kernel_pmap_present_mapping() 300 pmap_find_phys_nofault(kernel_pmap, vaddr)); in kernel_pmap_present_mapping() 516 ret = pmap_traverse_present_mappings(kernel_pmap, in kern_dump_save_summary() 543 ret = pmap_traverse_present_mappings(kernel_pmap, in kern_dump_save_seg_descriptions() 596 ret = pmap_traverse_present_mappings(kernel_pmap, in kern_dump_save_segment_data()
|
| /xnu-8020.121.3/osfmk/arm/pmap/ |
| H A D | pmap_data.c | 741 assert(ptdp->pmap != kernel_pmap); in ppr_find_eligible_pt_page() 1889 if ((pmap == NULL) || (kernel_pmap == pmap)) { in pv_alloc() 2709 if (pmap != kernel_pmap) { in ptd_alloc() 3361 if (pmap != kernel_pmap) { in validate_pmap_internal() 3389 if (pmap != kernel_pmap) { in validate_pmap_mutable_internal() 3439 if (pmap != kernel_pmap) { in pmap_require() 3730 pt_entry_t *ptep = pmap_pte(kernel_pmap, cur_va); in pmap_cpu_data_array_init()
|
| /xnu-8020.121.3/bsd/dev/ |
| H A D | memdev.c | 319 …pp = pmap_find_phys(kernel_pmap, (addr64_t)((uintptr_t)vaddr)); /* Get the sink physical address */ in mdevstrategy() 342 …pp = pmap_find_phys(kernel_pmap, (addr64_t)((uintptr_t)vaddr)); /* Get the source physical address… in mdevstrategy()
|