Home
last modified time | relevance | path

Searched refs:kernel_pmap (Results 1 – 25 of 49) sorted by relevance

12

/xnu-8020.121.3/osfmk/x86_64/
H A Dpmap.c232 const pmap_t kernel_pmap = &kernel_pmap_store; variable
284 kr = pmap_enter(kernel_pmap, (vm_map_offset_t)virt, in pmap_map()
325 cdp->cpu_kernel_cr3 = kernel_pmap->pm_cr3; in pmap_cpu_init()
327 cdp->cpu_active_cr3 = kernel_pmap->pm_cr3; in pmap_cpu_init()
500 os_ref_init(&kernel_pmap->ref_count, NULL); in pmap_bootstrap()
502 kernel_pmap->nx_enabled = TRUE; in pmap_bootstrap()
504 kernel_pmap->pm_task_map = TASK_MAP_64BIT; in pmap_bootstrap()
505 kernel_pmap->pm_obj = (vm_object_t) NULL; in pmap_bootstrap()
506 kernel_pmap->pm_pml4 = IdlePML4; in pmap_bootstrap()
507 kernel_pmap->pm_upml4 = IdlePML4; in pmap_bootstrap()
[all …]
H A Dpmap_pcid.c151 &(kernel_pmap->pmap_pcid_coherency_vector[ccpu]); in pmap_pcid_configure()
281 active_pmap = kernel_pmap; in pcid_for_pmap_cpu_tuple()
326 pmap_assert(((tpmap == kernel_pmap) && new_pcid == 0) || in pmap_pcid_activate()
332 if (pml4[KERNEL_PML4_INDEX] != kernel_pmap->pm_pml4[KERNEL_PML4_INDEX]) { in pmap_pcid_activate()
342 new_pcid = kernel_pmap->pmap_pcid_cpus[ccpu]; in pmap_pcid_activate()
343 ncr3 = kernel_pmap->pm_cr3; in pmap_pcid_activate()
345 cpu_datap(ccpu)->cpu_kernel_pcid = kernel_pmap->pmap_pcid_cpus[ccpu]; in pmap_pcid_activate()
H A Dcopyio.c224 if (__improbable(((pmap != kernel_pmap) && (use_kernel_map == 0)) && in copyio()
231 if (__improbable(pmap == kernel_pmap)) { in copyio()
283 if ((pmap != kernel_pmap) && !use_kernel_map && in copyio()
388 if (!recursive_CopyIOActive && (get_cr3_raw() != kernel_pmap->pm_cr3)) { in copyio()
392 set_cr3_raw(kernel_pmap->pm_cr3); in copyio()
/xnu-8020.121.3/osfmk/kdp/ml/i386/
H A Dkdp_x86_common.c96 pmap_t src_pmap = kernel_pmap; in kdp_machine_vm_read()
127 if (!(cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst))) { in kdp_machine_vm_read()
189 if (!(cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst))) { in kdp_machine_phys_read()
241 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_vm_write()
245 if ((cur_phys_src = kdp_vtophys(kernel_pmap, cur_virt_src)) == 0) { in kdp_machine_vm_write()
301 if ((cur_phys_src = kdp_vtophys(kernel_pmap, cur_virt_src)) == 0) { in kdp_machine_phys_write()
427 debugger_ptep = pmap_pte(kernel_pmap, debugger_window_kva); in kdp_map_debug_pagetable_window()
430 pmap_expand(kernel_pmap, debugger_window_kva, PMAP_EXPAND_OPTIONS_NONE); in kdp_map_debug_pagetable_window()
431 debugger_ptep = pmap_pte(kernel_pmap, debugger_window_kva); in kdp_map_debug_pagetable_window()
443 kdp_jtag_coredump.kernel_pmap_pml4 = (uint64_t) kernel_pmap->pm_pml4; in kdp_jtag_coredump_init()
/xnu-8020.121.3/osfmk/i386/
H A Dpmap_x86_common.c345 if ((pmap != kernel_pmap) && not_in_kdp) { in pmap_find_pa()
400 if ((pmap == kernel_pmap) || in pmap_find_phys_nofault()
427 if ((pmap != kernel_pmap) && not_in_kdp) { in pmap_get_prot()
533 …ked: Missing PTE, pmap: %p, pn: 0x%x vaddr: 0x%llx kernel_pmap: %p", pmap, pn, vaddr, kernel_pmap); in pmap_update_cache_attributes_locked()
698 if (__improbable(set_NX && (pmap == kernel_pmap) && in pmap_enter_options()
772 assert(pmap != kernel_pmap); in pmap_enter_options()
837 if (pmap != kernel_pmap && !is_ept) { in pmap_enter_options()
974 if (pmap != kernel_pmap) { in pmap_enter_options()
1011 if (pmap != kernel_pmap) { in pmap_enter_options()
1086 if (kernel_pmap == pmap) { in pmap_enter_options()
[all …]
H A Dmachine_routines.c172 if (pmap_get_prot(kernel_pmap, base + offset, &pageprot) == KERN_FAILURE) { in ml_static_verify_page_protections()
209 map_size = pmap_query_pagesize(kernel_pmap, vaddr_cur); in ml_static_mfree()
222 ppn = pmap_find_phys(kernel_pmap, vaddr_cur); in ml_static_mfree()
225 pmap_remove(kernel_pmap, vaddr_cur, vaddr_cur + map_size); in ml_static_mfree()
259 pmap_mark_range(kernel_pmap, vmaddr, size, NX, ro); in ml_static_protect()
H A Dx86_hypercall.c130 kernel_pmap->pm_cr3, /* args[3]: Kernel CR3 */ in hvg_hcall_set_coredump_data()
H A Dphys.c189 pa = ((pmap_paddr_t)pmap_find_phys(kernel_pmap, addr)) << INTEL_PGSHIFT; in kvtophys()
H A Dpmap_internal.h400 pmap_assert(p != kernel_pmap); in pmap_pv_throttle()
802 if (pmap == kernel_pmap) { in pmap_classify_pagetable_corruption()
817 } else if ((pmap != kernel_pmap) && (!is_ept) && ((cpte & INTEL_PTE_USER) == 0)) { in pmap_classify_pagetable_corruption()
H A DDiagnostics.c338 rval = pmap_permissions_verify(kernel_pmap, kernel_map, 0, ~0ULL); in diagCall64()
/xnu-8020.121.3/osfmk/kdp/ml/arm/
H A Dkdp_vm.c66 if (!pmap || ((pmap != kernel_pmap) && ((va < pmap->min) || (va >= pmap->max)))) { in kdp_vtophys()
118 pmap = kernel_pmap; /* otherwise, use kernel's */ in kdp_machine_vm_read()
125 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_vm_read()
185 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_phys_read()
232 if ((cur_phys_dst = kdp_vtophys(kernel_pmap, cur_virt_dst)) == 0) { in kdp_machine_vm_write()
236 if ((cur_phys_src = kdp_vtophys(kernel_pmap, cur_virt_src)) == 0) { in kdp_machine_vm_write()
/xnu-8020.121.3/osfmk/arm/
H A Dpmap.c406 const pmap_t kernel_pmap = &kernel_pmap_store; variable
563 if (pmap == kernel_pmap) { in pte_set_wired()
730 if ((current != kernel_pmap) && in current_pmap()
1366 pt_entry_t * const ptep = pmap_pte(kernel_pmap, kva); in pmap_set_xprr_perm()
1370 native_pt_ops.flush_tlb_region_async(kva, PAGE_SIZE, kernel_pmap, true); in pmap_set_xprr_perm()
1448 tt_entry_t *ttep = pmap_tte(kernel_pmap, va); in pmap_set_range_xprr_perm()
1487 PMAP_UPDATE_TLBS(kernel_pmap, start, end, false, true); in pmap_set_range_xprr_perm()
1512 if (pmap != kernel_pmap) { in pmap_tt_ledger_credit()
1523 if (pmap != kernel_pmap) { in pmap_tt_ledger_debit()
1671 kr = pmap_enter(kernel_pmap, virt, (ppnum_t)atop(start), in pmap_map()
[all …]
H A Dmachine_routines.c726 pmap_remove(kernel_pmap, addr, addr + sz); in ml_io_unmap()
847 ppn = pmap_find_phys(kernel_pmap, vaddr_cur); in ml_static_protect()
849 tt_entry_t *ttp = &kernel_pmap->tte[ttenum(vaddr_cur)]; in ml_static_protect()
905 ppn = pmap_find_phys(kernel_pmap, vaddr_cur); in ml_static_mfree()
H A Darm_vm_init.c369 pmap_init_pte_page(kernel_pmap, ppte, HIGH_EXC_VECTORS & ~ARM_TT_L1_PT_OFFMASK, 2, TRUE); in arm_vm_prot_init()
602 pmap_init_pte_page(kernel_pmap, ptp, va + off, 2, TRUE); in arm_vm_init()
H A Dpcb.c174 thread->machine.kptw_ttb = ((unsigned int) kernel_pmap->ttep) | TTBR_SETUP; in machine_thread_create()
H A Dmodel_dep.c591 print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, TRUE, filesetKC); in do_print_all_backtraces()
593 print_one_backtrace(kernel_pmap, cur_fp, nohilite_thread_marker, FALSE, filesetKC); in do_print_all_backtraces()
/xnu-8020.121.3/osfmk/arm64/
H A Dplatform_tests.c1552 ro_pn = pmap_find_phys(kernel_pmap, ro_test_va); in ctrr_test_cpu()
1553 nx_pn = pmap_find_phys(kernel_pmap, nx_test_va); in ctrr_test_cpu()
1559 prot = pmap_get_arm64_prot(kernel_pmap, ctrr_test_page); in ctrr_test_cpu()
1563 kr = pmap_enter(kernel_pmap, ctrr_test_page, ro_pn, in ctrr_test_cpu()
1569 prot = pmap_get_arm64_prot(kernel_pmap, ctrr_test_page); in ctrr_test_cpu()
1591 pmap_remove(kernel_pmap, ctrr_test_page, ctrr_test_page + PAGE_SIZE); in ctrr_test_cpu()
1595 kr = pmap_enter(kernel_pmap, ctrr_test_page, nx_pn, in ctrr_test_cpu()
1600 prot = pmap_get_arm64_prot(kernel_pmap, ctrr_test_page); in ctrr_test_cpu()
1624 pmap_remove(kernel_pmap, ctrr_test_page, ctrr_test_page + PAGE_SIZE); in ctrr_test_cpu()
H A Dcopyio.c114 return current_thread()->map->pmap == kernel_pmap; in is_kernel_to_kernel_copy()
/xnu-8020.121.3/osfmk/i386/vmx/
H A Dvmx_shims.c53 return ptoa_64(pmap_find_phys(kernel_pmap, (addr64_t)(uintptr_t)va)); in vmx_paddr()
/xnu-8020.121.3/osfmk/vm/
H A Dvm_kern.c366 assert(kernel_map && map->pmap == kernel_pmap); in kernel_memory_allocate()
651 PMAP_ENTER_CHECK(kernel_pmap, mem); in kernel_memory_populate_object_and_unlock()
665 pe_result = pmap_enter_options(kernel_pmap, in kernel_memory_populate_object_and_unlock()
675 pe_result = pmap_enter_options(kernel_pmap, in kernel_memory_populate_object_and_unlock()
783 pmap_protect(kernel_pmap, offset, offset + size, VM_PROT_NONE); in kernel_memory_depopulate()
1110 assert(map->pmap == kernel_pmap); in kmem_free()
1177 assert(parent->pmap == kernel_pmap); in kmem_suballoc()
H A Dpmap.h702 extern const pmap_t kernel_pmap; /* The kernel's map */
704 extern pmap_t kernel_pmap; /* The kernel's map */
706 #define pmap_kernel() (kernel_pmap)
H A Dvm_swapfile_pager.c387 retval = pmap_enter(kernel_pmap, in swapfile_pager_data_request()
409 pmap_remove(kernel_pmap, in swapfile_pager_data_request()
/xnu-8020.121.3/osfmk/kdp/
H A Dkdp_core.c299 pmap_find_phys(kernel_pmap, vaddr) : in kernel_pmap_present_mapping()
300 pmap_find_phys_nofault(kernel_pmap, vaddr)); in kernel_pmap_present_mapping()
516 ret = pmap_traverse_present_mappings(kernel_pmap, in kern_dump_save_summary()
543 ret = pmap_traverse_present_mappings(kernel_pmap, in kern_dump_save_seg_descriptions()
596 ret = pmap_traverse_present_mappings(kernel_pmap, in kern_dump_save_segment_data()
/xnu-8020.121.3/osfmk/arm/pmap/
H A Dpmap_data.c741 assert(ptdp->pmap != kernel_pmap); in ppr_find_eligible_pt_page()
1889 if ((pmap == NULL) || (kernel_pmap == pmap)) { in pv_alloc()
2709 if (pmap != kernel_pmap) { in ptd_alloc()
3361 if (pmap != kernel_pmap) { in validate_pmap_internal()
3389 if (pmap != kernel_pmap) { in validate_pmap_mutable_internal()
3439 if (pmap != kernel_pmap) { in pmap_require()
3730 pt_entry_t *ptep = pmap_pte(kernel_pmap, cur_va); in pmap_cpu_data_array_init()
/xnu-8020.121.3/bsd/dev/
H A Dmemdev.c319 …pp = pmap_find_phys(kernel_pmap, (addr64_t)((uintptr_t)vaddr)); /* Get the sink physical address */ in mdevstrategy()
342 …pp = pmap_find_phys(kernel_pmap, (addr64_t)((uintptr_t)vaddr)); /* Get the source physical address… in mdevstrategy()

12