Home
last modified time | relevance | path

Searched refs:VM_PAGE_NULL (Results 1 – 23 of 23) sorted by relevance

/xnu-10002.1.13/osfmk/tests/
H A Dpmap_tests.c99 if (m == VM_PAGE_NULL) { in test_pmap_enter_disconnect()
165 if (m == VM_PAGE_NULL) { in test_pmap_compress_remove()
272 vm_page_t m1 = VM_PAGE_NULL, m2 = VM_PAGE_NULL; in test_pmap_nesting()
276 if ((m1 == VM_PAGE_NULL) || (m2 == VM_PAGE_NULL)) { in test_pmap_nesting()
422 if (m1 != VM_PAGE_NULL) { in test_pmap_nesting()
425 if (m2 != VM_PAGE_NULL) { in test_pmap_nesting()
477 if (m != VM_PAGE_NULL) { in test_pmap_page_protect_overhead()
482 if ((new_pmap == NULL) || (m == VM_PAGE_NULL)) { in test_pmap_page_protect_overhead()
506 if (m != VM_PAGE_NULL) { in test_pmap_page_protect_overhead()
/xnu-10002.1.13/osfmk/arm/pmap/
H A Dpmap_ppl_interface.c129 vm_page_t mem = VM_PAGE_NULL; in pmap_release_ppl_pages_to_kernel()
130 vm_page_t local_freeq = VM_PAGE_NULL; in pmap_release_ppl_pages_to_kernel()
148 assert(mem != VM_PAGE_NULL); in pmap_release_ppl_pages_to_kernel()
H A Dpmap_data.c1223 while (mem != VM_PAGE_NULL) { in pmap_enqueue_pages()
1230 *(NEXT_PAGE_PTR(m_prev)) = VM_PAGE_NULL; in pmap_enqueue_pages()
1306 vm_page_t mem = VM_PAGE_NULL; in pmap_pages_alloc_zeroed()
1334 while ((mem = vm_page_grab()) == VM_PAGE_NULL) { in pmap_pages_alloc_zeroed()
1342 if (mem != VM_PAGE_NULL) { in pmap_pages_alloc_zeroed()
1370 if ((mem == VM_PAGE_NULL) && (options & PMAP_PAGE_RECLAIM_NOWAIT)) { in pmap_pages_alloc_zeroed()
1379 if (mem == VM_PAGE_NULL) { in pmap_pages_alloc_zeroed()
1419 vm_page_t mem = VM_PAGE_NULL; in pmap_alloc_page_for_kern()
1426 while ((mem = vm_page_grab()) == VM_PAGE_NULL) { in pmap_alloc_page_for_kern()
1538 vm_page_t mem = VM_PAGE_NULL; in pmap_pages_free()
[all …]
H A Dpmap.c14301 if ((unwired_vm_page == VM_PAGE_NULL) || (wired_vm_page == VM_PAGE_NULL)) {
/xnu-10002.1.13/osfmk/vm/
H A Dvm_fault.c363 if (top_page != VM_PAGE_NULL) { in vm_fault_cleanup()
787 if (m != VM_PAGE_NULL) { in vm_fault_check()
809 if (m != VM_PAGE_NULL) { in vm_fault_check()
1100 first_m = VM_PAGE_NULL; in vm_fault_page()
1147 m = VM_PAGE_NULL; in vm_fault_page()
1190 if (m != VM_PAGE_NULL) { in vm_fault_page()
1235 assert(first_m == VM_PAGE_NULL); in vm_fault_page()
1345 first_m = VM_PAGE_NULL; in vm_fault_page()
1432 if (m != VM_PAGE_NULL && m->vmp_cleaning) { in vm_fault_page()
1546 if (m == VM_PAGE_NULL) { in vm_fault_page()
[all …]
H A Dvm_resident.c307 SECURITY_READ_ONLY_LATE(vm_page_t) vm_pages = VM_PAGE_NULL;
1167 bucket->page_list = VM_PAGE_PACK_PTR(VM_PAGE_NULL); in vm_page_bootstrap()
1826 assert(vm_page_lookup(object, offset) == VM_PAGE_NULL); in vm_page_insert_internal()
2025 vm_page_t found_m = VM_PAGE_NULL; in vm_page_replace()
2358 return VM_PAGE_NULL; in kdp_vm_page_lookup()
2362 return VM_PAGE_NULL; in kdp_vm_page_lookup()
2392 return VM_PAGE_NULL; in vm_page_lookup()
2397 if (mem != VM_PAGE_NULL) { in vm_page_lookup()
2457 return VM_PAGE_NULL; in vm_page_lookup()
2490 mem != VM_PAGE_NULL; in vm_page_lookup()
[all …]
H A Dvm_fourk_pager.c943 src_page = VM_PAGE_NULL; in fourk_pager_data_request()
944 top_page = VM_PAGE_NULL; in fourk_pager_data_request()
1001 assert(src_page != VM_PAGE_NULL); in fourk_pager_data_request()
1113 src_page = VM_PAGE_NULL; in fourk_pager_data_request()
1122 top_page = VM_PAGE_NULL; in fourk_pager_data_request()
H A Dvm_object.c266 .memq_hint = VM_PAGE_NULL,
992 vm_page_t local_free_q = VM_PAGE_NULL; in vm_object_cache_evict()
1167 local_free_q = VM_PAGE_NULL; in vm_object_cache_evict()
1554 m != VM_PAGE_NULL; \
1562 _local_free_q = VM_PAGE_NULL; \
1574 vm_page_t local_free_q = VM_PAGE_NULL; in vm_object_reap_pages()
2183 if ((m = vm_page_lookup(object, offset)) != VM_PAGE_NULL) {
2601 if ((m) != VM_PAGE_NULL && \
2882 if (p != VM_PAGE_NULL) {
3048 == VM_PAGE_NULL) {
[all …]
H A Dvm_pageout.c410 p = VM_PAGE_NULL; in vm_pageout_object_terminate()
415 if (m == VM_PAGE_NULL) { in vm_pageout_object_terminate()
1927 secluded_page = VM_PAGE_NULL; in vps_deal_with_secluded_page_overflow()
5796 vm_page_t dst_page = VM_PAGE_NULL; in vm_object_upl_request()
6012 if (((dst_page = vm_page_lookup(object, dst_offset)) == VM_PAGE_NULL) || in vm_object_upl_request()
6240 if (dst_page != VM_PAGE_NULL) { in vm_object_upl_request()
6298 if (dst_page != VM_PAGE_NULL) { in vm_object_upl_request()
6305 if (dst_page == VM_PAGE_NULL) { in vm_object_upl_request()
6310 if (dst_page != VM_PAGE_NULL) { in vm_object_upl_request()
6314 if (dst_page == VM_PAGE_NULL) { in vm_object_upl_request()
[all …]
H A Dvm_shared_region_pager.c578 src_page = VM_PAGE_NULL; in shared_region_pager_data_request()
623 assert(src_page != VM_PAGE_NULL); in shared_region_pager_data_request()
770 src_page = VM_PAGE_NULL; in shared_region_pager_data_request()
774 if (top_page != VM_PAGE_NULL) { in shared_region_pager_data_request()
H A Dvm_apple_protect.c428 src_page = VM_PAGE_NULL; in apple_protect_pager_data_request()
473 assert(src_page != VM_PAGE_NULL); in apple_protect_pager_data_request()
640 src_page = VM_PAGE_NULL; in apple_protect_pager_data_request()
644 if (top_page != VM_PAGE_NULL) { in apple_protect_pager_data_request()
H A Dvm_dyld_pager.c804 src_page = VM_PAGE_NULL; in dyld_pager_data_request()
847 assert(src_page != VM_PAGE_NULL); in dyld_pager_data_request()
937 src_page = VM_PAGE_NULL; in dyld_pager_data_request()
941 if (top_page != VM_PAGE_NULL) { in dyld_pager_data_request()
H A Dvm_kern.c761 vm_page_t guard_left = VM_PAGE_NULL;
762 vm_page_t guard_right = VM_PAGE_NULL;
763 vm_page_t wired_page_list = VM_PAGE_NULL;
843 if (__improbable(guard_left == VM_PAGE_NULL)) {
850 if (__improbable(guard_right == VM_PAGE_NULL)) {
934 guard_left = VM_PAGE_NULL;
941 guard_right = VM_PAGE_NULL;
1582 vm_page_t guard_right = VM_PAGE_NULL; in kmem_realloc_shrink_guard()
1876 if (mem == VM_PAGE_NULL) { in kmem_realloc_guard()
1899 vm_page_t guard_right = VM_PAGE_NULL; in kmem_realloc_guard()
[all …]
H A Dvm_tests.c1194 m = VM_PAGE_NULL; in vm_test_collapse_overflow()
1195 while (m == VM_PAGE_NULL) { in vm_test_collapse_overflow()
1197 if (m == VM_PAGE_NULL) { in vm_test_collapse_overflow()
H A Dmemory_object.c547 while ((m = vm_page_lookup(object, offset)) != VM_PAGE_NULL) { in vm_object_update_extent()
809 page = VM_PAGE_NULL; in vm_object_update()
H A Dbsd_vm.c212 if ((dst_page = vm_page_lookup(object, offset)) == VM_PAGE_NULL) { in memory_object_control_uiomove()
H A Dvm_map.c2303 if (m == VM_PAGE_NULL || m->vmp_busy || m->vmp_fictitious || in vm_map_pmap_enter()
3260 *(NEXT_PAGE_PTR(m)) = VM_PAGE_NULL; in vm_map_enter()
5280 *(NEXT_PAGE_PTR(m)) = VM_PAGE_NULL; in vm_map_enter_cpm()
5362 assert(m != VM_PAGE_NULL); in vm_map_enter_cpm()
5392 if (m == VM_PAGE_NULL) { in vm_map_enter_cpm()
6961 assert(m != VM_PAGE_NULL); in vm_map_wire_nested()
6963 if (m != VM_PAGE_NULL && VM_PAGE_WIRED(m)) { in vm_map_wire_nested()
11649 if (m == VM_PAGE_NULL || !VM_PAGE_WIRED(m) || in vm_map_copyout_internal()
15708 if ((p = vm_page_lookup(object, offset)) != VM_PAGE_NULL) { in vm_map_region_look_for_page()
19593 vm_page_t m = VM_PAGE_NULL; in vm_map_page_range_info_internal()
[all …]
H A Dvm_page.h155 #define VM_PAGE_NULL ((vm_page_t) 0) macro
H A Dvm_user.c4445 if (dst_page == VM_PAGE_NULL) { in vm_map_get_phys_page()
/xnu-10002.1.13/osfmk/kdp/
H A Dkdp_core.c439 vm_page_t m = VM_PAGE_NULL; in pmap_traverse_present_mappings()
464 if (m != VM_PAGE_NULL) { in pmap_traverse_present_mappings()
478 m = VM_PAGE_NULL; in pmap_traverse_present_mappings()
491 if (m == VM_PAGE_NULL) { in pmap_traverse_present_mappings()
/xnu-10002.1.13/osfmk/x86_64/
H A Dpmap.c2084 while ((m = vm_page_grab()) == VM_PAGE_NULL) { in pmap_expand_pml4()
2189 while ((m = vm_page_grab()) == VM_PAGE_NULL) { in pmap_expand_pdpt()
2313 while ((m = vm_page_grab()) == VM_PAGE_NULL) { in pmap_expand()
/xnu-10002.1.13/osfmk/arm64/
H A Dplatform_tests.c1366 T_ASSERT(vm_page != VM_PAGE_NULL, NULL); in arm64_pan_test()
/xnu-10002.1.13/osfmk/i386/
H A Dpmap_x86_common.c1272 if (m == VM_PAGE_NULL) { in pmap_enter_options()