| /xnu-8792.61.2/osfmk/tests/ ! |
| H A D | pmap_tests.c | 91 if (m == VM_PAGE_NULL) { in test_pmap_enter_disconnect() 159 if (m != VM_PAGE_NULL) { in test_pmap_page_protect_overhead() 164 if ((new_pmap == NULL) || (m == VM_PAGE_NULL)) { in test_pmap_page_protect_overhead() 188 if (m != VM_PAGE_NULL) { in test_pmap_page_protect_overhead()
|
| /xnu-8792.61.2/osfmk/vm/ ! |
| H A D | vm_fault.c | 355 if (top_page != VM_PAGE_NULL) { in vm_fault_cleanup() 779 if (m != VM_PAGE_NULL) { in vm_fault_check() 801 if (m != VM_PAGE_NULL) { in vm_fault_check() 1092 first_m = VM_PAGE_NULL; in vm_fault_page() 1139 m = VM_PAGE_NULL; in vm_fault_page() 1182 if (m != VM_PAGE_NULL) { in vm_fault_page() 1227 assert(first_m == VM_PAGE_NULL); in vm_fault_page() 1337 first_m = VM_PAGE_NULL; in vm_fault_page() 1424 if (m != VM_PAGE_NULL && m->vmp_cleaning) { in vm_fault_page() 1538 if (m == VM_PAGE_NULL) { in vm_fault_page() [all …]
|
| H A D | vm_resident.c | 307 SECURITY_READ_ONLY_LATE(vm_page_t) vm_pages = VM_PAGE_NULL; 1167 bucket->page_list = VM_PAGE_PACK_PTR(VM_PAGE_NULL); in vm_page_bootstrap() 1824 assert(vm_page_lookup(object, offset) == VM_PAGE_NULL); in vm_page_insert_internal() 2023 vm_page_t found_m = VM_PAGE_NULL; in vm_page_replace() 2356 return VM_PAGE_NULL; in kdp_vm_page_lookup() 2360 return VM_PAGE_NULL; in kdp_vm_page_lookup() 2392 return VM_PAGE_NULL; in vm_page_lookup() 2397 if (mem != VM_PAGE_NULL) { in vm_page_lookup() 2457 return VM_PAGE_NULL; in vm_page_lookup() 2490 mem != VM_PAGE_NULL; in vm_page_lookup() [all …]
|
| H A D | vm_fourk_pager.c | 943 src_page = VM_PAGE_NULL; in fourk_pager_data_request() 944 top_page = VM_PAGE_NULL; in fourk_pager_data_request() 1001 assert(src_page != VM_PAGE_NULL); in fourk_pager_data_request() 1113 src_page = VM_PAGE_NULL; in fourk_pager_data_request() 1122 top_page = VM_PAGE_NULL; in fourk_pager_data_request()
|
| H A D | vm_object.c | 272 .memq_hint = VM_PAGE_NULL, 985 vm_page_t local_free_q = VM_PAGE_NULL; in vm_object_cache_evict() 1160 local_free_q = VM_PAGE_NULL; in vm_object_cache_evict() 1547 m != VM_PAGE_NULL; \ 1555 _local_free_q = VM_PAGE_NULL; \ 1567 vm_page_t local_free_q = VM_PAGE_NULL; in vm_object_reap_pages() 2157 if ((m = vm_page_lookup(object, offset)) != VM_PAGE_NULL) { in deactivate_pages_in_object() 2532 if ((m) != VM_PAGE_NULL && \ in vm_object_reuse_pages() 2813 if (p != VM_PAGE_NULL) { in vm_object_pmap_protect_options() 2979 == VM_PAGE_NULL) { in vm_object_copy_slowly() [all …]
|
| H A D | vm_kern.c | 717 *(NEXT_PAGE_PTR(m)) = VM_PAGE_NULL; in kmem_alloc_contig() 769 vm_page_t guard_left = VM_PAGE_NULL; in kmem_alloc_guard() 770 vm_page_t guard_right = VM_PAGE_NULL; in kmem_alloc_guard() 771 vm_page_t wired_page_list = VM_PAGE_NULL; in kmem_alloc_guard() 837 if (__improbable(guard_left == VM_PAGE_NULL)) { in kmem_alloc_guard() 844 if (__improbable(guard_right == VM_PAGE_NULL)) { in kmem_alloc_guard() 916 guard_left = VM_PAGE_NULL; in kmem_alloc_guard() 923 guard_right = VM_PAGE_NULL; in kmem_alloc_guard() 1455 vm_page_t guard_right = VM_PAGE_NULL; in kmem_realloc_shrink_guard() 1671 if (mem == VM_PAGE_NULL) { in kmem_realloc_guard() [all …]
|
| H A D | vm_pageout.c | 389 p = VM_PAGE_NULL; in vm_pageout_object_terminate() 394 if (m == VM_PAGE_NULL) { in vm_pageout_object_terminate() 1894 secluded_page = VM_PAGE_NULL; in vps_deal_with_secluded_page_overflow() 5739 vm_page_t dst_page = VM_PAGE_NULL; in vm_object_upl_request() 5974 if (((dst_page = vm_page_lookup(object, dst_offset)) == VM_PAGE_NULL) || in vm_object_upl_request() 6202 if (dst_page != VM_PAGE_NULL) { in vm_object_upl_request() 6260 if (dst_page != VM_PAGE_NULL) { in vm_object_upl_request() 6267 if (dst_page == VM_PAGE_NULL) { in vm_object_upl_request() 6272 if (dst_page != VM_PAGE_NULL) { in vm_object_upl_request() 6276 if (dst_page == VM_PAGE_NULL) { in vm_object_upl_request() [all …]
|
| H A D | vm_shared_region_pager.c | 578 src_page = VM_PAGE_NULL; in shared_region_pager_data_request() 623 assert(src_page != VM_PAGE_NULL); in shared_region_pager_data_request() 770 src_page = VM_PAGE_NULL; in shared_region_pager_data_request() 774 if (top_page != VM_PAGE_NULL) { in shared_region_pager_data_request()
|
| H A D | vm_apple_protect.c | 428 src_page = VM_PAGE_NULL; in apple_protect_pager_data_request() 473 assert(src_page != VM_PAGE_NULL); in apple_protect_pager_data_request() 640 src_page = VM_PAGE_NULL; in apple_protect_pager_data_request() 644 if (top_page != VM_PAGE_NULL) { in apple_protect_pager_data_request()
|
| H A D | vm_dyld_pager.c | 761 src_page = VM_PAGE_NULL; in dyld_pager_data_request() 804 assert(src_page != VM_PAGE_NULL); in dyld_pager_data_request() 893 src_page = VM_PAGE_NULL; in dyld_pager_data_request() 897 if (top_page != VM_PAGE_NULL) { in dyld_pager_data_request()
|
| H A D | vm_tests.c | 1198 m = VM_PAGE_NULL; in vm_test_collapse_overflow() 1199 while (m == VM_PAGE_NULL) { in vm_test_collapse_overflow() 1201 if (m == VM_PAGE_NULL) { in vm_test_collapse_overflow()
|
| H A D | memory_object.c | 547 while ((m = vm_page_lookup(object, offset)) != VM_PAGE_NULL) { in vm_object_update_extent() 809 page = VM_PAGE_NULL; in vm_object_update()
|
| H A D | bsd_vm.c | 235 if ((dst_page = vm_page_lookup(object, offset)) == VM_PAGE_NULL) { in memory_object_control_uiomove()
|
| H A D | vm_map.c | 2257 if (m == VM_PAGE_NULL || m->vmp_busy || m->vmp_fictitious || in vm_map_pmap_enter() 3160 *(NEXT_PAGE_PTR(m)) = VM_PAGE_NULL; in vm_map_enter() 5218 *(NEXT_PAGE_PTR(m)) = VM_PAGE_NULL; in vm_map_enter_cpm() 5301 assert(m != VM_PAGE_NULL); in vm_map_enter_cpm() 5331 if (m == VM_PAGE_NULL) { in vm_map_enter_cpm() 6776 assert(m != VM_PAGE_NULL); in vm_map_wire_nested() 6778 if (m != VM_PAGE_NULL && VM_PAGE_WIRED(m)) { in vm_map_wire_nested() 11420 if (m == VM_PAGE_NULL || !VM_PAGE_WIRED(m) || in vm_map_copyout_internal() 15425 if ((p = vm_page_lookup(object, offset)) != VM_PAGE_NULL) { in vm_map_region_look_for_page() 19238 vm_page_t m = VM_PAGE_NULL; in vm_map_page_range_info_internal() [all …]
|
| H A D | vm_page.h | 155 #define VM_PAGE_NULL ((vm_page_t) 0) macro
|
| H A D | vm_user.c | 4498 if (dst_page == VM_PAGE_NULL) { in vm_map_get_phys_page()
|
| /xnu-8792.61.2/osfmk/arm/pmap/ |
| H A D | pmap_ppl_interface.c | 129 vm_page_t mem = VM_PAGE_NULL; in pmap_release_ppl_pages_to_kernel() 130 vm_page_t local_freeq = VM_PAGE_NULL; in pmap_release_ppl_pages_to_kernel() 148 assert(mem != VM_PAGE_NULL); in pmap_release_ppl_pages_to_kernel()
|
| H A D | pmap_data.c | 1189 while (mem != VM_PAGE_NULL) { in pmap_enqueue_pages() 1196 *(NEXT_PAGE_PTR(m_prev)) = VM_PAGE_NULL; in pmap_enqueue_pages() 1274 vm_page_t mem = VM_PAGE_NULL; in pmap_pages_alloc_zeroed() 1302 while ((mem = vm_page_grab()) == VM_PAGE_NULL) { in pmap_pages_alloc_zeroed() 1310 if (mem != VM_PAGE_NULL) { in pmap_pages_alloc_zeroed() 1338 if ((mem == VM_PAGE_NULL) && (options & PMAP_PAGE_RECLAIM_NOWAIT)) { in pmap_pages_alloc_zeroed() 1347 if (mem == VM_PAGE_NULL) { in pmap_pages_alloc_zeroed() 1387 vm_page_t mem = VM_PAGE_NULL; in pmap_alloc_page_for_kern() 1389 while ((mem = vm_page_grab()) == VM_PAGE_NULL) { in pmap_alloc_page_for_kern() 1501 vm_page_t mem = VM_PAGE_NULL; in pmap_pages_free() [all …]
|
| H A D | pmap.c | 13711 if ((unwired_vm_page == VM_PAGE_NULL) || (wired_vm_page == VM_PAGE_NULL)) {
|
| /xnu-8792.61.2/osfmk/kdp/ |
| H A D | kdp_core.c | 437 vm_page_t m = VM_PAGE_NULL; in pmap_traverse_present_mappings() 462 if (m != VM_PAGE_NULL) { in pmap_traverse_present_mappings() 476 m = VM_PAGE_NULL; in pmap_traverse_present_mappings() 489 if (m == VM_PAGE_NULL) { in pmap_traverse_present_mappings()
|
| /xnu-8792.61.2/osfmk/x86_64/ |
| H A D | pmap.c | 2069 while ((m = vm_page_grab()) == VM_PAGE_NULL) { in pmap_expand_pml4() 2174 while ((m = vm_page_grab()) == VM_PAGE_NULL) { in pmap_expand_pdpt() 2298 while ((m = vm_page_grab()) == VM_PAGE_NULL) { in pmap_expand()
|
| /xnu-8792.61.2/osfmk/arm64/ |
| H A D | platform_tests.c | 1366 T_ASSERT(vm_page != VM_PAGE_NULL, NULL); in arm64_pan_test()
|
| /xnu-8792.61.2/osfmk/i386/ |
| H A D | pmap_x86_common.c | 1272 if (m == VM_PAGE_NULL) { in pmap_enter_options()
|