Lines Matching refs:VM_OBJECT_NULL

980 	if (object == VM_OBJECT_NULL) {  in vm_map_set_cache_attr()
1067 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
1083 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
2826 if (object != VM_OBJECT_NULL) { in vm_map_enter()
2962 if (object != VM_OBJECT_NULL && in vm_map_enter()
3004 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3084 (object != VM_OBJECT_NULL && in vm_map_enter()
3159 if ((obj2 == VM_OBJECT_NULL || obj2->internal) && in vm_map_enter()
3160 (object == VM_OBJECT_NULL || object->internal)) { in vm_map_enter()
3221 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3266 (object == VM_OBJECT_NULL) && in vm_map_enter()
3309 VM_OBJECT_NULL, in vm_map_enter()
3361 object == VM_OBJECT_NULL && in vm_map_enter()
3371 object != VM_OBJECT_NULL && in vm_map_enter()
3401 (object == VM_OBJECT_NULL || in vm_map_enter()
3408 object != VM_OBJECT_NULL && in vm_map_enter()
3422 object != VM_OBJECT_NULL && in vm_map_enter()
3502 assert(object == VM_OBJECT_NULL); in vm_map_enter()
3627 object != VM_OBJECT_NULL && in vm_map_enter()
3946 vm_object_t object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4014 object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4242 VM_OBJECT_NULL, in vm_map_enter_mem_object()
4264 vm_object_t copy_object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4296 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4346 if (copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4393 if (copy && copy_object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4556 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object()
4587 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4593 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4615 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4989 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
6588 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6694 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8537 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
9077 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9338 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9450 VME_OBJECT_SET(entry, VM_OBJECT_NULL, false, 0); in vm_map_copy_overwrite_nested()
10234 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10497 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10515 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10527 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10554 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10557 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10564 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10571 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10610 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10721 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10948 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
11901 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
11996 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
12052 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
12100 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
12153 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
12154 assert(new_object->vo_copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
12180 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12223 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL, false, 0); in vm_map_copyin_internal()
12694 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13276 VME_OBJECT(old_entry) != VM_OBJECT_NULL) { in vm_map_fork()
13565 VM_OBJECT_NULL, in vm_map_exec()
13868 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
13967 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
13987 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
13999 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14021 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14036 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14047 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14154 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14272 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14431 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
15504 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15507 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
15695 if (cur_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
15717 if (shadow_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16279 if ((object = find_vnode_object(entry)) == VM_OBJECT_NULL) { in vm_map_willneed()
16411 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
16430 object->vo_copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16431 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16509 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
16621 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
16644 object->vo_copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
16645 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
16816 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
16830 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
16949 if (object == VM_OBJECT_NULL) { in vm_map_zero()
17293 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17380 if (copy_object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17473 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17641 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
17674 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17899 VME_OBJECT(src_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
17908 VME_OBJECT(new_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
17923 object = VM_OBJECT_NULL; in vm_map_remap_extract()
19197 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
19547 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
19601 vm_object_t object = VM_OBJECT_NULL; in vm_map_footprint_query_page_info()
19618 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
19834 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19903 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20062 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20170 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20171 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20473 if (object == VM_OBJECT_NULL) { in vm_map_msync()
20509 (object->vo_copy == VM_OBJECT_NULL))) && in vm_map_msync()
20510 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
21377 if (object == VM_OBJECT_NULL) { in vm_map_sign()
21643 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_dump_entry_and_compressor_pager()
21972 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
22149 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()
22332 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_entry_cs_associate()
22438 cs_object->shadow != VM_OBJECT_NULL; in vm_map_entry_cs_associate()
22491 cs_object = VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23380 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
23725 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()
23748 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()