Lines Matching refs:VM_OBJECT_NULL

988 	if (object == VM_OBJECT_NULL) {  in vm_map_set_cache_attr()
1075 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
1091 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
2840 if (object != VM_OBJECT_NULL) { in vm_map_enter()
2976 if (object != VM_OBJECT_NULL && in vm_map_enter()
3018 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3098 (object != VM_OBJECT_NULL && in vm_map_enter()
3173 if ((obj2 == VM_OBJECT_NULL || obj2->internal) && in vm_map_enter()
3174 (object == VM_OBJECT_NULL || object->internal)) { in vm_map_enter()
3235 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3280 (object == VM_OBJECT_NULL) && in vm_map_enter()
3323 VM_OBJECT_NULL, in vm_map_enter()
3375 object == VM_OBJECT_NULL && in vm_map_enter()
3385 object != VM_OBJECT_NULL && in vm_map_enter()
3415 (object == VM_OBJECT_NULL || in vm_map_enter()
3422 object != VM_OBJECT_NULL && in vm_map_enter()
3436 object != VM_OBJECT_NULL && in vm_map_enter()
3516 assert(object == VM_OBJECT_NULL); in vm_map_enter()
3641 object != VM_OBJECT_NULL && in vm_map_enter()
3966 vm_object_t object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4034 object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4264 VM_OBJECT_NULL, in vm_map_enter_mem_object()
4286 vm_object_t copy_object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4318 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4368 if (copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4423 if (copy && copy_object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4583 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object()
4614 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4620 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4642 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
5015 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
6739 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6845 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8738 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
9296 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9559 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9671 VME_OBJECT_SET(entry, VM_OBJECT_NULL, false, 0); in vm_map_copy_overwrite_nested()
10467 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10737 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10756 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10768 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10795 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10798 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10805 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10812 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10851 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10962 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11189 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
12147 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
12242 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
12298 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
12346 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
12399 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
12400 assert(new_object->vo_copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
12426 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12469 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL, false, 0); in vm_map_copyin_internal()
12940 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13522 VME_OBJECT(old_entry) != VM_OBJECT_NULL) { in vm_map_fork()
13811 VM_OBJECT_NULL, in vm_map_exec()
14114 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
14213 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14233 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14245 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14267 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14282 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14293 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14400 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14523 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14682 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
15761 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15764 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
15954 if (cur_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
15976 if (shadow_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16559 if ((object = find_vnode_object(entry)) == VM_OBJECT_NULL) { in vm_map_willneed()
16691 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
16710 object->vo_copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16711 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16789 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
16901 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
16924 object->vo_copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
16925 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
17096 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17110 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17229 if (object == VM_OBJECT_NULL) { in vm_map_zero()
17573 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17660 if (copy_object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17753 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17915 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
17948 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
18193 VME_OBJECT(src_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18202 VME_OBJECT(new_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18217 object = VM_OBJECT_NULL; in vm_map_remap_extract()
19493 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
19902 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
19956 vm_object_t object = VM_OBJECT_NULL; in vm_map_footprint_query_page_info()
19973 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
20202 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20270 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20429 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20538 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20539 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20849 if (object == VM_OBJECT_NULL) { in vm_map_msync()
20885 (object->vo_copy == VM_OBJECT_NULL))) && in vm_map_msync()
20886 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
21753 if (object == VM_OBJECT_NULL) { in vm_map_sign()
22019 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_dump_entry_and_compressor_pager()
22348 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
22525 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()
22708 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_entry_cs_associate()
22814 cs_object->shadow != VM_OBJECT_NULL; in vm_map_entry_cs_associate()
22867 cs_object = VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23756 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
24101 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()
24124 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()