Lines Matching refs:VM_OBJECT_NULL

988 	if (object == VM_OBJECT_NULL) {  in vm_map_set_cache_attr()
1075 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
1091 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
2834 if (object != VM_OBJECT_NULL) { in vm_map_enter()
2970 if (object != VM_OBJECT_NULL && in vm_map_enter()
3012 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3092 (object != VM_OBJECT_NULL && in vm_map_enter()
3167 if ((obj2 == VM_OBJECT_NULL || obj2->internal) && in vm_map_enter()
3168 (object == VM_OBJECT_NULL || object->internal)) { in vm_map_enter()
3229 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3274 (object == VM_OBJECT_NULL) && in vm_map_enter()
3317 VM_OBJECT_NULL, in vm_map_enter()
3369 object == VM_OBJECT_NULL && in vm_map_enter()
3379 object != VM_OBJECT_NULL && in vm_map_enter()
3409 (object == VM_OBJECT_NULL || in vm_map_enter()
3416 object != VM_OBJECT_NULL && in vm_map_enter()
3430 object != VM_OBJECT_NULL && in vm_map_enter()
3510 assert(object == VM_OBJECT_NULL); in vm_map_enter()
3635 object != VM_OBJECT_NULL && in vm_map_enter()
3960 vm_object_t object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4028 object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4258 VM_OBJECT_NULL, in vm_map_enter_mem_object()
4280 vm_object_t copy_object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4312 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4362 if (copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4417 if (copy && copy_object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4577 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object()
4608 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4614 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4636 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
5009 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
6733 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6839 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8732 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
9290 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9553 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9665 VME_OBJECT_SET(entry, VM_OBJECT_NULL, false, 0); in vm_map_copy_overwrite_nested()
10461 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10731 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10750 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10762 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10789 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10792 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10799 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10806 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10845 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10956 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11183 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
12141 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
12236 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
12292 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
12340 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
12393 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
12394 assert(new_object->vo_copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
12420 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12463 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL, false, 0); in vm_map_copyin_internal()
12934 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13516 VME_OBJECT(old_entry) != VM_OBJECT_NULL) { in vm_map_fork()
13805 VM_OBJECT_NULL, in vm_map_exec()
14108 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
14207 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14227 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14239 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14261 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14276 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14287 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14394 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14517 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14676 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
15755 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15758 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
15948 if (cur_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
15970 if (shadow_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16553 if ((object = find_vnode_object(entry)) == VM_OBJECT_NULL) { in vm_map_willneed()
16685 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
16704 object->vo_copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16705 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16783 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
16895 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
16918 object->vo_copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
16919 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
17090 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17104 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17223 if (object == VM_OBJECT_NULL) { in vm_map_zero()
17567 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17654 if (copy_object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17747 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17909 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
17942 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
18187 VME_OBJECT(src_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18196 VME_OBJECT(new_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18211 object = VM_OBJECT_NULL; in vm_map_remap_extract()
19487 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
19896 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
19950 vm_object_t object = VM_OBJECT_NULL; in vm_map_footprint_query_page_info()
19967 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
20196 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20264 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20423 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20532 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20533 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20843 if (object == VM_OBJECT_NULL) { in vm_map_msync()
20879 (object->vo_copy == VM_OBJECT_NULL))) && in vm_map_msync()
20880 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
21747 if (object == VM_OBJECT_NULL) { in vm_map_sign()
22013 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_dump_entry_and_compressor_pager()
22342 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
22519 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()
22702 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_entry_cs_associate()
22808 cs_object->shadow != VM_OBJECT_NULL; in vm_map_entry_cs_associate()
22861 cs_object = VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23750 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
24095 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()
24118 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()