Lines Matching refs:VM_OBJECT_NULL

761 	if (object == VM_OBJECT_NULL) {  in vm_map_set_cache_attr()
850 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
866 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
1916 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL); in vm_map_find_space()
2298 if (object != VM_OBJECT_NULL && in vm_map_enter()
2318 if (object == VM_OBJECT_NULL) { in vm_map_enter()
2418 (object != VM_OBJECT_NULL && in vm_map_enter()
2785 if ((obj2 == VM_OBJECT_NULL || in vm_map_enter()
2787 (object == VM_OBJECT_NULL || in vm_map_enter()
2844 if (object == VM_OBJECT_NULL) { in vm_map_enter()
2877 (object == VM_OBJECT_NULL) && in vm_map_enter()
2917 VM_OBJECT_NULL, in vm_map_enter()
2963 if (object == VM_OBJECT_NULL && in vm_map_enter()
2973 object != VM_OBJECT_NULL && in vm_map_enter()
3013 (object == VM_OBJECT_NULL || in vm_map_enter()
3020 object != VM_OBJECT_NULL && in vm_map_enter()
3033 object != VM_OBJECT_NULL && in vm_map_enter()
3236 object != VM_OBJECT_NULL && in vm_map_enter()
3460 fourk_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
3616 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()
3623 while (cur_object->shadow != VM_OBJECT_NULL) { in vm_map_enter_fourk()
3628 shadow_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
3700 assert(copy_object != VM_OBJECT_NULL); in vm_map_enter_fourk()
3813 object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3818 old_object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3858 object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3879 fourk_object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3902 if (fourk_object != VM_OBJECT_NULL) { in vm_map_enter_fourk()
3904 fourk_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
4126 object = VM_OBJECT_NULL; in vm_map_enter_mem_object_helper()
4341 VM_OBJECT_NULL, in vm_map_enter_mem_object_helper()
4408 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object_helper()
4565 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object_helper()
4596 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_helper()
4602 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object_helper()
4624 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object_helper()
4972 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
5172 assert(cpm_obj != VM_OBJECT_NULL); in vm_map_enter_cpm()
5178 assert(cpm_obj->shadow == VM_OBJECT_NULL); in vm_map_enter_cpm()
5244 assert(cpm_obj->shadow == VM_OBJECT_NULL); in vm_map_enter_cpm()
5803 (object->copy == VM_OBJECT_NULL) && in vm_map_submap()
5804 (object->shadow == VM_OBJECT_NULL) && in vm_map_submap()
5807 VME_OBJECT_SET(entry, VM_OBJECT_NULL); in vm_map_submap()
6824 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6925 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8363 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
8759 copy->cpy_object = VM_OBJECT_NULL; in vm_map_copy_copy()
8927 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9199 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9311 VME_OBJECT_SET(entry, VM_OBJECT_NULL); in vm_map_copy_overwrite_nested()
10021 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10251 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10267 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10279 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10306 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10309 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10316 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10323 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10362 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10471 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10688 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
11689 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
11784 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
11840 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
11882 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
11930 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
11931 assert(new_object->copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
11957 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12000 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL); in vm_map_copyin_internal()
12491 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13148 VM_OBJECT_NULL, in vm_map_exec()
13441 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_locked()
13540 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13560 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13572 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13593 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13608 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13619 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_locked()
13707 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13808 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_locked()
13956 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_locked()
15014 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15017 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
15740 if ((object = find_vnode_object(entry)) == VM_OBJECT_NULL) { in vm_map_willneed()
15864 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
15884 object->copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
15885 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
15964 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
16072 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
16080 object->copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
16081 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
16247 object == VM_OBJECT_NULL || in vm_map_pageout()
16263 object == VM_OBJECT_NULL || in vm_map_pageout()
16701 object = VM_OBJECT_NULL; in vm_map_remap_extract()
16865 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
16897 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17076 object = VM_OBJECT_NULL; in vm_map_remap_extract()
18180 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
18768 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
18832 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
18840 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
19029 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19094 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19253 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
19361 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
19362 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19660 if (object == VM_OBJECT_NULL) { in vm_map_msync()
19697 (object->copy == VM_OBJECT_NULL))) && in vm_map_msync()
19698 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
20449 if (object == VM_OBJECT_NULL) { in vm_map_sign()
20937 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
21076 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()