Lines Matching refs:VM_OBJECT_NULL

770 	if (object == VM_OBJECT_NULL) {  in vm_map_set_cache_attr()
859 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
875 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
1918 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL); in vm_map_find_space()
2306 if (object != VM_OBJECT_NULL && in vm_map_enter()
2326 if (object == VM_OBJECT_NULL) { in vm_map_enter()
2426 (object != VM_OBJECT_NULL && in vm_map_enter()
2792 if ((obj2 == VM_OBJECT_NULL || in vm_map_enter()
2794 (object == VM_OBJECT_NULL || in vm_map_enter()
2851 if (object == VM_OBJECT_NULL) { in vm_map_enter()
2884 (object == VM_OBJECT_NULL) && in vm_map_enter()
2924 VM_OBJECT_NULL, in vm_map_enter()
2970 if (object == VM_OBJECT_NULL && in vm_map_enter()
2980 object != VM_OBJECT_NULL && in vm_map_enter()
3020 (object == VM_OBJECT_NULL || in vm_map_enter()
3027 object != VM_OBJECT_NULL && in vm_map_enter()
3040 object != VM_OBJECT_NULL && in vm_map_enter()
3243 object != VM_OBJECT_NULL && in vm_map_enter()
3466 fourk_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
3621 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()
3628 while (cur_object->shadow != VM_OBJECT_NULL) { in vm_map_enter_fourk()
3633 shadow_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
3705 assert(copy_object != VM_OBJECT_NULL); in vm_map_enter_fourk()
3818 object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3823 old_object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3863 object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3884 fourk_object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3907 if (fourk_object != VM_OBJECT_NULL) { in vm_map_enter_fourk()
3909 fourk_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
4130 object = VM_OBJECT_NULL; in vm_map_enter_mem_object_helper()
4340 VM_OBJECT_NULL, in vm_map_enter_mem_object_helper()
4406 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object_helper()
4562 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object_helper()
4592 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_helper()
4598 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object_helper()
4620 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object_helper()
4968 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
5168 assert(cpm_obj != VM_OBJECT_NULL); in vm_map_enter_cpm()
5174 assert(cpm_obj->shadow == VM_OBJECT_NULL); in vm_map_enter_cpm()
5240 assert(cpm_obj->shadow == VM_OBJECT_NULL); in vm_map_enter_cpm()
5799 (object->copy == VM_OBJECT_NULL) && in vm_map_submap()
5800 (object->shadow == VM_OBJECT_NULL) && in vm_map_submap()
5803 VME_OBJECT_SET(entry, VM_OBJECT_NULL); in vm_map_submap()
6820 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6921 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8360 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
8756 copy->cpy_object = VM_OBJECT_NULL; in vm_map_copy_copy()
8924 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9196 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9308 VME_OBJECT_SET(entry, VM_OBJECT_NULL); in vm_map_copy_overwrite_nested()
10018 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10248 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10264 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10276 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10303 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10306 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10313 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10320 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10359 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10468 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10685 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
11689 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
11784 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
11840 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
11886 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
11938 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
11939 assert(new_object->copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
11965 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12008 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL); in vm_map_copyin_internal()
12499 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13156 VM_OBJECT_NULL, in vm_map_exec()
13449 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_locked()
13548 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13568 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13580 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13601 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13616 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13627 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_locked()
13715 copy_object = VM_OBJECT_NULL; in vm_map_lookup_locked()
13816 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_locked()
13964 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_locked()
15022 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15025 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
15748 if ((object = find_vnode_object(entry)) == VM_OBJECT_NULL) { in vm_map_willneed()
15872 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
15892 object->copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
15893 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
15972 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
16080 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
16088 object->copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
16089 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
16255 object == VM_OBJECT_NULL || in vm_map_pageout()
16271 object == VM_OBJECT_NULL || in vm_map_pageout()
16711 object = VM_OBJECT_NULL; in vm_map_remap_extract()
16875 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
16907 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17089 object = VM_OBJECT_NULL; in vm_map_remap_extract()
18193 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
18768 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
18832 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
18840 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
19029 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19094 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19253 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
19361 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
19362 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19660 if (object == VM_OBJECT_NULL) { in vm_map_msync()
19697 (object->copy == VM_OBJECT_NULL))) && in vm_map_msync()
19698 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
20425 if (object == VM_OBJECT_NULL) { in vm_map_sign()
20907 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
21046 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()