Lines Matching refs:VM_OBJECT_NULL

805 	if (object == VM_OBJECT_NULL) {  in vm_map_set_cache_attr()
894 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
910 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
2533 if (object != VM_OBJECT_NULL && in vm_map_enter()
2554 if (object == VM_OBJECT_NULL) { in vm_map_enter()
2635 (object != VM_OBJECT_NULL && in vm_map_enter()
2803 if ((obj2 == VM_OBJECT_NULL || in vm_map_enter()
2805 (object == VM_OBJECT_NULL || in vm_map_enter()
2863 if (object == VM_OBJECT_NULL) { in vm_map_enter()
2896 (object == VM_OBJECT_NULL) && in vm_map_enter()
2936 VM_OBJECT_NULL, in vm_map_enter()
2988 object == VM_OBJECT_NULL && in vm_map_enter()
2998 object != VM_OBJECT_NULL && in vm_map_enter()
3032 (object == VM_OBJECT_NULL || in vm_map_enter()
3039 object != VM_OBJECT_NULL && in vm_map_enter()
3052 object != VM_OBJECT_NULL && in vm_map_enter()
3256 object != VM_OBJECT_NULL && in vm_map_enter()
3499 fourk_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
3626 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()
3633 while (cur_object->shadow != VM_OBJECT_NULL) { in vm_map_enter_fourk()
3638 shadow_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
3710 assert(copy_object != VM_OBJECT_NULL); in vm_map_enter_fourk()
3813 object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3818 old_object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3858 object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3879 fourk_object != VM_OBJECT_NULL && in vm_map_enter_fourk()
3902 if (fourk_object != VM_OBJECT_NULL) { in vm_map_enter_fourk()
3904 fourk_object = VM_OBJECT_NULL; in vm_map_enter_fourk()
4056 object = VM_OBJECT_NULL; in vm_map_enter_mem_object_helper()
4276 VM_OBJECT_NULL, in vm_map_enter_mem_object_helper()
4300 vm_object_t copy_object = VM_OBJECT_NULL; in vm_map_enter_mem_object_helper()
4347 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object_helper()
4388 if (copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object_helper()
4435 if (copy && copy_object != VM_OBJECT_NULL) { in vm_map_enter_mem_object_helper()
4594 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object_helper()
4625 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_helper()
4631 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object_helper()
4653 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object_helper()
5001 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
5202 assert(cpm_obj != VM_OBJECT_NULL); in vm_map_enter_cpm()
5208 assert(cpm_obj->shadow == VM_OBJECT_NULL); in vm_map_enter_cpm()
5274 assert(cpm_obj->shadow == VM_OBJECT_NULL); in vm_map_enter_cpm()
6762 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6864 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8583 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
8897 copy->cpy_object = VM_OBJECT_NULL; in vm_map_copy_copy()
9096 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9357 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9469 VME_OBJECT_SET(entry, VM_OBJECT_NULL, false, 0); in vm_map_copy_overwrite_nested()
10207 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10441 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10457 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10469 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10496 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10499 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10506 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10513 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10552 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10663 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10892 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
11829 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
11924 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
11980 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
12026 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
12078 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
12079 assert(new_object->copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
12105 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12148 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL, false, 0); in vm_map_copyin_internal()
12645 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13432 VM_OBJECT_NULL, in vm_map_exec()
13727 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
13826 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
13846 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
13858 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
13879 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
13894 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
13905 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
13996 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14098 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14247 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
15319 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15322 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
15502 if (entry->is_sub_map || VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_region_count_obj_refs()
16044 if ((object = find_vnode_object(entry)) == VM_OBJECT_NULL) { in vm_map_willneed()
16168 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
16188 object->copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16189 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16267 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
16374 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
16382 object->copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
16383 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
16553 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
16567 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
16975 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17137 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
17169 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17372 object = VM_OBJECT_NULL; in vm_map_remap_extract()
18561 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
18977 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
19031 vm_object_t object = VM_OBJECT_NULL; in vm_map_footprint_query_page_info()
19048 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
19237 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19302 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19461 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
19569 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
19570 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
19868 if (object == VM_OBJECT_NULL) { in vm_map_msync()
19905 (object->copy == VM_OBJECT_NULL))) && in vm_map_msync()
19906 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
20661 if (object == VM_OBJECT_NULL) { in vm_map_sign()
21128 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
21267 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()