Lines Matching refs:VM_OBJECT_NULL
1051 if (object == VM_OBJECT_NULL) { in vm_map_set_cache_attr()
1142 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
1158 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
2956 if (object != VM_OBJECT_NULL) { in vm_map_enter()
3099 if (object != VM_OBJECT_NULL && in vm_map_enter()
3144 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3198 (object != VM_OBJECT_NULL && in vm_map_enter()
3281 if ((obj2 == VM_OBJECT_NULL || obj2->internal) && in vm_map_enter()
3282 (object == VM_OBJECT_NULL || object->internal)) { in vm_map_enter()
3343 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3388 (object == VM_OBJECT_NULL) && in vm_map_enter()
3426 VM_OBJECT_NULL, in vm_map_enter()
3478 object == VM_OBJECT_NULL && in vm_map_enter()
3488 object != VM_OBJECT_NULL && in vm_map_enter()
3517 (object == VM_OBJECT_NULL || in vm_map_enter()
3524 object != VM_OBJECT_NULL && in vm_map_enter()
3538 object != VM_OBJECT_NULL && in vm_map_enter()
3631 assert(object == VM_OBJECT_NULL); in vm_map_enter()
3757 object != VM_OBJECT_NULL && in vm_map_enter()
4085 vm_object_t object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4157 object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4397 VM_OBJECT_NULL, in vm_map_enter_mem_object()
4420 vm_object_t copy_object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4452 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4502 if (copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4557 if (copy && copy_object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4717 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object()
4747 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4754 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4778 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
5207 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
6952 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
7058 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8985 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
9588 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9875 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9992 VME_OBJECT_SET(entry, VM_OBJECT_NULL, false, 0); in vm_map_copy_overwrite_nested()
10812 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
11091 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
11110 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11122 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
11149 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11152 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
11159 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11166 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11205 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11316 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11540 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
12558 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
12688 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
12746 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
12800 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
12853 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
12854 assert(new_object->vo_copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
12880 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12933 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL, false, 0); in vm_map_copyin_internal()
13406 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13898 VME_OBJECT(old_entry) != VM_OBJECT_NULL) { in vm_map_fork()
14174 VM_OBJECT_NULL, in vm_map_exec()
14487 assert(sub_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14536 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14558 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14572 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14582 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14635 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14758 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14927 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
15972 (VME_OBJECT(entry) == VM_OBJECT_NULL) || in vm_map_region_walk()
16066 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
16069 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
16259 if (cur_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16281 if (shadow_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16978 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
16997 object->vo_copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16998 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
17083 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
17204 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
17220 object->vo_copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
17221 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
17404 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17418 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17547 if (object == VM_OBJECT_NULL) { in vm_map_zero()
17895 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17982 if (copy_object == VM_OBJECT_NULL) { in vm_map_remap_extract()
18078 object = VM_OBJECT_NULL; in vm_map_remap_extract()
18240 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
18273 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
18522 VME_OBJECT(src_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18532 VME_OBJECT(new_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18548 object = VM_OBJECT_NULL; in vm_map_remap_extract()
19864 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
20332 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
20392 vm_object_t object = VM_OBJECT_NULL; in vm_map_footprint_query_page_info_exclusive()
20413 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info_exclusive()
20642 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20722 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20882 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20991 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20992 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
21325 if (object == VM_OBJECT_NULL) { in vm_map_msync()
21361 (object->vo_copy == VM_OBJECT_NULL))) && in vm_map_msync()
21362 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
22323 if (object == VM_OBJECT_NULL) { in vm_map_sign()
22612 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_dump_entry_and_compressor_pager()
22947 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
23124 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()
23322 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_entry_cs_associate()
23428 cs_object->shadow != VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23481 cs_object = VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23887 if ((!old_entry->is_sub_map && VME_OBJECT(old_entry) == VM_OBJECT_NULL) || in vm_map_corpse_footprint_collect()
24412 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
24764 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()
24792 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()
25062 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_seal()
25081 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_seal()