Lines Matching refs:VM_OBJECT_NULL

1011 	if (object == VM_OBJECT_NULL) {  in vm_map_set_cache_attr()
1098 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
1114 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
2876 if (object != VM_OBJECT_NULL) { in vm_map_enter()
3012 if (object != VM_OBJECT_NULL && in vm_map_enter()
3054 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3134 (object != VM_OBJECT_NULL && in vm_map_enter()
3214 if ((obj2 == VM_OBJECT_NULL || obj2->internal) && in vm_map_enter()
3215 (object == VM_OBJECT_NULL || object->internal)) { in vm_map_enter()
3276 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3321 (object == VM_OBJECT_NULL) && in vm_map_enter()
3364 VM_OBJECT_NULL, in vm_map_enter()
3416 object == VM_OBJECT_NULL && in vm_map_enter()
3426 object != VM_OBJECT_NULL && in vm_map_enter()
3456 (object == VM_OBJECT_NULL || in vm_map_enter()
3463 object != VM_OBJECT_NULL && in vm_map_enter()
3477 object != VM_OBJECT_NULL && in vm_map_enter()
3557 assert(object == VM_OBJECT_NULL); in vm_map_enter()
3683 object != VM_OBJECT_NULL && in vm_map_enter()
4008 vm_object_t object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4076 object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4307 VM_OBJECT_NULL, in vm_map_enter_mem_object()
4329 vm_object_t copy_object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4361 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4411 if (copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4466 if (copy && copy_object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4626 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object()
4656 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4662 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4684 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
5054 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
6782 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6888 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8813 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
9407 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9677 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9791 VME_OBJECT_SET(entry, VM_OBJECT_NULL, false, 0); in vm_map_copy_overwrite_nested()
10594 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10864 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10883 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10895 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10922 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10925 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10932 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10939 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10978 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11089 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11316 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
12278 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
12375 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
12431 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
12479 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
12532 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
12533 assert(new_object->vo_copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
12559 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12602 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL, false, 0); in vm_map_copyin_internal()
13073 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13656 VME_OBJECT(old_entry) != VM_OBJECT_NULL) { in vm_map_fork()
13945 VM_OBJECT_NULL, in vm_map_exec()
14267 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
14366 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14386 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14398 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14420 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14435 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14446 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14553 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14676 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14835 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
15918 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15921 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
16111 if (cur_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16133 if (shadow_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16793 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
16812 object->vo_copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16813 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16891 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
17003 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
17019 object->vo_copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
17020 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
17191 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17205 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17324 if (object == VM_OBJECT_NULL) { in vm_map_zero()
17668 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17755 if (copy_object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17848 object = VM_OBJECT_NULL; in vm_map_remap_extract()
18010 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
18043 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
18290 VME_OBJECT(src_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18299 VME_OBJECT(new_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18314 object = VM_OBJECT_NULL; in vm_map_remap_extract()
19595 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
20039 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
20093 vm_object_t object = VM_OBJECT_NULL; in vm_map_footprint_query_page_info()
20110 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
20339 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20408 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20568 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20677 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20678 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20989 if (object == VM_OBJECT_NULL) { in vm_map_msync()
21025 (object->vo_copy == VM_OBJECT_NULL))) && in vm_map_msync()
21026 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
21906 if (object == VM_OBJECT_NULL) { in vm_map_sign()
22177 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_dump_entry_and_compressor_pager()
22506 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
22683 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()
22866 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_entry_cs_associate()
22972 cs_object->shadow != VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23025 cs_object = VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23918 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
24263 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()
24286 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()