Lines Matching refs:VM_OBJECT_NULL
1012 if (object == VM_OBJECT_NULL) { in vm_map_set_cache_attr()
1099 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
1115 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected()
2910 if (object != VM_OBJECT_NULL) { in vm_map_enter()
3046 if (object != VM_OBJECT_NULL && in vm_map_enter()
3088 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3168 (object != VM_OBJECT_NULL && in vm_map_enter()
3249 if ((obj2 == VM_OBJECT_NULL || obj2->internal) && in vm_map_enter()
3250 (object == VM_OBJECT_NULL || object->internal)) { in vm_map_enter()
3311 if (object == VM_OBJECT_NULL) { in vm_map_enter()
3356 (object == VM_OBJECT_NULL) && in vm_map_enter()
3399 VM_OBJECT_NULL, in vm_map_enter()
3451 object == VM_OBJECT_NULL && in vm_map_enter()
3461 object != VM_OBJECT_NULL && in vm_map_enter()
3491 (object == VM_OBJECT_NULL || in vm_map_enter()
3498 object != VM_OBJECT_NULL && in vm_map_enter()
3512 object != VM_OBJECT_NULL && in vm_map_enter()
3592 assert(object == VM_OBJECT_NULL); in vm_map_enter()
3718 object != VM_OBJECT_NULL && in vm_map_enter()
4043 vm_object_t object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4111 object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4342 VM_OBJECT_NULL, in vm_map_enter_mem_object()
4364 vm_object_t copy_object = VM_OBJECT_NULL; in vm_map_enter_mem_object()
4396 copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4446 if (copy_object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
4501 if (copy && copy_object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4661 assert(object != VM_OBJECT_NULL); in vm_map_enter_mem_object()
4691 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4697 if (object != VM_OBJECT_NULL) { in vm_map_enter_mem_object()
4719 if (object != VM_OBJECT_NULL && in vm_map_enter_mem_object()
5089 if (object == VM_OBJECT_NULL) { in vm_map_enter_mem_object_control()
6822 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6928 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
8857 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
9451 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9722 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9836 VME_OBJECT_SET(entry, VM_OBJECT_NULL, false, 0); in vm_map_copy_overwrite_nested()
10639 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_unaligned()
10910 ((object == VM_OBJECT_NULL) || in vm_map_copy_overwrite_aligned()
10929 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10941 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10968 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10971 while (new_object != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10978 if (new_shadow == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
10985 if (new_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11024 if (old_object != VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11135 if (dst_object == VM_OBJECT_NULL) { in vm_map_copy_overwrite_aligned()
11364 VM_OBJECT_NULL, in vm_map_copyout_kernel_buffer()
12327 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
12424 if ((src_object == VM_OBJECT_NULL || in vm_map_copyin_internal()
12480 assert(src_object != VM_OBJECT_NULL); in vm_map_copyin_internal()
12528 if (new_object == VM_OBJECT_NULL) { in vm_map_copyin_internal()
12581 assert(new_object->shadow == VM_OBJECT_NULL); in vm_map_copyin_internal()
12582 assert(new_object->vo_copy == VM_OBJECT_NULL); in vm_map_copyin_internal()
12608 new_object = VM_OBJECT_NULL; in vm_map_copyin_internal()
12651 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL, false, 0); in vm_map_copyin_internal()
13122 } else if (object == VM_OBJECT_NULL) { in vm_map_fork_share()
13709 VME_OBJECT(old_entry) != VM_OBJECT_NULL) { in vm_map_fork()
13998 VM_OBJECT_NULL, in vm_map_exec()
14320 if (sub_object == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
14419 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14439 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14451 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14473 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14488 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14499 assert(copy_object != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14606 copy_object = VM_OBJECT_NULL; in vm_map_lookup_and_lock_object()
14729 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14888 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
15975 if (shadow_object != VM_OBJECT_NULL) { in vm_map_region_walk()
15978 shadow_object != VM_OBJECT_NULL; in vm_map_region_walk()
16168 if (cur_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16190 if (shadow_obj == VM_OBJECT_NULL) { in vm_map_region_has_obj_ref()
16852 if (object == VM_OBJECT_NULL) { in vm_map_entry_is_reusable()
16871 object->vo_copy == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16872 object->shadow == VM_OBJECT_NULL && in vm_map_entry_is_reusable()
16950 if (object != VM_OBJECT_NULL) { in vm_map_reuse_pages()
17062 if (object == VM_OBJECT_NULL) { in vm_map_reusable_pages()
17078 object->vo_copy == VM_OBJECT_NULL)) && in vm_map_reusable_pages()
17079 object->shadow == VM_OBJECT_NULL && in vm_map_reusable_pages()
17250 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17264 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_pageout()
17383 if (object == VM_OBJECT_NULL) { in vm_map_zero()
17727 object = VM_OBJECT_NULL; in vm_map_remap_extract()
17814 if (copy_object == VM_OBJECT_NULL) { in vm_map_remap_extract()
17907 object = VM_OBJECT_NULL; in vm_map_remap_extract()
18069 } else if (object != VM_OBJECT_NULL && in vm_map_remap_extract()
18102 if (object == VM_OBJECT_NULL) { in vm_map_remap_extract()
18349 VME_OBJECT(src_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18358 VME_OBJECT(new_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18373 object = VM_OBJECT_NULL; in vm_map_remap_extract()
19654 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
20098 if (object == VM_OBJECT_NULL || in vm_map_purgable_control()
20152 vm_object_t object = VM_OBJECT_NULL; in vm_map_footprint_query_page_info()
20169 if (object == VM_OBJECT_NULL) { in vm_map_footprint_query_page_info()
20398 vm_object_t object = VM_OBJECT_NULL, curr_object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20465 curr_object = object = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
20625 if (object == VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20734 if (curr_object->shadow != VM_OBJECT_NULL) { in vm_map_page_range_info_internal()
20735 vm_object_t shadow = VM_OBJECT_NULL; in vm_map_page_range_info_internal()
21047 if (object == VM_OBJECT_NULL) { in vm_map_msync()
21083 (object->vo_copy == VM_OBJECT_NULL))) && in vm_map_msync()
21084 (object->shadow == VM_OBJECT_NULL)) { in vm_map_msync()
21965 if (object == VM_OBJECT_NULL) { in vm_map_sign()
22236 if (object == VM_OBJECT_NULL || !object->internal) { in vm_map_dump_entry_and_compressor_pager()
22565 if (object == VM_OBJECT_NULL) { in vm_map_entry_should_cow_for_true_share()
22742 if (object == VM_OBJECT_NULL) { in vm_map_query_volatile()
22925 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_entry_cs_associate()
23031 cs_object->shadow != VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23084 cs_object = VM_OBJECT_NULL; in vm_map_entry_cs_associate()
23977 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
24322 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()
24345 0, VM_MAP_KERNEL_FLAGS_FIXED_PERMANENT(), VM_OBJECT_NULL, in vm_map_range_configure()