Lines Matching refs:VME_OBJECT

986 	object = VME_OBJECT(map_entry);  in vm_map_set_cache_attr()
1075 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
1090 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()
1900 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_zap_dispose()
3162 if (VME_OBJECT(entry) != object) { in vm_map_enter()
3166 obj2 = VME_OBJECT(entry); in vm_map_enter()
3316 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()
4289 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object()
4330 assert(copy_object != VME_OBJECT(copy_entry)); in vm_map_enter_mem_object()
4353 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object()
4469 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object()
5266 VME_OBJECT(entry) && in vm_map_clip_start()
5267 VME_OBJECT(entry)->phys_contiguous) { in vm_map_clip_start()
5348 vm_object_reference(VME_OBJECT(new_entry)); in _vm_map_clip_start()
5393 VME_OBJECT(entry) && in vm_map_clip_end()
5394 VME_OBJECT(entry)->phys_contiguous) { in vm_map_clip_end()
5471 vm_object_reference(VME_OBJECT(new_entry)); in _vm_map_clip_end()
6003 …if (current->is_sub_map || (VME_OBJECT(current) == NULL) || (VME_OBJECT(current) != compressor_obj… in vm_map_protect()
6006 assert(!VME_OBJECT(current)->code_signed); in vm_map_protect()
6007 assert(VME_OBJECT(current)->copy_strategy == MEMORY_OBJECT_COPY_NONE); in vm_map_protect()
6728 assert(VME_OBJECT(entry)); in vm_map_wire_nested()
6733 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6738 object = VME_OBJECT(entry); in vm_map_wire_nested()
6839 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6851 } else if (VME_OBJECT(entry)->copy_strategy == MEMORY_OBJECT_COPY_SYMMETRIC) { in vm_map_wire_nested()
6872 orig_object = VME_OBJECT(entry); in vm_map_wire_nested()
6875 if (VME_OBJECT(entry) != orig_object) { in vm_map_wire_nested()
6899 if (VME_OBJECT(entry)->copy_strategy == MEMORY_OBJECT_COPY_SYMMETRIC) { in vm_map_wire_nested()
6911 object = VME_OBJECT(entry); in vm_map_wire_nested()
7673 VME_OBJECT(entry) != NULL) { in vm_map_submap_pmap_clean()
7675 VME_OBJECT(entry), in vm_map_submap_pmap_clean()
7711 VME_OBJECT(entry) != NULL) { in vm_map_submap_pmap_clean()
7713 VME_OBJECT(entry), in vm_map_submap_pmap_clean()
8718 VME_OBJECT(entry) == compressor_object) { in vm_map_delete()
8725 VME_OBJECT(entry), VME_OFFSET(entry), in vm_map_delete()
8732 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
8994 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_copy_discard()
9290 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
9291 ((!VME_OBJECT(entry)->internal) || in vm_map_overwrite_submap_recurse()
9292 (VME_OBJECT(entry)->true_share))) { in vm_map_overwrite_submap_recurse()
9553 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
9554 ((!VME_OBJECT(entry)->internal) || in vm_map_copy_overwrite_nested()
9555 (VME_OBJECT(entry)->true_share))) { in vm_map_copy_overwrite_nested()
10456 dst_object = VME_OBJECT(entry); in vm_map_copy_overwrite_unaligned()
10486 VME_OBJECT(copy_entry), in vm_map_copy_overwrite_unaligned()
10519 vm_object_deallocate(VME_OBJECT(copy_entry)); in vm_map_copy_overwrite_unaligned()
10722 object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10736 vm_object_t old_object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10745 if (old_object == VME_OBJECT(copy_entry) && in vm_map_copy_overwrite_aligned()
10762 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10763 VME_OBJECT(copy_entry)->vo_size >= __TRADEOFF1_OBJ_SIZE && in vm_map_copy_overwrite_aligned()
10788 new_object = VME_OBJECT(copy_entry); in vm_map_copy_overwrite_aligned()
10874 VME_OBJECT(entry), in vm_map_copy_overwrite_aligned()
10902 VME_OBJECT_SET(entry, VME_OBJECT(copy_entry), false, 0); in vm_map_copy_overwrite_aligned()
10903 object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10947 dst_object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10990 VME_OBJECT(copy_entry), in vm_map_copy_overwrite_aligned()
11021 vm_object_deallocate(VME_OBJECT(copy_entry)); in vm_map_copy_overwrite_aligned()
11326 vm_object_reference(VME_OBJECT(new_entry)); in vm_map_copy_remap()
11609 object = VME_OBJECT(entry); in vm_map_copyout_internal()
12141 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
12142 (VME_OBJECT(tmp_entry)->phys_contiguous)) { in vm_map_copyin_internal()
12193 src_object = VME_OBJECT(src_entry); in vm_map_copyin_internal()
12240 VME_OBJECT(new_entry), in vm_map_copyin_internal()
12389 new_object = VME_OBJECT(new_entry); in vm_map_copyin_internal()
12462 vm_object_deallocate(VME_OBJECT(new_entry)); in vm_map_copyin_internal()
12511 if ((VME_OBJECT(src_entry) != src_object) || in vm_map_copyin_internal()
12522 vm_object_deallocate(VME_OBJECT(new_entry)); in vm_map_copyin_internal()
12914 object = VME_OBJECT(old_entry); in vm_map_fork_share()
13084 VME_OBJECT(old_entry), in vm_map_fork_share()
13101 object = VME_OBJECT(old_entry); in vm_map_fork_share()
13507 VME_OBJECT(old_entry) != NULL && in vm_map_fork()
13508 VME_OBJECT(old_entry)->pager != NULL && in vm_map_fork()
13510 VME_OBJECT(old_entry)->pager->mo_pager_ops))) { in vm_map_fork()
13516 VME_OBJECT(old_entry) != VM_OBJECT_NULL) { in vm_map_fork()
13519 object = VME_OBJECT(old_entry); in vm_map_fork()
13573 ((VME_OBJECT(old_entry) != NULL) && in vm_map_fork()
13574 (VME_OBJECT(old_entry)->true_share))) { in vm_map_fork()
13608 VME_OBJECT(new_entry), in vm_map_fork()
13651 VME_OBJECT(old_entry), in vm_map_fork()
14107 sub_object = VME_OBJECT(submap_entry); in vm_map_lookup_and_lock_object()
14517 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_and_lock_object()
14518 assert(VME_OBJECT(entry)->copy_strategy == MEMORY_OBJECT_COPY_SYMMETRIC); in vm_map_lookup_and_lock_object()
14637 if (VME_OBJECT(entry)->shadowed == FALSE) { in vm_map_lookup_and_lock_object()
14638 vm_object_lock(VME_OBJECT(entry)); in vm_map_lookup_and_lock_object()
14639 VM_OBJECT_SET_SHADOWED(VME_OBJECT(entry), TRUE); in vm_map_lookup_and_lock_object()
14640 vm_object_unlock(VME_OBJECT(entry)); in vm_map_lookup_and_lock_object()
14676 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_and_lock_object()
14698 *object = VME_OBJECT(entry); in vm_map_lookup_and_lock_object()
15216 submap_info->object_id = VM_OBJECT_ID(VME_OBJECT(curr_entry)); in vm_map_region_recurse_64()
15230 short_info->object_id = VM_OBJECT_ID(VME_OBJECT(curr_entry)); in vm_map_region_recurse_64()
15290 } else if (VME_OBJECT(curr_entry)) { in vm_map_region_recurse_64()
15291 submap_info->object_id_full = (vm_object_id_t)VM_KERNEL_ADDRHASH(VME_OBJECT(curr_entry)); in vm_map_region_recurse_64()
15554 if (entry->is_sub_map || VME_OBJECT(entry) == 0) { in vm_map_region_top_walk()
15568 obj = VME_OBJECT(entry); in vm_map_region_top_walk()
15661 (VME_OBJECT(entry) == 0) || in vm_map_region_walk()
15662 (VME_OBJECT(entry)->phys_contiguous && in vm_map_region_walk()
15687 obj = VME_OBJECT(entry); in vm_map_region_walk()
15816 obj = VME_OBJECT(entry); in vm_map_region_walk()
15947 cur_obj = VME_OBJECT(entry); in vm_map_region_has_obj_ref()
16087 vm_object_deallocate(VME_OBJECT(prev_entry)); in vm_map_simplify_entry()
16243 } else if (VME_OBJECT(entry)) { in vm_map_machine_attribute()
16256 object = VME_OBJECT(entry); in vm_map_machine_attribute()
16521 object = VME_OBJECT(entry); in vm_map_willneed()
16684 object = VME_OBJECT(entry); in vm_map_entry_is_reusable()
16782 object = VME_OBJECT(entry); in vm_map_reuse_pages()
16894 object = VME_OBJECT(entry); in vm_map_reusable_pages()
17089 object = VME_OBJECT(submap_entry); in vm_map_pageout()
17103 object = VME_OBJECT(entry); in vm_map_pageout()
17215 vm_object_t object = VME_OBJECT(entry); in vm_map_zero()
17652 copy_object = VME_OBJECT(copy_entry); in vm_map_remap_extract()
17693 copy_object = VME_OBJECT(copy_entry); in vm_map_remap_extract()
17888 object = VME_OBJECT(src_entry); in vm_map_remap_extract()
18033 VME_OBJECT(src_entry), in vm_map_remap_extract()
18061 object = VME_OBJECT(src_entry); in vm_map_remap_extract()
18187 VME_OBJECT(src_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18196 VME_OBJECT(new_entry) == VM_OBJECT_NULL) { in vm_map_remap_extract()
18214 vm_object_copy_quickly(VME_OBJECT(new_entry), in vm_map_remap_extract()
18361 vm_object_deallocate(VME_OBJECT(new_entry)); in vm_map_remap_extract()
18419 vm_object_deallocate(VME_OBJECT(src_entry)); in vm_map_remap_extract()
18607 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_copy_to_physcopy()
18656 vm_object_reference(VME_OBJECT(target_entry)); in vm_map_copy_adjust_get_target_copy_map()
18713 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_copy_trim()
18954 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
18957 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
18982 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
18999 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
19487 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
19488 VME_OBJECT(entry)->internal)) { in vm_map_remap()
19895 object = VME_OBJECT(entry); in vm_map_purgable_control()
19966 object = VME_OBJECT(map_entry); in vm_map_footprint_query_page_info()
20421 object = VME_OBJECT(map_entry); in vm_map_page_range_info_internal()
20837 object = VME_OBJECT(entry); in vm_map_msync()
21025 object = VME_OBJECT(copy_entry); in vm_named_entry_to_vm_object()
21746 object = VME_OBJECT(entry); in vm_map_sign()
21820 VME_OBJECT(entry) && in vm_map_partial_reap()
21821 (VME_OBJECT(entry)->internal == TRUE) && in vm_map_partial_reap()
21822 (os_ref_get_count_raw(&VME_OBJECT(entry)->ref_count) == 1)) { in vm_map_partial_reap()
21823 *reclaimed_resident += VME_OBJECT(entry)->resident_page_count; in vm_map_partial_reap()
21824 *reclaimed_compressed += vm_compressor_pager_get_count(VME_OBJECT(entry)->pager); in vm_map_partial_reap()
21877 if (!entry->is_sub_map && ((VME_OBJECT(entry) == 0) || in vm_map_disconnect_page_mappings()
21878 (VME_OBJECT(entry)->phys_contiguous))) { in vm_map_disconnect_page_mappings()
22012 vm_object_t object = VME_OBJECT(entry); in vm_map_dump_entry_and_compressor_pager()
22127 src_object = VME_OBJECT(entry2); in vm_map_freeze()
22340 object = VME_OBJECT(entry); in vm_map_entry_should_cow_for_true_share()
22518 object = VME_OBJECT(entry); in vm_map_query_volatile()
22645 object = VME_OBJECT(entry); in vm_map_shadow_max()
22702 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_entry_cs_associate()
22800 cs_object = VME_OBJECT(entry); in vm_map_entry_cs_associate()
23750 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
23754 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_get_phys_page()
23759 if (VME_OBJECT(entry)->vo_shadow_offset == 0) { in vm_map_get_phys_page()
23771 ((VME_OBJECT(entry)->vo_shadow_offset in vm_map_get_phys_page()
23776 object = VME_OBJECT(entry); in vm_map_get_phys_page()
24340 object = VME_OBJECT(entry); in vm_map_entry_has_device_pager()