Lines Matching refs:VME_OBJECT

800 	object = VME_OBJECT(map_entry);  in vm_map_set_cache_attr()
891 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
906 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()
1742 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_zap_dispose()
2811 if (VME_OBJECT(entry) != object) { in vm_map_enter()
2815 obj2 = VME_OBJECT(entry); in vm_map_enter()
2947 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()
3598 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()
3603 cur_object = VME_OBJECT(entry); in vm_map_enter_fourk()
4289 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()
4341 assert(copy_object != VME_OBJECT(copy_entry)); in vm_map_enter_mem_object_helper()
4358 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()
4465 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()
5497 if (VME_OBJECT(entry) && in vm_map_clip_start()
5499 VME_OBJECT(entry)->phys_contiguous) { in vm_map_clip_start()
5571 vm_object_reference(VME_OBJECT(new_entry)); in _vm_map_clip_start()
5615 if (VME_OBJECT(entry) && in vm_map_clip_end()
5617 VME_OBJECT(entry)->phys_contiguous) { in vm_map_clip_end()
5685 vm_object_reference(VME_OBJECT(new_entry)); in _vm_map_clip_end()
6052 …if (current->is_sub_map || (VME_OBJECT(current) == NULL) || (VME_OBJECT(current) != compressor_obj… in vm_map_protect()
6055 assert(!VME_OBJECT(current)->code_signed); in vm_map_protect()
6056 assert(VME_OBJECT(current)->copy_strategy == MEMORY_OBJECT_COPY_NONE); in vm_map_protect()
6101 if (VME_OBJECT(current) == compressor_object) { in vm_map_protect()
6730 assert(VME_OBJECT(entry)); in vm_map_wire_nested()
6735 VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
6740 object = VME_OBJECT(entry); in vm_map_wire_nested()
6836 } else if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_wire_nested()
7524 VME_OBJECT(entry) != NULL) { in vm_map_submap_pmap_clean()
7526 VME_OBJECT(entry), in vm_map_submap_pmap_clean()
7562 VME_OBJECT(entry) != NULL) { in vm_map_submap_pmap_clean()
7564 VME_OBJECT(entry), in vm_map_submap_pmap_clean()
8130 if (VME_OBJECT(&tmp_entry) == kernel_object) { in vm_map_delete()
8140 VME_OBJECT(&tmp_entry) == kernel_object, in vm_map_delete()
8251 } else if (VME_OBJECT(entry) == kernel_object || in vm_map_delete()
8252 VME_OBJECT(entry) == compressor_object) { in vm_map_delete()
8259 VME_OBJECT(entry), VME_OFFSET(entry), in vm_map_delete()
8266 } else if ((VME_OBJECT(entry) != VM_OBJECT_NULL) || in vm_map_delete()
8473 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_copy_discard()
8727 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_overwrite_submap_recurse()
8728 ((!VME_OBJECT(entry)->internal) || in vm_map_overwrite_submap_recurse()
8729 (VME_OBJECT(entry)->true_share))) { in vm_map_overwrite_submap_recurse()
8988 if ((VME_OBJECT(entry) != VM_OBJECT_NULL) && in vm_map_copy_overwrite_nested()
8989 ((!VME_OBJECT(entry)->internal) || in vm_map_copy_overwrite_nested()
8990 (VME_OBJECT(entry)->true_share))) { in vm_map_copy_overwrite_nested()
9821 dst_object = VME_OBJECT(entry); in vm_map_copy_overwrite_unaligned()
9851 VME_OBJECT(copy_entry), in vm_map_copy_overwrite_unaligned()
9884 vm_object_deallocate(VME_OBJECT(copy_entry)); in vm_map_copy_overwrite_unaligned()
10058 object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10063 vm_object_t old_object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10071 if (old_object == VME_OBJECT(copy_entry) && in vm_map_copy_overwrite_aligned()
10088 if (VME_OBJECT(copy_entry) != VM_OBJECT_NULL && in vm_map_copy_overwrite_aligned()
10089 VME_OBJECT(copy_entry)->vo_size >= __TRADEOFF1_OBJ_SIZE && in vm_map_copy_overwrite_aligned()
10114 new_object = VME_OBJECT(copy_entry); in vm_map_copy_overwrite_aligned()
10199 VME_OBJECT(entry), in vm_map_copy_overwrite_aligned()
10227 VME_OBJECT_SET(entry, VME_OBJECT(copy_entry)); in vm_map_copy_overwrite_aligned()
10228 object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10271 dst_object = VME_OBJECT(entry); in vm_map_copy_overwrite_aligned()
10314 VME_OBJECT(copy_entry), in vm_map_copy_overwrite_aligned()
10345 vm_object_deallocate(VME_OBJECT(copy_entry)); in vm_map_copy_overwrite_aligned()
10649 vm_object_reference(VME_OBJECT(new_entry)); in vm_map_copy_remap()
11001 object = VME_OBJECT(entry); in vm_map_copyout_internal()
11437 if ((VME_OBJECT(tmp_entry) != VM_OBJECT_NULL) && in vm_map_copyin_internal()
11438 (VME_OBJECT(tmp_entry)->phys_contiguous)) { in vm_map_copyin_internal()
11489 src_object = VME_OBJECT(src_entry); in vm_map_copyin_internal()
11536 VME_OBJECT(new_entry), in vm_map_copyin_internal()
11682 new_object = VME_OBJECT(new_entry); in vm_map_copyin_internal()
11755 vm_object_deallocate(VME_OBJECT(new_entry)); in vm_map_copyin_internal()
11804 if ((VME_OBJECT(src_entry) != src_object) || in vm_map_copyin_internal()
11815 vm_object_deallocate(VME_OBJECT(new_entry)); in vm_map_copyin_internal()
12228 object = VME_OBJECT(old_entry); in vm_map_fork_share()
12370 VME_OBJECT(old_entry), in vm_map_fork_share()
12387 object = VME_OBJECT(old_entry); in vm_map_fork_share()
12661 VME_OBJECT(old_entry) != NULL && in vm_map_fork()
12662 VME_OBJECT(old_entry)->pager != NULL && in vm_map_fork()
12664 VME_OBJECT(old_entry)->pager->mo_pager_ops))) { in vm_map_fork()
12704 ((VME_OBJECT(old_entry) != NULL) && in vm_map_fork()
12705 (VME_OBJECT(old_entry)->true_share))) { in vm_map_fork()
12739 VME_OBJECT(new_entry), in vm_map_fork()
12768 VME_OBJECT(old_entry), in vm_map_fork()
13192 sub_object = VME_OBJECT(submap_entry); in vm_map_lookup_locked()
13560 assert(VME_OBJECT(entry) != VM_OBJECT_NULL); in vm_map_lookup_locked()
13561 assert(VME_OBJECT(entry)->copy_strategy == MEMORY_OBJECT_COPY_SYMMETRIC); in vm_map_lookup_locked()
13670 if (VME_OBJECT(entry)->shadowed == FALSE) { in vm_map_lookup_locked()
13671 vm_object_lock(VME_OBJECT(entry)); in vm_map_lookup_locked()
13672 VME_OBJECT(entry)->shadowed = TRUE; in vm_map_lookup_locked()
13673 vm_object_unlock(VME_OBJECT(entry)); in vm_map_lookup_locked()
13708 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_lookup_locked()
13730 *object = VME_OBJECT(entry); in vm_map_lookup_locked()
14235 submap_info->object_id = VM_OBJECT_ID(VME_OBJECT(curr_entry)); in vm_map_region_recurse_64()
14245 short_info->object_id = VM_OBJECT_ID(VME_OBJECT(curr_entry)); in vm_map_region_recurse_64()
14307 …map_info->object_id_full = (vm_object_id_t) (VME_OBJECT(curr_entry) != NULL) ? VM_KERNEL_ADDRPERM( in vm_map_region_recurse_64()
14569 if (VME_OBJECT(entry) == 0 || entry->is_sub_map) { in vm_map_region_top_walk()
14583 obj = VME_OBJECT(entry); in vm_map_region_top_walk()
14672 if ((VME_OBJECT(entry) == 0) || in vm_map_region_walk()
14674 (VME_OBJECT(entry)->phys_contiguous && in vm_map_region_walk()
14699 obj = VME_OBJECT(entry); in vm_map_region_walk()
14824 obj = VME_OBJECT(entry); in vm_map_region_walk()
14949 if (VME_OBJECT(entry) == 0) { in vm_map_region_count_obj_refs()
14958 chk_obj = VME_OBJECT(entry); in vm_map_region_count_obj_refs()
15006 (VME_OBJECT(prev_entry) == VME_OBJECT(this_entry)) && in vm_map_simplify_entry()
15059 vm_object_deallocate(VME_OBJECT(prev_entry)); in vm_map_simplify_entry()
15192 if (VME_OBJECT(entry)) { in vm_map_machine_attribute()
15205 object = VME_OBJECT(entry); in vm_map_machine_attribute()
15460 object = VME_OBJECT(entry); in vm_map_willneed()
15615 object = VME_OBJECT(entry); in vm_map_entry_is_reusable()
15715 object = VME_OBJECT(entry); in vm_map_reuse_pages()
15823 object = VME_OBJECT(entry); in vm_map_reusable_pages()
15997 object = VME_OBJECT(submap_entry); in vm_map_pageout()
16013 object = VME_OBJECT(entry); in vm_map_pageout()
16523 object = VME_OBJECT(copy_entry); in vm_map_remap_extract()
16567 object = VME_OBJECT(src_entry); in vm_map_remap_extract()
16670 VME_OBJECT(src_entry), in vm_map_remap_extract()
16697 object = VME_OBJECT(src_entry); in vm_map_remap_extract()
16805 vm_object_copy_quickly(VME_OBJECT(new_entry), in vm_map_remap_extract()
16937 vm_object_deallocate(VME_OBJECT(new_entry)); in vm_map_remap_extract()
16971 VME_OBJECT(src_entry->vme_next) == VME_OBJECT(src_entry)) { in vm_map_remap_extract()
16993 vm_object_deallocate(VME_OBJECT(src_entry)); in vm_map_remap_extract()
17181 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_copy_to_physcopy()
17232 vm_object_reference(VME_OBJECT(target_entry)); in vm_map_copy_adjust_get_target_copy_map()
17289 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_copy_trim()
17513 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
17516 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
17541 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
17558 …y, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), V… in vm_map_copy_adjust_to_target()
17918 (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_remap()
17919 VME_OBJECT(entry)->internal)) { in vm_map_remap()
17930 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_remap()
18322 object = VME_OBJECT(entry); in vm_map_purgable_control()
18386 object = VME_OBJECT(map_entry); in vm_map_footprint_query_page_info()
18806 object = VME_OBJECT(map_entry); in vm_map_page_range_info_internal()
19209 object = VME_OBJECT(entry); in vm_map_msync()
19400 object = VME_OBJECT(copy_entry); in vm_named_entry_to_vm_object()
19984 object = VME_OBJECT(entry); in vm_map_sign()
20057 if (VME_OBJECT(entry) && in vm_map_partial_reap()
20059 (VME_OBJECT(entry)->internal == TRUE) && in vm_map_partial_reap()
20060 (VME_OBJECT(entry)->ref_count == 1)) { in vm_map_partial_reap()
20061 *reclaimed_resident += VME_OBJECT(entry)->resident_page_count; in vm_map_partial_reap()
20062 *reclaimed_compressed += vm_compressor_pager_get_count(VME_OBJECT(entry)->pager); in vm_map_partial_reap()
20116 if (!entry->is_sub_map && ((VME_OBJECT(entry) == 0) || in vm_map_disconnect_page_mappings()
20117 (VME_OBJECT(entry)->phys_contiguous))) { in vm_map_disconnect_page_mappings()
20251 vm_object_t src_object = VME_OBJECT(entry2); in vm_map_freeze()
20446 object = VME_OBJECT(entry); in vm_map_entry_should_cow_for_true_share()
20586 object = VME_OBJECT(entry); in vm_map_query_volatile()
20714 object = VME_OBJECT(entry); in vm_map_shadow_max()