Searched refs:VME_OBJECT (Results 1 – 13 of 13) sorted by relevance
1009 object = VME_OBJECT(map_entry); in vm_map_set_cache_attr()1098 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()1113 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()1924 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_zap_dispose()3209 if (VME_OBJECT(entry) != object) { in vm_map_enter()3213 obj2 = VME_OBJECT(entry); in vm_map_enter()3363 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()4338 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object()4379 assert(copy_object != VME_OBJECT(copy_entry)); in vm_map_enter_mem_object()4402 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object()[all …]
599 VME_OBJECT(entry) == VM_OBJECT_NULL || in mach_make_memory_entry_copy()600 VM_OBJECT_OWNER(VME_OBJECT(entry)) == TASK_NULL) { in mach_make_memory_entry_copy()829 VME_OBJECT(copy_entry) == VME_OBJECT(parent_copy_entry) && in mach_make_memory_entry_share()851 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_share()888 if (VM_OBJECT_OWNER(VME_OBJECT(copy_entry)) == TASK_NULL) { in mach_make_memory_entry_share()889 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_share()1688 object = VME_OBJECT(entry); in mach_memory_entry_ownership()
179 object = VME_OBJECT(entry); in vm32_mach_vm_region_info()392 object = VME_OBJECT(entry); in vm32_mach_vm_region_info_64()
407 object = VME_OBJECT(entry); in VME_OBJECT_SHADOW()410 if (object != VME_OBJECT(entry)) { in VME_OBJECT_SHADOW()
6704 VME_OBJECT(entry) != VM_OBJECT_NULL) {6705 if (VME_OBJECT(entry)->private) {6709 if (VME_OBJECT(entry)->phys_contiguous) {6733 if (VME_OBJECT(entry) == VM_OBJECT_NULL ||6734 !VME_OBJECT(entry)->phys_contiguous) {6743 if (VME_OBJECT(entry) == VM_OBJECT_NULL) {6843 local_object = VME_OBJECT(entry);6881 assert(VME_OBJECT(entry) == local_object);7008 (VME_OBJECT(entry)->shadow ||7009 VME_OBJECT(entry)->vo_copy)) {[all …]
242 vm_object_t object = VME_OBJECT(entry); in __kmem_entry_orig_size()613 (VME_OBJECT(entry) == compressor_object); in __kmem_entry_validate_object()1650 vm_object_t object = VME_OBJECT(entry); in __kmem_realloc_invalid_object_size_panic()1666 vm_object_t object = VME_OBJECT(entry); in __kmem_realloc_invalid_pager_panic()1715 object = VME_OBJECT(entry); in kmem_realloc_shrink_guard()1974 object = VME_OBJECT(oldentry); in kmem_realloc_guard()2015 object = VME_OBJECT(oldentry); in kmem_realloc_guard()2387 vm_object_t object = VME_OBJECT(entry); in __kmem_free_invalid_object_size_panic()
1315 top_object = VME_OBJECT(entry); in fill_vnodeinfoforaddr()1423 top_object = VME_OBJECT(entry); in find_vnode_object()
1088 vm_object_reference(VME_OBJECT(found)); in find_mapping_to_slide()1290 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()1297 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()2720 sr_pager = shared_region_pager_setup(VME_OBJECT(tmp_entry), VME_OFFSET(tmp_entry), si, 0); in vm_shared_region_slide_mapping()2778 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_slide_mapping()
320 #define VME_OBJECT(entry) ({ assert(!(entry)->is_sub_map); _VME_OBJECT(entry); }) macro
6584 (VME_OBJECT(entry) == object)) { in vm_fault_internal()6775 VME_OBJECT(entry) != VM_OBJECT_NULL && in vm_fault_wire()6776 VME_OBJECT(entry)->phys_contiguous) { in vm_fault_wire()6855 object = (entry->is_sub_map) ? VM_OBJECT_NULL : VME_OBJECT(entry); in vm_fault_unwire()7132 object = VME_OBJECT(entry); in vm_fault_wire_fast()7829 object = VME_OBJECT(entry); in kdp_lightweight_fault()
10761 if (is_kernel_object(VME_OBJECT(entry))) { in vm_page_diagnose()10763 vm_object_lock(VME_OBJECT(entry)); in vm_page_diagnose()10765 page = vm_page_lookup(VME_OBJECT(entry), offset); in vm_page_diagnose()10770 vm_object_unlock(VME_OBJECT(entry)); in vm_page_diagnose()
2251 vm_object_t obj = VME_OBJECT(entry); in vm_get_wimg_mode()
170 VME_OBJECT(entry)->copy_strategy = MEMORY_OBJECT_COPY_NONE; in commpage_allocate()