Searched refs:VME_OBJECT (Results 1 – 11 of 11) sorted by relevance
803 object = VME_OBJECT(map_entry); in vm_map_set_cache_attr()894 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()909 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()1711 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_zap_dispose()2798 if (VME_OBJECT(entry) != object) { in vm_map_enter()2802 obj2 = VME_OBJECT(entry); in vm_map_enter()2935 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()3626 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()3631 cur_object = VME_OBJECT(entry); in vm_map_enter_fourk()4314 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()[all …]
3030 VME_OBJECT(entry) == VM_OBJECT_NULL || in mach_make_memory_entry_internal()3031 VM_OBJECT_OWNER(VME_OBJECT(entry)) == TASK_NULL) { in mach_make_memory_entry_internal()3220 VME_OBJECT(copy_entry) == VME_OBJECT(parent_copy_entry) && in mach_make_memory_entry_internal()3241 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_internal()3274 VM_OBJECT_OWNER(VME_OBJECT(copy_entry)) == TASK_NULL) { in mach_make_memory_entry_internal()3863 object = VME_OBJECT(entry); in mach_memory_entry_ownership()4468 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()4472 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_get_phys_page()4477 if (VME_OBJECT(entry)->vo_shadow_offset == 0) { in vm_map_get_phys_page()4489 ((VME_OBJECT(entry)->vo_shadow_offset in vm_map_get_phys_page()[all …]
177 object = VME_OBJECT(entry); in vm32_region_info()388 object = VME_OBJECT(entry); in vm32_region_info_64()
403 #define VME_OBJECT(entry) ({ assert(!(entry)->is_sub_map); _VME_OBJECT(entry); }) macro485 object = VME_OBJECT(entry); in VME_OBJECT_SHADOW()488 if (object != VME_OBJECT(entry)) { in VME_OBJECT_SHADOW()
6698 VME_OBJECT(entry) != VM_OBJECT_NULL) { in vm_map_create_upl()6699 if (VME_OBJECT(entry)->private) { in vm_map_create_upl()6703 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6727 if (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_create_upl()6728 !VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6737 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_create_upl()6830 local_object = VME_OBJECT(entry); in vm_map_create_upl()6868 assert(VME_OBJECT(entry) == local_object); in vm_map_create_upl()6995 (VME_OBJECT(entry)->shadow || in vm_map_create_upl()6996 VME_OBJECT(entry)->copy)) { in vm_map_create_upl()[all …]
1255 top_object = VME_OBJECT(entry); in fill_vnodeinfoforaddr()1360 top_object = VME_OBJECT(entry); in find_vnode_object()
1075 vm_object_reference(VME_OBJECT(found)); in find_mapping_to_slide()1259 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()1266 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()2701 sr_pager = shared_region_pager_setup(VME_OBJECT(tmp_entry), VME_OFFSET(tmp_entry), si, 0); in vm_shared_region_slide_mapping()2760 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_slide_mapping()
512 (VME_OBJECT(entry) == compressor_object); in __kmem_entry_validate_object()1431 object = VME_OBJECT(entry); in kmem_realloc_shrink_guard()1628 object = VME_OBJECT(oldentry); in kmem_realloc_guard()
6092 (VME_OBJECT(entry) == object)) { in vm_fault_internal()6288 VME_OBJECT(entry) != VM_OBJECT_NULL && in vm_fault_wire()6289 VME_OBJECT(entry)->phys_contiguous) { in vm_fault_wire()6360 object = (entry->is_sub_map) ? VM_OBJECT_NULL : VME_OBJECT(entry); in vm_fault_unwire()6622 object = VME_OBJECT(entry); in vm_fault_wire_fast()7281 object = VME_OBJECT(entry); in kdp_lightweight_fault()
9837 if (VME_OBJECT(entry) == kernel_object) { in vm_page_diagnose()9839 vm_object_lock(VME_OBJECT(entry)); in vm_page_diagnose()9841 page = vm_page_lookup(VME_OBJECT(entry), offset); in vm_page_diagnose()9846 vm_object_unlock(VME_OBJECT(entry)); in vm_page_diagnose()
171 VME_OBJECT(entry)->copy_strategy = MEMORY_OBJECT_COPY_NONE; in commpage_allocate()