Searched refs:VME_OBJECT (Results 1 – 11 of 11) sorted by relevance
759 object = VME_OBJECT(map_entry); in vm_map_set_cache_attr()850 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()865 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()2780 if (VME_OBJECT(entry) != object) { in vm_map_enter()2784 obj2 = VME_OBJECT(entry); in vm_map_enter()2916 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()3616 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()3621 cur_object = VME_OBJECT(entry); in vm_map_enter_fourk()4373 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()4444 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()[all …]
177 object = VME_OBJECT(entry); in vm32_region_info()400 object = VME_OBJECT(entry); in vm32_region_info_64()
342 #define VME_OBJECT(entry) \ macro403 object = VME_OBJECT(entry); in VME_OBJECT_SHADOW()406 if (object != VME_OBJECT(entry)) { in VME_OBJECT_SHADOW()
3126 VME_OBJECT(copy_entry) == VME_OBJECT(parent_copy_entry) && in mach_make_memory_entry_internal()3147 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_internal()4364 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()4378 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_get_phys_page()4383 if (VME_OBJECT(entry)->vo_shadow_offset == 0) { in vm_map_get_phys_page()4395 ((VME_OBJECT(entry)->vo_shadow_offset in vm_map_get_phys_page()4400 object = VME_OBJECT(entry); in vm_map_get_phys_page()
1045 vm_object_reference(VME_OBJECT(found)); in find_mapping_to_slide()1190 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()1197 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()2404 sr_pager = shared_region_pager_setup(VME_OBJECT(tmp_entry), VME_OFFSET(tmp_entry), si, 0); in vm_shared_region_slide_mapping()2461 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_slide_mapping()
6497 VME_OBJECT(entry) != VM_OBJECT_NULL) { in vm_map_create_upl()6498 if (VME_OBJECT(entry)->private) { in vm_map_create_upl()6502 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6525 if (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_create_upl()6526 !VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6535 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_create_upl()6628 local_object = VME_OBJECT(entry); in vm_map_create_upl()6665 assert(VME_OBJECT(entry) == local_object); in vm_map_create_upl()6792 (VME_OBJECT(entry)->shadow || in vm_map_create_upl()6793 VME_OBJECT(entry)->copy)) { in vm_map_create_upl()[all …]
6089 (VME_OBJECT(entry) != NULL) && in vm_fault_internal()6090 (VME_OBJECT(entry) == object)) { in vm_fault_internal()6120 (pmap_paddr_t)(((vm_map_offset_t) (VME_OBJECT(entry)->vo_shadow_offset)) + in vm_fault_internal()6135 (pmap_paddr_t)(((vm_map_offset_t)(VME_OBJECT(entry)->vo_shadow_offset)) + in vm_fault_internal()6276 if ((VME_OBJECT(entry) != NULL) && in vm_fault_wire()6278 VME_OBJECT(entry)->phys_contiguous) { in vm_fault_wire()6349 object = (entry->is_sub_map) ? VM_OBJECT_NULL : VME_OBJECT(entry); in vm_fault_unwire()6606 object = VME_OBJECT(entry); in vm_fault_wire_fast()7262 object = VME_OBJECT(entry); in kdp_lightweight_fault()
1282 top_object = VME_OBJECT(entry); in fill_vnodeinfoforaddr()1386 top_object = VME_OBJECT(entry); in find_vnode_object()
970 object = VME_OBJECT(oldentry); in kmem_realloc()
9711 if (VME_OBJECT(entry) == kernel_object) { in vm_page_diagnose()9713 vm_object_lock(VME_OBJECT(entry)); in vm_page_diagnose()9715 page = vm_page_lookup(VME_OBJECT(entry), offset); in vm_page_diagnose()9720 vm_object_unlock(VME_OBJECT(entry)); in vm_page_diagnose()
171 VME_OBJECT(entry)->copy_strategy = MEMORY_OBJECT_COPY_NONE; in commpage_allocate()