Searched refs:VME_OBJECT (Results 1 – 11 of 11) sorted by relevance
768 object = VME_OBJECT(map_entry); in vm_map_set_cache_attr()859 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()874 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()2787 if (VME_OBJECT(entry) != object) { in vm_map_enter()2791 obj2 = VME_OBJECT(entry); in vm_map_enter()2923 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()3621 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()3626 cur_object = VME_OBJECT(entry); in vm_map_enter_fourk()4371 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()4442 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()[all …]
177 object = VME_OBJECT(entry); in vm32_region_info()388 object = VME_OBJECT(entry); in vm32_region_info_64()
326 #define VME_OBJECT(entry) \ macro395 object = VME_OBJECT(entry); in VME_OBJECT_SHADOW()398 if (object != VME_OBJECT(entry)) { in VME_OBJECT_SHADOW()
3105 VME_OBJECT(copy_entry) == VME_OBJECT(parent_copy_entry) && in mach_make_memory_entry_internal()3126 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_internal()4177 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()4191 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_get_phys_page()4196 if (VME_OBJECT(entry)->vo_shadow_offset == 0) { in vm_map_get_phys_page()4208 ((VME_OBJECT(entry)->vo_shadow_offset in vm_map_get_phys_page()4213 object = VME_OBJECT(entry); in vm_map_get_phys_page()
1031 vm_object_reference(VME_OBJECT(found)); in find_mapping_to_slide()1180 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()1187 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()2394 sr_pager = shared_region_pager_setup(VME_OBJECT(tmp_entry), VME_OFFSET(tmp_entry), si, 0); in vm_shared_region_slide_mapping()2451 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_slide_mapping()
6489 VME_OBJECT(entry) != VM_OBJECT_NULL) { in vm_map_create_upl()6490 if (VME_OBJECT(entry)->private) { in vm_map_create_upl()6494 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6517 if (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_create_upl()6518 !VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6527 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_create_upl()6620 local_object = VME_OBJECT(entry); in vm_map_create_upl()6657 assert(VME_OBJECT(entry) == local_object); in vm_map_create_upl()6784 (VME_OBJECT(entry)->shadow || in vm_map_create_upl()6785 VME_OBJECT(entry)->copy)) { in vm_map_create_upl()[all …]
6122 (VME_OBJECT(entry) != NULL) && in vm_fault_internal()6123 (VME_OBJECT(entry) == object)) { in vm_fault_internal()6153 (pmap_paddr_t)(((vm_map_offset_t) (VME_OBJECT(entry)->vo_shadow_offset)) + in vm_fault_internal()6168 (pmap_paddr_t)(((vm_map_offset_t)(VME_OBJECT(entry)->vo_shadow_offset)) + in vm_fault_internal()6309 if ((VME_OBJECT(entry) != NULL) && in vm_fault_wire()6311 VME_OBJECT(entry)->phys_contiguous) { in vm_fault_wire()6382 object = (entry->is_sub_map) ? VM_OBJECT_NULL : VME_OBJECT(entry); in vm_fault_unwire()6639 object = VME_OBJECT(entry); in vm_fault_wire_fast()7297 object = VME_OBJECT(entry); in kdp_lightweight_fault()
1282 top_object = VME_OBJECT(entry); in fill_vnodeinfoforaddr()1386 top_object = VME_OBJECT(entry); in find_vnode_object()
897 object = VME_OBJECT(oldentry); in kmem_realloc()
9732 if (VME_OBJECT(entry) == kernel_object) { in vm_page_diagnose()9734 vm_object_lock(VME_OBJECT(entry)); in vm_page_diagnose()9736 page = vm_page_lookup(VME_OBJECT(entry), offset); in vm_page_diagnose()9741 vm_object_unlock(VME_OBJECT(entry)); in vm_page_diagnose()
171 VME_OBJECT(entry)->copy_strategy = MEMORY_OBJECT_COPY_NONE; in commpage_allocate()