Searched refs:VME_OBJECT (Results 1 – 11 of 11) sorted by relevance
800 object = VME_OBJECT(map_entry); in vm_map_set_cache_attr()891 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()906 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()1742 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_zap_dispose()2811 if (VME_OBJECT(entry) != object) { in vm_map_enter()2815 obj2 = VME_OBJECT(entry); in vm_map_enter()2947 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()3598 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()3603 cur_object = VME_OBJECT(entry); in vm_map_enter_fourk()4289 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()[all …]
177 object = VME_OBJECT(entry); in vm32_region_info()388 object = VME_OBJECT(entry); in vm32_region_info_64()
346 #define VME_OBJECT(entry) \ macro419 object = VME_OBJECT(entry); in VME_OBJECT_SHADOW()422 if (object != VME_OBJECT(entry)) { in VME_OBJECT_SHADOW()
3120 VME_OBJECT(copy_entry) == VME_OBJECT(parent_copy_entry) && in mach_make_memory_entry_internal()3141 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_internal()4251 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()4265 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_get_phys_page()4270 if (VME_OBJECT(entry)->vo_shadow_offset == 0) { in vm_map_get_phys_page()4282 ((VME_OBJECT(entry)->vo_shadow_offset in vm_map_get_phys_page()4287 object = VME_OBJECT(entry); in vm_map_get_phys_page()
1031 vm_object_reference(VME_OBJECT(found)); in find_mapping_to_slide()1180 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()1187 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()2394 sr_pager = shared_region_pager_setup(VME_OBJECT(tmp_entry), VME_OFFSET(tmp_entry), si, 0); in vm_shared_region_slide_mapping()2451 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_slide_mapping()
6487 VME_OBJECT(entry) != VM_OBJECT_NULL) { in vm_map_create_upl()6488 if (VME_OBJECT(entry)->private) { in vm_map_create_upl()6492 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6515 if (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_create_upl()6516 !VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()6525 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_create_upl()6616 local_object = VME_OBJECT(entry); in vm_map_create_upl()6653 assert(VME_OBJECT(entry) == local_object); in vm_map_create_upl()6780 (VME_OBJECT(entry)->shadow || in vm_map_create_upl()6781 VME_OBJECT(entry)->copy)) { in vm_map_create_upl()[all …]
6132 (VME_OBJECT(entry) != NULL) && in vm_fault_internal()6133 (VME_OBJECT(entry) == object)) { in vm_fault_internal()6163 (pmap_paddr_t)(((vm_map_offset_t) (VME_OBJECT(entry)->vo_shadow_offset)) + in vm_fault_internal()6178 (pmap_paddr_t)(((vm_map_offset_t)(VME_OBJECT(entry)->vo_shadow_offset)) + in vm_fault_internal()6319 if ((VME_OBJECT(entry) != NULL) && in vm_fault_wire()6321 VME_OBJECT(entry)->phys_contiguous) { in vm_fault_wire()6392 object = (entry->is_sub_map) ? VM_OBJECT_NULL : VME_OBJECT(entry); in vm_fault_unwire()6649 object = VME_OBJECT(entry); in vm_fault_wire_fast()7307 object = VME_OBJECT(entry); in kdp_lightweight_fault()
1282 top_object = VME_OBJECT(entry); in fill_vnodeinfoforaddr()1386 top_object = VME_OBJECT(entry); in find_vnode_object()
888 object = VME_OBJECT(oldentry); in kmem_realloc()995 object = VME_OBJECT(entry); in kmem_realloc_down()
9725 if (VME_OBJECT(entry) == kernel_object) { in vm_page_diagnose()9727 vm_object_lock(VME_OBJECT(entry)); in vm_page_diagnose()9729 page = vm_page_lookup(VME_OBJECT(entry), offset); in vm_page_diagnose()9734 vm_object_unlock(VME_OBJECT(entry)); in vm_page_diagnose()
171 VME_OBJECT(entry)->copy_strategy = MEMORY_OBJECT_COPY_NONE; in commpage_allocate()