Home
last modified time | relevance | path

Searched refs:VME_OBJECT (Results 1 – 11 of 11) sorted by relevance

/xnu-8796.101.5/osfmk/vm/ !
H A Dvm_map.c817 object = VME_OBJECT(map_entry); in vm_map_set_cache_attr()
903 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
918 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()
1720 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_zap_dispose()
2822 if (VME_OBJECT(entry) != object) { in vm_map_enter()
2826 obj2 = VME_OBJECT(entry); in vm_map_enter()
2962 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()
3647 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()
3652 cur_object = VME_OBJECT(entry); in vm_map_enter_fourk()
4322 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()
[all …]
H A Dvm_user.c2901 VME_OBJECT(entry) == VM_OBJECT_NULL || in mach_make_memory_entry_internal()
2902 VM_OBJECT_OWNER(VME_OBJECT(entry)) == TASK_NULL) { in mach_make_memory_entry_internal()
3091 VME_OBJECT(copy_entry) == VME_OBJECT(parent_copy_entry) && in mach_make_memory_entry_internal()
3112 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_internal()
3145 VM_OBJECT_OWNER(VME_OBJECT(copy_entry)) == TASK_NULL) { in mach_make_memory_entry_internal()
3734 object = VME_OBJECT(entry); in mach_memory_entry_ownership()
4339 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
4343 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_get_phys_page()
4348 if (VME_OBJECT(entry)->vo_shadow_offset == 0) { in vm_map_get_phys_page()
4360 ((VME_OBJECT(entry)->vo_shadow_offset in vm_map_get_phys_page()
[all …]
H A Dvm_debug.c177 object = VME_OBJECT(entry); in vm32_region_info()
388 object = VME_OBJECT(entry); in vm32_region_info_64()
H A Dvm_map.h368 #define VME_OBJECT(entry) ({ assert(!(entry)->is_sub_map); _VME_OBJECT(entry); }) macro
438 object = VME_OBJECT(entry); in VME_OBJECT_SHADOW()
441 if (object != VME_OBJECT(entry)) { in VME_OBJECT_SHADOW()
H A Dvm_pageout.c6695 VME_OBJECT(entry) != VM_OBJECT_NULL) { in vm_map_create_upl()
6696 if (VME_OBJECT(entry)->private) { in vm_map_create_upl()
6700 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()
6724 if (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_create_upl()
6725 !VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()
6734 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_create_upl()
6827 local_object = VME_OBJECT(entry); in vm_map_create_upl()
6865 assert(VME_OBJECT(entry) == local_object); in vm_map_create_upl()
6992 (VME_OBJECT(entry)->shadow || in vm_map_create_upl()
6993 VME_OBJECT(entry)->copy)) { in vm_map_create_upl()
[all …]
H A Dvm_kern.c233 vm_object_t object = VME_OBJECT(entry); in __kmem_entry_orig_size()
608 (VME_OBJECT(entry) == compressor_object); in __kmem_entry_validate_object()
1493 vm_object_t object = VME_OBJECT(entry); in __kmem_realloc_invalid_object_size_panic()
1539 object = VME_OBJECT(entry); in kmem_realloc_shrink_guard()
1789 object = VME_OBJECT(oldentry); in kmem_realloc_guard()
1830 object = VME_OBJECT(oldentry); in kmem_realloc_guard()
2026 vm_object_t object = VME_OBJECT(entry); in __kmem_free_invalid_object_size_panic()
H A Dbsd_vm.c1232 top_object = VME_OBJECT(entry); in fill_vnodeinfoforaddr()
1337 top_object = VME_OBJECT(entry); in find_vnode_object()
H A Dvm_shared_region.c1075 vm_object_reference(VME_OBJECT(found)); in find_mapping_to_slide()
1255 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()
1262 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()
2682 sr_pager = shared_region_pager_setup(VME_OBJECT(tmp_entry), VME_OFFSET(tmp_entry), si, 0); in vm_shared_region_slide_mapping()
2739 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_slide_mapping()
H A Dvm_fault.c6161 (VME_OBJECT(entry) == object)) { in vm_fault_internal()
6357 VME_OBJECT(entry) != VM_OBJECT_NULL && in vm_fault_wire()
6358 VME_OBJECT(entry)->phys_contiguous) { in vm_fault_wire()
6429 object = (entry->is_sub_map) ? VM_OBJECT_NULL : VME_OBJECT(entry); in vm_fault_unwire()
6699 object = VME_OBJECT(entry); in vm_fault_wire_fast()
7366 object = VME_OBJECT(entry); in kdp_lightweight_fault()
H A Dvm_resident.c9803 if (VME_OBJECT(entry) == kernel_object) { in vm_page_diagnose()
9805 vm_object_lock(VME_OBJECT(entry)); in vm_page_diagnose()
9807 page = vm_page_lookup(VME_OBJECT(entry), offset); in vm_page_diagnose()
9812 vm_object_unlock(VME_OBJECT(entry)); in vm_page_diagnose()
/xnu-8796.101.5/osfmk/i386/commpage/ !
H A Dcommpage.c169 VME_OBJECT(entry)->copy_strategy = MEMORY_OBJECT_COPY_NONE; in commpage_allocate()