Home
last modified time | relevance | path

Searched refs:VME_OBJECT (Results 1 – 11 of 11) sorted by relevance

/xnu-10063.141.1/osfmk/vm/
H A Dvm_map.c850 object = VME_OBJECT(map_entry); in vm_map_set_cache_attr()
939 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected()
954 protected_object = VME_OBJECT(map_entry); in vm_map_apple_protected()
1776 vm_object_deallocate(VME_OBJECT(entry)); in vm_map_zap_dispose()
2965 if (VME_OBJECT(entry) != object) { in vm_map_enter()
2969 obj2 = VME_OBJECT(entry); in vm_map_enter()
3113 if (vm_object_coalesce(VME_OBJECT(entry), in vm_map_enter()
3798 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_enter_fourk()
3803 cur_object = VME_OBJECT(entry); in vm_map_enter_fourk()
4479 copy_object = VME_OBJECT(copy_entry); in vm_map_enter_mem_object_helper()
[all …]
H A Dvm_user.c2938 VME_OBJECT(entry) == VM_OBJECT_NULL || in mach_make_memory_entry_internal()
2939 VM_OBJECT_OWNER(VME_OBJECT(entry)) == TASK_NULL) { in mach_make_memory_entry_internal()
3135 VME_OBJECT(copy_entry) == VME_OBJECT(parent_copy_entry) && in mach_make_memory_entry_internal()
3156 object = VME_OBJECT(copy_entry); in mach_make_memory_entry_internal()
3189 VM_OBJECT_OWNER(VME_OBJECT(copy_entry)) == TASK_NULL) { in mach_make_memory_entry_internal()
3834 object = VME_OBJECT(entry); in mach_memory_entry_ownership()
4439 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_get_phys_page()
4443 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_get_phys_page()
4448 if (VME_OBJECT(entry)->vo_shadow_offset == 0) { in vm_map_get_phys_page()
4460 ((VME_OBJECT(entry)->vo_shadow_offset in vm_map_get_phys_page()
[all …]
H A Dvm_debug.c177 object = VME_OBJECT(entry); in vm32_region_info()
388 object = VME_OBJECT(entry); in vm32_region_info_64()
H A Dvm_map.h384 #define VME_OBJECT(entry) ({ assert(!(entry)->is_sub_map); _VME_OBJECT(entry); }) macro
466 object = VME_OBJECT(entry); in VME_OBJECT_SHADOW()
469 if (object != VME_OBJECT(entry)) { in VME_OBJECT_SHADOW()
H A Dvm_kern.c239 vm_object_t object = VME_OBJECT(entry); in __kmem_entry_orig_size()
610 (VME_OBJECT(entry) == compressor_object); in __kmem_entry_validate_object()
1505 vm_object_t object = VME_OBJECT(entry); in __kmem_realloc_invalid_object_size_panic()
1521 vm_object_t object = VME_OBJECT(entry); in __kmem_realloc_invalid_pager_panic()
1570 object = VME_OBJECT(entry); in kmem_realloc_shrink_guard()
1820 object = VME_OBJECT(oldentry); in kmem_realloc_guard()
1862 object = VME_OBJECT(oldentry); in kmem_realloc_guard()
2154 vm_object_t object = VME_OBJECT(entry); in __kmem_free_invalid_object_size_panic()
H A Dvm_pageout.c6719 VME_OBJECT(entry) != VM_OBJECT_NULL) { in vm_map_create_upl()
6720 if (VME_OBJECT(entry)->private) { in vm_map_create_upl()
6724 if (VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()
6748 if (VME_OBJECT(entry) == VM_OBJECT_NULL || in vm_map_create_upl()
6749 !VME_OBJECT(entry)->phys_contiguous) { in vm_map_create_upl()
6758 if (VME_OBJECT(entry) == VM_OBJECT_NULL) { in vm_map_create_upl()
6851 local_object = VME_OBJECT(entry); in vm_map_create_upl()
6889 assert(VME_OBJECT(entry) == local_object); in vm_map_create_upl()
7016 (VME_OBJECT(entry)->shadow || in vm_map_create_upl()
7017 VME_OBJECT(entry)->vo_copy)) { in vm_map_create_upl()
[all …]
H A Dbsd_vm.c1248 top_object = VME_OBJECT(entry); in fill_vnodeinfoforaddr()
1353 top_object = VME_OBJECT(entry); in find_vnode_object()
H A Dvm_shared_region.c1079 vm_object_reference(VME_OBJECT(found)); in find_mapping_to_slide()
1282 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()
1289 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_auth_remap()
2711 sr_pager = shared_region_pager_setup(VME_OBJECT(tmp_entry), VME_OFFSET(tmp_entry), si, 0); in vm_shared_region_slide_mapping()
2769 vm_object_deallocate(VME_OBJECT(tmp_entry)); in vm_shared_region_slide_mapping()
H A Dvm_fault.c6393 (VME_OBJECT(entry) == object)) { in vm_fault_internal()
6588 VME_OBJECT(entry) != VM_OBJECT_NULL && in vm_fault_wire()
6589 VME_OBJECT(entry)->phys_contiguous) { in vm_fault_wire()
6660 object = (entry->is_sub_map) ? VM_OBJECT_NULL : VME_OBJECT(entry); in vm_fault_unwire()
6931 object = VME_OBJECT(entry); in vm_fault_wire_fast()
7603 object = VME_OBJECT(entry); in kdp_lightweight_fault()
H A Dvm_resident.c10061 if (is_kernel_object(VME_OBJECT(entry))) { in vm_page_diagnose()
10063 vm_object_lock(VME_OBJECT(entry)); in vm_page_diagnose()
10065 page = vm_page_lookup(VME_OBJECT(entry), offset); in vm_page_diagnose()
10070 vm_object_unlock(VME_OBJECT(entry)); in vm_page_diagnose()
/xnu-10063.141.1/osfmk/i386/commpage/
H A Dcommpage.c170 VME_OBJECT(entry)->copy_strategy = MEMORY_OBJECT_COPY_NONE; in commpage_allocate()