| /xnu-8020.101.4/osfmk/vm/ |
| H A D | memory_object.c | 322 if (object == VM_OBJECT_NULL) { in memory_object_lock_request() 385 if (object == VM_OBJECT_NULL) { in memory_object_release_name() 409 if (object == VM_OBJECT_NULL) { in memory_object_destroy() 711 vm_object_t copy_object = VM_OBJECT_NULL; in vm_object_update() 753 while ((copy_object = object->copy) != VM_OBJECT_NULL) { in vm_object_update() 782 if ((copy_object != VM_OBJECT_NULL && update_cow) || (flags & MEMORY_OBJECT_DATA_SYNC)) { in vm_object_update() 792 if (copy_object != VM_OBJECT_NULL) { in vm_object_update() 910 if (copy_object != VM_OBJECT_NULL && copy_object != object) { in vm_object_update() 917 if (copy_object != VM_OBJECT_NULL && copy_object != object) { in vm_object_update() 1073 if (object == VM_OBJECT_NULL) { in vm_object_set_attributes_common() [all …]
|
| H A D | vm_object.c | 290 .copy = VM_OBJECT_NULL, 291 .shadow = VM_OBJECT_NULL, 642 vm_object_t shadow = VM_OBJECT_NULL; in vm_object_deallocate() 647 if (object == VM_OBJECT_NULL) { in vm_object_deallocate() 678 object->shadow != VM_OBJECT_NULL) { in vm_object_deallocate() 712 object->shadow != VM_OBJECT_NULL) { in vm_object_deallocate() 748 while (object != VM_OBJECT_NULL) { in vm_object_deallocate() 794 object->shadow != VM_OBJECT_NULL) { in vm_object_deallocate() 841 shadow = object->pageout?VM_OBJECT_NULL:object->shadow; in vm_object_deallocate() 846 if (shadow != VM_OBJECT_NULL) { in vm_object_deallocate() [all …]
|
| H A D | vm_fault.c | 1292 if (next_object == VM_OBJECT_NULL) { in vm_fault_page() 1392 && ((object != first_object) || (object->copy != VM_OBJECT_NULL)) in vm_fault_page() 2001 if (next_object == VM_OBJECT_NULL) { in vm_fault_page() 2241 while ((copy_object = first_object->copy) != VM_OBJECT_NULL) { in vm_fault_page() 2838 shadow != VM_OBJECT_NULL; in vm_fault_cs_handle_violation() 3770 assertf(VM_PAGE_OBJECT(m) != VM_OBJECT_NULL, "m=%p", m); in vm_fault_enter() 3916 if (top_object != VM_OBJECT_NULL) { in vm_fault_complete() 3944 if (top_object != VM_OBJECT_NULL) { in vm_fault_complete() 3955 top_object = VM_OBJECT_NULL; in vm_fault_complete() 3984 assert(*written_on_object == VM_OBJECT_NULL); in vm_fault_complete() [all …]
|
| H A D | vm_map.c | 770 if (object == VM_OBJECT_NULL) { in vm_map_set_cache_attr() 859 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected() 875 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected() 1918 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL); in vm_map_find_space() 2306 if (object != VM_OBJECT_NULL && in vm_map_enter() 2326 if (object == VM_OBJECT_NULL) { in vm_map_enter() 2426 (object != VM_OBJECT_NULL && in vm_map_enter() 2792 if ((obj2 == VM_OBJECT_NULL || in vm_map_enter() 2794 (object == VM_OBJECT_NULL || in vm_map_enter() 2851 if (object == VM_OBJECT_NULL) { in vm_map_enter() [all …]
|
| H A D | vm_apple_protect.c | 370 src_top_object = VM_OBJECT_NULL; in apple_protect_pager_data_request() 371 src_page_object = VM_OBJECT_NULL; in apple_protect_pager_data_request() 411 assert(dst_object != VM_OBJECT_NULL); in apple_protect_pager_data_request() 419 assert(src_top_object != VM_OBJECT_NULL); in apple_protect_pager_data_request() 729 if (src_top_object != VM_OBJECT_NULL) { in apple_protect_pager_data_request() 799 if (pager->backing_object != VM_OBJECT_NULL) { in apple_protect_pager_terminate_internal() 801 pager->backing_object = VM_OBJECT_NULL; in apple_protect_pager_terminate_internal() 1150 pager->backing_object = VM_OBJECT_NULL; in apple_protect_pager_create()
|
| H A D | bsd_vm.c | 210 if (object == VM_OBJECT_NULL) { in memory_object_control_uiomove() 217 if (mark_dirty && object->copy != VM_OBJECT_NULL) { in memory_object_control_uiomove() 1283 if (top_object == VM_OBJECT_NULL) { in fill_vnodeinfoforaddr() 1284 object = VM_OBJECT_NULL; in fill_vnodeinfoforaddr() 1289 object->shadow != VM_OBJECT_NULL; in fill_vnodeinfoforaddr() 1297 if (object == VM_OBJECT_NULL) { in fill_vnodeinfoforaddr() 1391 for (object = top_object; object->shadow != VM_OBJECT_NULL; object = object->shadow) { in find_vnode_object() 1416 return VM_OBJECT_NULL; in find_vnode_object()
|
| H A D | vm_shared_region_pager.c | 518 src_top_object = VM_OBJECT_NULL; in shared_region_pager_data_request() 519 src_page_object = VM_OBJECT_NULL; in shared_region_pager_data_request() 558 assert(dst_object != VM_OBJECT_NULL); in shared_region_pager_data_request() 566 assert(src_top_object != VM_OBJECT_NULL); in shared_region_pager_data_request() 833 if (src_top_object != VM_OBJECT_NULL) { in shared_region_pager_data_request() 904 if (pager->srp_backing_object != VM_OBJECT_NULL) { in shared_region_pager_terminate_internal() 906 pager->srp_backing_object = VM_OBJECT_NULL; in shared_region_pager_terminate_internal()
|
| H A D | vm_debug.c | 192 if (object == VM_OBJECT_NULL) { in vm32_region_info() 254 if (nobject == VM_OBJECT_NULL) { in vm32_region_info() 403 if (object == VM_OBJECT_NULL) { in vm32_region_info_64() 465 if (nobject == VM_OBJECT_NULL) { in vm32_region_info_64()
|
| H A D | vm_fourk_pager.c | 380 if (pager->slots[i].backing_object != VM_OBJECT_NULL && in fourk_pager_terminate_internal() 684 return VM_OBJECT_NULL; in fourk_pager_to_vm_object() 690 assert(object != VM_OBJECT_NULL); in fourk_pager_to_vm_object() 825 assert(dst_object != VM_OBJECT_NULL); in fourk_pager_data_request() 964 if (src_object == VM_OBJECT_NULL || in fourk_pager_data_request()
|
| H A D | vm_user.c | 230 VM_OBJECT_NULL, in mach_vm_allocate_kernel() 314 VM_OBJECT_NULL, in vm_allocate_external() 2692 if (parent_is_object && object != VM_OBJECT_NULL) { in mach_make_memory_entry_internal() 2752 assert(object != VM_OBJECT_NULL); in mach_make_memory_entry_internal() 3242 assert(object != VM_OBJECT_NULL); in mach_make_memory_entry_internal() 3408 if (object != VM_OBJECT_NULL) { in mach_memory_object_memory_entry_64() 3412 if (object == VM_OBJECT_NULL) { in mach_memory_object_memory_entry_64() 3495 if (object == VM_OBJECT_NULL) { in memory_entry_purgeable_control_internal() 3549 if (object == VM_OBJECT_NULL) { in memory_entry_access_tracking_internal() 3671 if (object == VM_OBJECT_NULL) { in mach_memory_entry_ownership() [all …]
|
| H A D | vm_purgeable.c | 692 best_object = VM_OBJECT_NULL; in vm_purgeable_object_find_and_lock() 751 if (best_object != VM_OBJECT_NULL) { in vm_purgeable_object_find_and_lock() 777 if (object == VM_OBJECT_NULL) { in vm_purgeable_object_find_and_lock() 778 return VM_OBJECT_NULL; in vm_purgeable_object_find_and_lock() 844 if (object == VM_OBJECT_NULL) { in vm_purgeable_object_purge_all() 1298 vm_object_t object = VM_OBJECT_NULL; in vm_purgeable_queue_purge_task_owned()
|
| H A D | vm_tests.c | 86 assert(backing_object != VM_OBJECT_NULL); in vm_test_collapse_compressor() 136 assert(top_object != VM_OBJECT_NULL); in vm_test_collapse_compressor() 202 if (top_object->shadow != VM_OBJECT_NULL) { in vm_test_collapse_compressor() 414 assert(anon_object != VM_OBJECT_NULL); in vm_test_device_pager_transpose() 418 assert(device_object != VM_OBJECT_NULL); in vm_test_device_pager_transpose()
|
| H A D | vm_pageout.c | 334 vm_object_t vm_pageout_scan_wants_object = VM_OBJECT_NULL; 765 assert(m_object != VM_OBJECT_NULL); in vm_pageout_throttle_up() 2223 vm_pageout_scan_wants_object = VM_OBJECT_NULL; in vps_flow_control() 2309 assert(vm_pageout_scan_wants_object == VM_OBJECT_NULL); in vps_flow_control() 2340 vm_object_t m_object = VM_OBJECT_NULL; in vps_choose_victim_page() 2352 m_object = VM_OBJECT_NULL; in vps_choose_victim_page() 2659 vm_pageout_scan_wants_object = VM_OBJECT_NULL; in vps_switch_object() 2678 vm_object_t cur_object = VM_OBJECT_NULL; in vps_deal_with_throttled_queues() 2736 cur_object = VM_OBJECT_NULL; in vps_deal_with_throttled_queues() 2742 *object = VM_OBJECT_NULL; in vps_deal_with_throttled_queues() [all …]
|
| H A D | vm_shared_region.c | 1542 if (object == VM_OBJECT_NULL) { in vm_shared_region_map_file_setup() 1561 object = VM_OBJECT_NULL; /* no anonymous memory here */ in vm_shared_region_map_file_setup() 1618 object = VM_OBJECT_NULL; in vm_shared_region_map_file_setup() 2271 vm_object_t object = VM_OBJECT_NULL; in vm_shared_region_slide_mapping() 2323 if (object == VM_OBJECT_NULL || object->internal) { in vm_shared_region_slide_mapping() 2324 object = VM_OBJECT_NULL; in vm_shared_region_slide_mapping() 2460 si->si_slide_object = VM_OBJECT_NULL; in vm_shared_region_slide_mapping()
|
| H A D | vm_resident.c | 3420 if (object == VM_OBJECT_NULL) { in vm_page_grab_secluded() 3427 assert(VM_PAGE_OBJECT(mem) == VM_OBJECT_NULL); in vm_page_grab_secluded() 3485 object = VM_OBJECT_NULL; in vm_page_grab_secluded() 3494 assert(VM_PAGE_OBJECT(mem) == VM_OBJECT_NULL); in vm_page_grab_secluded() 4620 assert(m_object != VM_OBJECT_NULL); in vm_page_unwire() 5441 if (src_m_object != VM_OBJECT_NULL && in vm_page_copy() 6014 vm_object_t locked_object = VM_OBJECT_NULL; in vm_page_find_contiguous() 6112 locked_object = VM_OBJECT_NULL; in vm_page_find_contiguous() 6118 if (locked_object == VM_OBJECT_NULL || in vm_page_find_contiguous() 6125 locked_object = VM_OBJECT_NULL; in vm_page_find_contiguous() [all …]
|
| H A D | device_vm.c | 166 assert(object != VM_OBJECT_NULL); in device_pager_setup()
|
| H A D | vm_kern.c | 1297 VM_OBJECT_NULL, in kmem_init() 1332 VM_OBJECT_NULL, in kmem_init()
|
| H A D | vm_map.h | 343 entry->vme_object.vmo_object = VM_OBJECT_NULL; in VME_OBJECT_SET() 348 if (object != VM_OBJECT_NULL && !object->internal) { in VME_OBJECT_SET()
|
| H A D | vm_swapfile_pager.c | 352 assert(dst_object != VM_OBJECT_NULL); in swapfile_pager_data_request()
|
| /xnu-8020.101.4/osfmk/mach/ |
| H A D | vm_types.h | 106 #define VM_OBJECT_NULL ((vm_object_t) NULL) macro
|