| /xnu-8019.80.24/osfmk/vm/ |
| H A D | memory_object.c | 322 if (object == VM_OBJECT_NULL) { in memory_object_lock_request() 385 if (object == VM_OBJECT_NULL) { in memory_object_release_name() 409 if (object == VM_OBJECT_NULL) { in memory_object_destroy() 711 vm_object_t copy_object = VM_OBJECT_NULL; in vm_object_update() 753 while ((copy_object = object->copy) != VM_OBJECT_NULL) { in vm_object_update() 782 if ((copy_object != VM_OBJECT_NULL && update_cow) || (flags & MEMORY_OBJECT_DATA_SYNC)) { in vm_object_update() 792 if (copy_object != VM_OBJECT_NULL) { in vm_object_update() 910 if (copy_object != VM_OBJECT_NULL && copy_object != object) { in vm_object_update() 917 if (copy_object != VM_OBJECT_NULL && copy_object != object) { in vm_object_update() 1073 if (object == VM_OBJECT_NULL) { in vm_object_set_attributes_common() [all …]
|
| H A D | vm_object.c | 294 .copy = VM_OBJECT_NULL, 295 .shadow = VM_OBJECT_NULL, 652 vm_object_t shadow = VM_OBJECT_NULL; in vm_object_deallocate() 657 if (object == VM_OBJECT_NULL) { in vm_object_deallocate() 688 object->shadow != VM_OBJECT_NULL) { in vm_object_deallocate() 722 object->shadow != VM_OBJECT_NULL) { in vm_object_deallocate() 758 while (object != VM_OBJECT_NULL) { in vm_object_deallocate() 804 object->shadow != VM_OBJECT_NULL) { in vm_object_deallocate() 851 shadow = object->pageout?VM_OBJECT_NULL:object->shadow; in vm_object_deallocate() 856 if (shadow != VM_OBJECT_NULL) { in vm_object_deallocate() [all …]
|
| H A D | vm_fault.c | 1287 if (next_object == VM_OBJECT_NULL) { in vm_fault_page() 1387 && ((object != first_object) || (object->copy != VM_OBJECT_NULL)) in vm_fault_page() 1996 if (next_object == VM_OBJECT_NULL) { in vm_fault_page() 2236 while ((copy_object = first_object->copy) != VM_OBJECT_NULL) { in vm_fault_page() 2833 shadow != VM_OBJECT_NULL; in vm_fault_cs_handle_violation() 3910 if (top_object != VM_OBJECT_NULL) { in vm_fault_complete() 3938 if (top_object != VM_OBJECT_NULL) { in vm_fault_complete() 3949 top_object = VM_OBJECT_NULL; in vm_fault_complete() 3978 assert(*written_on_object == VM_OBJECT_NULL); in vm_fault_complete() 4059 vm_object_t top_object = VM_OBJECT_NULL; in vm_fault_internal() [all …]
|
| H A D | vm_map.c | 761 if (object == VM_OBJECT_NULL) { in vm_map_set_cache_attr() 850 VME_OBJECT(map_entry) == VM_OBJECT_NULL) { in vm_map_apple_protected() 866 if (protected_object == VM_OBJECT_NULL) { in vm_map_apple_protected() 1916 VME_OBJECT_SET(new_entry, VM_OBJECT_NULL); in vm_map_find_space() 2298 if (object != VM_OBJECT_NULL && in vm_map_enter() 2318 if (object == VM_OBJECT_NULL) { in vm_map_enter() 2418 (object != VM_OBJECT_NULL && in vm_map_enter() 2785 if ((obj2 == VM_OBJECT_NULL || in vm_map_enter() 2787 (object == VM_OBJECT_NULL || in vm_map_enter() 2844 if (object == VM_OBJECT_NULL) { in vm_map_enter() [all …]
|
| H A D | bsd_vm.c | 210 if (object == VM_OBJECT_NULL) { in memory_object_control_uiomove() 217 if (mark_dirty && object->copy != VM_OBJECT_NULL) { in memory_object_control_uiomove() 1283 if (top_object == VM_OBJECT_NULL) { in fill_vnodeinfoforaddr() 1284 object = VM_OBJECT_NULL; in fill_vnodeinfoforaddr() 1289 object->shadow != VM_OBJECT_NULL; in fill_vnodeinfoforaddr() 1297 if (object == VM_OBJECT_NULL) { in fill_vnodeinfoforaddr() 1391 for (object = top_object; object->shadow != VM_OBJECT_NULL; object = object->shadow) { in find_vnode_object() 1416 return VM_OBJECT_NULL; in find_vnode_object()
|
| H A D | vm_apple_protect.c | 370 src_top_object = VM_OBJECT_NULL; in apple_protect_pager_data_request() 371 src_page_object = VM_OBJECT_NULL; in apple_protect_pager_data_request() 411 assert(dst_object != VM_OBJECT_NULL); in apple_protect_pager_data_request() 419 assert(src_top_object != VM_OBJECT_NULL); in apple_protect_pager_data_request() 729 if (src_top_object != VM_OBJECT_NULL) { in apple_protect_pager_data_request() 799 if (pager->backing_object != VM_OBJECT_NULL) { in apple_protect_pager_terminate_internal() 801 pager->backing_object = VM_OBJECT_NULL; in apple_protect_pager_terminate_internal() 1150 pager->backing_object = VM_OBJECT_NULL; in apple_protect_pager_create()
|
| H A D | vm_shared_region_pager.c | 518 src_top_object = VM_OBJECT_NULL; in shared_region_pager_data_request() 519 src_page_object = VM_OBJECT_NULL; in shared_region_pager_data_request() 558 assert(dst_object != VM_OBJECT_NULL); in shared_region_pager_data_request() 566 assert(src_top_object != VM_OBJECT_NULL); in shared_region_pager_data_request() 833 if (src_top_object != VM_OBJECT_NULL) { in shared_region_pager_data_request() 904 if (pager->srp_backing_object != VM_OBJECT_NULL) { in shared_region_pager_terminate_internal() 906 pager->srp_backing_object = VM_OBJECT_NULL; in shared_region_pager_terminate_internal()
|
| H A D | vm_debug.c | 192 if (object == VM_OBJECT_NULL) { in vm32_region_info() 254 if (nobject == VM_OBJECT_NULL) { in vm32_region_info() 415 if (object == VM_OBJECT_NULL) { in vm32_region_info_64() 477 if (nobject == VM_OBJECT_NULL) { in vm32_region_info_64()
|
| H A D | vm_fourk_pager.c | 380 if (pager->slots[i].backing_object != VM_OBJECT_NULL && in fourk_pager_terminate_internal() 684 return VM_OBJECT_NULL; in fourk_pager_to_vm_object() 690 assert(object != VM_OBJECT_NULL); in fourk_pager_to_vm_object() 825 assert(dst_object != VM_OBJECT_NULL); in fourk_pager_data_request() 965 if (src_object == VM_OBJECT_NULL || in fourk_pager_data_request()
|
| H A D | vm_tests.c | 83 assert(backing_object != VM_OBJECT_NULL); in vm_test_collapse_compressor() 133 assert(top_object != VM_OBJECT_NULL); in vm_test_collapse_compressor() 199 if (top_object->shadow != VM_OBJECT_NULL) { in vm_test_collapse_compressor() 411 assert(anon_object != VM_OBJECT_NULL); in vm_test_device_pager_transpose() 415 assert(device_object != VM_OBJECT_NULL); in vm_test_device_pager_transpose()
|
| H A D | vm_user.c | 231 VM_OBJECT_NULL, in mach_vm_allocate_kernel() 320 VM_OBJECT_NULL, in vm_allocate_kernel() 2696 if (parent_is_object && object != VM_OBJECT_NULL) { in mach_make_memory_entry_internal() 2762 assert(object != VM_OBJECT_NULL); in mach_make_memory_entry_internal() 3278 assert(object != VM_OBJECT_NULL); in mach_make_memory_entry_internal() 3507 if (object != VM_OBJECT_NULL) { in mach_memory_object_memory_entry_64() 3511 if (object == VM_OBJECT_NULL) { in mach_memory_object_memory_entry_64() 3607 if (object == VM_OBJECT_NULL) { in memory_entry_purgeable_control_internal() 3669 if (object == VM_OBJECT_NULL) { in memory_entry_access_tracking_internal() 3788 if (object == VM_OBJECT_NULL) { in mach_memory_entry_ownership() [all …]
|
| H A D | vm_purgeable.c | 692 best_object = VM_OBJECT_NULL; in vm_purgeable_object_find_and_lock() 751 if (best_object != VM_OBJECT_NULL) { in vm_purgeable_object_find_and_lock() 777 if (object == VM_OBJECT_NULL) { in vm_purgeable_object_find_and_lock() 778 return VM_OBJECT_NULL; in vm_purgeable_object_find_and_lock() 844 if (object == VM_OBJECT_NULL) { in vm_purgeable_object_purge_all() 1298 vm_object_t object = VM_OBJECT_NULL; in vm_purgeable_queue_purge_task_owned()
|
| H A D | vm_pageout.c | 334 vm_object_t vm_pageout_scan_wants_object = VM_OBJECT_NULL; 765 assert(m_object != VM_OBJECT_NULL); in vm_pageout_throttle_up() 2223 vm_pageout_scan_wants_object = VM_OBJECT_NULL; in vps_flow_control() 2309 assert(vm_pageout_scan_wants_object == VM_OBJECT_NULL); in vps_flow_control() 2340 vm_object_t m_object = VM_OBJECT_NULL; in vps_choose_victim_page() 2352 m_object = VM_OBJECT_NULL; in vps_choose_victim_page() 2659 vm_pageout_scan_wants_object = VM_OBJECT_NULL; in vps_switch_object() 2678 vm_object_t cur_object = VM_OBJECT_NULL; in vps_deal_with_throttled_queues() 2736 cur_object = VM_OBJECT_NULL; in vps_deal_with_throttled_queues() 2742 *object = VM_OBJECT_NULL; in vps_deal_with_throttled_queues() [all …]
|
| H A D | vm_shared_region.c | 1552 if (object == VM_OBJECT_NULL) { in vm_shared_region_map_file_setup() 1571 object = VM_OBJECT_NULL; /* no anonymous memory here */ in vm_shared_region_map_file_setup() 1628 object = VM_OBJECT_NULL; in vm_shared_region_map_file_setup() 2281 vm_object_t object = VM_OBJECT_NULL; in vm_shared_region_slide_mapping() 2333 if (object == VM_OBJECT_NULL || object->internal) { in vm_shared_region_slide_mapping() 2334 object = VM_OBJECT_NULL; in vm_shared_region_slide_mapping() 2470 si->si_slide_object = VM_OBJECT_NULL; in vm_shared_region_slide_mapping()
|
| H A D | vm_kern.c | 1145 VM_OBJECT_NULL, (vm_object_offset_t) 0, FALSE, in kmem_alloc_pageable() 1432 VM_OBJECT_NULL, in kmem_init() 1466 VM_OBJECT_NULL, in kmem_init()
|
| H A D | vm_resident.c | 3426 if (object == VM_OBJECT_NULL) { in vm_page_grab_secluded() 3433 assert(VM_PAGE_OBJECT(mem) == VM_OBJECT_NULL); in vm_page_grab_secluded() 3491 object = VM_OBJECT_NULL; in vm_page_grab_secluded() 3500 assert(VM_PAGE_OBJECT(mem) == VM_OBJECT_NULL); in vm_page_grab_secluded() 4626 assert(m_object != VM_OBJECT_NULL); in vm_page_unwire() 5447 if (src_m_object != VM_OBJECT_NULL && in vm_page_copy() 6020 vm_object_t locked_object = VM_OBJECT_NULL; in vm_page_find_contiguous() 6118 locked_object = VM_OBJECT_NULL; in vm_page_find_contiguous() 6124 if (locked_object == VM_OBJECT_NULL || in vm_page_find_contiguous() 6131 locked_object = VM_OBJECT_NULL; in vm_page_find_contiguous() [all …]
|
| H A D | device_vm.c | 167 assert(object != VM_OBJECT_NULL); in device_pager_setup()
|
| H A D | vm_swapfile_pager.c | 352 assert(dst_object != VM_OBJECT_NULL); in swapfile_pager_data_request()
|
| H A D | vm_map.h | 356 if (object != VM_OBJECT_NULL && !object->internal) { in VME_OBJECT_SET()
|
| /xnu-8019.80.24/osfmk/mach/ |
| H A D | vm_types.h | 98 #define VM_OBJECT_NULL ((vm_object_t) NULL) macro
|