| /xnu-12377.1.9/osfmk/vm/ |
| H A D | vm_map_store_rb.c | 53 if (vme_c->vme_start >= vme_p->vme_end) { in rb_node_compare() 70 if (address < cur->vme_end) { in vm_map_store_lookup_entry_rb() 97 (uintptr_t)entry->vme_end, in vm_map_store_entry_link_rb() 99 (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_end); in vm_map_store_entry_link_rb() 141 hole_entry->vme_end = hole_entry->vme_next->vme_end; in vm_map_combine_hole() 147 assert(middle_hole_entry->vme_end != last_hole_entry->vme_start); in vm_map_combine_hole() 157 assert(hole_entry->vme_start < hole_entry->vme_end); in vm_map_combine_hole() 158 assert(last_hole_entry->vme_start < last_hole_entry->vme_end); in vm_map_combine_hole() 233 if (map_entry->vme_end != map_entry->vme_next->vme_start) { in check_map_sanity() 238 if (hole_entry->vme_start != map_entry->vme_end) { in check_map_sanity() [all …]
|
| H A D | vm_map_store.c | 108 if (__improbable(entry->vme_end <= entry->vme_start)) { in _vm_map_store_entry_link() 109 …p start 0x%llx end 0x%llx\n", mapHdr, entry, (uint64_t)entry->vme_start, (uint64_t)entry->vme_end); in _vm_map_store_entry_link() 112 assert(entry->vme_start < entry->vme_end); in _vm_map_store_entry_link() 118 vm_address_t, entry->vme_end); in _vm_map_store_entry_link() 130 entry->vme_end_original = entry->vme_end; in _vm_map_store_entry_link() 177 map->highest_entry_end < entry->vme_end) { in vm_map_store_entry_link() 178 map->highest_entry_end = entry->vme_end; in vm_map_store_entry_link() 207 vm_address_t, entry->vme_end); in _vm_map_store_entry_unlink() 218 (uint64_t)entry->vme_start, (uint64_t)entry->vme_end, in _vm_map_store_entry_unlink() 297 entry, entry->vme_start, entry->vme_end, map, start); in __vm_map_store_find_space_holelist_corruption() [all …]
|
| H A D | vm_map.c | 649 map, new, new->vme_start, new->vme_end); in vm_map_entry_copy_csm_assoc() 680 uint64_t, new->vme_end, in vm_map_entry_copy_code_signing() 1133 map_addr = tmp_entry.vme_end) { in vm_map_apple_protected() 1181 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected() 1190 if (tmp_entry.vme_end > end) { in vm_map_apple_protected() 1191 if (tmp_entry.vme_end != end_aligned) { in vm_map_apple_protected() 1196 crypto_end -= (tmp_entry.vme_end - end); in vm_map_apple_protected() 1252 (tmp_entry.vme_end - in vm_map_apple_protected() 1278 (uint64_t) (map_addr + (tmp_entry.vme_end - in vm_map_apple_protected() 1299 crypto_backing_offset += (tmp_entry.vme_end - in vm_map_apple_protected() [all …]
|
| H A D | vm_map_store_ll.c | 40 vm_map_trunc_page(entry->vme_end, map_page_mask) || in first_free_is_valid_ll() 73 assert(VM_MAP_PAGE_ALIGNED(entry->vme_end, in vm_map_store_entry_link_ll() 120 vm_map_trunc_page(new_first_free->vme_end, map_page_mask) || in update_first_free_ll()
|
| H A D | bsd_vm.c | 1018 address == tmp_entry->vme_end) { in fill_procregioninfo() 1111 …vm_map_region_walk(map, start, entry, VME_OFFSET(entry), entry->vme_end - start, &extended, TRUE, … in fill_procregioninfo() 1131 pinfo->pri_size = (uint64_t)(entry->vme_end - start); in fill_procregioninfo() 1216 pinfo->pri_size = (uint64_t)(entry->vme_end - entry->vme_start); in fill_procregioninfo_onlymappedvnodes() 1310 *len_p = entry->vme_end - entry->vme_start; in task_find_region_details()
|
| H A D | vm_debug.c | 188 region.vir_end = (natural_t) entry->vme_end; in vm32_mach_vm_region_info() 409 region.vir_end = (natural_t) entry->vme_end; in vm32_mach_vm_region_info_64()
|
| H A D | vm_lock_perf.h | 314 vmlp_range_event(map, entry->vme_start, entry->vme_end - entry->vme_start); in vmlp_range_event_entry()
|
| H A D | vm_pageout.c | 6927 if (entry->vme_end - original_offset < adjusted_size) { 6928 adjusted_size = entry->vme_end - original_offset; 6979 vm_object_round_page((entry->vme_end - entry->vme_start)), map->serial_id), 7097 if ((entry->vme_end - offset) < *upl_size) { 7098 *upl_size = (upl_size_t) (entry->vme_end - offset); 7099 assert(*upl_size == entry->vme_end - offset); 7108 entry->vme_end - entry->vme_start, 7160 (entry->vme_end - entry->vme_start) / PAGE_SIZE; 7289 (uint64_t) entry->vme_end); 7293 uint64_t, (uint64_t)entry->vme_end); [all …]
|
| H A D | vm_kern.c | 248 return entry->vme_end - entry->vme_start - in __kmem_entry_orig_size() 494 } else if ((flags & KMF_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_panic() 530 if ((flags & KMEM_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_guard() 3652 prev_entry->vme_end <= slot->min_address)) && in kmem_free_space() 4380 used += (entry->vme_end - entry->vme_start); in kmem_get_gobj_stats() 4849 assert3u(addr + ptoa(10), <=, e->vme_end); in kmem_alloc_basic_test()
|
| H A D | vm_fault.c | 6691 if (hdelta > (entry->vme_end - laddr)) { in vm_fault_internal() 6692 hdelta = entry->vme_end - laddr; in vm_fault_internal() 6726 (entry->vme_end - entry->vme_start == object->vo_size) && in vm_fault_internal() 6905 vm_map_offset_t end_addr = entry->vme_end; in vm_fault_wire() 6960 tmp_entry.vme_end = va; in vm_fault_wire() 6962 pmap, pmap_addr, tmp_entry.vme_end); in vm_fault_wire() 7012 fault_info.hi_offset = (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_fault_unwire()
|
| H A D | vm_map_xnu.h | 172 #define vme_end links.end macro
|
| H A D | vm_memory_entry.c | 796 map_size = copy_entry->vme_end - copy_entry->vme_start; in mach_make_memory_entry_share() 1681 entry->vme_end - entry->vme_start != object->vo_size) { in mach_memory_entry_ownership()
|
| H A D | vm_shared_region.c | 1404 if (tmp_entry->vme_end - tmp_entry->vme_start != si->si_end - si->si_start) { in vm_shared_region_auth_remap() 2832 assert(tmp_entry->vme_end - tmp_entry->vme_start == size); in vm_shared_region_slide_mapping() 2863 tmp_entry->vme_end - tmp_entry->vme_start, in vm_shared_region_slide_mapping()
|
| H A D | vm_user.c | 1235 map->highest_entry_end = map->first_free->vme_end; in vm_toggle_entry_reuse()
|
| H A D | vm_resident.c | 11125 for (offset = entry->vme_start; offset < entry->vme_end; offset += page_size) { in vm_page_diagnose() 11199 *size = (entry->vme_end - addr); in vm_kern_allocation_info() 11254 textSize = entry->vme_end - entry->vme_start; in vm_task_evict_shared_cache()
|
| H A D | vm_tests.c | 2252 …assert3u((entry->vme_end - entry->vme_start), ==, expected_size); /* 4k entries combined into a si… in vm_map_4k_16k_test()
|
| /xnu-12377.1.9/osfmk/kdp/ |
| H A D | kdp_common.c | 128 …for (vcur = entry->vme_start; ret == KERN_SUCCESS && vcur < entry->vme_end; vcur += task_page_size… in kdp_traverse_mappings()
|
| /xnu-12377.1.9/osfmk/kern/ |
| H A D | bsd_kern.c | 929 entry->vme_end - in get_vmsubmap_entries() 964 entry->vme_end - in get_vmmap_entries()
|
| H A D | zalloc.c | 4083 vmlp_range_event(map, first->vme_end, size); in zone_submap_alloc_sequestered_va() 4086 if (first->vme_end + size > last->vme_start) { in zone_submap_alloc_sequestered_va() 4097 addr = first->vme_end; in zone_submap_alloc_sequestered_va() 4098 first->vme_end += size; in zone_submap_alloc_sequestered_va() 9670 reloc_base = first->vme_end; in zone_metadata_init() 9671 first->vme_end += early_sz; in zone_metadata_init()
|
| /xnu-12377.1.9/tools/lldbmacros/ |
| H A D | memory.py | 3418 vme_end = links.end 3420 vme_end_val = unsigned(vme_end) 3424 if unsigned(vme_end) <= start_vaddr: 3540 …d}{:<4s} {:3s}/{:3s}/{:<8s} {:<18s} {:<#18x} {:s}".format(vme,vme_start,vme_end,(vme_end_val-vme_…
|