| /xnu-8020.140.41/osfmk/vm/ |
| H A D | vm_map_store_rb.c | 53 if (vme_c->vme_start >= vme_p->vme_end) { in rb_node_compare() 87 if (address < cur->vme_end) { in vm_map_store_lookup_entry_rb() 111 …INSERT FAILED: 0x%lx, 0x%lx, 0x%lx, 0x%lx", (uintptr_t)entry->vme_start, (uintptr_t)entry->vme_end, in vm_map_store_entry_link_rb() 112 … (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_start, (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_end); in vm_map_store_entry_link_rb() 156 hole_entry->vme_end = hole_entry->vme_next->vme_end; in vm_map_combine_hole() 162 assert(middle_hole_entry->vme_end != last_hole_entry->vme_start); in vm_map_combine_hole() 172 assert(hole_entry->vme_start < hole_entry->vme_end); in vm_map_combine_hole() 173 assert(last_hole_entry->vme_start < last_hole_entry->vme_end); in vm_map_combine_hole() 248 if (map_entry->vme_end != map_entry->vme_next->vme_start) { in check_map_sanity() 253 if (hole_entry->vme_start != map_entry->vme_end) { in check_map_sanity() [all …]
|
| H A D | vm_map_store.c | 125 assert(entry->vme_start < entry->vme_end); in _vm_map_store_entry_link() 139 entry->vme_end_original = entry->vme_end; in _vm_map_store_entry_link() 253 entry, (uint64_t)entry->vme_start, (uint64_t)entry->vme_end, in __vm_map_store_find_space_holelist_corruption() 306 if (entry->vme_end < end) { in vm_map_store_find_space_backwards() 307 end = entry->vme_end; in vm_map_store_find_space_backwards() 356 end = entry->vme_end; in vm_map_store_find_space_backwards() 368 if (entry->vme_end <= start) { in vm_map_store_find_space_backwards() 411 while (entry->vme_end <= start) { in vm_map_store_find_space_forward() 431 first_free_start = entry->vme_end; in vm_map_store_find_space_forward() 437 start = entry->vme_end; in vm_map_store_find_space_forward() [all …]
|
| H A D | vm_map.c | 879 map_addr = tmp_entry.vme_end) { in vm_map_apple_protected() 926 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected() 933 if (tmp_entry.vme_end > end) { in vm_map_apple_protected() 934 if (tmp_entry.vme_end != end_aligned) { in vm_map_apple_protected() 937 crypto_end -= (tmp_entry.vme_end - end); in vm_map_apple_protected() 1000 (tmp_entry.vme_end - in vm_map_apple_protected() 1028 (uint64_t) (map_addr + (tmp_entry.vme_end - in vm_map_apple_protected() 1049 crypto_backing_offset += (tmp_entry.vme_end - in vm_map_apple_protected() 2188 new_entry->vme_end = hint_address + size; in vm_map_find_space() 2203 assert(page_aligned(new_entry->vme_end)); in vm_map_find_space() [all …]
|
| H A D | vm_map_store_ll.c | 39 vm_map_trunc_page(entry->vme_end, in first_free_is_valid_ll() 78 vm_map_trunc_page(UFF_first_free->vme_end, \ 100 assert(VM_MAP_PAGE_ALIGNED((entry->vme_end), \ 191 if ((cur != last) && (cur->vme_end > address)) { in vm_map_store_lookup_entry_ll() 208 if (cur->vme_end > address) { in vm_map_store_lookup_entry_ll()
|
| H A D | vm_map_store.h | 91 UHEE_map->highest_entry_end < UHEE_entry->vme_end) { \ 92 UHEE_map->highest_entry_end = UHEE_entry->vme_end; \ 106 VMHE_start = tmp_entry->vme_end + PAGE_SIZE_64; \
|
| H A D | bsd_vm.c | 93 return vm_map_last_entry(map)->vme_end; in mach_get_vm_end() 1001 address == tmp_entry->vme_end) { in fill_procregioninfo() 1090 …vm_map_region_walk(map, start, entry, VME_OFFSET(entry), entry->vme_end - start, &extended, TRUE, … in fill_procregioninfo() 1114 pinfo->pri_size = (uint64_t)(entry->vme_end - start); in fill_procregioninfo() 1192 pinfo->pri_size = (uint64_t)(entry->vme_end - entry->vme_start); in fill_procregioninfo_onlymappedvnodes() 1247 *len = entry->vme_end - entry->vme_start; in find_region_details()
|
| H A D | vm_debug.c | 179 region.vir_end = (natural_t) entry->vme_end; in vm32_region_info() 390 region.vir_end = (natural_t) entry->vme_end; in vm32_region_info_64()
|
| H A D | vm_kern.c | 309 } else if ((flags & KMF_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_panic() 347 if ((flags & KMEM_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_guard() 441 size = (vm_size_t)(entry->vme_end - entry->vme_start); in kmem_size_guard() 2579 assert3u(addr + ptoa(10), <=, e->vme_end); in kmem_alloc_basic_test()
|
| H A D | vm_fault.c | 6110 if (hdelta > (entry->vme_end - laddr)) { in vm_fault_internal() 6111 hdelta = entry->vme_end - laddr; in vm_fault_internal() 6140 (entry->vme_end - entry->vme_start == object->vo_size) && in vm_fault_internal() 6314 vm_map_offset_t end_addr = entry->vme_end; in vm_fault_wire() 6363 tmp_entry.vme_end = va; in vm_fault_wire() 6387 vm_map_offset_t end_addr = entry->vme_end; in vm_fault_unwire() 6413 fault_info.hi_offset = (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_fault_unwire()
|
| H A D | vm_pageout.c | 6476 if (entry->vme_end - original_offset < adjusted_size) { in vm_map_create_upl() 6477 adjusted_size = entry->vme_end - original_offset; in vm_map_create_upl() 6533 vm_object_round_page((entry->vme_end - entry->vme_start))), in vm_map_create_upl() 6667 if ((entry->vme_end - offset) < *upl_size) { in vm_map_create_upl() 6668 *upl_size = (upl_size_t) (entry->vme_end - offset); in vm_map_create_upl() 6669 assert(*upl_size == entry->vme_end - offset); in vm_map_create_upl() 6678 entry->vme_end - entry->vme_start, in vm_map_create_upl() 6730 (entry->vme_end - entry->vme_start) / PAGE_SIZE; in vm_map_create_upl() 6860 (uint64_t) entry->vme_end); in vm_map_create_upl() 6864 uint64_t, (uint64_t)entry->vme_end); in vm_map_create_upl()
|
| H A D | vm_shared_region.c | 1141 if (tmp_entry->vme_end - tmp_entry->vme_start != si->si_end - si->si_start) { in vm_shared_region_auth_remap() 2391 assert(tmp_entry->vme_end - tmp_entry->vme_start == size); in vm_shared_region_slide_mapping() 2419 (tmp_entry->vme_end - tmp_entry->vme_start), in vm_shared_region_slide_mapping()
|
| H A D | vm_map.h | 265 #define vme_end links.end macro
|
| H A D | vm_user.c | 1847 map->highest_entry_end = map->first_free->vme_end; in vm_toggle_entry_reuse() 3184 map_size = copy_entry->vme_end - copy_entry->vme_start; in mach_make_memory_entry_internal()
|
| H A D | vm_resident.c | 9804 *size = (entry->vme_end - addr); in vm_kern_allocation_info()
|
| /xnu-8020.140.41/osfmk/kern/ |
| H A D | gzalloc.c | 571 "vme: %p, start: %llu end: %llu", gzvme, gzvme->vme_start, gzvme->vme_end); in gzalloc_element_size() 579 gzh = (gzhdr_t *)(gzvme->vme_end - GZHEADER_SIZE); in gzalloc_element_size() 586 while (p < (uint32_t *) gzvme->vme_end) { in gzalloc_element_size() 593 if (p >= (uint32_t *) gzvme->vme_end) { in gzalloc_element_size()
|
| H A D | bsd_kern.c | 861 entry->vme_end - in get_vmsubmap_entries() 892 entry->vme_end - in get_vmmap_entries()
|
| H A D | zalloc.c | 4148 if (first->vme_end + size > last->vme_start) { in zone_submap_alloc_sequestered_va() 4158 addr = first->vme_end; in zone_submap_alloc_sequestered_va() 4159 first->vme_end += size; in zone_submap_alloc_sequestered_va() 9618 reloc_base = first->vme_end; in zone_metadata_init() 9619 first->vme_end += early_sz; in zone_metadata_init()
|