Home
last modified time | relevance | path

Searched refs:vme_end (Results 1 – 17 of 17) sorted by relevance

/xnu-8792.81.2/osfmk/vm/ !
H A Dvm_map_store_rb.c53 if (vme_c->vme_start >= vme_p->vme_end) { in rb_node_compare()
87 if (address < cur->vme_end) { in vm_map_store_lookup_entry_rb()
111 …INSERT FAILED: 0x%lx, 0x%lx, 0x%lx, 0x%lx", (uintptr_t)entry->vme_start, (uintptr_t)entry->vme_end, in vm_map_store_entry_link_rb()
112 … (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_start, (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_end); in vm_map_store_entry_link_rb()
154 hole_entry->vme_end = hole_entry->vme_next->vme_end; in vm_map_combine_hole()
160 assert(middle_hole_entry->vme_end != last_hole_entry->vme_start); in vm_map_combine_hole()
170 assert(hole_entry->vme_start < hole_entry->vme_end); in vm_map_combine_hole()
171 assert(last_hole_entry->vme_start < last_hole_entry->vme_end); in vm_map_combine_hole()
246 if (map_entry->vme_end != map_entry->vme_next->vme_start) { in check_map_sanity()
251 if (hole_entry->vme_start != map_entry->vme_end) { in check_map_sanity()
[all …]
H A Dvm_map_store.c125 assert(entry->vme_start < entry->vme_end); in _vm_map_store_entry_link()
139 entry->vme_end_original = entry->vme_end; in _vm_map_store_entry_link()
195 …ubmap %d\n", mapHdr, entry, (uint64_t)entry->vme_start, (uint64_t)entry->vme_end, entry->protectio… in _vm_map_store_entry_unlink()
267 entry, entry->vme_start, entry->vme_end, map, start); in __vm_map_store_find_space_holelist_corruption()
319 if (entry->vme_end < end) { in vm_map_store_find_space_backwards()
320 end = entry->vme_end; in vm_map_store_find_space_backwards()
369 end = entry->vme_end; in vm_map_store_find_space_backwards()
381 if (entry->vme_end <= start) { in vm_map_store_find_space_backwards()
424 while (entry->vme_end <= start) { in vm_map_store_find_space_forward()
444 first_free_start = entry->vme_end; in vm_map_store_find_space_forward()
[all …]
H A Dvm_map.c885 map_addr = tmp_entry.vme_end) { in vm_map_apple_protected()
932 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected()
939 if (tmp_entry.vme_end > end) { in vm_map_apple_protected()
940 if (tmp_entry.vme_end != end_aligned) { in vm_map_apple_protected()
943 crypto_end -= (tmp_entry.vme_end - end); in vm_map_apple_protected()
1006 (tmp_entry.vme_end - in vm_map_apple_protected()
1034 (uint64_t) (map_addr + (tmp_entry.vme_end - in vm_map_apple_protected()
1055 crypto_backing_offset += (tmp_entry.vme_end - in vm_map_apple_protected()
2168 new_entry->vme_end = hint_address + size; in vm_map_find_space()
2183 assert(page_aligned(new_entry->vme_end)); in vm_map_find_space()
[all …]
H A Dvm_map_store_ll.c39 vm_map_trunc_page(entry->vme_end, in first_free_is_valid_ll()
78 vm_map_trunc_page(UFF_first_free->vme_end, \
100 assert(VM_MAP_PAGE_ALIGNED((entry->vme_end), \
191 if ((cur != last) && (cur->vme_end > address)) { in vm_map_store_lookup_entry_ll()
208 if (cur->vme_end > address) { in vm_map_store_lookup_entry_ll()
H A Dvm_map_store.h91 UHEE_map->highest_entry_end < UHEE_entry->vme_end) { \
92 UHEE_map->highest_entry_end = UHEE_entry->vme_end; \
106 VMHE_start = tmp_entry->vme_end + PAGE_SIZE_64; \
H A Dbsd_vm.c93 return vm_map_last_entry(map)->vme_end; in mach_get_vm_end()
974 address == tmp_entry->vme_end) { in fill_procregioninfo()
1063 …vm_map_region_walk(map, start, entry, VME_OFFSET(entry), entry->vme_end - start, &extended, TRUE, … in fill_procregioninfo()
1087 pinfo->pri_size = (uint64_t)(entry->vme_end - start); in fill_procregioninfo()
1165 pinfo->pri_size = (uint64_t)(entry->vme_end - entry->vme_start); in fill_procregioninfo_onlymappedvnodes()
1220 *len = entry->vme_end - entry->vme_start; in find_region_details()
H A Dvm_debug.c179 region.vir_end = (natural_t) entry->vme_end; in vm32_region_info()
390 region.vir_end = (natural_t) entry->vme_end; in vm32_region_info_64()
H A Dvm_kern.c403 } else if ((flags & KMF_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_panic()
441 if ((flags & KMEM_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_guard()
535 size = (vm_size_t)(entry->vme_end - entry->vme_start); in kmem_size_guard()
2732 assert3u(addr + ptoa(10), <=, e->vme_end); in kmem_alloc_basic_test()
H A Dvm_fault.c6068 if (hdelta > (entry->vme_end - laddr)) { in vm_fault_internal()
6069 hdelta = entry->vme_end - laddr; in vm_fault_internal()
6098 (entry->vme_end - entry->vme_start == object->vo_size) && in vm_fault_internal()
6281 vm_map_offset_t end_addr = entry->vme_end; in vm_fault_wire()
6330 tmp_entry.vme_end = va; in vm_fault_wire()
6354 vm_map_offset_t end_addr = entry->vme_end; in vm_fault_unwire()
6380 fault_info.hi_offset = (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_fault_unwire()
H A Dvm_pageout.c6687 if (entry->vme_end - original_offset < adjusted_size) { in vm_map_create_upl()
6688 adjusted_size = entry->vme_end - original_offset; in vm_map_create_upl()
6744 vm_object_round_page((entry->vme_end - entry->vme_start))), in vm_map_create_upl()
6878 if ((entry->vme_end - offset) < *upl_size) { in vm_map_create_upl()
6879 *upl_size = (upl_size_t) (entry->vme_end - offset); in vm_map_create_upl()
6880 assert(*upl_size == entry->vme_end - offset); in vm_map_create_upl()
6889 entry->vme_end - entry->vme_start, in vm_map_create_upl()
6941 (entry->vme_end - entry->vme_start) / PAGE_SIZE; in vm_map_create_upl()
7071 (uint64_t) entry->vme_end); in vm_map_create_upl()
7075 uint64_t, (uint64_t)entry->vme_end); in vm_map_create_upl()
[all …]
H A Dvm_shared_region.c1217 if (tmp_entry->vme_end - tmp_entry->vme_start != si->si_end - si->si_start) { in vm_shared_region_auth_remap()
2698 assert(tmp_entry->vme_end - tmp_entry->vme_start == size); in vm_shared_region_slide_mapping()
2728 (tmp_entry->vme_end - tmp_entry->vme_start), in vm_shared_region_slide_mapping()
H A Dvm_user.c1856 map->highest_entry_end = map->first_free->vme_end; in vm_toggle_entry_reuse()
3216 map_size = copy_entry->vme_end - copy_entry->vme_start; in mach_make_memory_entry_internal()
3871 entry->vme_end - entry->vme_start != object->vo_size) { in mach_memory_entry_ownership()
H A Dvm_map.h267 #define vme_end links.end macro
H A Dvm_resident.c9907 *size = (entry->vme_end - addr); in vm_kern_allocation_info()
/xnu-8792.81.2/osfmk/kdp/ !
H A Dkdp_common.c128 …for (vcur = entry->vme_start; ret == KERN_SUCCESS && vcur < entry->vme_end; vcur += task_page_size… in kdp_traverse_mappings()
/xnu-8792.81.2/osfmk/kern/ !
H A Dbsd_kern.c869 entry->vme_end - in get_vmsubmap_entries()
900 entry->vme_end - in get_vmmap_entries()
H A Dzalloc.c4192 if (first->vme_end + size > last->vme_start) { in zone_submap_alloc_sequestered_va()
4202 addr = first->vme_end; in zone_submap_alloc_sequestered_va()
4203 first->vme_end += size; in zone_submap_alloc_sequestered_va()
9665 reloc_base = first->vme_end; in zone_metadata_init()
9666 first->vme_end += early_sz; in zone_metadata_init()