Home
last modified time | relevance | path

Searched refs:vme_end (Results 1 – 16 of 16) sorted by relevance

/xnu-8796.101.5/osfmk/vm/
H A Dvm_map_store_rb.c53 if (vme_c->vme_start >= vme_p->vme_end) { in rb_node_compare()
70 if (address < cur->vme_end) { in vm_map_store_lookup_entry_rb()
97 (uintptr_t)entry->vme_end, in vm_map_store_entry_link_rb()
99 (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_end); in vm_map_store_entry_link_rb()
141 hole_entry->vme_end = hole_entry->vme_next->vme_end; in vm_map_combine_hole()
147 assert(middle_hole_entry->vme_end != last_hole_entry->vme_start); in vm_map_combine_hole()
157 assert(hole_entry->vme_start < hole_entry->vme_end); in vm_map_combine_hole()
158 assert(last_hole_entry->vme_start < last_hole_entry->vme_end); in vm_map_combine_hole()
233 if (map_entry->vme_end != map_entry->vme_next->vme_start) { in check_map_sanity()
238 if (hole_entry->vme_start != map_entry->vme_end) { in check_map_sanity()
[all …]
H A Dvm_map_store.c108 assert(entry->vme_start < entry->vme_end); in _vm_map_store_entry_link()
114 vm_address_t, entry->vme_end); in _vm_map_store_entry_link()
126 entry->vme_end_original = entry->vme_end; in _vm_map_store_entry_link()
173 map->highest_entry_end < entry->vme_end) { in vm_map_store_entry_link()
174 map->highest_entry_end = entry->vme_end; in vm_map_store_entry_link()
203 vm_address_t, entry->vme_end); in _vm_map_store_entry_unlink()
214 (uint64_t)entry->vme_start, (uint64_t)entry->vme_end, in _vm_map_store_entry_unlink()
293 entry, entry->vme_start, entry->vme_end, map, start); in __vm_map_store_find_space_holelist_corruption()
345 if (entry->vme_end < end) { in vm_map_store_find_space_backwards()
346 end = entry->vme_end; in vm_map_store_find_space_backwards()
[all …]
H A Dvm_map.c465 map, new, new->vme_start, new->vme_end); in vm_map_entry_copy_csm_assoc()
894 map_addr = tmp_entry.vme_end) { in vm_map_apple_protected()
941 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected()
948 if (tmp_entry.vme_end > end) { in vm_map_apple_protected()
949 if (tmp_entry.vme_end != end_aligned) { in vm_map_apple_protected()
952 crypto_end -= (tmp_entry.vme_end - end); in vm_map_apple_protected()
1018 (tmp_entry.vme_end - in vm_map_apple_protected()
1044 (uint64_t) (map_addr + (tmp_entry.vme_end - in vm_map_apple_protected()
1065 crypto_backing_offset += (tmp_entry.vme_end - in vm_map_apple_protected()
2168 new_entry->vme_end = hint_address + size; in vm_map_find_space()
[all …]
H A Dvm_map_store_ll.c40 vm_map_trunc_page(entry->vme_end, map_page_mask) || in first_free_is_valid_ll()
73 assert(VM_MAP_PAGE_ALIGNED(entry->vme_end, in vm_map_store_entry_link_ll()
121 vm_map_trunc_page(new_first_free->vme_end, map_page_mask) || in update_first_free_ll()
H A Dbsd_vm.c951 address == tmp_entry->vme_end) { in fill_procregioninfo()
1040 …vm_map_region_walk(map, start, entry, VME_OFFSET(entry), entry->vme_end - start, &extended, TRUE, … in fill_procregioninfo()
1064 pinfo->pri_size = (uint64_t)(entry->vme_end - start); in fill_procregioninfo()
1142 pinfo->pri_size = (uint64_t)(entry->vme_end - entry->vme_start); in fill_procregioninfo_onlymappedvnodes()
1197 *len = entry->vme_end - entry->vme_start; in find_region_details()
H A Dvm_debug.c179 region.vir_end = (natural_t) entry->vme_end; in vm32_region_info()
390 region.vir_end = (natural_t) entry->vme_end; in vm32_region_info_64()
H A Dvm_kern.c236 return entry->vme_end - entry->vme_start - in __kmem_entry_orig_size()
503 } else if ((flags & KMF_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_panic()
539 if ((flags & KMEM_GUESS_SIZE) == 0 && addr + size != entry->vme_end) { in __kmem_entry_validate_guard()
3238 prev_entry->vme_end <= slot->min_address)) && in kmem_free_space()
3741 used += (entry->vme_end - entry->vme_start); in kmem_get_gobj_stats()
4182 assert3u(addr + ptoa(10), <=, e->vme_end); in kmem_alloc_basic_test()
H A Dvm_fault.c6137 if (hdelta > (entry->vme_end - laddr)) { in vm_fault_internal()
6138 hdelta = entry->vme_end - laddr; in vm_fault_internal()
6167 (entry->vme_end - entry->vme_start == object->vo_size) && in vm_fault_internal()
6350 vm_map_offset_t end_addr = entry->vme_end; in vm_fault_wire()
6399 tmp_entry.vme_end = va; in vm_fault_wire()
6401 pmap, pmap_addr, tmp_entry.vme_end); in vm_fault_wire()
6449 fault_info.hi_offset = (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_fault_unwire()
H A Dvm_pageout.c6684 if (entry->vme_end - original_offset < adjusted_size) { in vm_map_create_upl()
6685 adjusted_size = entry->vme_end - original_offset; in vm_map_create_upl()
6741 vm_object_round_page((entry->vme_end - entry->vme_start))), in vm_map_create_upl()
6875 if ((entry->vme_end - offset) < *upl_size) { in vm_map_create_upl()
6876 *upl_size = (upl_size_t) (entry->vme_end - offset); in vm_map_create_upl()
6877 assert(*upl_size == entry->vme_end - offset); in vm_map_create_upl()
6886 entry->vme_end - entry->vme_start, in vm_map_create_upl()
6938 (entry->vme_end - entry->vme_start) / PAGE_SIZE; in vm_map_create_upl()
7068 (uint64_t) entry->vme_end); in vm_map_create_upl()
7072 uint64_t, (uint64_t)entry->vme_end); in vm_map_create_upl()
[all …]
H A Dvm_shared_region.c1216 if (tmp_entry->vme_end - tmp_entry->vme_start != si->si_end - si->si_start) { in vm_shared_region_auth_remap()
2679 assert(tmp_entry->vme_end - tmp_entry->vme_start == size); in vm_shared_region_slide_mapping()
2709 (tmp_entry->vme_end - tmp_entry->vme_start), in vm_shared_region_slide_mapping()
H A Dvm_user.c1742 map->highest_entry_end = map->first_free->vme_end; in vm_toggle_entry_reuse()
3087 map_size = copy_entry->vme_end - copy_entry->vme_start; in mach_make_memory_entry_internal()
3742 entry->vme_end - entry->vme_start != object->vo_size) { in mach_memory_entry_ownership()
H A Dvm_map.h238 #define vme_end links.end macro
H A Dvm_resident.c9806 for (offset = entry->vme_start; offset < entry->vme_end; offset += page_size) { in vm_page_diagnose()
9873 *size = (entry->vme_end - addr); in vm_kern_allocation_info()
/xnu-8796.101.5/osfmk/kdp/
H A Dkdp_common.c128 …for (vcur = entry->vme_start; ret == KERN_SUCCESS && vcur < entry->vme_end; vcur += task_page_size… in kdp_traverse_mappings()
/xnu-8796.101.5/osfmk/kern/
H A Dbsd_kern.c869 entry->vme_end - in get_vmsubmap_entries()
900 entry->vme_end - in get_vmmap_entries()
H A Dzalloc.c4073 if (first->vme_end + size > last->vme_start) { in zone_submap_alloc_sequestered_va()
4083 addr = first->vme_end; in zone_submap_alloc_sequestered_va()
4084 first->vme_end += size; in zone_submap_alloc_sequestered_va()
9820 reloc_base = first->vme_end; in zone_metadata_init()
9821 first->vme_end += early_sz; in zone_metadata_init()