Home
last modified time | relevance | path

Searched refs:vme_start (Results 1 – 18 of 18) sorted by relevance

/xnu-8020.121.3/osfmk/vm/
H A Dvm_map_store_rb.c50 if (vme_c->vme_start < vme_p->vme_start) { in rb_node_compare()
53 if (vme_c->vme_start >= vme_p->vme_end) { in rb_node_compare()
86 if (address >= cur->vme_start) { in vm_map_store_lookup_entry_rb()
111 …panic("VMSEL: INSERT FAILED: 0x%lx, 0x%lx, 0x%lx, 0x%lx", (uintptr_t)entry->vme_start, (uintptr_t)… in vm_map_store_entry_link_rb()
112 … (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_start, (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_end); in vm_map_store_entry_link_rb()
162 assert(middle_hole_entry->vme_end != last_hole_entry->vme_start); in vm_map_combine_hole()
172 assert(hole_entry->vme_start < hole_entry->vme_end); in vm_map_combine_hole()
173 assert(last_hole_entry->vme_start < last_hole_entry->vme_end); in vm_map_combine_hole()
234 while (map_entry->vme_start > hole_entry->vme_start) { in check_map_sanity()
244 if (map_entry->vme_start >= map->max_offset) { in check_map_sanity()
[all …]
H A Dvm_map_store_ll.c37 while (vm_map_trunc_page(next->vme_start, in first_free_is_valid_ll()
41 (vm_map_trunc_page(next->vme_start, in first_free_is_valid_ll()
43 vm_map_trunc_page(entry->vme_start, in first_free_is_valid_ll()
76 while (vm_map_trunc_page(UFF_next_entry->vme_start, \
80 (vm_map_trunc_page(UFF_next_entry->vme_start, \
82 vm_map_trunc_page(UFF_first_free->vme_start, \
98 assert(VM_MAP_PAGE_ALIGNED((entry->vme_start), \
178 if (address >= cur->vme_start) { in vm_map_store_lookup_entry_ll()
209 if (address >= cur->vme_start) { in vm_map_store_lookup_entry_ll()
H A Dvm_map_store.c125 assert(entry->vme_start < entry->vme_end); in _vm_map_store_entry_link()
138 entry->vme_start_original = entry->vme_start; in _vm_map_store_entry_link()
205 if (VMEU_entry->vme_start <= VMEU_map->first_free->vme_start) { in vm_map_store_entry_unlink()
253 entry, (uint64_t)entry->vme_start, (uint64_t)entry->vme_end, in __vm_map_store_find_space_holelist_corruption()
265 if (_vm_map_store_lookup_entry(map, entry->vme_start, entry_p)) { in vm_map_store_convert_hole_to_entry()
298 while (end <= entry->vme_start) { in vm_map_store_find_space_backwards()
314 end = entry->vme_start; in vm_map_store_find_space_backwards()
340 if (entry->vme_start <= start) { in vm_map_store_find_space_backwards()
376 end = entry->vme_start; in vm_map_store_find_space_backwards()
419 if (start < entry->vme_start) { in vm_map_store_find_space_forward()
[all …]
H A Dvm_map.c929 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected()
930 if (tmp_entry.vme_start < start) { in vm_map_apple_protected()
931 if (tmp_entry.vme_start != start_aligned) { in vm_map_apple_protected()
934 crypto_start += (start - tmp_entry.vme_start); in vm_map_apple_protected()
994 proc_selfpid(), tmp_entry.vme_start); in vm_map_apple_protected()
1000 map_addr = tmp_entry.vme_start; in vm_map_apple_protected()
1004 tmp_entry.vme_start), in vm_map_apple_protected()
1017 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected()
1020 (uint64_t) tmp_entry.vme_start, in vm_map_apple_protected()
1032 tmp_entry.vme_start)), in vm_map_apple_protected()
[all …]
H A Dvm_map_store.h89 (UHEE_entry->vme_start < SHARED_REGION_BASE || \
90 UHEE_entry->vme_start >= (SHARED_REGION_BASE + SHARED_REGION_SIZE)) && \
H A Dbsd_vm.c87 return vm_map_first_entry(map)->vme_start; in mach_get_vm_start()
1062 start = entry->vme_start; in fill_procregioninfo()
1191 pinfo->pri_address = (uint64_t)entry->vme_start; in fill_procregioninfo_onlymappedvnodes()
1192 pinfo->pri_size = (uint64_t)(entry->vme_end - entry->vme_start); in fill_procregioninfo_onlymappedvnodes()
1246 *start = entry->vme_start; in find_region_details()
1247 *len = entry->vme_end - entry->vme_start; in find_region_details()
H A Dvm_fault.c6107 if (ldelta > (laddr - entry->vme_start)) { in vm_fault_internal()
6108 ldelta = laddr - entry->vme_start; in vm_fault_internal()
6114 laddr = ((laddr - entry->vme_start) in vm_fault_internal()
6139 (entry->vme_end - entry->vme_start == object->vo_size) && in vm_fault_internal()
6140 VM_MAP_PAGE_ALIGNED(entry->vme_start, (object->vo_size - 1))) { in vm_fault_internal()
6152 + (laddr - entry->vme_start)) in vm_fault_internal()
6164 VME_OFFSET(entry) + (laddr - entry->vme_start) - ldelta), in vm_fault_internal()
6179 VME_OFFSET(entry) + (laddr - entry->vme_start) - ldelta), in vm_fault_internal()
6332 pmap_addr + (end_addr - entry->vme_start), FALSE); in vm_fault_wire()
6340 for (va = entry->vme_start; in vm_fault_wire()
[all …]
H A Dvm_kern.c242 map_addr = entry->vme_start; in kmem_alloc_contig()
496 map_addr = entry->vme_start; in kernel_memory_allocate()
931 newmapaddr = newentry->vme_start; in kmem_realloc()
1005 assert(entry->vme_start == addr && in kmem_realloc_down()
1011 vm_map_clip_end(map, entry, entry->vme_start + newsize); in kmem_realloc_down()
1556 VME_OFFSET_SET(entry, entry->vme_start); in kmem_scramble_ranges()
H A Dvm_debug.c178 region.vir_start = (natural_t) entry->vme_start; in vm32_region_info()
389 region.vir_start = (natural_t) entry->vme_start; in vm32_region_info_64()
H A Dvm_shared_region.c1141 if (tmp_entry->vme_end - tmp_entry->vme_start != si->si_end - si->si_start) { in vm_shared_region_auth_remap()
2391 assert(tmp_entry->vme_end - tmp_entry->vme_start == size); in vm_shared_region_slide_mapping()
2416 map_addr = tmp_entry->vme_start; in vm_shared_region_slide_mapping()
2419 (tmp_entry->vme_end - tmp_entry->vme_start), in vm_shared_region_slide_mapping()
2431 assertf(map_addr == tmp_entry->vme_start, in vm_shared_region_slide_mapping()
2434 (uint64_t) tmp_entry->vme_start, in vm_shared_region_slide_mapping()
H A Dvm_pageout.c6469 local_entry_start = entry->vme_start; in vm_map_create_upl()
6532 vm_object_round_page((entry->vme_end - entry->vme_start)))); in vm_map_create_upl()
6674 entry->vme_end - entry->vme_start, in vm_map_create_upl()
6680 entry->vme_start, in vm_map_create_upl()
6726 (entry->vme_end - entry->vme_start) / PAGE_SIZE; in vm_map_create_upl()
6760 local_start = entry->vme_start; in vm_map_create_upl()
6783 local_start = entry->vme_start; in vm_map_create_upl()
6806 local_start = entry->vme_start; in vm_map_create_upl()
6838 local_start = entry->vme_start; in vm_map_create_upl()
6855 (uint64_t) entry->vme_start, in vm_map_create_upl()
[all …]
H A Dvm_user.c3116 map_size = copy_entry->vme_end - copy_entry->vme_start; in mach_make_memory_entry_internal()
4261 (map_offset - entry->vme_start)); in vm_map_get_phys_page()
4280 (map_offset - entry->vme_start)); in vm_map_get_phys_page()
4286 offset = (VME_OFFSET(entry) + (map_offset - entry->vme_start)); in vm_map_get_phys_page()
H A Dvm_map.h273 #define vme_start links.start macro
H A Dvm_resident.c9791 if (entry->vme_start != addr) { in vm_kern_allocation_info()
/xnu-8020.121.3/osfmk/kern/
H A Dgzalloc.c571 "vme: %p, start: %llu end: %llu", gzvme, gzvme->vme_start, gzvme->vme_end); in gzalloc_element_size()
585 uint32_t *p = (uint32_t*) gzvme->vme_start; in gzalloc_element_size()
H A Dbsd_kern.c851 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < start)) { in get_vmsubmap_entries()
855 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) { in get_vmsubmap_entries()
862 entry->vme_start)); in get_vmsubmap_entries()
893 entry->vme_start)); in get_vmmap_entries()
H A Dkalloc.c1735 if (vm_entry->vme_start != addr) { in vm_map_lookup_kalloc_entry_locked()
1737 ptr, vm_entry, (void *)vm_entry->vme_start, in vm_map_lookup_kalloc_entry_locked()
2226 round_page(size) != entry->vme_end - entry->vme_start) { in kfree_large()
2228 (uint64_t)(entry->vme_end - entry->vme_start), in kfree_large()
2232 size = entry->vme_end - entry->vme_start; in kfree_large()
2234 vm_map_remove_and_unlock(map, entry->vme_start, entry->vme_end, in kfree_large()
2651 osize = entry->vme_end - entry->vme_start; in kern_os_realloc_external()
H A Dzalloc.c4148 if (first->vme_end + size > last->vme_start) { in zone_submap_alloc_sequestered_va()
4154 last->vme_start -= size; in zone_submap_alloc_sequestered_va()
4155 addr = last->vme_start; in zone_submap_alloc_sequestered_va()