Home
last modified time | relevance | path

Searched refs:vme_start (Results 1 – 17 of 17) sorted by relevance

/xnu-11417.140.69/osfmk/vm/
H A Dvm_map_store_rb.c50 if (vme_c->vme_start < vme_p->vme_start) { in rb_node_compare()
53 if (vme_c->vme_start >= vme_p->vme_end) { in rb_node_compare()
69 if (address >= cur->vme_start) { in vm_map_store_lookup_entry_rb()
96 (uintptr_t)entry->vme_start, in vm_map_store_entry_link_rb()
98 (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_start, in vm_map_store_entry_link_rb()
147 assert(middle_hole_entry->vme_end != last_hole_entry->vme_start); in vm_map_combine_hole()
157 assert(hole_entry->vme_start < hole_entry->vme_end); in vm_map_combine_hole()
158 assert(last_hole_entry->vme_start < last_hole_entry->vme_end); in vm_map_combine_hole()
219 while (map_entry->vme_start > hole_entry->vme_start) { in check_map_sanity()
229 if (map_entry->vme_start >= map->max_offset) { in check_map_sanity()
[all …]
H A Dvm_map_store.c108 if (__improbable(entry->vme_end <= entry->vme_start)) { in _vm_map_store_entry_link()
109 …panic("maphdr %p entry %p start 0x%llx end 0x%llx\n", mapHdr, entry, (uint64_t)entry->vme_start, (… in _vm_map_store_entry_link()
112 assert(entry->vme_start < entry->vme_end); in _vm_map_store_entry_link()
117 vm_address_t, entry->vme_start, in _vm_map_store_entry_link()
129 entry->vme_start_original = entry->vme_start; in _vm_map_store_entry_link()
175 (entry->vme_start < SHARED_REGION_BASE || in vm_map_store_entry_link()
176 entry->vme_start >= (SHARED_REGION_BASE + SHARED_REGION_SIZE)) && in vm_map_store_entry_link()
206 vm_address_t, entry->vme_start, in _vm_map_store_entry_unlink()
218 (uint64_t)entry->vme_start, (uint64_t)entry->vme_end, in _vm_map_store_entry_unlink()
246 if (VMEU_entry->vme_start <= VMEU_map->first_free->vme_start) { in vm_map_store_entry_unlink()
[all …]
H A Dvm_map_store_ll.c39 while (vm_map_trunc_page(next->vme_start, map_page_mask) == in first_free_is_valid_ll()
41 (vm_map_trunc_page(next->vme_start, map_page_mask) == in first_free_is_valid_ll()
42 vm_map_trunc_page(entry->vme_start, map_page_mask) && in first_free_is_valid_ll()
71 assert(VM_MAP_PAGE_ALIGNED(entry->vme_start, in vm_map_store_entry_link_ll()
120 while (vm_map_trunc_page(next->vme_start, map_page_mask) == in update_first_free_ll()
122 (vm_map_trunc_page(next->vme_start, map_page_mask) == in update_first_free_ll()
123 vm_map_trunc_page(new_first_free->vme_start, map_page_mask) && in update_first_free_ll()
H A Dvm_map.c625 map, new, new->vme_start, new->vme_end); in vm_map_entry_copy_csm_assoc()
655 uint64_t, new->vme_start, in vm_map_entry_copy_code_signing()
1137 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected()
1138 if (tmp_entry.vme_start < start) { in vm_map_apple_protected()
1139 if (tmp_entry.vme_start != start_aligned) { in vm_map_apple_protected()
1144 crypto_start += (start - tmp_entry.vme_start); in vm_map_apple_protected()
1205 map_addr = tmp_entry.vme_start; in vm_map_apple_protected()
1209 tmp_entry.vme_start), in vm_map_apple_protected()
1220 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected()
1223 (uint64_t) tmp_entry.vme_start, in vm_map_apple_protected()
[all …]
H A Dbsd_vm.c1065 start = entry->vme_start; in fill_procregioninfo()
1190 pinfo->pri_address = (uint64_t)entry->vme_start; in fill_procregioninfo_onlymappedvnodes()
1191 pinfo->pri_size = (uint64_t)(entry->vme_end - entry->vme_start); in fill_procregioninfo_onlymappedvnodes()
1276 *start_p = entry->vme_start; in task_find_region_details()
1277 *len_p = entry->vme_end - entry->vme_start; in task_find_region_details()
H A Dvm_fault.c6565 if (ldelta > (laddr - entry->vme_start)) { in vm_fault_internal()
6566 ldelta = laddr - entry->vme_start; in vm_fault_internal()
6575 laddr = ((laddr - entry->vme_start) in vm_fault_internal()
6603 (entry->vme_end - entry->vme_start == object->vo_size) && in vm_fault_internal()
6604 VM_MAP_PAGE_ALIGNED(entry->vme_start, (object->vo_size - 1))) { in vm_fault_internal()
6616 + (laddr - entry->vme_start)) in vm_fault_internal()
6627 VME_OFFSET(entry) + (laddr - entry->vme_start) - ldelta); in vm_fault_internal()
6800 pmap_addr + (end_addr - entry->vme_start), FALSE); in vm_fault_wire()
6808 for (va = entry->vme_start; in vm_fault_wire()
6812 pmap_addr + (va - entry->vme_start), in vm_fault_wire()
[all …]
H A Dvm_debug.c180 region.vir_start = (natural_t) entry->vme_start; in vm32_mach_vm_region_info()
393 region.vir_start = (natural_t) entry->vme_start; in vm32_mach_vm_region_info_64()
H A Dvm_kern.c244 return entry->vme_end - entry->vme_start - in __kmem_entry_orig_size()
488 } else if (addr != entry->vme_start) { in __kmem_entry_validate_panic()
522 if (addr != entry->vme_start) { in __kmem_entry_validate_guard()
923 map_addr = entry->vme_start;
1707 vm_map_clip_end(map, entry, entry->vme_start + newsize); in kmem_realloc_shrink_guard()
2001 newaddr = newentry->vme_start; in kmem_realloc_guard()
3115 ((*entry)->vme_next->vme_start < (addr + size))) { in kmem_get_addr_from_meta()
3589 (next_entry->vme_start >= slot->max_address))) { in kmem_free_space()
4146 VME_OFFSET_SET(entry, entry->vme_start); in kmem_scramble_ranges()
4269 entry->vme_start < range.max_address) { in kmem_get_gobj_stats()
[all …]
H A Dvm_shared_region.c1246 if (tmp_entry->vme_end - tmp_entry->vme_start != si->si_end - si->si_start) { in vm_shared_region_auth_remap()
2717 assert(tmp_entry->vme_end - tmp_entry->vme_start == size); in vm_shared_region_slide_mapping()
2739 map_addr = tmp_entry->vme_start; in vm_shared_region_slide_mapping()
2748 tmp_entry->vme_end - tmp_entry->vme_start, in vm_shared_region_slide_mapping()
2758 assertf(map_addr == tmp_entry->vme_start, in vm_shared_region_slide_mapping()
2761 (uint64_t) tmp_entry->vme_start, in vm_shared_region_slide_mapping()
H A Dvm_pageout.c6697 local_entry_start = entry->vme_start;
6768 vm_object_round_page((entry->vme_end - entry->vme_start)), map->serial_id),
6913 entry->vme_end - entry->vme_start,
6919 entry->vme_start,
6965 (entry->vme_end - entry->vme_start) / PAGE_SIZE;
6999 local_start = entry->vme_start;
7022 local_start = entry->vme_start;
7045 local_start = entry->vme_start;
7077 local_start = entry->vme_start;
7095 (uint64_t) entry->vme_start,
[all …]
H A Dvm_map_xnu.h171 #define vme_start links.start macro
H A Dvm_memory_entry.c825 map_size = copy_entry->vme_end - copy_entry->vme_start; in mach_make_memory_entry_share()
1697 entry->vme_end - entry->vme_start != object->vo_size) { in mach_memory_entry_ownership()
H A Dvm_resident.c11039 for (offset = entry->vme_start; offset < entry->vme_end; offset += page_size) { in vm_page_diagnose()
11102 if (entry->vme_start != addr) { in vm_kern_allocation_info()
/xnu-11417.140.69/osfmk/kern/
H A Dbsd_kern.c917 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < start)) { in get_vmsubmap_entries()
921 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) { in get_vmsubmap_entries()
928 entry->vme_start)); in get_vmsubmap_entries()
959 entry->vme_start)); in get_vmmap_entries()
H A Dzalloc.c4165 if (first->vme_end + size > last->vme_start) { in zone_submap_alloc_sequestered_va()
4171 last->vme_start -= size; in zone_submap_alloc_sequestered_va()
4172 addr = last->vme_start; in zone_submap_alloc_sequestered_va()
/xnu-11417.140.69/osfmk/kdp/
H A Dkdp_common.c128 …for (vcur = entry->vme_start; ret == KERN_SUCCESS && vcur < entry->vme_end; vcur += task_page_size… in kdp_traverse_mappings()
/xnu-11417.140.69/tools/lldbmacros/
H A Dmemory.py3462 vme_start = links.start
3464 vme_start_val = unsigned(vme_start)
3479 …print("{:18s} {:#018x}:{:#018x} {:>10d}".format("------------------",last_end,vme_start,(vme_start…
3587 …10d} {:>3d}{:<4s} {:3s}/{:3s}/{:<8s} {:<18s} {:<#18x} {:s}".format(vme,vme_start,vme_end,(vme_end…