| /xnu-12377.61.12/osfmk/vm/ |
| H A D | vm_map_store_rb.c | 50 if (vme_c->vme_start < vme_p->vme_start) { in rb_node_compare() 53 if (vme_c->vme_start >= vme_p->vme_end) { in rb_node_compare() 69 if (address >= cur->vme_start) { in vm_map_store_lookup_entry_rb() 96 (uintptr_t)entry->vme_start, in vm_map_store_entry_link_rb() 98 (uintptr_t)(VME_FOR_STORE(tmp_store))->vme_start, in vm_map_store_entry_link_rb() 147 assert(middle_hole_entry->vme_end != last_hole_entry->vme_start); in vm_map_combine_hole() 157 assert(hole_entry->vme_start < hole_entry->vme_end); in vm_map_combine_hole() 158 assert(last_hole_entry->vme_start < last_hole_entry->vme_end); in vm_map_combine_hole() 219 while (map_entry->vme_start > hole_entry->vme_start) { in check_map_sanity() 229 if (map_entry->vme_start >= map->max_offset) { in check_map_sanity() [all …]
|
| H A D | vm_map_store.c | 108 if (__improbable(entry->vme_end <= entry->vme_start)) { in _vm_map_store_entry_link() 109 …panic("maphdr %p entry %p start 0x%llx end 0x%llx\n", mapHdr, entry, (uint64_t)entry->vme_start, (… in _vm_map_store_entry_link() 112 assert(entry->vme_start < entry->vme_end); in _vm_map_store_entry_link() 117 vm_address_t, entry->vme_start, in _vm_map_store_entry_link() 129 entry->vme_start_original = entry->vme_start; in _vm_map_store_entry_link() 175 (entry->vme_start < SHARED_REGION_BASE || in vm_map_store_entry_link() 176 entry->vme_start >= (SHARED_REGION_BASE + SHARED_REGION_SIZE)) && in vm_map_store_entry_link() 206 vm_address_t, entry->vme_start, in _vm_map_store_entry_unlink() 218 (uint64_t)entry->vme_start, (uint64_t)entry->vme_end, in _vm_map_store_entry_unlink() 246 if (VMEU_entry->vme_start <= VMEU_map->first_free->vme_start) { in vm_map_store_entry_unlink() [all …]
|
| H A D | vm_map_store_ll.c | 39 while (vm_map_trunc_page(next->vme_start, map_page_mask) == in first_free_is_valid_ll() 41 (vm_map_trunc_page(next->vme_start, map_page_mask) == in first_free_is_valid_ll() 42 vm_map_trunc_page(entry->vme_start, map_page_mask) && in first_free_is_valid_ll() 71 assert(VM_MAP_PAGE_ALIGNED(entry->vme_start, in vm_map_store_entry_link_ll() 119 while (vm_map_trunc_page(next->vme_start, map_page_mask) == in update_first_free_ll() 121 (vm_map_trunc_page(next->vme_start, map_page_mask) == in update_first_free_ll() 122 vm_map_trunc_page(new_first_free->vme_start, map_page_mask) && in update_first_free_ll()
|
| H A D | vm_map.c | 680 map, new, new->vme_start, new->vme_end); in vm_map_entry_copy_csm_assoc() 710 uint64_t, new->vme_start, in vm_map_entry_copy_code_signing() 1212 tmp_entry.vme_start += map_addr; in vm_map_apple_protected() 1225 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected() 1226 if (tmp_entry.vme_start < start) { in vm_map_apple_protected() 1227 if (tmp_entry.vme_start != start_aligned) { in vm_map_apple_protected() 1232 crypto_start += (start - tmp_entry.vme_start); in vm_map_apple_protected() 1292 map_addr = tmp_entry.vme_start; in vm_map_apple_protected() 1296 tmp_entry.vme_start), in vm_map_apple_protected() 1307 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected() [all …]
|
| H A D | vm_lock_perf.h | 314 vmlp_range_event(map, entry->vme_start, entry->vme_end - entry->vme_start); in vmlp_range_event_entry()
|
| H A D | bsd_vm.c | 1083 start = entry->vme_start; in fill_procregioninfo() 1216 pinfo->pri_address = (uint64_t)entry->vme_start; in fill_procregioninfo_onlymappedvnodes() 1217 pinfo->pri_size = (uint64_t)(entry->vme_end - entry->vme_start); in fill_procregioninfo_onlymappedvnodes() 1310 *start_p = entry->vme_start; in task_find_region_details() 1311 *len_p = entry->vme_end - entry->vme_start; in task_find_region_details()
|
| H A D | vm_fault.c | 6793 if (ldelta > (laddr - entry->vme_start)) { in vm_fault_internal() 6794 ldelta = laddr - entry->vme_start; in vm_fault_internal() 6803 laddr = ((laddr - entry->vme_start) in vm_fault_internal() 6831 (entry->vme_end - entry->vme_start == object->vo_size) && in vm_fault_internal() 6832 VM_MAP_PAGE_ALIGNED(entry->vme_start, (object->vo_size - 1))) { in vm_fault_internal() 6844 + (laddr - entry->vme_start)) in vm_fault_internal() 6855 VME_OFFSET(entry) + (laddr - entry->vme_start) - ldelta); in vm_fault_internal() 7039 pmap_addr + (end_addr - entry->vme_start), FALSE); in vm_fault_wire() 7047 for (va = entry->vme_start; in vm_fault_wire() 7051 pmap_addr + (va - entry->vme_start), in vm_fault_wire() [all …]
|
| H A D | vm_debug.c | 187 region.vir_start = (natural_t) entry->vme_start; in vm32_mach_vm_region_info() 408 region.vir_start = (natural_t) entry->vme_start; in vm32_mach_vm_region_info_64()
|
| H A D | vm_kern.c | 256 return entry->vme_end - entry->vme_start - in __kmem_entry_orig_size() 500 } else if (addr != entry->vme_start) { in __kmem_entry_validate_panic() 534 if (addr != entry->vme_start) { in __kmem_entry_validate_guard() 967 map_addr = entry->vme_start; 1780 vm_map_clip_end(map, entry, entry->vme_start + newsize); in kmem_realloc_shrink_guard() 2093 newaddr = newentry->vme_start; in kmem_realloc_guard() 3263 ((*entry)->vme_next->vme_start < (addr + size))) { in kmem_get_addr_from_meta() 3743 (next_entry->vme_start >= slot->max_address))) { in kmem_free_space() 4331 VME_OFFSET_SET(entry, entry->vme_start); in kmem_scramble_ranges() 4468 entry->vme_start < range.max_address) { in kmem_get_gobj_stats() [all …]
|
| H A D | vm_shared_region.c | 1404 if (tmp_entry->vme_end - tmp_entry->vme_start != si->si_end - si->si_start) { in vm_shared_region_auth_remap() 2832 assert(tmp_entry->vme_end - tmp_entry->vme_start == size); in vm_shared_region_slide_mapping() 2859 map_addr = tmp_entry->vme_start; in vm_shared_region_slide_mapping() 2868 tmp_entry->vme_end - tmp_entry->vme_start, in vm_shared_region_slide_mapping() 2878 assertf(map_addr == tmp_entry->vme_start, in vm_shared_region_slide_mapping() 2881 (uint64_t) tmp_entry->vme_start, in vm_shared_region_slide_mapping()
|
| H A D | vm_pageout.c | 7135 local_entry_start = entry->vme_start; 7194 vm_object_round_page((entry->vme_end - entry->vme_start)), map->serial_id), 7323 entry->vme_end - entry->vme_start, 7329 entry->vme_start, 7375 (entry->vme_end - entry->vme_start) / PAGE_SIZE; 7409 local_start = entry->vme_start; 7430 local_start = entry->vme_start; 7453 local_start = entry->vme_start; 7485 local_start = entry->vme_start; 7489 vm_size_t size = entry->vme_end - entry->vme_start; [all …]
|
| H A D | vm_tests.c | 1986 (VME_OFFSET(entry) + (canonical_addr - entry->vme_start))); in vm_page_relocate_test() 2004 (VME_OFFSET(entry) + (canonical_addr - entry->vme_start))); in vm_page_relocate_test() 2571 …assert3u((entry->vme_end - entry->vme_start), ==, expected_size); /* 4k entries combined into a si… in vm_map_4k_16k_test() 3588 (vm_object_offset_t)(args.ptr - entry->vme_start + VME_OFFSET(entry)), in vm_upl_object_test() 3603 m = vm_page_lookup(object, (VME_OFFSET(entry) + ((vm_map_address_t)args.ptr - entry->vme_start))); in vm_upl_object_test()
|
| H A D | vm_map_xnu.h | 171 #define vme_start links.start macro
|
| H A D | vm_memory_entry.c | 812 map_size = copy_entry->vme_end - copy_entry->vme_start; in mach_make_memory_entry_share() 1706 entry->vme_end - entry->vme_start != object->vo_size) { in mach_memory_entry_ownership()
|
| H A D | vm_resident.c | 12013 for (offset = entry->vme_start; offset < entry->vme_end; offset += page_size) { in vm_page_diagnose() 12080 if (entry->vme_start != addr) { in vm_kern_allocation_info() 12142 textSize = entry->vme_end - entry->vme_start; in vm_task_evict_shared_cache()
|
| /xnu-12377.61.12/osfmk/kern/ |
| H A D | bsd_kern.c | 934 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < start)) { in get_vmsubmap_entries() 938 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) { in get_vmsubmap_entries() 945 entry->vme_start)); in get_vmsubmap_entries() 980 entry->vme_start)); in get_vmmap_entries()
|
| H A D | zalloc.c | 4159 vmlp_range_event(map, last->vme_start - size, size); in zone_submap_alloc_sequestered_va() 4164 if (first->vme_end + size > last->vme_start) { in zone_submap_alloc_sequestered_va() 4171 last->vme_start -= size; in zone_submap_alloc_sequestered_va() 4172 addr = last->vme_start; in zone_submap_alloc_sequestered_va()
|
| /xnu-12377.61.12/osfmk/kdp/ |
| H A D | kdp_common.c | 131 …for (vcur = entry->vme_start; ret == KERN_SUCCESS && vcur < entry->vme_end; vcur += task_page_size… in kdp_traverse_mappings()
|
| /xnu-12377.61.12/tools/lldbmacros/ |
| H A D | memory.py | 3421 vme_start = links.start 3423 vme_start_val = unsigned(vme_start) 3438 …print("{:18s} {:#018x}:{:#018x} {:>10d}".format("------------------",last_end,vme_start,(vme_start… 3544 …10d} {:>3d}{:<4s} {:3s}/{:3s}/{:<8s} {:<18s} {:<#18x} {:s}".format(vme,vme_start,vme_end,(vme_end…
|