Lines Matching refs:vme_start

932 		crypto_end = tmp_entry.vme_end - tmp_entry.vme_start;  in vm_map_apple_protected()
933 if (tmp_entry.vme_start < start) { in vm_map_apple_protected()
934 if (tmp_entry.vme_start != start_aligned) { in vm_map_apple_protected()
937 crypto_start += (start - tmp_entry.vme_start); in vm_map_apple_protected()
997 proc_selfpid(), tmp_entry.vme_start); in vm_map_apple_protected()
1003 map_addr = tmp_entry.vme_start; in vm_map_apple_protected()
1007 tmp_entry.vme_start), in vm_map_apple_protected()
1020 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected()
1023 (uint64_t) tmp_entry.vme_start, in vm_map_apple_protected()
1035 tmp_entry.vme_start)), in vm_map_apple_protected()
1056 tmp_entry.vme_start); in vm_map_apple_protected()
2167 new_entry->vme_start = hint_address; in vm_map_find_space()
2182 assert(page_aligned(new_entry->vme_start)); in vm_map_find_space()
2184 assert(VM_MAP_PAGE_ALIGNED(new_entry->vme_start, VM_MAP_PAGE_MASK(map))); in vm_map_find_space()
2340 hole_end = next_entry->vme_start; in vm_map_random_address_for_size()
2765 if (entry->vme_start < start) { in vm_map_enter()
2766 tmp_start -= start - entry->vme_start; in vm_map_enter()
2767 tmp_offset -= start - entry->vme_start; in vm_map_enter()
2769 for (; entry->vme_start < end; in vm_map_enter()
2776 entry->vme_start != tmp_start || in vm_map_enter()
2818 tmp_offset += entry->vme_end - entry->vme_start; in vm_map_enter()
2819 tmp_start += entry->vme_end - entry->vme_start; in vm_map_enter()
2835 (entry->vme_next->vme_start < end)) { in vm_map_enter()
2929 ((entry->vme_end - entry->vme_start) + size <= in vm_map_enter()
2939 (vm_map_size_t)(entry->vme_end - entry->vme_start), in vm_map_enter()
2947 assert(entry->vme_start < end); in vm_map_enter()
2951 …ntry_extend, vm_map_t, map, vm_map_entry_t, entry, vm_address_t, entry->vme_start, vm_address_t, e… in vm_map_enter()
3062 vm_map_offset_t, new_entry->vme_start, in vm_map_enter()
3068 new_entry->vme_start)); in vm_map_enter()
3145 sp_object = vm_object_allocate((vm_map_size_t)(entry->vme_end - entry->vme_start)); in vm_map_enter()
3364 start = vm_map_zap_first_entry(&zap_old_list)->vme_start; in vm_map_enter()
3386 entry2->vme_start); in vm_map_enter()
3619 if (entry->vme_end - entry->vme_start in vm_map_enter_fourk()
3678 (entry->vme_next->vme_start < end)) { in vm_map_enter_fourk()
3736 (uint64_t) entry->vme_start, in vm_map_enter_fourk()
3781 (uint64_t) entry->vme_start, in vm_map_enter_fourk()
3796 (uint64_t) entry->vme_start, in vm_map_enter_fourk()
4318 copy_entry->vme_start); in vm_map_enter_mem_object_helper()
5420 (entry->vme_start + in vm_map_clip_unnest()
5425 if (entry->vme_start > start_unnest || in vm_map_clip_unnest()
5430 (long long)entry->vme_start, (long long)entry->vme_end); in vm_map_clip_unnest()
5433 if (start_unnest > entry->vme_start) { in vm_map_clip_unnest()
5455 entry->vme_start, in vm_map_clip_unnest()
5456 entry->vme_end - entry->vme_start); in vm_map_clip_unnest()
5460 map, entry->vme_start, in vm_map_clip_unnest()
5482 (uint64_t)entry->vme_start, in __vm_map_clip_atomic_entry_panic()
5503 startaddr >= entry->vme_start) { in vm_map_clip_start()
5518 if (startaddr > entry->vme_start) { in vm_map_clip_start()
5523 (addr64_t)(entry->vme_start), in vm_map_clip_start()
5533 vm_map_offset_t, entry->vme_start, in vm_map_clip_start()
5550 if ((startaddr) > (entry)->vme_start) \
5583 assert(new_entry->vme_start < new_entry->vme_end); in _vm_map_clip_start()
5584 VME_OFFSET_SET(entry, VME_OFFSET(entry) + (start - entry->vme_start)); in _vm_map_clip_start()
5586 entry->vme_start = start; in _vm_map_clip_start()
5629 start_unnest = entry->vme_start; in vm_map_clip_end()
5641 (addr64_t)(entry->vme_start), in vm_map_clip_end()
5650 vm_map_offset_t, entry->vme_start, in vm_map_clip_end()
5696 assert(entry->vme_start < end); in _vm_map_clip_end()
5697 new_entry->vme_start = entry->vme_end = end; in _vm_map_clip_end()
5699 VME_OFFSET(new_entry) + (end - entry->vme_start)); in _vm_map_clip_end()
5700 assert(new_entry->vme_start < new_entry->vme_end); in _vm_map_clip_end()
5784 while ((cur != vm_map_to_entry(map)) && (prev == cur->vme_start)) { in vm_map_range_check()
5920 prev = current->vme_start; in vm_map_protect()
5922 (current->vme_start < end)) { in vm_map_protect()
5926 if (current->vme_start != prev) { in vm_map_protect()
5971 uint64_t, (uint64_t) current->vme_start, in vm_map_protect()
6045 (current->vme_start < end)) { in vm_map_protect()
6109 …0x%x - <rdar://23770418> code-signing bypass?", map, (uint64_t)current->vme_start, (uint64_t)curre… in vm_map_protect()
6116 map, (uint64_t)current->vme_start, (uint64_t)current->vme_end, prot, current->wired_count); in vm_map_protect()
6137 current->vme_start, in vm_map_protect()
6142 current->vme_start, in vm_map_protect()
6154 (current->vme_start <= end)) { in vm_map_protect()
6194 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) { in vm_map_inherit()
6211 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) { in vm_map_inherit()
6249 size = entry->vme_end - entry->vme_start; in add_wire_counts()
6335 map->user_wire_size -= entry->vme_end - entry->vme_start; in subtract_wire_counts()
6554 sub_end += VME_OFFSET(entry) - entry->vme_start; in vm_map_wire_nested()
6591 s = entry->vme_start; in vm_map_wire_nested()
6598 local_start = entry->vme_start; in vm_map_wire_nested()
6651 local_start = entry->vme_start; in vm_map_wire_nested()
6688 (entry->vme_start < e)) { in vm_map_wire_nested()
6705 s = entry->vme_start; in vm_map_wire_nested()
6753 assert((entry->vme_end - entry->vme_start) in vm_map_wire_nested()
6758 if (((entry->vme_end - entry->vme_start) in vm_map_wire_nested()
6794 s = entry->vme_start; in vm_map_wire_nested()
6828 (uint64_t) entry->vme_start, in vm_map_wire_nested()
6832 uint64_t, (uint64_t)entry->vme_start, in vm_map_wire_nested()
6845 size = entry->vme_end - entry->vme_start; in vm_map_wire_nested()
6893 size = entry->vme_end - entry->vme_start; in vm_map_wire_nested()
6920 entry->vme_start, in vm_map_wire_nested()
6943 (uint64_t)entry->vme_start, in vm_map_wire_nested()
6976 (entry->vme_next->vme_start > entry->vme_end))) { in vm_map_wire_nested()
7026 tmp_entry.vme_start, in vm_map_wire_nested()
7041 if (!vm_map_lookup_entry(map, tmp_entry.vme_start, in vm_map_wire_nested()
7052 (entry->vme_start < tmp_entry.vme_end)) { in vm_map_wire_nested()
7071 (entry->vme_start != tmp_entry.vme_end)) { /* AND, the next entry is not contiguous. */ in vm_map_wire_nested()
7078 s = entry->vme_start; in vm_map_wire_nested()
7258 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) { in vm_map_unwire_nested()
7312 sub_end = entry->vme_end - entry->vme_start; in vm_map_unwire_nested()
7339 (entry->vme_next->vme_start in vm_map_unwire_nested()
7375 tmp_entry.vme_start, in vm_map_unwire_nested()
7393 (entry->vme_start < tmp_entry.vme_end)) { in vm_map_unwire_nested()
7417 tmp_entry.vme_start, in vm_map_unwire_nested()
7455 (entry->vme_next->vme_start > entry->vme_end)))) { in vm_map_unwire_nested()
7488 tmp_entry.vme_start); in vm_map_unwire_nested()
7497 if (!vm_map_lookup_entry(map, tmp_entry.vme_start, in vm_map_unwire_nested()
7515 (entry->vme_start < tmp_entry.vme_end)) { in vm_map_unwire_nested()
7572 s = entry->vme_start; in vm_map_entry_zap()
7612 remove_size = (entry->vme_end - entry->vme_start); in vm_map_submap_pmap_clean()
7613 if (offset > entry->vme_start) { in vm_map_submap_pmap_clean()
7614 remove_size -= offset - entry->vme_start; in vm_map_submap_pmap_clean()
7637 entry->vme_start), in vm_map_submap_pmap_clean()
7641 entry->vme_start, in vm_map_submap_pmap_clean()
7655 && (entry->vme_start < submap_end)) { in vm_map_submap_pmap_clean()
7656 remove_size = (entry->vme_end - entry->vme_start); in vm_map_submap_pmap_clean()
7663 (start + entry->vme_start) - offset, in vm_map_submap_pmap_clean()
7664 ((start + entry->vme_start) - offset) + remove_size, in vm_map_submap_pmap_clean()
7677 entry->vme_start, in vm_map_submap_pmap_clean()
7682 (addr64_t)((start + entry->vme_start) in vm_map_submap_pmap_clean()
7684 (addr64_t)(((start + entry->vme_start) in vm_map_submap_pmap_clean()
7817 submap_entry->vme_start < submap_end; in vm_map_delete_submap_recurse()
7876 (uint64_t)entry->vme_start, in __vm_map_delete_permanent_panic()
8062 it->vme_start < clear_in_transition_end; in vm_map_delete()
8081 if (entry == vm_map_to_entry(map) || s < entry->vme_start) { in vm_map_delete()
8103 end <= entry->vme_start) { in vm_map_delete()
8107 s = entry->vme_start; in vm_map_delete()
8180 submap_start = s - entry->vme_start; in vm_map_delete()
8182 submap_end = end - entry->vme_start; in vm_map_delete()
8200 (uint64_t)entry->vme_start, in vm_map_delete()
8206 vm_map_offset_t, entry->vme_start, in vm_map_delete()
8224 if (entry->vme_start < s) { in vm_map_delete()
8262 vm_map_offset_t, entry->vme_start, in vm_map_delete()
8270 assert(s == entry->vme_start); in vm_map_delete()
8420 tmp_entry.vme_start); in vm_map_delete()
8423 pmap_addr = tmp_entry.vme_start; in vm_map_delete()
8426 pmap_addr = tmp_entry.vme_start; in vm_map_delete()
8436 tmp_entry.vme_start, in vm_map_delete()
8444 map->pmap, tmp_entry.vme_start); in vm_map_delete()
8486 (uint64_t)entry->vme_start, in vm_map_delete()
8500 (uint64_t)entry->vme_start, in vm_map_delete()
8509 vm_map_offset_t, entry->vme_start, in vm_map_delete()
8549 (addr64_t)entry->vme_start, in vm_map_delete()
8550 entry->vme_end - entry->vme_start, in vm_map_delete()
8557 map, entry->vme_start, in vm_map_delete()
8564 map, entry->vme_start, entry->vme_end, in vm_map_delete()
8577 entry->vme_end - entry->vme_start, in vm_map_delete()
8580 entry->vme_start, in vm_map_delete()
8597 (addr64_t)entry->vme_start, in vm_map_delete()
8608 entry->vme_start, in vm_map_delete()
8616 vm_map_offset_t, entry->vme_start, in vm_map_delete()
8621 entry->vme_start)); in vm_map_delete()
8628 ret.kmr_size += entry->vme_end - entry->vme_start; in vm_map_delete()
8644 (uint64_t)entry->vme_start, in vm_map_delete()
8652 vm_map_offset_t, entry->vme_start, in vm_map_delete()
8947 vm_map_offset_t, entry->vme_start, in vm_map_entry_is_overwritable()
9026 sub_end -= entry->vme_start; in vm_map_overwrite_submap_recurse()
9088 (next->vme_start != entry->vme_end)) { in vm_map_overwrite_submap_recurse()
9288 sub_end -= entry->vme_start; in vm_map_copy_overwrite_nested()
9348 (next->vme_start != entry->vme_end)) { in vm_map_copy_overwrite_nested()
9455 if (entry->vme_start < base_addr) { in vm_map_copy_overwrite_nested()
9458 sub_start = entry->vme_start; in vm_map_copy_overwrite_nested()
9485 if (base_addr < entry->vme_start) { in vm_map_copy_overwrite_nested()
9488 entry->vme_start - base_addr; in vm_map_copy_overwrite_nested()
9498 sub_end -= entry->vme_start; in vm_map_copy_overwrite_nested()
9515 copy_entry->vme_start; in vm_map_copy_overwrite_nested()
9521 copy_entry->vme_start + in vm_map_copy_overwrite_nested()
9524 copy_entry->vme_start; in vm_map_copy_overwrite_nested()
9630 (next->vme_start != entry->vme_end)) { in vm_map_copy_overwrite_nested()
9650 copy_entry->vme_start; in vm_map_copy_overwrite_nested()
9655 copy_entry->vme_start + in vm_map_copy_overwrite_nested()
9658 copy_entry->vme_start; in vm_map_copy_overwrite_nested()
9912 entry->vme_start < dst_addr + copy_size); in vm_map_copy_overwrite()
10148 assert((start >= entry->vme_start) && (start < entry->vme_end)); in vm_map_copy_overwrite_unaligned()
10162 dst_offset = start - entry->vme_start; in vm_map_copy_overwrite_unaligned()
10167 (copy_entry->vme_start + src_offset); in vm_map_copy_overwrite_unaligned()
10197 - entry->vme_start), in vm_map_copy_overwrite_unaligned()
10213 entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_unaligned()
10255 if ((copy_entry->vme_start + src_offset) == copy_entry->vme_end in vm_map_copy_overwrite_unaligned()
10299 if (start != entry->vme_start) { in vm_map_copy_overwrite_unaligned()
10364 copy_size = (copy_entry->vme_end - copy_entry->vme_start); in vm_map_copy_overwrite_aligned()
10375 size = (entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_aligned()
10383 if ((entry->vme_start != start) || ((entry->is_sub_map) in vm_map_copy_overwrite_aligned()
10410 !VM_MAP_PAGE_ALIGNED(entry->vme_start + copy_size, in vm_map_copy_overwrite_aligned()
10415 vm_map_clip_end(dst_map, entry, entry->vme_start + copy_size); in vm_map_copy_overwrite_aligned()
10425 copy_entry->vme_start + size); in vm_map_copy_overwrite_aligned()
10429 assert((entry->vme_end - entry->vme_start) == size); in vm_map_copy_overwrite_aligned()
10430 assert((tmp_entry->vme_end - tmp_entry->vme_start) == size); in vm_map_copy_overwrite_aligned()
10431 assert((copy_entry->vme_end - copy_entry->vme_start) == size); in vm_map_copy_overwrite_aligned()
10558 (addr64_t)entry->vme_start, in vm_map_copy_overwrite_aligned()
10559 entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_aligned()
10565 dst_map, entry->vme_start, in vm_map_copy_overwrite_aligned()
10572 dst_map, entry->vme_start, in vm_map_copy_overwrite_aligned()
10584 - entry->vme_start, in vm_map_copy_overwrite_aligned()
10587 entry->vme_start, in vm_map_copy_overwrite_aligned()
10593 (addr64_t)(entry->vme_start), in vm_map_copy_overwrite_aligned()
10649 entry->vme_start), in vm_map_copy_overwrite_aligned()
10677 entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_aligned()
10726 copy_entry->vme_start + copy_size); in vm_map_copy_overwrite_aligned()
11023 new_entry->vme_start += adjustment; in vm_map_copy_remap()
11358 entry->vme_start += adjustment; in vm_map_copyout_internal()
11362 assert(VM_MAP_PAGE_ALIGNED(entry->vme_start, in vm_map_copyout_internal()
11389 va = entry->vme_start; in vm_map_copyout_internal()
11392 entry->vme_start, in vm_map_copyout_internal()
11751 if (src_start < tmp_entry->vme_start) { in vm_map_copyin_internal()
11756 src_start = tmp_entry->vme_start; in vm_map_copyin_internal()
11808 src_start -= tmp_entry->vme_start; in vm_map_copyin_internal()
11957 src_entry->vme_start, in vm_map_copyin_internal()
12307 if ((tmp_entry->vme_start != src_start) || in vm_map_copyin_internal()
12352 original_start = tmp_entry->vme_start; in vm_map_copyin_internal()
12356 adjustment = (tmp_entry->vme_start - in vm_map_copyin_internal()
12358 tmp_entry->vme_start, in vm_map_copyin_internal()
12360 tmp_entry->vme_start -= adjustment; in vm_map_copyin_internal()
12364 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12374 tmp_entry->vme_start += adjustment; in vm_map_copyin_internal()
12379 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12386 assert(tmp_entry->vme_start >= original_start); in vm_map_copyin_internal()
12392 assert(vm_map_trunc_page(tmp_entry->vme_start, in vm_map_copyin_internal()
12421 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12449 tmp_entry->vme_start), in vm_map_copyin_internal()
12462 (tmp_entry->vme_end - tmp_entry->vme_start); in vm_map_copyin_internal()
12463 tmp_entry->vme_start = copy_addr; in vm_map_copyin_internal()
12464 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12465 copy_addr += tmp_entry->vme_end - tmp_entry->vme_start; in vm_map_copyin_internal()
12466 copy_size += tmp_entry->vme_end - tmp_entry->vme_start; in vm_map_copyin_internal()
12637 (addr64_t)old_entry->vme_start, in vm_map_fork_share()
12638 (uint64_t)(old_entry->vme_end - old_entry->vme_start)); in vm_map_fork_share()
12647 old_entry->vme_start)); in vm_map_fork_share()
12667 old_entry->vme_start)))) { in vm_map_fork_share()
12746 old_entry->vme_start), in vm_map_fork_share()
12775 old_entry->vme_start), in vm_map_fork_share()
12778 old_entry->vme_start, in vm_map_fork_share()
12782 old_entry->vme_start, in vm_map_fork_share()
12858 pmap_copy(new_map->pmap, old_map->pmap, new_entry->vme_start, in vm_map_fork_share()
12859 old_entry->vme_end - old_entry->vme_start, in vm_map_fork_share()
12860 old_entry->vme_start); in vm_map_fork_share()
12872 vm_map_size_t entry_size = old_entry->vme_end - old_entry->vme_start; in vm_map_fork_copy()
12873 vm_map_offset_t start = old_entry->vme_start; in vm_map_fork_copy()
12932 if (last->vme_start == start) { in vm_map_fork_copy()
13136 entry_size = old_entry->vme_end - old_entry->vme_start; in vm_map_fork()
13148 if (prev_end < old_entry->vme_start) { in vm_map_fork()
13152 prev_end, old_entry->vme_start); in vm_map_fork()
13157 …x%llx nested\n", __FUNCTION__, __LINE__, new_pmap, (uint64_t)old_entry->vme_start, (uint64_t)old_e… in vm_map_fork()
13163 old_entry->vme_start, old_entry->vme_end); in vm_map_fork()
13258 old_entry->vme_start), in vm_map_fork()
13287 old_entry->vme_start), in vm_map_fork()
13293 old_entry->vme_start, in vm_map_fork()
13548 (vaddr < entry->vme_start) || (vaddr >= entry->vme_end)) { in vm_map_lookup_and_lock_object()
13569 old_start = entry->vme_start; in vm_map_lookup_and_lock_object()
13628 old_start = entry->vme_start; in vm_map_lookup_and_lock_object()
13659 local_vaddr = (local_vaddr - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
13662 (uint64_t)local_vaddr, (uint64_t)entry->vme_start, (uint64_t)fault_page_mask); in vm_map_lookup_and_lock_object()
13687 start_delta = submap_entry->vme_start > VME_OFFSET(entry) ? in vm_map_lookup_and_lock_object()
13688 submap_entry->vme_start - VME_OFFSET(entry) : 0; in vm_map_lookup_and_lock_object()
13732 submap_entry->vme_start)); in vm_map_lookup_and_lock_object()
13754 submap_entry_size = submap_entry->vme_end - submap_entry->vme_start; in vm_map_lookup_and_lock_object()
13871 submap_entry->vme_end - submap_entry->vme_start, in vm_map_lookup_and_lock_object()
13940 submap_entry->vme_start, in vm_map_lookup_and_lock_object()
13945 submap_entry->vme_start, in vm_map_lookup_and_lock_object()
13958 submap_entry->vme_start + in vm_map_lookup_and_lock_object()
13999 uint64_t, (uint64_t)entry->vme_start, in vm_map_lookup_and_lock_object()
14081 uint64_t, (uint64_t)entry->vme_start, in vm_map_lookup_and_lock_object()
14215 entry->vme_start), in vm_map_lookup_and_lock_object()
14256 entry->vme_start)), false, 0); in vm_map_lookup_and_lock_object()
14268 *offset = (vaddr - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14286 (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14508 (curr_entry->vme_start >= in vm_map_region_recurse_64()
14525 skip = curr_entry->vme_start - curr_address; in vm_map_region_recurse_64()
14526 curr_address = curr_entry->vme_start; in vm_map_region_recurse_64()
14541 } else if (tmp_entry->vme_start >= in vm_map_region_recurse_64()
14548 (tmp_entry->vme_start + curr_offset <= in vm_map_region_recurse_64()
14549 next_entry->vme_start + next_offset)) { in vm_map_region_recurse_64()
14564 next_address = next_entry->vme_start; in vm_map_region_recurse_64()
14574 next_address - next_entry->vme_start); in vm_map_region_recurse_64()
14588 curr_address - curr_entry->vme_start); in vm_map_region_recurse_64()
14636 (VME_OFFSET(curr_entry) - curr_entry->vme_start); in vm_map_region_recurse_64()
14809 curr_entry->vme_start); in vm_map_region_recurse_64()
14817 curr_entry->vme_start)), in vm_map_region_recurse_64()
14931 start = entry->vme_start; in vm_map_region()
14979 start = entry->vme_start; in vm_map_region()
15034 start = entry->vme_start; in vm_map_region()
15090 start = entry->vme_start; in vm_map_region()
15134 entry_size = (uint32_t) ((entry->vme_end - entry->vme_start) / PAGE_SIZE_64); in vm_map_region_top_walk()
15552 (prev_entry->vme_end == this_entry->vme_start) && in vm_map_simplify_entry()
15558 prev_entry->vme_start)) in vm_map_simplify_entry()
15598 assert(prev_entry->vme_start < this_entry->vme_end); in vm_map_simplify_entry()
15600 assert(VM_MAP_PAGE_ALIGNED(prev_entry->vme_start, in vm_map_simplify_entry()
15603 this_entry->vme_start = prev_entry->vme_start; in vm_map_simplify_entry()
15669 entry->vme_start <= end) { in vm_map_simplify_range()
15737 sub_start = (start - entry->vme_start) in vm_map_machine_attribute()
15754 offset = (start - entry->vme_start) in vm_map_machine_attribute()
15864 while ((entry != vm_map_to_entry(map)) && (entry->vme_start < end)) { in vm_map_behavior_set()
15974 offset = (start - entry->vme_start) + VME_OFFSET(entry); in vm_map_willneed()
16008 start = entry->vme_start; in vm_map_willneed()
16046 start = entry->vme_start; in vm_map_willneed()
16241 for (; entry != vm_map_to_entry(map) && entry->vme_start < end; in vm_map_reuse_pages()
16257 if (entry->vme_start < start) { in vm_map_reuse_pages()
16258 start_offset = start - entry->vme_start; in vm_map_reuse_pages()
16262 end_offset = MIN(end, entry->vme_end) - entry->vme_start; in vm_map_reuse_pages()
16336 for (; entry != vm_map_to_entry(map) && entry->vme_start < end; in vm_map_reusable_pages()
16362 if (entry->vme_start < start) { in vm_map_reusable_pages()
16363 start_offset = start - entry->vme_start; in vm_map_reusable_pages()
16367 pmap_offset = entry->vme_start; in vm_map_reusable_pages()
16369 end_offset = MIN(end, entry->vme_end) - entry->vme_start; in vm_map_reusable_pages()
16470 for (; entry != vm_map_to_entry(map) && entry->vme_start < end; in vm_map_can_reuse()
16518 for (; entry != vm_map_to_entry(map) && entry->vme_start < end; in vm_map_pageout()
16534 entry->vme_start); in vm_map_pageout()
16634 new_entry->vme_start = start; in vm_map_entry_insert()
16840 if (src_start < src_entry->vme_start || in vm_map_remap_extract()
16841 (mapped_size && src_start != src_entry->vme_start)) { in vm_map_remap_extract()
16852 src_entry->vme_start); in vm_map_remap_extract()
16871 submap_start = VME_OFFSET(src_entry) + src_start - src_entry->vme_start; in vm_map_remap_extract()
16980 submap_start = VME_OFFSET(src_entry) + src_start - src_entry->vme_start; in vm_map_remap_extract()
17093 copy_entry_size = copy_entry->vme_end - copy_entry->vme_start; in vm_map_remap_extract()
17094 copy_entry->vme_start = map_address; in vm_map_remap_extract()
17148 (uint64_t)src_entry->vme_start, in vm_map_remap_extract()
17162 (uint64_t)src_entry->vme_start, in vm_map_remap_extract()
17240 src_entry->vme_start, in vm_map_remap_extract()
17256 src_entry->vme_start, in vm_map_remap_extract()
17291 (src_start - src_entry->vme_start)); in vm_map_remap_extract()
17315 new_entry->vme_start = map_address; in vm_map_remap_extract()
17317 assert(new_entry->vme_start < new_entry->vme_end); in vm_map_remap_extract()
17378 new_entry->vme_start), in vm_map_remap_extract()
17410 src_entry->vme_start, in vm_map_remap_extract()
17449 new_entry->vme_start), in vm_map_remap_extract()
17466 new_entry->vme_start), in vm_map_remap_extract()
17664 new_entry->vme_start = 0; in vm_map_copy_to_physcopy()
17833 trim_start += vm_map_copy_first_entry(copy_map)->vme_start; in vm_map_copy_trim()
17834 trim_end += vm_map_copy_first_entry(copy_map)->vme_start; in vm_map_copy_trim()
17848 if (entry->vme_start >= trim_end) { in vm_map_copy_trim()
17856 copy_map->size -= entry->vme_end - entry->vme_start; in vm_map_copy_trim()
17996 object_offset_end += entry->vme_end - entry->vme_start; in vm_map_copy_adjust_to_target()
18085 … (uint64_t)target_copy_map->size, target_entry, (uint64_t)target_entry->vme_start, (uint64_t)targe… in vm_map_copy_adjust_to_target()
18088 … (uint64_t)target_copy_map->size, target_entry, (uint64_t)target_entry->vme_start, (uint64_t)targe… in vm_map_copy_adjust_to_target()
18107 target_entry->vme_start += overmap_start; in vm_map_copy_adjust_to_target()
18111 object_offset_end = VME_OFFSET(target_entry) + target_entry->vme_end - target_entry->vme_start; in vm_map_copy_adjust_to_target()
18113 … (uint64_t)target_copy_map->size, target_entry, (uint64_t)target_entry->vme_start, (uint64_t)targe… in vm_map_copy_adjust_to_target()
18128 target_entry->vme_start -= addr_adjustment; in vm_map_copy_adjust_to_target()
18130 … (uint64_t)target_copy_map->size, target_entry, (uint64_t)target_entry->vme_start, (uint64_t)targe… in vm_map_copy_adjust_to_target()
18532 vm_map_offset_t, entry->vme_start, in vm_map_remap()
18556 entry->vme_start += *address; in vm_map_remap()
18565 assert(VM_MAP_PAGE_ALIGNED(entry->vme_start, MIN(target_page_mask, PAGE_MASK))); in vm_map_remap()
18737 (entry->vme_next->vme_start < end)) { in vm_map_remap_range_allocate()
18895 if (start < entry->vme_start) { in vm_map_check_protection()
18990 entry->vme_end - entry->vme_start != object->vo_size) { in vm_map_purgable_control()
19037 assert(curr_s_offset >= map_entry->vme_start); in vm_map_footprint_query_page_info()
19082 - map_entry->vme_start in vm_map_footprint_query_page_info()
19109 - map_entry->vme_start in vm_map_footprint_query_page_info()
19383 curr_e_offset = MIN(map_entry->vme_next->vme_start, end); in vm_map_page_range_info_internal()
19392 curr_e_offset = MIN(map_entry->vme_next->vme_start, end); in vm_map_page_range_info_internal()
19412 offset_in_object = curr_s_offset - map_entry->vme_start; in vm_map_page_range_info_internal()
19810 (entry->vme_next->vme_start < address)) { in vm_map_msync()
19817 skip = (entry->vme_next->vme_start - address); in vm_map_msync()
19823 address = entry->vme_next->vme_start; in vm_map_msync()
19828 offset = address - entry->vme_start; in vm_map_msync()
19835 if (amount_left + entry->vme_start + offset > entry->vme_end) { in vm_map_msync()
19837 (entry->vme_start + offset); in vm_map_msync()
20016 copy_entry->vme_start = VM_MAP_TRUNC_PAGE(offset, PAGE_MASK); in vm_named_entry_associate_vm_object()
20054 …%llx prot 0x%x\n", named_entry, copy, copy_entry, (uint64_t)copy_entry->vme_start, (uint64_t)copy_… in vm_named_entry_to_vm_object()
20450 first_entry->vme_start < new_min_offset) { in vm_map_raise_min_offset()
20651 if ((entry->vme_start > start) || (entry->vme_end < end)) { in vm_map_sign()
20676 start - entry->vme_start + VME_OFFSET(entry)); in vm_map_sign()
20740 (void)vm_map_delete(map, entry->vme_start, in vm_map_partial_reap()
20778 vm_map_clip_unnest(map, entry, entry->vme_start, entry->vme_end); in vm_map_disconnect_page_mappings()
20799 pmap_remove_options(map->pmap, entry->vme_start, entry->vme_end, 0); in vm_map_disconnect_page_mappings()
21291 volatile_virtual_size += entry->vme_end - entry->vme_start; in vm_map_query_volatile()
21300 entry->vme_start, in vm_map_query_volatile()
21348 free = entry->vme_start - prev; in vm_map_sizes()
21648 va = old_entry->vme_start; in vm_map_corpse_footprint_collect()
21734 != old_entry->vme_start) { in vm_map_corpse_footprint_collect()
21744 num_pages_delta = ((old_entry->vme_start - in vm_map_corpse_footprint_collect()
21783 footprint_region->cfr_vaddr = old_entry->vme_start; in vm_map_corpse_footprint_collect()
21808 for (va = old_entry->vme_start; in vm_map_corpse_footprint_collect()