Lines Matching refs:vme_end

885 	    map_addr = tmp_entry.vme_end) {  in vm_map_apple_protected()
932 crypto_end = tmp_entry.vme_end - tmp_entry.vme_start; in vm_map_apple_protected()
939 if (tmp_entry.vme_end > end) { in vm_map_apple_protected()
940 if (tmp_entry.vme_end != end_aligned) { in vm_map_apple_protected()
943 crypto_end -= (tmp_entry.vme_end - end); in vm_map_apple_protected()
1006 (tmp_entry.vme_end - in vm_map_apple_protected()
1034 (uint64_t) (map_addr + (tmp_entry.vme_end - in vm_map_apple_protected()
1055 crypto_backing_offset += (tmp_entry.vme_end - in vm_map_apple_protected()
2168 new_entry->vme_end = hint_address + size; in vm_map_find_space()
2183 assert(page_aligned(new_entry->vme_end)); in vm_map_find_space()
2185 assert(VM_MAP_PAGE_ALIGNED(new_entry->vme_end, VM_MAP_PAGE_MASK(map))); in vm_map_find_space()
2818 tmp_offset += entry->vme_end - entry->vme_start; in vm_map_enter()
2819 tmp_start += entry->vme_end - entry->vme_start; in vm_map_enter()
2820 if (entry->vme_end >= end) { in vm_map_enter()
2898 (entry->vme_end == start) && in vm_map_enter()
2929 ((entry->vme_end - entry->vme_start) + size <= in vm_map_enter()
2939 (vm_map_size_t)(entry->vme_end - entry->vme_start), in vm_map_enter()
2940 (vm_map_size_t)(end - entry->vme_end))) { in vm_map_enter()
2946 map->size += (end - entry->vme_end); in vm_map_enter()
2951 …map_entry_t, entry, vm_address_t, entry->vme_start, vm_address_t, entry->vme_end, vm_address_t, en… in vm_map_enter()
2953 entry->vme_end = end; in vm_map_enter()
3063 vm_map_offset_t, new_entry->vme_end, in vm_map_enter()
3067 (new_entry->vme_end - in vm_map_enter()
3145 sp_object = vm_object_allocate((vm_map_size_t)(entry->vme_end - entry->vme_start)); in vm_map_enter()
3365 end = vm_map_zap_last_entry(&zap_old_list)->vme_end; in vm_map_enter()
3385 entry_size = (entry2->vme_end - in vm_map_enter()
3619 if (entry->vme_end - entry->vme_start in vm_map_enter_fourk()
3737 (uint64_t) entry->vme_end, in vm_map_enter_fourk()
3782 (uint64_t) entry->vme_end, in vm_map_enter_fourk()
3797 (uint64_t) entry->vme_end, in vm_map_enter_fourk()
4317 copy_size = (copy_entry->vme_end - in vm_map_enter_mem_object_helper()
5426 entry->vme_end < end_unnest) { in vm_map_clip_unnest()
5430 (long long)entry->vme_start, (long long)entry->vme_end); in vm_map_clip_unnest()
5443 if (entry->vme_end > end_unnest) { in vm_map_clip_unnest()
5456 entry->vme_end - entry->vme_start); in vm_map_clip_unnest()
5461 entry->vme_end, in vm_map_clip_unnest()
5483 (uint64_t)entry->vme_end, in __vm_map_clip_atomic_entry_panic()
5524 (addr64_t)(entry->vme_end)); in vm_map_clip_start()
5534 vm_map_offset_t, entry->vme_end, in vm_map_clip_start()
5582 new_entry->vme_end = start; in _vm_map_clip_start()
5583 assert(new_entry->vme_start < new_entry->vme_end); in _vm_map_clip_start()
5585 assert(start < entry->vme_end); in _vm_map_clip_start()
5611 if (endaddr > entry->vme_end) { in vm_map_clip_end()
5616 endaddr = entry->vme_end; in vm_map_clip_end()
5636 if (endaddr < entry->vme_end) { in vm_map_clip_end()
5642 (addr64_t)(entry->vme_end)); in vm_map_clip_end()
5651 vm_map_offset_t, entry->vme_end, in vm_map_clip_end()
5667 if ((endaddr) < (entry)->vme_end) \
5697 new_entry->vme_start = entry->vme_end = end; in _vm_map_clip_end()
5700 assert(new_entry->vme_start < new_entry->vme_end); in _vm_map_clip_end()
5774 if (end <= cur->vme_end) { in vm_map_range_check()
5782 prev = cur->vme_end; in vm_map_range_check()
5785 if (end <= cur->vme_end) { in vm_map_range_check()
5788 prev = cur->vme_end; in vm_map_range_check()
5972 uint64_t, (uint64_t) current->vme_end, in vm_map_protect()
6001 prev = current->vme_end; in vm_map_protect()
6013 (vm_map_round_page(prev_entry->vme_end, in vm_map_protect()
6109 …e-signing bypass?", map, (uint64_t)current->vme_start, (uint64_t)current->vme_end, old_prot, prot); in vm_map_protect()
6116 map, (uint64_t)current->vme_start, (uint64_t)current->vme_end, prot, current->wired_count); in vm_map_protect()
6138 current->vme_end, in vm_map_protect()
6143 current->vme_end, in vm_map_protect()
6249 size = entry->vme_end - entry->vme_start; in add_wire_counts()
6335 map->user_wire_size -= entry->vme_end - entry->vme_start; in subtract_wire_counts()
6461 e = entry->vme_end; in vm_map_wire_nested()
6553 sub_end = entry->vme_end; in vm_map_wire_nested()
6556 local_end = entry->vme_end; in vm_map_wire_nested()
6640 e = entry->vme_end; in vm_map_wire_nested()
6681 e = entry->vme_end; in vm_map_wire_nested()
6753 assert((entry->vme_end - entry->vme_start) in vm_map_wire_nested()
6758 if (((entry->vme_end - entry->vme_start) in vm_map_wire_nested()
6829 (uint64_t) entry->vme_end); in vm_map_wire_nested()
6833 uint64_t, (uint64_t)entry->vme_end); in vm_map_wire_nested()
6845 size = entry->vme_end - entry->vme_start; in vm_map_wire_nested()
6893 size = entry->vme_end - entry->vme_start; in vm_map_wire_nested()
6944 (uint64_t)entry->vme_end, in vm_map_wire_nested()
6962 e = entry->vme_end; in vm_map_wire_nested()
6974 if ((entry->vme_end < end) && in vm_map_wire_nested()
6976 (entry->vme_next->vme_start > entry->vme_end))) { in vm_map_wire_nested()
7052 (entry->vme_start < tmp_entry.vme_end)) { in vm_map_wire_nested()
7070 (tmp_entry.vme_end != end) && /* AND, we are not at the end of the requested range */ in vm_map_wire_nested()
7071 (entry->vme_start != tmp_entry.vme_end)) { /* AND, the next entry is not contiguous. */ in vm_map_wire_nested()
7073 s = tmp_entry.vme_end; in vm_map_wire_nested()
7312 sub_end = entry->vme_end - entry->vme_start; in vm_map_unwire_nested()
7314 local_end = entry->vme_end; in vm_map_unwire_nested()
7337 if (((entry->vme_end < end) && in vm_map_unwire_nested()
7340 > entry->vme_end)))) { in vm_map_unwire_nested()
7393 (entry->vme_start < tmp_entry.vme_end)) { in vm_map_unwire_nested()
7453 if (((entry->vme_end < end) && in vm_map_unwire_nested()
7455 (entry->vme_next->vme_start > entry->vme_end)))) { in vm_map_unwire_nested()
7515 (entry->vme_start < tmp_entry.vme_end)) { in vm_map_unwire_nested()
7573 e = entry->vme_end; in vm_map_entry_zap()
7612 remove_size = (entry->vme_end - entry->vme_start); in vm_map_submap_pmap_clean()
7618 if (submap_end < entry->vme_end) { in vm_map_submap_pmap_clean()
7620 entry->vme_end - submap_end; in vm_map_submap_pmap_clean()
7656 remove_size = (entry->vme_end - entry->vme_start); in vm_map_submap_pmap_clean()
7657 if (submap_end < entry->vme_end) { in vm_map_submap_pmap_clean()
7658 remove_size -= entry->vme_end - submap_end; in vm_map_submap_pmap_clean()
7877 (uint64_t)entry->vme_end); in __vm_map_delete_permanent_panic()
8121 end = entry->vme_end; in vm_map_delete()
8201 (uint64_t)entry->vme_end, in vm_map_delete()
8207 vm_map_offset_t, entry->vme_end, in vm_map_delete()
8238 if (end < entry->vme_end) { in vm_map_delete()
8263 vm_map_offset_t, entry->vme_end, in vm_map_delete()
8271 assert(entry->vme_end <= end); in vm_map_delete()
8419 sub_end = sub_start + (tmp_entry.vme_end - in vm_map_delete()
8437 tmp_entry.vme_end, in vm_map_delete()
8456 clear_in_transition_end = tmp_entry.vme_end; in vm_map_delete()
8487 (uint64_t)entry->vme_end, in vm_map_delete()
8501 (uint64_t)entry->vme_end, in vm_map_delete()
8510 vm_map_offset_t, entry->vme_end, in vm_map_delete()
8550 entry->vme_end - entry->vme_start, in vm_map_delete()
8558 entry->vme_end, in vm_map_delete()
8564 map, entry->vme_start, entry->vme_end, in vm_map_delete()
8577 entry->vme_end - entry->vme_start, in vm_map_delete()
8598 (addr64_t)entry->vme_end, in vm_map_delete()
8609 entry->vme_end)); in vm_map_delete()
8617 vm_map_offset_t, entry->vme_end, in vm_map_delete()
8620 (entry->vme_end - in vm_map_delete()
8626 s = entry->vme_end; in vm_map_delete()
8628 ret.kmr_size += entry->vme_end - entry->vme_start; in vm_map_delete()
8645 (uint64_t)entry->vme_end, in vm_map_delete()
8653 vm_map_offset_t, entry->vme_end, in vm_map_delete()
8948 vm_map_offset_t, entry->vme_end, in vm_map_entry_is_overwritable()
9021 if (entry->vme_end < dst_end) { in vm_map_overwrite_submap_recurse()
9022 sub_end = entry->vme_end; in vm_map_overwrite_submap_recurse()
9028 local_end = entry->vme_end; in vm_map_overwrite_submap_recurse()
9039 if (dst_end <= entry->vme_end) { in vm_map_overwrite_submap_recurse()
9080 if (dst_end <= entry->vme_end) { in vm_map_overwrite_submap_recurse()
9088 (next->vme_start != entry->vme_end)) { in vm_map_overwrite_submap_recurse()
9274 local_end = entry->vme_end; in vm_map_copy_overwrite_nested()
9283 if (entry->vme_end < dst_end) { in vm_map_copy_overwrite_nested()
9284 sub_end = entry->vme_end; in vm_map_copy_overwrite_nested()
9302 if (dst_end <= entry->vme_end) { in vm_map_copy_overwrite_nested()
9341 if (dst_end <= entry->vme_end) { in vm_map_copy_overwrite_nested()
9348 (next->vme_start != entry->vme_end)) { in vm_map_copy_overwrite_nested()
9450 if (entry->vme_end < dst_end) { in vm_map_copy_overwrite_nested()
9451 sub_end = entry->vme_end; in vm_map_copy_overwrite_nested()
9493 if (entry->vme_end < dst_end) { in vm_map_copy_overwrite_nested()
9494 sub_end = entry->vme_end; in vm_map_copy_overwrite_nested()
9500 local_end = entry->vme_end; in vm_map_copy_overwrite_nested()
9514 entry_size = copy_entry->vme_end - in vm_map_copy_overwrite_nested()
9523 entry_size = copy_entry->vme_end - in vm_map_copy_overwrite_nested()
9624 if (dst_end <= entry->vme_end) { in vm_map_copy_overwrite_nested()
9630 (next->vme_start != entry->vme_end)) { in vm_map_copy_overwrite_nested()
9649 entry_size = copy_entry->vme_end - in vm_map_copy_overwrite_nested()
9657 entry_size = copy_entry->vme_end - in vm_map_copy_overwrite_nested()
9938 if (entry->vme_end < copy->offset + head_size) { in vm_map_copy_overwrite()
9939 head_size = entry->vme_end - copy->offset; in vm_map_copy_overwrite()
10148 assert((start >= entry->vme_start) && (start < entry->vme_end)); in vm_map_copy_overwrite_unaligned()
10164 dst_size = entry->vme_end - start; in vm_map_copy_overwrite_unaligned()
10166 src_size = copy_entry->vme_end - in vm_map_copy_overwrite_unaligned()
10196 (vm_map_size_t)(entry->vme_end in vm_map_copy_overwrite_unaligned()
10213 entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_unaligned()
10226 entry_end = entry->vme_end; in vm_map_copy_overwrite_unaligned()
10255 if ((copy_entry->vme_start + src_offset) == copy_entry->vme_end in vm_map_copy_overwrite_unaligned()
10364 copy_size = (copy_entry->vme_end - copy_entry->vme_start); in vm_map_copy_overwrite_aligned()
10375 size = (entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_aligned()
10429 assert((entry->vme_end - entry->vme_start) == size); in vm_map_copy_overwrite_aligned()
10430 assert((tmp_entry->vme_end - tmp_entry->vme_start) == size); in vm_map_copy_overwrite_aligned()
10431 assert((copy_entry->vme_end - copy_entry->vme_start) == size); in vm_map_copy_overwrite_aligned()
10461 start = tmp_entry->vme_end; in vm_map_copy_overwrite_aligned()
10559 entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_aligned()
10566 entry->vme_end, in vm_map_copy_overwrite_aligned()
10573 entry->vme_end, in vm_map_copy_overwrite_aligned()
10583 entry->vme_end in vm_map_copy_overwrite_aligned()
10594 (addr64_t)(entry->vme_end), in vm_map_copy_overwrite_aligned()
10637 start = tmp_entry->vme_end; in vm_map_copy_overwrite_aligned()
10648 (entry->vme_end - in vm_map_copy_overwrite_aligned()
10677 entry->vme_end - entry->vme_start); in vm_map_copy_overwrite_aligned()
11024 new_entry->vme_end += adjustment; in vm_map_copy_remap()
11359 entry->vme_end += adjustment; in vm_map_copyout_internal()
11364 assert(VM_MAP_PAGE_ALIGNED(entry->vme_end, in vm_map_copyout_internal()
11393 entry->vme_end, in vm_map_copyout_internal()
11396 while (va < entry->vme_end) { in vm_map_copyout_internal()
11802 submap_len = tmp_entry->vme_end - src_start; in vm_map_copyin_internal()
11880 src_size = src_entry->vme_end - src_start; in vm_map_copyin_internal()
12165 if (src_entry->vme_end < new_entry->vme_end) { in vm_map_copyin_internal()
12176 assert(VM_MAP_PAGE_ALIGNED(src_entry->vme_end, in vm_map_copyin_internal()
12178 new_entry->vme_end = src_entry->vme_end; in vm_map_copyin_internal()
12179 src_size = new_entry->vme_end - src_start; in vm_map_copyin_internal()
12180 } else if (src_entry->vme_end > new_entry->vme_end) { in vm_map_copyin_internal()
12198 (src_entry->vme_end > new_entry->vme_end)) { in vm_map_copyin_internal()
12240 src_start = new_entry->vme_end; in vm_map_copyin_internal()
12364 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12379 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12403 original_end = tmp_entry->vme_end; in vm_map_copyin_internal()
12406 tmp_entry->vme_end = in vm_map_copyin_internal()
12407 vm_map_round_page(tmp_entry->vme_end, in vm_map_copyin_internal()
12420 tmp_entry->vme_end -= adjustment; in vm_map_copyin_internal()
12421 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12428 assert(tmp_entry->vme_end <= original_end); in vm_map_copyin_internal()
12433 assert(vm_map_round_page(tmp_entry->vme_end, in vm_map_copyin_internal()
12448 copy_addr + (tmp_entry->vme_end - in vm_map_copyin_internal()
12461 tmp_entry->vme_end = copy_addr + in vm_map_copyin_internal()
12462 (tmp_entry->vme_end - tmp_entry->vme_start); in vm_map_copyin_internal()
12464 assert(tmp_entry->vme_start < tmp_entry->vme_end); in vm_map_copyin_internal()
12465 copy_addr += tmp_entry->vme_end - tmp_entry->vme_start; in vm_map_copyin_internal()
12466 copy_size += tmp_entry->vme_end - tmp_entry->vme_start; in vm_map_copyin_internal()
12638 (uint64_t)(old_entry->vme_end - old_entry->vme_start)); in vm_map_fork_share()
12646 object = vm_object_allocate((vm_map_size_t)(old_entry->vme_end - in vm_map_fork_share()
12666 (vm_map_size_t)(old_entry->vme_end - in vm_map_fork_share()
12745 (vm_map_size_t) (old_entry->vme_end - in vm_map_fork_share()
12774 (old_entry->vme_end - in vm_map_fork_share()
12783 old_entry->vme_end, in vm_map_fork_share()
12859 old_entry->vme_end - old_entry->vme_start, in vm_map_fork_share()
12872 vm_map_size_t entry_size = old_entry->vme_end - old_entry->vme_start; in vm_map_fork_copy()
13120 } else if (new_entry->vme_end < vm_map_max(new_map)) { in vm_map_fork()
13124 new_entry->vme_end, vm_map_max(new_map)); in vm_map_fork()
13136 entry_size = old_entry->vme_end - old_entry->vme_start; in vm_map_fork()
13146 prev_end = old_entry->vme_prev->vme_end; in vm_map_fork()
13157 …", __FUNCTION__, __LINE__, new_pmap, (uint64_t)old_entry->vme_start, (uint64_t)old_entry->vme_end); in vm_map_fork()
13163 old_entry->vme_start, old_entry->vme_end); in vm_map_fork()
13257 (old_entry->vme_end - in vm_map_fork()
13286 (old_entry->vme_end - in vm_map_fork()
13337 } else if (new_entry->vme_end < vm_map_max(new_map)) { in vm_map_fork()
13341 new_entry->vme_end, vm_map_max(new_map)); in vm_map_fork()
13548 (vaddr < entry->vme_start) || (vaddr >= entry->vme_end)) { in vm_map_lookup_and_lock_object()
13570 old_end = entry->vme_end; in vm_map_lookup_and_lock_object()
13629 old_end = entry->vme_end; in vm_map_lookup_and_lock_object()
13692 submap_entry->vme_end ? in vm_map_lookup_and_lock_object()
13695 - submap_entry->vme_end; in vm_map_lookup_and_lock_object()
13731 (submap_entry->vme_end - in vm_map_lookup_and_lock_object()
13754 submap_entry_size = submap_entry->vme_end - submap_entry->vme_start; in vm_map_lookup_and_lock_object()
13871 submap_entry->vme_end - submap_entry->vme_start, in vm_map_lookup_and_lock_object()
13939 submap_entry->vme_end - in vm_map_lookup_and_lock_object()
14000 uint64_t, (uint64_t)entry->vme_end, in vm_map_lookup_and_lock_object()
14082 uint64_t, (uint64_t)entry->vme_end, in vm_map_lookup_and_lock_object()
14214 (vm_map_size_t) (entry->vme_end - in vm_map_lookup_and_lock_object()
14255 (vm_map_size_t)(entry->vme_end - in vm_map_lookup_and_lock_object()
14286 (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14571 next_entry->vme_end - next_address); in vm_map_region_recurse_64()
14586 curr_entry->vme_end - curr_address); in vm_map_region_recurse_64()
14654 user_address <= vm_map_last_entry(map)->vme_end) { in vm_map_region_recurse_64()
14715 *address = vm_map_last_entry(map)->vme_end; in vm_map_region_recurse_64()
14811 curr_entry->vme_end); in vm_map_region_recurse_64()
14941 *size = (entry->vme_end - start); in vm_map_region()
14989 *size = (entry->vme_end - start); in vm_map_region()
15053 …vm_map_region_walk(map, start, entry, VME_OFFSET(entry), entry->vme_end - start, extended, TRUE, *… in vm_map_region()
15063 *size = (entry->vme_end - start); in vm_map_region()
15101 *size = (entry->vme_end - start); in vm_map_region()
15134 entry_size = (uint32_t) ((entry->vme_end - entry->vme_start) / PAGE_SIZE_64); in vm_map_region_top_walk()
15552 (prev_entry->vme_end == this_entry->vme_start) && in vm_map_simplify_entry()
15557 ((VME_OFFSET(prev_entry) + (prev_entry->vme_end - in vm_map_simplify_entry()
15598 assert(prev_entry->vme_start < this_entry->vme_end); in vm_map_simplify_entry()
15726 if ((entry->vme_end - start) > sync_size) { in vm_map_machine_attribute()
15730 sub_size = entry->vme_end - start; in vm_map_machine_attribute()
15986 len = MIN(entry->vme_end - start, end - start); in vm_map_willneed()
16262 end_offset = MIN(end, entry->vme_end) - entry->vme_start; in vm_map_reuse_pages()
16369 end_offset = MIN(end, entry->vme_end) - entry->vme_start; in vm_map_reusable_pages()
16533 submap_end = submap_start + (entry->vme_end - in vm_map_pageout()
16635 new_entry->vme_end = end; in vm_map_entry_insert()
16847 if (src_end > src_entry->vme_end) { in vm_map_remap_extract()
16848 tmp_size -= (src_end - src_entry->vme_end); in vm_map_remap_extract()
16851 entry_size = (vm_map_size_t)(src_entry->vme_end - in vm_map_remap_extract()
16867 submap_size = src_entry->vme_end - src_start; in vm_map_remap_extract()
17093 copy_entry_size = copy_entry->vme_end - copy_entry->vme_start; in vm_map_remap_extract()
17095 copy_entry->vme_end = map_address + copy_entry_size; in vm_map_remap_extract()
17149 (uint64_t)src_entry->vme_end, in vm_map_remap_extract()
17163 (uint64_t)src_entry->vme_end, in vm_map_remap_extract()
17257 src_entry->vme_end, in vm_map_remap_extract()
17316 new_entry->vme_end = map_address + tmp_size; in vm_map_remap_extract()
17317 assert(new_entry->vme_start < new_entry->vme_end); in vm_map_remap_extract()
17377 (new_entry->vme_end - in vm_map_remap_extract()
17448 (new_entry->vme_end - in vm_map_remap_extract()
17465 (new_entry->vme_end - in vm_map_remap_extract()
17665 new_entry->vme_end = size; in vm_map_copy_to_physcopy()
17844 if (entry->vme_end <= trim_start) { in vm_map_copy_trim()
17856 copy_map->size -= entry->vme_end - entry->vme_start; in vm_map_copy_trim()
17996 object_offset_end += entry->vme_end - entry->vme_start; in vm_map_copy_adjust_to_target()
18085 …target_entry, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(targe… in vm_map_copy_adjust_to_target()
18088 …target_entry, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(targe… in vm_map_copy_adjust_to_target()
18109 target_entry->vme_end += overmap_start; in vm_map_copy_adjust_to_target()
18111 object_offset_end = VME_OFFSET(target_entry) + target_entry->vme_end - target_entry->vme_start; in vm_map_copy_adjust_to_target()
18113 …target_entry, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(targe… in vm_map_copy_adjust_to_target()
18120 target_entry->vme_end += overmap_end; in vm_map_copy_adjust_to_target()
18129 target_entry->vme_end -= addr_adjustment; in vm_map_copy_adjust_to_target()
18130 …target_entry, (uint64_t)target_entry->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(targe… in vm_map_copy_adjust_to_target()
18533 vm_map_offset_t, entry->vme_end, in vm_map_remap()
18557 entry->vme_end += *address; in vm_map_remap()
18566 assert(VM_MAP_PAGE_ALIGNED(entry->vme_end, MIN(target_page_mask, PAGE_MASK))); in vm_map_remap()
18911 start = entry->vme_end; in vm_map_check_protection()
18990 entry->vme_end - entry->vme_start != object->vo_size) { in vm_map_purgable_control()
19038 assert(curr_s_offset < map_entry->vme_end); in vm_map_footprint_query_page_info()
19308 curr_s_offset >= vm_map_last_entry(map)->vme_end) { in vm_map_page_range_info_internal()
19318 if (curr_s_offset - vm_map_last_entry(map)->vme_end <= in vm_map_page_range_info_internal()
19422 range_len = MIN(map_entry->vme_end, end) - curr_s_offset; in vm_map_page_range_info_internal()
19468 curr_e_offset = MIN(map_entry->vme_end, end); in vm_map_page_range_info_internal()
19528 curr_e_offset = MIN(map_entry->vme_end, end); in vm_map_page_range_info_internal()
19835 if (amount_left + entry->vme_start + offset > entry->vme_end) { in vm_map_msync()
19836 flush_size = entry->vme_end - in vm_map_msync()
20017 copy_entry->vme_end = VM_MAP_ROUND_PAGE(offset + size, PAGE_MASK); in vm_named_entry_associate_vm_object()
20054 …copy, copy_entry, (uint64_t)copy_entry->vme_start, (uint64_t)copy_entry->vme_end, copy_entry->vme_… in vm_named_entry_to_vm_object()
20279 if (map->holes_list->prev->vme_end == old_max_offset) { in vm_map_set_max_addr()
20283 map->holes_list->prev->vme_end = map->max_offset; in vm_map_set_max_addr()
20651 if ((entry->vme_start > start) || (entry->vme_end < end)) { in vm_map_sign()
20741 entry->vme_end, VM_MAP_REMOVE_NO_YIELD, in vm_map_partial_reap()
20778 vm_map_clip_unnest(map, entry, entry->vme_start, entry->vme_end); in vm_map_disconnect_page_mappings()
20799 pmap_remove_options(map->pmap, entry->vme_start, entry->vme_end, 0); in vm_map_disconnect_page_mappings()
21291 volatile_virtual_size += entry->vme_end - entry->vme_start; in vm_map_query_volatile()
21301 entry->vme_end, in vm_map_query_volatile()
21346 free = entry->vme_end - prev; in vm_map_sizes()
21359 prev = entry->vme_end; in vm_map_sizes()
21809 va < old_entry->vme_end; in vm_map_corpse_footprint_collect()