Lines Matching refs:VME_OFFSET

952 			crypto_backing_offset = VME_OFFSET(&tmp_entry);  in vm_map_apple_protected()
971 VME_OFFSET(&tmp_entry), in vm_map_apple_protected()
1038 VME_OFFSET(&tmp_entry), in vm_map_apple_protected()
2778 VME_OFFSET(entry) != tmp_offset || in vm_map_enter()
2937 VME_OFFSET(entry), in vm_map_enter()
3784 VME_OFFSET(entry), in vm_map_enter_fourk()
3799 VME_OFFSET(entry), in vm_map_enter_fourk()
4316 copy_offset = VME_OFFSET(copy_entry); in vm_map_enter_mem_object_helper()
4380 copy_offset = VME_OFFSET(copy_entry); in vm_map_enter_mem_object_helper()
5422 VME_OFFSET(entry))); in vm_map_clip_unnest()
5463 VME_OFFSET(entry)); in vm_map_clip_unnest()
5584 VME_OFFSET_SET(entry, VME_OFFSET(entry) + (start - entry->vme_start)); in _vm_map_clip_start()
5699 VME_OFFSET(new_entry) + (end - entry->vme_start)); in _vm_map_clip_end()
6552 sub_start = VME_OFFSET(entry); in vm_map_wire_nested()
6554 sub_end += VME_OFFSET(entry) - entry->vme_start; in vm_map_wire_nested()
6768 offset = VME_OFFSET(entry); in vm_map_wire_nested()
6898 orig_offset = VME_OFFSET(entry); in vm_map_wire_nested()
6939 …assertf(vm_object_round_page(VME_OFFSET(entry) + size) - vm_object_trunc_page(VME_OFFSET(entry)) =… in vm_map_wire_nested()
6945 (uint64_t)VME_OFFSET(entry), in vm_map_wire_nested()
7311 sub_start = VME_OFFSET(entry); in vm_map_unwire_nested()
7313 sub_end += VME_OFFSET(entry); in vm_map_unwire_nested()
7628 VME_OFFSET(entry)); in vm_map_submap_pmap_clean()
7635 (VME_OFFSET(entry) + in vm_map_submap_pmap_clean()
7666 VME_OFFSET(entry)); in vm_map_submap_pmap_clean()
7673 VME_OFFSET(entry), in vm_map_submap_pmap_clean()
8181 submap_start += VME_OFFSET(entry); in vm_map_delete()
8183 submap_end += VME_OFFSET(entry); in vm_map_delete()
8418 sub_start = VME_OFFSET(&tmp_entry); in vm_map_delete()
8560 VME_OFFSET(entry)); in vm_map_delete()
8566 VME_OFFSET(entry)); in vm_map_delete()
8576 VME_OBJECT(entry), VME_OFFSET(entry), in vm_map_delete()
9019 sub_start = VME_OFFSET(entry); in vm_map_overwrite_submap_recurse()
9027 sub_end += VME_OFFSET(entry); in vm_map_overwrite_submap_recurse()
9281 sub_start = VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
9289 sub_end += VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
9491 sub_start = VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
9499 sub_end += VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
10225 entry_offset = VME_OFFSET(entry); in vm_map_copy_overwrite_unaligned()
10233 VME_OFFSET(copy_entry) + src_offset, in vm_map_copy_overwrite_unaligned()
10445 vm_object_offset_t old_offset = VME_OFFSET(entry); in vm_map_copy_overwrite_aligned()
10453 old_offset == VME_OFFSET(copy_entry)) { in vm_map_copy_overwrite_aligned()
10568 VME_OFFSET(entry)); in vm_map_copy_overwrite_aligned()
10575 VME_OFFSET(entry)); in vm_map_copy_overwrite_aligned()
10582 VME_OFFSET(entry), in vm_map_copy_overwrite_aligned()
10614 offset = VME_OFFSET(copy_entry); in vm_map_copy_overwrite_aligned()
10655 dst_offset = VME_OFFSET(entry); in vm_map_copy_overwrite_aligned()
10698 VME_OFFSET(copy_entry), in vm_map_copy_overwrite_aligned()
11388 offset = VME_OFFSET(entry); in vm_map_copyout_internal()
11809 src_start += VME_OFFSET(tmp_entry); in vm_map_copyin_internal()
11882 src_offset = VME_OFFSET(src_entry); in vm_map_copyin_internal()
12039 new_offset = VME_OFFSET(new_entry); in vm_map_copyin_internal()
12050 if (new_offset != VME_OFFSET(new_entry)) { in vm_map_copyin_internal()
12197 (VME_OFFSET(src_entry) != src_offset) || in vm_map_copyin_internal()
12353 original_offset = VME_OFFSET(tmp_entry); in vm_map_copyin_internal()
12362 VME_OFFSET(tmp_entry) - adjustment); in vm_map_copyin_internal()
12376 (VME_OFFSET(tmp_entry) + in vm_map_copyin_internal()
12387 assert(VME_OFFSET(tmp_entry) >= original_offset); in vm_map_copyin_internal()
12773 VME_OFFSET(old_entry), in vm_map_fork_share()
13256 VME_OFFSET(old_entry), in vm_map_fork()
13285 VME_OFFSET(old_entry), in vm_map_fork()
13659 local_vaddr = (local_vaddr - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
13687 start_delta = submap_entry->vme_start > VME_OFFSET(entry) ? in vm_map_lookup_and_lock_object()
13688 submap_entry->vme_start - VME_OFFSET(entry) : 0; in vm_map_lookup_and_lock_object()
13691 (VME_OFFSET(entry) + start_delta + (old_end - old_start)) <= in vm_map_lookup_and_lock_object()
13693 0 : (VME_OFFSET(entry) + in vm_map_lookup_and_lock_object()
13753 submap_entry_offset = VME_OFFSET(submap_entry); in vm_map_lookup_and_lock_object()
13783 assertf(VM_MAP_PAGE_ALIGNED(VME_OFFSET(submap_entry), VM_MAP_PAGE_MASK(map)), in vm_map_lookup_and_lock_object()
13785 submap_entry, VME_OFFSET(submap_entry)); in vm_map_lookup_and_lock_object()
13857 submap_entry_offset = VME_OFFSET(submap_entry); in vm_map_lookup_and_lock_object()
13938 VME_OFFSET(submap_entry), in vm_map_lookup_and_lock_object()
13959 VME_OFFSET(submap_entry)); in vm_map_lookup_and_lock_object()
14268 *offset = (vaddr - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14284 fault_info->lo_offset = VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14286 (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14636 (VME_OFFSET(curr_entry) - curr_entry->vme_start); in vm_map_region_recurse_64()
14767 submap_info->offset = VME_OFFSET(curr_entry); in vm_map_region_recurse_64()
14781 short_info->offset = VME_OFFSET(curr_entry); in vm_map_region_recurse_64()
14815 (VME_OFFSET(curr_entry) + in vm_map_region_recurse_64()
14933 basic->offset = (uint32_t)VME_OFFSET(entry); in vm_map_region()
14981 basic->offset = VME_OFFSET(entry); in vm_map_region()
15053 …vm_map_region_walk(map, start, entry, VME_OFFSET(entry), entry->vme_end - start, extended, TRUE, *… in vm_map_region()
15557 ((VME_OFFSET(prev_entry) + (prev_entry->vme_end - in vm_map_simplify_entry()
15559 == VME_OFFSET(this_entry)) && in vm_map_simplify_entry()
15604 VME_OFFSET_SET(this_entry, VME_OFFSET(prev_entry)); in vm_map_simplify_entry()
15738 + VME_OFFSET(entry); in vm_map_machine_attribute()
15755 + VME_OFFSET(entry); in vm_map_machine_attribute()
15974 offset = (start - entry->vme_start) + VME_OFFSET(entry); in vm_map_willneed()
16263 start_offset += VME_OFFSET(entry); in vm_map_reuse_pages()
16264 end_offset += VME_OFFSET(entry); in vm_map_reuse_pages()
16370 start_offset += VME_OFFSET(entry); in vm_map_reusable_pages()
16371 end_offset += VME_OFFSET(entry); in vm_map_reusable_pages()
16532 submap_start = VME_OFFSET(entry); in vm_map_pageout()
16871 submap_start = VME_OFFSET(src_entry) + src_start - src_entry->vme_start; in vm_map_remap_extract()
16980 submap_start = VME_OFFSET(src_entry) + src_start - src_entry->vme_start; in vm_map_remap_extract()
17236 VME_OFFSET(src_entry), in vm_map_remap_extract()
17290 offset = (VME_OFFSET(src_entry) + in vm_map_remap_extract()
17376 VME_OFFSET(new_entry), in vm_map_remap_extract()
17461 new_offset = VME_OFFSET(new_entry); in vm_map_remap_extract()
17474 if (new_offset != VME_OFFSET(new_entry)) { in vm_map_remap_extract()
17994 object_offset_start = VME_OFFSET(entry); in vm_map_copy_adjust_to_target()
18085 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
18086 object_offset_start = VME_OFFSET(target_entry); in vm_map_copy_adjust_to_target()
18088 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
18095 …try %p offset 0x%llx copy %d -> overmap_start 0x%llx\n", target_entry, VME_OFFSET(target_entry), c… in vm_map_copy_adjust_to_target()
18096 VME_OFFSET_SET(target_entry, VME_OFFSET(target_entry) - overmap_start); in vm_map_copy_adjust_to_target()
18099 …DEBUG4K_ADJUST("entry %p offset 0x%llx copy %d -> misalignments %d\n", target_entry, VME_OFFSET(ta… in vm_map_copy_adjust_to_target()
18111 object_offset_end = VME_OFFSET(target_entry) + target_entry->vme_end - target_entry->vme_start; in vm_map_copy_adjust_to_target()
18113 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
18119 …entry %p offset 0x%llx copy %d -> overmap_end 0x%llx\n", target_entry, VME_OFFSET(target_entry), c… in vm_map_copy_adjust_to_target()
18124 …DEBUG4K_ADJUST("entry %p offset 0x%llx copy %d -> misalignments %d\n", target_entry, VME_OFFSET(ta… in vm_map_copy_adjust_to_target()
18130 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
18567 assert(VM_MAP_PAGE_ALIGNED(VME_OFFSET(entry), MIN(target_page_mask, PAGE_MASK))); in vm_map_remap()
18989 if (VME_OFFSET(entry) != 0 || in vm_map_purgable_control()
19083 + VME_OFFSET(map_entry)) in vm_map_footprint_query_page_info()
19110 + VME_OFFSET(map_entry)) in vm_map_footprint_query_page_info()
19415 offset_in_object += VME_OFFSET(map_entry); in vm_map_page_range_info_internal()
19849 local_offset = VME_OFFSET(entry); in vm_map_msync()
19872 offset += VME_OFFSET(entry); in vm_map_msync()
20676 start - entry->vme_start + VME_OFFSET(entry)); in vm_map_sign()
21274 if (VME_OFFSET(entry)) { in vm_map_query_volatile()
21285 if ((VME_OFFSET(entry) / PAGE_SIZE) >= resident_count) { in vm_map_query_volatile()
21288 resident_count -= (VME_OFFSET(entry) / PAGE_SIZE); in vm_map_query_volatile()