Lines Matching refs:VME_OFFSET

1137 			crypto_backing_offset = VME_OFFSET(&tmp_entry);  in vm_map_apple_protected()
1156 VME_OFFSET(&tmp_entry), in vm_map_apple_protected()
1214 VME_OFFSET(&tmp_entry), in vm_map_apple_protected()
3142 VME_OFFSET(entry) != tmp_offset || in vm_map_enter()
3318 VME_OFFSET(entry), in vm_map_enter()
4291 copy_offset = VME_OFFSET(copy_entry); in vm_map_enter_mem_object()
4354 copy_offset = VME_OFFSET(copy_entry); in vm_map_enter_mem_object()
5168 VME_OFFSET(entry))); in vm_map_clip_unnest()
5209 VME_OFFSET(entry)); in vm_map_clip_unnest()
5330 VME_OFFSET_SET(entry, VME_OFFSET(entry) + (start - entry->vme_start)); in _vm_map_clip_start()
5457 VME_OFFSET(new_entry) + (end - entry->vme_start)); in _vm_map_clip_end()
6523 sub_start = VME_OFFSET(entry); in vm_map_wire_nested()
6525 sub_end += VME_OFFSET(entry) - entry->vme_start; in vm_map_wire_nested()
6739 offset = VME_OFFSET(entry); in vm_map_wire_nested()
6873 orig_offset = VME_OFFSET(entry); in vm_map_wire_nested()
6914 …assertf(vm_object_round_page(VME_OFFSET(entry) + size) - vm_object_trunc_page(VME_OFFSET(entry)) =… in vm_map_wire_nested()
6920 (uint64_t)VME_OFFSET(entry), in vm_map_wire_nested()
7306 sub_start = VME_OFFSET(entry); in vm_map_unwire_nested()
7308 sub_end += VME_OFFSET(entry); in vm_map_unwire_nested()
7669 VME_OFFSET(entry)); in vm_map_submap_pmap_clean()
7676 (VME_OFFSET(entry) + in vm_map_submap_pmap_clean()
7707 VME_OFFSET(entry)); in vm_map_submap_pmap_clean()
7714 VME_OFFSET(entry), in vm_map_submap_pmap_clean()
8300 submap_start += VME_OFFSET(entry); in vm_map_delete()
8302 submap_end += VME_OFFSET(entry); in vm_map_delete()
8538 sub_start = VME_OFFSET(&tmp_entry); in vm_map_delete()
8709 VME_OFFSET(entry)); in vm_map_delete()
8715 VME_OFFSET(entry)); in vm_map_delete()
8725 VME_OBJECT(entry), VME_OFFSET(entry), in vm_map_delete()
9211 sub_start = VME_OFFSET(entry); in vm_map_overwrite_submap_recurse()
9219 sub_end += VME_OFFSET(entry); in vm_map_overwrite_submap_recurse()
9475 sub_start = VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
9483 sub_end += VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
9687 sub_start = VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
9695 sub_end += VME_OFFSET(entry); in vm_map_copy_overwrite_nested()
10479 entry_offset = VME_OFFSET(entry); in vm_map_copy_overwrite_unaligned()
10487 VME_OFFSET(copy_entry) + src_offset, in vm_map_copy_overwrite_unaligned()
10737 vm_object_offset_t old_offset = VME_OFFSET(entry); in vm_map_copy_overwrite_aligned()
10746 old_offset == VME_OFFSET(copy_entry)) { in vm_map_copy_overwrite_aligned()
10861 VME_OFFSET(entry)); in vm_map_copy_overwrite_aligned()
10868 VME_OFFSET(entry)); in vm_map_copy_overwrite_aligned()
10875 VME_OFFSET(entry), in vm_map_copy_overwrite_aligned()
10907 offset = VME_OFFSET(copy_entry); in vm_map_copy_overwrite_aligned()
10948 dst_offset = VME_OFFSET(entry); in vm_map_copy_overwrite_aligned()
10991 VME_OFFSET(copy_entry), in vm_map_copy_overwrite_aligned()
11610 offset = VME_OFFSET(entry); in vm_map_copyout_internal()
12121 src_start += VME_OFFSET(tmp_entry); in vm_map_copyin_internal()
12194 src_offset = VME_OFFSET(src_entry); in vm_map_copyin_internal()
12353 new_offset = VME_OFFSET(new_entry); in vm_map_copyin_internal()
12365 if (new_offset != VME_OFFSET(new_entry)) { in vm_map_copyin_internal()
12512 (VME_OFFSET(src_entry) != src_offset) || in vm_map_copyin_internal()
12668 original_offset = VME_OFFSET(tmp_entry); in vm_map_copyin_internal()
12677 VME_OFFSET(tmp_entry) - adjustment); in vm_map_copyin_internal()
12691 (VME_OFFSET(tmp_entry) + in vm_map_copyin_internal()
12702 assert(VME_OFFSET(tmp_entry) >= original_offset); in vm_map_copyin_internal()
13085 VME_OFFSET(old_entry), in vm_map_fork_share()
13609 VME_OFFSET(old_entry), in vm_map_fork()
13652 VME_OFFSET(old_entry), in vm_map_fork()
13980 top_entry_saved_offset = VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14134 submap_entry_offset = VME_OFFSET(submap_entry); in vm_map_lookup_and_lock_object()
14164 assertf(VM_MAP_PAGE_ALIGNED(VME_OFFSET(submap_entry), VM_MAP_PAGE_MASK(map)), in vm_map_lookup_and_lock_object()
14166 submap_entry, VME_OFFSET(submap_entry)); in vm_map_lookup_and_lock_object()
14238 submap_entry_offset = VME_OFFSET(submap_entry); in vm_map_lookup_and_lock_object()
14334 VME_OFFSET(submap_entry), in vm_map_lookup_and_lock_object()
14355 VME_OFFSET(submap_entry)); in vm_map_lookup_and_lock_object()
14697 *offset = (vaddr - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14713 fault_info->lo_offset = VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
14715 (entry->vme_end - entry->vme_start) + VME_OFFSET(entry); in vm_map_lookup_and_lock_object()
15075 (VME_OFFSET(curr_entry) - curr_entry->vme_start); in vm_map_region_recurse_64()
15206 submap_info->offset = VME_OFFSET(curr_entry); in vm_map_region_recurse_64()
15220 short_info->offset = VME_OFFSET(curr_entry); in vm_map_region_recurse_64()
15254 (VME_OFFSET(curr_entry) + in vm_map_region_recurse_64()
15368 basic->offset = (uint32_t)VME_OFFSET(entry); in vm_map_region()
15416 basic->offset = VME_OFFSET(entry); in vm_map_region()
15488 …vm_map_region_walk(map, start, entry, VME_OFFSET(entry), entry->vme_end - start, extended, TRUE, *… in vm_map_region()
16026 ((VME_OFFSET(prev_entry) + (prev_entry->vme_end - in vm_map_simplify_entry()
16028 == VME_OFFSET(this_entry)) && in vm_map_simplify_entry()
16078 VME_OFFSET_SET(this_entry, VME_OFFSET(prev_entry)); in vm_map_simplify_entry()
16236 + VME_OFFSET(entry); in vm_map_machine_attribute()
16253 + VME_OFFSET(entry); in vm_map_machine_attribute()
16482 offset = (start - entry->vme_start) + VME_OFFSET(entry); in vm_map_willneed()
16779 start_offset += VME_OFFSET(entry); in vm_map_reuse_pages()
16780 end_offset += VME_OFFSET(entry); in vm_map_reuse_pages()
16891 start_offset += VME_OFFSET(entry); in vm_map_reusable_pages()
16892 end_offset += VME_OFFSET(entry); in vm_map_reusable_pages()
17069 submap_start = VME_OFFSET(entry); in vm_map_pageout()
17163 *start_offset += VME_OFFSET(entry); in vm_map_get_bounds_in_object()
17164 *end_offset += VME_OFFSET(entry); in vm_map_get_bounds_in_object()
17573 submap_start = VME_OFFSET(src_entry) + src_start - src_entry->vme_start; in vm_map_remap_extract()
17653 copy_offset = VME_OFFSET(copy_entry); in vm_map_remap_extract()
17694 copy_offset = VME_OFFSET(copy_entry); in vm_map_remap_extract()
17752 submap_start = VME_OFFSET(src_entry) + src_start - src_entry->vme_start; in vm_map_remap_extract()
18034 VME_OFFSET(src_entry), in vm_map_remap_extract()
18089 offset = (VME_OFFSET(src_entry) + in vm_map_remap_extract()
18215 VME_OFFSET(new_entry), in vm_map_remap_extract()
18314 new_offset = VME_OFFSET(new_entry); in vm_map_remap_extract()
18328 if (new_offset != VME_OFFSET(new_entry)) { in vm_map_remap_extract()
18863 object_offset_start = VME_OFFSET(entry); in vm_map_copy_adjust_to_target()
18954 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
18955 object_offset_start = VME_OFFSET(target_entry); in vm_map_copy_adjust_to_target()
18957 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
18964 …try %p offset 0x%llx copy %d -> overmap_start 0x%llx\n", target_entry, VME_OFFSET(target_entry), c… in vm_map_copy_adjust_to_target()
18965 VME_OFFSET_SET(target_entry, VME_OFFSET(target_entry) - overmap_start); in vm_map_copy_adjust_to_target()
18968 …DEBUG4K_ADJUST("entry %p offset 0x%llx copy %d -> misalignments %d\n", target_entry, VME_OFFSET(ta… in vm_map_copy_adjust_to_target()
18980 object_offset_end = VME_OFFSET(target_entry) + target_entry->vme_end - target_entry->vme_start; in vm_map_copy_adjust_to_target()
18982 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
18988 …entry %p offset 0x%llx copy %d -> overmap_end 0x%llx\n", target_entry, VME_OFFSET(target_entry), c… in vm_map_copy_adjust_to_target()
18993 …DEBUG4K_ADJUST("entry %p offset 0x%llx copy %d -> misalignments %d\n", target_entry, VME_OFFSET(ta… in vm_map_copy_adjust_to_target()
18999 …y->vme_start, (uint64_t)target_entry->vme_end, VME_OBJECT(target_entry), VME_OFFSET(target_entry)); in vm_map_copy_adjust_to_target()
19493 assert(VM_MAP_PAGE_ALIGNED(VME_OFFSET(entry), MIN(target_page_mask, PAGE_MASK))); in vm_map_remap()
19908 if (VME_OFFSET(entry) != 0 || in vm_map_purgable_control()
20002 + VME_OFFSET(map_entry)) in vm_map_footprint_query_page_info()
20026 + VME_OFFSET(map_entry)) in vm_map_footprint_query_page_info()
20056 + VME_OFFSET(map_entry)) in vm_map_footprint_query_page_info()
20377 offset_in_object += VME_OFFSET(map_entry); in vm_map_page_range_info_internal()
20824 local_offset = VME_OFFSET(entry); in vm_map_msync()
20847 offset += VME_OFFSET(entry); in vm_map_msync()
21762 start - entry->vme_start + VME_OFFSET(entry)); in vm_map_sign()
22000 out_entry->vmei_offset = VME_OFFSET(entry); in vm_map_dump_entry_and_compressor_pager()
22526 if (VME_OFFSET(entry)) { in vm_map_query_volatile()
22537 if ((VME_OFFSET(entry) / PAGE_SIZE) >= resident_count) { in vm_map_query_volatile()
22540 resident_count -= (VME_OFFSET(entry) / PAGE_SIZE); in vm_map_query_volatile()
22802 cs_offset = VME_OFFSET(entry); in vm_map_entry_cs_associate()
23745 map_offset = (VME_OFFSET(entry) + in vm_map_get_phys_page()
23768 offset = (VME_OFFSET(entry) + in vm_map_get_phys_page()
23775 offset = (VME_OFFSET(entry) + (map_offset - entry->vme_start)); in vm_map_get_phys_page()