| /xnu-11215.1.10/osfmk/vm/ |
| H A D | vm_map_store_ll.c | 39 while (vm_map_trunc_page(next->vme_start, map_page_mask) == in first_free_is_valid_ll() 40 vm_map_trunc_page(entry->vme_end, map_page_mask) || in first_free_is_valid_ll() 41 (vm_map_trunc_page(next->vme_start, map_page_mask) == in first_free_is_valid_ll() 42 vm_map_trunc_page(entry->vme_start, map_page_mask) && in first_free_is_valid_ll() 120 while (vm_map_trunc_page(next->vme_start, map_page_mask) == in update_first_free_ll() 121 vm_map_trunc_page(new_first_free->vme_end, map_page_mask) || in update_first_free_ll() 122 (vm_map_trunc_page(next->vme_start, map_page_mask) == in update_first_free_ll() 123 vm_map_trunc_page(new_first_free->vme_start, map_page_mask) && in update_first_free_ll()
|
| H A D | vm_user.c | 264 vm_map_trunc_page(start, in mach_vm_inherit() 293 vm_map_trunc_page(start, in vm_inherit() 324 vm_map_trunc_page(start, in mach_vm_protect() 361 vm_map_trunc_page(start, in vm_protect() 392 vm_map_trunc_page(addr, in mach_vm_machine_attribute() 424 vm_map_trunc_page(addr, in vm_machine_attribute() 1297 vm_map_trunc_page(start, align_mask), in mach_vm_behavior_set() 1624 vm_map_trunc_page(address, VM_MAP_PAGE_MASK(map)), in mach_vm_purgable_control() 1668 vm_map_trunc_page(address, VM_MAP_PAGE_MASK(map)), in vm_purgable_control() 1709 vm_map_trunc_page(offset, PAGE_MASK), in mach_vm_page_query() [all …]
|
| H A D | vm_map.h | 262 #define vm_map_trunc_page(offset, pgmask) \ macro
|
| H A D | vm_map_store.c | 368 end = vm_map_trunc_page(end, map_mask); in vm_map_store_find_space_backwards() 370 start = vm_map_trunc_page(start, map_mask); in vm_map_store_find_space_backwards()
|
| H A D | vm32_user.c | 541 vm_map_trunc_page(address, PAGE_MASK), in vm32_purgable_control() 559 vm_map_trunc_page(offset, PAGE_MASK), in vm32_map_page_query()
|
| H A D | vm_debug.c | 566 vm_map_trunc_page(addr, in vm32_mapped_pages_info() 605 vm_map_trunc_page(addr, in vm32_mapped_pages_info()
|
| H A D | vm_upl.c | 88 aligned_offset_to_map = vm_map_trunc_page(offset_to_map, vm_map_page_mask(map)); in vm_upl_map_range() 111 aligned_offset_to_unmap = vm_map_trunc_page(offset_to_unmap, vm_map_page_mask(map)); in vm_upl_unmap_range()
|
| H A D | vm_map.c | 1034 start_aligned = vm_map_trunc_page(start, PAGE_MASK_64); in vm_map_apple_protected() 1036 start_aligned = vm_map_trunc_page(start_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected() 2695 random_addr = vm_map_trunc_page( in vm_map_random_address_for_size() 8973 vm_map_trunc_page(dst_addr, in vm_map_overwrite_submap_recurse() 9235 vm_map_trunc_page(dst_addr, in vm_map_copy_overwrite_nested() 9743 vm_map_trunc_page(base_addr, in vm_map_copy_overwrite_nested() 10884 vm_map_trunc_page(src_addr, VM_MAP_PAGE_MASK(src_map)), in vm_map_copyin_kernel_buffer() 11004 vm_map_trunc_page(*addr, in vm_map_copyout_kernel_buffer() 11247 vm_copy_start = vm_map_trunc_page((vm_map_size_t)copy->offset, in vm_map_copyout_internal() 11664 *src_start = vm_map_trunc_page(*src_addr_unaligned, in vm_map_copyin_sanitize() [all …]
|
| H A D | vm_map_store_rb.c | 652 if (vm_map_trunc_page(entry->vme_start, VM_MAP_PAGE_MASK(map)) >= max_valid_offset) { in update_first_free_rb()
|
| H A D | vm_reclaim.c | 958 vm_map_trunc_page(entry->address, in reclaim_chunk() 977 vm_map_trunc_page(entry->address, VM_MAP_PAGE_MASK(map)), in reclaim_chunk()
|
| H A D | vm_fault.c | 4301 vaddr = vm_map_trunc_page(vaddr, fault_page_mask); in vm_fault_internal() 4308 vaddr = vm_map_trunc_page(vaddr, PAGE_MASK); in vm_fault_internal() 4579 …fault_phys_offset = (vm_map_offset_t)offset - vm_map_trunc_page((vm_map_offset_t)offset, PAGE_MASK… in vm_fault_internal() 8274 if (copyin((user_addr_t)vm_map_trunc_page(code_addr, size - 1), new_code_buffer, size) != 0) { in vmtc_text_page_diagnose()
|
| H A D | vm_pageout.c | 6804 offset = vm_map_trunc_page(original_offset, VM_MAP_PAGE_MASK(map)); 6962 vm_map_trunc_page(offset, 7612 vm_map_trunc_page(addr, VM_MAP_PAGE_MASK(map)),
|
| /xnu-11215.1.10/san/memory/ |
| H A D | kasan-arm64.c | 162 vm_offset_t addr_aligned = vm_map_trunc_page(*addrp, ARM_PGMASK); in kasan_arm64_align_to_page() 262 vm_offset_t shadow_base = vm_map_trunc_page(SHADOW_FOR_ADDRESS(address), ARM_PGMASK); in kasan_map_shadow_internal() 356 tosteal = vm_map_trunc_page(tosteal, ARM_PGMASK); in kasan_bootstrap()
|
| H A D | kasan-x86_64.c | 131 address = vm_map_trunc_page(address, I386_LPGMASK); in kasan_map_shadow_superpage_zero() 188 vm_offset_t shadow_base = vm_map_trunc_page(SHADOW_FOR_ADDRESS(address), PAGE_MASK); in kasan_map_shadow()
|
| H A D | kasan.c | 397 shadow_base = vm_map_trunc_page(shadow_base, PAGE_MASK); in kasan_traverse_mappings()
|
| /xnu-11215.1.10/osfmk/kern/ |
| H A D | kext_alloc.c | 83 text_start = vm_map_trunc_page(text->vmaddr, in kext_alloc_init() 155 addr = vm_map_trunc_page(addr, in get_address_from_kext_map()
|
| H A D | stack.c | 338 stack = (vm_offset_t)vm_map_trunc_page( in stack_collect()
|
| H A D | host.c | 1216 result = vm_map_unwire(ipc_kernel_map, vm_map_trunc_page(addr, VM_MAP_PAGE_MASK(ipc_kernel_map)), in host_processor_info()
|
| /xnu-11215.1.10/bsd/kern/ |
| H A D | kern_mman.c | 554 vm_map_trunc_page(file_pos, user_mask), in mmap() 932 control, vm_map_trunc_page(file_pos, user_mask), in mmap() 1193 vm_map_trunc_page(user_addr, in mprotect() 1392 first_addr = addr = vm_map_trunc_page(uap->addr, in mincore()
|
| H A D | kern_exec.c | 7001 addr = vm_map_trunc_page(load_result->user_stack - size, in create_unix_stack() 7030 prot_size = vm_map_trunc_page(size - load_result->user_stack_size, vm_map_page_mask(map)); in create_unix_stack() 7951 vm_map_trunc_page(load_result->entry_point, in exec_prefault_data() 7957 …DEBUG4K_ERROR("map %p va 0x%llx -> 0x%x\n", current_map(), (uint64_t)vm_map_trunc_page(load_result… in exec_prefault_data() 7984 vm_map_trunc_page(load_result->all_image_info_addr, in exec_prefault_data() 7995 vm_map_trunc_page(load_result->all_image_info_addr + expected_all_image_infos_size - 1, in exec_prefault_data() 8059 vm_map_trunc_page(notification_address + dyld_slide_amount, in exec_prefault_data() 8071 vm_map_trunc_page(dyld_image_address + dyld_slide_amount, in exec_prefault_data() 8083 vm_map_trunc_page(dyld_version_address + dyld_slide_amount, in exec_prefault_data() 8095 vm_map_trunc_page(dyld_all_image_infos_address + dyld_slide_amount, in exec_prefault_data()
|
| H A D | mach_loader.c | 2121 vm_end_aligned = vm_map_trunc_page(vm_end, in load_segment() 2150 file_start = vm_map_trunc_page(file_offset, in load_segment() 2154 vm_start = vm_map_trunc_page(vm_offset, in load_segment() 2184 file_start = vm_map_trunc_page(file_offset, in load_segment() 2188 vm_start = vm_map_trunc_page(vm_offset, in load_segment()
|
| /xnu-11215.1.10/osfmk/i386/commpage/ |
| H A D | commpage.c | 167 …if (!(kr = vm_map_lookup_entry( kernel_map, vm_map_trunc_page(kernel_addr, VM_MAP_PAGE_MASK(kernel… in commpage_allocate()
|
| /xnu-11215.1.10/bsd/vfs/ |
| H A D | vfs_cluster.c | 3397 vm_map_trunc_page(iov_base, vm_map_page_mask(map)), in cluster_write_contig() 5649 vm_pre_fault(vm_map_trunc_page(curr_iov_base, current_page_mask), VM_PROT_READ); in cluster_read_direct() 5765 vm_map_trunc_page(iov_base, vm_map_page_mask(map)), in cluster_read_contig() 5944 vm_map_trunc_page(iov_base, vm_map_page_mask(map)), in cluster_io_type()
|
| /xnu-11215.1.10/bsd/vm/ |
| H A D | vm_unix.c | 558 vm_map_trunc_page(addr, in useracc() 657 for (vaddr = vm_map_trunc_page(addr, PAGE_MASK); in vsunlock()
|
| /xnu-11215.1.10/osfmk/tests/ |
| H A D | vm_parameter_validation_kern.c | 1417 mach_vm_address_t start = vm_map_trunc_page(*address, VM_MAP_PAGE_MASK(target_map)); in vm_map_enter_mem_object_control_wrapped()
|