| /xnu-8019.80.24/osfmk/vm/ |
| H A D | vm_map_store_ll.c | 37 while (vm_map_trunc_page(next->vme_start, in first_free_is_valid_ll() 39 vm_map_trunc_page(entry->vme_end, in first_free_is_valid_ll() 41 (vm_map_trunc_page(next->vme_start, in first_free_is_valid_ll() 43 vm_map_trunc_page(entry->vme_start, in first_free_is_valid_ll() 76 while (vm_map_trunc_page(UFF_next_entry->vme_start, \ 78 vm_map_trunc_page(UFF_first_free->vme_end, \ 80 (vm_map_trunc_page(UFF_next_entry->vme_start, \ 82 vm_map_trunc_page(UFF_first_free->vme_start, \
|
| H A D | vm_debug.c | 284 vm_map_trunc_page(addr, in vm32_region_info() 309 vm_map_trunc_page(addr, in vm32_region_info() 507 vm_map_trunc_page(addr, in vm32_region_info_64() 532 vm_map_trunc_page(addr, in vm32_region_info_64() 589 vm_map_trunc_page(addr, in vm32_mapped_pages_info() 628 vm_map_trunc_page(addr, in vm32_mapped_pages_info()
|
| H A D | vm_user.c | 214 map_addr = vm_map_trunc_page(*addr, in mach_vm_allocate_kernel() 303 map_addr = vm_map_trunc_page(*addr, in vm_allocate_kernel() 357 vm_map_trunc_page(start, in mach_vm_deallocate() 385 vm_map_trunc_page(start, in vm_deallocate() 414 vm_map_trunc_page(start, in mach_vm_inherit() 443 vm_map_trunc_page(start, in vm_inherit() 474 vm_map_trunc_page(start, in mach_vm_protect() 511 vm_map_trunc_page(start, in vm_protect() 542 vm_map_trunc_page(addr, in mach_vm_machine_attribute() 574 vm_map_trunc_page(addr, in vm_machine_attribute() [all …]
|
| H A D | vm_kern.c | 177 vm_map_trunc_page(map_addr, in kmem_alloc_contig() 198 vm_map_trunc_page(map_addr, in kmem_alloc_contig() 212 vm_map_trunc_page(map_addr, in kmem_alloc_contig() 945 oldmapmin = vm_map_trunc_page(oldaddr, in kmem_realloc() 1187 vm_map_trunc_page(addr, in kmem_free() 1281 : vm_map_trunc_page(*addr, in kmem_suballoc() 1405 map_start = vm_map_trunc_page(start, in kmem_init()
|
| H A D | vm32_user.c | 496 vm_map_trunc_page(address, PAGE_MASK), in vm32_purgable_control() 514 vm_map_trunc_page(offset, PAGE_MASK), in vm32_map_page_query()
|
| H A D | vm_map.c | 817 start_aligned = vm_map_trunc_page(start, PAGE_MASK_64); in vm_map_apple_protected() 819 start_aligned = vm_map_trunc_page(start_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected() 2088 random_addr = vm_map_trunc_page( in vm_map_random_address_for_size() 3555 start = vm_map_trunc_page(*address, VM_MAP_PAGE_MASK(map)); in vm_map_enter_fourk() 3706 vm_map_trunc_page(start, in vm_map_enter_fourk() 4110 map_addr = vm_map_trunc_page(*address, FOURK_PAGE_MASK); in vm_map_enter_mem_object_helper() 4115 map_addr = vm_map_trunc_page(*address, in vm_map_enter_mem_object_helper() 4956 map_addr = vm_map_trunc_page(*address, in vm_map_enter_mem_object_control() 4963 map_addr = vm_map_trunc_page(*address, in vm_map_enter_mem_object_control() 5152 *addr = vm_map_trunc_page(*addr, in vm_map_enter_cpm() [all …]
|
| H A D | vm_map_store_rb.c | 671 if (vm_map_trunc_page(entry->vme_start, VM_MAP_PAGE_MASK(map)) >= max_valid_offset) { in update_first_free_rb()
|
| H A D | vm_fourk_pager.c | 957 (vm_map_trunc_page(offset, in fourk_pager_data_request()
|
| H A D | vm_fault.c | 4098 vaddr = vm_map_trunc_page(vaddr, fault_page_mask); in vm_fault_internal() 4105 vaddr = vm_map_trunc_page(vaddr, PAGE_MASK); in vm_fault_internal() 4366 …fault_phys_offset = (vm_map_offset_t)offset - vm_map_trunc_page((vm_map_offset_t)offset, PAGE_MASK… in vm_fault_internal() 7862 if (copyin((user_addr_t)vm_map_trunc_page(code_addr, size - 1), new_code_buffer, size) != 0) { in vmtc_text_page_diagnose()
|
| H A D | vm_map.h | 1704 #define vm_map_trunc_page(x, pgmask) ((vm_map_offset_t)(x) & ~((signed)(pgmask))) macro
|
| /xnu-8019.80.24/san/memory/ |
| H A D | kasan-arm64.c | 153 vm_offset_t addr_aligned = vm_map_trunc_page(*addrp, ARM_PGMASK); in kasan_arm64_align_to_page() 250 vm_offset_t shadow_base = vm_map_trunc_page(SHADOW_FOR_ADDRESS(address), ARM_PGMASK); in kasan_map_shadow_internal() 340 tosteal = vm_map_trunc_page(tosteal, ARM_PGMASK); in kasan_bootstrap()
|
| H A D | kasan-x86_64.c | 127 address = vm_map_trunc_page(address, I386_LPGMASK); in kasan_map_shadow_superpage_zero() 184 vm_offset_t shadow_base = vm_map_trunc_page(SHADOW_FOR_ADDRESS(address), PAGE_MASK); in kasan_map_shadow()
|
| H A D | kasan.c | 428 shadow_base = vm_map_trunc_page(shadow_base, PAGE_MASK); in kasan_traverse_mappings()
|
| /xnu-8019.80.24/osfmk/kern/ |
| H A D | kext_alloc.c | 82 text_start = vm_map_trunc_page(text->vmaddr, in kext_alloc_init() 157 addr = vm_map_trunc_page(addr, in get_address_from_kext_map()
|
| H A D | arcade.c | 318 vm_map_trunc_page((vm_offset_t)path, VM_MAP_PAGE_MASK(ipc_kernel_map)), in arcade_ast()
|
| H A D | stack.c | 349 stack = (vm_offset_t)vm_map_trunc_page( in stack_collect()
|
| H A D | host.c | 1180 result = vm_map_unwire(ipc_kernel_map, vm_map_trunc_page(addr, VM_MAP_PAGE_MASK(ipc_kernel_map)), in host_processor_info()
|
| /xnu-8019.80.24/bsd/kern/ |
| H A D | mach_loader.c | 1891 vm_map_trunc_page(vm_start + (file_end - file_start), in map_segment() 1925 cur_end = vm_map_trunc_page(vm_start + (file_end - in map_segment() 2230 vm_end_aligned = vm_map_trunc_page(vm_end, in load_segment() 2283 file_start = vm_map_trunc_page(file_offset, in load_segment() 2287 vm_start = vm_map_trunc_page(vm_offset, in load_segment() 2317 file_start = vm_map_trunc_page(file_offset, in load_segment() 2321 vm_start = vm_map_trunc_page(vm_offset, in load_segment()
|
| H A D | kern_mman.c | 624 if (user_addr != vm_map_trunc_page(user_addr, in mmap() 1059 vm_map_trunc_page(user_addr, in mprotect() 1256 first_addr = addr = vm_map_trunc_page(uap->addr, in mincore()
|
| H A D | kern_exec.c | 6319 addr = vm_map_trunc_page(load_result->user_stack - size, in create_unix_stack() 6348 prot_size = vm_map_trunc_page(size - load_result->user_stack_size, vm_map_page_mask(map)); in create_unix_stack() 7146 vm_map_trunc_page(load_result->entry_point, in exec_prefault_data() 7152 …DEBUG4K_ERROR("map %p va 0x%llx -> 0x%x\n", current_map(), (uint64_t)vm_map_trunc_page(load_result… in exec_prefault_data() 7178 vm_map_trunc_page(load_result->all_image_info_addr, in exec_prefault_data() 7189 vm_map_trunc_page(load_result->all_image_info_addr + expected_all_image_infos_size - 1, in exec_prefault_data() 7253 vm_map_trunc_page(notification_address + dyld_slide_amount, in exec_prefault_data() 7265 vm_map_trunc_page(dyld_image_address + dyld_slide_amount, in exec_prefault_data() 7277 vm_map_trunc_page(dyld_version_address + dyld_slide_amount, in exec_prefault_data() 7289 vm_map_trunc_page(dyld_all_image_infos_address + dyld_slide_amount, in exec_prefault_data()
|
| /xnu-8019.80.24/osfmk/ipc/ |
| H A D | mach_port.c | 288 vm_map_trunc_page(addr1, in mach_port_names() 302 vm_map_trunc_page(addr2, in mach_port_names() 365 vm_map_trunc_page(addr1, in mach_port_names() 374 vm_map_trunc_page(addr2, in mach_port_names() 1374 vm_map_trunc_page(addr, in mach_port_get_set_status()
|
| H A D | mach_debug.c | 302 vm_map_trunc_page(table_addr, in mach_port_space_info()
|
| /xnu-8019.80.24/osfmk/i386/commpage/ |
| H A D | commpage.c | 168 …if (!(kr = vm_map_lookup_entry( kernel_map, vm_map_trunc_page(kernel_addr, VM_MAP_PAGE_MASK(kernel… in commpage_allocate()
|
| /xnu-8019.80.24/bsd/vm/ |
| H A D | vm_unix.c | 475 vm_map_trunc_page(addr, in useracc() 492 vm_map_trunc_page(addr, in vslock() 532 for (vaddr = vm_map_trunc_page(addr, PAGE_MASK); in vsunlock() 545 vm_map_trunc_page(addr, in vsunlock()
|
| /xnu-8019.80.24/bsd/vfs/ |
| H A D | vfs_cluster.c | 3036 vm_map_trunc_page(iov_base, vm_map_page_mask(map)), in cluster_write_contig() 5250 vm_pre_fault(vm_map_trunc_page(curr_iov_base, current_page_mask), VM_PROT_READ); in cluster_read_direct() 5366 vm_map_trunc_page(iov_base, vm_map_page_mask(map)), in cluster_read_contig() 5545 vm_map_trunc_page(iov_base, vm_map_page_mask(map)), in cluster_io_type()
|