| /xnu-12377.41.6/osfmk/vm/ |
| H A D | vm_debug.c | 283 size = vm_map_round_page(2 * used * sizeof(vm_info_object_t), in vm32_mach_vm_region_info() 304 vm_size_t vmsize_used = vm_map_round_page(size_used, in vm32_mach_vm_region_info() 504 size = vm_map_round_page(2 * used * sizeof(vm_info_object_t), in vm32_mach_vm_region_info_64() 525 vm_size_t vmsize_used = vm_map_round_page(size_used, in vm32_mach_vm_region_info_64() 579 size = vm_map_round_page(size, in vm32_vm_mapped_pages_info() 589 vm_map_round_page(addr + size, in vm32_vm_mapped_pages_info() 611 size = vm_map_round_page(actual * sizeof(vm_offset_t), in vm32_vm_mapped_pages_info() 622 vmsize_used = vm_map_round_page(size_used, in vm32_vm_mapped_pages_info() 628 vm_map_round_page(addr + size, in vm32_vm_mapped_pages_info() 699 size = vm_map_round_page(actual * sizeof *info, in host_virtual_physical_table_info() [all …]
|
| H A D | vm_sanitize_error_compat.c | 419 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_vm_protect() 459 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_useracc() 507 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_vm_behavior_set() 628 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_mincore()
|
| H A D | vm_map.h | 224 #define vm_map_round_page(x, pgmask) \ macro
|
| H A D | vm_map_store.c | 498 start = vm_map_round_page(start, map_mask); in vm_map_store_find_space_forward() 505 desired_empty_end = vm_map_round_page(end, map_mask); in vm_map_store_find_space_forward()
|
| H A D | vm_shared_region.c | 1546 end = vm_map_round_page(end, VM_MAP_PAGE_MASK(sr_map)); in vm_shared_region_undo_mappings() 1751 vm_map_round_page(mappings[0].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup() 1773 (vm_map_round_page(sfm_end, VM_MAP_PAGE_MASK(sr_map)) < in vm_shared_region_map_file_setup() 1830 (vm_map_round_page(sfm_end, VM_MAP_PAGE_MASK(sr_map)) < in vm_shared_region_map_file_setup() 1899 obj_size = vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)); in vm_shared_region_map_file_setup() 1907 vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup() 1922 vm_map_round_page( in vm_shared_region_map_file_setup()
|
| H A D | vm_upl.c | 95 size_to_map = vm_map_round_page(size_to_map + adjusted_offset, vm_map_page_mask(map)); in vm_upl_map_range() 118 size_to_unmap = vm_map_round_page(size_to_unmap + page_offset, vm_map_page_mask(map)); in vm_upl_unmap_range()
|
| H A D | vm_reclaim.c | 275 return vm_map_round_page(metadata_size + entries_size, vm_map_page_mask(map)); in vmdr_round_len_to_size() 1080 vm_map_address_t end = vm_map_round_page(entry->address + entry->size, in reclaim_chunk() 1578 *trim_threshold_out = vm_map_round_page(unneeded_bytes, in vmdr_sample_working_set()
|
| H A D | vm_map.c | 1141 end_aligned = vm_map_round_page(end, PAGE_MASK_64); in vm_map_apple_protected() 1143 end_aligned = vm_map_round_page(end_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected() 4383 (vm_map_round_page(obj_size, map_mask) == named_entry->size)) { in vm_map_enter_mem_object() 4576 vm_map_round_page(target_size, map_mask), in vm_map_enter_mem_object() 4902 map_size = vm_map_round_page(initial_size + in vm_map_enter_mem_object() 9936 dst_end = vm_map_round_page(dst_addr + dst_size, in vm_map_overwrite_submap_recurse() 10218 dst_end = vm_map_round_page(dst_addr + copy->size, in vm_map_copy_overwrite_nested() 11994 vm_map_round_page(src_addr + len, VM_MAP_PAGE_MASK(src_map)), in vm_map_copyin_kernel_buffer() 12068 vm_map_round_page(copy_size, in vm_map_copyout_kernel_buffer() 12134 vm_map_round_page((*addr + in vm_map_copyout_kernel_buffer() [all …]
|
| H A D | vm_user.c | 1827 vm_map_round_page(start + curr_sz, effective_page_mask), in mach_vm_page_range_query() 1864 curr_sz = MIN(vm_map_round_page(size, effective_page_mask), MAX_PAGE_RANGE_QUERY); in mach_vm_page_range_query()
|
| H A D | vm_tests.c | 678 dealloc_addr = vm_map_round_page(alloc5_addr, PAGE_SHIFT); in vm_test_4k() 685 remap_src_addr = vm_map_round_page(alloc3_addr, PAGE_SHIFT); in vm_test_4k()
|
| /xnu-12377.41.6/osfmk/arm64/ |
| H A D | alternate_debugger.c | 128 alt_size = vm_map_round_page(alt_size, in alternate_debugger_init() 130 alt_pages_size = vm_map_round_page(alt_pages_size, in alternate_debugger_init()
|
| /xnu-12377.41.6/san/memory/ |
| H A D | kasan-arm64.c | 191 *sizep = vm_map_round_page(*sizep + (*addrp - addr_aligned), ARM_PGMASK); in kasan_arm64_align_to_page() 319 vm_offset_t shadow_top = vm_map_round_page(SHADOW_FOR_ADDRESS(address + size), ARM_PGMASK); in kasan_map_shadow_internal() 414 shadow_pbase = vm_map_round_page(pbase + memSize, ARM_PGMASK); in kasan_bootstrap() 481 shadow_pbase = vm_map_round_page(pbase + args->memSize, ARM_PGMASK); in kasan_bootstrap()
|
| H A D | kasan-report.c | 60 uptr shadow_page = vm_map_round_page(shadow_p, PAGE_MASK); in kasan_dump_shadow() 70 if ((vm_map_round_page(shadow, PAGE_MASK) != shadow_page) && !kasan_is_shadow_mapped(shadow)) { in kasan_dump_shadow()
|
| H A D | kasan-x86_64.c | 132 size = vm_map_round_page(size, I386_LPGMASK); in kasan_map_shadow_superpage_zero() 189 vm_offset_t shadow_top = vm_map_round_page(SHADOW_FOR_ADDRESS(address + size), PAGE_MASK); in kasan_map_shadow()
|
| H A D | kasan.c | 398 shadow_top = vm_map_round_page(shadow_top, PAGE_MASK); in kasan_traverse_mappings()
|
| /xnu-12377.41.6/osfmk/kern/ |
| H A D | kext_alloc.c | 86 text_end = vm_map_round_page(text->vmaddr + text->vmsize, in kext_alloc_init() 101 vm_map_round_page(kext_alloc_base + prelinkTextSegment->vmsize, in kext_alloc_init()
|
| H A D | lock_group.c | 357 vmsize = vm_map_round_page(size, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info() 394 vmused = vm_map_round_page(used, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info()
|
| H A D | host.c | 1212 size = vm_map_round_page(needed, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_processor_info() 1239 vm_map_round_page(addr + size, VM_MAP_PAGE_MASK(ipc_kernel_map)), FALSE); in host_processor_info()
|
| /xnu-12377.41.6/bsd/kern/ |
| H A D | posix_shm.c | 607 total_size = vm_map_round_page(length, vm_map_page_mask(user_map)); in pshm_truncate() 795 pshm_size = vm_map_round_page((vm_map_size_t)pinfo->pshm_length, vm_map_page_mask(user_map)); in pshm_mmap() 835 user_addr = vm_map_round_page(user_addr, in pshm_mmap() 839 if (user_addr != vm_map_round_page(user_addr, in pshm_mmap()
|
| H A D | mach_loader.c | 1293 slide = vm_map_round_page(slide + max_vm_addr, effective_page_mask); in parse_machfile() 1775 vm_end = vm_map_round_page(result->dynlinker_max_vm_addr, mask); in parse_machfile() 1777 vm_end = vm_map_round_page(result->max_vm_addr, mask); in parse_machfile() 1780 mach_vm_size_t user_stack_size = vm_map_round_page(result->user_stack_alloc_size, mask); in parse_machfile() 1781 …result->user_stack = vm_map_round_page(vm_end + user_stack_size + reserved_aot_size + slide, mask); in parse_machfile() 2409 vm_end = vm_map_round_page(vm_end, in load_segment() 2435 file_end = vm_map_round_page(file_offset + file_size, in load_segment() 2439 vm_end = vm_map_round_page(vm_offset + vm_size, in load_segment() 2459 file_end = vm_map_round_page(file_end, in load_segment() 2461 vm_end = vm_map_round_page(vm_end, in load_segment() [all …]
|
| H A D | sysv_shm.c | 280 size = vm_map_round_page(shmseg->u.shm_segsz, in shm_deallocate_segment() 301 size = vm_map_round_page(shmseg->u.shm_segsz, in shm_delete_mapping() 462 map_size = vm_map_round_page(shmseg->u.shm_segsz, in shmat()
|
| /xnu-12377.41.6/osfmk/ipc/ |
| H A D | mach_debug.c | 192 vm_map_round_page(tsize * sizeof(ipc_info_name_t), in mach_port_space_info() 283 vm_map_size_t keep = vm_map_round_page(used, in mach_port_space_info()
|
| H A D | mach_port.c | 267 size_needed = vm_map_round_page( in mach_port_names() 336 vm_map_round_page(size_used, in mach_port_names() 1312 size = vm_map_round_page(actual * sizeof(mach_port_name_t), in mach_port_get_set_status() 1330 vm_size_used = vm_map_round_page(size_used, in mach_port_get_set_status()
|
| /xnu-12377.41.6/osfmk/console/ |
| H A D | serial_console.c | 108 #define KERN_CONSOLE_RING_SIZE vm_map_round_page(CPU_CONS_BUF_SIZE * (MAX_CPUS + 1), PAGE_SIZE - 1)
|
| /xnu-12377.41.6/bsd/nfs/ |
| H A D | nfs_gss.c | 1397 tbuflen = vm_map_round_page(buflen, vm_map_page_mask(ipc_kernel_map)); in nfs_gss_mach_alloc_buffer()
|