| /xnu-11417.101.15/osfmk/vm/ |
| H A D | vm_debug.c | 276 size = vm_map_round_page(2 * used * sizeof(vm_info_object_t), in vm32_mach_vm_region_info() 296 vm_size_t vmsize_used = vm_map_round_page(size_used, in vm32_mach_vm_region_info() 489 size = vm_map_round_page(2 * used * sizeof(vm_info_object_t), in vm32_mach_vm_region_info_64() 509 vm_size_t vmsize_used = vm_map_round_page(size_used, in vm32_mach_vm_region_info_64() 562 size = vm_map_round_page(size, in vm32_vm_mapped_pages_info() 572 vm_map_round_page(addr + size, in vm32_vm_mapped_pages_info() 594 size = vm_map_round_page(actual * sizeof(vm_offset_t), in vm32_vm_mapped_pages_info() 605 vmsize_used = vm_map_round_page(size_used, in vm32_vm_mapped_pages_info() 611 vm_map_round_page(addr + size, in vm32_vm_mapped_pages_info() 682 size = vm_map_round_page(actual * sizeof *info, in host_virtual_physical_table_info() [all …]
|
| H A D | vm_sanitize_error_compat.c | 419 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_vm_protect() 459 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_useracc() 507 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_vm_behavior_set() 628 vm_address_t aligned_end = vm_map_round_page(start + size, pgmask); in vm_sanitize_err_compat_addr_size_mincore()
|
| H A D | vm_map.h | 224 #define vm_map_round_page(x, pgmask) \ macro
|
| H A D | vm_map_store.c | 498 start = vm_map_round_page(start, map_mask); in vm_map_store_find_space_forward() 505 desired_empty_end = vm_map_round_page(end, map_mask); in vm_map_store_find_space_forward()
|
| H A D | vm_shared_region.c | 1410 end = vm_map_round_page(end, VM_MAP_PAGE_MASK(sr_map)); in vm_shared_region_undo_mappings() 1655 vm_map_round_page(mappings[0].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup() 1677 (vm_map_round_page(sfm_end, VM_MAP_PAGE_MASK(sr_map)) < in vm_shared_region_map_file_setup() 1734 (vm_map_round_page(sfm_end, VM_MAP_PAGE_MASK(sr_map)) < in vm_shared_region_map_file_setup() 1803 obj_size = vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)); in vm_shared_region_map_file_setup() 1811 vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup() 1826 vm_map_round_page( in vm_shared_region_map_file_setup()
|
| H A D | vm_user.c | 1732 vm_map_round_page(start + curr_sz, effective_page_mask), in mach_vm_page_range_query() 1769 curr_sz = MIN(vm_map_round_page(size, effective_page_mask), MAX_PAGE_RANGE_QUERY); in mach_vm_page_range_query() 1858 vm_map_round_page(size, VM_MAP_PAGE_MASK(target_map)), in vm_region_object_create()
|
| H A D | vm_map.c | 1066 end_aligned = vm_map_round_page(end, PAGE_MASK_64); in vm_map_apple_protected() 1068 end_aligned = vm_map_round_page(end_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected() 4121 (vm_map_round_page(obj_size, map_mask) == named_entry->size)) { in vm_map_enter_mem_object() 4304 vm_map_round_page(target_size, map_mask), in vm_map_enter_mem_object() 4621 map_size = vm_map_round_page(initial_size + in vm_map_enter_mem_object() 5883 end == vm_map_round_page(prev, VM_MAP_PAGE_MASK(map))) { in vm_map_protect() 5889 (vm_map_round_page(prev_entry->vme_end, in vm_map_protect() 9283 dst_end = vm_map_round_page(dst_addr + dst_size, in vm_map_overwrite_submap_recurse() 9541 dst_end = vm_map_round_page(dst_addr + copy->size, in vm_map_copy_overwrite_nested() 11253 vm_map_round_page(src_addr + len, VM_MAP_PAGE_MASK(src_map)), in vm_map_copyin_kernel_buffer() [all …]
|
| H A D | vm_reclaim.c | 307 return vm_map_round_page(metadata_size + entries_size, vm_map_page_mask(map)); in vmdr_round_len_to_size() 1066 vm_map_address_t end = vm_map_round_page(entry->address + entry->size, 1491 *trim_threshold_out = vm_map_round_page(trim_threshold_bytes,
|
| H A D | vm_upl.c | 90 size_to_map = vm_map_round_page(size_to_map + adjusted_offset, vm_map_page_mask(map)); in vm_upl_map_range() 113 size_to_unmap = vm_map_round_page(size_to_unmap + page_offset, vm_map_page_mask(map)); in vm_upl_unmap_range()
|
| H A D | vm_tests.c | 669 dealloc_addr = vm_map_round_page(alloc5_addr, PAGE_SHIFT); in vm_test_4k() 676 remap_src_addr = vm_map_round_page(alloc3_addr, PAGE_SHIFT); in vm_test_4k()
|
| /xnu-11417.101.15/osfmk/arm64/ |
| H A D | alternate_debugger.c | 128 alt_size = vm_map_round_page(alt_size, in alternate_debugger_init() 130 alt_pages_size = vm_map_round_page(alt_pages_size, in alternate_debugger_init()
|
| /xnu-11417.101.15/san/memory/ |
| H A D | kasan-arm64.c | 191 *sizep = vm_map_round_page(*sizep + (*addrp - addr_aligned), ARM_PGMASK); in kasan_arm64_align_to_page() 319 vm_offset_t shadow_top = vm_map_round_page(SHADOW_FOR_ADDRESS(address + size), ARM_PGMASK); in kasan_map_shadow_internal() 414 shadow_pbase = vm_map_round_page(pbase + memSize, ARM_PGMASK); in kasan_bootstrap() 481 shadow_pbase = vm_map_round_page(pbase + args->memSize, ARM_PGMASK); in kasan_bootstrap()
|
| H A D | kasan-report.c | 60 uptr shadow_page = vm_map_round_page(shadow_p, PAGE_MASK); in kasan_dump_shadow() 70 if ((vm_map_round_page(shadow, PAGE_MASK) != shadow_page) && !kasan_is_shadow_mapped(shadow)) { in kasan_dump_shadow()
|
| H A D | kasan-x86_64.c | 132 size = vm_map_round_page(size, I386_LPGMASK); in kasan_map_shadow_superpage_zero() 189 vm_offset_t shadow_top = vm_map_round_page(SHADOW_FOR_ADDRESS(address + size), PAGE_MASK); in kasan_map_shadow()
|
| H A D | kasan.c | 398 shadow_top = vm_map_round_page(shadow_top, PAGE_MASK); in kasan_traverse_mappings()
|
| /xnu-11417.101.15/osfmk/kern/ |
| H A D | kext_alloc.c | 86 text_end = vm_map_round_page(text->vmaddr + text->vmsize, in kext_alloc_init() 101 vm_map_round_page(kext_alloc_base + prelinkTextSegment->vmsize, in kext_alloc_init()
|
| H A D | lock_group.c | 357 vmsize = vm_map_round_page(size, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info() 394 vmused = vm_map_round_page(used, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info()
|
| H A D | host.c | 1189 size = vm_map_round_page(needed, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_processor_info() 1216 vm_map_round_page(addr + size, VM_MAP_PAGE_MASK(ipc_kernel_map)), FALSE); in host_processor_info()
|
| /xnu-11417.101.15/bsd/kern/ |
| H A D | posix_shm.c | 607 total_size = vm_map_round_page(length, vm_map_page_mask(user_map)); in pshm_truncate() 795 pshm_size = vm_map_round_page((vm_map_size_t)pinfo->pshm_length, vm_map_page_mask(user_map)); in pshm_mmap() 835 user_addr = vm_map_round_page(user_addr, in pshm_mmap() 839 if (user_addr != vm_map_round_page(user_addr, in pshm_mmap()
|
| H A D | mach_loader.c | 1011 slide = vm_map_round_page(slide + max_vm_addr, effective_page_mask); in parse_machfile() 1493 vm_end = vm_map_round_page(result->dynlinker_max_vm_addr, mask); in parse_machfile() 1495 vm_end = vm_map_round_page(result->max_vm_addr, mask); in parse_machfile() 1498 mach_vm_size_t user_stack_size = vm_map_round_page(result->user_stack_alloc_size, mask); in parse_machfile() 1499 …result->user_stack = vm_map_round_page(vm_end + user_stack_size + reserved_aot_size + slide, mask); in parse_machfile() 2127 vm_end = vm_map_round_page(vm_end, in load_segment() 2153 file_end = vm_map_round_page(file_offset + file_size, in load_segment() 2157 vm_end = vm_map_round_page(vm_offset + vm_size, in load_segment() 2177 file_end = vm_map_round_page(file_end, in load_segment() 2179 vm_end = vm_map_round_page(vm_end, in load_segment() [all …]
|
| H A D | sysv_shm.c | 280 size = vm_map_round_page(shmseg->u.shm_segsz, in shm_deallocate_segment() 301 size = vm_map_round_page(shmseg->u.shm_segsz, in shm_delete_mapping() 462 map_size = vm_map_round_page(shmseg->u.shm_segsz, in shmat()
|
| /xnu-11417.101.15/osfmk/ipc/ |
| H A D | mach_debug.c | 191 vm_map_round_page(tsize * sizeof(ipc_info_name_t), in mach_port_space_info() 282 vm_map_size_t keep = vm_map_round_page(used, in mach_port_space_info()
|
| H A D | mach_port.c | 270 size_needed = vm_map_round_page( in mach_port_names() 339 vm_map_round_page(size_used, in mach_port_names() 1311 size = vm_map_round_page(actual * sizeof(mach_port_name_t), in mach_port_get_set_status() 1329 vm_size_used = vm_map_round_page(size_used, in mach_port_get_set_status()
|
| /xnu-11417.101.15/osfmk/console/ |
| H A D | serial_console.c | 108 #define KERN_CONSOLE_RING_SIZE vm_map_round_page(CPU_CONS_BUF_SIZE * (MAX_CPUS + 1), PAGE_SIZE - 1)
|
| /xnu-11417.101.15/bsd/nfs/ |
| H A D | nfs_gss.c | 1395 tbuflen = vm_map_round_page(buflen, vm_map_page_mask(ipc_kernel_map)); in nfs_gss_mach_alloc_buffer()
|