| /xnu-8020.140.41/osfmk/vm/ |
| H A D | vm_map_store_ll.c | 38 VM_MAP_PAGE_MASK(map)) == in first_free_is_valid_ll() 40 VM_MAP_PAGE_MASK(map)) || in first_free_is_valid_ll() 42 VM_MAP_PAGE_MASK(map)) == in first_free_is_valid_ll() 44 VM_MAP_PAGE_MASK(map)) && in first_free_is_valid_ll() 77 VM_MAP_PAGE_MASK(UFF_map)) == \ 79 VM_MAP_PAGE_MASK(UFF_map)) || \ 81 VM_MAP_PAGE_MASK(UFF_map)) == \ 83 VM_MAP_PAGE_MASK(UFF_map)) && \
|
| H A D | vm_debug.c | 275 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info() 295 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info() 486 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info_64() 506 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info_64() 559 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_mapped_pages_info() 566 VM_MAP_PAGE_MASK(ipc_kernel_map)), in vm32_mapped_pages_info() 568 VM_MAP_PAGE_MASK(ipc_kernel_map)), in vm32_mapped_pages_info() 590 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_mapped_pages_info() 601 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_mapped_pages_info() 605 VM_MAP_PAGE_MASK(ipc_kernel_map)), in vm32_mapped_pages_info() [all …]
|
| H A D | vm_user.c | 202 VM_MAP_PAGE_MASK(map)); in mach_vm_allocate_kernel() 205 VM_MAP_PAGE_MASK(map)); in mach_vm_allocate_kernel() 281 VM_MAP_PAGE_MASK(map)); in vm_allocate_external() 284 VM_MAP_PAGE_MASK(map)); in vm_allocate_external() 335 VM_MAP_PAGE_MASK(map)), in mach_vm_deallocate() 337 VM_MAP_PAGE_MASK(map)), in mach_vm_deallocate() 364 VM_MAP_PAGE_MASK(map)), in vm_deallocate() 366 VM_MAP_PAGE_MASK(map)), in vm_deallocate() 394 VM_MAP_PAGE_MASK(map)), in mach_vm_inherit() 396 VM_MAP_PAGE_MASK(map)), in mach_vm_inherit() [all …]
|
| H A D | vm_map.c | 857 start_aligned = vm_map_trunc_page(start_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected() 858 end_aligned = vm_map_round_page(end_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected() 2204 assert(VM_MAP_PAGE_ALIGNED(new_entry->vme_start, VM_MAP_PAGE_MASK(map))); in vm_map_find_space() 2205 assert(VM_MAP_PAGE_ALIGNED(new_entry->vme_end, VM_MAP_PAGE_MASK(map))); in vm_map_find_space() 2332 assert(VM_MAP_PAGE_ALIGNED(size, VM_MAP_PAGE_MASK(map))); in vm_map_random_address_for_size() 2343 VM_MAP_PAGE_MASK(map)); in vm_map_random_address_for_size() 2592 (offset & MIN(VM_MAP_PAGE_MASK(map), PAGE_MASK_64)) != 0) { in vm_map_enter() 2611 if (VM_MAP_PAGE_MASK(map) >= PAGE_MASK) { in vm_map_enter() 2616 if (VM_MAP_PAGE_MASK(map) >= PAGE_MASK && in vm_map_enter() 2617 !VM_MAP_PAGE_ALIGNED(size, VM_MAP_PAGE_MASK(map))) { in vm_map_enter() [all …]
|
| H A D | vm_kern.c | 559 map_size = vm_map_round_page(size, VM_MAP_PAGE_MASK(map)); in kmem_alloc_contig() 610 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig() 612 VM_MAP_PAGE_MASK(map))); in kmem_alloc_contig() 637 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig() 639 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig() 651 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig() 653 VM_MAP_PAGE_MASK(map))); in kmem_alloc_contig() 2165 VM_MAP_PAGE_MASK(kernel_map)); in kmem_init() 2167 VM_MAP_PAGE_MASK(kernel_map)); in kmem_init() 2189 VM_MAP_PAGE_MASK(kernel_map)), in kmem_init()
|
| H A D | vm_map_store.c | 280 const vm_map_offset_t map_mask = VM_MAP_PAGE_MASK(map); in vm_map_store_find_space_backwards() 394 const vm_map_offset_t map_mask = VM_MAP_PAGE_MASK(map); in vm_map_store_find_space_forward()
|
| H A D | vm_pageout.c | 6503 offset = vm_map_trunc_page(original_offset, VM_MAP_PAGE_MASK(map)); in vm_map_create_upl() 6506 VM_MAP_PAGE_MASK(map)) in vm_map_create_upl() 6662 VM_MAP_PAGE_MASK(map))); in vm_map_create_upl() 6666 VM_MAP_PAGE_MASK(map))); in vm_map_create_upl() 6985 if (VM_MAP_PAGE_MASK(map) < PAGE_MASK) { in vm_map_enter_upl_range() 7033 size = upl_adjusted_size(upl, VM_MAP_PAGE_MASK(map)); in vm_map_enter_upl_range() 7116 offset = upl_adjusted_offset(upl, VM_MAP_PAGE_MASK(map)) - upl->map_object->paging_offset; in vm_map_enter_upl_range() 7123 size = upl_adjusted_size(upl, VM_MAP_PAGE_MASK(map)); in vm_map_enter_upl_range() 7125 size = MIN(upl_adjusted_size(upl, VM_MAP_PAGE_MASK(map)), size_to_map); in vm_map_enter_upl_range() 7197 …addr_adjustment = (vm_map_offset_t)(upl->u_offset - upl_adjusted_offset(upl, VM_MAP_PAGE_MASK(map)… in vm_map_enter_upl_range() [all …]
|
| H A D | vm_shared_region.c | 1484 (vm_map_round_page(sfm_end, VM_MAP_PAGE_MASK(sr_map)) < in vm_shared_region_map_file_setup() 1540 obj_size = vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)); in vm_shared_region_map_file_setup() 1548 vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup() 1565 vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup()
|
| H A D | vm_map_store_rb.c | 667 if (vm_map_trunc_page(entry->vme_start, VM_MAP_PAGE_MASK(map)) >= max_valid_offset) { in update_first_free_rb()
|
| H A D | vm_fault.c | 4114 fault_page_mask = VM_MAP_PAGE_MASK(original_map); in vm_fault_internal() 4425 cur_offset < VM_MAP_ROUND_PAGE(cur_object->vo_size, VM_MAP_PAGE_MASK(map)) && in vm_fault_internal() 4941 if (VM_MAP_PAGE_MASK(map) != PAGE_MASK) { in vm_fault_internal() 7278 effective_page_mask = VM_MAP_PAGE_MASK(map); in kdp_lightweight_fault()
|
| H A D | vm_map.h | 1575 #define VM_MAP_PAGE_MASK(map) (VM_MAP_PAGE_SIZE((map)) - 1) macro
|
| H A D | vm_compressor.c | 832 …_size = vm_map_round_page((sizeof(union c_segu) * c_segments_limit), VM_MAP_PAGE_MASK(kernel_map)); in vm_compressor_set_size() 833 …und_page(((vm_size_t)c_seg_allocsize * (vm_size_t)c_segments_limit), VM_MAP_PAGE_MASK(kernel_map)); in vm_compressor_set_size()
|
| /xnu-8020.140.41/osfmk/arm64/ |
| H A D | alternate_debugger.c | 129 VM_MAP_PAGE_MASK(kernel_map)); in alternate_debugger_init() 131 VM_MAP_PAGE_MASK(kernel_map)); in alternate_debugger_init() 135 VM_MAP_PAGE_MASK(kernel_map), 0, 0, in alternate_debugger_init()
|
| /xnu-8020.140.41/osfmk/kern/ |
| H A D | kext_alloc.c | 82 VM_MAP_PAGE_MASK(kernel_map)); in kext_alloc_init() 85 VM_MAP_PAGE_MASK(kernel_map)); in kext_alloc_init() 100 VM_MAP_PAGE_MASK(kernel_map)); in kext_alloc_init() 154 VM_MAP_PAGE_MASK(g_kext_map)); in get_address_from_kext_map()
|
| H A D | lock_group.c | 308 vmsize = vm_map_round_page(size, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info() 358 vmused = vm_map_round_page(used, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info()
|
| H A D | stack.c | 345 VM_MAP_PAGE_MASK(kernel_map)); in stack_collect()
|
| H A D | host.c | 1155 size = vm_map_round_page(needed, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_processor_info() 1181 result = vm_map_unwire(ipc_kernel_map, vm_map_trunc_page(addr, VM_MAP_PAGE_MASK(ipc_kernel_map)), in host_processor_info() 1182 vm_map_round_page(addr + size, VM_MAP_PAGE_MASK(ipc_kernel_map)), FALSE); in host_processor_info()
|
| H A D | debug.c | 495 VM_MAP_PAGE_MASK(kernel_map), 0, 0, KMA_PERMANENT | KMA_NOPAGEWAIT | KMA_DATA, in phys_carveout_init() 515 VM_MAP_PAGE_MASK(kernel_map), 0, 0, in phys_carveout_init()
|
| H A D | kern_stackshot.c | 3366 *effective_page_mask = VM_MAP_PAGE_MASK(map); in _stackshot_get_page_size()
|
| /xnu-8020.140.41/osfmk/ipc/ |
| H A D | mach_debug.c | 232 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_space_info() 314 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_space_info()
|
| H A D | mach_port.c | 273 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_names() 341 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_names() 1324 VM_MAP_PAGE_MASK(ipc_kernel_map)) + in mach_port_get_set_status() 1341 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_get_set_status()
|
| H A D | ipc_kmsg.c | 4811 effective_page_mask = VM_MAP_PAGE_MASK(map); in ipc_kmsg_copyout_ool_descriptor()
|
| /xnu-8020.140.41/osfmk/kdp/output_stages/ |
| H A D | out_shmem.c | 404 kdp_hw_shmem_dbg_bufsize, VM_MAP_PAGE_MASK(kernel_map), in shmem_stage_initialize()
|
| /xnu-8020.140.41/osfmk/i386/commpage/ |
| H A D | commpage.c | 168 …if (!(kr = vm_map_lookup_entry( kernel_map, vm_map_trunc_page(kernel_addr, VM_MAP_PAGE_MASK(kernel… in commpage_allocate()
|