Home
last modified time | relevance | path

Searched refs:VM_MAP_PAGE_MASK (Results 1 – 24 of 24) sorted by relevance

/xnu-8020.140.41/osfmk/vm/
H A Dvm_map_store_ll.c38 VM_MAP_PAGE_MASK(map)) == in first_free_is_valid_ll()
40 VM_MAP_PAGE_MASK(map)) || in first_free_is_valid_ll()
42 VM_MAP_PAGE_MASK(map)) == in first_free_is_valid_ll()
44 VM_MAP_PAGE_MASK(map)) && in first_free_is_valid_ll()
77 VM_MAP_PAGE_MASK(UFF_map)) == \
79 VM_MAP_PAGE_MASK(UFF_map)) || \
81 VM_MAP_PAGE_MASK(UFF_map)) == \
83 VM_MAP_PAGE_MASK(UFF_map)) && \
H A Dvm_debug.c275 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info()
295 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info()
486 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info_64()
506 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_region_info_64()
559 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_mapped_pages_info()
566 VM_MAP_PAGE_MASK(ipc_kernel_map)), in vm32_mapped_pages_info()
568 VM_MAP_PAGE_MASK(ipc_kernel_map)), in vm32_mapped_pages_info()
590 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_mapped_pages_info()
601 VM_MAP_PAGE_MASK(ipc_kernel_map)); in vm32_mapped_pages_info()
605 VM_MAP_PAGE_MASK(ipc_kernel_map)), in vm32_mapped_pages_info()
[all …]
H A Dvm_user.c202 VM_MAP_PAGE_MASK(map)); in mach_vm_allocate_kernel()
205 VM_MAP_PAGE_MASK(map)); in mach_vm_allocate_kernel()
281 VM_MAP_PAGE_MASK(map)); in vm_allocate_external()
284 VM_MAP_PAGE_MASK(map)); in vm_allocate_external()
335 VM_MAP_PAGE_MASK(map)), in mach_vm_deallocate()
337 VM_MAP_PAGE_MASK(map)), in mach_vm_deallocate()
364 VM_MAP_PAGE_MASK(map)), in vm_deallocate()
366 VM_MAP_PAGE_MASK(map)), in vm_deallocate()
394 VM_MAP_PAGE_MASK(map)), in mach_vm_inherit()
396 VM_MAP_PAGE_MASK(map)), in mach_vm_inherit()
[all …]
H A Dvm_map.c857 start_aligned = vm_map_trunc_page(start_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected()
858 end_aligned = vm_map_round_page(end_aligned, VM_MAP_PAGE_MASK(map)); in vm_map_apple_protected()
2204 assert(VM_MAP_PAGE_ALIGNED(new_entry->vme_start, VM_MAP_PAGE_MASK(map))); in vm_map_find_space()
2205 assert(VM_MAP_PAGE_ALIGNED(new_entry->vme_end, VM_MAP_PAGE_MASK(map))); in vm_map_find_space()
2332 assert(VM_MAP_PAGE_ALIGNED(size, VM_MAP_PAGE_MASK(map))); in vm_map_random_address_for_size()
2343 VM_MAP_PAGE_MASK(map)); in vm_map_random_address_for_size()
2592 (offset & MIN(VM_MAP_PAGE_MASK(map), PAGE_MASK_64)) != 0) { in vm_map_enter()
2611 if (VM_MAP_PAGE_MASK(map) >= PAGE_MASK) { in vm_map_enter()
2616 if (VM_MAP_PAGE_MASK(map) >= PAGE_MASK && in vm_map_enter()
2617 !VM_MAP_PAGE_ALIGNED(size, VM_MAP_PAGE_MASK(map))) { in vm_map_enter()
[all …]
H A Dvm_kern.c559 map_size = vm_map_round_page(size, VM_MAP_PAGE_MASK(map)); in kmem_alloc_contig()
610 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig()
612 VM_MAP_PAGE_MASK(map))); in kmem_alloc_contig()
637 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig()
639 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig()
651 VM_MAP_PAGE_MASK(map)), in kmem_alloc_contig()
653 VM_MAP_PAGE_MASK(map))); in kmem_alloc_contig()
2165 VM_MAP_PAGE_MASK(kernel_map)); in kmem_init()
2167 VM_MAP_PAGE_MASK(kernel_map)); in kmem_init()
2189 VM_MAP_PAGE_MASK(kernel_map)), in kmem_init()
H A Dvm_map_store.c280 const vm_map_offset_t map_mask = VM_MAP_PAGE_MASK(map); in vm_map_store_find_space_backwards()
394 const vm_map_offset_t map_mask = VM_MAP_PAGE_MASK(map); in vm_map_store_find_space_forward()
H A Dvm_pageout.c6503 offset = vm_map_trunc_page(original_offset, VM_MAP_PAGE_MASK(map)); in vm_map_create_upl()
6506 VM_MAP_PAGE_MASK(map)) in vm_map_create_upl()
6662 VM_MAP_PAGE_MASK(map))); in vm_map_create_upl()
6666 VM_MAP_PAGE_MASK(map))); in vm_map_create_upl()
6985 if (VM_MAP_PAGE_MASK(map) < PAGE_MASK) { in vm_map_enter_upl_range()
7033 size = upl_adjusted_size(upl, VM_MAP_PAGE_MASK(map)); in vm_map_enter_upl_range()
7116 offset = upl_adjusted_offset(upl, VM_MAP_PAGE_MASK(map)) - upl->map_object->paging_offset; in vm_map_enter_upl_range()
7123 size = upl_adjusted_size(upl, VM_MAP_PAGE_MASK(map)); in vm_map_enter_upl_range()
7125 size = MIN(upl_adjusted_size(upl, VM_MAP_PAGE_MASK(map)), size_to_map); in vm_map_enter_upl_range()
7197 …addr_adjustment = (vm_map_offset_t)(upl->u_offset - upl_adjusted_offset(upl, VM_MAP_PAGE_MASK(map)… in vm_map_enter_upl_range()
[all …]
H A Dvm_shared_region.c1484 (vm_map_round_page(sfm_end, VM_MAP_PAGE_MASK(sr_map)) < in vm_shared_region_map_file_setup()
1540 obj_size = vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)); in vm_shared_region_map_file_setup()
1548 vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup()
1565 vm_map_round_page(mappings[i].sms_size, VM_MAP_PAGE_MASK(sr_map)), in vm_shared_region_map_file_setup()
H A Dvm_map_store_rb.c667 if (vm_map_trunc_page(entry->vme_start, VM_MAP_PAGE_MASK(map)) >= max_valid_offset) { in update_first_free_rb()
H A Dvm_fault.c4114 fault_page_mask = VM_MAP_PAGE_MASK(original_map); in vm_fault_internal()
4425 cur_offset < VM_MAP_ROUND_PAGE(cur_object->vo_size, VM_MAP_PAGE_MASK(map)) && in vm_fault_internal()
4941 if (VM_MAP_PAGE_MASK(map) != PAGE_MASK) { in vm_fault_internal()
7278 effective_page_mask = VM_MAP_PAGE_MASK(map); in kdp_lightweight_fault()
H A Dvm_map.h1575 #define VM_MAP_PAGE_MASK(map) (VM_MAP_PAGE_SIZE((map)) - 1) macro
H A Dvm_compressor.c832 …_size = vm_map_round_page((sizeof(union c_segu) * c_segments_limit), VM_MAP_PAGE_MASK(kernel_map)); in vm_compressor_set_size()
833 …und_page(((vm_size_t)c_seg_allocsize * (vm_size_t)c_segments_limit), VM_MAP_PAGE_MASK(kernel_map)); in vm_compressor_set_size()
/xnu-8020.140.41/osfmk/arm64/
H A Dalternate_debugger.c129 VM_MAP_PAGE_MASK(kernel_map)); in alternate_debugger_init()
131 VM_MAP_PAGE_MASK(kernel_map)); in alternate_debugger_init()
135 VM_MAP_PAGE_MASK(kernel_map), 0, 0, in alternate_debugger_init()
/xnu-8020.140.41/osfmk/kern/
H A Dkext_alloc.c82 VM_MAP_PAGE_MASK(kernel_map)); in kext_alloc_init()
85 VM_MAP_PAGE_MASK(kernel_map)); in kext_alloc_init()
100 VM_MAP_PAGE_MASK(kernel_map)); in kext_alloc_init()
154 VM_MAP_PAGE_MASK(g_kext_map)); in get_address_from_kext_map()
H A Dlock_group.c308 vmsize = vm_map_round_page(size, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info()
358 vmused = vm_map_round_page(used, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_lockgroup_info()
H A Dstack.c345 VM_MAP_PAGE_MASK(kernel_map)); in stack_collect()
H A Dhost.c1155 size = vm_map_round_page(needed, VM_MAP_PAGE_MASK(ipc_kernel_map)); in host_processor_info()
1181 result = vm_map_unwire(ipc_kernel_map, vm_map_trunc_page(addr, VM_MAP_PAGE_MASK(ipc_kernel_map)), in host_processor_info()
1182 vm_map_round_page(addr + size, VM_MAP_PAGE_MASK(ipc_kernel_map)), FALSE); in host_processor_info()
H A Ddebug.c495 VM_MAP_PAGE_MASK(kernel_map), 0, 0, KMA_PERMANENT | KMA_NOPAGEWAIT | KMA_DATA, in phys_carveout_init()
515 VM_MAP_PAGE_MASK(kernel_map), 0, 0, in phys_carveout_init()
H A Dkern_stackshot.c3366 *effective_page_mask = VM_MAP_PAGE_MASK(map); in _stackshot_get_page_size()
/xnu-8020.140.41/osfmk/ipc/
H A Dmach_debug.c232 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_space_info()
314 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_space_info()
H A Dmach_port.c273 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_names()
341 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_names()
1324 VM_MAP_PAGE_MASK(ipc_kernel_map)) + in mach_port_get_set_status()
1341 VM_MAP_PAGE_MASK(ipc_kernel_map)); in mach_port_get_set_status()
H A Dipc_kmsg.c4811 effective_page_mask = VM_MAP_PAGE_MASK(map); in ipc_kmsg_copyout_ool_descriptor()
/xnu-8020.140.41/osfmk/kdp/output_stages/
H A Dout_shmem.c404 kdp_hw_shmem_dbg_bufsize, VM_MAP_PAGE_MASK(kernel_map), in shmem_stage_initialize()
/xnu-8020.140.41/osfmk/i386/commpage/
H A Dcommpage.c168 …if (!(kr = vm_map_lookup_entry( kernel_map, vm_map_trunc_page(kernel_addr, VM_MAP_PAGE_MASK(kernel… in commpage_allocate()