| /xnu-10063.141.1/iokit/Tests/ |
| H A D | TestIOMemoryDescriptor.cpp | 88 data = (typeof(data))IOMallocAligned(ptoa(8), page_size); in IOMultMemoryDescriptorTest() 89 for (i = 0; i < ptoa(8); i++) { in IOMultMemoryDescriptorTest() 93 ranges[0].address = (IOVirtualAddress)(data + ptoa(4)); in IOMultMemoryDescriptorTest() 94 ranges[0].length = ptoa(4); in IOMultMemoryDescriptorTest() 95 ranges[1].address = (IOVirtualAddress)(data + ptoa(0)); in IOMultMemoryDescriptorTest() 96 ranges[1].length = ptoa(4); in IOMultMemoryDescriptorTest() 104 assert(ptoa(1) == dmaLen); in IOMultMemoryDescriptorTest() 113 assert(ptoa(2) == dmaLen); in IOMultMemoryDescriptorTest() 122 assert(ptoa(8) == dmaLen); in IOMultMemoryDescriptorTest() 124 mds[1] = IOSubMemoryDescriptor::withSubRange(mds[0], ptoa(3), ptoa(2), kIODirectionOutIn); in IOMultMemoryDescriptorTest() [all …]
|
| H A D | Tests.cpp | 227 round_page(items * size) + ptoa(2), in TestZLib_alloc() 235 return (void *)(uintptr_t) (result + ptoa(1)); in TestZLib_alloc()
|
| /xnu-10063.141.1/iokit/bsddev/skywalk/ |
| H A D | IOSkywalkSupport.cpp | 1834 regions[0] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(1), 2); in IOSkywalkSupportTest() 1837 regions[1] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(2), 3); in IOSkywalkSupportTest() 1839 regions[2] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(3), 4); in IOSkywalkSupportTest() 1854 buffers[0] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[0]); in IOSkywalkSupportTest() 1858 buffers[1] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[1]); in IOSkywalkSupportTest() 1860 buffers[2] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[2]); in IOSkywalkSupportTest() 1862 buffers[3] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[3]); in IOSkywalkSupportTest() 1864 buffers[4] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[4]); in IOSkywalkSupportTest() 1866 buffers[5] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[5]); in IOSkywalkSupportTest() 1898 assert(size == ptoa(20)); in IOSkywalkSupportTest() [all …]
|
| /xnu-10063.141.1/osfmk/kern/ |
| H A D | exclaves_memory.c | 159 ptoa(VM_PAGE_GET_PHYS_PAGE(mem)), in exclaves_memory_alloc() 166 sptm_retype(ptoa(VM_PAGE_GET_PHYS_PAGE(mem)), in exclaves_memory_alloc() 202 m = vm_page_lookup(exclaves_object, ptoa(pages[p])); in exclaves_memory_free() 208 assert3u(sptm_get_frame_type(ptoa(VM_PAGE_GET_PHYS_PAGE(m))), in exclaves_memory_free()
|
| H A D | stack.c | 173 kernel_stack_size + ptoa(2), stack_addr_mask, in stack_alloc_internal() 342 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in stack_collect()
|
| H A D | zalloc.c | 970 return ptoa((int32_t)(meta - zone_info.zi_meta_base)); in zone_meta_to_addr() 1171 page -= ptoa(meta->zm_page_index); in zone_invalid_element_panic() 1208 offs += ptoa(meta->zm_page_index); in zone_element_resolve() 1271 page -= ptoa(meta->zm_page_index); in zone_element_bounds_check_panic() 1315 offs += ptoa(meta->zm_page_index); in zone_element_bounds_check() 1488 offs += ptoa(meta->zm_page_index); in zone_id_require_aligned() 2535 elem_count = ptoa(z->z_percpu ? 1 : z->z_chunk_pages) / in zone_alloc_pages_for_nelems() 3488 zleak_max_zonemap_size = ptoa(zone_pages_wired_max); in zleak_init() 3872 vm_memtag_set_tag(tagged_address + ptoa(index), elem_size); in zone_tag_element() 3950 free_end = (uint32_t)(ptoa(chunk_len) - oob_offs) / elem_size; in zcram_and_lock() [all …]
|
| H A D | zalloc.h | 1189 __zpcpu_cast(ptr, __zpcpu_demangle(ptr) + ptoa((unsigned)(cpu))) 2582 __zpcpu_cast(ptr, __zpcpu_mangle(__zpcpu_addr(ptr) - ptoa(cpu_number()))); \
|
| H A D | zalloc_internal.h | 707 vm_size_t size = ptoa(os_atomic_load(&zone->z_wired_cur, relaxed)); in zone_size_wired()
|
| /xnu-10063.141.1/osfmk/vm/ |
| H A D | vm_kern.c | 837 delta = ptoa(2); 1557 delta = ptoa(2); in kmem_realloc_shrink_guard() 1719 delta = ptoa(2); in kmem_realloc_guard() 2185 delta = ptoa(2); in kmem_free_guard() 3479 vm_map_size_t kmapoff_size = ptoa(kmapoff_pgcnt); in kmem_fuzz_start() 4268 assert3u((map)->size, ==, ptoa(pg)); \ 4277 return verify_write(&zero, (void *)(offs + ptoa(page) + 128), 1) == 0; in can_write_at() 4288 (*(uint32_t *)((offs) + ptoa(page))) 4291 (*(uint32_t *)((offs) + ptoa(page)) = (v)) 4308 addr = kmem_alloc_guard(map, ptoa(10), ptoa(2) - 1, in kmem_alloc_basic_test() [all …]
|
| H A D | vm_fourk_pager.c | 803 kr = kmem_alloc(kernel_map, &kernel_mapping, ptoa(2), in fourk_pager_data_request() 1216 kmem_free(kernel_map, kernel_mapping, ptoa(2)); in fourk_pager_data_request()
|
| H A D | vm_tests.c | 379 kernel_stack_size + ptoa(2), in vm_test_kernel_object_fault() 391 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in vm_test_kernel_object_fault()
|
| /xnu-10063.141.1/tests/ |
| H A D | sysctl_wire_limits.c | 21 ptoa(size_t num_pages) in ptoa() function 158 offset_from_limit = ptoa(current_wired + current_free + wiggle_room_pages); in wire_to_limit()
|
| /xnu-10063.141.1/osfmk/arm/ |
| H A D | caches.c | 261 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page() 440 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page()
|
| /xnu-10063.141.1/osfmk/arm64/sptm/pmap/ |
| H A D | pmap.c | 2162 const pmap_paddr_t pa = ptoa(ppnum); in pmap_verify_free() 2191 const pmap_paddr_t pa = ptoa(ppnum); in pmap_assert_free() 4125 pmap_paddr_t phys = ptoa(ppnum); 4645 pmap_paddr_t phys = ptoa(ppnum); 6481 addr = (unsigned int *) phystokv(ptoa(pn)); 6523 pmap_paddr_t pa = ptoa(pn); 6803 pmap_paddr_t pa = ptoa(pn); 6830 pmap_paddr_t pa = ptoa(pn); 7044 if (ptoa(first) >= vm_last_phys) { 7047 if (ptoa(last) < vm_first_phys) { [all …]
|
| H A D | pmap.h | 263 #define pte_increment_pa(p) ((p) += ptoa(1))
|
| H A D | pmap_data.c | 406 (vm_object_offset_t) ((ptoa(VM_PAGE_GET_PHYS_PAGE(mem))) - gPhysBase); in pmap_enqueue_pages() 513 pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_page_alloc()
|
| /xnu-10063.141.1/osfmk/arm/pmap/ |
| H A D | pmap.c | 1503 const vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai)); in pmap_set_xprr_perm() 2801 const pmap_paddr_t pa = ptoa(ppnum); in pmap_verify_free() 2831 const pmap_paddr_t pa = ptoa(ppnum); in pmap_assert_free() 4047 vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai)); 4689 pmap_paddr_t phys = ptoa(ppnum); 5230 pmap_paddr_t phys = ptoa(ppnum); 7089 addr = (unsigned int *) phystokv(ptoa(pn)); 7131 pmap_paddr_t pa = ptoa(pn); 7371 pmap_paddr_t pa = ptoa(pn); 7411 pmap_paddr_t pa = ptoa(pn); [all …]
|
| H A D | pmap_data.c | 1229 (vm_object_offset_t) ((ptoa(VM_PAGE_GET_PHYS_PAGE(mem))) - gPhysBase); in pmap_enqueue_pages() 1387 *pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_pages_alloc_zeroed() 1442 pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_alloc_page_for_kern() 2630 pmap_flush_noncoherent_page((pmap_paddr_t)ptoa(pai) + vm_first_phys); in pmap_remove_pv()
|
| H A D | pmap.h | 315 #define pte_increment_pa(p) ((p) += ptoa(1))
|
| /xnu-10063.141.1/osfmk/i386/ |
| H A D | mp_desc.c | 556 round_page(size) + ptoa(2), flags, VM_KERN_MEMORY_CPU); in cpu_data_startup_init() 603 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_alloc() 774 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_realloc()
|
| /xnu-10063.141.1/osfmk/mach/ |
| H A D | vm_param.h | 113 #define ptoa(x) ((vm_address_t)(x) << PAGE_SHIFT) macro 116 #define ptoa(x) (0UL = 0) macro
|
| /xnu-10063.141.1/osfmk/mach/i386/ |
| H A D | vm_param.h | 302 #define vm_to_i386(p) (i386_btop(ptoa(p)))
|
| /xnu-10063.141.1/osfmk/console/ |
| H A D | serial_console.c | 200 KERN_CONSOLE_RING_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | in console_init()
|
| /xnu-10063.141.1/osfmk/arm64/ |
| H A D | cpu.c | 758 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc() 766 EXCEPSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc()
|
| /xnu-10063.141.1/bsd/kern/ |
| H A D | subr_log.c | 523 kmem_alloc(kernel_map, &kernel_firehose_addr, size + ptoa(2), in oslog_init_firehose()
|