| /xnu-8020.140.41/iokit/Tests/ |
| H A D | TestIOMemoryDescriptor.cpp | 88 data = (typeof(data))IOMallocAligned(ptoa(8), page_size); in IOMultMemoryDescriptorTest() 89 for (i = 0; i < ptoa(8); i++) { in IOMultMemoryDescriptorTest() 93 ranges[0].address = (IOVirtualAddress)(data + ptoa(4)); in IOMultMemoryDescriptorTest() 94 ranges[0].length = ptoa(4); in IOMultMemoryDescriptorTest() 95 ranges[1].address = (IOVirtualAddress)(data + ptoa(0)); in IOMultMemoryDescriptorTest() 96 ranges[1].length = ptoa(4); in IOMultMemoryDescriptorTest() 104 assert(ptoa(1) == dmaLen); in IOMultMemoryDescriptorTest() 113 assert(ptoa(2) == dmaLen); in IOMultMemoryDescriptorTest() 122 assert(ptoa(8) == dmaLen); in IOMultMemoryDescriptorTest() 124 mds[1] = IOSubMemoryDescriptor::withSubRange(mds[0], ptoa(3), ptoa(2), kIODirectionOutIn); in IOMultMemoryDescriptorTest() [all …]
|
| /xnu-8020.140.41/iokit/bsddev/skywalk/ |
| H A D | IOSkywalkSupport.cpp | 1831 regions[0] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(1), 2); in IOSkywalkSupportTest() 1834 regions[1] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(2), 3); in IOSkywalkSupportTest() 1836 regions[2] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(3), 4); in IOSkywalkSupportTest() 1851 buffers[0] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[0]); in IOSkywalkSupportTest() 1855 buffers[1] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[1]); in IOSkywalkSupportTest() 1857 buffers[2] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[2]); in IOSkywalkSupportTest() 1859 buffers[3] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[3]); in IOSkywalkSupportTest() 1861 buffers[4] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[4]); in IOSkywalkSupportTest() 1863 buffers[5] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[5]); in IOSkywalkSupportTest() 1895 assert(size == ptoa(20)); in IOSkywalkSupportTest() [all …]
|
| /xnu-8020.140.41/osfmk/vm/ |
| H A D | vm_kern.c | 1825 vm_map_size_t kmapoff_size = ptoa(kmapoff_pgcnt); in kmem_fuzz_start() 2528 assert3u((map)->size, ==, ptoa(pg)); \ 2537 return verify_write(&zero, (void *)(offs + ptoa(page) + 128), 1) == 0; in can_write_at() 2548 (*(uint32_t *)((offs) + ptoa(page))) 2551 (*(uint32_t *)((offs) + ptoa(page)) = (v)) 2568 addr = kmem_alloc_guard(map, ptoa(10), ptoa(2) - 1, in kmem_alloc_basic_test() 2571 assert3u((addr + PAGE_SIZE) % ptoa(2), ==, 0); in kmem_alloc_basic_test() 2579 assert3u(addr + ptoa(10), <=, e->vme_end); in kmem_alloc_basic_test() 2588 kmem_free(map, addr, ptoa(10)); in kmem_alloc_basic_test() 2594 addr = kmem_alloc_guard(map, ptoa(10), 0, in kmem_alloc_basic_test() [all …]
|
| H A D | vm_fourk_pager.c | 836 kr = kmem_alloc(kernel_map, &kernel_mapping, ptoa(2), in fourk_pager_data_request() 1248 kmem_free(kernel_map, kernel_mapping, ptoa(2)); in fourk_pager_data_request()
|
| H A D | vm_tests.c | 379 kernel_stack_size + ptoa(2), in vm_test_kernel_object_fault() 391 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in vm_test_kernel_object_fault()
|
| H A D | vm_fault.c | 7352 return ptoa(VM_PAGE_GET_PHYS_PAGE(m)); in kdp_lightweight_fault() 7843 page_paddr = ptoa(VM_PAGE_GET_PHYS_PAGE(page)); in vmtc_get_page_data() 8290 pa = ptoa(VM_PAGE_GET_PHYS_PAGE(page)) + (va - vm_object_trunc_page(va)); in vm_corrupt_text_addr() 8357 …printf("corrupt_text_addr: ptoa(PHYS_PAGE) 0x%llx\n", (uint64_t)ptoa(VM_PAGE_GET_PHYS_PAGE(page))); in vm_corrupt_text_addr()
|
| /xnu-8020.140.41/osfmk/kern/ |
| H A D | zalloc.c | 1059 return ptoa((int32_t)(meta - zone_info.zi_meta_base)); in zone_meta_to_addr() 1283 page -= ptoa(meta->zm_page_index); in zone_invalid_element_panic() 1339 page -= ptoa(meta->zm_page_index); in zone_element_resolve() 2693 tagbase[idx] = block + (uint32_t)((ptoa(idx) + esize - 1) / esize); in ztMemoryAdd() 2856 elem_count = ptoa(z->z_percpu ? 1 : z->z_chunk_pages) / in zone_alloc_pages_for_nelems() 3618 zleak_max_zonemap_size = ptoa(zone_pages_wired_max); in zleak_init() 3988 free_end = (uint32_t)(ptoa(chunk_len) - oob_offs) / elem_size; in zcram_and_lock() 3995 free_end = (uint32_t)(ptoa(pg_end) - oob_offs) / elem_size; in zcram_and_lock() 3996 free_start = (uint32_t)(ptoa(pg_start) - oob_offs) / elem_size; in zcram_and_lock() 4002 ztMemoryAdd(zone, addr + ptoa(pg_start), in zcram_and_lock() [all …]
|
| H A D | stack.c | 178 kernel_stack_size + ptoa(2), stack_addr_mask, in stack_alloc_internal() 347 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in stack_collect()
|
| H A D | zalloc.h | 1053 __zpcpu_cast(ptr, __zpcpu_demangle(ptr) + ptoa((unsigned)cpu)) 1606 __zpcpu_cast(ptr, __zpcpu_mangle(__zpcpu_addr(ptr) - ptoa(cpu_number()))); \
|
| H A D | zalloc_internal.h | 624 vm_size_t size = ptoa(os_atomic_load(&zone->z_wired_cur, relaxed)); in zone_size_wired()
|
| H A D | kalloc.c | 1960 size = round_page(kasan_alloc_resize(req_size) + ptoa(2)); in kalloc_large() 2670 new_size = round_page(kasan_alloc_resize(new_req_size) + ptoa(2)); in krealloc_large() 2841 new_bucket_size + ptoa(old_z ? 0 : 2), kr.size, in krealloc_ext()
|
| /xnu-8020.140.41/tests/ |
| H A D | sysctl_wire_limits.c | 21 ptoa(size_t num_pages) in ptoa() function 158 offset_from_limit = ptoa(current_wired + current_free + wiggle_room_pages); in wire_to_limit()
|
| /xnu-8020.140.41/osfmk/arm/ |
| H A D | pmap.c | 1365 const vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai)); in pmap_set_xprr_perm() 2576 pmap_paddr_t pa = ptoa(ppn); in pmap_is_bad_ram() 2769 const pmap_paddr_t pa = ptoa(ppnum); in pmap_verify_free() 2799 const pmap_paddr_t pa = ptoa(ppnum); in pmap_assert_free() 4056 vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai)); 4691 pmap_paddr_t phys = ptoa(ppnum); 5211 pmap_paddr_t phys = ptoa(ppnum); 7216 addr = (unsigned int *) phystokv(ptoa(pn)); 7258 pmap_paddr_t pa = ptoa(pn); 7489 pmap_paddr_t pa = ptoa(pn); [all …]
|
| H A D | caches.c | 276 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page() 651 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page()
|
| H A D | pmap.h | 336 #define pte_increment_pa(p) ((p) += ptoa(1)) 353 #define pte_increment_pa(p) ((p) += ptoa(1))
|
| /xnu-8020.140.41/osfmk/mach/ |
| H A D | vm_param.h | 113 #define ptoa(x) ((vm_address_t)(x) << PAGE_SHIFT) macro 116 #define ptoa(x) (0UL = 0) macro
|
| /xnu-8020.140.41/osfmk/i386/ |
| H A D | mp_desc.c | 560 round_page(size) + ptoa(2), flags, VM_KERN_MEMORY_CPU); in cpu_data_startup_init() 607 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_alloc() 778 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_realloc()
|
| /xnu-8020.140.41/osfmk/mach/i386/ |
| H A D | vm_param.h | 303 #define vm_to_i386(p) (i386_btop(ptoa(p)))
|
| /xnu-8020.140.41/san/memory/ |
| H A D | kasan-tbi.c | 249 (void)kasan_tbi_tag_range(addr + ptoa(index), elem_size, tag); in kasan_tbi_do_tag_zone_object()
|
| /xnu-8020.140.41/osfmk/console/ |
| H A D | serial_console.c | 194 KERN_CONSOLE_RING_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | in console_init()
|
| /xnu-8020.140.41/osfmk/arm64/ |
| H A D | cpu.c | 757 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc() 765 EXCEPSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc()
|
| /xnu-8020.140.41/osfmk/arm/pmap/ |
| H A D | pmap_data.c | 1148 (vm_object_offset_t) ((ptoa(VM_PAGE_GET_PHYS_PAGE(mem))) - gPhysBase); in pmap_enqueue_pages() 1287 *pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_pages_alloc_zeroed() 1337 pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_alloc_page_for_kern()
|
| /xnu-8020.140.41/libkern/os/ |
| H A D | log.c | 131 if (kmem_alloc(kernel_map, &addr, logmem_size + ptoa(2), in oslog_init_logmem()
|
| /xnu-8020.140.41/bsd/kern/ |
| H A D | subr_log.c | 1191 kmem_alloc(kernel_map, &kernel_firehose_addr, size + ptoa(2), in oslog_init_firehose()
|
| /xnu-8020.140.41/osfmk/kdp/ |
| H A D | kdp_core.c | 381 vcur = phystokv(ptoa(ppn)); in pmap_traverse_present_mappings()
|