Home
last modified time | relevance | path

Searched refs:ptoa (Results 1 – 25 of 30) sorted by relevance

12

/xnu-8020.140.41/iokit/Tests/
H A DTestIOMemoryDescriptor.cpp88 data = (typeof(data))IOMallocAligned(ptoa(8), page_size); in IOMultMemoryDescriptorTest()
89 for (i = 0; i < ptoa(8); i++) { in IOMultMemoryDescriptorTest()
93 ranges[0].address = (IOVirtualAddress)(data + ptoa(4)); in IOMultMemoryDescriptorTest()
94 ranges[0].length = ptoa(4); in IOMultMemoryDescriptorTest()
95 ranges[1].address = (IOVirtualAddress)(data + ptoa(0)); in IOMultMemoryDescriptorTest()
96 ranges[1].length = ptoa(4); in IOMultMemoryDescriptorTest()
104 assert(ptoa(1) == dmaLen); in IOMultMemoryDescriptorTest()
113 assert(ptoa(2) == dmaLen); in IOMultMemoryDescriptorTest()
122 assert(ptoa(8) == dmaLen); in IOMultMemoryDescriptorTest()
124 mds[1] = IOSubMemoryDescriptor::withSubRange(mds[0], ptoa(3), ptoa(2), kIODirectionOutIn); in IOMultMemoryDescriptorTest()
[all …]
/xnu-8020.140.41/iokit/bsddev/skywalk/
H A DIOSkywalkSupport.cpp1831 regions[0] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(1), 2); in IOSkywalkSupportTest()
1834 regions[1] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(2), 3); in IOSkywalkSupportTest()
1836 regions[2] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(3), 4); in IOSkywalkSupportTest()
1851 buffers[0] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[0]); in IOSkywalkSupportTest()
1855 buffers[1] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[1]); in IOSkywalkSupportTest()
1857 buffers[2] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[2]); in IOSkywalkSupportTest()
1859 buffers[3] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[3]); in IOSkywalkSupportTest()
1861 buffers[4] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[4]); in IOSkywalkSupportTest()
1863 buffers[5] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[5]); in IOSkywalkSupportTest()
1895 assert(size == ptoa(20)); in IOSkywalkSupportTest()
[all …]
/xnu-8020.140.41/osfmk/vm/
H A Dvm_kern.c1825 vm_map_size_t kmapoff_size = ptoa(kmapoff_pgcnt); in kmem_fuzz_start()
2528 assert3u((map)->size, ==, ptoa(pg)); \
2537 return verify_write(&zero, (void *)(offs + ptoa(page) + 128), 1) == 0; in can_write_at()
2548 (*(uint32_t *)((offs) + ptoa(page)))
2551 (*(uint32_t *)((offs) + ptoa(page)) = (v))
2568 addr = kmem_alloc_guard(map, ptoa(10), ptoa(2) - 1, in kmem_alloc_basic_test()
2571 assert3u((addr + PAGE_SIZE) % ptoa(2), ==, 0); in kmem_alloc_basic_test()
2579 assert3u(addr + ptoa(10), <=, e->vme_end); in kmem_alloc_basic_test()
2588 kmem_free(map, addr, ptoa(10)); in kmem_alloc_basic_test()
2594 addr = kmem_alloc_guard(map, ptoa(10), 0, in kmem_alloc_basic_test()
[all …]
H A Dvm_fourk_pager.c836 kr = kmem_alloc(kernel_map, &kernel_mapping, ptoa(2), in fourk_pager_data_request()
1248 kmem_free(kernel_map, kernel_mapping, ptoa(2)); in fourk_pager_data_request()
H A Dvm_tests.c379 kernel_stack_size + ptoa(2), in vm_test_kernel_object_fault()
391 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in vm_test_kernel_object_fault()
H A Dvm_fault.c7352 return ptoa(VM_PAGE_GET_PHYS_PAGE(m)); in kdp_lightweight_fault()
7843 page_paddr = ptoa(VM_PAGE_GET_PHYS_PAGE(page)); in vmtc_get_page_data()
8290 pa = ptoa(VM_PAGE_GET_PHYS_PAGE(page)) + (va - vm_object_trunc_page(va)); in vm_corrupt_text_addr()
8357 …printf("corrupt_text_addr: ptoa(PHYS_PAGE) 0x%llx\n", (uint64_t)ptoa(VM_PAGE_GET_PHYS_PAGE(page))); in vm_corrupt_text_addr()
/xnu-8020.140.41/osfmk/kern/
H A Dzalloc.c1059 return ptoa((int32_t)(meta - zone_info.zi_meta_base)); in zone_meta_to_addr()
1283 page -= ptoa(meta->zm_page_index); in zone_invalid_element_panic()
1339 page -= ptoa(meta->zm_page_index); in zone_element_resolve()
2693 tagbase[idx] = block + (uint32_t)((ptoa(idx) + esize - 1) / esize); in ztMemoryAdd()
2856 elem_count = ptoa(z->z_percpu ? 1 : z->z_chunk_pages) / in zone_alloc_pages_for_nelems()
3618 zleak_max_zonemap_size = ptoa(zone_pages_wired_max); in zleak_init()
3988 free_end = (uint32_t)(ptoa(chunk_len) - oob_offs) / elem_size; in zcram_and_lock()
3995 free_end = (uint32_t)(ptoa(pg_end) - oob_offs) / elem_size; in zcram_and_lock()
3996 free_start = (uint32_t)(ptoa(pg_start) - oob_offs) / elem_size; in zcram_and_lock()
4002 ztMemoryAdd(zone, addr + ptoa(pg_start), in zcram_and_lock()
[all …]
H A Dstack.c178 kernel_stack_size + ptoa(2), stack_addr_mask, in stack_alloc_internal()
347 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in stack_collect()
H A Dzalloc.h1053 __zpcpu_cast(ptr, __zpcpu_demangle(ptr) + ptoa((unsigned)cpu))
1606 __zpcpu_cast(ptr, __zpcpu_mangle(__zpcpu_addr(ptr) - ptoa(cpu_number()))); \
H A Dzalloc_internal.h624 vm_size_t size = ptoa(os_atomic_load(&zone->z_wired_cur, relaxed)); in zone_size_wired()
H A Dkalloc.c1960 size = round_page(kasan_alloc_resize(req_size) + ptoa(2)); in kalloc_large()
2670 new_size = round_page(kasan_alloc_resize(new_req_size) + ptoa(2)); in krealloc_large()
2841 new_bucket_size + ptoa(old_z ? 0 : 2), kr.size, in krealloc_ext()
/xnu-8020.140.41/tests/
H A Dsysctl_wire_limits.c21 ptoa(size_t num_pages) in ptoa() function
158 offset_from_limit = ptoa(current_wired + current_free + wiggle_room_pages); in wire_to_limit()
/xnu-8020.140.41/osfmk/arm/
H A Dpmap.c1365 const vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai)); in pmap_set_xprr_perm()
2576 pmap_paddr_t pa = ptoa(ppn); in pmap_is_bad_ram()
2769 const pmap_paddr_t pa = ptoa(ppnum); in pmap_verify_free()
2799 const pmap_paddr_t pa = ptoa(ppnum); in pmap_assert_free()
4056 vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai));
4691 pmap_paddr_t phys = ptoa(ppnum);
5211 pmap_paddr_t phys = ptoa(ppnum);
7216 addr = (unsigned int *) phystokv(ptoa(pn));
7258 pmap_paddr_t pa = ptoa(pn);
7489 pmap_paddr_t pa = ptoa(pn);
[all …]
H A Dcaches.c276 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page()
651 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page()
H A Dpmap.h336 #define pte_increment_pa(p) ((p) += ptoa(1))
353 #define pte_increment_pa(p) ((p) += ptoa(1))
/xnu-8020.140.41/osfmk/mach/
H A Dvm_param.h113 #define ptoa(x) ((vm_address_t)(x) << PAGE_SHIFT) macro
116 #define ptoa(x) (0UL = 0) macro
/xnu-8020.140.41/osfmk/i386/
H A Dmp_desc.c560 round_page(size) + ptoa(2), flags, VM_KERN_MEMORY_CPU); in cpu_data_startup_init()
607 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_alloc()
778 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_realloc()
/xnu-8020.140.41/osfmk/mach/i386/
H A Dvm_param.h303 #define vm_to_i386(p) (i386_btop(ptoa(p)))
/xnu-8020.140.41/san/memory/
H A Dkasan-tbi.c249 (void)kasan_tbi_tag_range(addr + ptoa(index), elem_size, tag); in kasan_tbi_do_tag_zone_object()
/xnu-8020.140.41/osfmk/console/
H A Dserial_console.c194 KERN_CONSOLE_RING_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | in console_init()
/xnu-8020.140.41/osfmk/arm64/
H A Dcpu.c757 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc()
765 EXCEPSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc()
/xnu-8020.140.41/osfmk/arm/pmap/
H A Dpmap_data.c1148 (vm_object_offset_t) ((ptoa(VM_PAGE_GET_PHYS_PAGE(mem))) - gPhysBase); in pmap_enqueue_pages()
1287 *pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_pages_alloc_zeroed()
1337 pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_alloc_page_for_kern()
/xnu-8020.140.41/libkern/os/
H A Dlog.c131 if (kmem_alloc(kernel_map, &addr, logmem_size + ptoa(2), in oslog_init_logmem()
/xnu-8020.140.41/bsd/kern/
H A Dsubr_log.c1191 kmem_alloc(kernel_map, &kernel_firehose_addr, size + ptoa(2), in oslog_init_firehose()
/xnu-8020.140.41/osfmk/kdp/
H A Dkdp_core.c381 vcur = phystokv(ptoa(ppn)); in pmap_traverse_present_mappings()

12