Home
last modified time | relevance | path

Searched refs:ptoa (Results 1 – 25 of 36) sorted by relevance

12

/xnu-10063.141.1/iokit/Tests/ !
H A DTestIOMemoryDescriptor.cpp88 data = (typeof(data))IOMallocAligned(ptoa(8), page_size); in IOMultMemoryDescriptorTest()
89 for (i = 0; i < ptoa(8); i++) { in IOMultMemoryDescriptorTest()
93 ranges[0].address = (IOVirtualAddress)(data + ptoa(4)); in IOMultMemoryDescriptorTest()
94 ranges[0].length = ptoa(4); in IOMultMemoryDescriptorTest()
95 ranges[1].address = (IOVirtualAddress)(data + ptoa(0)); in IOMultMemoryDescriptorTest()
96 ranges[1].length = ptoa(4); in IOMultMemoryDescriptorTest()
104 assert(ptoa(1) == dmaLen); in IOMultMemoryDescriptorTest()
113 assert(ptoa(2) == dmaLen); in IOMultMemoryDescriptorTest()
122 assert(ptoa(8) == dmaLen); in IOMultMemoryDescriptorTest()
124 mds[1] = IOSubMemoryDescriptor::withSubRange(mds[0], ptoa(3), ptoa(2), kIODirectionOutIn); in IOMultMemoryDescriptorTest()
[all …]
H A DTests.cpp227 round_page(items * size) + ptoa(2), in TestZLib_alloc()
235 return (void *)(uintptr_t) (result + ptoa(1)); in TestZLib_alloc()
/xnu-10063.141.1/iokit/bsddev/skywalk/ !
H A DIOSkywalkSupport.cpp1834 regions[0] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(1), 2); in IOSkywalkSupportTest()
1837 regions[1] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(2), 3); in IOSkywalkSupportTest()
1839 regions[2] = IOSKRegionCreate(&rspec, (IOSKSize) ptoa(3), 4); in IOSkywalkSupportTest()
1854 buffers[0] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[0]); in IOSkywalkSupportTest()
1858 buffers[1] = IOSKMemoryBufferCreate(ptoa(1), &bspec, &bufkvas[1]); in IOSkywalkSupportTest()
1860 buffers[2] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[2]); in IOSkywalkSupportTest()
1862 buffers[3] = IOSKMemoryBufferCreate(ptoa(2), &bspec, &bufkvas[3]); in IOSkywalkSupportTest()
1864 buffers[4] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[4]); in IOSkywalkSupportTest()
1866 buffers[5] = IOSKMemoryBufferCreate(ptoa(3), &bspec, &bufkvas[5]); in IOSkywalkSupportTest()
1898 assert(size == ptoa(20)); in IOSkywalkSupportTest()
[all …]
/xnu-10063.141.1/osfmk/kern/ !
H A Dexclaves_memory.c159 ptoa(VM_PAGE_GET_PHYS_PAGE(mem)), in exclaves_memory_alloc()
166 sptm_retype(ptoa(VM_PAGE_GET_PHYS_PAGE(mem)), in exclaves_memory_alloc()
202 m = vm_page_lookup(exclaves_object, ptoa(pages[p])); in exclaves_memory_free()
208 assert3u(sptm_get_frame_type(ptoa(VM_PAGE_GET_PHYS_PAGE(m))), in exclaves_memory_free()
H A Dstack.c173 kernel_stack_size + ptoa(2), stack_addr_mask, in stack_alloc_internal()
342 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in stack_collect()
H A Dzalloc.c970 return ptoa((int32_t)(meta - zone_info.zi_meta_base)); in zone_meta_to_addr()
1171 page -= ptoa(meta->zm_page_index); in zone_invalid_element_panic()
1208 offs += ptoa(meta->zm_page_index); in zone_element_resolve()
1271 page -= ptoa(meta->zm_page_index); in zone_element_bounds_check_panic()
1315 offs += ptoa(meta->zm_page_index); in zone_element_bounds_check()
1488 offs += ptoa(meta->zm_page_index); in zone_id_require_aligned()
2535 elem_count = ptoa(z->z_percpu ? 1 : z->z_chunk_pages) / in zone_alloc_pages_for_nelems()
3488 zleak_max_zonemap_size = ptoa(zone_pages_wired_max); in zleak_init()
3872 vm_memtag_set_tag(tagged_address + ptoa(index), elem_size); in zone_tag_element()
3950 free_end = (uint32_t)(ptoa(chunk_len) - oob_offs) / elem_size; in zcram_and_lock()
[all …]
H A Dzalloc.h1189 __zpcpu_cast(ptr, __zpcpu_demangle(ptr) + ptoa((unsigned)(cpu)))
2582 __zpcpu_cast(ptr, __zpcpu_mangle(__zpcpu_addr(ptr) - ptoa(cpu_number()))); \
H A Dzalloc_internal.h707 vm_size_t size = ptoa(os_atomic_load(&zone->z_wired_cur, relaxed)); in zone_size_wired()
/xnu-10063.141.1/osfmk/vm/ !
H A Dvm_kern.c837 delta = ptoa(2);
1557 delta = ptoa(2); in kmem_realloc_shrink_guard()
1719 delta = ptoa(2); in kmem_realloc_guard()
2185 delta = ptoa(2); in kmem_free_guard()
3479 vm_map_size_t kmapoff_size = ptoa(kmapoff_pgcnt); in kmem_fuzz_start()
4268 assert3u((map)->size, ==, ptoa(pg)); \
4277 return verify_write(&zero, (void *)(offs + ptoa(page) + 128), 1) == 0; in can_write_at()
4288 (*(uint32_t *)((offs) + ptoa(page)))
4291 (*(uint32_t *)((offs) + ptoa(page)) = (v))
4308 addr = kmem_alloc_guard(map, ptoa(10), ptoa(2) - 1, in kmem_alloc_basic_test()
[all …]
H A Dvm_fourk_pager.c803 kr = kmem_alloc(kernel_map, &kernel_mapping, ptoa(2), in fourk_pager_data_request()
1216 kmem_free(kernel_map, kernel_mapping, ptoa(2)); in fourk_pager_data_request()
H A Dvm_tests.c379 kernel_stack_size + ptoa(2), in vm_test_kernel_object_fault()
391 kmem_free(kernel_map, stack, kernel_stack_size + ptoa(2)); in vm_test_kernel_object_fault()
/xnu-10063.141.1/tests/ !
H A Dsysctl_wire_limits.c21 ptoa(size_t num_pages) in ptoa() function
158 offset_from_limit = ptoa(current_wired + current_free + wiggle_room_pages); in wire_to_limit()
/xnu-10063.141.1/osfmk/arm/ !
H A Dcaches.c261 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page()
440 pmap_paddr_t paddr = ptoa(pp); in cache_sync_page()
/xnu-10063.141.1/osfmk/arm64/sptm/pmap/ !
H A Dpmap.c2162 const pmap_paddr_t pa = ptoa(ppnum); in pmap_verify_free()
2191 const pmap_paddr_t pa = ptoa(ppnum); in pmap_assert_free()
4125 pmap_paddr_t phys = ptoa(ppnum);
4645 pmap_paddr_t phys = ptoa(ppnum);
6481 addr = (unsigned int *) phystokv(ptoa(pn));
6523 pmap_paddr_t pa = ptoa(pn);
6803 pmap_paddr_t pa = ptoa(pn);
6830 pmap_paddr_t pa = ptoa(pn);
7044 if (ptoa(first) >= vm_last_phys) {
7047 if (ptoa(last) < vm_first_phys) {
[all …]
H A Dpmap.h263 #define pte_increment_pa(p) ((p) += ptoa(1))
H A Dpmap_data.c406 (vm_object_offset_t) ((ptoa(VM_PAGE_GET_PHYS_PAGE(mem))) - gPhysBase); in pmap_enqueue_pages()
513 pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_page_alloc()
/xnu-10063.141.1/osfmk/arm/pmap/ !
H A Dpmap.c1503 const vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai)); in pmap_set_xprr_perm()
2801 const pmap_paddr_t pa = ptoa(ppnum); in pmap_verify_free()
2831 const pmap_paddr_t pa = ptoa(ppnum); in pmap_assert_free()
4047 vm_offset_t kva = phystokv(vm_first_phys + (pmap_paddr_t)ptoa(pai));
4689 pmap_paddr_t phys = ptoa(ppnum);
5230 pmap_paddr_t phys = ptoa(ppnum);
7089 addr = (unsigned int *) phystokv(ptoa(pn));
7131 pmap_paddr_t pa = ptoa(pn);
7371 pmap_paddr_t pa = ptoa(pn);
7411 pmap_paddr_t pa = ptoa(pn);
[all …]
H A Dpmap_data.c1229 (vm_object_offset_t) ((ptoa(VM_PAGE_GET_PHYS_PAGE(mem))) - gPhysBase); in pmap_enqueue_pages()
1387 *pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_pages_alloc_zeroed()
1442 pa = (pmap_paddr_t)ptoa(VM_PAGE_GET_PHYS_PAGE(mem)); in pmap_alloc_page_for_kern()
2630 pmap_flush_noncoherent_page((pmap_paddr_t)ptoa(pai) + vm_first_phys); in pmap_remove_pv()
H A Dpmap.h315 #define pte_increment_pa(p) ((p) += ptoa(1))
/xnu-10063.141.1/osfmk/i386/ !
H A Dmp_desc.c556 round_page(size) + ptoa(2), flags, VM_KERN_MEMORY_CPU); in cpu_data_startup_init()
603 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_alloc()
774 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_data_realloc()
/xnu-10063.141.1/osfmk/mach/ !
H A Dvm_param.h113 #define ptoa(x) ((vm_address_t)(x) << PAGE_SHIFT) macro
116 #define ptoa(x) (0UL = 0) macro
/xnu-10063.141.1/osfmk/mach/i386/ !
H A Dvm_param.h302 #define vm_to_i386(p) (i386_btop(ptoa(p)))
/xnu-10063.141.1/osfmk/console/ !
H A Dserial_console.c200 KERN_CONSOLE_RING_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | in console_init()
/xnu-10063.141.1/osfmk/arm64/ !
H A Dcpu.c758 INTSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc()
766 EXCEPSTACK_SIZE + ptoa(2), KMA_NOFAIL | KMA_PERMANENT | KMA_ZERO | in cpu_stack_alloc()
/xnu-10063.141.1/bsd/kern/ !
H A Dsubr_log.c523 kmem_alloc(kernel_map, &kernel_firehose_addr, size + ptoa(2), in oslog_init_firehose()

12