Lines Matching refs:ptoa

983 	return ptoa((int32_t)(meta - zone_info.zi_meta_base));  in zone_meta_to_addr()
1184 page -= ptoa(meta->zm_page_index); in zone_invalid_element_panic()
1221 offs += ptoa(meta->zm_page_index); in zone_element_resolve()
1284 page -= ptoa(meta->zm_page_index); in zone_element_bounds_check_panic()
1328 offs += ptoa(meta->zm_page_index); in zone_element_bounds_check()
1501 offs += ptoa(meta->zm_page_index); in zone_id_require_aligned()
2546 elem_count = ptoa(z->z_percpu ? 1 : z->z_chunk_pages) / in zone_alloc_pages_for_nelems()
3499 zleak_max_zonemap_size = ptoa(zone_pages_wired_max); in zleak_init()
3877 vm_memtag_set_tag(tagged_address + ptoa(index), elem_size); in zone_tag_element()
3974 free_end = (uint32_t)(ptoa(chunk_len) - oob_offs) / elem_size; in zcram_and_lock()
4007 free_end = (uint32_t)(ptoa(pg_end) - oob_offs) / elem_size; in zcram_and_lock()
4008 free_start = (uint32_t)(ptoa(pg_start) - oob_offs) / elem_size; in zcram_and_lock()
4016 kasan_poison_range(addr, ptoa(pg_end), ASAN_VALID); in zcram_and_lock()
4019 kasan_zmem_add(addr + ptoa(i), PAGE_SIZE, in zcram_and_lock()
4025 kasan_zmem_add(addr, ptoa(pg_end), in zcram_and_lock()
4090 for (; pages > 0; pages -= chunk_pages, addr += ptoa(chunk_pages)) { in zcram()
4104 assert3u(size % ptoa(zone->z_chunk_pages), ==, 0); in zone_cram_early()
4142 vm_size_t size = ptoa(pages); in zone_submap_alloc_sequestered_va()
4193 kernel_memory_populate(addr, ptoa(pages), in zone_fill_initially()
4197 kmem_alloc(zone_submap(zsflags), &addr, ptoa(pages), in zone_fill_initially()
4201 zone_meta_populate(addr, ptoa(pages)); in zone_fill_initially()
4419 guards = (uint32_t)ptoa(pages) / ZONE_GUARD_SPARSE; in zone_allocate_va_locked()
4420 rest = (uint32_t)ptoa(pages) % ZONE_GUARD_SPARSE; in zone_allocate_va_locked()
4422 } else if (ptoa(chunk_pages) >= ZONE_GUARD_DENSE) { in zone_allocate_va_locked()
4437 guards = (uint32_t)ptoa(pages) / ZONE_GUARD_DENSE; in zone_allocate_va_locked()
4438 rest = (uint32_t)ptoa(pages) % ZONE_GUARD_DENSE; in zone_allocate_va_locked()
4494 ptoa(pages + guards), kmaflags, VM_KERN_MEMORY_ZONE); in zone_allocate_va_locked()
4511 zone_meta_populate(addr, ptoa(pages + guards)); in zone_allocate_va_locked()
4868 addr -= ptoa(cur_pages); in zone_expand_locked()
4892 ZONE_TRACE_VM_KERN_REQUEST_START(ptoa(z->z_chunk_pages - cur_pages)); in zone_expand_locked()
4963 addr + ptoa(cur_pages), addr + ptoa(cur_pages), ptoa(pages), page_list, in zone_expand_locked()
5222 return zone_info.zi_pgz_range.min_address + ptoa(2 * slot + 1); in pgz_addr()
5590 return ptoa(2 * pgz_slots + 1); in pgz_get_size()
5798 vm_offset_t max_size = ptoa(meta->zm_chunk_len) + ZM_ALLOC_SIZE_LOCK; in zfree_drop()
6025 offs += ptoa(meta->zm_page_index); in __zcache_mark_invalid()
6289 vm_memtag_bzero((char *)addr + ptoa(i), esize); in zfree_percpu()
6484 vm_offset_t max_size = ptoa(meta->zm_chunk_len) + ZM_ALLOC_SIZE_LOCK; in zalloc_import()
6532 offs += ptoa(meta->zm_page_index); in __zcache_mark_valid()
7487 size_to_free = ptoa(z->z_chunk_pages); in zone_reclaim_chunk()
7497 size_to_free = ptoa(page_count); in zone_reclaim_chunk()
7555 kasan_zmem_remove(page_addr + ptoa(i), PAGE_SIZE, in zone_reclaim_chunk()
7575 ptoa(z->z_chunk_pages + oob_guard)); in zone_reclaim_chunk()
7883 count = (uint32_t)ptoa(meta->zm_chunk_len) / zone_elem_outer_size(z); in zone_reclaim()
8308 (uintptr_t)ptoa(inuse_ptepages_count)); in panic_display_zprint()
9928 ptoa(z->z_chunk_pages)); in zdestroy()
10315 vm_address_t base = reloc_base + ptoa(i) + zone_elem_inner_offs(z); in zone_metadata_init()
10439 zone_bits_size = round_page(ptoa(zone_pages_wired_max) / in zone_set_map_sizes()