Lines Matching refs:ptoa

1059 	return ptoa((int32_t)(meta - zone_info.zi_meta_base));  in zone_meta_to_addr()
1283 page -= ptoa(meta->zm_page_index); in zone_invalid_element_panic()
1339 page -= ptoa(meta->zm_page_index); in zone_element_resolve()
2693 tagbase[idx] = block + (uint32_t)((ptoa(idx) + esize - 1) / esize); in ztMemoryAdd()
2856 elem_count = ptoa(z->z_percpu ? 1 : z->z_chunk_pages) / in zone_alloc_pages_for_nelems()
3618 zleak_max_zonemap_size = ptoa(zone_pages_wired_max); in zleak_init()
3988 free_end = (uint32_t)(ptoa(chunk_len) - oob_offs) / elem_size; in zcram_and_lock()
3995 free_end = (uint32_t)(ptoa(pg_end) - oob_offs) / elem_size; in zcram_and_lock()
3996 free_start = (uint32_t)(ptoa(pg_start) - oob_offs) / elem_size; in zcram_and_lock()
4002 ztMemoryAdd(zone, addr + ptoa(pg_start), in zcram_and_lock()
4003 ptoa(pg_end - pg_start)); in zcram_and_lock()
4065 for (; pages > 0; pages -= chunk_pages, addr += ptoa(chunk_pages)) { in zcram()
4078 assert3u(size % ptoa(zone->z_chunk_pages), ==, 0); in zone_cram_early()
4138 vm_size_t size = ptoa(pages); in zone_submap_alloc_sequestered_va()
4189 kernel_memory_populate(addr, ptoa(pages), in zone_fill_initially()
4193 kmem_alloc(zone_submap(zsflags), &addr, ptoa(pages), in zone_fill_initially()
4197 zone_meta_populate(addr, ptoa(pages)); in zone_fill_initially()
4415 guards = (uint32_t)ptoa(pages) / ZONE_GUARD_SPARSE; in zone_allocate_va_locked()
4416 rest = (uint32_t)ptoa(pages) % ZONE_GUARD_SPARSE; in zone_allocate_va_locked()
4418 } else if (ptoa(chunk_pages) >= ZONE_GUARD_DENSE) { in zone_allocate_va_locked()
4433 guards = (uint32_t)ptoa(pages) / ZONE_GUARD_DENSE; in zone_allocate_va_locked()
4434 rest = (uint32_t)ptoa(pages) % ZONE_GUARD_DENSE; in zone_allocate_va_locked()
4490 ptoa(pages + guards), kmaflags, VM_KERN_MEMORY_ZONE); in zone_allocate_va_locked()
4506 zone_meta_populate(addr, ptoa(pages + guards)); in zone_allocate_va_locked()
4727 addr -= ptoa(cur_pages); in zone_expand_locked()
4750 ZONE_TRACE_VM_KERN_REQUEST_START(ptoa(z->z_chunk_pages - cur_pages)); in zone_expand_locked()
4798 addr + ptoa(cur_pages), addr + ptoa(cur_pages), ptoa(pages), page_list, in zone_expand_locked()
5084 return zone_info.zi_pgz_range.min_address + ptoa(2 * slot + 1); in pgz_addr()
5437 return ptoa(2 * pgz_slots + 1); in pgz_get_size()
5700 vm_offset_t max_size = ptoa(meta->zm_chunk_len) + ZM_ALLOC_SIZE_LOCK; in zfree_drop()
5955 kasan_poison_range(elem + ptoa(i), elem_size, in zfree_ext()
6001 bzero((char *)addr + ptoa(i), esize); in zfree_percpu()
6153 vm_offset_t max_size = ptoa(meta->zm_chunk_len) + ZM_ALLOC_SIZE_LOCK; in zalloc_import()
6247 kasan_poison_range(addr + ptoa(i), elem_size, ASAN_VALID); in zalloc_return()
6248 __nosan_bzero((char *)addr + ptoa(i), elem_size); in zalloc_return()
7142 size_to_free = ptoa(z->z_chunk_pages); in zone_reclaim_chunk()
7152 size_to_free = ptoa(page_count); in zone_reclaim_chunk()
7219 ptoa(z->z_chunk_pages + oob_guard)); in zone_reclaim_chunk()
7479 count = (uint32_t)ptoa(meta->zm_chunk_len) / zone_elem_size(z); in zone_reclaim()
7858 (uintptr_t)ptoa(inuse_ptepages_count)); in panic_display_zprint()
9245 ptoa(z->z_chunk_pages)); in zdestroy()
9684 vm_address_t base = reloc_base + ptoa(i); in zone_metadata_init()
9754 zone_map_size = ptoa(zone_pages_wired_max * in zone_set_map_sizes()
9775 zone_bits_size = round_page(16 * (ptoa(zone_pages_wired_max) >> 10)); in zone_set_map_sizes()