| /xnu-8019.80.24/osfmk/mach/i386/ |
| H A D | vm_param.h | 310 #define trunc_i386_to_vm(p) (atop(trunc_page(i386_ptob(p)))) 311 #define round_i386_to_vm(p) (atop(round_page(i386_ptob(p))))
|
| /xnu-8019.80.24/osfmk/i386/ |
| H A D | hibernate_i386.c | 85 if (atop(shadow_ptop) > pnmax) { in hibernate_page_list_allocate() 86 pnmax = (ppnum_t)atop(shadow_ptop); in hibernate_page_list_allocate()
|
| /xnu-8019.80.24/osfmk/mach/ |
| H A D | vm_param.h | 112 #define atop(x) ((vm_address_t)(x) >> PAGE_SHIFT) macro 115 #define atop(x) (0UL = 0) macro
|
| /xnu-8019.80.24/osfmk/kern/ |
| H A D | stack.c | 117 uint32_t kernel_stack_pages = atop(KERNEL_STACK_SIZE); in stack_init()
|
| H A D | zalloc.c | 1121 struct zone_page_metadata *to = from + atop(size); in zone_meta_populate() 2184 (&((uint32_t *)zone_tagbase_min)[atop((element) - \ 2477 zone_tagbase_map_size = atop(max_zonemap_size) * sizeof(uint32_t); in zone_tagging_init() 2527 pages = atop(size); in ztMemoryAdd() 2566 pages = atop(size); in ztMemoryRemove() 4296 uint32_t pages = (uint32_t)atop(size); in zone_cram_foreign() 4501 pages = roundup(atop(ZONE_CHUNK_ALLOC_SIZE), chunk_pages); in zone_allocate_va_locked() 4737 min_pages = (uint32_t)atop(round_page(zone_elem_size(z))); in zone_expand_locked() 8058 z->z_chunk_pages = (uint16_t)atop(alloc); 8722 meta_size = round_page(atop(r1.max_address - r0.min_address) * in zone_metadata_init() [all …]
|
| H A D | kern_stackshot.c | 3072 if (!pmap_valid_page((ppnum_t) atop(cur_phys_addr))) { in kdp_find_phys() 3098 if (!pmap_valid_page((ppnum_t) atop(cur_phys_addr))) { in kdp_find_phys() 3115 unsigned int cur_wimg_bits = pmap_cache_attributes((ppnum_t) atop(cur_phys_addr)); in kdp_find_phys()
|
| /xnu-8019.80.24/osfmk/kdp/ |
| H A D | kdp_core.c | 354 ppn = (ppnum_t)atop(avail_end); in pmap_traverse_present_mappings() 357 if (VM_PAGE_GET_PHYS_PAGE(m) >= atop(avail_start)) { in pmap_traverse_present_mappings() 364 if (ppn == atop(avail_end)) { in pmap_traverse_present_mappings()
|
| /xnu-8019.80.24/osfmk/arm/ |
| H A D | pmap.c | 1655 kr = pmap_enter(kernel_pmap, virt, (ppnum_t)atop(start), in pmap_map() 2584 avail_page_count = atop(end - first); in initialize_ram_ranges() 2636 return (unsigned int)atop(avail_end - first_avail); in pmap_free_pages_span() 2669 *pnum = (ppnum_t)atop(first_avail); in pmap_next_page() 4130 unsigned int cacheattr = pmap_cache_attributes((ppnum_t)atop(pa)); 4757 if (__improbable((pmap == NULL) || (atop(pte_to_pa(*pte_p)) != ppnum))) { 4806 assertf(atop(pte_to_pa(spte)) == ppnum, "unexpected value 0x%llx for pte %p mapping ppnum 0x%x", 5969 ppnum_t pn = (ppnum_t)atop(pa); 6635 ppn = atop(pmap_find_pa_nofault(pmap, va)); 6645 ppn = atop(pmap_find_pa(pmap, va)); [all …]
|
| /xnu-8019.80.24/iokit/Kernel/ |
| H A D | IOLib.cpp | 685 alignMask, (ppnum_t) atop(maxPhys), (ppnum_t) atop(alignMask), in IOKernelAllocateWithPhysicalRestrict()
|
| /xnu-8019.80.24/bsd/kern/ |
| H A D | kern_exit.c | 1320 atop(bt[0]) != atop(bt[1]) && /* don't recheck PC page */ in proc_prepareexit() 1321 atop(bt[0]) - 1 != atop(bt[1])) { /* don't recheck page before */ in proc_prepareexit()
|
| /xnu-8019.80.24/osfmk/arm/pmap/ |
| H A D | pmap_data.h | 103 return (unsigned int)atop(pa - vm_first_phys); in pa_index()
|
| H A D | pmap_data.c | 615 const unsigned int npages = (unsigned int)atop(mem_size); in pmap_data_bootstrap() 1044 if (!pmap_verify_free((ppnum_t)atop(pa))) { in pmap_mark_page_as_ppl_page_internal()
|
| /xnu-8019.80.24/osfmk/vm/ |
| H A D | vm_object.c | 4901 size = (unsigned int)atop(object->vo_size); in vm_object_collapse() 4942 if ((int)backing_rcount - (int)(atop(backing_object->vo_size) - size) > (int)rcount) { in vm_object_collapse() 7917 io_upl_reprio_info = kalloc_data(sizeof(uint64_t) * atop(io_upl_size), Z_WAITOK); in vm_decmp_upl_reprioritize() 7933 sizeof(uint64_t) * atop(io_upl_size)); in vm_decmp_upl_reprioritize() 7955 blkno = io_upl_reprio_info[atop(offset)] & UPL_REPRIO_INFO_MASK; in vm_decmp_upl_reprioritize() 7956 len = (io_upl_reprio_info[atop(offset)] >> UPL_REPRIO_INFO_SHIFT) & UPL_REPRIO_INFO_MASK; in vm_decmp_upl_reprioritize() 7988 kfree_data(io_upl_reprio_info, sizeof(uint64_t) * atop(io_upl_size)); in vm_decmp_upl_reprioritize()
|
| H A D | vm_pageout.c | 5232 page_field_size = (atop(size) + 7) >> 3; in upl_create() 5238 upl_size += sizeof(struct upl_page_info) * atop(size); in upl_create() 5271 upl->upl_reprio_info = kalloc_data(sizeof(uint64_t) * atop(size), Z_WAITOK | Z_ZERO); in upl_create()
|
| H A D | vm_resident.c | 1656 vm_page_zone_pages = atop(round_page((vm_offset_t)vm_page_array_zone_data_size)); in vm_page_module_init_delayed() 3029 if (cpm_allocate(PAGE_SIZE, &mem, atop(PPNUM_MAX), 0, FALSE, KMA_LOMEM) != KERN_SUCCESS) { in vm_page_grablo()
|
| H A D | vm_compressor.c | 959 kdp_compressor_decompressed_page_ppnum = (ppnum_t) atop(kdp_compressor_decompressed_page_paddr); in vm_compressor_init()
|
| H A D | vm_map.c | 1200 kentry_initial_pages = (uint16_t)atop(10 * 4096); in vm_map_steal_memory()
|
| /xnu-8019.80.24/iokit/Tests/ |
| H A D | TestIOMemoryDescriptor.cpp | 90 data[i] = ((uint8_t) atop(i)) | 0xD0; in IOMultMemoryDescriptorTest()
|
| /xnu-8019.80.24/osfmk/x86_64/ |
| H A D | pmap.c | 3583 pai = ppn_to_pai(atop(pa)); in pmap_test_text_corruption()
|