| /xnu-11215.1.10/osfmk/mach/i386/ |
| H A D | vm_param.h | 300 #define trunc_i386_to_vm(p) (atop(trunc_page(i386_ptob(p)))) 301 #define round_i386_to_vm(p) (atop(round_page(i386_ptob(p))))
|
| /xnu-11215.1.10/osfmk/i386/ |
| H A D | hibernate_i386.c | 85 if (atop(shadow_ptop) > pnmax) { in hibernate_page_list_allocate() 86 pnmax = (ppnum_t)atop(shadow_ptop); in hibernate_page_list_allocate()
|
| /xnu-11215.1.10/osfmk/kdp/ |
| H A D | kdp_common.c | 156 if (!pmap_valid_page((ppnum_t) atop(cur_phys_addr))) { in kdp_find_phys() 183 if (!pmap_valid_page((ppnum_t) atop(cur_phys_addr))) { in kdp_find_phys() 198 unsigned int cur_wimg_bits = pmap_cache_attributes((ppnum_t) atop(cur_phys_addr)); in kdp_find_phys()
|
| H A D | kdp_core.c | 470 ppn = (ppnum_t)atop(avail_end); in pmap_traverse_present_mappings() 473 if (VM_PAGE_GET_PHYS_PAGE(m) >= atop(avail_start)) { in pmap_traverse_present_mappings() 480 if (ppn == atop(avail_end)) { in pmap_traverse_present_mappings()
|
| /xnu-11215.1.10/osfmk/mach/ |
| H A D | vm_param.h | 114 #define atop(x) ((vm_address_t)(x) >> PAGE_SHIFT) macro 117 #define atop(x) (0UL = 0) macro
|
| /xnu-11215.1.10/iokit/Kernel/ |
| H A D | IOKitDebug.cpp | 651 hashIdx = atop(memAddr->address) % queue->numSiteQs; in IOTrackingAdd() 689 && (addressIdx != (atop(nextAddress->address) % queue->numSiteQs))) { in IOTrackingRemoveInternal() 732 addressIdx = atop(address) % queue->numSiteQs; in IOTrackingRemoveAddress() 774 hashIdx = atop(address) % queue->numSiteQs; in IOTrackingFree() 787 && (hashIdx != (atop(nextAddress->address) % queue->numSiteQs))) { in IOTrackingFree()
|
| H A D | IOLib.cpp | 777 alignMask, (ppnum_t) atop(maxPhys), (ppnum_t) atop(alignMask), in IOKernelAllocateWithPhysicalRestrict()
|
| /xnu-11215.1.10/osfmk/kern/ |
| H A D | stack.c | 112 uint32_t kernel_stack_pages = atop(KERNEL_STACK_SIZE); in stack_init()
|
| H A D | zalloc_internal.h | 637 return atop(addr ^ (addr + size - 1)) != 0; in zone_addr_size_crosses_page()
|
| H A D | zalloc.c | 1113 struct zone_page_metadata *to = from + atop(size); in zone_meta_populate() 1362 if (oob_offs && meta->zm_guarded && atop(addr ^ end)) { in zone_element_size() 3519 zleak_pages_per_zone_wired_threshold = atop(value); in zleak_update_threshold() 4100 uint32_t pages = (uint32_t)atop(size); in zone_cram_early() 4377 pages = atop(ZONE_CHUNK_ALLOC_SIZE); in zone_allocate_va_locked() 4879 min_pages = (uint32_t)atop(round_page(zone_elem_outer_offs(z) + in zone_expand_locked() 9568 z->z_chunk_pages = (uint16_t)atop(alloc); 10265 atop(early_sz) * sizeof(struct zone_page_metadata)); in zone_metadata_init() 10266 for (uint32_t i = 0; i < atop(early_sz); i++) { in zone_metadata_init() 10281 zba_meta()->zbam_right = (uint32_t)atop(zone_bits_size); in zone_metadata_init() [all …]
|
| H A D | kalloc.c | 2113 return addr | atop(size); in __kalloc_array_encode_vm()
|
| /xnu-11215.1.10/osfmk/vm/ |
| H A D | vm_kern.c | 872 kmr.kmr_return = vm_page_alloc_list(atop(fill_size), flags, 980 atop(fill_size), 0, 0, 0); 1493 vm_page_wire_count -= atop(size); in kernel_memory_populate_object_and_unlock() 1495 vm_page_gobble_count -= atop(size); in kernel_memory_populate_object_and_unlock() 1499 vm_page_wire_count += atop(size); in kernel_memory_populate_object_and_unlock() 1872 kmr.kmr_return = vm_page_alloc_list(atop(newsize - oldsize), in kmem_realloc_guard() 2241 atop(newsize - oldsize), 0, 0, 0); in kmem_realloc_guard() 4070 pte_sz = round_page(atop(va - used) * 8); in kmem_get_gobj_stats()
|
| H A D | vm_upl.c | 489 uint32_t pages = (uint32_t)atop(upl_adjusted_size(upl, PAGE_MASK)); in upl_abort_range() 1267 uint32_t pages = (uint32_t)atop(upl_adjusted_size(upl, PAGE_MASK)); in upl_commit_range()
|
| H A D | vm_object.c | 7884 io_upl_reprio_info = kalloc_data(sizeof(uint64_t) * atop(io_upl_size), Z_WAITOK); 7900 sizeof(uint64_t) * atop(io_upl_size)); 7922 blkno = io_upl_reprio_info[atop(offset)] & UPL_REPRIO_INFO_MASK; 7923 len = (io_upl_reprio_info[atop(offset)] >> UPL_REPRIO_INFO_SHIFT) & UPL_REPRIO_INFO_MASK; 7950 kfree_data(io_upl_reprio_info, sizeof(uint64_t) * atop(io_upl_size));
|
| H A D | vm_compressor.c | 1214 …vm_compressor_kdp_state.kc_panic_decompressed_page_ppnum = (ppnum_t) atop(vm_compressor_kdp_state.… in vm_compressor_init() 1341 …vm_compressor_kdp_state.kc_decompressed_pages_ppnum[i] = (ppnum_t) atop(vm_compressor_kdp_state.kc… in vm_compressor_kdp_init() 6373 uncompressed_file0_free_pages = (uint32_t)atop(size); in vm_uncompressed_put() 6385 uncompressed_file1_free_pages = (uint32_t)atop(size); in vm_uncompressed_put()
|
| H A D | vm_pageout.c | 5545 uint32_t pages = (uint32_t)atop(round_page_32(size)); 5656 pages = (uint32_t)atop(upl_adjusted_size(upl, PAGE_MASK)); 9379 kfree_type(struct upl_page_info, atop(vector_upl->size), vector_upl->pagelist); 9450 atop(vector_upl->size), Z_WAITOK);
|
| /xnu-11215.1.10/osfmk/arm64/sptm/pmap/ |
| H A D | pmap.c | 1354 kr = pmap_enter(kernel_pmap, virt, (ppnum_t)atop(start), in pmap_map() 2078 avail_page_count = atop(end - first); in initialize_ram_ranges() 2101 return (unsigned int)atop(avail_end - first_avail); in pmap_free_pages_span() 2124 *pnum = (ppnum_t)atop(first_avail); in pmap_next_page() 4423 (((spte & ARM_PTE_TYPE_MASK) != ARM_PTE_TYPE_FAULT) && (atop(pte_to_pa(spte)) != ppnum)))) { 5315 pmap_page_protect_options_with_flush_range((ppnum_t)atop(pa), VM_PROT_NONE, 5571 ppnum_t pn = (ppnum_t)atop(pa); 6121 ppn = atop(pmap_find_pa_nofault(pmap, va)); 6131 ppn = atop(pmap_find_pa(pmap, va)); 6673 const ppnum_t pn = (ppnum_t) atop(pa); [all …]
|
| H A D | pmap_data.h | 112 return (unsigned int)atop(pa - vm_first_phys); in pa_index()
|
| H A D | pmap_data.c | 348 unsigned int npages = (unsigned int)atop(mem_size); in pmap_data_bootstrap()
|
| /xnu-11215.1.10/osfmk/arm/pmap/ |
| H A D | pmap.c | 1865 kr = pmap_enter(kernel_pmap, virt, (ppnum_t)atop(start), in pmap_map() 1897 const unsigned int wimg_bits = pmap_cache_attributes((ppnum_t)atop(paddr)); in pmap_construct_io_pte() 2739 avail_page_count = atop(end - first); in initialize_ram_ranges() 2762 return (unsigned int)atop(avail_end - first_avail); in pmap_free_pages_span() 2785 *pnum = (ppnum_t)atop(first_avail); in pmap_next_page() 4214 unsigned int cacheattr = pmap_cache_attributes((ppnum_t)atop(pa)); 4931 if (__improbable((pmap == NULL) || (atop(pte_to_pa(*pte_p)) != ppnum))) { 4994 assertf(atop(pte_to_pa(spte)) == ppnum, "unexpected value 0x%llx for pte %p mapping ppnum 0x%x", 6116 ppnum_t pn = (ppnum_t)atop(pa); 6857 ppn = atop(pmap_find_pa_nofault(pmap, va)); [all …]
|
| H A D | pmap_data.h | 101 return (unsigned int)atop(pa - vm_first_phys); in pa_index()
|
| H A D | pmap_data.c | 634 unsigned int npages = (unsigned int)atop(mem_size); in pmap_data_bootstrap() 1130 if (__improbable(!pmap_verify_free((ppnum_t)atop(pa)))) { in pmap_mark_page_as_ppl_page_internal()
|
| /xnu-11215.1.10/bsd/kern/ |
| H A D | kern_exit.c | 1995 atop(bt[0]) != atop(bt[1]) && /* don't recheck PC page */ in proc_prepareexit() 1996 atop(bt[0]) - 1 != atop(bt[1])) { /* don't recheck page before */ in proc_prepareexit()
|
| /xnu-11215.1.10/iokit/Tests/ |
| H A D | TestIOMemoryDescriptor.cpp | 91 data[i] = ((uint8_t) atop(i)) | 0xD0; in IOMultMemoryDescriptorTest()
|
| /xnu-11215.1.10/osfmk/x86_64/ |
| H A D | pmap.c | 3454 pai = ppn_to_pai(atop(pa)); in pmap_test_text_corruption()
|