Home
last modified time | relevance | path

Searched refs:atop (Results 1 – 25 of 25) sorted by relevance

/xnu-10063.141.1/osfmk/mach/i386/
H A Dvm_param.h300 #define trunc_i386_to_vm(p) (atop(trunc_page(i386_ptob(p))))
301 #define round_i386_to_vm(p) (atop(round_page(i386_ptob(p))))
/xnu-10063.141.1/osfmk/i386/
H A Dhibernate_i386.c85 if (atop(shadow_ptop) > pnmax) { in hibernate_page_list_allocate()
86 pnmax = (ppnum_t)atop(shadow_ptop); in hibernate_page_list_allocate()
/xnu-10063.141.1/osfmk/kdp/
H A Dkdp_common.c156 if (!pmap_valid_page((ppnum_t) atop(cur_phys_addr))) { in kdp_find_phys()
183 if (!pmap_valid_page((ppnum_t) atop(cur_phys_addr))) { in kdp_find_phys()
198 unsigned int cur_wimg_bits = pmap_cache_attributes((ppnum_t) atop(cur_phys_addr)); in kdp_find_phys()
H A Dkdp_core.c472 ppn = (ppnum_t)atop(avail_end); in pmap_traverse_present_mappings()
475 if (VM_PAGE_GET_PHYS_PAGE(m) >= atop(avail_start)) { in pmap_traverse_present_mappings()
482 if (ppn == atop(avail_end)) { in pmap_traverse_present_mappings()
/xnu-10063.141.1/osfmk/mach/
H A Dvm_param.h112 #define atop(x) ((vm_address_t)(x) >> PAGE_SHIFT) macro
115 #define atop(x) (0UL = 0) macro
/xnu-10063.141.1/iokit/Kernel/
H A DIOKitDebug.cpp651 hashIdx = atop(memAddr->address) % queue->numSiteQs; in IOTrackingAdd()
689 && (addressIdx != (atop(nextAddress->address) % queue->numSiteQs))) { in IOTrackingRemoveInternal()
732 addressIdx = atop(address) % queue->numSiteQs; in IOTrackingRemoveAddress()
774 hashIdx = atop(address) % queue->numSiteQs; in IOTrackingFree()
787 && (hashIdx != (atop(nextAddress->address) % queue->numSiteQs))) { in IOTrackingFree()
H A DIOLib.cpp776 alignMask, (ppnum_t) atop(maxPhys), (ppnum_t) atop(alignMask), in IOKernelAllocateWithPhysicalRestrict()
/xnu-10063.141.1/osfmk/kern/
H A Dstack.c112 uint32_t kernel_stack_pages = atop(KERNEL_STACK_SIZE); in stack_init()
H A Dzalloc_internal.h628 return atop(addr ^ (addr + size - 1)) != 0; in zone_addr_size_crosses_page()
H A Dzalloc.c1100 struct zone_page_metadata *to = from + atop(size); in zone_meta_populate()
1349 if (oob_offs && meta->zm_guarded && atop(addr ^ end)) { in zone_element_size()
3508 zleak_pages_per_zone_wired_threshold = atop(value); in zleak_update_threshold()
4080 uint32_t pages = (uint32_t)atop(size); in zone_cram_early()
4357 pages = atop(ZONE_CHUNK_ALLOC_SIZE); in zone_allocate_va_locked()
4859 min_pages = (uint32_t)atop(round_page(zone_elem_outer_offs(z) + in zone_expand_locked()
9512 z->z_chunk_pages = (uint16_t)atop(alloc);
10197 atop(early_sz) * sizeof(struct zone_page_metadata)); in zone_metadata_init()
10198 for (uint32_t i = 0; i < atop(early_sz); i++) { in zone_metadata_init()
10213 zba_meta()->zbam_right = (uint32_t)atop(zone_bits_size); in zone_metadata_init()
[all …]
H A Dkalloc.c2114 return addr | atop(size); in __kalloc_array_encode_vm()
/xnu-10063.141.1/osfmk/vm/
H A Dvm_kern.c871 kmr.kmr_return = vm_page_alloc_list(atop(fill_size), flags,
977 atop(fill_size), 0, 0, 0);
1377 vm_page_wire_count -= atop(size); in kernel_memory_populate_object_and_unlock()
1379 vm_page_gobble_count -= atop(size); in kernel_memory_populate_object_and_unlock()
1383 vm_page_wire_count += atop(size); in kernel_memory_populate_object_and_unlock()
1755 kmr.kmr_return = vm_page_alloc_list(atop(newsize - oldsize), in kmem_realloc_guard()
2117 atop(newsize - oldsize), 0, 0, 0); in kmem_realloc_guard()
3880 pte_sz = round_page(atop(va - used) * 8); in kmem_get_gobj_stats()
H A Dvm_object.c7835 io_upl_reprio_info = kalloc_data(sizeof(uint64_t) * atop(io_upl_size), Z_WAITOK);
7851 sizeof(uint64_t) * atop(io_upl_size));
7873 blkno = io_upl_reprio_info[atop(offset)] & UPL_REPRIO_INFO_MASK;
7874 len = (io_upl_reprio_info[atop(offset)] >> UPL_REPRIO_INFO_SHIFT) & UPL_REPRIO_INFO_MASK;
7906 kfree_data(io_upl_reprio_info, sizeof(uint64_t) * atop(io_upl_size));
H A Dvm_pageout.c5476 uint32_t pages = (uint32_t)atop(round_page_32(size)); in upl_create()
5587 pages = (uint32_t)atop(upl_adjusted_size(upl, PAGE_MASK)); in upl_destroy()
8254 uint32_t pages = (uint32_t)atop(upl_adjusted_size(upl, PAGE_MASK)); in upl_commit_range()
8661 uint32_t pages = (uint32_t)atop(upl_adjusted_size(upl, PAGE_MASK)); in upl_abort_range()
10480 kfree_type(struct upl_page_info, atop(vector_upl->size), vector_upl->pagelist); in vector_upl_deallocate()
10551 atop(vector_upl->size), Z_WAITOK); in vector_upl_set_pagelist()
H A Dvm_compressor.c1173 kdp_compressor_decompressed_page_ppnum = (ppnum_t) atop(kdp_compressor_decompressed_page_paddr); in vm_compressor_init()
6111 uncompressed_file0_free_pages = (uint32_t)atop(size); in vm_uncompressed_put()
6123 uncompressed_file1_free_pages = (uint32_t)atop(size); in vm_uncompressed_put()
H A Dvm_resident.c1651 vm_page_zone_pages = atop(round_page((vm_offset_t)vm_page_array_zone_data_size)); in vm_page_module_init_delayed()
3084 if (cpm_allocate(PAGE_SIZE, &mem, atop(PPNUM_MAX), 0, FALSE, KMA_LOMEM) != KERN_SUCCESS) { in vm_page_grablo()
/xnu-10063.141.1/osfmk/arm64/sptm/pmap/
H A Dpmap.c1311 kr = pmap_enter(kernel_pmap, virt, (ppnum_t)atop(start), in pmap_map()
2039 avail_page_count = atop(end - first); in initialize_ram_ranges()
2061 return (unsigned int)atop(avail_end - first_avail); in pmap_free_pages_span()
2084 *pnum = (ppnum_t)atop(first_avail); in pmap_next_page()
4361 (((spte & ARM_PTE_TYPE_MASK) != ARM_PTE_TYPE_FAULT) && (atop(pte_to_pa(spte)) != ppnum)))) {
5274 pmap_page_protect_options_with_flush_range((ppnum_t)atop(pa), VM_PROT_NONE,
5545 ppnum_t pn = (ppnum_t)atop(pa);
6094 ppn = atop(pmap_find_pa_nofault(pmap, va));
6104 ppn = atop(pmap_find_pa(pmap, va));
6643 const ppnum_t pn = (ppnum_t) atop(pa);
[all …]
H A Dpmap_data.h104 return (unsigned int)atop(pa - vm_first_phys); in pa_index()
H A Dpmap_data.c339 const unsigned int npages = (unsigned int)atop(mem_size); in pmap_data_bootstrap()
/xnu-10063.141.1/osfmk/arm/pmap/
H A Dpmap.c1828 kr = pmap_enter(kernel_pmap, virt, (ppnum_t)atop(start), in pmap_map()
1860 const unsigned int wimg_bits = pmap_cache_attributes((ppnum_t)atop(paddr)); in pmap_construct_io_pte()
2687 avail_page_count = atop(end - first); in initialize_ram_ranges()
2709 return (unsigned int)atop(avail_end - first_avail); in pmap_free_pages_span()
2732 *pnum = (ppnum_t)atop(first_avail); in pmap_next_page()
4176 unsigned int cacheattr = pmap_cache_attributes((ppnum_t)atop(pa));
4827 if (__improbable((pmap == NULL) || (atop(pte_to_pa(*pte_p)) != ppnum))) {
4890 assertf(atop(pte_to_pa(spte)) == ppnum, "unexpected value 0x%llx for pte %p mapping ppnum 0x%x",
6006 ppnum_t pn = (ppnum_t)atop(pa);
6742 ppn = atop(pmap_find_pa_nofault(pmap, va));
[all …]
H A Dpmap_data.h100 return (unsigned int)atop(pa - vm_first_phys); in pa_index()
H A Dpmap_data.c634 const unsigned int npages = (unsigned int)atop(mem_size); in pmap_data_bootstrap()
1129 if (__improbable(!pmap_verify_free((ppnum_t)atop(pa)))) { in pmap_mark_page_as_ppl_page_internal()
/xnu-10063.141.1/bsd/kern/
H A Dkern_exit.c2002 atop(bt[0]) != atop(bt[1]) && /* don't recheck PC page */ in proc_prepareexit()
2003 atop(bt[0]) - 1 != atop(bt[1])) { /* don't recheck page before */ in proc_prepareexit()
/xnu-10063.141.1/iokit/Tests/
H A DTestIOMemoryDescriptor.cpp90 data[i] = ((uint8_t) atop(i)) | 0xD0; in IOMultMemoryDescriptorTest()
/xnu-10063.141.1/osfmk/x86_64/
H A Dpmap.c3454 pai = ppn_to_pai(atop(pa)); in pmap_test_text_corruption()