Home
last modified time | relevance | path

Searched refs:ptoa_64 (Results 1 – 22 of 22) sorted by relevance

/xnu-11215.1.10/osfmk/arm64/
H A Dhibernate_restore.c186 uint64_t result = ptoa_64(ppnum); in allocate_page()
302 uint64_t header_phys = ptoa_64(headerPpnum); in pal_hib_resume_tramp()
305 uint64_t hib_text_start = ptoa_64(header->restore1CodePhysPage); in pal_hib_resume_tramp()
320 phys_start = ptoa_64(bank_bitmap->first_page); in pal_hib_resume_tramp()
321 phys_end = ptoa_64(bank_bitmap->last_page) + PAGE_SIZE; in pal_hib_resume_tramp()
352 uint64_t size = ptoa_64(seg_info->segments[i].pageCount); in pal_hib_resume_tramp()
354 uint64_t seg_start = ptoa_64(seg_info->segments[i].physPage); in pal_hib_resume_tramp()
388 uint64_t bank_start = ptoa_64(bank_bitmap->first_page); in pal_hib_resume_tramp()
389 uint64_t bank_end = ptoa_64(bank_bitmap->last_page) + PAGE_SIZE; in pal_hib_resume_tramp()
412 image_start = ptoa_64(header->handoffPages); in pal_hib_resume_tramp()
[all …]
H A Dhibernate_arm64.c138 ptoa_64(bitmap->first_page), bitmap->first_page, in hibernate_page_list_allocate()
139 ptoa_64(bitmap->last_page), bitmap->last_page); in hibernate_page_list_allocate()
/xnu-11215.1.10/iokit/Kernel/
H A DIOMapper.cpp243 NULL, 0, ptoa_64(pages), in IOMapperIOVMAlloc()
264 IOMapper::gSystem->iovmUnmapMemory(NULL, NULL, ptoa_64(addr), ptoa_64(pages)); in IOMapperIOVMFree()
278 ptoa_64(addr), ptoa_64(offset), ptoa_64(page), ptoa_64(1)); in IOMapperInsertPage()
H A DIOHibernateRestoreKernel.c460 scratch->curPage = (uint8_t *)pal_hib_map(SCRATCH_AREA, ptoa_64(scratch->headPage)); in hibernate_scratch_init()
470 result.curPage = (uint8_t *)pal_hib_map(SCRATCH_AREA, ptoa_64(result.headPage)); in hibernate_scratch_start_read()
491 scratch->curPage = (uint8_t *)pal_hib_map(SCRATCH_AREA, ptoa_64(*nextPage)); in hibernate_scratch_io()
528 uint64_t dst = ptoa_64(ppnum); in store_one_page()
615 headerPhys = ptoa_64(p1); in hibernate_kernel_entrypoint()
650 debug_code(kIOHibernateRestoreCodeImageEnd, ptoa_64(lastImagePage)); in hibernate_kernel_entrypoint()
652 debug_code(kIOHibernateRestoreCodeMapEnd, ptoa_64(lastMapPage)); in hibernate_kernel_entrypoint()
655 debug_code(kIOHibernateRestoreCodeHandoffPages, ptoa_64(handoffPages)); in hibernate_kernel_entrypoint()
670 wkdmScratch = (uint8_t *)pal_hib_map(WKDM_AREA, ptoa_64(hibernate_page_list_grab(map, &nextFree))); in hibernate_kernel_entrypoint()
700 debug_code(kIOHibernateRestoreCodePageIndexEnd, ptoa_64(lastPageIndexPage)); in hibernate_kernel_entrypoint()
[all …]
H A DIOHibernateIO.cpp294 memcpy(&copyDest[i * PAGE_SIZE], (void *)phystokv(ptoa_64(page_array[i])), PAGE_SIZE); in HibernationCopyHandoffRegionFromPageArray()
537 ptoa_64(gIOHibernateHandoffPageCount), page_size); in IOHibernateSystemSleep()
595 setFileSize = ((ptoa_64((106 * pageCount) / 100) * gIOHibernateCompression) >> 8) in IOHibernateSystemSleep()
1971 uint64_t physAddr = ptoa_64(page); in hibernate_write_image()
1972 uint64_t size = ptoa_64(count); in hibernate_write_image()
2090 err = IOMemoryDescriptorWriteFromPhysical(vars->srcBuffer, 0, ptoa_64(page), page_size); in hibernate_write_image()
2670 err = IOMemoryDescriptorReadToPhysical(vars->srcBuffer, decoOffset, ptoa_64(ppnum), page_size); in hibernate_machine_init()
2795 element = (element & page_mask) | ptoa_64(pmap_find_phys(kernel_pmap, element)); in IOHibernateSystemRestart()
H A DIOMemoryDescriptor.cpp741 vm_size_t size = ptoa_64(_pages); in memoryReferenceCreate()
3283 address = trunc_page_64(mappedBase) + ptoa_64(ioplInfo.fMappedPage) + offset; in dmaCommandOperation()
3301 address = ptoa_64(pageList->phys_addr) + offset; in dmaCommandOperation()
3315 address = ptoa_64(pageAddr) + offset; in dmaCommandOperation()
4056 *upl_size = (upl_size_t) ptoa_64(pageCount); in io_get_kernel_static_upl()
4062 phys = pmap_find_phys(kernel_pmap, ((addr64_t)offset) + ptoa_64(page)); in io_get_kernel_static_upl()
4357 ioplSize = ((typeof(ioplSize))ptoa_64(numPageInfo)); in wireVirtual()
4625 if ((_length == ptoa_64(_pages)) && !(page_mask & ioplList->fPageOffset)) { in dmaMap()
H A DIODMACommand.cpp587 remapAddr = ptoa_64(vm_page_get_phys_page(lastPage)); in segmentOp()
1513 curSeg.fIOVMAddr = ptoa_64(vm_page_get_phys_page(remap)) in genIOVMSegments()
H A DIOLib.cpp1474 ret = IOMapPages( get_task_map(task), address, ptoa_64(pagenum), page_size, cacheMode ); in IOSetProcessorCacheMode()
/xnu-11215.1.10/osfmk/vm/
H A Dvm_purgeable.c1533 ptoa_64(resident_page_count - wired_page_count)); in vm_purgeable_accounting()
1537 ptoa_64(compressed_page_count)); in vm_purgeable_accounting()
1542 ptoa_64(resident_page_count - wired_page_count)); in vm_purgeable_accounting()
1546 ptoa_64(compressed_page_count)); in vm_purgeable_accounting()
1551 ptoa_64(resident_page_count in vm_purgeable_accounting()
1557 ptoa_64(resident_page_count in vm_purgeable_accounting()
1565 ptoa_64(resident_page_count - wired_page_count)); in vm_purgeable_accounting()
1569 ptoa_64(compressed_page_count)); in vm_purgeable_accounting()
1574 ptoa_64(resident_page_count in vm_purgeable_accounting()
1580 ptoa_64(resident_page_count in vm_purgeable_accounting()
[all …]
H A Dvm_object_xnu.h471 vm_tag_update_size((object)->wire_tag, -ptoa_64((object)->wired_page_count), (object)); \
499 vm_tag_update_size((object)->wire_tag, ptoa_64(__wireddelta), (object)); \
502 vm_tag_update_size((object)->wire_tag, ptoa_64(__wireddelta), (object)); \
H A Dvm_object.c2944 if (ptoa_64(object->resident_page_count) > size_in_object / 2 && pmap != PMAP_NULL) {
2966 if (ptoa_64(object->resident_page_count / 4) < size_in_object) {
8605 ptoa_64(resident_count));
8612 ptoa_64(resident_count));
8616 ptoa_64(resident_count));
8624 ptoa_64(wired_count));
8628 ptoa_64(wired_count));
8632 ptoa_64(wired_count));
8642 ptoa_64(compressed_count));
8648 ptoa_64(compressed_count));
[all …]
H A Dvm_resident.c7053 if ((flags & KMA_NOFAIL) == 0 && ptoa_64(page_size) > max_mem / 4) { in vm_page_alloc_list()
7054 uint64_t unavailable = ptoa_64(vm_page_wire_count + vm_page_free_target); in vm_page_alloc_list()
7055 if (unavailable > max_mem || ptoa_64(page_count) > (max_mem - unavailable)) { in vm_page_alloc_list()
8037 …force_to_wired_list = hibernate_sptm_should_force_page_to_wired_pagelist(ptoa_64(VM_PAGE_GET_PHYS_… in hibernate_page_list_setall()
8116 …force_to_wired_list = hibernate_sptm_should_force_page_to_wired_pagelist(ptoa_64(VM_PAGE_GET_PHYS_… in hibernate_page_list_setall()
8165 …force_to_wired_list = hibernate_sptm_should_force_page_to_wired_pagelist(ptoa_64(VM_PAGE_GET_PHYS_… in hibernate_page_list_setall()
9784 info[object->wire_tag].size += ptoa_64(object->wired_page_count); in vm_page_count_object()
10108 wired_size = ptoa_64(vm_page_wire_count); in vm_page_diagnose()
10109 wired_reserved_size = ptoa_64(vm_page_wire_count_initial - vm_page_stolen_count); in vm_page_diagnose()
10111 wired_size = ptoa_64(vm_page_wire_count + vm_lopage_free_count + vm_page_throttled_count); in vm_page_diagnose()
[all …]
H A Dvm_kern.c1608 vm_tag_update_size(tag, -ptoa_64(pages_unwired), NULL); in kernel_memory_depopulate()
H A Dvm_compressor.c850 compressor_pool_size = ptoa_64(vm_compression_limit); in vm_compressor_set_size()
H A Dvm_fault.c6878 vm_tag_update_size((vm_tag_t) fault_info.user_tag, -ptoa_64(unwired_pages), NULL); in vm_fault_unwire()
H A Dvm_map.c6089 size + ptoa_64(total_wire_count) > vm_global_user_wire_limit) { in add_wire_counts()
6090 if (size + ptoa_64(total_wire_count) > vm_global_user_wire_limit) { in add_wire_counts()
6093 …bal wire limit. %llu bytes wired and requested to wire %llu bytes more", ptoa_64(total_wire_count)… in add_wire_counts()
/xnu-11215.1.10/osfmk/i386/vmx/
H A Dvmx_shims.c53 return ptoa_64(pmap_find_phys(kernel_pmap, (addr64_t)(uintptr_t)va)); in vmx_paddr()
/xnu-11215.1.10/osfmk/mach/
H A Dvm_param.h102 #define ptoa_64(x) ((uint64_t)(x) << PAGE_SHIFT) macro
190 #undef ptoa_64
225 #define ptoa_64(x) ((uint64_t)((x) + (uint8_t *)0))
/xnu-11215.1.10/osfmk/kern/
H A Dzalloc.c5069 *current_size = ptoa_64(phys_pages); in get_zone_map_size()
5070 *capacity = ptoa_64(zone_pages_wired_max); in get_zone_map_size()
5115 ptoa_64(os_atomic_load(&zone_pages_wired, relaxed)), in kill_process_in_largest_zone()
5116 ptoa_64(zone_pages_wired_max), in kill_process_in_largest_zone()
8555 .mzi_cur_size = ptoa_64(zone_scale_for_percpu(&zcopy, zcopy.z_wired_cur)), in get_zone_info()
8557 .mzi_max_size = ptoa_64(zone_scale_for_percpu(&zcopy, zcopy.z_wired_hwm)), in get_zone_info()
8559 .mzi_alloc_size = ptoa_64(zcopy.z_chunk_pages), in get_zone_info()
8571 ptoa_64(zone_scale_for_percpu(&zcopy, zcopy.z_wired_empty))); in get_zone_info()
/xnu-11215.1.10/osfmk/i386/AT386/
H A Dmodel_dep.c1598 uint64_t phys_src = ptoa_64(upn) | (uaddr & PAGE_MASK); in debug_copyin()
/xnu-11215.1.10/osfmk/kdp/
H A Dkdp_core.c413 uint64_t phys = ptoa_64(ppn); in kernel_pmap_present_mapping()
/xnu-11215.1.10/bsd/kern/
H A Duipc_mbuf.c7454 return (uint64_t)(ptoa_64(base_phys) | ((uint64_t)addr & PAGE_MASK));