| /xnu-11215.1.10/osfmk/arm64/ |
| H A D | hibernate_restore.c | 186 uint64_t result = ptoa_64(ppnum); in allocate_page() 302 uint64_t header_phys = ptoa_64(headerPpnum); in pal_hib_resume_tramp() 305 uint64_t hib_text_start = ptoa_64(header->restore1CodePhysPage); in pal_hib_resume_tramp() 320 phys_start = ptoa_64(bank_bitmap->first_page); in pal_hib_resume_tramp() 321 phys_end = ptoa_64(bank_bitmap->last_page) + PAGE_SIZE; in pal_hib_resume_tramp() 352 uint64_t size = ptoa_64(seg_info->segments[i].pageCount); in pal_hib_resume_tramp() 354 uint64_t seg_start = ptoa_64(seg_info->segments[i].physPage); in pal_hib_resume_tramp() 388 uint64_t bank_start = ptoa_64(bank_bitmap->first_page); in pal_hib_resume_tramp() 389 uint64_t bank_end = ptoa_64(bank_bitmap->last_page) + PAGE_SIZE; in pal_hib_resume_tramp() 412 image_start = ptoa_64(header->handoffPages); in pal_hib_resume_tramp() [all …]
|
| H A D | hibernate_arm64.c | 138 ptoa_64(bitmap->first_page), bitmap->first_page, in hibernate_page_list_allocate() 139 ptoa_64(bitmap->last_page), bitmap->last_page); in hibernate_page_list_allocate()
|
| /xnu-11215.1.10/iokit/Kernel/ |
| H A D | IOMapper.cpp | 243 NULL, 0, ptoa_64(pages), in IOMapperIOVMAlloc() 264 IOMapper::gSystem->iovmUnmapMemory(NULL, NULL, ptoa_64(addr), ptoa_64(pages)); in IOMapperIOVMFree() 278 ptoa_64(addr), ptoa_64(offset), ptoa_64(page), ptoa_64(1)); in IOMapperInsertPage()
|
| H A D | IOHibernateRestoreKernel.c | 460 scratch->curPage = (uint8_t *)pal_hib_map(SCRATCH_AREA, ptoa_64(scratch->headPage)); in hibernate_scratch_init() 470 result.curPage = (uint8_t *)pal_hib_map(SCRATCH_AREA, ptoa_64(result.headPage)); in hibernate_scratch_start_read() 491 scratch->curPage = (uint8_t *)pal_hib_map(SCRATCH_AREA, ptoa_64(*nextPage)); in hibernate_scratch_io() 528 uint64_t dst = ptoa_64(ppnum); in store_one_page() 615 headerPhys = ptoa_64(p1); in hibernate_kernel_entrypoint() 650 debug_code(kIOHibernateRestoreCodeImageEnd, ptoa_64(lastImagePage)); in hibernate_kernel_entrypoint() 652 debug_code(kIOHibernateRestoreCodeMapEnd, ptoa_64(lastMapPage)); in hibernate_kernel_entrypoint() 655 debug_code(kIOHibernateRestoreCodeHandoffPages, ptoa_64(handoffPages)); in hibernate_kernel_entrypoint() 670 wkdmScratch = (uint8_t *)pal_hib_map(WKDM_AREA, ptoa_64(hibernate_page_list_grab(map, &nextFree))); in hibernate_kernel_entrypoint() 700 debug_code(kIOHibernateRestoreCodePageIndexEnd, ptoa_64(lastPageIndexPage)); in hibernate_kernel_entrypoint() [all …]
|
| H A D | IOHibernateIO.cpp | 294 memcpy(©Dest[i * PAGE_SIZE], (void *)phystokv(ptoa_64(page_array[i])), PAGE_SIZE); in HibernationCopyHandoffRegionFromPageArray() 537 ptoa_64(gIOHibernateHandoffPageCount), page_size); in IOHibernateSystemSleep() 595 setFileSize = ((ptoa_64((106 * pageCount) / 100) * gIOHibernateCompression) >> 8) in IOHibernateSystemSleep() 1971 uint64_t physAddr = ptoa_64(page); in hibernate_write_image() 1972 uint64_t size = ptoa_64(count); in hibernate_write_image() 2090 err = IOMemoryDescriptorWriteFromPhysical(vars->srcBuffer, 0, ptoa_64(page), page_size); in hibernate_write_image() 2670 err = IOMemoryDescriptorReadToPhysical(vars->srcBuffer, decoOffset, ptoa_64(ppnum), page_size); in hibernate_machine_init() 2795 element = (element & page_mask) | ptoa_64(pmap_find_phys(kernel_pmap, element)); in IOHibernateSystemRestart()
|
| H A D | IOMemoryDescriptor.cpp | 741 vm_size_t size = ptoa_64(_pages); in memoryReferenceCreate() 3283 address = trunc_page_64(mappedBase) + ptoa_64(ioplInfo.fMappedPage) + offset; in dmaCommandOperation() 3301 address = ptoa_64(pageList->phys_addr) + offset; in dmaCommandOperation() 3315 address = ptoa_64(pageAddr) + offset; in dmaCommandOperation() 4056 *upl_size = (upl_size_t) ptoa_64(pageCount); in io_get_kernel_static_upl() 4062 phys = pmap_find_phys(kernel_pmap, ((addr64_t)offset) + ptoa_64(page)); in io_get_kernel_static_upl() 4357 ioplSize = ((typeof(ioplSize))ptoa_64(numPageInfo)); in wireVirtual() 4625 if ((_length == ptoa_64(_pages)) && !(page_mask & ioplList->fPageOffset)) { in dmaMap()
|
| H A D | IODMACommand.cpp | 587 remapAddr = ptoa_64(vm_page_get_phys_page(lastPage)); in segmentOp() 1513 curSeg.fIOVMAddr = ptoa_64(vm_page_get_phys_page(remap)) in genIOVMSegments()
|
| H A D | IOLib.cpp | 1474 ret = IOMapPages( get_task_map(task), address, ptoa_64(pagenum), page_size, cacheMode ); in IOSetProcessorCacheMode()
|
| /xnu-11215.1.10/osfmk/vm/ |
| H A D | vm_purgeable.c | 1533 ptoa_64(resident_page_count - wired_page_count)); in vm_purgeable_accounting() 1537 ptoa_64(compressed_page_count)); in vm_purgeable_accounting() 1542 ptoa_64(resident_page_count - wired_page_count)); in vm_purgeable_accounting() 1546 ptoa_64(compressed_page_count)); in vm_purgeable_accounting() 1551 ptoa_64(resident_page_count in vm_purgeable_accounting() 1557 ptoa_64(resident_page_count in vm_purgeable_accounting() 1565 ptoa_64(resident_page_count - wired_page_count)); in vm_purgeable_accounting() 1569 ptoa_64(compressed_page_count)); in vm_purgeable_accounting() 1574 ptoa_64(resident_page_count in vm_purgeable_accounting() 1580 ptoa_64(resident_page_count in vm_purgeable_accounting() [all …]
|
| H A D | vm_object_xnu.h | 471 vm_tag_update_size((object)->wire_tag, -ptoa_64((object)->wired_page_count), (object)); \ 499 vm_tag_update_size((object)->wire_tag, ptoa_64(__wireddelta), (object)); \ 502 vm_tag_update_size((object)->wire_tag, ptoa_64(__wireddelta), (object)); \
|
| H A D | vm_object.c | 2944 if (ptoa_64(object->resident_page_count) > size_in_object / 2 && pmap != PMAP_NULL) { 2966 if (ptoa_64(object->resident_page_count / 4) < size_in_object) { 8605 ptoa_64(resident_count)); 8612 ptoa_64(resident_count)); 8616 ptoa_64(resident_count)); 8624 ptoa_64(wired_count)); 8628 ptoa_64(wired_count)); 8632 ptoa_64(wired_count)); 8642 ptoa_64(compressed_count)); 8648 ptoa_64(compressed_count)); [all …]
|
| H A D | vm_resident.c | 7053 if ((flags & KMA_NOFAIL) == 0 && ptoa_64(page_size) > max_mem / 4) { in vm_page_alloc_list() 7054 uint64_t unavailable = ptoa_64(vm_page_wire_count + vm_page_free_target); in vm_page_alloc_list() 7055 if (unavailable > max_mem || ptoa_64(page_count) > (max_mem - unavailable)) { in vm_page_alloc_list() 8037 …force_to_wired_list = hibernate_sptm_should_force_page_to_wired_pagelist(ptoa_64(VM_PAGE_GET_PHYS_… in hibernate_page_list_setall() 8116 …force_to_wired_list = hibernate_sptm_should_force_page_to_wired_pagelist(ptoa_64(VM_PAGE_GET_PHYS_… in hibernate_page_list_setall() 8165 …force_to_wired_list = hibernate_sptm_should_force_page_to_wired_pagelist(ptoa_64(VM_PAGE_GET_PHYS_… in hibernate_page_list_setall() 9784 info[object->wire_tag].size += ptoa_64(object->wired_page_count); in vm_page_count_object() 10108 wired_size = ptoa_64(vm_page_wire_count); in vm_page_diagnose() 10109 wired_reserved_size = ptoa_64(vm_page_wire_count_initial - vm_page_stolen_count); in vm_page_diagnose() 10111 wired_size = ptoa_64(vm_page_wire_count + vm_lopage_free_count + vm_page_throttled_count); in vm_page_diagnose() [all …]
|
| H A D | vm_kern.c | 1608 vm_tag_update_size(tag, -ptoa_64(pages_unwired), NULL); in kernel_memory_depopulate()
|
| H A D | vm_compressor.c | 850 compressor_pool_size = ptoa_64(vm_compression_limit); in vm_compressor_set_size()
|
| H A D | vm_fault.c | 6878 vm_tag_update_size((vm_tag_t) fault_info.user_tag, -ptoa_64(unwired_pages), NULL); in vm_fault_unwire()
|
| H A D | vm_map.c | 6089 size + ptoa_64(total_wire_count) > vm_global_user_wire_limit) { in add_wire_counts() 6090 if (size + ptoa_64(total_wire_count) > vm_global_user_wire_limit) { in add_wire_counts() 6093 …bal wire limit. %llu bytes wired and requested to wire %llu bytes more", ptoa_64(total_wire_count)… in add_wire_counts()
|
| /xnu-11215.1.10/osfmk/i386/vmx/ |
| H A D | vmx_shims.c | 53 return ptoa_64(pmap_find_phys(kernel_pmap, (addr64_t)(uintptr_t)va)); in vmx_paddr()
|
| /xnu-11215.1.10/osfmk/mach/ |
| H A D | vm_param.h | 102 #define ptoa_64(x) ((uint64_t)(x) << PAGE_SHIFT) macro 190 #undef ptoa_64 225 #define ptoa_64(x) ((uint64_t)((x) + (uint8_t *)0))
|
| /xnu-11215.1.10/osfmk/kern/ |
| H A D | zalloc.c | 5069 *current_size = ptoa_64(phys_pages); in get_zone_map_size() 5070 *capacity = ptoa_64(zone_pages_wired_max); in get_zone_map_size() 5115 ptoa_64(os_atomic_load(&zone_pages_wired, relaxed)), in kill_process_in_largest_zone() 5116 ptoa_64(zone_pages_wired_max), in kill_process_in_largest_zone() 8555 .mzi_cur_size = ptoa_64(zone_scale_for_percpu(&zcopy, zcopy.z_wired_cur)), in get_zone_info() 8557 .mzi_max_size = ptoa_64(zone_scale_for_percpu(&zcopy, zcopy.z_wired_hwm)), in get_zone_info() 8559 .mzi_alloc_size = ptoa_64(zcopy.z_chunk_pages), in get_zone_info() 8571 ptoa_64(zone_scale_for_percpu(&zcopy, zcopy.z_wired_empty))); in get_zone_info()
|
| /xnu-11215.1.10/osfmk/i386/AT386/ |
| H A D | model_dep.c | 1598 uint64_t phys_src = ptoa_64(upn) | (uaddr & PAGE_MASK); in debug_copyin()
|
| /xnu-11215.1.10/osfmk/kdp/ |
| H A D | kdp_core.c | 413 uint64_t phys = ptoa_64(ppn); in kernel_pmap_present_mapping()
|
| /xnu-11215.1.10/bsd/kern/ |
| H A D | uipc_mbuf.c | 7454 return (uint64_t)(ptoa_64(base_phys) | ((uint64_t)addr & PAGE_MASK));
|