| /xnu-11215.1.10/osfmk/kdp/ |
| H A D | sk_core.c | 420 for (vm_map_address_t vaddr = physmap_base; vaddr < physmap_end; vaddr += PAGE_SIZE_64) { in papt_walk() 431 seg_end = paddr + PAGE_SIZE_64; in papt_walk() 437 if (paddr == last_paddr + PAGE_SIZE_64) { in papt_walk() 438 seg_end = paddr + PAGE_SIZE_64; in papt_walk() 452 seg_end = paddr + PAGE_SIZE_64; in papt_walk()
|
| H A D | kdp_core.c | 381 uint64_t vincr = PAGE_SIZE_64; in kernel_pmap_present_mapping() 479 vincr = PAGE_SIZE_64; in pmap_traverse_present_mappings()
|
| /xnu-11215.1.10/bsd/vm/ |
| H A D | dp_backing_file.c | 218 *pagesize_p = (vm_size_t)PAGE_SIZE_64; in macx_swapinfo()
|
| H A D | vm_compressor_backing_file.c | 185 upl_size_t io_size = (upl_size_t) (npages * PAGE_SIZE_64); in vm_swapfile_io()
|
| /xnu-11215.1.10/osfmk/vm/ |
| H A D | vm_object.c | 2053 #define CHUNK_SIZE (PAGES_IN_A_CHUNK * PAGE_SIZE_64) /* Size of a chunk in bytes */ 2071 for (p = (len) / PAGE_SIZE_64; p < PAGES_IN_A_CHUNK; p++) \ 2196 …NK_NOT_COMPLETE(*chunk_state); p++, size -= PAGE_SIZE_64, offset += PAGE_SIZE_64, pmap_offset += P… 2685 cur_offset += PAGE_SIZE_64) { 2813 cur_offset += PAGE_SIZE_64; 2926 phys_addr += PAGE_SIZE_64) { 2985 curr < start + PAGE_SIZE_64; 3184 src_offset += PAGE_SIZE_64, 3185 new_offset += PAGE_SIZE_64, size -= PAGE_SIZE_64 5112 hint_offset = (hint_offset > 8 * PAGE_SIZE_64) ? [all …]
|
| H A D | vm_map_store.c | 449 start = map->highest_entry_end + PAGE_SIZE_64; in vm_map_store_find_space_forward() 451 start = entry->vme_end + PAGE_SIZE_64; in vm_map_store_find_space_forward()
|
| H A D | vm_compressor_backing_store.c | 380 vm_page_encrypt_counter += (size / PAGE_SIZE_64); in vm_swap_encrypt() 411 vm_page_decrypt_counter += (size / PAGE_SIZE_64); in vm_swap_decrypt() 1678 …>swp_vp, file_offset, (uint64_t)c_seg->c_store.c_buffer, (int)(size / PAGE_SIZE_64), SWAP_READ, NU… in vm_swap_get() 1846 …error = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int) (size / PAGE_SIZE_64), SWAP_WRITE, up… in vm_swap_put() 2262 if (vm_swapfile_io(swf->swp_vp, f_offset, addr, (int)(c_size / PAGE_SIZE_64), SWAP_READ, NULL)) { in vm_swap_reclaim()
|
| H A D | vm_compressor.c | 470 uint64_t bytes_total = VM_PAGE_COMPRESSOR_COUNT * PAGE_SIZE_64; in vm_compressor_needs_to_minor_compact() 482 return ((uint64_t)AVAILABLE_NON_COMPRESSED_MEMORY) * PAGE_SIZE_64; in vm_available_memory() 514 return c_segment_pages_compressed * PAGE_SIZE_64; in vm_compressor_pages_compressed() 1672 uint64_t uncompressed_bytes = ((c_seg->c_slots_used) * PAGE_SIZE_64); in c_seg_update_task_owner() 5357 task_update_frozen_to_swap_acct(c_seg->c_task_owner, PAGE_SIZE_64, DEBIT_FROM_SWAP); in c_decompress_page() 6265 uint64_t pgidx = free_offset / PAGE_SIZE_64; in vm_uncompressed_return_space_to_swap() 6282 uint64_t pgidx = free_offset / PAGE_SIZE_64; in vm_uncompressed_return_space_to_swap() 6322 uncompressed_file0_free_offset = ((chunkidx * 64) + chunkoffset) * PAGE_SIZE_64; in vm_uncompressed_reserve_space_in_swap() 6340 uncompressed_file1_free_offset = ((chunkidx * 64) + chunkoffset) * PAGE_SIZE_64; in vm_uncompressed_reserve_space_in_swap() 6378 …kmem_alloc(kernel_map, (vm_offset_t*)&uncompressor_io_req[i].addr, PAGE_SIZE_64, KMA_NOFAIL | KMA_… in vm_uncompressed_put() [all …]
|
| H A D | vm_fault.c | 413 #define ALIGNED(x) (((x) & (PAGE_SIZE_64 - 1)) == 0) 467 if (offset && last_alloc == offset - PAGE_SIZE_64) { in vm_fault_is_sequential() 483 if (last_alloc && last_alloc == offset + PAGE_SIZE_64) { in vm_fault_is_sequential() 500 if (offset && last_alloc == (offset - PAGE_SIZE_64)) { in vm_fault_is_sequential() 510 } else if (last_alloc && last_alloc == (offset + PAGE_SIZE_64)) { in vm_fault_is_sequential() 614 run_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind() 620 run_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind() 626 { vm_object_offset_t behind = vm_default_behind * PAGE_SIZE_64; in vm_fault_deactivate_behind() 642 pg_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind() 648 pg_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind() [all …]
|
| H A D | vm_pageout.c | 1584 (m)->vmp_offset + PAGE_SIZE_64, \ 6599 dst_offset += PAGE_SIZE_64; 7397 offset += PAGE_SIZE_64; 7398 new_offset += PAGE_SIZE_64; 7470 kasan_notify_address(addr, PAGE_SIZE_64); 7473 offset += PAGE_SIZE_64; 7960 *dst_offset += PAGE_SIZE_64; 8710 dst_offset += PAGE_SIZE_64; 9198 map_size -= PAGE_SIZE_64, page_map_offset += PAGE_SIZE_64) { 9226 vm_paging_pages_mapped_slow += (unsigned long) (map_size / PAGE_SIZE_64); [all …]
|
| H A D | vm_upl.c | 456 target_offset += PAGE_SIZE_64; in upl_abort_range() 1158 target_offset += PAGE_SIZE_64; in upl_commit_range()
|
| H A D | vm_swapfile_pager.c | 392 (addr64_t) (kernel_mapping + PAGE_SIZE_64)); in swapfile_pager_data_request()
|
| H A D | memory_object.c | 533 offset += PAGE_SIZE_64) { in vm_object_update_extent() 596 next_offset = offset + PAGE_SIZE_64; in vm_object_update_extent()
|
| H A D | vm_kern.c | 1426 pg_offset += PAGE_SIZE_64) { in kernel_memory_populate_object_and_unlock() 1570 pg_offset += PAGE_SIZE_64) { in kernel_memory_depopulate() 2038 offset += PAGE_SIZE_64) { in kmem_realloc_guard() 2093 offset += PAGE_SIZE_64) { in kmem_realloc_guard() 2119 offset += PAGE_SIZE_64) { in kmem_realloc_guard() 2206 offset += PAGE_SIZE_64) { in kmem_realloc_guard()
|
| H A D | bsd_vm.c | 257 offset += PAGE_SIZE_64; in memory_object_control_uiomove()
|
| H A D | vm_map.c | 2652 offset += PAGE_SIZE_64; in vm_map_pmap_enter() 11456 offset += PAGE_SIZE_64; in vm_map_copyout_internal() 15318 entry_size = (uint32_t) ((entry->vme_end - entry->vme_start) / PAGE_SIZE_64); in vm_map_region_top_walk() 21876 *shared_count = (unsigned int) ((dirty_shared_count * PAGE_SIZE_64) / (1024 * 1024ULL)); in vm_map_freeze() 21878 …_pages_threshold = (memorystatus_freeze_shared_mb_per_process_max * 1024 * 1024ULL) / PAGE_SIZE_64; in vm_map_freeze()
|
| H A D | vm_resident.c | 4776 mem->vmp_offset + PAGE_SIZE_64, in vm_page_wire() 8551 for (addr = saddr_aligned; addr < eaddr_aligned; addr += PAGE_SIZE_64) { in hibernate_mark_as_unneeded()
|
| /xnu-11215.1.10/bsd/vfs/ |
| H A D | vfs_cluster.c | 2520 f_offset = (off_t)(r_addr * PAGE_SIZE_64); in cluster_read_ahead() 2524 ubc_range_op(vp, f_offset, f_offset + PAGE_SIZE_64, UPL_ROP_PRESENT, &size_of_prefetch); in cluster_read_ahead() 3560 zero_pg_index = (int)((zero_off - upl_f_offset) / PAGE_SIZE_64); in cluster_zero_range() 3588 cl.b_addr = (daddr64_t)(s_offset / PAGE_SIZE_64); in cluster_update_state() 3589 cl.e_addr = (daddr64_t)(e_offset / PAGE_SIZE_64); in cluster_update_state() 3963 zero_cnt1 = PAGE_SIZE_64 - (zero_off1 & PAGE_MASK_64); in cluster_write_copy() 4005 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64); in cluster_write_copy() 4038 upl_f_offset += PAGE_SIZE_64; in cluster_write_copy() 4256 cl.e_addr = (daddr64_t)((upl_f_offset + (off_t)upl_size) / PAGE_SIZE_64); in cluster_write_copy() 4500 extent.b_addr = uio->uio_offset / PAGE_SIZE_64; in cluster_read_copy() [all …]
|
| /xnu-11215.1.10/bsd/kern/ |
| H A D | kern_memorystatus.c | 1021 memorystatus_log_debug("memorystatus_debug_dump ***START*(PAGE_SIZE_64=%llu)**\n", PAGE_SIZE_64); in memorystatus_debug_dump_bucket_locked() 1029 (bytes / PAGE_SIZE_64), /* task's footprint converted from bytes to pages */ in memorystatus_debug_dump_bucket_locked() 4601 pages = (get_task_phys_footprint(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts() 4606 pages = (get_task_phys_footprint_lifetime_max(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts() 4611 pages = (get_task_purgeable_size(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts() 4628 *internal_pages = (get_task_internal(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4632 *internal_compressed_pages = (get_task_internal_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4636 *purgeable_nonvolatile_pages = (get_task_purgeable_nonvolatile(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4640 …le_nonvolatile_compressed_pages = (get_task_purgeable_nonvolatile_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4644 *alternate_accounting_pages = (get_task_alternate_accounting(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() [all …]
|
| H A D | uipc_syscalls.c | 3917 xfsize = PAGE_SIZE_64 - pgoff; in sendfile()
|
| /xnu-11215.1.10/osfmk/mach/ |
| H A D | vm_param.h | 90 #define PAGE_SIZE_64 (unsigned long long)PAGE_SIZE /* pagesize in addr units */ macro
|
| /xnu-11215.1.10/osfmk/i386/ |
| H A D | pmap_x86_common.c | 1354 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options() 1410 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options()
|
| /xnu-11215.1.10/osfmk/x86_64/ |
| H A D | pmap.c | 1524 offset += PAGE_SIZE_64) { in pmap_is_empty()
|
| /xnu-11215.1.10/osfmk/kern/ |
| H A D | coalition.c | 2560 return bytes / PAGE_SIZE_64; in coalition_get_page_count()
|
| H A D | task.c | 2516 task_page_count = (get_task_phys_footprint(task) / PAGE_SIZE_64); in task_partial_reap() 5034 …*dirty_count = (uint32_t) (freezer_context_global.freezer_ctx_swapped_bytes / PAGE_SIZE_64); … in task_freeze()
|