| /xnu-8019.80.24/osfmk/vm/ |
| H A D | vm_kern.c | 339 fill_start += PAGE_SIZE_64; in kernel_memory_allocate_prot() 340 fill_size -= PAGE_SIZE_64; in kernel_memory_allocate_prot() 350 fill_size -= PAGE_SIZE_64; in kernel_memory_allocate_prot() 358 wired_page_count = (int) (fill_size / PAGE_SIZE_64); in kernel_memory_allocate_prot() 359 assert(wired_page_count * PAGE_SIZE_64 == fill_size); in kernel_memory_allocate_prot() 455 pg_offset += PAGE_SIZE_64; in kernel_memory_allocate_prot() 470 for (pg_offset = fill_start; pg_offset < fill_start + fill_size; pg_offset += PAGE_SIZE_64) { in kernel_memory_allocate_prot() 622 pg_offset += PAGE_SIZE_64) { in kernel_memory_populate_with_pages() 711 pg_offset = page_count * PAGE_SIZE_64; in kernel_memory_populate() 729 pg_offset -= PAGE_SIZE_64; in kernel_memory_populate() [all …]
|
| H A D | vm_map_store.h | 104 VMHE_start= VMHE_map->highest_entry_end + PAGE_SIZE_64; \ 107 VMHE_start = VMHE_map->highest_entry_end + PAGE_SIZE_64; \
|
| H A D | vm_fourk_pager.c | 703 if (PAGE_SIZE_64 == FOURK_PAGE_SIZE) { in fourk_pager_create() 840 2 * PAGE_SIZE_64, in fourk_pager_data_request() 854 dst_vaddr = CAST_DOWN(vm_offset_t, kernel_mapping + PAGE_SIZE_64); in fourk_pager_data_request() 1146 (addr64_t) src_vaddr + PAGE_SIZE_64); in fourk_pager_data_request() 1262 kernel_mapping + (2 * PAGE_SIZE_64), in fourk_pager_data_request()
|
| H A D | vm_swapfile_pager.c | 362 PAGE_SIZE_64, in swapfile_pager_data_request() 425 (addr64_t) (kernel_mapping + PAGE_SIZE_64)); in swapfile_pager_data_request() 463 kernel_mapping + PAGE_SIZE_64, in swapfile_pager_data_request()
|
| H A D | vm_object.c | 2019 #define CHUNK_SIZE (PAGES_IN_A_CHUNK * PAGE_SIZE_64) /* Size of a chunk in bytes */ 2037 for (p = (len) / PAGE_SIZE_64; p < PAGES_IN_A_CHUNK; p++) \ 2155 …NK_NOT_COMPLETE(*chunk_state); p++, size -= PAGE_SIZE_64, offset += PAGE_SIZE_64, pmap_offset += P… in deactivate_pages_in_object() 2601 cur_offset += PAGE_SIZE_64) { in vm_object_reuse_pages() 2730 phys_addr += PAGE_SIZE_64) { in vm_object_pmap_protect_options() 2789 curr < start + PAGE_SIZE_64; in vm_object_pmap_protect_options() 2983 src_offset += PAGE_SIZE_64, in vm_object_copy_slowly() 2984 new_offset += PAGE_SIZE_64, size -= PAGE_SIZE_64 in vm_object_copy_slowly() 4982 hint_offset = (hint_offset > 8 * PAGE_SIZE_64) ? in vm_object_collapse() 4983 (hint_offset - 8 * PAGE_SIZE_64) : 0; in vm_object_collapse() [all …]
|
| H A D | vm_compressor_backing_store.c | 366 vm_page_encrypt_counter += (size / PAGE_SIZE_64); in vm_swap_encrypt() 397 vm_page_decrypt_counter += (size / PAGE_SIZE_64); in vm_swap_decrypt() 1524 …>swp_vp, file_offset, (uint64_t)c_seg->c_store.c_buffer, (int)(size / PAGE_SIZE_64), SWAP_READ, NU… in vm_swap_get() 1692 …error = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int) (size / PAGE_SIZE_64), SWAP_WRITE, up… in vm_swap_put() 2110 if (vm_swapfile_io(swf->swp_vp, f_offset, addr, (int)(c_size / PAGE_SIZE_64), SWAP_READ, NULL)) { in vm_swap_reclaim()
|
| H A D | vm_fault.c | 380 #define ALIGNED(x) (((x) & (PAGE_SIZE_64 - 1)) == 0) 434 if (offset && last_alloc == offset - PAGE_SIZE_64) { in vm_fault_is_sequential() 450 if (last_alloc && last_alloc == offset + PAGE_SIZE_64) { in vm_fault_is_sequential() 467 if (offset && last_alloc == (offset - PAGE_SIZE_64)) { in vm_fault_is_sequential() 477 } else if (last_alloc && last_alloc == (offset + PAGE_SIZE_64)) { in vm_fault_is_sequential() 580 run_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind() 586 run_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind() 592 { vm_object_offset_t behind = vm_default_behind * PAGE_SIZE_64; in vm_fault_deactivate_behind() 608 pg_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind() 614 pg_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind() [all …]
|
| H A D | vm_pageout.c | 1438 (m)->vmp_offset + PAGE_SIZE_64, \ 6309 dst_offset += PAGE_SIZE_64; in vm_object_upl_request() 7122 offset += PAGE_SIZE_64; in vm_map_enter_upl_range() 7123 new_offset += PAGE_SIZE_64; in vm_map_enter_upl_range() 7193 kasan_notify_address(addr, PAGE_SIZE_64); in vm_map_enter_upl_range() 7196 offset += PAGE_SIZE_64; in vm_map_enter_upl_range() 7936 target_offset += PAGE_SIZE_64; in upl_commit_range() 8449 target_offset += PAGE_SIZE_64; in upl_abort_range() 8921 *dst_offset += PAGE_SIZE_64; in vm_object_iopl_wire_empty() 9665 dst_offset += PAGE_SIZE_64; in vm_object_iopl_request() [all …]
|
| H A D | memory_object.c | 565 offset += PAGE_SIZE_64) { in vm_object_update_extent() 628 next_offset = offset + PAGE_SIZE_64; in vm_object_update_extent()
|
| H A D | vm_compressor.c | 403 return ((uint64_t)AVAILABLE_NON_COMPRESSED_MEMORY) * PAGE_SIZE_64; in vm_available_memory() 416 return c_segment_pages_compressed * PAGE_SIZE_64; in vm_compressor_pages_compressed() 1342 uint64_t uncompressed_bytes = ((c_seg->c_slots_used) * PAGE_SIZE_64); in c_seg_update_task_owner() 2591 total_bytes = compressor_object->resident_page_count * PAGE_SIZE_64; in vm_consider_waking_compactor_swapper() 4530 task_update_frozen_to_swap_acct(c_seg->c_task_owner, PAGE_SIZE_64, DEBIT_FROM_SWAP); in c_decompress_page()
|
| H A D | bsd_vm.c | 284 offset += PAGE_SIZE_64; in memory_object_control_uiomove()
|
| H A D | vm_map.c | 2055 offset += PAGE_SIZE_64; in vm_map_pmap_enter() 2535 start += PAGE_SIZE_64; in vm_map_enter() 11325 offset += PAGE_SIZE_64; in vm_map_copyout_internal() 14829 entry_size = (uint32_t) ((entry->vme_end - entry->vme_start) / PAGE_SIZE_64); in vm_map_region_top_walk() 20841 *shared_count = (unsigned int) ((dirty_shared_count * PAGE_SIZE_64) / (1024 * 1024ULL)); in vm_map_freeze() 20843 …_pages_threshold = (memorystatus_freeze_shared_mb_per_process_max * 1024 * 1024ULL) / PAGE_SIZE_64; in vm_map_freeze()
|
| H A D | vm_resident.c | 4576 mem->vmp_offset + PAGE_SIZE_64, in vm_page_wire() 8134 for (addr = saddr_aligned; addr < eaddr_aligned; addr += PAGE_SIZE_64) { in hibernate_mark_as_unneeded()
|
| /xnu-8019.80.24/bsd/vm/ |
| H A D | dp_backing_file.c | 206 *pagesize_p = (vm_size_t)PAGE_SIZE_64; in macx_swapinfo()
|
| H A D | vm_compressor_backing_file.c | 196 upl_size_t io_size = (upl_size_t) (npages * PAGE_SIZE_64); in vm_swapfile_io()
|
| /xnu-8019.80.24/bsd/vfs/ |
| H A D | vfs_cluster.c | 2183 f_offset = (off_t)(r_addr * PAGE_SIZE_64); in cluster_read_ahead() 2187 ubc_range_op(vp, f_offset, f_offset + PAGE_SIZE_64, UPL_ROP_PRESENT, &size_of_prefetch); in cluster_read_ahead() 3199 zero_pg_index = (int)((zero_off - upl_f_offset) / PAGE_SIZE_64); in cluster_zero_range() 3227 cl.b_addr = (daddr64_t)(s_offset / PAGE_SIZE_64); in cluster_update_state() 3228 cl.e_addr = (daddr64_t)(e_offset / PAGE_SIZE_64); in cluster_update_state() 3602 zero_cnt1 = PAGE_SIZE_64 - (zero_off1 & PAGE_MASK_64); in cluster_write_copy() 3644 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64); in cluster_write_copy() 3677 upl_f_offset += PAGE_SIZE_64; in cluster_write_copy() 3895 cl.e_addr = (daddr64_t)((upl_f_offset + (off_t)upl_size) / PAGE_SIZE_64); in cluster_write_copy() 4116 extent.b_addr = uio->uio_offset / PAGE_SIZE_64; in cluster_read_copy() [all …]
|
| /xnu-8019.80.24/bsd/kern/ |
| H A D | kern_memorystatus.c | 976 printf("memorystatus_debug_dump ***START*(PAGE_SIZE_64=%llu)**\n", PAGE_SIZE_64); in memorystatus_debug_dump_bucket_locked() 984 (bytes / PAGE_SIZE_64), /* task's footprint converted from bytes to pages */ in memorystatus_debug_dump_bucket_locked() 4812 pages = (get_task_phys_footprint(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts() 4817 pages = (get_task_phys_footprint_lifetime_max(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts() 4822 pages = (get_task_purgeable_size(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts() 4838 *internal_pages = (get_task_internal(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4842 *internal_compressed_pages = (get_task_internal_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4846 *purgeable_nonvolatile_pages = (get_task_purgeable_nonvolatile(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4850 …le_nonvolatile_compressed_pages = (get_task_purgeable_nonvolatile_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() 4854 *alternate_accounting_pages = (get_task_alternate_accounting(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts() [all …]
|
| H A D | ubc_subr.c | 1170 lastpg += PAGE_SIZE_64; in ubc_setsize_ex()
|
| /xnu-8019.80.24/osfmk/mach/ |
| H A D | vm_param.h | 88 #define PAGE_SIZE_64 (unsigned long long)PAGE_SIZE /* pagesize in addr units */ macro
|
| /xnu-8019.80.24/osfmk/kdp/ |
| H A D | kdp_core.c | 265 uint64_t vincr = PAGE_SIZE_64; in kernel_pmap_present_mapping() 363 vincr = PAGE_SIZE_64; in pmap_traverse_present_mappings()
|
| /xnu-8019.80.24/osfmk/i386/ |
| H A D | pmap_x86_common.c | 1345 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options() 1401 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options()
|
| /xnu-8019.80.24/bsd/nfs/ |
| H A D | nfs_bio.c | 536 for (i = 0; i < npages; i++, fileoffset += PAGE_SIZE_64) { in nfs_buf_upl_check() 1173 start = trunc_page_64(np->n_size) + PAGE_SIZE_64; in nfs_buf_release() 2356 auio = uio_create(1, (NBOFF(bp) + firstpg * PAGE_SIZE_64), UIO_SYSSPACE, UIO_READ); in nfs_bioread()
|
| /xnu-8019.80.24/osfmk/kern/ |
| H A D | coalition.c | 2292 return bytes / PAGE_SIZE_64; in coalition_get_page_count()
|
| H A D | task.c | 2288 task_page_count = (get_task_phys_footprint(task) / PAGE_SIZE_64); in task_partial_reap() 4641 …*dirty_count = (uint32_t) (freezer_context_global.freezer_ctx_swapped_bytes / PAGE_SIZE_64); … in task_freeze()
|
| /xnu-8019.80.24/osfmk/x86_64/ |
| H A D | pmap.c | 1470 offset += PAGE_SIZE_64) { in pmap_is_empty()
|