Home
last modified time | relevance | path

Searched refs:PAGE_SIZE_64 (Results 1 – 25 of 26) sorted by relevance

12

/xnu-8019.80.24/osfmk/vm/
H A Dvm_kern.c339 fill_start += PAGE_SIZE_64; in kernel_memory_allocate_prot()
340 fill_size -= PAGE_SIZE_64; in kernel_memory_allocate_prot()
350 fill_size -= PAGE_SIZE_64; in kernel_memory_allocate_prot()
358 wired_page_count = (int) (fill_size / PAGE_SIZE_64); in kernel_memory_allocate_prot()
359 assert(wired_page_count * PAGE_SIZE_64 == fill_size); in kernel_memory_allocate_prot()
455 pg_offset += PAGE_SIZE_64; in kernel_memory_allocate_prot()
470 for (pg_offset = fill_start; pg_offset < fill_start + fill_size; pg_offset += PAGE_SIZE_64) { in kernel_memory_allocate_prot()
622 pg_offset += PAGE_SIZE_64) { in kernel_memory_populate_with_pages()
711 pg_offset = page_count * PAGE_SIZE_64; in kernel_memory_populate()
729 pg_offset -= PAGE_SIZE_64; in kernel_memory_populate()
[all …]
H A Dvm_map_store.h104 VMHE_start= VMHE_map->highest_entry_end + PAGE_SIZE_64; \
107 VMHE_start = VMHE_map->highest_entry_end + PAGE_SIZE_64; \
H A Dvm_fourk_pager.c703 if (PAGE_SIZE_64 == FOURK_PAGE_SIZE) { in fourk_pager_create()
840 2 * PAGE_SIZE_64, in fourk_pager_data_request()
854 dst_vaddr = CAST_DOWN(vm_offset_t, kernel_mapping + PAGE_SIZE_64); in fourk_pager_data_request()
1146 (addr64_t) src_vaddr + PAGE_SIZE_64); in fourk_pager_data_request()
1262 kernel_mapping + (2 * PAGE_SIZE_64), in fourk_pager_data_request()
H A Dvm_swapfile_pager.c362 PAGE_SIZE_64, in swapfile_pager_data_request()
425 (addr64_t) (kernel_mapping + PAGE_SIZE_64)); in swapfile_pager_data_request()
463 kernel_mapping + PAGE_SIZE_64, in swapfile_pager_data_request()
H A Dvm_object.c2019 #define CHUNK_SIZE (PAGES_IN_A_CHUNK * PAGE_SIZE_64) /* Size of a chunk in bytes */
2037 for (p = (len) / PAGE_SIZE_64; p < PAGES_IN_A_CHUNK; p++) \
2155 …NK_NOT_COMPLETE(*chunk_state); p++, size -= PAGE_SIZE_64, offset += PAGE_SIZE_64, pmap_offset += P… in deactivate_pages_in_object()
2601 cur_offset += PAGE_SIZE_64) { in vm_object_reuse_pages()
2730 phys_addr += PAGE_SIZE_64) { in vm_object_pmap_protect_options()
2789 curr < start + PAGE_SIZE_64; in vm_object_pmap_protect_options()
2983 src_offset += PAGE_SIZE_64, in vm_object_copy_slowly()
2984 new_offset += PAGE_SIZE_64, size -= PAGE_SIZE_64 in vm_object_copy_slowly()
4982 hint_offset = (hint_offset > 8 * PAGE_SIZE_64) ? in vm_object_collapse()
4983 (hint_offset - 8 * PAGE_SIZE_64) : 0; in vm_object_collapse()
[all …]
H A Dvm_compressor_backing_store.c366 vm_page_encrypt_counter += (size / PAGE_SIZE_64); in vm_swap_encrypt()
397 vm_page_decrypt_counter += (size / PAGE_SIZE_64); in vm_swap_decrypt()
1524 …>swp_vp, file_offset, (uint64_t)c_seg->c_store.c_buffer, (int)(size / PAGE_SIZE_64), SWAP_READ, NU… in vm_swap_get()
1692 …error = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int) (size / PAGE_SIZE_64), SWAP_WRITE, up… in vm_swap_put()
2110 if (vm_swapfile_io(swf->swp_vp, f_offset, addr, (int)(c_size / PAGE_SIZE_64), SWAP_READ, NULL)) { in vm_swap_reclaim()
H A Dvm_fault.c380 #define ALIGNED(x) (((x) & (PAGE_SIZE_64 - 1)) == 0)
434 if (offset && last_alloc == offset - PAGE_SIZE_64) { in vm_fault_is_sequential()
450 if (last_alloc && last_alloc == offset + PAGE_SIZE_64) { in vm_fault_is_sequential()
467 if (offset && last_alloc == (offset - PAGE_SIZE_64)) { in vm_fault_is_sequential()
477 } else if (last_alloc && last_alloc == (offset + PAGE_SIZE_64)) { in vm_fault_is_sequential()
580 run_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind()
586 run_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind()
592 { vm_object_offset_t behind = vm_default_behind * PAGE_SIZE_64; in vm_fault_deactivate_behind()
608 pg_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind()
614 pg_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind()
[all …]
H A Dvm_pageout.c1438 (m)->vmp_offset + PAGE_SIZE_64, \
6309 dst_offset += PAGE_SIZE_64; in vm_object_upl_request()
7122 offset += PAGE_SIZE_64; in vm_map_enter_upl_range()
7123 new_offset += PAGE_SIZE_64; in vm_map_enter_upl_range()
7193 kasan_notify_address(addr, PAGE_SIZE_64); in vm_map_enter_upl_range()
7196 offset += PAGE_SIZE_64; in vm_map_enter_upl_range()
7936 target_offset += PAGE_SIZE_64; in upl_commit_range()
8449 target_offset += PAGE_SIZE_64; in upl_abort_range()
8921 *dst_offset += PAGE_SIZE_64; in vm_object_iopl_wire_empty()
9665 dst_offset += PAGE_SIZE_64; in vm_object_iopl_request()
[all …]
H A Dmemory_object.c565 offset += PAGE_SIZE_64) { in vm_object_update_extent()
628 next_offset = offset + PAGE_SIZE_64; in vm_object_update_extent()
H A Dvm_compressor.c403 return ((uint64_t)AVAILABLE_NON_COMPRESSED_MEMORY) * PAGE_SIZE_64; in vm_available_memory()
416 return c_segment_pages_compressed * PAGE_SIZE_64; in vm_compressor_pages_compressed()
1342 uint64_t uncompressed_bytes = ((c_seg->c_slots_used) * PAGE_SIZE_64); in c_seg_update_task_owner()
2591 total_bytes = compressor_object->resident_page_count * PAGE_SIZE_64; in vm_consider_waking_compactor_swapper()
4530 task_update_frozen_to_swap_acct(c_seg->c_task_owner, PAGE_SIZE_64, DEBIT_FROM_SWAP); in c_decompress_page()
H A Dbsd_vm.c284 offset += PAGE_SIZE_64; in memory_object_control_uiomove()
H A Dvm_map.c2055 offset += PAGE_SIZE_64; in vm_map_pmap_enter()
2535 start += PAGE_SIZE_64; in vm_map_enter()
11325 offset += PAGE_SIZE_64; in vm_map_copyout_internal()
14829 entry_size = (uint32_t) ((entry->vme_end - entry->vme_start) / PAGE_SIZE_64); in vm_map_region_top_walk()
20841 *shared_count = (unsigned int) ((dirty_shared_count * PAGE_SIZE_64) / (1024 * 1024ULL)); in vm_map_freeze()
20843 …_pages_threshold = (memorystatus_freeze_shared_mb_per_process_max * 1024 * 1024ULL) / PAGE_SIZE_64; in vm_map_freeze()
H A Dvm_resident.c4576 mem->vmp_offset + PAGE_SIZE_64, in vm_page_wire()
8134 for (addr = saddr_aligned; addr < eaddr_aligned; addr += PAGE_SIZE_64) { in hibernate_mark_as_unneeded()
/xnu-8019.80.24/bsd/vm/
H A Ddp_backing_file.c206 *pagesize_p = (vm_size_t)PAGE_SIZE_64; in macx_swapinfo()
H A Dvm_compressor_backing_file.c196 upl_size_t io_size = (upl_size_t) (npages * PAGE_SIZE_64); in vm_swapfile_io()
/xnu-8019.80.24/bsd/vfs/
H A Dvfs_cluster.c2183 f_offset = (off_t)(r_addr * PAGE_SIZE_64); in cluster_read_ahead()
2187 ubc_range_op(vp, f_offset, f_offset + PAGE_SIZE_64, UPL_ROP_PRESENT, &size_of_prefetch); in cluster_read_ahead()
3199 zero_pg_index = (int)((zero_off - upl_f_offset) / PAGE_SIZE_64); in cluster_zero_range()
3227 cl.b_addr = (daddr64_t)(s_offset / PAGE_SIZE_64); in cluster_update_state()
3228 cl.e_addr = (daddr64_t)(e_offset / PAGE_SIZE_64); in cluster_update_state()
3602 zero_cnt1 = PAGE_SIZE_64 - (zero_off1 & PAGE_MASK_64); in cluster_write_copy()
3644 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64); in cluster_write_copy()
3677 upl_f_offset += PAGE_SIZE_64; in cluster_write_copy()
3895 cl.e_addr = (daddr64_t)((upl_f_offset + (off_t)upl_size) / PAGE_SIZE_64); in cluster_write_copy()
4116 extent.b_addr = uio->uio_offset / PAGE_SIZE_64; in cluster_read_copy()
[all …]
/xnu-8019.80.24/bsd/kern/
H A Dkern_memorystatus.c976 printf("memorystatus_debug_dump ***START*(PAGE_SIZE_64=%llu)**\n", PAGE_SIZE_64); in memorystatus_debug_dump_bucket_locked()
984 (bytes / PAGE_SIZE_64), /* task's footprint converted from bytes to pages */ in memorystatus_debug_dump_bucket_locked()
4812 pages = (get_task_phys_footprint(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts()
4817 pages = (get_task_phys_footprint_lifetime_max(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts()
4822 pages = (get_task_purgeable_size(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts()
4838 *internal_pages = (get_task_internal(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4842 *internal_compressed_pages = (get_task_internal_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4846 *purgeable_nonvolatile_pages = (get_task_purgeable_nonvolatile(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4850 …le_nonvolatile_compressed_pages = (get_task_purgeable_nonvolatile_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4854 *alternate_accounting_pages = (get_task_alternate_accounting(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
[all …]
H A Dubc_subr.c1170 lastpg += PAGE_SIZE_64; in ubc_setsize_ex()
/xnu-8019.80.24/osfmk/mach/
H A Dvm_param.h88 #define PAGE_SIZE_64 (unsigned long long)PAGE_SIZE /* pagesize in addr units */ macro
/xnu-8019.80.24/osfmk/kdp/
H A Dkdp_core.c265 uint64_t vincr = PAGE_SIZE_64; in kernel_pmap_present_mapping()
363 vincr = PAGE_SIZE_64; in pmap_traverse_present_mappings()
/xnu-8019.80.24/osfmk/i386/
H A Dpmap_x86_common.c1345 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options()
1401 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options()
/xnu-8019.80.24/bsd/nfs/
H A Dnfs_bio.c536 for (i = 0; i < npages; i++, fileoffset += PAGE_SIZE_64) { in nfs_buf_upl_check()
1173 start = trunc_page_64(np->n_size) + PAGE_SIZE_64; in nfs_buf_release()
2356 auio = uio_create(1, (NBOFF(bp) + firstpg * PAGE_SIZE_64), UIO_SYSSPACE, UIO_READ); in nfs_bioread()
/xnu-8019.80.24/osfmk/kern/
H A Dcoalition.c2292 return bytes / PAGE_SIZE_64; in coalition_get_page_count()
H A Dtask.c2288 task_page_count = (get_task_phys_footprint(task) / PAGE_SIZE_64); in task_partial_reap()
4641 …*dirty_count = (uint32_t) (freezer_context_global.freezer_ctx_swapped_bytes / PAGE_SIZE_64); … in task_freeze()
/xnu-8019.80.24/osfmk/x86_64/
H A Dpmap.c1470 offset += PAGE_SIZE_64) { in pmap_is_empty()

12