Home
last modified time | relevance | path

Searched refs:PAGE_SIZE_64 (Results 1 – 25 of 26) sorted by relevance

12

/xnu-11215.1.10/osfmk/kdp/
H A Dsk_core.c420 for (vm_map_address_t vaddr = physmap_base; vaddr < physmap_end; vaddr += PAGE_SIZE_64) { in papt_walk()
431 seg_end = paddr + PAGE_SIZE_64; in papt_walk()
437 if (paddr == last_paddr + PAGE_SIZE_64) { in papt_walk()
438 seg_end = paddr + PAGE_SIZE_64; in papt_walk()
452 seg_end = paddr + PAGE_SIZE_64; in papt_walk()
H A Dkdp_core.c381 uint64_t vincr = PAGE_SIZE_64; in kernel_pmap_present_mapping()
479 vincr = PAGE_SIZE_64; in pmap_traverse_present_mappings()
/xnu-11215.1.10/bsd/vm/
H A Ddp_backing_file.c218 *pagesize_p = (vm_size_t)PAGE_SIZE_64; in macx_swapinfo()
H A Dvm_compressor_backing_file.c185 upl_size_t io_size = (upl_size_t) (npages * PAGE_SIZE_64); in vm_swapfile_io()
/xnu-11215.1.10/osfmk/vm/
H A Dvm_object.c2053 #define CHUNK_SIZE (PAGES_IN_A_CHUNK * PAGE_SIZE_64) /* Size of a chunk in bytes */
2071 for (p = (len) / PAGE_SIZE_64; p < PAGES_IN_A_CHUNK; p++) \
2196 …NK_NOT_COMPLETE(*chunk_state); p++, size -= PAGE_SIZE_64, offset += PAGE_SIZE_64, pmap_offset += P…
2685 cur_offset += PAGE_SIZE_64) {
2813 cur_offset += PAGE_SIZE_64;
2926 phys_addr += PAGE_SIZE_64) {
2985 curr < start + PAGE_SIZE_64;
3184 src_offset += PAGE_SIZE_64,
3185 new_offset += PAGE_SIZE_64, size -= PAGE_SIZE_64
5112 hint_offset = (hint_offset > 8 * PAGE_SIZE_64) ?
[all …]
H A Dvm_map_store.c449 start = map->highest_entry_end + PAGE_SIZE_64; in vm_map_store_find_space_forward()
451 start = entry->vme_end + PAGE_SIZE_64; in vm_map_store_find_space_forward()
H A Dvm_compressor_backing_store.c380 vm_page_encrypt_counter += (size / PAGE_SIZE_64); in vm_swap_encrypt()
411 vm_page_decrypt_counter += (size / PAGE_SIZE_64); in vm_swap_decrypt()
1678 …>swp_vp, file_offset, (uint64_t)c_seg->c_store.c_buffer, (int)(size / PAGE_SIZE_64), SWAP_READ, NU… in vm_swap_get()
1846 …error = vm_swapfile_io(swf->swp_vp, file_offset, addr, (int) (size / PAGE_SIZE_64), SWAP_WRITE, up… in vm_swap_put()
2262 if (vm_swapfile_io(swf->swp_vp, f_offset, addr, (int)(c_size / PAGE_SIZE_64), SWAP_READ, NULL)) { in vm_swap_reclaim()
H A Dvm_compressor.c470 uint64_t bytes_total = VM_PAGE_COMPRESSOR_COUNT * PAGE_SIZE_64; in vm_compressor_needs_to_minor_compact()
482 return ((uint64_t)AVAILABLE_NON_COMPRESSED_MEMORY) * PAGE_SIZE_64; in vm_available_memory()
514 return c_segment_pages_compressed * PAGE_SIZE_64; in vm_compressor_pages_compressed()
1672 uint64_t uncompressed_bytes = ((c_seg->c_slots_used) * PAGE_SIZE_64); in c_seg_update_task_owner()
5357 task_update_frozen_to_swap_acct(c_seg->c_task_owner, PAGE_SIZE_64, DEBIT_FROM_SWAP); in c_decompress_page()
6265 uint64_t pgidx = free_offset / PAGE_SIZE_64; in vm_uncompressed_return_space_to_swap()
6282 uint64_t pgidx = free_offset / PAGE_SIZE_64; in vm_uncompressed_return_space_to_swap()
6322 uncompressed_file0_free_offset = ((chunkidx * 64) + chunkoffset) * PAGE_SIZE_64; in vm_uncompressed_reserve_space_in_swap()
6340 uncompressed_file1_free_offset = ((chunkidx * 64) + chunkoffset) * PAGE_SIZE_64; in vm_uncompressed_reserve_space_in_swap()
6378 …kmem_alloc(kernel_map, (vm_offset_t*)&uncompressor_io_req[i].addr, PAGE_SIZE_64, KMA_NOFAIL | KMA_… in vm_uncompressed_put()
[all …]
H A Dvm_fault.c413 #define ALIGNED(x) (((x) & (PAGE_SIZE_64 - 1)) == 0)
467 if (offset && last_alloc == offset - PAGE_SIZE_64) { in vm_fault_is_sequential()
483 if (last_alloc && last_alloc == offset + PAGE_SIZE_64) { in vm_fault_is_sequential()
500 if (offset && last_alloc == (offset - PAGE_SIZE_64)) { in vm_fault_is_sequential()
510 } else if (last_alloc && last_alloc == (offset + PAGE_SIZE_64)) { in vm_fault_is_sequential()
614 run_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind()
620 run_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind()
626 { vm_object_offset_t behind = vm_default_behind * PAGE_SIZE_64; in vm_fault_deactivate_behind()
642 pg_offset = PAGE_SIZE_64; in vm_fault_deactivate_behind()
648 pg_offset = 0 - PAGE_SIZE_64; in vm_fault_deactivate_behind()
[all …]
H A Dvm_pageout.c1584 (m)->vmp_offset + PAGE_SIZE_64, \
6599 dst_offset += PAGE_SIZE_64;
7397 offset += PAGE_SIZE_64;
7398 new_offset += PAGE_SIZE_64;
7470 kasan_notify_address(addr, PAGE_SIZE_64);
7473 offset += PAGE_SIZE_64;
7960 *dst_offset += PAGE_SIZE_64;
8710 dst_offset += PAGE_SIZE_64;
9198 map_size -= PAGE_SIZE_64, page_map_offset += PAGE_SIZE_64) {
9226 vm_paging_pages_mapped_slow += (unsigned long) (map_size / PAGE_SIZE_64);
[all …]
H A Dvm_upl.c456 target_offset += PAGE_SIZE_64; in upl_abort_range()
1158 target_offset += PAGE_SIZE_64; in upl_commit_range()
H A Dvm_swapfile_pager.c392 (addr64_t) (kernel_mapping + PAGE_SIZE_64)); in swapfile_pager_data_request()
H A Dmemory_object.c533 offset += PAGE_SIZE_64) { in vm_object_update_extent()
596 next_offset = offset + PAGE_SIZE_64; in vm_object_update_extent()
H A Dvm_kern.c1426 pg_offset += PAGE_SIZE_64) { in kernel_memory_populate_object_and_unlock()
1570 pg_offset += PAGE_SIZE_64) { in kernel_memory_depopulate()
2038 offset += PAGE_SIZE_64) { in kmem_realloc_guard()
2093 offset += PAGE_SIZE_64) { in kmem_realloc_guard()
2119 offset += PAGE_SIZE_64) { in kmem_realloc_guard()
2206 offset += PAGE_SIZE_64) { in kmem_realloc_guard()
H A Dbsd_vm.c257 offset += PAGE_SIZE_64; in memory_object_control_uiomove()
H A Dvm_map.c2652 offset += PAGE_SIZE_64; in vm_map_pmap_enter()
11456 offset += PAGE_SIZE_64; in vm_map_copyout_internal()
15318 entry_size = (uint32_t) ((entry->vme_end - entry->vme_start) / PAGE_SIZE_64); in vm_map_region_top_walk()
21876 *shared_count = (unsigned int) ((dirty_shared_count * PAGE_SIZE_64) / (1024 * 1024ULL)); in vm_map_freeze()
21878 …_pages_threshold = (memorystatus_freeze_shared_mb_per_process_max * 1024 * 1024ULL) / PAGE_SIZE_64; in vm_map_freeze()
H A Dvm_resident.c4776 mem->vmp_offset + PAGE_SIZE_64, in vm_page_wire()
8551 for (addr = saddr_aligned; addr < eaddr_aligned; addr += PAGE_SIZE_64) { in hibernate_mark_as_unneeded()
/xnu-11215.1.10/bsd/vfs/
H A Dvfs_cluster.c2520 f_offset = (off_t)(r_addr * PAGE_SIZE_64); in cluster_read_ahead()
2524 ubc_range_op(vp, f_offset, f_offset + PAGE_SIZE_64, UPL_ROP_PRESENT, &size_of_prefetch); in cluster_read_ahead()
3560 zero_pg_index = (int)((zero_off - upl_f_offset) / PAGE_SIZE_64); in cluster_zero_range()
3588 cl.b_addr = (daddr64_t)(s_offset / PAGE_SIZE_64); in cluster_update_state()
3589 cl.e_addr = (daddr64_t)(e_offset / PAGE_SIZE_64); in cluster_update_state()
3963 zero_cnt1 = PAGE_SIZE_64 - (zero_off1 & PAGE_MASK_64); in cluster_write_copy()
4005 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64); in cluster_write_copy()
4038 upl_f_offset += PAGE_SIZE_64; in cluster_write_copy()
4256 cl.e_addr = (daddr64_t)((upl_f_offset + (off_t)upl_size) / PAGE_SIZE_64); in cluster_write_copy()
4500 extent.b_addr = uio->uio_offset / PAGE_SIZE_64; in cluster_read_copy()
[all …]
/xnu-11215.1.10/bsd/kern/
H A Dkern_memorystatus.c1021 memorystatus_log_debug("memorystatus_debug_dump ***START*(PAGE_SIZE_64=%llu)**\n", PAGE_SIZE_64); in memorystatus_debug_dump_bucket_locked()
1029 (bytes / PAGE_SIZE_64), /* task's footprint converted from bytes to pages */ in memorystatus_debug_dump_bucket_locked()
4601 pages = (get_task_phys_footprint(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts()
4606 pages = (get_task_phys_footprint_lifetime_max(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts()
4611 pages = (get_task_purgeable_size(task) / PAGE_SIZE_64); in memorystatus_get_task_page_counts()
4628 *internal_pages = (get_task_internal(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4632 *internal_compressed_pages = (get_task_internal_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4636 *purgeable_nonvolatile_pages = (get_task_purgeable_nonvolatile(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4640 …le_nonvolatile_compressed_pages = (get_task_purgeable_nonvolatile_compressed(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
4644 *alternate_accounting_pages = (get_task_alternate_accounting(task) / PAGE_SIZE_64); in memorystatus_get_task_phys_footprint_page_counts()
[all …]
H A Duipc_syscalls.c3917 xfsize = PAGE_SIZE_64 - pgoff; in sendfile()
/xnu-11215.1.10/osfmk/mach/
H A Dvm_param.h90 #define PAGE_SIZE_64 (unsigned long long)PAGE_SIZE /* pagesize in addr units */ macro
/xnu-11215.1.10/osfmk/i386/
H A Dpmap_x86_common.c1354 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options()
1410 cpte++, vaddr += PAGE_SIZE_64) { in pmap_remove_range_options()
/xnu-11215.1.10/osfmk/x86_64/
H A Dpmap.c1524 offset += PAGE_SIZE_64) { in pmap_is_empty()
/xnu-11215.1.10/osfmk/kern/
H A Dcoalition.c2560 return bytes / PAGE_SIZE_64; in coalition_get_page_count()
H A Dtask.c2516 task_page_count = (get_task_phys_footprint(task) / PAGE_SIZE_64); in task_partial_reap()
5034 …*dirty_count = (uint32_t) (freezer_context_global.freezer_ctx_swapped_bytes / PAGE_SIZE_64); … in task_freeze()

12