1 /* 2 * Copyright (c) 2004-2020 Apple Inc. All rights reserved. 3 * 4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ 5 * 6 * This file contains Original Code and/or Modifications of Original Code 7 * as defined in and that are subject to the Apple Public Source License 8 * Version 2.0 (the 'License'). You may not use this file except in 9 * compliance with the License. The rights granted to you under the License 10 * may not be used to create, or enable the creation or redistribution of, 11 * unlawful or unlicensed copies of an Apple operating system, or to 12 * circumvent, violate, or enable the circumvention or violation of, any 13 * terms of an Apple operating system software license agreement. 14 * 15 * Please obtain a copy of the License at 16 * http://www.opensource.apple.com/apsl/ and read it before using this file. 17 * 18 * The Original Code and all software distributed under the License are 19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER 20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, 21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, 22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. 23 * Please see the License for the specific language governing rights and 24 * limitations under the License. 25 * 26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ 27 */ 28 29 #ifdef XNU_KERNEL_PRIVATE 30 31 #ifndef _VM_VM_PROTOS_H_ 32 #define _VM_VM_PROTOS_H_ 33 34 #include <mach/mach_types.h> 35 #include <kern/kern_types.h> 36 #include <ipc/ipc_types.h> 37 #include <vm/vm_options.h> 38 39 #ifdef __cplusplus 40 extern "C" { 41 #endif 42 43 /* 44 * This file contains various type definitions and routine prototypes 45 * that are needed to avoid compilation warnings for VM code (in osfmk, 46 * default_pager and bsd). 47 * Most of these should eventually go into more appropriate header files. 48 * 49 * Include it after all other header files since it doesn't include any 50 * type definitions and it works around some conflicts with other header 51 * files. 52 */ 53 54 55 /* 56 * osfmk 57 */ 58 #ifndef _IPC_IPC_PORT_H_ 59 extern mach_port_name_t ipc_port_copyout_send( 60 ipc_port_t sright, 61 ipc_space_t space); 62 extern mach_port_name_t ipc_port_copyout_send_pinned( 63 ipc_port_t sright, 64 ipc_space_t space); 65 extern kern_return_t mach_port_deallocate_kernel( 66 ipc_space_t space, 67 mach_port_name_t name, 68 ipc_object_type_t otype); 69 #endif /* _IPC_IPC_PORT_H_ */ 70 71 #ifndef _KERN_IPC_TT_H_ 72 73 #define port_name_to_task(name) port_name_to_task_kernel(name) 74 75 extern task_t port_name_to_task_kernel( 76 mach_port_name_t name); 77 extern task_t port_name_to_task_read( 78 mach_port_name_t name); 79 extern task_t port_name_to_task_name( 80 mach_port_name_t name); 81 extern void ipc_port_release_send( 82 ipc_port_t port); 83 #endif /* _KERN_IPC_TT_H_ */ 84 85 extern ipc_space_t get_task_ipcspace( 86 task_t t); 87 88 #if CONFIG_MEMORYSTATUS 89 extern int max_task_footprint_mb; /* Per-task limit on physical memory consumption in megabytes */ 90 #endif /* CONFIG_MEMORYSTATUS */ 91 92 /* Some loose-ends VM stuff */ 93 94 extern const vm_size_t msg_ool_size_small; 95 96 extern kern_return_t vm_tests(void); 97 extern void consider_machine_adjust(void); 98 extern vm_map_offset_t get_map_min(vm_map_t); 99 extern vm_map_offset_t get_map_max(vm_map_t); 100 extern vm_map_size_t get_vmmap_size(vm_map_t); 101 extern int get_task_page_size(task_t); 102 #if CONFIG_COREDUMP 103 extern int get_vmmap_entries(vm_map_t); 104 #endif 105 extern int get_map_nentries(vm_map_t); 106 107 extern vm_map_offset_t vm_map_page_mask(vm_map_t); 108 #if HAS_MTE 109 extern kern_return_t vm_map_page_tags_get(vm_map_t map, vm_address_t page_addr, uint64_t *buf, vm_size_t size); 110 #endif /* HAS_MTE */ 111 112 #if MACH_ASSERT 113 extern void vm_map_pmap_set_process( 114 vm_map_t map, 115 int pid, 116 char *procname); 117 extern void vm_map_pmap_check_ledgers( 118 pmap_t pmap, 119 ledger_t ledger, 120 int pid, 121 char *procname); 122 #endif /* MACH_ASSERT */ 123 124 #if CONFIG_COREDUMP 125 extern boolean_t coredumpok(vm_map_t map, mach_vm_offset_t va); 126 #endif 127 128 #if XNU_PLATFORM_MacOSX 129 /* 130 * VM routines that used to be published to 131 * user space, and are now restricted to the kernel. 132 * 133 * No longer supported and always returns an error. 134 */ 135 extern kern_return_t vm_region_object_create 136 ( 137 vm_map_t target_task, 138 vm_size_t size, 139 ipc_port_t *object_handle 140 ); 141 #endif /* XNU_PLATFORM_MacOSX */ 142 143 #if CONFIG_CODE_DECRYPTION 144 #define VM_MAP_DEBUG_APPLE_PROTECT MACH_ASSERT 145 #if VM_MAP_DEBUG_APPLE_PROTECT 146 extern int vm_map_debug_apple_protect; 147 #endif /* VM_MAP_DEBUG_APPLE_PROTECT */ 148 struct pager_crypt_info; 149 extern kern_return_t vm_map_apple_protected( 150 vm_map_t map, 151 vm_map_offset_t start, 152 vm_map_offset_t end, 153 vm_object_offset_t crypto_backing_offset, 154 struct pager_crypt_info *crypt_info, 155 uint32_t cryptid); 156 #endif /* CONFIG_CODE_DECRYPTION */ 157 158 struct vm_shared_region_slide_info; 159 160 #if __has_feature(ptrauth_calls) 161 extern void shared_region_key_alloc( 162 char *shared_region_id, 163 bool inherit, 164 uint64_t inherited_key); 165 extern void shared_region_key_dealloc( 166 char *shared_region_id); 167 extern uint64_t generate_jop_key(void); 168 #endif /* __has_feature(ptrauth_calls) */ 169 extern bool vm_shared_region_is_reslide(struct task *task); 170 171 struct vnode; 172 extern memory_object_t swapfile_pager_setup(struct vnode *vp); 173 extern memory_object_control_t swapfile_pager_control(memory_object_t mem_obj); 174 175 #if __arm64__ || (__ARM_ARCH_7K__ >= 2) 176 #define SIXTEENK_PAGE_SIZE 0x4000 177 #define SIXTEENK_PAGE_MASK 0x3FFF 178 #define SIXTEENK_PAGE_SHIFT 14 179 #endif /* __arm64__ || (__ARM_ARCH_7K__ >= 2) */ 180 181 #define FOURK_PAGE_SIZE 0x1000 182 #define FOURK_PAGE_MASK 0xFFF 183 #define FOURK_PAGE_SHIFT 12 184 185 #if __arm64__ 186 extern unsigned int page_shift_user32; 187 #endif /* __arm64__ */ 188 189 /* 190 * bsd 191 */ 192 struct vnode; 193 194 extern void vnode_setswapmount(struct vnode *); 195 extern int64_t vnode_getswappin_avail(struct vnode *); 196 197 #if CHECK_CS_VALIDATION_BITMAP 198 /* used by the vnode_pager_cs_validation_bitmap routine*/ 199 #define CS_BITMAP_SET 1 200 #define CS_BITMAP_CLEAR 2 201 #define CS_BITMAP_CHECK 3 202 203 #endif /* CHECK_CS_VALIDATION_BITMAP */ 204 205 extern kern_return_t vnode_pager_init( 206 memory_object_t, 207 memory_object_control_t, 208 memory_object_cluster_size_t); 209 210 #if CONFIG_IOSCHED 211 extern kern_return_t vnode_pager_get_object_devvp( 212 memory_object_t, 213 uintptr_t *); 214 #endif 215 216 /* 217 * Functions defined in ubc_subr.c used by the vm code 218 */ 219 extern kern_return_t ubc_cs_check_validation_bitmap( 220 struct vnode *vp, 221 memory_object_offset_t offset, 222 int optype); 223 extern int ubc_map( 224 struct vnode *vp, 225 int flags); 226 extern void ubc_unmap( 227 struct vnode *vp); 228 229 230 extern void device_pager_reference(memory_object_t); 231 extern void device_pager_deallocate(memory_object_t); 232 extern kern_return_t device_pager_init(memory_object_t, 233 memory_object_control_t, 234 memory_object_cluster_size_t); 235 extern kern_return_t device_pager_terminate(memory_object_t); 236 extern kern_return_t device_pager_data_request(memory_object_t, 237 memory_object_offset_t, 238 memory_object_cluster_size_t, 239 vm_prot_t, 240 memory_object_fault_info_t); 241 extern kern_return_t device_pager_data_return(memory_object_t, 242 memory_object_offset_t, 243 memory_object_cluster_size_t, 244 memory_object_offset_t *, 245 int *, 246 boolean_t, 247 boolean_t, 248 int); 249 extern kern_return_t device_pager_data_initialize(memory_object_t, 250 memory_object_offset_t, 251 memory_object_cluster_size_t); 252 extern kern_return_t device_pager_map(memory_object_t, vm_prot_t); 253 extern kern_return_t device_pager_last_unmap(memory_object_t); 254 255 extern kern_return_t pager_map_to_phys_contiguous( 256 memory_object_control_t object, 257 memory_object_offset_t offset, 258 addr64_t base_vaddr, 259 vm_size_t size); 260 261 struct macx_triggers_args; 262 263 extern int macx_swapinfo( 264 memory_object_size_t *total_p, 265 memory_object_size_t *avail_p, 266 vm_size_t *pagesize_p, 267 boolean_t *encrypted_p); 268 269 270 struct proc; 271 struct proc *current_proc(void); 272 extern int cs_allow_invalid(struct proc *p); 273 extern int cs_invalid_page(addr64_t vaddr, boolean_t *cs_killed); 274 275 #define CS_VALIDATE_TAINTED 0x00000001 276 #define CS_VALIDATE_NX 0x00000002 277 extern boolean_t cs_validate_range(struct vnode *vp, 278 memory_object_t pager, 279 memory_object_offset_t offset, 280 const void *data, 281 vm_size_t size, 282 unsigned *result); 283 extern void cs_validate_page( 284 struct vnode *vp, 285 memory_object_t pager, 286 memory_object_offset_t offset, 287 const void *data, 288 int *validated_p, 289 int *tainted_p, 290 int *nx_p); 291 292 293 extern kern_return_t mach_memory_entry_purgable_control( 294 ipc_port_t entry_port, 295 vm_purgable_t control, 296 int *state); 297 298 extern unsigned int vmtc_total; /* total # of text page corruptions detected */ 299 300 extern kern_return_t revalidate_text_page(task_t, vm_map_offset_t); 301 302 #define VM_TOGGLE_CLEAR 0 303 #define VM_TOGGLE_SET 1 304 #define VM_TOGGLE_GETVALUE 999 305 int vm_toggle_entry_reuse(int, int*); 306 307 #define SWAP_WRITE 0x00000000 /* Write buffer (pseudo flag). */ 308 #define SWAP_READ 0x00000001 /* Read buffer. */ 309 #define SWAP_ASYNC 0x00000002 /* Start I/O, do not wait. */ 310 311 void do_fastwake_warmup_all(void); 312 313 #if CONFIG_JETSAM 314 extern int proc_get_memstat_priority(struct proc*, boolean_t); 315 #endif /* CONFIG_JETSAM */ 316 317 /* the object purger. purges the next eligible object from memory. */ 318 /* returns TRUE if an object was purged, otherwise FALSE. */ 319 boolean_t vm_purgeable_object_purge_one_unlocked(int force_purge_below_group); 320 void vm_owned_objects_disown(task_t task); 321 void vm_object_wired_page_update_ledgers( 322 vm_object_t object, 323 int64_t wired_delta); 324 325 struct trim_list { 326 uint64_t tl_offset; 327 uint64_t tl_length; 328 struct trim_list *tl_next; 329 }; 330 331 #define MAX_SWAPFILENAME_LEN 1024 332 #define SWAPFILENAME_INDEX_LEN 2 /* Doesn't include the terminating NULL character */ 333 334 extern char swapfilename[MAX_SWAPFILENAME_LEN + 1]; 335 336 struct vm_counters { 337 unsigned int do_collapse_compressor; 338 unsigned int do_collapse_compressor_pages; 339 unsigned int do_collapse_terminate; 340 unsigned int do_collapse_terminate_failure; 341 unsigned int should_cow_but_wired; 342 unsigned int create_upl_extra_cow; 343 unsigned int create_upl_extra_cow_pages; 344 unsigned int create_upl_lookup_failure_write; 345 unsigned int create_upl_lookup_failure_copy; 346 }; 347 extern struct vm_counters vm_counters; 348 349 #if CONFIG_SECLUDED_MEMORY 350 struct vm_page_secluded_data { 351 int eligible_for_secluded; 352 int grab_success_free; 353 int grab_success_other; 354 int grab_failure_locked; 355 int grab_failure_state; 356 int grab_failure_realtime; 357 int grab_failure_dirty; 358 int grab_for_iokit; 359 int grab_for_iokit_success; 360 }; 361 extern struct vm_page_secluded_data vm_page_secluded; 362 363 extern int num_tasks_can_use_secluded_mem; 364 365 /* boot-args */ 366 367 __enum_decl(secluded_filecache_mode_t, uint8_t, { 368 /* 369 * SECLUDED_FILECACHE_NONE: 370 * + no file contents in secluded pool 371 */ 372 SECLUDED_FILECACHE_NONE = 0, 373 /* 374 * SECLUDED_FILECACHE_APPS 375 * + no files from / 376 * + files from /Applications/ are OK 377 * + files from /Applications/Camera are not OK 378 * + no files that are open for write 379 */ 380 SECLUDED_FILECACHE_APPS = 1, 381 /* 382 * SECLUDED_FILECACHE_RDONLY 383 * + all read-only files OK, except: 384 * + dyld_shared_cache_arm64* 385 * + Camera 386 * + mediaserverd 387 * + cameracaptured 388 */ 389 SECLUDED_FILECACHE_RDONLY = 2, 390 }); 391 392 extern secluded_filecache_mode_t secluded_for_filecache; 393 extern bool secluded_for_apps; 394 extern bool secluded_for_iokit; 395 396 extern uint64_t vm_page_secluded_drain(void); 397 extern void memory_object_mark_eligible_for_secluded( 398 memory_object_control_t control, 399 boolean_t eligible_for_secluded); 400 401 #endif /* CONFIG_SECLUDED_MEMORY */ 402 403 extern void memory_object_mark_for_realtime( 404 memory_object_control_t control, 405 bool for_realtime); 406 407 #define MAX_PAGE_RANGE_QUERY (1ULL * 1024 * 1024 * 1024) /* 1 GB */ 408 409 extern uint64_t vm_purge_filebacked_pagers(void); 410 411 #define roundup(x, y) ((((x) % (y)) == 0) ? \ 412 (x) : ((x) + ((y) - ((x) % (y))))) 413 414 #define rounddown(x, y) (((x)/(y))*(y)) 415 416 #ifdef __cplusplus 417 } 418 #endif 419 420 /* 421 * Flags for the VM swapper/reclaimer. 422 * Used by vm_swap_consider_defragment() 423 * to force defrag/reclaim by the swap 424 * GC thread. 425 */ 426 #define VM_SWAP_FLAGS_NONE 0 427 #define VM_SWAP_FLAGS_FORCE_DEFRAG 1 428 #define VM_SWAP_FLAGS_FORCE_RECLAIM 2 429 430 #if __arm64__ 431 /* 432 * Flags to control the behavior of 433 * the legacy footprint entitlement. 434 */ 435 #define LEGACY_FOOTPRINT_ENTITLEMENT_IGNORE (1) 436 #define LEGACY_FOOTPRINT_ENTITLEMENT_IOS11_ACCT (2) 437 #define LEGACY_FOOTPRINT_ENTITLEMENT_LIMIT_INCREASE (3) 438 439 #endif /* __arm64__ */ 440 441 #if DEVELOPMENT || DEBUG 442 struct proc; 443 extern struct proc *current_proc(void); 444 extern int proc_pid(struct proc *); 445 extern const char *proc_best_name(struct proc *); 446 struct thread; 447 extern uint64_t thread_tid(struct thread *); 448 extern int debug4k_filter; 449 extern int debug4k_proc_filter; 450 extern char debug4k_proc_name[]; 451 extern const char *debug4k_category_name[]; 452 453 #define __DEBUG4K(category, fmt, ...) \ 454 MACRO_BEGIN \ 455 int __category = (category); \ 456 struct thread *__t = NULL; \ 457 struct proc *__p = NULL; \ 458 const char *__pname = "?"; \ 459 boolean_t __do_log = FALSE; \ 460 \ 461 if ((1 << __category) & debug4k_filter) { \ 462 __do_log = TRUE; \ 463 } else if (((1 << __category) & debug4k_proc_filter) && \ 464 debug4k_proc_name[0] != '\0') { \ 465 __p = current_proc(); \ 466 if (__p != NULL) { \ 467 __pname = proc_best_name(__p); \ 468 } \ 469 if (!strcmp(debug4k_proc_name, __pname)) { \ 470 __do_log = TRUE; \ 471 } \ 472 } \ 473 if (__do_log) { \ 474 if (__p == NULL) { \ 475 __p = current_proc(); \ 476 if (__p != NULL) { \ 477 __pname = proc_best_name(__p); \ 478 } \ 479 } \ 480 __t = current_thread(); \ 481 printf("DEBUG4K(%s) %d[%s] %p(0x%llx) %s:%d: " fmt, \ 482 debug4k_category_name[__category], \ 483 __p ? proc_pid(__p) : 0, \ 484 __pname, \ 485 __t, \ 486 thread_tid(__t), \ 487 __FUNCTION__, \ 488 __LINE__, \ 489 ##__VA_ARGS__); \ 490 } \ 491 MACRO_END 492 493 #define __DEBUG4K_ERROR 0 494 #define __DEBUG4K_LIFE 1 495 #define __DEBUG4K_LOAD 2 496 #define __DEBUG4K_FAULT 3 497 #define __DEBUG4K_COPY 4 498 #define __DEBUG4K_SHARE 5 499 #define __DEBUG4K_ADJUST 6 500 #define __DEBUG4K_PMAP 7 501 #define __DEBUG4K_MEMENTRY 8 502 #define __DEBUG4K_IOKIT 9 503 #define __DEBUG4K_UPL 10 504 #define __DEBUG4K_EXC 11 505 #define __DEBUG4K_VFS 12 506 507 #define DEBUG4K_ERROR(...) __DEBUG4K(__DEBUG4K_ERROR, ##__VA_ARGS__) 508 #define DEBUG4K_LIFE(...) __DEBUG4K(__DEBUG4K_LIFE, ##__VA_ARGS__) 509 #define DEBUG4K_LOAD(...) __DEBUG4K(__DEBUG4K_LOAD, ##__VA_ARGS__) 510 #define DEBUG4K_FAULT(...) __DEBUG4K(__DEBUG4K_FAULT, ##__VA_ARGS__) 511 #define DEBUG4K_COPY(...) __DEBUG4K(__DEBUG4K_COPY, ##__VA_ARGS__) 512 #define DEBUG4K_SHARE(...) __DEBUG4K(__DEBUG4K_SHARE, ##__VA_ARGS__) 513 #define DEBUG4K_ADJUST(...) __DEBUG4K(__DEBUG4K_ADJUST, ##__VA_ARGS__) 514 #define DEBUG4K_PMAP(...) __DEBUG4K(__DEBUG4K_PMAP, ##__VA_ARGS__) 515 #define DEBUG4K_MEMENTRY(...) __DEBUG4K(__DEBUG4K_MEMENTRY, ##__VA_ARGS__) 516 #define DEBUG4K_IOKIT(...) __DEBUG4K(__DEBUG4K_IOKIT, ##__VA_ARGS__) 517 #define DEBUG4K_UPL(...) __DEBUG4K(__DEBUG4K_UPL, ##__VA_ARGS__) 518 #define DEBUG4K_EXC(...) __DEBUG4K(__DEBUG4K_EXC, ##__VA_ARGS__) 519 #define DEBUG4K_VFS(...) __DEBUG4K(__DEBUG4K_VFS, ##__VA_ARGS__) 520 521 #else /* DEVELOPMENT || DEBUG */ 522 523 #define DEBUG4K_ERROR(...) 524 #define DEBUG4K_LIFE(...) 525 #define DEBUG4K_LOAD(...) 526 #define DEBUG4K_FAULT(...) 527 #define DEBUG4K_COPY(...) 528 #define DEBUG4K_SHARE(...) 529 #define DEBUG4K_ADJUST(...) 530 #define DEBUG4K_PMAP(...) 531 #define DEBUG4K_MEMENTRY(...) 532 #define DEBUG4K_IOKIT(...) 533 #define DEBUG4K_UPL(...) 534 #define DEBUG4K_EXC(...) 535 #define DEBUG4K_VFS(...) 536 537 #endif /* DEVELOPMENT || DEBUG */ 538 539 540 __enum_decl(vm_object_destroy_reason_t, uint8_t, { 541 VM_OBJECT_DESTROY_UNKNOWN_REASON = 0, 542 VM_OBJECT_DESTROY_RECLAIM = 1, 543 VM_OBJECT_DESTROY_UNMOUNT = 2, 544 VM_OBJECT_DESTROY_FORCED_UNMOUNT = 3, 545 VM_OBJECT_DESTROY_UNGRAFT = 4, 546 VM_OBJECT_DESTROY_PAGER = 5, 547 VM_OBJECT_DESTROY_MAX = 5, 548 }); 549 _Static_assert(VM_OBJECT_DESTROY_MAX < 8, "Need to fit in `no_pager_reason`'s number of bits"); 550 551 /* From vm_resident.c */ 552 void vm_update_darkwake_mode(boolean_t); 553 554 #if FBDP_DEBUG_OBJECT_NO_PAGER 555 extern kern_return_t memory_object_mark_as_tracked( 556 memory_object_control_t control, 557 bool new_value, 558 bool *old_value); 559 #endif /* FBDP_DEBUG_OBJECT_NO_PAGER */ 560 561 #endif /* _VM_VM_PROTOS_H_ */ 562 563 #endif /* XNU_KERNEL_PRIVATE */ 564