/* * Copyright (c) 2004-2020 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in * compliance with the License. The rights granted to you under the License * may not be used to create, or enable the creation or redistribution of, * unlawful or unlicensed copies of an Apple operating system, or to * circumvent, violate, or enable the circumvention or violation of, any * terms of an Apple operating system software license agreement. * * Please obtain a copy of the License at * http://www.opensource.apple.com/apsl/ and read it before using this file. * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. * Please see the License for the specific language governing rights and * limitations under the License. * * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ #ifdef XNU_KERNEL_PRIVATE #ifndef _VM_VM_PROTOS_H_ #define _VM_VM_PROTOS_H_ #include #include #include #ifdef __cplusplus extern "C" { #endif /* * This file contains various type definitions and routine prototypes * that are needed to avoid compilation warnings for VM code (in osfmk, * default_pager and bsd). * Most of these should eventually go into more appropriate header files. * * Include it after all other header files since it doesn't include any * type definitions and it works around some conflicts with other header * files. */ /* * iokit */ extern kern_return_t device_data_action( uintptr_t device_handle, ipc_port_t device_pager, vm_prot_t protection, vm_object_offset_t offset, vm_size_t size); extern kern_return_t device_close( uintptr_t device_handle); extern boolean_t vm_swap_files_pinned(void); /* * osfmk */ #ifndef _IPC_IPC_PORT_H_ extern mach_port_name_t ipc_port_copyout_send( ipc_port_t sright, ipc_space_t space); extern mach_port_name_t ipc_port_copyout_send_pinned( ipc_port_t sright, ipc_space_t space); #endif /* _IPC_IPC_PORT_H_ */ #ifndef _KERN_IPC_TT_H_ #define port_name_to_task(name) port_name_to_task_kernel(name) extern task_t port_name_to_task_kernel( mach_port_name_t name); extern task_t port_name_to_task_read( mach_port_name_t name); extern task_t port_name_to_task_name( mach_port_name_t name); extern void ipc_port_release_send( ipc_port_t port); #endif /* _KERN_IPC_TT_H_ */ extern ipc_space_t get_task_ipcspace( task_t t); #if CONFIG_MEMORYSTATUS extern int max_task_footprint_mb; /* Per-task limit on physical memory consumption in megabytes */ #endif /* CONFIG_MEMORYSTATUS */ /* Some loose-ends VM stuff */ extern const vm_size_t msg_ool_size_small; extern kern_return_t vm_tests(void); extern void consider_machine_adjust(void); extern vm_map_offset_t get_map_min(vm_map_t); extern vm_map_offset_t get_map_max(vm_map_t); extern vm_map_size_t get_vmmap_size(vm_map_t); extern int get_task_page_size(task_t); #if CONFIG_COREDUMP extern int get_vmmap_entries(vm_map_t); #endif extern int get_map_nentries(vm_map_t); extern vm_map_offset_t vm_map_page_mask(vm_map_t); extern kern_return_t vm_map_purgable_control( vm_map_t map, vm_map_offset_t address, vm_purgable_t control, int *state); #if MACH_ASSERT extern void vm_map_pmap_set_process( vm_map_t map, int pid, char *procname); extern void vm_map_pmap_check_ledgers( pmap_t pmap, ledger_t ledger, int pid, char *procname); #endif /* MACH_ASSERT */ extern kern_return_t vnode_pager_get_object_vnode( memory_object_t mem_obj, uintptr_t * vnodeaddr, uint32_t * vid); #if CONFIG_COREDUMP extern boolean_t coredumpok(vm_map_t map, mach_vm_offset_t va); #endif /* * VM routines that used to be published to * user space, and are now restricted to the kernel. * * They should eventually go away entirely - * to be replaced with standard vm_map() and * vm_deallocate() calls. */ extern kern_return_t vm_upl_map ( vm_map_t target_task, upl_t upl, vm_address_t *address ); extern kern_return_t vm_upl_unmap ( vm_map_t target_task, upl_t upl ); extern kern_return_t vm_upl_map_range ( vm_map_t target_task, upl_t upl, vm_offset_t offset, vm_size_t size, vm_prot_t prot, vm_address_t *address ); extern kern_return_t vm_upl_unmap_range ( vm_map_t target_task, upl_t upl, vm_offset_t offset, vm_size_t size ); extern kern_return_t vm_region_object_create ( vm_map_t target_task, vm_size_t size, ipc_port_t *object_handle ); #if CONFIG_CODE_DECRYPTION #define VM_MAP_DEBUG_APPLE_PROTECT MACH_ASSERT #if VM_MAP_DEBUG_APPLE_PROTECT extern int vm_map_debug_apple_protect; #endif /* VM_MAP_DEBUG_APPLE_PROTECT */ struct pager_crypt_info; extern kern_return_t vm_map_apple_protected( vm_map_t map, vm_map_offset_t start, vm_map_offset_t end, vm_object_offset_t crypto_backing_offset, struct pager_crypt_info *crypt_info, uint32_t cryptid); extern memory_object_t apple_protect_pager_setup( vm_object_t backing_object, vm_object_offset_t backing_offset, vm_object_offset_t crypto_backing_offset, struct pager_crypt_info *crypt_info, vm_object_offset_t crypto_start, vm_object_offset_t crypto_end, boolean_t cache_pager); #endif /* CONFIG_CODE_DECRYPTION */ struct vm_shared_region_slide_info; extern kern_return_t vm_map_shared_region( vm_map_t map, vm_map_offset_t start, vm_map_offset_t end, vm_object_offset_t backing_offset, struct vm_shared_region_slide_info *slide_info); extern memory_object_t shared_region_pager_setup( vm_object_t backing_object, vm_object_offset_t backing_offset, struct vm_shared_region_slide_info *slide_info, uint64_t jop_key); extern uint64_t apple_protect_pager_purge_all(void); extern uint64_t shared_region_pager_purge_all(void); extern uint64_t dyld_pager_purge_all(void); #if __has_feature(ptrauth_calls) extern memory_object_t shared_region_pager_match( vm_object_t backing_object, vm_object_offset_t backing_offset, struct vm_shared_region_slide_info *slide_info, uint64_t jop_key); extern void shared_region_key_alloc( char *shared_region_id, bool inherit, uint64_t inherited_key); extern void shared_region_key_dealloc( char *shared_region_id); extern uint64_t generate_jop_key(void); extern void shared_region_pager_match_task_key(memory_object_t memobj, task_t task); #endif /* __has_feature(ptrauth_calls) */ extern bool vm_shared_region_is_reslide(struct task *task); struct vnode; extern memory_object_t swapfile_pager_setup(struct vnode *vp); extern memory_object_control_t swapfile_pager_control(memory_object_t mem_obj); #if __arm64__ || (__ARM_ARCH_7K__ >= 2) #define SIXTEENK_PAGE_SIZE 0x4000 #define SIXTEENK_PAGE_MASK 0x3FFF #define SIXTEENK_PAGE_SHIFT 14 #endif /* __arm64__ || (__ARM_ARCH_7K__ >= 2) */ #define FOURK_PAGE_SIZE 0x1000 #define FOURK_PAGE_MASK 0xFFF #define FOURK_PAGE_SHIFT 12 #if __arm64__ extern unsigned int page_shift_user32; #define VM_MAP_DEBUG_FOURK MACH_ASSERT #if VM_MAP_DEBUG_FOURK extern int vm_map_debug_fourk; #endif /* VM_MAP_DEBUG_FOURK */ extern memory_object_t fourk_pager_create(void); extern vm_object_t fourk_pager_to_vm_object(memory_object_t mem_obj); extern kern_return_t fourk_pager_populate( memory_object_t mem_obj, boolean_t overwrite, int index, vm_object_t new_backing_object, vm_object_offset_t new_backing_offset, vm_object_t *old_backing_object, vm_object_offset_t *old_backing_offset); #endif /* __arm64__ */ /* * bsd */ struct vnode; extern void vnode_setswapmount(struct vnode *); extern int64_t vnode_getswappin_avail(struct vnode *); extern void vnode_pager_was_dirtied( struct vnode *, vm_object_offset_t, vm_object_offset_t); typedef int pager_return_t; extern pager_return_t vnode_pagein( struct vnode *, upl_t, upl_offset_t, vm_object_offset_t, upl_size_t, int, int *); extern pager_return_t vnode_pageout( struct vnode *, upl_t, upl_offset_t, vm_object_offset_t, upl_size_t, int, int *); extern uint32_t vnode_trim(struct vnode *, int64_t offset, unsigned long len); extern memory_object_t vnode_pager_setup( struct vnode *, memory_object_t); extern vm_object_offset_t vnode_pager_get_filesize( struct vnode *); extern uint32_t vnode_pager_isinuse( struct vnode *); extern boolean_t vnode_pager_isSSD( struct vnode *); #if FBDP_DEBUG_OBJECT_NO_PAGER extern bool vnode_pager_forced_unmount( struct vnode *); #endif /* FBDP_DEBUG_OBJECT_NO_PAGER */ extern void vnode_pager_throttle( void); extern uint32_t vnode_pager_return_throttle_io_limit( struct vnode *, uint32_t *); extern kern_return_t vnode_pager_get_name( struct vnode *vp, char *pathname, vm_size_t pathname_len, char *filename, vm_size_t filename_len, boolean_t *truncated_path_p); struct timespec; extern kern_return_t vnode_pager_get_mtime( struct vnode *vp, struct timespec *mtime, struct timespec *cs_mtime); extern kern_return_t vnode_pager_get_cs_blobs( struct vnode *vp, void **blobs); #if CONFIG_IOSCHED void vnode_pager_issue_reprioritize_io( struct vnode *devvp, uint64_t blkno, uint32_t len, int priority); #endif #if CHECK_CS_VALIDATION_BITMAP /* used by the vnode_pager_cs_validation_bitmap routine*/ #define CS_BITMAP_SET 1 #define CS_BITMAP_CLEAR 2 #define CS_BITMAP_CHECK 3 #endif /* CHECK_CS_VALIDATION_BITMAP */ extern kern_return_t vnode_pager_data_unlock( memory_object_t mem_obj, memory_object_offset_t offset, memory_object_size_t size, vm_prot_t desired_access); extern kern_return_t vnode_pager_init( memory_object_t, memory_object_control_t, memory_object_cluster_size_t); extern kern_return_t vnode_pager_get_object_size( memory_object_t, memory_object_offset_t *); #if CONFIG_IOSCHED extern kern_return_t vnode_pager_get_object_devvp( memory_object_t, uintptr_t *); #endif extern void vnode_pager_dirtied( memory_object_t, vm_object_offset_t, vm_object_offset_t); extern kern_return_t vnode_pager_get_isinuse( memory_object_t, uint32_t *); extern kern_return_t vnode_pager_get_isSSD( memory_object_t, boolean_t *); #if FBDP_DEBUG_OBJECT_NO_PAGER extern kern_return_t vnode_pager_get_forced_unmount( memory_object_t, bool *); #endif /* FBDP_DEBUG_OBJECT_NO_PAGER */ extern kern_return_t vnode_pager_get_throttle_io_limit( memory_object_t, uint32_t *); extern kern_return_t vnode_pager_get_object_name( memory_object_t mem_obj, char *pathname, vm_size_t pathname_len, char *filename, vm_size_t filename_len, boolean_t *truncated_path_p); extern kern_return_t vnode_pager_get_object_mtime( memory_object_t mem_obj, struct timespec *mtime, struct timespec *cs_mtime); #if CHECK_CS_VALIDATION_BITMAP extern kern_return_t vnode_pager_cs_check_validation_bitmap( memory_object_t mem_obj, memory_object_offset_t offset, int optype); #endif /*CHECK_CS_VALIDATION_BITMAP*/ extern kern_return_t ubc_cs_check_validation_bitmap( struct vnode *vp, memory_object_offset_t offset, int optype); extern kern_return_t vnode_pager_data_request( memory_object_t, memory_object_offset_t, memory_object_cluster_size_t, vm_prot_t, memory_object_fault_info_t); extern kern_return_t vnode_pager_data_return( memory_object_t, memory_object_offset_t, memory_object_cluster_size_t, memory_object_offset_t *, int *, boolean_t, boolean_t, int); extern kern_return_t vnode_pager_data_initialize( memory_object_t, memory_object_offset_t, memory_object_cluster_size_t); extern void vnode_pager_reference( memory_object_t mem_obj); extern kern_return_t vnode_pager_map( memory_object_t mem_obj, vm_prot_t prot); extern kern_return_t vnode_pager_last_unmap( memory_object_t mem_obj); extern void vnode_pager_deallocate( memory_object_t); extern kern_return_t vnode_pager_terminate( memory_object_t); extern void vnode_pager_vrele( struct vnode *vp); extern struct vnode *vnode_pager_lookup_vnode( memory_object_t); extern int ubc_map( struct vnode *vp, int flags); extern void ubc_unmap( struct vnode *vp); struct vm_map_entry; extern struct vm_object *find_vnode_object(struct vm_map_entry *entry); extern void device_pager_reference(memory_object_t); extern void device_pager_deallocate(memory_object_t); extern kern_return_t device_pager_init(memory_object_t, memory_object_control_t, memory_object_cluster_size_t); extern kern_return_t device_pager_terminate(memory_object_t); extern kern_return_t device_pager_data_request(memory_object_t, memory_object_offset_t, memory_object_cluster_size_t, vm_prot_t, memory_object_fault_info_t); extern kern_return_t device_pager_data_return(memory_object_t, memory_object_offset_t, memory_object_cluster_size_t, memory_object_offset_t *, int *, boolean_t, boolean_t, int); extern kern_return_t device_pager_data_initialize(memory_object_t, memory_object_offset_t, memory_object_cluster_size_t); extern kern_return_t device_pager_map(memory_object_t, vm_prot_t); extern kern_return_t device_pager_last_unmap(memory_object_t); extern kern_return_t device_pager_populate_object( memory_object_t device, memory_object_offset_t offset, ppnum_t page_num, vm_size_t size); extern memory_object_t device_pager_setup( memory_object_t, uintptr_t, vm_size_t, int); extern boolean_t is_device_pager_ops(const struct memory_object_pager_ops *pager_ops); extern kern_return_t pager_map_to_phys_contiguous( memory_object_control_t object, memory_object_offset_t offset, addr64_t base_vaddr, vm_size_t size); extern kern_return_t memory_object_create_named( memory_object_t pager, memory_object_offset_t size, memory_object_control_t *control); struct macx_triggers_args; extern int mach_macx_triggers( struct macx_triggers_args *args); extern int macx_swapinfo( memory_object_size_t *total_p, memory_object_size_t *avail_p, vm_size_t *pagesize_p, boolean_t *encrypted_p); extern void log_stack_execution_failure(addr64_t vaddr, vm_prot_t prot); extern void log_unnest_badness( vm_map_t map, vm_map_offset_t start_unnest, vm_map_offset_t end_unnest, boolean_t is_nested_map, vm_map_offset_t lowest_unnestable_addr); struct proc; struct proc *current_proc(void); extern int cs_allow_invalid(struct proc *p); extern int cs_invalid_page(addr64_t vaddr, boolean_t *cs_killed); #define CS_VALIDATE_TAINTED 0x00000001 #define CS_VALIDATE_NX 0x00000002 extern boolean_t cs_validate_range(struct vnode *vp, memory_object_t pager, memory_object_offset_t offset, const void *data, vm_size_t size, unsigned *result); extern void cs_validate_page( struct vnode *vp, memory_object_t pager, memory_object_offset_t offset, const void *data, int *validated_p, int *tainted_p, int *nx_p); extern kern_return_t memory_entry_purgeable_control_internal( ipc_port_t entry_port, vm_purgable_t control, int *state); extern kern_return_t memory_entry_access_tracking_internal( ipc_port_t entry_port, int *access_tracking, uint32_t *access_tracking_reads, uint32_t *access_tracking_writes); extern kern_return_t mach_memory_object_memory_entry_64( host_t host, boolean_t internal, vm_object_offset_t size, vm_prot_t permission, memory_object_t pager, ipc_port_t *entry_handle); extern kern_return_t mach_memory_entry_purgable_control( ipc_port_t entry_port, vm_purgable_t control, int *state); extern kern_return_t mach_memory_entry_get_page_counts( ipc_port_t entry_port, unsigned int *resident_page_count, unsigned int *dirty_page_count); extern kern_return_t mach_memory_entry_phys_page_offset( ipc_port_t entry_port, vm_object_offset_t *offset_p); extern kern_return_t mach_memory_entry_map_size( ipc_port_t entry_port, vm_map_t map, memory_object_offset_t offset, memory_object_offset_t size, mach_vm_size_t *map_size); extern kern_return_t vm_map_range_physical_size( vm_map_t map, vm_map_address_t start, mach_vm_size_t size, mach_vm_size_t * phys_size); extern kern_return_t mach_memory_entry_page_op( ipc_port_t entry_port, vm_object_offset_t offset, int ops, ppnum_t *phys_entry, int *flags); extern kern_return_t mach_memory_entry_range_op( ipc_port_t entry_port, vm_object_offset_t offset_beg, vm_object_offset_t offset_end, int ops, int *range); extern void mach_memory_entry_port_release(ipc_port_t port); extern vm_named_entry_t mach_memory_entry_from_port(ipc_port_t port); extern struct vm_named_entry *mach_memory_entry_allocate(ipc_port_t *user_handle_p); extern vm_object_t vm_named_entry_to_vm_object( vm_named_entry_t named_entry); extern void vm_named_entry_associate_vm_object( vm_named_entry_t named_entry, vm_object_t object, vm_object_offset_t offset, vm_object_size_t size, vm_prot_t prot); extern int macx_backing_store_compaction(int flags); extern unsigned int mach_vm_ctl_page_free_wanted(void); extern int no_paging_space_action(void); extern unsigned int vmtc_total; /* total # of text page corruptions detected */ extern kern_return_t revalidate_text_page(task_t, vm_map_offset_t); #define VM_TOGGLE_CLEAR 0 #define VM_TOGGLE_SET 1 #define VM_TOGGLE_GETVALUE 999 int vm_toggle_entry_reuse(int, int*); #define SWAP_WRITE 0x00000000 /* Write buffer (pseudo flag). */ #define SWAP_READ 0x00000001 /* Read buffer. */ #define SWAP_ASYNC 0x00000002 /* Start I/O, do not wait. */ extern kern_return_t compressor_memory_object_create( memory_object_size_t, memory_object_t *); extern boolean_t vm_compressor_low_on_space(void); extern bool vm_compressor_compressed_pages_nearing_limit(void); extern boolean_t vm_compressor_out_of_space(void); extern int vm_swap_low_on_space(void); extern int vm_swap_out_of_space(void); void do_fastwake_warmup_all(void); #if defined(__arm64__) extern void vm_panic_hibernate_write_image_failed(int err); #endif /* __arm64__ */ #if CONFIG_JETSAM extern int proc_get_memstat_priority(struct proc*, boolean_t); #endif /* CONFIG_JETSAM */ /* the object purger. purges the next eligible object from memory. */ /* returns TRUE if an object was purged, otherwise FALSE. */ boolean_t vm_purgeable_object_purge_one_unlocked(int force_purge_below_group); void vm_purgeable_nonvolatile_owner_update(task_t owner, int delta); void vm_purgeable_volatile_owner_update(task_t owner, int delta); void vm_owned_objects_disown(task_t task); struct trim_list { uint64_t tl_offset; uint64_t tl_length; struct trim_list *tl_next; }; u_int32_t vnode_trim_list(struct vnode *vp, struct trim_list *tl, boolean_t route_only); #define MAX_SWAPFILENAME_LEN 1024 #define SWAPFILENAME_INDEX_LEN 2 /* Doesn't include the terminating NULL character */ extern char swapfilename[MAX_SWAPFILENAME_LEN + 1]; struct vm_counters { unsigned int do_collapse_compressor; unsigned int do_collapse_compressor_pages; unsigned int do_collapse_terminate; unsigned int do_collapse_terminate_failure; unsigned int should_cow_but_wired; unsigned int create_upl_extra_cow; unsigned int create_upl_extra_cow_pages; unsigned int create_upl_lookup_failure_write; unsigned int create_upl_lookup_failure_copy; }; extern struct vm_counters vm_counters; #if CONFIG_SECLUDED_MEMORY struct vm_page_secluded_data { int eligible_for_secluded; int grab_success_free; int grab_success_other; int grab_failure_locked; int grab_failure_state; int grab_failure_realtime; int grab_failure_dirty; int grab_for_iokit; int grab_for_iokit_success; }; extern struct vm_page_secluded_data vm_page_secluded; extern int num_tasks_can_use_secluded_mem; /* boot-args */ __enum_decl(secluded_filecache_mode_t, uint8_t, { /* * SECLUDED_FILECACHE_NONE: * + no file contents in secluded pool */ SECLUDED_FILECACHE_NONE = 0, /* * SECLUDED_FILECACHE_APPS * + no files from / * + files from /Applications/ are OK * + files from /Applications/Camera are not OK * + no files that are open for write */ SECLUDED_FILECACHE_APPS = 1, /* * SECLUDED_FILECACHE_RDONLY * + all read-only files OK, except: * + dyld_shared_cache_arm64* * + Camera * + mediaserverd */ SECLUDED_FILECACHE_RDONLY = 2, }); extern secluded_filecache_mode_t secluded_for_filecache; extern bool secluded_for_apps; extern bool secluded_for_iokit; extern uint64_t vm_page_secluded_drain(void); extern void memory_object_mark_eligible_for_secluded( memory_object_control_t control, boolean_t eligible_for_secluded); #endif /* CONFIG_SECLUDED_MEMORY */ extern void memory_object_mark_for_realtime( memory_object_control_t control, bool for_realtime); #if FBDP_DEBUG_OBJECT_NO_PAGER extern kern_return_t memory_object_mark_as_tracked( memory_object_control_t control, bool new_value, bool *old_value); #endif /* FBDP_DEBUG_OBJECT_NO_PAGER */ #define MAX_PAGE_RANGE_QUERY (1ULL * 1024 * 1024 * 1024) /* 1 GB */ extern kern_return_t mach_make_memory_entry_internal( vm_map_t target_map, memory_object_size_t *size, memory_object_offset_t offset, vm_prot_t permission, vm_named_entry_kernel_flags_t vmne_kflags, ipc_port_t *object_handle, ipc_port_t parent_handle); extern kern_return_t memory_entry_check_for_adjustment( vm_map_t src_map, ipc_port_t port, vm_map_offset_t *overmap_start, vm_map_offset_t *overmap_end); extern uint64_t vm_purge_filebacked_pagers(void); #define roundup(x, y) ((((x) % (y)) == 0) ? \ (x) : ((x) + ((y) - ((x) % (y))))) #ifdef __cplusplus } #endif /* * Flags for the VM swapper/reclaimer. * Used by vm_swap_consider_defragment() * to force defrag/reclaim by the swap * GC thread. */ #define VM_SWAP_FLAGS_NONE 0 #define VM_SWAP_FLAGS_FORCE_DEFRAG 1 #define VM_SWAP_FLAGS_FORCE_RECLAIM 2 #if __arm64__ /* * Flags to control the behavior of * the legacy footprint entitlement. */ #define LEGACY_FOOTPRINT_ENTITLEMENT_IGNORE (1) #define LEGACY_FOOTPRINT_ENTITLEMENT_IOS11_ACCT (2) #define LEGACY_FOOTPRINT_ENTITLEMENT_LIMIT_INCREASE (3) #endif /* __arm64__ */ #if MACH_ASSERT struct proc; extern struct proc *current_proc(void); extern int proc_pid(struct proc *); extern char *proc_best_name(struct proc *); struct thread; extern uint64_t thread_tid(struct thread *); extern int debug4k_filter; extern int debug4k_proc_filter; extern char debug4k_proc_name[]; extern const char *debug4k_category_name[]; #define __DEBUG4K(category, fmt, ...) \ MACRO_BEGIN \ int __category = (category); \ struct thread *__t = NULL; \ struct proc *__p = NULL; \ const char *__pname = "?"; \ boolean_t __do_log = FALSE; \ \ if ((1 << __category) & debug4k_filter) { \ __do_log = TRUE; \ } else if (((1 << __category) & debug4k_proc_filter) && \ debug4k_proc_name[0] != '\0') { \ __p = current_proc(); \ if (__p != NULL) { \ __pname = proc_best_name(__p); \ } \ if (!strcmp(debug4k_proc_name, __pname)) { \ __do_log = TRUE; \ } \ } \ if (__do_log) { \ if (__p == NULL) { \ __p = current_proc(); \ if (__p != NULL) { \ __pname = proc_best_name(__p); \ } \ } \ __t = current_thread(); \ printf("DEBUG4K(%s) %d[%s] %p(0x%llx) %s:%d: " fmt, \ debug4k_category_name[__category], \ __p ? proc_pid(__p) : 0, \ __pname, \ __t, \ thread_tid(__t), \ __FUNCTION__, \ __LINE__, \ ##__VA_ARGS__); \ } \ MACRO_END #define __DEBUG4K_ERROR 0 #define __DEBUG4K_LIFE 1 #define __DEBUG4K_LOAD 2 #define __DEBUG4K_FAULT 3 #define __DEBUG4K_COPY 4 #define __DEBUG4K_SHARE 5 #define __DEBUG4K_ADJUST 6 #define __DEBUG4K_PMAP 7 #define __DEBUG4K_MEMENTRY 8 #define __DEBUG4K_IOKIT 9 #define __DEBUG4K_UPL 10 #define __DEBUG4K_EXC 11 #define __DEBUG4K_VFS 12 #define DEBUG4K_ERROR(...) __DEBUG4K(__DEBUG4K_ERROR, ##__VA_ARGS__) #define DEBUG4K_LIFE(...) __DEBUG4K(__DEBUG4K_LIFE, ##__VA_ARGS__) #define DEBUG4K_LOAD(...) __DEBUG4K(__DEBUG4K_LOAD, ##__VA_ARGS__) #define DEBUG4K_FAULT(...) __DEBUG4K(__DEBUG4K_FAULT, ##__VA_ARGS__) #define DEBUG4K_COPY(...) __DEBUG4K(__DEBUG4K_COPY, ##__VA_ARGS__) #define DEBUG4K_SHARE(...) __DEBUG4K(__DEBUG4K_SHARE, ##__VA_ARGS__) #define DEBUG4K_ADJUST(...) __DEBUG4K(__DEBUG4K_ADJUST, ##__VA_ARGS__) #define DEBUG4K_PMAP(...) __DEBUG4K(__DEBUG4K_PMAP, ##__VA_ARGS__) #define DEBUG4K_MEMENTRY(...) __DEBUG4K(__DEBUG4K_MEMENTRY, ##__VA_ARGS__) #define DEBUG4K_IOKIT(...) __DEBUG4K(__DEBUG4K_IOKIT, ##__VA_ARGS__) #define DEBUG4K_UPL(...) __DEBUG4K(__DEBUG4K_UPL, ##__VA_ARGS__) #define DEBUG4K_EXC(...) __DEBUG4K(__DEBUG4K_EXC, ##__VA_ARGS__) #define DEBUG4K_VFS(...) __DEBUG4K(__DEBUG4K_VFS, ##__VA_ARGS__) #else /* MACH_ASSERT */ #define DEBUG4K_ERROR(...) #define DEBUG4K_LIFE(...) #define DEBUG4K_LOAD(...) #define DEBUG4K_FAULT(...) #define DEBUG4K_COPY(...) #define DEBUG4K_SHARE(...) #define DEBUG4K_ADJUST(...) #define DEBUG4K_PMAP(...) #define DEBUG4K_MEMENTRY(...) #define DEBUG4K_IOKIT(...) #define DEBUG4K_UPL(...) #define DEBUG4K_EXC(...) #define DEBUG4K_VFS(...) #endif /* MACH_ASSERT */ __enum_decl(vm_object_destroy_reason_t, uint8_t, { VM_OBJECT_DESTROY_UNKNOWN_REASON = 0, VM_OBJECT_DESTROY_FORCED_UNMOUNT = 1, VM_OBJECT_DESTROY_UNGRAFT = 2, VM_OBJECT_DESTROY_MAX = 2, }); _Static_assert(VM_OBJECT_DESTROY_MAX < 4, "Need to fit in `no_pager_reason`'s number of bits"); #endif /* _VM_VM_PROTOS_H_ */ #endif /* XNU_KERNEL_PRIVATE */