| /xnu-12377.81.4/osfmk/mach/ |
| H A D | vm_prot.h | 89 #define VM_PROT_EXECUTE ((vm_prot_t) 0x04) /* execute permission */ macro 101 #define VM_PROT_ALL (VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE) 163 #define VM_PROT_EXECUTE_ONLY (VM_PROT_EXECUTE|VM_PROT_STRIP_READ) 194 #define VM_PROT_ALLEXEC (VM_PROT_EXECUTE | VM_PROT_UEXEC) 196 #define VM_PROT_ALLEXEC (VM_PROT_EXECUTE)
|
| H A D | dyld_kernel_fixups.h | 485 if (!(seg->maxprot & VM_PROT_EXECUTE)) { in kernel_collection_adjust_mh_addrs()
|
| /xnu-12377.81.4/bsd/kern/ |
| H A D | kern_mman.c | 349 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 419 int reject_prot = ((flags & MAP_PRIVATE) ? VM_PROT_EXECUTE : (VM_PROT_WRITE | VM_PROT_EXECUTE)); in mmap() 456 if ((prot & VM_PROT_EXECUTE) || in mmap() 604 maxprot = VM_PROT_EXECUTE; /* TODO: Remove this and restrict maxprot? */ in mmap() 769 prot |= VM_PROT_EXECUTE; in mmap() 772 maxprot |= VM_PROT_EXECUTE; in mmap() 778 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 781 if (maxprot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 889 prot |= VM_PROT_EXECUTE; in mmap() 892 maxprot |= VM_PROT_EXECUTE; in mmap() [all …]
|
| H A D | mach_loader.c | 1408 if (scp->initprot & VM_PROT_EXECUTE) { in parse_machfile() 1457 ((scp->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) != (VM_PROT_READ | VM_PROT_EXECUTE))) { in parse_machfile() 1512 … ((scp64->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) != (VM_PROT_READ | VM_PROT_EXECUTE))) { in parse_machfile() 2170 if (result->is_rosetta && (initprot & VM_PROT_EXECUTE) == VM_PROT_EXECUTE) { in map_segment() 2669 if ((scp->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) == (VM_PROT_READ | VM_PROT_EXECUTE)) { in load_segment()
|
| /xnu-12377.81.4/bsd/dev/dtrace/ |
| H A D | dtrace_ptss.c | 184 vm_prot_t cur_protection = VM_PROT_READ | VM_PROT_EXECUTE; in dtrace_ptss_allocate_page() 185 vm_prot_t max_protection = VM_PROT_READ | VM_PROT_EXECUTE; in dtrace_ptss_allocate_page() 208 VM_PROT_READ | VM_PROT_EXECUTE); in dtrace_ptss_allocate_page()
|
| H A D | dtrace_glue.c | 180 …tect(map, (mach_vm_offset_t)a, (mach_vm_size_t)len, 0, (reprotect & ~VM_PROT_EXECUTE) | VM_PROT_WR… in uwrite() 210 ASSERT(reprotect & VM_PROT_EXECUTE); in uwrite()
|
| /xnu-12377.81.4/osfmk/arm64/ |
| H A D | alternate_debugger.c | 99 pmap_protect(kernel_map->pmap, alt_code, alt_code + alt_size, VM_PROT_READ | VM_PROT_EXECUTE); in alternate_debugger_enter()
|
| H A D | sleh.c | 1578 *fault_type = (VM_PROT_READ | VM_PROT_EXECUTE); in inspect_instruction_abort() 1956 if (!(fault_type & VM_PROT_EXECUTE)) { in handle_user_abort() 2055 if (is_vm_fault(fault_code) && (fault_type & VM_PROT_EXECUTE)) { in handle_user_abort() 2544 if (!((fault_type) & VM_PROT_EXECUTE)) { in handle_kernel_abort() 2566 if (!((fault_type) & VM_PROT_EXECUTE)) { in handle_kernel_abort()
|
| H A D | hibernate_restore.c | 375 bool executable = (protection & VM_PROT_EXECUTE); in pal_hib_resume_tramp()
|
| H A D | machine_routines.c | 1931 if ((new_prot & VM_PROT_WRITE) && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 1934 if (lockdown_done && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 1950 if (!(new_prot & VM_PROT_EXECUTE)) { in ml_static_protect()
|
| /xnu-12377.81.4/osfmk/i386/commpage/ |
| H A D | commpage.c | 183 if (uperm == (VM_PROT_READ | VM_PROT_EXECUTE)) { in commpage_allocate() 670 …locate(commpage_text32_map, (vm_size_t) _COMM_PAGE_TEXT_AREA_USED, VM_PROT_READ | VM_PROT_EXECUTE); in commpage_text_populate() 691 …locate(commpage_text64_map, (vm_size_t) _COMM_PAGE_TEXT_AREA_USED, VM_PROT_READ | VM_PROT_EXECUTE); in commpage_text_populate()
|
| /xnu-12377.81.4/osfmk/i386/ |
| H A D | pmap_x86_common.c | 466 prot |= VM_PROT_EXECUTE; in pmap_get_prot() 478 prot |= VM_PROT_EXECUTE; in pmap_get_prot() 708 if ((prot & VM_PROT_EXECUTE) || __improbable(is_ept && (prot & VM_PROT_UEXEC))) { in pmap_enter_options() 877 if (prot & VM_PROT_EXECUTE) { in pmap_enter_options() 1229 if (prot & VM_PROT_EXECUTE) { in pmap_enter_options() 1763 case VM_PROT_READ | VM_PROT_EXECUTE: in pmap_page_protect_options() 2390 if ((prot & VM_PROT_EXECUTE) == 0) { in pmap_map_bd() 2437 if ((prot & VM_PROT_EXECUTE) == 0) { in pmap_alias()
|
| H A D | trap.c | 869 prot |= VM_PROT_EXECUTE; in kernel_trap() 1257 code = VM_PROT_READ | VM_PROT_EXECUTE; in user_trap() 1307 prot |= VM_PROT_EXECUTE; in user_trap()
|
| /xnu-12377.81.4/san/memory/ |
| H A D | kasan_dynamic_denylist.c | 202 bool is_exec = seg->initprot & VM_PROT_EXECUTE; in kasan_dyn_denylist_load_kext() 243 bool is_exec = seg->initprot & VM_PROT_EXECUTE; in kasan_dyn_denylist_unload_kext()
|
| /xnu-12377.81.4/tests/vm/ |
| H A D | upl.c | 557 .fault_prot = VM_PROT_EXECUTE | VM_PROT_READ }; 627 …protect(mach_task_self(), (mach_vm_address_t)buf, buf_size, FALSE, VM_PROT_READ | VM_PROT_EXECUTE);
|
| H A D | configurator_vm_wire.c | 167 if (prot_contains_all(checker->protection, VM_PROT_EXECUTE)) { in checker_perform_failed_vm_wire()
|
| /xnu-12377.81.4/osfmk/x86_64/ |
| H A D | pmap.c | 1331 boolean_t NXbit = !(seg->initprot & VM_PROT_EXECUTE), in pmap_lowmem_finalize() 1932 if ((prot & VM_PROT_EXECUTE) || __improbable(is_ept && (prot & VM_PROT_UEXEC))) { in pmap_protect_options() 2007 set_bits |= ((prot & VM_PROT_EXECUTE) ? INTEL_EPT_EX : 0) | in pmap_protect_options() 3272 if ((prot & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in pmap_permissions_verify() 3291 tprot |= VM_PROT_EXECUTE; in pmap_permissions_verify()
|
| /xnu-12377.81.4/osfmk/arm64/sptm/pmap/ |
| H A D | pmap.c | 4773 case VM_PROT_READ | VM_PROT_EXECUTE: 5086 if ((!(prot & VM_PROT_EXECUTE) && nx_enabled && pmap->nx_enabled) || 5089 if (!(prot & VM_PROT_EXECUTE) || 5397 case VM_PROT_READ | VM_PROT_EXECUTE: 5402 case VM_PROT_EXECUTE: 5426 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) 5428 if ((prot & VM_PROT_EXECUTE)) 5666 case VM_PROT_READ | VM_PROT_EXECUTE: 5671 case VM_PROT_EXECUTE: 5967 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) [all …]
|
| /xnu-12377.81.4/osfmk/arm/pmap/ |
| H A D | pmap.c | 2606 false, VM_PROT_READ | VM_PROT_EXECUTE); in pmap_lockdown_ppl() 4873 case VM_PROT_READ | VM_PROT_EXECUTE: 5006 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) 5008 if ((prot & VM_PROT_EXECUTE)) 5503 case VM_PROT_READ | VM_PROT_EXECUTE: 5508 case VM_PROT_EXECUTE: 5528 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) 5530 if ((prot & VM_PROT_EXECUTE)) 5756 case VM_PROT_READ | VM_PROT_EXECUTE: 5761 case VM_PROT_EXECUTE: [all …]
|
| /xnu-12377.81.4/osfmk/vm/ |
| H A D | vm_fault.c | 2790 (prot & VM_PROT_EXECUTE) && in vm_fault_cs_check_violation() 2791 (caller_prot & VM_PROT_EXECUTE)) { in vm_fault_cs_check_violation() 2834 (prot & VM_PROT_EXECUTE)) { in vm_fault_cs_check_violation() 2881 (prot & VM_PROT_EXECUTE) in vm_fault_cs_check_violation() 3463 if ((prot & VM_PROT_EXECUTE) && !m->vmp_xpmapped) { in vm_fault_enter_set_mapped() 3770 (*prot & VM_PROT_EXECUTE) && in vm_fault_attempt_pmap_enter() 3771 !(caller_prot & VM_PROT_EXECUTE)) { in vm_fault_attempt_pmap_enter() 3772 *prot &= ~VM_PROT_EXECUTE; in vm_fault_attempt_pmap_enter() 4091 *prot &= ~VM_PROT_EXECUTE; in vm_fault_enter_prepare() 4827 (prot & VM_PROT_EXECUTE) && in vm_fault_internal() [all …]
|
| H A D | vm_map.c | 720 new->protection &= ~VM_PROT_EXECUTE; in vm_map_entry_copy_code_signing() 721 new->max_protection &= ~VM_PROT_EXECUTE; in vm_map_entry_copy_code_signing() 1175 cur_prot |= VM_PROT_EXECUTE; in vm_map_apple_protected() 3170 (cur_protection & VM_PROT_EXECUTE) && in vm_map_enter() 3201 cur_protection &= ~VM_PROT_EXECUTE; in vm_map_enter() 3260 if (cur_protection & VM_PROT_EXECUTE) { in vm_map_enter() 3659 !((entry->protection & VM_PROT_EXECUTE) && in vm_map_enter() 4425 if (cur_protection & VM_PROT_EXECUTE) { in vm_map_enter_mem_object() 6271 …if (set_max && (new_prot & VM_PROT_EXECUTE) && (csm_address_space_exempt(map->pmap) == KERN_SUCCES… in vm_map_protect() 6272 new_max |= VM_PROT_EXECUTE; in vm_map_protect() [all …]
|
| H A D | vm_user.c | 958 if ((max_protection & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == in mach_vm_remap_new_external() 959 (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in mach_vm_remap_new_external()
|
| /xnu-12377.81.4/osfmk/i386/AT386/ |
| H A D | model_dep.c | 461 …(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_… in efi_init() 554 …(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_… in hibernate_newruntime_map()
|
| /xnu-12377.81.4/osfmk/tests/ |
| H A D | vm_parameter_validation.h | 1658 VM_PROT_TRIAL(VM_PROT_READ | VM_PROT_EXECUTE), 1660 VM_PROT_TRIAL(VM_PROT_READ | VM_PROT_WRITE | VM_PROT_EXECUTE), 1669 VM_PROT_TRIAL(VM_PROT_READ | VM_PROT_EXECUTE | 1u << 7), 1672 VM_PROT_TRIAL(VM_PROT_READ | VM_PROT_EXECUTE | 1u << 8), 1682 VM_PROT_TRIAL(VM_PROT_READ | VM_PROT_EXECUTE | 1u << 16), 1694 VM_PROT_TRIAL(VM_PROT_READ | VM_PROT_EXECUTE | 1u << 25), 1702 VM_PROT_TRIAL(VM_PROT_READ | VM_PROT_EXECUTE | 1u << 31),
|
| /xnu-12377.81.4/tests/ |
| H A D | vm_test_mach_map.c | 2045 T_ASSERT_EQ((region_info.protection & VM_PROT_EXECUTE), 2318 VM_PROT_EXECUTE, VM_PROT_EXECUTE,
|