| /xnu-8020.101.4/osfmk/mach/ |
| H A D | vm_prot.h | 85 #define VM_PROT_EXECUTE ((vm_prot_t) 0x04) /* execute permission */ macro 97 #define VM_PROT_ALL (VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE) 159 #define VM_PROT_EXECUTE_ONLY (VM_PROT_EXECUTE|VM_PROT_STRIP_READ) 180 #define VM_PROT_ALLEXEC (VM_PROT_EXECUTE | VM_PROT_UEXEC) 182 #define VM_PROT_ALLEXEC (VM_PROT_EXECUTE)
|
| H A D | dyld_kernel_fixups.h | 485 if (!(seg->maxprot & VM_PROT_EXECUTE)) { in kernel_collection_adjust_mh_addrs()
|
| /xnu-8020.101.4/bsd/kern/ |
| H A D | kern_mman.c | 229 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 339 int reject_prot = ((flags & MAP_PRIVATE) ? VM_PROT_EXECUTE : (VM_PROT_WRITE | VM_PROT_EXECUTE)); in mmap() 514 maxprot = VM_PROT_EXECUTE; /* TODO: Remove this and restrict maxprot? */ in mmap() 678 prot |= VM_PROT_EXECUTE; in mmap() 681 maxprot |= VM_PROT_EXECUTE; in mmap() 687 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 690 if (maxprot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 753 prot |= VM_PROT_EXECUTE; in mmap() 756 maxprot |= VM_PROT_EXECUTE; in mmap() 762 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() [all …]
|
| H A D | mach_loader.c | 1100 if (scp->initprot & VM_PROT_EXECUTE) { in parse_machfile() 1149 ((scp->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) != (VM_PROT_READ | VM_PROT_EXECUTE))) { in parse_machfile() 1204 … ((scp64->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) != (VM_PROT_READ | VM_PROT_EXECUTE))) { in parse_machfile() 1920 if (result->is_cambria && (initprot & VM_PROT_EXECUTE) == VM_PROT_EXECUTE) { in map_segment() 2482 if ((scp->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) == (VM_PROT_READ | VM_PROT_EXECUTE)) { in load_segment()
|
| /xnu-8020.101.4/bsd/dev/dtrace/ |
| H A D | dtrace_ptss.c | 183 vm_prot_t cur_protection = VM_PROT_READ | VM_PROT_EXECUTE; in dtrace_ptss_allocate_page() 184 vm_prot_t max_protection = VM_PROT_READ | VM_PROT_EXECUTE | VM_PROT_WRITE; in dtrace_ptss_allocate_page()
|
| H A D | dtrace_glue.c | 183 …tect(map, (mach_vm_offset_t)a, (mach_vm_size_t)len, 0, (reprotect & ~VM_PROT_EXECUTE) | VM_PROT_WR… in uwrite() 213 ASSERT(reprotect & VM_PROT_EXECUTE); in uwrite()
|
| /xnu-8020.101.4/osfmk/arm64/ |
| H A D | alternate_debugger.c | 99 pmap_protect(kernel_map->pmap, alt_code, alt_code + alt_size, VM_PROT_READ | VM_PROT_EXECUTE); in alternate_debugger_enter()
|
| H A D | hibernate_restore.c | 376 bool executable = (protection & VM_PROT_EXECUTE); in pal_hib_resume_tramp()
|
| H A D | machine_routines.c | 1542 if ((new_prot & VM_PROT_WRITE) && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 1545 if (lockdown_done && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 1561 if (!(new_prot & VM_PROT_EXECUTE)) { in ml_static_protect()
|
| H A D | sleh.c | 1013 *fault_type = (VM_PROT_READ | VM_PROT_EXECUTE); in inspect_instruction_abort() 1318 if (!(fault_type & VM_PROT_EXECUTE)) { in handle_user_abort()
|
| /xnu-8020.101.4/osfmk/i386/commpage/ |
| H A D | commpage.c | 183 if (uperm == (VM_PROT_READ | VM_PROT_EXECUTE)) { in commpage_allocate() 628 …locate(commpage_text32_map, (vm_size_t) _COMM_PAGE_TEXT_AREA_USED, VM_PROT_READ | VM_PROT_EXECUTE); in commpage_text_populate() 649 …locate(commpage_text64_map, (vm_size_t) _COMM_PAGE_TEXT_AREA_USED, VM_PROT_READ | VM_PROT_EXECUTE); in commpage_text_populate()
|
| /xnu-8020.101.4/osfmk/i386/ |
| H A D | pmap_x86_common.c | 443 prot |= VM_PROT_EXECUTE; in pmap_get_prot() 455 prot |= VM_PROT_EXECUTE; in pmap_get_prot() 682 if ((prot & VM_PROT_EXECUTE) || __improbable(is_ept && (prot & VM_PROT_UEXEC))) { in pmap_enter_options() 851 if (prot & VM_PROT_EXECUTE) { in pmap_enter_options() 1203 if (prot & VM_PROT_EXECUTE) { in pmap_enter_options() 1723 case VM_PROT_READ | VM_PROT_EXECUTE: in pmap_page_protect_options() 2350 if ((prot & VM_PROT_EXECUTE) == 0) { in pmap_map_bd() 2397 if ((prot & VM_PROT_EXECUTE) == 0) { in pmap_alias()
|
| H A D | trap.c | 717 prot |= VM_PROT_EXECUTE; in kernel_trap() 1062 code = VM_PROT_READ | VM_PROT_EXECUTE; in user_trap() 1112 prot |= VM_PROT_EXECUTE; in user_trap()
|
| H A D | machine_routines.c | 255 boolean_t NX = !!!(prot & VM_PROT_EXECUTE), ro = !!!(prot & VM_PROT_WRITE); in ml_static_protect()
|
| /xnu-8020.101.4/san/memory/ |
| H A D | kasan_dynamic_blacklist.c | 202 bool is_exec = seg->initprot & VM_PROT_EXECUTE; in kasan_dybl_load_kext() 243 bool is_exec = seg->initprot & VM_PROT_EXECUTE; in kasan_dybl_unload_kext()
|
| /xnu-8020.101.4/osfmk/arm/ |
| H A D | machine_routines.c | 823 if ((new_prot & VM_PROT_WRITE) && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 826 if (lockdown_done && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 839 if (!(new_prot & VM_PROT_EXECUTE)) { in ml_static_protect()
|
| H A D | pmap.c | 4702 case VM_PROT_READ | VM_PROT_EXECUTE: 4831 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) 4833 if ((prot & VM_PROT_EXECUTE)) 5307 case VM_PROT_EXECUTE: 5312 case VM_PROT_READ | VM_PROT_EXECUTE: 5330 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) 5332 if ((prot & VM_PROT_EXECUTE)) 5556 case VM_PROT_EXECUTE: 5558 case VM_PROT_READ | VM_PROT_EXECUTE: 5877 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) [all …]
|
| H A D | trap.c | 327 fault_type = VM_PROT_READ | VM_PROT_EXECUTE; in sleh_abort()
|
| /xnu-8020.101.4/osfmk/x86_64/ |
| H A D | pmap.c | 1306 boolean_t NXbit = !(seg->initprot & VM_PROT_EXECUTE), in pmap_lowmem_finalize() 1876 if ((prot & VM_PROT_EXECUTE) || __improbable(is_ept && (prot & VM_PROT_UEXEC))) { in pmap_protect_options() 1944 set_bits |= ((prot & VM_PROT_EXECUTE) ? INTEL_EPT_EX : 0) | in pmap_protect_options() 3192 if ((prot & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in pmap_permissions_verify() 3211 tprot |= VM_PROT_EXECUTE; in pmap_permissions_verify()
|
| /xnu-8020.101.4/osfmk/vm/ |
| H A D | vm_map.c | 868 !(map_entry->protection & VM_PROT_EXECUTE)) { in vm_map_apple_protected() 2251 (cur_protection & VM_PROT_EXECUTE) && in vm_map_enter() 2279 cur_protection &= ~VM_PROT_EXECUTE; in vm_map_enter() 2290 if (cur_protection & VM_PROT_EXECUTE) { in vm_map_enter() 2900 !((entry->protection & VM_PROT_EXECUTE) && in vm_map_enter() 3473 (cur_protection & VM_PROT_EXECUTE) && in vm_map_enter_fourk() 3494 cur_protection &= ~VM_PROT_EXECUTE; in vm_map_enter_fourk() 3502 if (cur_protection & VM_PROT_EXECUTE) { in vm_map_enter_fourk() 6050 ((current->protection & VM_PROT_EXECUTE) && (new_prot & VM_PROT_WRITE))) { in vm_map_protect() 6145 prot |= VM_PROT_EXECUTE; in vm_map_protect() [all …]
|
| H A D | vm_fault.c | 2701 (prot & VM_PROT_EXECUTE)) { in vm_fault_cs_check_violation() 2748 (prot & VM_PROT_EXECUTE) in vm_fault_cs_check_violation() 3325 if ((prot & VM_PROT_EXECUTE) && !m->vmp_xpmapped) { in vm_fault_enter_set_mapped() 3710 *prot &= ~VM_PROT_EXECUTE; in vm_fault_enter_prepare() 4378 (prot & VM_PROT_EXECUTE) && in vm_fault_internal() 8024 !(prot & VM_PROT_EXECUTE)) { in vmtc_revalidate_lookup()
|
| H A D | vm_user.c | 1247 if ((*max_protection & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == in mach_vm_remap_new_external() 1248 (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in mach_vm_remap_new_external() 1414 if ((*max_protection & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == in vm_remap_new_external() 1415 (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in vm_remap_new_external()
|
| H A D | vm_shared_region.c | 3257 VM_PROT_READ | VM_PROT_EXECUTE, in vm_commpage_enter() 3258 VM_PROT_READ | VM_PROT_EXECUTE, in vm_commpage_enter()
|
| /xnu-8020.101.4/osfmk/i386/AT386/ |
| H A D | model_dep.c | 465 …(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_… in efi_init() 558 …(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_… in hibernate_newruntime_map()
|
| /xnu-8020.101.4/libkern/kxld/ |
| H A D | kxld_seg.c | 50 #define TEXT_SEG_PROT (VM_PROT_READ | VM_PROT_EXECUTE)
|