| /xnu-8020.121.3/osfmk/mach/ |
| H A D | vm_prot.h | 85 #define VM_PROT_EXECUTE ((vm_prot_t) 0x04) /* execute permission */ macro 97 #define VM_PROT_ALL (VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE) 159 #define VM_PROT_EXECUTE_ONLY (VM_PROT_EXECUTE|VM_PROT_STRIP_READ) 180 #define VM_PROT_ALLEXEC (VM_PROT_EXECUTE | VM_PROT_UEXEC) 182 #define VM_PROT_ALLEXEC (VM_PROT_EXECUTE)
|
| H A D | dyld_kernel_fixups.h | 485 if (!(seg->maxprot & VM_PROT_EXECUTE)) { in kernel_collection_adjust_mh_addrs()
|
| /xnu-8020.121.3/bsd/kern/ |
| H A D | kern_mman.c | 229 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 339 int reject_prot = ((flags & MAP_PRIVATE) ? VM_PROT_EXECUTE : (VM_PROT_WRITE | VM_PROT_EXECUTE)); in mmap() 514 maxprot = VM_PROT_EXECUTE; /* TODO: Remove this and restrict maxprot? */ in mmap() 678 prot |= VM_PROT_EXECUTE; in mmap() 681 maxprot |= VM_PROT_EXECUTE; in mmap() 687 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 690 if (maxprot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() 753 prot |= VM_PROT_EXECUTE; in mmap() 756 maxprot |= VM_PROT_EXECUTE; in mmap() 762 if (prot & (VM_PROT_EXECUTE | VM_PROT_WRITE)) { in mmap() [all …]
|
| H A D | mach_loader.c | 1100 if (scp->initprot & VM_PROT_EXECUTE) { in parse_machfile() 1149 ((scp->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) != (VM_PROT_READ | VM_PROT_EXECUTE))) { in parse_machfile() 1204 … ((scp64->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) != (VM_PROT_READ | VM_PROT_EXECUTE))) { in parse_machfile() 1920 if (result->is_cambria && (initprot & VM_PROT_EXECUTE) == VM_PROT_EXECUTE) { in map_segment() 2482 if ((scp->initprot & (VM_PROT_READ | VM_PROT_EXECUTE)) == (VM_PROT_READ | VM_PROT_EXECUTE)) { in load_segment()
|
| /xnu-8020.121.3/bsd/dev/dtrace/ |
| H A D | dtrace_ptss.c | 183 vm_prot_t cur_protection = VM_PROT_READ | VM_PROT_EXECUTE; in dtrace_ptss_allocate_page() 184 vm_prot_t max_protection = VM_PROT_READ | VM_PROT_EXECUTE | VM_PROT_WRITE; in dtrace_ptss_allocate_page()
|
| H A D | dtrace_glue.c | 183 …tect(map, (mach_vm_offset_t)a, (mach_vm_size_t)len, 0, (reprotect & ~VM_PROT_EXECUTE) | VM_PROT_WR… in uwrite() 213 ASSERT(reprotect & VM_PROT_EXECUTE); in uwrite()
|
| /xnu-8020.121.3/osfmk/arm64/ |
| H A D | alternate_debugger.c | 99 pmap_protect(kernel_map->pmap, alt_code, alt_code + alt_size, VM_PROT_READ | VM_PROT_EXECUTE); in alternate_debugger_enter()
|
| H A D | hibernate_restore.c | 376 bool executable = (protection & VM_PROT_EXECUTE); in pal_hib_resume_tramp()
|
| H A D | machine_routines.c | 1545 if ((new_prot & VM_PROT_WRITE) && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 1548 if (lockdown_done && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 1564 if (!(new_prot & VM_PROT_EXECUTE)) { in ml_static_protect()
|
| H A D | sleh.c | 1056 *fault_type = (VM_PROT_READ | VM_PROT_EXECUTE); in inspect_instruction_abort() 1325 if (!(fault_type & VM_PROT_EXECUTE)) { in handle_user_abort()
|
| /xnu-8020.121.3/osfmk/i386/commpage/ |
| H A D | commpage.c | 183 if (uperm == (VM_PROT_READ | VM_PROT_EXECUTE)) { in commpage_allocate() 628 …locate(commpage_text32_map, (vm_size_t) _COMM_PAGE_TEXT_AREA_USED, VM_PROT_READ | VM_PROT_EXECUTE); in commpage_text_populate() 649 …locate(commpage_text64_map, (vm_size_t) _COMM_PAGE_TEXT_AREA_USED, VM_PROT_READ | VM_PROT_EXECUTE); in commpage_text_populate()
|
| /xnu-8020.121.3/osfmk/i386/ |
| H A D | pmap_x86_common.c | 448 prot |= VM_PROT_EXECUTE; in pmap_get_prot() 460 prot |= VM_PROT_EXECUTE; in pmap_get_prot() 687 if ((prot & VM_PROT_EXECUTE) || __improbable(is_ept && (prot & VM_PROT_UEXEC))) { in pmap_enter_options() 856 if (prot & VM_PROT_EXECUTE) { in pmap_enter_options() 1208 if (prot & VM_PROT_EXECUTE) { in pmap_enter_options() 1742 case VM_PROT_READ | VM_PROT_EXECUTE: in pmap_page_protect_options() 2369 if ((prot & VM_PROT_EXECUTE) == 0) { in pmap_map_bd() 2416 if ((prot & VM_PROT_EXECUTE) == 0) { in pmap_alias()
|
| H A D | trap.c | 717 prot |= VM_PROT_EXECUTE; in kernel_trap() 1054 code = VM_PROT_READ | VM_PROT_EXECUTE; in user_trap() 1104 prot |= VM_PROT_EXECUTE; in user_trap()
|
| H A D | machine_routines.c | 255 boolean_t NX = !!!(prot & VM_PROT_EXECUTE), ro = !!!(prot & VM_PROT_WRITE); in ml_static_protect()
|
| /xnu-8020.121.3/san/memory/ |
| H A D | kasan_dynamic_blacklist.c | 202 bool is_exec = seg->initprot & VM_PROT_EXECUTE; in kasan_dybl_load_kext() 243 bool is_exec = seg->initprot & VM_PROT_EXECUTE; in kasan_dybl_unload_kext()
|
| /xnu-8020.121.3/osfmk/arm/ |
| H A D | machine_routines.c | 823 if ((new_prot & VM_PROT_WRITE) && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 826 if (lockdown_done && (new_prot & VM_PROT_EXECUTE)) { in ml_static_protect() 839 if (!(new_prot & VM_PROT_EXECUTE)) { in ml_static_protect()
|
| H A D | pmap.c | 4720 case VM_PROT_READ | VM_PROT_EXECUTE: 4849 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) 4851 if ((prot & VM_PROT_EXECUTE)) 5325 case VM_PROT_EXECUTE: 5330 case VM_PROT_READ | VM_PROT_EXECUTE: 5348 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) 5350 if ((prot & VM_PROT_EXECUTE)) 5574 case VM_PROT_EXECUTE: 5576 case VM_PROT_READ | VM_PROT_EXECUTE: 5895 if ((prot & VM_PROT_EXECUTE) || !nx_enabled || !pmap->nx_enabled) [all …]
|
| H A D | trap.c | 327 fault_type = VM_PROT_READ | VM_PROT_EXECUTE; in sleh_abort()
|
| /xnu-8020.121.3/osfmk/x86_64/ |
| H A D | pmap.c | 1320 boolean_t NXbit = !(seg->initprot & VM_PROT_EXECUTE), in pmap_lowmem_finalize() 1890 if ((prot & VM_PROT_EXECUTE) || __improbable(is_ept && (prot & VM_PROT_UEXEC))) { in pmap_protect_options() 1965 set_bits |= ((prot & VM_PROT_EXECUTE) ? INTEL_EPT_EX : 0) | in pmap_protect_options() 3213 if ((prot & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in pmap_permissions_verify() 3232 tprot |= VM_PROT_EXECUTE; in pmap_permissions_verify()
|
| /xnu-8020.121.3/osfmk/vm/ |
| H A D | vm_map.c | 900 !(map_entry->protection & VM_PROT_EXECUTE)) { in vm_map_apple_protected() 2499 (cur_protection & VM_PROT_EXECUTE) && in vm_map_enter() 2527 cur_protection &= ~VM_PROT_EXECUTE; in vm_map_enter() 2538 if (cur_protection & VM_PROT_EXECUTE) { in vm_map_enter() 2924 !((entry->protection & VM_PROT_EXECUTE) && in vm_map_enter() 3478 (cur_protection & VM_PROT_EXECUTE) && in vm_map_enter_fourk() 3499 cur_protection &= ~VM_PROT_EXECUTE; in vm_map_enter_fourk() 3507 if (cur_protection & VM_PROT_EXECUTE) { in vm_map_enter_fourk() 5965 ((current->protection & VM_PROT_EXECUTE) && (new_prot & VM_PROT_WRITE))) { in vm_map_protect() 6060 prot |= VM_PROT_EXECUTE; in vm_map_protect() [all …]
|
| H A D | vm_fault.c | 2696 (prot & VM_PROT_EXECUTE)) { in vm_fault_cs_check_violation() 2743 (prot & VM_PROT_EXECUTE) in vm_fault_cs_check_violation() 3320 if ((prot & VM_PROT_EXECUTE) && !m->vmp_xpmapped) { in vm_fault_enter_set_mapped() 3705 *prot &= ~VM_PROT_EXECUTE; in vm_fault_enter_prepare() 4373 (prot & VM_PROT_EXECUTE) && in vm_fault_internal() 8034 !(prot & VM_PROT_EXECUTE)) { in vmtc_revalidate_lookup()
|
| H A D | vm_user.c | 1247 if ((*max_protection & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == in mach_vm_remap_new_external() 1248 (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in mach_vm_remap_new_external() 1420 if ((*max_protection & (VM_PROT_WRITE | VM_PROT_EXECUTE)) == in vm_remap_new_external() 1421 (VM_PROT_WRITE | VM_PROT_EXECUTE)) { in vm_remap_new_external()
|
| H A D | vm_shared_region.c | 3257 VM_PROT_READ | VM_PROT_EXECUTE, in vm_commpage_enter() 3258 VM_PROT_READ | VM_PROT_EXECUTE, in vm_commpage_enter()
|
| /xnu-8020.121.3/osfmk/i386/AT386/ |
| H A D | model_dep.c | 465 …(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_… in efi_init() 558 …(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_… in hibernate_newruntime_map()
|
| /xnu-8020.121.3/libkern/kxld/ |
| H A D | kxld_seg.c | 50 #define TEXT_SEG_PROT (VM_PROT_READ | VM_PROT_EXECUTE)
|