Lines Matching refs:VM_PROT_WRITE
2531 if ((cur_protection & VM_PROT_WRITE) && in vm_map_enter()
2581 int reject_prot = (needs_copy ? VM_PROT_ALLEXEC : (VM_PROT_WRITE | VM_PROT_ALLEXEC)); in vm_map_enter()
2613 !(max_protection & VM_PROT_WRITE) || in vm_map_enter()
3125 int reject_prot = (needs_copy ? VM_PROT_ALLEXEC : (VM_PROT_WRITE | VM_PROT_ALLEXEC)); in vm_map_enter()
3353 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter()
3600 if ((cur_protection & VM_PROT_WRITE) && in vm_map_enter_fourk()
3760 if ((entry->protection & VM_PROT_WRITE) && in vm_map_enter_fourk()
3950 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_fourk()
4441 copy_entry->protection & VM_PROT_WRITE) { in vm_map_enter_mem_object_helper()
4444 prot = copy_entry->protection & ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
4486 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
4757 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
5095 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_control()
5370 VM_PROT_ALL, VM_PROT_WRITE, in vm_map_enter_cpm()
6052 if ((new_prot & VM_PROT_WRITE) && in vm_map_protect()
6099 ((current->protection & VM_PROT_EXECUTE) && (new_prot & VM_PROT_WRITE))) { in vm_map_protect()
6197 !(old_prot & VM_PROT_WRITE) && in vm_map_protect()
6198 (current->protection & VM_PROT_WRITE)))) { in vm_map_protect()
6245 prot &= ~VM_PROT_WRITE; in vm_map_protect()
6249 if (prot & VM_PROT_WRITE) { in vm_map_protect()
6955 if (entry->protection & VM_PROT_WRITE) { in vm_map_wire_nested()
6965 if (entry->protection & VM_PROT_WRITE) { in vm_map_wire_nested()
7110 entry->protection & ~VM_PROT_WRITE); in vm_map_wire_nested()
9151 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_entry_is_overwritable()
9293 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_overwrite_submap_recurse()
9554 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_nested()
10396 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_unaligned()
10652 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_aligned()
12199 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_copyin_internal()
12984 (old_entry->protection & VM_PROT_WRITE)) { in vm_map_fork_share()
12989 prot = old_entry->protection & ~VM_PROT_WRITE; in vm_map_fork_share()
13066 new_entry->protection &= ~VM_PROT_WRITE; in vm_map_fork_share()
13067 new_entry->max_protection &= ~VM_PROT_WRITE; in vm_map_fork_share()
13519 prot = old_entry->protection & ~VM_PROT_WRITE; in vm_map_fork()
13860 !((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
13871 ((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
13964 if (((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
14186 prot = prot & ~VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
14334 if ((entry->protection & VM_PROT_WRITE) && in vm_map_lookup_and_lock_object()
14420 if (entry->used_for_tpro && (fault_type & VM_PROT_WRITE)) { in vm_map_lookup_and_lock_object()
14421 prot |= VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
14482 if ((fault_type & VM_PROT_WRITE) || *wired || force_copy) { in vm_map_lookup_and_lock_object()
14508 if ((fault_type & VM_PROT_WRITE) == 0 && *wired == 0) { in vm_map_lookup_and_lock_object()
14514 prot &= (~VM_PROT_WRITE); in vm_map_lookup_and_lock_object()
14518 if (submap_needed_copy && (prot & VM_PROT_WRITE)) { in vm_map_lookup_and_lock_object()
14524 assert(!(fault_type & VM_PROT_WRITE)); in vm_map_lookup_and_lock_object()
14528 prot &= ~VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
16341 VM_PROT_READ | VM_PROT_WRITE); in vm_map_willneed()
16657 if (!(entry->protection & VM_PROT_WRITE) && !entry->used_for_jit in vm_map_reusable_pages()
17549 (src_entry->protection & VM_PROT_WRITE)) { in vm_map_remap_extract()
17554 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_remap_extract()
17729 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_remap_extract()
18898 entry->max_protection |= VM_PROT_WRITE; in vm_map_remap()
19335 if ((entry->protection & VM_PROT_WRITE) == 0 && in vm_map_purgable_control()
20462 (named_entry->protection & VM_PROT_WRITE)) { in convert_port_entry_to_map()
21753 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_query_volatile()