Lines Matching refs:VM_PROT_WRITE
2531 if ((cur_protection & VM_PROT_WRITE) && in vm_map_enter()
2581 int reject_prot = (needs_copy ? VM_PROT_ALLEXEC : (VM_PROT_WRITE | VM_PROT_ALLEXEC)); in vm_map_enter()
2613 !(max_protection & VM_PROT_WRITE) || in vm_map_enter()
3125 int reject_prot = (needs_copy ? VM_PROT_ALLEXEC : (VM_PROT_WRITE | VM_PROT_ALLEXEC)); in vm_map_enter()
3353 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter()
3600 if ((cur_protection & VM_PROT_WRITE) && in vm_map_enter_fourk()
3760 if ((entry->protection & VM_PROT_WRITE) && in vm_map_enter_fourk()
3950 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_fourk()
4441 copy_entry->protection & VM_PROT_WRITE) { in vm_map_enter_mem_object_helper()
4444 prot = copy_entry->protection & ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
4486 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
4757 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
5095 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_control()
5370 VM_PROT_ALL, VM_PROT_WRITE, in vm_map_enter_cpm()
6052 if ((new_prot & VM_PROT_WRITE) && in vm_map_protect()
6099 ((current->protection & VM_PROT_EXECUTE) && (new_prot & VM_PROT_WRITE))) { in vm_map_protect()
6197 !(old_prot & VM_PROT_WRITE) && in vm_map_protect()
6198 (current->protection & VM_PROT_WRITE)))) { in vm_map_protect()
6245 prot &= ~VM_PROT_WRITE; in vm_map_protect()
6249 if (prot & VM_PROT_WRITE) { in vm_map_protect()
6955 if (entry->protection & VM_PROT_WRITE) { in vm_map_wire_nested()
6965 if (entry->protection & VM_PROT_WRITE) { in vm_map_wire_nested()
7110 entry->protection & ~VM_PROT_WRITE); in vm_map_wire_nested()
9138 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_entry_is_overwritable()
9280 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_overwrite_submap_recurse()
9541 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_nested()
10383 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_unaligned()
10639 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_aligned()
12186 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_copyin_internal()
12971 (old_entry->protection & VM_PROT_WRITE)) { in vm_map_fork_share()
12976 prot = old_entry->protection & ~VM_PROT_WRITE; in vm_map_fork_share()
13053 new_entry->protection &= ~VM_PROT_WRITE; in vm_map_fork_share()
13054 new_entry->max_protection &= ~VM_PROT_WRITE; in vm_map_fork_share()
13506 prot = old_entry->protection & ~VM_PROT_WRITE; in vm_map_fork()
13847 !((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
13858 ((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
13951 if (((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
14173 prot = prot & ~VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
14321 if ((entry->protection & VM_PROT_WRITE) && in vm_map_lookup_and_lock_object()
14407 if (entry->used_for_tpro && (fault_type & VM_PROT_WRITE)) { in vm_map_lookup_and_lock_object()
14408 prot |= VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
14469 if ((fault_type & VM_PROT_WRITE) || *wired || force_copy) { in vm_map_lookup_and_lock_object()
14495 if ((fault_type & VM_PROT_WRITE) == 0 && *wired == 0) { in vm_map_lookup_and_lock_object()
14501 prot &= (~VM_PROT_WRITE); in vm_map_lookup_and_lock_object()
14505 if (submap_needed_copy && (prot & VM_PROT_WRITE)) { in vm_map_lookup_and_lock_object()
14511 assert(!(fault_type & VM_PROT_WRITE)); in vm_map_lookup_and_lock_object()
14515 prot &= ~VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
16328 VM_PROT_READ | VM_PROT_WRITE); in vm_map_willneed()
16644 if (!(entry->protection & VM_PROT_WRITE) && !entry->used_for_jit in vm_map_reusable_pages()
17536 (src_entry->protection & VM_PROT_WRITE)) { in vm_map_remap_extract()
17541 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_remap_extract()
17716 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_remap_extract()
18885 entry->max_protection |= VM_PROT_WRITE; in vm_map_remap()
19322 if ((entry->protection & VM_PROT_WRITE) == 0 && in vm_map_purgable_control()
20449 (named_entry->protection & VM_PROT_WRITE)) { in convert_port_entry_to_map()
21740 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_query_volatile()