Lines Matching refs:VM_PROT_WRITE

2477 	if ((cur_protection & VM_PROT_WRITE) &&  in vm_map_enter()
2524 int reject_prot = (needs_copy ? VM_PROT_ALLEXEC : (VM_PROT_WRITE | VM_PROT_ALLEXEC)); in vm_map_enter()
3025 int reject_prot = (needs_copy ? VM_PROT_ALLEXEC : (VM_PROT_WRITE | VM_PROT_ALLEXEC)); in vm_map_enter()
3253 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter()
3505 if ((cur_protection & VM_PROT_WRITE) && in vm_map_enter_fourk()
3662 if ((entry->protection & VM_PROT_WRITE) && in vm_map_enter_fourk()
3855 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_fourk()
4365 copy_entry->protection & VM_PROT_WRITE) { in vm_map_enter_mem_object_helper()
4368 prot = copy_entry->protection & ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
4410 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
4676 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_helper()
5025 pager_prot &= ~VM_PROT_WRITE; in vm_map_enter_mem_object_control()
5309 VM_PROT_ALL, VM_PROT_WRITE, in vm_map_enter_cpm()
5959 if ((new_prot & VM_PROT_WRITE) && in vm_map_protect()
5995 ((current->protection & VM_PROT_EXECUTE) && (new_prot & VM_PROT_WRITE))) { in vm_map_protect()
6083 prot &= ~VM_PROT_WRITE; in vm_map_protect()
6087 if (prot & VM_PROT_WRITE) { in vm_map_protect()
6770 if (entry->protection & VM_PROT_WRITE) { in vm_map_wire_nested()
6780 if (entry->protection & VM_PROT_WRITE) { in vm_map_wire_nested()
6921 entry->protection & ~VM_PROT_WRITE); in vm_map_wire_nested()
8910 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_entry_is_overwritable()
9052 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_overwrite_submap_recurse()
9313 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_nested()
10153 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_unaligned()
10394 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_copy_overwrite_aligned()
11942 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_copyin_internal()
12756 (old_entry->protection & VM_PROT_WRITE)) { in vm_map_fork_share()
12761 prot = old_entry->protection & ~VM_PROT_WRITE; in vm_map_fork_share()
12838 new_entry->protection &= ~VM_PROT_WRITE; in vm_map_fork_share()
12839 new_entry->max_protection &= ~VM_PROT_WRITE; in vm_map_fork_share()
13274 prot = old_entry->protection & ~VM_PROT_WRITE; in vm_map_fork()
13603 !((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
13614 ((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
13706 if (((fault_type & VM_PROT_WRITE) || in vm_map_lookup_and_lock_object()
13927 prot = prot & ~VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
14068 if ((entry->protection & VM_PROT_WRITE) && in vm_map_lookup_and_lock_object()
14195 if ((fault_type & VM_PROT_WRITE) || *wired || force_copy) { in vm_map_lookup_and_lock_object()
14221 if ((fault_type & VM_PROT_WRITE) == 0 && *wired == 0) { in vm_map_lookup_and_lock_object()
14227 prot &= (~VM_PROT_WRITE); in vm_map_lookup_and_lock_object()
14231 if (submap_needed_copy && (prot & VM_PROT_WRITE)) { in vm_map_lookup_and_lock_object()
14237 assert(!(fault_type & VM_PROT_WRITE)); in vm_map_lookup_and_lock_object()
14241 prot &= ~VM_PROT_WRITE; in vm_map_lookup_and_lock_object()
16033 VM_PROT_READ | VM_PROT_WRITE); in vm_map_willneed()
16349 if (!(entry->protection & VM_PROT_WRITE) && !entry->used_for_jit) { in vm_map_reusable_pages()
17218 (src_entry->protection & VM_PROT_WRITE)) { in vm_map_remap_extract()
17223 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_remap_extract()
17393 prot = src_entry->protection & ~VM_PROT_WRITE; in vm_map_remap_extract()
18547 entry->max_protection |= VM_PROT_WRITE; in vm_map_remap()
18967 if ((entry->protection & VM_PROT_WRITE) == 0 && in vm_map_purgable_control()
20088 (named_entry->protection & VM_PROT_WRITE)) { in convert_port_entry_to_map()
21263 if (!(entry->protection & VM_PROT_WRITE)) { in vm_map_query_volatile()