Lines Matching refs:TRUE

307 	boolean_t need_default_val = TRUE;  in vm_fault_init()
416 boolean_t vm_page_deactivate_behind = TRUE;
693 return TRUE; in vm_fault_deactivate_behind()
843 if (page_throttle == TRUE) { in vm_fault_check()
896 vm_page_queues_remove(m, TRUE); in vm_fault_enqueue_throttled_locked()
934 m->vmp_pmapped = TRUE; in vm_fault_zero_page()
936 if (no_zero_fill == TRUE) { in vm_fault_zero_page()
1150 while (TRUE) { in vm_fault_page()
1214 if (caller_lookup == TRUE) { in vm_fault_page()
1273 m->vmp_busy = TRUE; in vm_fault_page()
1361 …error = vm_fault_check(object, m, first_m, interruptible_state, (type_of_fault == NULL) ? TRUE : F… in vm_fault_page()
1397 m->vmp_busy = TRUE; in vm_fault_page()
1399 if (fault_info->mark_zf_absent && no_zero_fill == TRUE) { in vm_fault_page()
1400 m->vmp_absent = TRUE; in vm_fault_page()
1419 m->vmp_busy = TRUE; in vm_fault_page()
1535 m->vmp_busy = TRUE; in vm_fault_page()
1569 …look_for_page = (object->pager_created && (MUST_ASK_PAGER(object, offset, external_state) == TRUE)… in vm_fault_page()
1589 if (fault_info && fault_info->batch_pmap_op == TRUE) { in vm_fault_page()
1592 VM_KERN_MEMORY_NONE, FALSE, TRUE, TRUE, FALSE, NULL); in vm_fault_page()
1690 m->vmp_absent = TRUE; in vm_fault_page()
1691 if (fault_info && fault_info->batch_pmap_op == TRUE) { in vm_fault_page()
1692 …rnal(m, object, vm_object_trunc_page(offset), VM_KERN_MEMORY_NONE, FALSE, TRUE, TRUE, FALSE, NULL); in vm_fault_page()
1699 m->vmp_absent = TRUE; in vm_fault_page()
1728 if ((throttle_delay = vm_page_throttled(TRUE))) { in vm_fault_page()
1746 m->vmp_dirty = TRUE; in vm_fault_page()
1759 m->vmp_written_by_kernel = TRUE; in vm_fault_page()
1798 m->vmp_unusual = TRUE; in vm_fault_page()
1799 m->vmp_error = TRUE; in vm_fault_page()
1915 force_fault_retry = TRUE; in vm_fault_page()
1920 if (data_already_requested == TRUE) { in vm_fault_page()
1937 if (data_already_requested == TRUE) { in vm_fault_page()
1941 data_already_requested = TRUE; in vm_fault_page()
1982 if (force_fault_retry == TRUE) { in vm_fault_page()
2066 …error = vm_fault_check(object, m, first_m, interruptible_state, (type_of_fault == NULL) ? TRUE : F… in vm_fault_page()
2083 if (fault_info->mark_zf_absent && no_zero_fill == TRUE) { in vm_fault_page()
2084 m->vmp_absent = TRUE; in vm_fault_page()
2254 SET_PAGE_DIRTY(copy_m, TRUE); in vm_fault_page()
2264 vm_object_collapse(object, vm_object_trunc_page(offset), TRUE); in vm_fault_page()
2431 SET_PAGE_DIRTY(copy_m, TRUE); in vm_fault_page()
2434 assert(copy_m->vmp_busy == TRUE); in vm_fault_page()
2440 SET_PAGE_DIRTY(copy_m, TRUE); in vm_fault_page()
2780 *cs_violation = TRUE; in vm_fault_cs_check_violation()
2802 *cs_violation = TRUE; in vm_fault_cs_check_violation()
2813 *cs_violation = TRUE; in vm_fault_cs_check_violation()
3122 *must_disconnect = TRUE; in vm_fault_cs_handle_violation()
3123 VMP_CS_SET_TAINTED(m, fault_page_size, fault_phys_offset, TRUE); in vm_fault_cs_handle_violation()
3205 page_queues_locked = TRUE; \ in vm_fault_enqueue_page()
3229 vm_page_wire(m, wire_tag, TRUE); in vm_fault_enqueue_page()
3232 vm_page_unwire(m, TRUE); in vm_fault_enqueue_page()
3236 if (object->internal == TRUE) { in vm_fault_enqueue_page()
3349 m->vmp_no_cache = TRUE; in vm_fault_enqueue_page()
3390 m->vmp_pmapped = TRUE; in vm_fault_enter_set_mapped()
3393 m->vmp_xpmapped = TRUE; in vm_fault_enter_set_mapped()
3429 m->vmp_pmapped = TRUE; in vm_fault_enter_set_mapped()
3440 m->vmp_wpmapped = TRUE; in vm_fault_enter_set_mapped()
3588 *need_retry = TRUE; in vm_fault_pmap_enter()
3656 *need_retry = TRUE; in vm_fault_pmap_enter_with_object_lock()
3671 m->vmp_busy = TRUE; in vm_fault_pmap_enter_with_object_lock()
3698 m->vmp_busy = TRUE; in vm_fault_pmap_enter_with_object_lock()
3769 cs_bypass = TRUE; in vm_fault_enter_prepare()
4124 m->vmp_dirty = TRUE; in vm_fault_complete()
4142 if (need_collapse == TRUE) { in vm_fault_complete()
4143 vm_object_collapse(object, vm_object_trunc_page(offset), TRUE); in vm_fault_complete()
4374 need_copy = TRUE; in vm_fault_internal()
4418 need_copy = TRUE; in vm_fault_internal()
4618 while (TRUE) { in vm_fault_internal()
4894 need_copy = TRUE; in vm_fault_internal()
5081 if (need_retry == TRUE) { in vm_fault_internal()
5228 need_collapse = TRUE; in vm_fault_internal()
5275 if (MUST_ASK_PAGER(cur_object, cur_offset, compressor_external_state) == TRUE) { in vm_fault_internal()
5304 insert_cur_object = TRUE; in vm_fault_internal()
5307 if (insert_cur_object == TRUE) { in vm_fault_internal()
5391 shared_lock = TRUE; in vm_fault_internal()
5433 m->vmp_dirty = TRUE; in vm_fault_internal()
5692 m->vmp_pmapped = TRUE; in vm_fault_internal()
5737 m->vmp_busy = TRUE; in vm_fault_internal()
5868 if (need_retry == TRUE) { in vm_fault_internal()
6140 object_locks_dropped = TRUE; in vm_fault_internal()
6153 object_locks_dropped = TRUE; in vm_fault_internal()
6266 if (object_locks_dropped == TRUE) { in vm_fault_internal()
6328 need_copy_on_read = TRUE; in vm_fault_internal()
6329 need_copy = TRUE; in vm_fault_internal()
6444 m->vmp_dirty = TRUE; in vm_fault_internal()
6628 if ((throttle_delay = vm_page_throttled(TRUE))) { in vm_fault_internal()
6728 rc = vm_fault_internal(map, va, prot, TRUE, wire_tag, in vm_fault_wire()
6793 fault_info.stealth = TRUE; in vm_fault_unwire()
6820 TRUE, VM_KERN_MEMORY_NONE, THREAD_UNINT, pmap, pmap_addr); in vm_fault_unwire()
6845 VM_PROT_NONE, TRUE, in vm_fault_unwire()
6907 vm_page_unwire(result_page, TRUE); in vm_fault_unwire()
6929 pmap_addr + (end_addr - entry->vme_start), TRUE); in vm_fault_unwire()
7004 vm_page_unwire(m, TRUE); \ in vm_fault_wire_fast()
7095 vm_page_wire(m, wire_tag, TRUE); in vm_fault_wire_fast()
7102 m->vmp_busy = TRUE; in vm_fault_wire_fast()
7147 TRUE, /* wired */ in vm_fault_wire_fast()
7172 m->vmp_dirty = TRUE; in vm_fault_wire_fast()
7223 vm_page_unwire(page, TRUE); in vm_fault_copy_dst_cleanup()
7306 fault_info_src.stealth = TRUE; in vm_fault_copy()
7312 fault_info_dst.stealth = TRUE; in vm_fault_copy()
7394 vm_page_wire(dst_page, VM_KERN_MEMORY_OSFMK, TRUE); in vm_fault_copy()
7548 SET_PAGE_DIRTY(dst_page, TRUE); in vm_fault_copy()
7564 SET_PAGE_DIRTY(dst_page, TRUE); in vm_fault_copy()
7620 while (TRUE) { in vm_fault_classify()
7735 while (TRUE) { in kdp_lightweight_fault()
7841 VMP_CS_SET_VALIDATED(page, fault_page_size, fault_phys_offset, TRUE); in vm_page_validate_cs_fast()
7842 VMP_CS_SET_TAINTED(page, fault_page_size, fault_phys_offset, TRUE); in vm_page_validate_cs_fast()
7855 return TRUE; in vm_page_validate_cs_fast()
7870 return TRUE; in vm_page_validate_cs_fast()
7879 return TRUE; in vm_page_validate_cs_fast()
7978 page->vmp_busy = TRUE; in vm_page_map_and_validate_cs()
8103 *validated_p = TRUE; in vm_page_validate_cs_mapped_chunk()
8191 early_exit = TRUE; in vmrtf_extract()
8655 page->vmp_reference = TRUE; in revalidate_text_page()