| /xnu-12377.41.6/bsd/nfs/ ! |
| H A D | nfs_srvcache.c | 111 FALSE, 112 FALSE, 114 FALSE, 115 FALSE, 116 FALSE, 117 FALSE, 127 FALSE, 128 FALSE, 129 FALSE, 130 FALSE, [all …]
|
| /xnu-12377.41.6/pexpert/gen/ ! |
| H A D | bootargs.c | 76 return FALSE; in PE_parse_boot_argn_internal() 81 return FALSE; in PE_parse_boot_argn_internal() 85 arg_found = FALSE; in PE_parse_boot_argn_internal() 95 arg_boolean = FALSE; in PE_parse_boot_argn_internal() 141 switch (force_string ? STR : getval(cp, &val, isargsep, FALSE)) { in PE_parse_boot_argn_internal() 197 return PE_parse_boot_argn_internal(PE_boot_args(), arg_string, arg_ptr, max_len, FALSE); in PE_parse_boot_argn() 237 {"-x -ab -aa y=42", "-a", NULL, FALSE}, in parse_boot_arg_test() 238 {"-x b=96 y=42", "bx", NULL, FALSE}, in parse_boot_arg_test() 239 {"-x ab=96 y=42", "a", NULL, FALSE}, in parse_boot_arg_test() 241 FALSE}, in parse_boot_arg_test() [all …]
|
| /xnu-12377.41.6/osfmk/arm/ ! |
| H A D | arm_timer.c | 104 mytimer->has_expired = FALSE; in timer_intr() 201 mytimer->has_expired = FALSE; in timer_queue_expire_local() 213 return FALSE; in timer_resort_threshold() 295 .latency_tier_rate_limited = {FALSE, FALSE, FALSE, FALSE, FALSE, FALSE}, 300 .latency_tier_rate_limited = {FALSE, FALSE, FALSE, FALSE, TRUE, TRUE},
|
| /xnu-12377.41.6/osfmk/vm/ ! |
| H A D | vm_compressor_backing_store.c | 51 boolean_t compressor_store_stop_compaction = FALSE; 54 boolean_t vm_swapfile_create_needed = FALSE; 55 boolean_t vm_swapfile_gc_needed = FALSE; 69 bool vm_swapout_thread_running = FALSE; 77 boolean_t vm_swappin_enabled = FALSE; 115 boolean_t swp_trim_supported = FALSE; 123 boolean_t hibernate_in_progress_with_pinned_swap = FALSE; 137 boolean_t vm_swap_force_defrag = FALSE, vm_swap_force_reclaim = FALSE; 165 #define VM_SWAP_SHOULD_PIN(_size) FALSE 180 #define VM_SWAP_SHOULD_ABORT_RECLAIM() (((vm_swap_force_reclaim == FALSE) && ((vm_swapfile_total_s… [all …]
|
| H A D | vm_upl.c | 193 *empty = FALSE; in upl_abort_range() 209 dwp_finish_ctx = FALSE; in upl_abort_range() 240 *empty = FALSE; in upl_abort_range() 285 shadow_object->blocked_access = FALSE; in upl_abort_range() 307 needed = FALSE; in upl_abort_range() 327 t->vmp_free_when_done = FALSE; in upl_abort_range() 353 m->vmp_absent = FALSE; in upl_abort_range() 355 must_free = FALSE; in upl_abort_range() 357 m->vmp_restart = FALSE; in upl_abort_range() 359 must_free = FALSE; in upl_abort_range() [all …]
|
| H A D | vm_compressor.c | 123 if (__probable(popcount_c_segs == FALSE)) { in vmc_pop() 195 if (__probable(checksum_c_segs == FALSE)) { in vmc_hash() 258 boolean_t hibernate_no_swapspace = FALSE; 259 boolean_t hibernate_flush_timed_out = FALSE; 346 boolean_t decompressions_blocked = FALSE; 367 boolean_t vm_swapout_ripe_segments = FALSE; 376 boolean_t fastwake_warmup = FALSE; 377 boolean_t fastwake_recording_in_progress = FALSE; 383 boolean_t hibernate_flushing = FALSE; 609 return FALSE; in vm_compressor_out_of_space() [all …]
|
| H A D | vm_pageout.c | 326 boolean_t vm_pageout_running = FALSE; 332 static boolean_t vm_pageout_waiter = FALSE; 417 p->vmp_free_when_done = FALSE; in vm_pageout_object_terminate() 456 m->vmp_cleaning = FALSE; in vm_pageout_object_terminate() 457 m->vmp_free_when_done = FALSE; in vm_pageout_object_terminate() 469 SET_PAGE_DIRTY(m, FALSE); in vm_pageout_object_terminate() 471 m->vmp_dirty = FALSE; in vm_pageout_object_terminate() 513 m->vmp_busy = FALSE; in vm_pageout_object_terminate() 514 m->vmp_absent = FALSE; in vm_pageout_object_terminate() 524 m->vmp_overwriting = FALSE; in vm_pageout_object_terminate() [all …]
|
| H A D | vm_object.c | 319 .pager_created = FALSE, 320 .pager_initialized = FALSE, 321 .pager_ready = FALSE, 322 .pager_trusted = FALSE, 323 .can_persist = FALSE, 325 .private = FALSE, 326 .pageout = FALSE, 329 .purgeable_when_ripe = FALSE, 330 .purgeable_only_by_kernel = FALSE, 331 .shadowed = FALSE, [all …]
|
| /xnu-12377.41.6/osfmk/tests/ ! |
| H A D | test_thread_call.c | 117 T_ASSERT_EQ_INT(canceled, FALSE, "thread_call_cancel should not succeed"); in test_once_thread_call() 120 T_ASSERT_EQ_INT(pending, FALSE, "call should not be pending"); in test_once_thread_call() 129 T_ASSERT_EQ_INT(pending, FALSE, "call should not be pending"); in test_once_thread_call() 132 T_ASSERT_EQ_INT(freed, FALSE, "thread_call_free should not succeed"); in test_once_thread_call() 143 T_ASSERT_EQ_INT(canceled, FALSE, "thread_call_cancel should not succeed"); in test_once_thread_call() 181 T_ASSERT_EQ_INT(pending, FALSE, "call should not be pending"); in test_signal_thread_call() 188 T_ASSERT_EQ_INT(pending, FALSE, "call should not be pending"); in test_signal_thread_call() 196 T_ASSERT_EQ_INT(canceled, FALSE, "thread_call_cancel should not succeed"); in test_signal_thread_call()
|
| /xnu-12377.41.6/osfmk/i386/ ! |
| H A D | mp.c | 120 boolean_t smp_initialized = FALSE; 122 volatile boolean_t force_immediate_debugger_NMI = FALSE; 123 volatile boolean_t pmap_tlb_flush_timeout = FALSE; 448 istate = ml_set_interrupts_enabled(FALSE); in intel_startCPU() 530 mp_kdp_wait(TRUE, FALSE); in cpu_signal_handler() 636 mp_kdp_wait(FALSE, TRUE); in NMIInterruptHandler() 646 if (__sync_bool_compare_and_swap(&mp_kdp_is_NMI, FALSE, TRUE)) { in NMIInterruptHandler() 653 mp_kdp_is_NMI = FALSE; in NMIInterruptHandler() 655 mp_kdp_wait(FALSE, FALSE); in NMIInterruptHandler() 658 mp_kdp_wait(FALSE, FALSE); in NMIInterruptHandler() [all …]
|
| H A D | mp_native.c | 58 return FALSE; in i386_smp_init() 80 static boolean_t NMIPIs_enabled = FALSE; 91 boolean_t state = ml_set_interrupts_enabled(FALSE); in i386_send_NMI() 93 if (NMIPIs_enabled == FALSE) { in i386_send_NMI() 106 if (i_bit(MP_TLB_FLUSH, my_word) && (pmap_tlb_flush_timeout == FALSE)) { in handle_pending_TLB_flushes()
|
| H A D | i386_timer.c | 108 mytimer->has_expired = FALSE; in timer_intr() 281 mytimer->has_expired = FALSE; in timer_queue_expire_local() 299 assert(ml_get_interrupts_enabled() == FALSE); in timer_queue_expire_rescan() 307 mytimer->has_expired = FALSE; in timer_queue_expire_rescan() 328 return FALSE; in timer_resort_threshold() 457 .latency_tier_rate_limited = {FALSE, FALSE, FALSE, FALSE, TRUE, TRUE},
|
| /xnu-12377.41.6/osfmk/kdp/ ! |
| H A D | kdp.c | 150 return FALSE; in kdp_packet() 159 return FALSE; in kdp_packet() 166 return FALSE; in kdp_packet() 174 return FALSE; in kdp_packet() 196 return FALSE; in kdp_unknown() 214 return FALSE; in kdp_connect() 267 return FALSE; in kdp_disconnect() 271 return FALSE; in kdp_disconnect() 279 kdp.is_halted = kdp.is_conn = FALSE; in kdp_disconnect() 332 return FALSE; in kdp_hostinfo() [all …]
|
| /xnu-12377.41.6/osfmk/x86_64/ ! |
| H A D | pmap_pcid.c | 59 boolean_t pmap_pcid_disabled = FALSE; 69 boolean_t pcid_present = FALSE; in pmap_pcid_configure() 72 pmap_assert(ml_get_interrupts_enabled() == FALSE || get_preemption_level() != 0); in pmap_pcid_configure() 83 if (pmap_pcid_disabled == FALSE) { in pmap_pcid_configure() 84 no_shared_cr3 = FALSE; in pmap_pcid_configure() 101 cpu_datap(i)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 105 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 114 cpu_datap(ccpu)->cpu_pmap_pcid_enabled = FALSE; in pmap_pcid_configure() 266 pmap_assert(ml_get_interrupts_enabled() == FALSE || get_preemption_level() != 0); in pmap_destroy_pcid_sync() 300 boolean_t pcid_conflict = FALSE, pending_flush = FALSE; in pmap_pcid_activate() [all …]
|
| H A D | loose_ends.c | 318 boolean_t istate = TRUE, timeread = FALSE; in ml_phys_read_data() 326 istate = ml_set_interrupts_enabled(FALSE); in ml_phys_read_data() 362 if (phy_read_panic && (machine_timeout_suspended() == FALSE)) { in ml_phys_read_data() 460 boolean_t istate = TRUE, timewrite = FALSE; in ml_phys_write_data() 468 istate = ml_set_interrupts_enabled(FALSE); in ml_phys_write_data() 502 if (phy_write_panic && (machine_timeout_suspended() == FALSE)) { in ml_phys_write_data() 589 boolean_t istate, timeread = FALSE; in ml_port_io_read() 592 istate = ml_set_interrupts_enabled(FALSE); in ml_port_io_read() 624 if (phy_read_panic && (machine_timeout_suspended() == FALSE)) { in ml_port_io_read() 651 boolean_t istate, timewrite = FALSE; in ml_port_io_write() [all …]
|
| /xnu-12377.41.6/osfmk/kperf/ ! |
| H A D | task_samplers.c | 58 boolean_t dirty = FALSE, dirty_tracked = FALSE, allow_idle_exit = FALSE; in kperf_task_snapshot_sample() 59 boolean_t is_active = FALSE, is_managed = FALSE, has_assertion = FALSE; in kperf_task_snapshot_sample()
|
| /xnu-12377.41.6/tools/tests/superpages/ ! |
| H A D | testsp.c | 54 return FALSE; in check_kr() 64 return FALSE; in check_addr0() 74 return FALSE; in check_addr() 84 return FALSE; in check_align() 97 return FALSE; in check_r() 120 return FALSE; in check_nr() 134 return FALSE; in check_w() 153 return FALSE; in check_nw() 172 return FALSE; in check_rw() 226 return FALSE; in test_deallocate() [all …]
|
| /xnu-12377.41.6/bsd/kern/ ! |
| H A D | netboot.c | 134 return FALSE; in parse_booter_path() 138 return FALSE; in parse_booter_path() 145 return FALSE; in parse_booter_path() 220 return FALSE; in parse_netboot_path() 227 return FALSE; in parse_netboot_path() 231 return FALSE; in parse_netboot_path() 267 boolean_t found = FALSE; in get_root_path() 273 return FALSE; in get_root_path() 318 boolean_t have_root_path = FALSE; in netboot_info_init() 339 if (have_root_path == FALSE) { in netboot_info_init() [all …]
|
| H A D | kern_memorystatus_notify.c | 480 boolean_t ret = FALSE; in memorystatus_warn_process() 481 boolean_t found_knote = FALSE; in memorystatus_warn_process() 668 boolean_t set_privilege = FALSE; in memorystatus_low_mem_privileged_listener() 675 set_privilege = FALSE; in memorystatus_low_mem_privileged_listener() 705 *is_dirty = FALSE; in memorystatus_proc_flags_unsafe() 706 *is_dirty_tracked = FALSE; in memorystatus_proc_flags_unsafe() 707 *allow_idle_exit = FALSE; in memorystatus_proc_flags_unsafe() 708 *is_active = FALSE; in memorystatus_proc_flags_unsafe() 709 *is_managed = FALSE; in memorystatus_proc_flags_unsafe() 710 *has_assertion = FALSE; in memorystatus_proc_flags_unsafe() [all …]
|
| /xnu-12377.41.6/tests/skywalk/ ! |
| H A D | skt_xfer.c | 775 assert(connect == FALSE); in frame_populate() 1197 0, payload_length, 0, FALSE); 1299 (void *)payload, payload_length, flow_id, FALSE, in channel_port_send_fragments() 1301 fragment_count * payload_length, 0, FALSE); in channel_port_send_fragments() 1305 (void *)payload, payload_length, flow_id, FALSE, in channel_port_send_fragments() 1453 (void) frame_process(pkt, &payload, pkt_len, FALSE); in channel_port_receive() 1605 rx_complete = tx_complete = FALSE; in send_and_receive() 1688 next_batch, FALSE, FALSE, in send_and_receive() 1779 bool expect_flowadv = FALSE; in ping_pong() 1780 bool expect_stall = FALSE; in ping_pong() [all …]
|
| /xnu-12377.41.6/osfmk/console/ ! |
| H A D | video_console.c | 124 static boolean_t gc_enabled = FALSE; 125 static boolean_t gc_initialized = FALSE; 126 static boolean_t vm_initialized = FALSE; 351 if (enable == FALSE) { in gc_enable() 353 if (console_is_serial() == FALSE) { in gc_enable() 356 gc_enabled = FALSE; in gc_enable() 357 gc_ops.enable(FALSE); in gc_enable() 445 disableConsoleOutput = FALSE; in gc_enable() 460 gc_update_color(COLOR_CODE_GET(colorcode, FALSE), FALSE); in gc_hide_cursor() 465 gc_update_color(COLOR_CODE_GET(colorcodesave, FALSE), FALSE); in gc_hide_cursor() [all …]
|
| /xnu-12377.41.6/osfmk/i386/vmx/ ! |
| H A D | vmx_cpu.c | 43 boolean_t vmx_exclusive = FALSE; 186 if (FALSE == cpu->specs.vmx_on) { in vmx_on() 223 cpu->specs.vmx_on = FALSE; in vmx_off() 283 available = FALSE; in vmx_globally_available() 341 vmx_exclusive = FALSE; in host_vmxoff() 392 cpu->specs.vmx_on = FALSE; in vmx_resume() 406 return FALSE; in vmx_hv_support() 409 #define CHK(msr, shift, mask) if (!VMX_CAP(msr, shift, mask)) return FALSE; in vmx_hv_support()
|
| /xnu-12377.41.6/osfmk/kern/ ! |
| H A D | hibernate.c | 48 boolean_t need_to_unlock_decompressor = FALSE; 68 page_list_wired = hibernate_page_list_allocate(FALSE); in hibernate_alloc_page_lists() 75 page_list_pal = hibernate_page_list_allocate(FALSE); in hibernate_alloc_page_lists() 125 need_to_unlock_decompressor = FALSE; in hibernate_setup() 148 need_to_unlock_decompressor = FALSE; in hibernate_teardown()
|
| H A D | spl.h | 35 #define splhigh() (spl_t) ml_set_interrupts_enabled(FALSE) 36 #define splsched() (spl_t) ml_set_interrupts_enabled(FALSE) 37 #define splclock() (spl_t) ml_set_interrupts_enabled(FALSE)
|
| H A D | spl.c | 40 return ml_set_interrupts_enabled(FALSE); in splhigh() 47 return ml_set_interrupts_enabled(FALSE); in splsched() 54 return ml_set_interrupts_enabled(FALSE); in splclock()
|