| /xnu-8020.140.41/bsd/kern/ |
| H A D | sys_ulock.c | 302 ull->ull_owner = THREAD_NULL; in ull_alloc() 314 assert(ull->ull_owner == THREAD_NULL); in ull_free() 457 if (*owner == THREAD_NULL) { in ulock_resolve_owner() 496 thread_t owner_thread = THREAD_NULL; in sys_ulock_wait2() 497 thread_t old_owner = THREAD_NULL; in sys_ulock_wait2() 655 if (owner_thread == THREAD_NULL) { in sys_ulock_wait2() 684 ull->ull_owner = THREAD_NULL; in sys_ulock_wait2() 686 if (owner_thread != THREAD_NULL) { in sys_ulock_wait2() 736 if (set_owner && owner_thread != THREAD_NULL) { in sys_ulock_wait2() 739 assert(owner_thread == THREAD_NULL); in sys_ulock_wait2() [all …]
|
| H A D | kern_resource.c | 802 if (thread != THREAD_NULL && in do_background_socket() 809 if (thread != THREAD_NULL) { in do_background_socket() 821 if (thread == THREAD_NULL) { in do_background_socket() 1709 thread = THREAD_NULL; in iopolicysys_disk() 1750 thread = THREAD_NULL; in iopolicysys_disk() 1798 if (thread != THREAD_NULL) { in iopolicysys_disk() 1805 if (thread != THREAD_NULL) { in iopolicysys_disk() 1934 thread = THREAD_NULL; in iopolicysys_vfs_atime_updates() 1956 if (thread != THREAD_NULL) { in iopolicysys_vfs_atime_updates() 1963 if (thread != THREAD_NULL) { in iopolicysys_vfs_atime_updates() [all …]
|
| /xnu-8020.140.41/osfmk/kern/ |
| H A D | thread_act.c | 207 if (thread == THREAD_NULL) { in thread_terminate() 246 if (thread == THREAD_NULL) { in thread_terminate_from_user() 268 if (thread == THREAD_NULL) { in thread_terminate_pinned() 338 if (thread == THREAD_NULL || get_threadtask(thread) == kernel_task) { in thread_suspend() 366 if (thread == THREAD_NULL || get_threadtask(thread) == kernel_task) { in thread_resume() 399 if (thread == THREAD_NULL) { in thread_depress_abort_from_user() 449 if (thread == THREAD_NULL) { in thread_abort() 473 if (thread == THREAD_NULL) { in thread_abort_safely() 516 if (thread == THREAD_NULL) { in thread_info() 545 if (thread == THREAD_NULL) { in thread_get_state_internal() [all …]
|
| H A D | sync_sema.c | 369 if (thread != THREAD_NULL) { in semaphore_signal_internal() 471 if (thread == THREAD_NULL) { in semaphore_signal_thread_trap() 475 thread = THREAD_NULL; in semaphore_signal_thread_trap() 485 if (thread != THREAD_NULL) { in semaphore_signal_thread_trap() 515 THREAD_NULL, in semaphore_signal() 546 THREAD_NULL, in semaphore_signal_internal_trap() 573 THREAD_NULL, in semaphore_signal_all() 597 THREAD_NULL, in semaphore_signal_all_trap() 654 assert(self->handoff_thread == THREAD_NULL); in semaphore_wait_continue() 682 thread_t handoff_thread = THREAD_NULL; in semaphore_wait_internal() [all …]
|
| H A D | sched_dualq.c | 227 return THREAD_NULL; in sched_dualq_choose_thread() 241 return THREAD_NULL; in sched_dualq_choose_thread() 250 if (potential_thread == THREAD_NULL) { in sched_dualq_choose_thread() 251 return THREAD_NULL; in sched_dualq_choose_thread() 260 return THREAD_NULL; in sched_dualq_choose_thread() 266 return THREAD_NULL; in sched_dualq_choose_thread() 278 return THREAD_NULL; in sched_dualq_choose_thread() 480 return THREAD_NULL; in sched_dualq_steal_thread() 514 if (thread != THREAD_NULL && thread->sched_stamp != sched_tick) { in sched_dualq_thread_update_scan()
|
| H A D | sched_grrr.c | 336 return THREAD_NULL; in sched_grrr_steal_thread() 380 while ((thread = sched_grrr_choose_thread(processor, IDLEPRI, AST_NONE)) != THREAD_NULL) { in sched_grrr_processor_queue_shutdown() 388 while ((thread = (thread_t)(void *)dequeue_head(&bqueue)) != THREAD_NULL) { in sched_grrr_processor_queue_shutdown() 394 while ((thread = (thread_t)(void *)dequeue_head(&tqueue)) != THREAD_NULL) { in sched_grrr_processor_queue_shutdown() 597 return THREAD_NULL; in grrr_intragroup_schedule() 601 if (thread == THREAD_NULL) { in grrr_intragroup_schedule() 624 return THREAD_NULL; in grrr_intergroup_schedule() 694 runq->groups[index].current_client = THREAD_NULL; in grrr_runqueue_init() 736 if (group->current_client == THREAD_NULL || in grrr_enqueue() 764 if (thread != THREAD_NULL) { in grrr_select() [all …]
|
| H A D | syscall_subr.c | 201 thread_t thread = THREAD_NULL; in thread_switch() 255 if (thread != THREAD_NULL) { in thread_switch() 272 if (thread != THREAD_NULL) { in thread_switch() 282 if (pulled_thread != THREAD_NULL) { in thread_switch() 357 if (thread != THREAD_NULL) { in thread_handoff_internal() 372 if (pulled_thread != THREAD_NULL) { in thread_handoff_internal()
|
| H A D | sched_traditional.c | 332 if (thread != THREAD_NULL) { in sched_traditional_choose_thread() 384 return THREAD_NULL; in sched_traditional_choose_thread_from_runq() 570 while ((thread = cqe_dequeue_head(&tqueue, struct thread, runq_links)) != THREAD_NULL) { in sched_traditional_processor_queue_shutdown() 691 return THREAD_NULL; in sched_traditional_steal_processor_thread() 725 if (thread != THREAD_NULL) { in sched_traditional_steal_thread() 745 return THREAD_NULL; in sched_traditional_steal_thread() 778 if (thread != THREAD_NULL && thread->sched_stamp != sched_tick) { in sched_traditional_thread_update_scan()
|
| H A D | thread.c | 686 assert(thread->handoff_thread == THREAD_NULL); in thread_terminate_self() 708 if (thread == THREAD_NULL) { in thread_ref_release() 1246 assert(vm_pageout_scan_thread == THREAD_NULL); in thread_create_internal() 1930 if (thread == THREAD_NULL) { in thread_info_internal() 2041 quantum_time = SCHED(initial_quantum_size)(THREAD_NULL); in thread_info_internal() 2911 return thread != THREAD_NULL? thread->thread_id: 0; in thread_tid() 3076 if (thread == THREAD_NULL) { in thread_dispatchqaddr() 3104 if (thread == THREAD_NULL) { in thread_wqquantum_addr() 3131 if (thread == THREAD_NULL) { in thread_rettokern_addr() 3178 if (thread != THREAD_NULL) { in thread_reference() [all …]
|
| H A D | waitq.c | 837 thread_t thread = THREAD_NULL; in waitq_select_queue_flush() 868 thread_t thread = THREAD_NULL; in waitq_queue_iterate_locked() 869 thread_t first_thread = THREAD_NULL; in waitq_queue_iterate_locked() 872 thread_t t = THREAD_NULL; in waitq_queue_iterate_locked() 883 if (first_thread == THREAD_NULL) { in waitq_queue_iterate_locked() 891 if (t != THREAD_NULL) { in waitq_queue_iterate_locked() 908 if (t == THREAD_NULL) { in waitq_queue_iterate_locked() 945 thread_t first_thread = THREAD_NULL; in waitq_prioq_iterate_locked() 946 thread_t thread = THREAD_NULL; in waitq_prioq_iterate_locked() 980 if (first_thread == THREAD_NULL) { in waitq_prioq_iterate_locked() [all …]
|
| H A D | sched_amp.c | 248 return THREAD_NULL; in sched_amp_choose_thread() 258 return THREAD_NULL; in sched_amp_choose_thread() 486 thread_t thread = THREAD_NULL; in sched_amp_steal_thread() 498 return THREAD_NULL; in sched_amp_steal_thread() 564 if (thread != THREAD_NULL && thread->sched_stamp != sched_tick) { in sched_amp_thread_update_scan()
|
| H A D | lock_rw.c | 1038 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_exclusive_check_contended() 1072 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_exclusive_internal_inline() 1281 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_shared_internal_inline() 1547 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_shared_to_exclusive() 1632 ordered_store_rw_owner(lock, THREAD_NULL); in lck_rw_lock_exclusive_to_shared() 1752 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_try_lock_shared_internal_inline() 1835 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_try_lock_exclusive_internal_inline() 2035 …assertf(lock->lck_rw_owner == THREAD_NULL, "state=0x%x, owner=%p", lock->lck_rw_data, lock->lck_rw… in lck_rw_done() 2053 ordered_store_rw_owner(lock, THREAD_NULL); in lck_rw_done() 2098 …assertf(lck->lck_rw_owner == THREAD_NULL, "state=0x%x, owner=%p", lck->lck_rw_data, lck->lck_rw_ow… in lck_rw_unlock_shared() [all …]
|
| H A D | ux_handler.c | 140 thread_t target_thread = THREAD_NULL; in catch_mach_exception_raise() 143 if ((target_thread = convert_port_to_thread(thread_port)) == THREAD_NULL) { in catch_mach_exception_raise()
|
| H A D | mk_sp.c | 84 if (thread == THREAD_NULL || in thread_set_policy() 198 if (thread == THREAD_NULL) { in thread_policy()
|
| H A D | ipc_tt.c | 1634 if (thread == THREAD_NULL) { in thread_get_special_port() 1668 if (thread == THREAD_NULL) { in thread_get_special_port_from_user() 1772 if (thread == THREAD_NULL) { in thread_set_special_port() 3022 thread_t thread = THREAD_NULL; in convert_port_to_thread_with_flavor_locked() 3030 return THREAD_NULL; in convert_port_to_thread_with_flavor_locked() 3035 if (thread == THREAD_NULL) { in convert_port_to_thread_with_flavor_locked() 3036 return THREAD_NULL; in convert_port_to_thread_with_flavor_locked() 3041 return THREAD_NULL; in convert_port_to_thread_with_flavor_locked() 3049 return THREAD_NULL; in convert_port_to_thread_with_flavor_locked() 3054 return THREAD_NULL; in convert_port_to_thread_with_flavor_locked() [all …]
|
| H A D | sched_clutch.c | 2545 return THREAD_NULL; in sched_clutch_thread_highest_remove() 2548 thread_t highest_thread = THREAD_NULL; in sched_clutch_thread_highest_remove() 2638 thread_t thread = THREAD_NULL; in sched_clutch_root_highest_foreign_thread_remove() 2818 if (thread == THREAD_NULL) { in sched_clutch_initial_quantum_size() 2882 return THREAD_NULL; in sched_clutch_choose_thread() 2894 return THREAD_NULL; in sched_clutch_choose_thread() 2897 thread_t thread = THREAD_NULL; in sched_clutch_choose_thread() 3080 return THREAD_NULL; in sched_clutch_steal_thread() 3114 if (thread != THREAD_NULL && thread->sched_stamp != sched_tick) { in sched_clutch_thread_update_scan() 3578 return THREAD_NULL; in sched_edge_choose_thread() [all …]
|
| H A D | extmod_statistics.c | 94 if ((ctask == kernel_task) || (target == THREAD_NULL)) { in extmod_statistics_incr_thread_set_state()
|
| /xnu-8020.140.41/bsd/net/ |
| H A D | flowadv.c | 108 static thread_t fadv_thread = THREAD_NULL; 165 if (!fadv_active && fadv_thread != THREAD_NULL) { in flowadv_add() 179 if (!fadv_active && fadv_thread != THREAD_NULL) { in flowadv_add_entry()
|
| H A D | nwk_wq.c | 53 thread_t nwk_wq_thread = THREAD_NULL; in nwk_wq_init()
|
| /xnu-8020.140.41/osfmk/corpses/ |
| H A D | corpse.c | 445 if (thread != THREAD_NULL) { in task_generate_corpse() 493 thread_t thread = THREAD_NULL; in task_enqueue_exception_with_corpse() 504 if (thread == THREAD_NULL) { in task_enqueue_exception_with_corpse() 538 thread_t thread = THREAD_NULL; in task_generate_corpse_internal() 539 thread_t thread_next = THREAD_NULL; in task_generate_corpse_internal() 659 if (thread != THREAD_NULL) { in task_generate_corpse_internal()
|
| /xnu-8020.140.41/osfmk/ipc/ |
| H A D | ipc_eventlink.c | 182 ipc_eventlink->el_thread = THREAD_NULL; in ipc_eventlink_initialize() 292 temp_ipc_eventlink->el_thread = THREAD_NULL; in ipc_eventlink_destroy_internal() 331 if (associated_thread[i] != THREAD_NULL && in ipc_eventlink_destroy_internal() 713 thread_t handoff_thread = THREAD_NULL; in ipc_eventlink_signal_wait_internal() 773 assert(self->handoff_thread == THREAD_NULL); in ipc_eventlink_signal_wait_internal() 783 self->handoff_thread = THREAD_NULL; in ipc_eventlink_signal_wait_internal() 793 assert(self->handoff_thread == THREAD_NULL); in ipc_eventlink_signal_wait_internal() 807 assert(self->handoff_thread == THREAD_NULL); in ipc_eventlink_signal_wait_internal()
|
| /xnu-8020.140.41/tests/ |
| H A D | contextswitch.c | 145 thread_switch(THREAD_NULL, SWITCH_OPTION_NONE, 0); in thread() 148 if (thread_switch(THREAD_NULL, SWITCH_OPTION_NONE, 0)) { in thread() 155 if (thread_switch(THREAD_NULL, SWITCH_OPTION_NONE, 0)) { in thread()
|
| /xnu-8020.140.41/bsd/nfs/ |
| H A D | nfs_upcall.c | 186 nfsrv_uc_queue_tbl[i].ucq_thd = THREAD_NULL; in nfsrv_uc_init() 264 if (nfsrv_uc_queue_tbl[i].ucq_thd != THREAD_NULL) { in nfsrv_uc_stop() 267 nfsrv_uc_queue_tbl[i].ucq_thd = THREAD_NULL; in nfsrv_uc_stop()
|
| /xnu-8020.140.41/osfmk/mach/ |
| H A D | mach_types.h | 284 #define THREAD_NULL ((thread_t) NULL) macro 313 #define THREAD_NULL ((thread_t) 0) macro
|
| /xnu-8020.140.41/bsd/net/pktsched/ |
| H A D | pktsched_netem.c | 1054 if (__probable(ne->netem_output_thread != THREAD_NULL)) { in netem_enqueue() 1224 ne->netem_output_thread = THREAD_NULL; in netem_create() 1248 if (ne->netem_output_thread != THREAD_NULL) { in netem_destroy() 1274 ne->netem_output_thread = THREAD_NULL; in netem_destroy() 1276 ASSERT(ne->netem_output_thread == THREAD_NULL); in netem_destroy()
|