Searched refs:nset (Results 1 – 9 of 9) sorted by relevance
| /xnu-8020.140.41/osfmk/kern/ |
| H A D | sched_amp.c | 487 processor_set_t nset = pset; in sched_amp_steal_thread() local 501 nset = pcore_set; in sched_amp_steal_thread() 503 assert(nset != pset); in sched_amp_steal_thread() 505 if (sched_get_pset_load_average(nset, 0) >= sched_amp_steal_threshold(nset, spill_pending)) { in sched_amp_steal_thread() 508 pset = nset; in sched_amp_steal_thread() 655 processor_set_t nset = pset; in sched_amp_choose_processor() local 663 nset = pcore_set; in sched_amp_choose_processor() 664 assert(nset != NULL); in sched_amp_choose_processor() 666 nset = ecore_set; in sched_amp_choose_processor() 667 assert(nset != NULL); in sched_amp_choose_processor() [all …]
|
| H A D | sched_amp_common.c | 215 should_spill_to_ecores(processor_set_t nset, thread_t thread) in should_spill_to_ecores() argument 217 if (nset->pset_cluster_type == PSET_AMP_E) { in should_spill_to_ecores() 237 if ((nset->recommended_bitmask & nset->cpu_state_map[PROCESSOR_IDLE]) != 0) { in should_spill_to_ecores() 242 …if ((sched_get_pset_load_average(nset, 0) >= sched_amp_spill_threshold(nset)) && /* There is alre… in should_spill_to_ecores()
|
| H A D | sched_dualq.c | 457 processor_set_t nset = next_pset(cset); in sched_dualq_steal_thread() local 463 while (nset != pset) { in sched_dualq_steal_thread() 465 cset = nset; in sched_dualq_steal_thread() 475 nset = next_pset(cset); in sched_dualq_steal_thread()
|
| H A D | sched_traditional.c | 714 processor_set_t nset, cset = pset; in sched_traditional_steal_thread() local 733 nset = next_pset(cset); in sched_traditional_steal_thread() 735 if (nset != pset) { in sched_traditional_steal_thread() 738 cset = nset; in sched_traditional_steal_thread() 741 } while (nset != pset); in sched_traditional_steal_thread()
|
| H A D | sched_amp_common.h | 56 bool should_spill_to_ecores(processor_set_t nset, thread_t thread);
|
| H A D | sched_prim.c | 2211 processor_set_t nset = pset_array[pset_id]; in other_psets_have_earlier_rt_threads_pending() local 2213 …if (deadline_add(nset->stealable_rt_threads_earliest_deadline, rt_deadline_epsilon) < earliest_dea… in other_psets_have_earlier_rt_threads_pending() 2241 processor_set_t nset = next_pset(pset); in choose_next_rt_processor_for_IPI() local 2242 assert(nset != starting_pset); in choose_next_rt_processor_for_IPI() 2243 pset = change_locked_pset(pset, nset); in choose_next_rt_processor_for_IPI() 2253 processor_set_t nset = next_pset(pset); in choose_next_rt_processor_for_IPI() local 2254 if (nset == starting_pset) { in choose_next_rt_processor_for_IPI() 2257 pset = change_locked_pset(pset, nset); in choose_next_rt_processor_for_IPI() 4329 processor_set_t nset = pset_array[pset_id]; in sched_rtlocal_steal_thread() local 4331 if (nset->stealable_rt_threads_earliest_deadline < target_deadline) { in sched_rtlocal_steal_thread() [all …]
|
| H A D | sfi.c | 298 processor_set_t pset, nset; in sfi_timer_global_off() local 349 nset = processor->processor_set; in sfi_timer_global_off() 350 if (nset != pset) { in sfi_timer_global_off() 352 pset = nset; in sfi_timer_global_off()
|
| /xnu-8020.140.41/osfmk/ipc/ |
| H A D | mach_port.c | 1453 ipc_pset_t nset = ips_object_to_pset(ps_obj); in mach_port_move_member() local 1455 ipc_mqueue_add_locked(&port->ip_messages, nset, &link); in mach_port_move_member() 1456 ips_mq_unlock(nset); in mach_port_move_member() 1458 keep_waitq_set = &nset->ips_wqset; in mach_port_move_member()
|
| /xnu-8020.140.41/bsd/kern/ |
| H A D | kern_sig.c | 1232 sigset_t nset; in __pthread_sigmask() local 1245 error = copyin(set, &nset, sizeof(sigset_t)); in __pthread_sigmask() 1252 ut->uu_sigmask |= (nset & ~sigcantmask); in __pthread_sigmask() 1256 ut->uu_sigmask &= ~(nset); in __pthread_sigmask() 1261 ut->uu_sigmask = (nset & ~sigcantmask); in __pthread_sigmask()
|