| /xnu-10002.81.5/osfmk/kern/ |
| H A D | processor.h | 619 assert(bit_test(pset->cpu_bitmask, cpuid)); in pset_update_processor_state() 629 if (bit_test(pset->cpu_available_map, cpuid) && (new_state < PROCESSOR_IDLE)) { in pset_update_processor_state() 632 } else if (!bit_test(pset->cpu_available_map, cpuid) && (new_state >= PROCESSOR_IDLE)) { in pset_update_processor_state() 657 if (!bit_test(atomic_load(&node->pset_non_rt_primary_map), pset->pset_id)) { in pset_update_processor_state() 660 if (!bit_test(atomic_load(&node->pset_idle_primary_map), pset->pset_id)) { in pset_update_processor_state() 664 if (!bit_test(atomic_load(&node->pset_non_rt_map), pset->pset_id)) { in pset_update_processor_state() 667 if (!bit_test(atomic_load(&node->pset_idle_map), pset->pset_id)) { in pset_update_processor_state() 674 if (bit_test(atomic_load(&node->pset_idle_map), pset->pset_id)) { in pset_update_processor_state() 682 if (bit_test(atomic_load(&node->pset_idle_primary_map), pset->pset_id)) { in pset_update_processor_state()
|
| H A D | bits.h | 60 #define bit_test(x, b) ((bool)((x) & BIT(b))) macro 92 bool _bit_is_set = bit_test(*_map, _n); \ 102 bool _bit_is_set = bit_test(*_map, _n); \ 195 return bit_test(prev, n); in atomic_bit_set() 203 return bit_test(prev, n); in atomic_bit_clear() 310 return bit_test(map[bitmap_index(n)], bitmap_bit(n)); in bitmap_test()
|
| H A D | sched_amp.c | 242 if (pset == ecore_set && bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in sched_amp_choose_thread() 295 bool spill_pending = bit_test(pset->pending_spill_cpu_mask, processor->cpu_id); in sched_amp_processor_queue_empty() 332 if (pset == ecore_set && bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in sched_amp_processor_csw_check() 377 if (pset == ecore_set && bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in sched_amp_processor_queue_has_priority() 500 bool spill_pending = bit_test(pset->pending_spill_cpu_mask, processor->cpu_id); in sched_amp_steal_thread()
|
| H A D | sched_prim.c | 2366 if (!bit_test(atomic_load(&node->pset_non_rt_map), pset->pset_id)) { in pset_commit_processor_to_new_thread() 2372 if (!bit_test(atomic_load(&node->pset_non_rt_primary_map), pset->pset_id)) { in pset_commit_processor_to_new_thread() 2537 pending_AST_URGENT = bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id); in thread_select() 2538 pending_AST_PREEMPT = bit_test(pset->pending_AST_PREEMPT_cpu_mask, processor->cpu_id); in thread_select() 2663 if (bit_test(pset->rt_pending_spill_cpu_mask, processor->cpu_id)) { in thread_select() 2820 if (!pending_AST_URGENT && bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id)) { in thread_select() 2826 bool spill_pending = bit_test(pset->rt_pending_spill_cpu_mask, processor->cpu_id); in thread_select() 2896 if (!pending_AST_URGENT && bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id)) { in thread_select() 2931 if ((!pending_AST_URGENT && bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id)) || in thread_select() 2932 (!pending_AST_PREEMPT && bit_test(pset->pending_AST_PREEMPT_cpu_mask, processor->cpu_id))) { in thread_select() [all …]
|
| H A D | sched_amp_common.c | 149 if (bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in pset_signal_spill()
|
| H A D | sched_clutch.c | 3879 …if (bit_test(target_pset->cpu_running_cluster_shared_rsrc_thread[CLUSTER_SHARED_RSRC_TYPE_RR], for… in sched_edge_cpu_running_foreign_shared_rsrc_available() 3884 …if (bit_test(target_pset->cpu_running_cluster_shared_rsrc_thread[CLUSTER_SHARED_RSRC_TYPE_NATIVE_F… in sched_edge_cpu_running_foreign_shared_rsrc_available() 4089 if (bit_test(processor->processor_set->perfcontrol_cpu_migration_bitmask, processor->cpu_id) && in sched_edge_thread_avoid_processor() 4239 if (bit_test(local_candidate_map, previous_cluster)) { in sched_edge_iterate_clusters_ordered()
|
| /xnu-10002.81.5/iokit/Kernel/arm/ |
| H A D | AppleARMSMP.cpp | 368 bit_test(cpu_power_state_mask, i)) { in is_cluster_powering_down() 423 return bit_test(online_clusters_mask, cluster_id); in PE_cpu_power_check_kdp() 435 if (!bit_test(online_clusters_mask, cluster_id)) { in PE_cpu_power_enable()
|
| /xnu-10002.81.5/osfmk/arm64/ |
| H A D | lock_ticket_pv.c | 57 if (!bit_test(wmask, tcpunum)) { in hw_lck_ticket_unlock_kick_pv()
|
| /xnu-10002.81.5/bsd/skywalk/mem/ |
| H A D | skmem_region.c | 1520 ASSERT(bit_test(skr->skr_seg_bmap[i / BMAPSZ], i % BMAPSZ)); in skmem_region_depopulate() 1552 ASSERT(bit_test(*bmap, i % BMAPSZ)); in sksegment_create() 1592 ASSERT(!bit_test(*bmap, i % BMAPSZ)); in sksegment_destroy() 1871 if (!bit_test(skr->skr_seg_bmap[idx / BMAPSZ], idx % BMAPSZ)) { in sksegment_alloc_with_idx() 1882 VERIFY(!bit_test(skr->skr_seg_bmap[idx / BMAPSZ], idx % BMAPSZ)); in sksegment_alloc_with_idx()
|
| /xnu-10002.81.5/bsd/skywalk/nexus/ |
| H A D | nexus.c | 2634 if (bit_test(*bmap, j)) { in nx_port_alloc() 2695 ASSERT(!bit_test(nx->nx_ports_bmap[nx_port / NX_PORT_CHUNK], in nx_port_alloc() 2733 ASSERT(!bit_test(*bmap, j)); in nx_port_free() 2778 if (bit_test(*bmap, j)) { in nx_port_bind_info() 2849 (!bit_test(*bmap, j) && nx->nx_active_ports > 0)); in nx_port_unbind() 2861 ASSERT(!bit_test(*bmap, j)); in nx_port_unbind() 2972 if (bit_test(bmap, j)) {
|
| /xnu-10002.81.5/bsd/skywalk/nexus/flowswitch/flow/ |
| H A D | flow_owner.c | 502 ASSERT(!bit_test(bmap[chunk_idx], bit_pos)); in flow_owner_flowadv_index_free()
|