| /xnu-8796.121.2/osfmk/kern/ |
| H A D | processor.h | 614 assert(bit_test(pset->cpu_bitmask, cpuid)); in pset_update_processor_state() 624 if (bit_test(pset->cpu_available_map, cpuid) && (new_state < PROCESSOR_IDLE)) { in pset_update_processor_state() 627 } else if (!bit_test(pset->cpu_available_map, cpuid) && (new_state >= PROCESSOR_IDLE)) { in pset_update_processor_state() 652 if (!bit_test(atomic_load(&node->pset_non_rt_primary_map), pset->pset_id)) { in pset_update_processor_state() 655 if (!bit_test(atomic_load(&node->pset_idle_primary_map), pset->pset_id)) { in pset_update_processor_state() 659 if (!bit_test(atomic_load(&node->pset_non_rt_map), pset->pset_id)) { in pset_update_processor_state() 662 if (!bit_test(atomic_load(&node->pset_idle_map), pset->pset_id)) { in pset_update_processor_state() 669 if (bit_test(atomic_load(&node->pset_idle_map), pset->pset_id)) { in pset_update_processor_state() 677 if (bit_test(atomic_load(&node->pset_idle_primary_map), pset->pset_id)) { in pset_update_processor_state()
|
| H A D | bits.h | 57 #define bit_test(x, b) ((bool)((x) & BIT(b))) macro 89 bool _bit_is_set = bit_test(*_map, _n); \ 99 bool _bit_is_set = bit_test(*_map, _n); \ 192 return bit_test(prev, n); in atomic_bit_set() 200 return bit_test(prev, n); in atomic_bit_clear() 307 return bit_test(map[bitmap_index(n)], bitmap_bit(n)); in bitmap_test()
|
| H A D | sched_amp.c | 242 if (pset == ecore_set && bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in sched_amp_choose_thread() 295 bool spill_pending = bit_test(pset->pending_spill_cpu_mask, processor->cpu_id); in sched_amp_processor_queue_empty() 332 if (pset == ecore_set && bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in sched_amp_processor_csw_check() 377 if (pset == ecore_set && bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in sched_amp_processor_queue_has_priority() 500 bool spill_pending = bit_test(pset->pending_spill_cpu_mask, processor->cpu_id); in sched_amp_steal_thread()
|
| H A D | sched_prim.c | 2311 if (!bit_test(atomic_load(&node->pset_non_rt_map), pset->pset_id)) { in pset_commit_processor_to_new_thread() 2317 if (!bit_test(atomic_load(&node->pset_non_rt_primary_map), pset->pset_id)) { in pset_commit_processor_to_new_thread() 2482 pending_AST_URGENT = bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id); in thread_select() 2483 pending_AST_PREEMPT = bit_test(pset->pending_AST_PREEMPT_cpu_mask, processor->cpu_id); in thread_select() 2608 if (bit_test(pset->rt_pending_spill_cpu_mask, processor->cpu_id)) { in thread_select() 2765 if (!pending_AST_URGENT && bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id)) { in thread_select() 2771 bool spill_pending = bit_test(pset->rt_pending_spill_cpu_mask, processor->cpu_id); in thread_select() 2841 if (!pending_AST_URGENT && bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id)) { in thread_select() 2876 if ((!pending_AST_URGENT && bit_test(pset->pending_AST_URGENT_cpu_mask, processor->cpu_id)) || in thread_select() 2877 (!pending_AST_PREEMPT && bit_test(pset->pending_AST_PREEMPT_cpu_mask, processor->cpu_id))) { in thread_select() [all …]
|
| H A D | sched_amp_common.c | 148 if (bit_test(pset->pending_spill_cpu_mask, processor->cpu_id)) { in pset_signal_spill()
|
| H A D | sched_clutch.c | 3911 …if (bit_test(target_pset->cpu_running_cluster_shared_rsrc_thread[CLUSTER_SHARED_RSRC_TYPE_RR], for… in sched_edge_cpu_running_foreign_shared_rsrc_available() 3916 …if (bit_test(target_pset->cpu_running_cluster_shared_rsrc_thread[CLUSTER_SHARED_RSRC_TYPE_NATIVE_F… in sched_edge_cpu_running_foreign_shared_rsrc_available() 4256 if (bit_test(local_candidate_map, previous_cluster)) { in sched_edge_iterate_clusters_ordered()
|
| /xnu-8796.121.2/osfmk/arm64/ |
| H A D | lock_ticket_pv.c | 57 if (!bit_test(wmask, tcpunum)) { in hw_lck_ticket_unlock_kick_pv()
|
| /xnu-8796.121.2/iokit/Kernel/arm/ |
| H A D | AppleARMSMP.cpp | 368 bit_test(cpu_power_state_mask, i)) { in is_cluster_powering_down() 424 if (!bit_test(online_clusters_mask, cluster_id)) { in PE_cpu_power_enable()
|
| /xnu-8796.121.2/bsd/skywalk/mem/ |
| H A D | skmem_region.c | 1517 ASSERT(bit_test(skr->skr_seg_bmap[i / BMAPSZ], i % BMAPSZ)); in skmem_region_depopulate() 1549 ASSERT(bit_test(*bmap, i % BMAPSZ)); in sksegment_create() 1589 ASSERT(!bit_test(*bmap, i % BMAPSZ)); in sksegment_destroy() 1868 if (!bit_test(skr->skr_seg_bmap[idx / BMAPSZ], idx % BMAPSZ)) { in sksegment_alloc_with_idx() 1879 VERIFY(!bit_test(skr->skr_seg_bmap[idx / BMAPSZ], idx % BMAPSZ)); in sksegment_alloc_with_idx()
|
| /xnu-8796.121.2/bsd/skywalk/nexus/ |
| H A D | nexus.c | 2616 if (bit_test(*bmap, j)) { in nx_port_alloc() 2677 ASSERT(!bit_test(nx->nx_ports_bmap[nx_port / NX_PORT_CHUNK], in nx_port_alloc() 2715 ASSERT(!bit_test(*bmap, j)); in nx_port_free() 2760 if (bit_test(*bmap, j)) { in nx_port_bind_info() 2831 (!bit_test(*bmap, j) && nx->nx_active_ports > 0)); in nx_port_unbind() 2843 ASSERT(!bit_test(*bmap, j)); in nx_port_unbind() 2954 if (bit_test(bmap, j)) {
|
| /xnu-8796.121.2/bsd/skywalk/nexus/flowswitch/flow/ |
| H A D | flow_owner.c | 502 ASSERT(!bit_test(bmap[chunk_idx], bit_pos)); in flow_owner_flowadv_index_free()
|