| /xnu-11417.121.6/tools/cocci/ ! |
| H A D | OSAtomic_rewrite.cocci | 11 + os_atomic_inc_orig(E, relaxed) 14 + os_atomic_inc_orig(E, relaxed) 17 + os_atomic_inc_orig(E, relaxed) 20 + os_atomic_inc_orig(E, relaxed) 23 + os_atomic_inc_orig(E, relaxed) 26 + os_atomic_inc_orig(E, relaxed) 29 + os_atomic_inc_orig(E, relaxed) 32 + os_atomic_inc_orig(E, relaxed) 35 + os_atomic_inc_orig(E, relaxed) 38 + os_atomic_inc_orig(E, relaxed) [all …]
|
| H A D | hw_atomic_rewrite.cocci | 11 + os_atomic_dec_orig(E, relaxed) 14 + os_atomic_dec(E, relaxed) 17 + os_atomic_sub_orig(E, F, relaxed) 20 + os_atomic_sub(E, F, relaxed) 23 + os_atomic_inc_orig(E, relaxed) 26 + os_atomic_inc(E, relaxed) 29 + os_atomic_add_orig(E, F, relaxed) 32 + os_atomic_add(E, F, relaxed) 39 + os_atomic_inc_orig(E, relaxed) 42 + os_atomic_inc(E, relaxed) [all …]
|
| /xnu-11417.121.6/tests/ ! |
| H A D | os_atomic.cpp | 16 T_ASSERT_EQ(os_atomic_inc_orig(&i, relaxed), 0, "atomic inc"); 17 T_ASSERT_EQ(os_atomic_cmpxchg(&i, 1, 0, relaxed), true, "os_atomic_cmpxchg"); 18 os_atomic_rmw_loop(&i, a, b, relaxed, { 22 T_ASSERT_EQ(os_atomic_inc_orig(&old_i, relaxed), 0, "atomic inc"); 23 T_ASSERT_EQ(os_atomic_cmpxchg(&old_i, 1, 0, relaxed), true, "os_atomic_cmpxchg"); 24 os_atomic_rmw_loop(&old_i, a, b, relaxed, { 28 T_ASSERT_EQ(os_atomic_inc_orig(&v_i, relaxed), 0, "atomic inc"); 29 T_ASSERT_EQ(os_atomic_cmpxchg(&v_i, 1, 0, relaxed), true, "os_atomic_cmpxchg"); 30 os_atomic_rmw_loop(&v_i, a, b, relaxed, {
|
| H A D | os_refcnt.c | 430 while (os_atomic_load(&pcpu_perf_step, relaxed) == 0) { in worker_ref() 433 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in worker_ref() 439 os_atomic_add(count, n, relaxed); in worker_ref() 447 while (os_atomic_load(&pcpu_perf_step, relaxed) == 0) { in worker_pcpu_ref() 450 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in worker_pcpu_ref() 456 os_atomic_add(count, n, relaxed); in worker_pcpu_ref() 464 os_atomic_store(&pcpu_perf_step, 1, relaxed); in warmup_thread_pool() 468 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in warmup_thread_pool() 473 os_atomic_store(&pcpu_perf_step, 0, relaxed); in warmup_thread_pool() 496 os_atomic_store(&pcpu_perf_step, 1, relaxed); [all …]
|
| /xnu-11417.121.6/libkern/gen/ ! |
| H A D | OSAtomicOperations.c | 101 return os_atomic_add_orig(address, (SInt8)amount, relaxed); in OSAddAtomic8() 107 return os_atomic_add_orig(address, (SInt16)amount, relaxed); in OSAddAtomic16() 115 return os_atomic_add_orig(address, amount, relaxed); in OSAddAtomic() 125 return os_atomic_add_orig(aligned_address, amount, relaxed); in OSAddAtomic64() 132 return os_atomic_add_orig(address, theAmount, relaxed); in OSAddAtomicLong() 139 return os_atomic_inc_orig(value, relaxed); in OSIncrementAtomic() 146 return os_atomic_dec_orig(value, relaxed); in OSDecrementAtomic() 153 return os_atomic_and_orig(value, mask, relaxed); in OSBitAndAtomic() 160 return os_atomic_or_orig(value, mask, relaxed); in OSBitOrAtomic() 167 return os_atomic_xor_orig(value, mask, relaxed); in OSBitXorAtomic() [all …]
|
| /xnu-11417.121.6/osfmk/kern/ ! |
| H A D | exclaves_memory.c | 94 e->pages_alloced = os_atomic_load(&exclaves_allocation_statistics.pages_alloced, relaxed); in exclaves_memory_report_accounting() 95 e->pages_freed = os_atomic_load(&exclaves_allocation_statistics.pages_freed, relaxed); in exclaves_memory_report_accounting() 96 e->time_allocating = os_atomic_load(&exclaves_allocation_statistics.time_allocating, relaxed); in exclaves_memory_report_accounting() 97 e->max_alloc_latency = os_atomic_load(&exclaves_allocation_statistics.max_alloc_latency, relaxed); in exclaves_memory_report_accounting() 98 …ncy_highbit0 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[0], relaxed); in exclaves_memory_report_accounting() 99 …ncy_highbit1 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[1], relaxed); in exclaves_memory_report_accounting() 100 …ncy_highbit2 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[2], relaxed); in exclaves_memory_report_accounting() 101 …ncy_highbit3 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[3], relaxed); in exclaves_memory_report_accounting() 102 …ncy_highbit4 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[4], relaxed); in exclaves_memory_report_accounting() 103 …ncy_highbit5 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[5], relaxed); in exclaves_memory_report_accounting() [all …]
|
| H A D | counter_common.c | 67 uint64_t current_value = os_atomic_load_wide(zpercpu_get(*counter), relaxed); in scalable_counter_static_init() 73 os_atomic_store_wide(zpercpu_get(*counter), current_value, relaxed); in scalable_counter_static_init() 88 os_atomic_store_wide(counter, 0, relaxed); in counter_alloc() 109 os_atomic_add(counter, amount, relaxed); in counter_add() 116 os_atomic_inc(counter, relaxed); in counter_inc() 123 os_atomic_dec(counter, relaxed); in counter_dec() 151 return os_atomic_load_wide(counter, relaxed); in counter_load() 160 value += os_atomic_load_wide(it, relaxed); in counter_load()
|
| H A D | mpsc_queue.c | 48 mpsc_queue_chain_t head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch() 50 os_atomic_store(&last->mpqc_next, head, relaxed); in mpsc_queue_restore_batch() 54 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch() 58 os_atomic_store(&last->mpqc_next, head, relaxed); in mpsc_queue_restore_batch() 61 os_atomic_store(&q->mpqh_head.mpqc_next, first, relaxed); in mpsc_queue_restore_batch() 72 tail = os_atomic_load(&q->mpqh_tail, relaxed); in mpsc_queue_dequeue_batch() 78 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_dequeue_batch() 82 os_atomic_store(&q->mpqh_head.mpqc_next, NULL, relaxed); in mpsc_queue_dequeue_batch() 111 elm = os_atomic_load(&cur->mpqc_next, relaxed); in mpsc_queue_batch_next() 291 os_atomic_andnot(&dq->mpd_state, MPSC_QUEUE_STATE_WAKEUP, relaxed); in _mpsc_daemon_queue_drain() [all …]
|
| H A D | exclaves_boot.c | 147 EXCLAVES_BS_NOT_STARTED, relaxed); in exclaves_check_sk() 150 EXCLAVES_BS_NOT_SUPPORTED, relaxed); in exclaves_check_sk() 160 while (os_atomic_load(&exclaves_boot_status, relaxed) < status) { in exclaves_boot_status_wait() 176 assert3u(status, >, os_atomic_load(&exclaves_boot_status, relaxed)); in exclaves_boot_status_set() 206 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_exclavecore() 258 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_exclavekit() 413 exclaves_boot_status_t boot_status = os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_get_boot_status_string() 438 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_get_boot_stage() 465 status = os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_supported()
|
| H A D | sched_average.c | 190 load_now[TH_BUCKET_RUN] = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in compute_sched_load() 191 load_now[TH_BUCKET_FIXPRI] = os_atomic_load(&sched_run_buckets[TH_BUCKET_FIXPRI], relaxed); in compute_sched_load() 192 load_now[TH_BUCKET_SHARE_FG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_FG], relaxed); in compute_sched_load() 193 load_now[TH_BUCKET_SHARE_DF] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_DF], relaxed); in compute_sched_load() 194 load_now[TH_BUCKET_SHARE_UT] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_UT], relaxed); in compute_sched_load() 195 load_now[TH_BUCKET_SHARE_BG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_BG], relaxed); in compute_sched_load() 288 uint32_t nthreads = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed) - 1; in compute_averages()
|
| H A D | ast.h | 188 #define thread_ast_set(act, reason) ((void)os_atomic_or(&(act)->ast, (reason), relaxed)) 189 #define thread_ast_clear(act, reason) ((void)os_atomic_andnot(&(act)->ast, (reason), relaxed)) 190 #define thread_ast_peek(act, reason) (os_atomic_load(&(act)->ast, relaxed) & (reason)) 191 #define thread_ast_get(act) os_atomic_load(&(act)->ast, relaxed)
|
| H A D | lock_ptr.c | 62 hw_lck_ptr_t tmp = os_atomic_load(lck, relaxed); in __hw_lck_ptr_invalid_panic() 113 hw_lck_ptr_t tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_destroy() 128 os_atomic_store(lck, tmp, relaxed); in hw_lck_ptr_destroy() 134 return os_atomic_load(lck, relaxed).lck_ptr_locked; in hw_lck_ptr_held() 147 tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_timeout_panic() 196 os_atomic_store(&pnode->lsm_next, txn.txn_slot, relaxed); in hw_lck_ptr_contended() 235 os_atomic_store(&nnode->lsm_ready, 1, relaxed); in hw_lck_ptr_contended() 271 tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_lock_fastpath()
|
| H A D | lock_group.c | 276 os_atomic_inc(cnt, relaxed); in lck_grp_reference() 287 os_atomic_dec(cnt, relaxed); in lck_grp_deallocate() 308 os_atomic_or(&lck_debug_state.lds_value, bit, relaxed); in lck_grp_enable_feature() 325 os_atomic_andnot(&lck_debug_state.lds_value, bit, relaxed); in lck_grp_disable_feature() 354 needed = os_atomic_load(&lck_grp_table.cidt_count, relaxed); in host_lockgroup_info() 445 os_atomic_or(&attr->lck_attr_val, LCK_ATTR_DEBUG, relaxed); in lck_attr_setdebug() 451 os_atomic_andnot(&attr->lck_attr_val, LCK_ATTR_DEBUG, relaxed); in lck_attr_cleardebug() 457 os_atomic_or(&attr->lck_attr_val, LCK_ATTR_RW_SHARED_PRIORITY, relaxed); in lck_attr_rw_shared_priority() 496 __unused uint64_t val = os_atomic_inc_orig(&stat->lgs_count, relaxed); in lck_grp_stat_inc() 509 __unused uint64_t val = os_atomic_add_orig(&stat->lgs_count, time, relaxed); in lck_grp_inc_time_stats()
|
| H A D | smr.c | 651 s_wr_seq = os_atomic_load(&smr->smr_clock.s_wr_seq, relaxed); in __smr_enter() 657 os_atomic_store(&pcpu->c_rd_seq, s_wr_seq | sleepable, relaxed); in __smr_enter() 724 os_atomic_store(&smrw->sect_waiter, NULL, relaxed); in __smr_wake_oncore_sleepers() 776 os_atomic_store(&pcpu->stall_rd_seq, t->smrt_seq, relaxed); in smr_mark_active_trackers_stalled() 899 os_atomic_rmw_loop(&smr->smr_clock.s_rd_seq, o_seq, rd_seq, relaxed, { in __smr_rd_advance() 1015 if (lock_cmpxchg(&smrw->sect_waiter, NULL, self, relaxed)) { in __smr_wait_for_oncore() 1032 return os_atomic_load(&pcpu->c_rd_seq, relaxed); in __smr_wait_for_oncore() 1067 clk.s_wr_seq, goal, &clk.s_wr_seq, relaxed)) { in __smr_scan() 1104 smr_seq_t seq = os_atomic_load(&pcpu->c_rd_seq, relaxed); in __smr_scan() 1150 smr_seq_t seq = os_atomic_load(&pcpu->stall_rd_seq, relaxed); in __smr_scan() [all …]
|
| H A D | socd_client.c | 61 bool already_initialized = os_atomic_load(&socd_client_trace_available, relaxed); in socd_client_init() 94 long available = os_atomic_load(&socd_client_trace_available, relaxed); in socd_client_set_primary_kernelcache_uuid() 137 trace_idx = os_atomic_inc_orig(&socd_client_state.trace_idx, relaxed); in socd_client_trace() 161 os_atomic_store(&socd_client_trace_has_sticky_events, true, relaxed); in socd_client_trace()
|
| H A D | restartable.c | 355 (uint8_t)TRR_FAULT_PENDING, relaxed); in thread_reset_pcs_will_fault() 371 state.trr_value, relaxed); in thread_reset_pcs_done_faulting() 390 trrs.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_reset_pcs_ack_IPI() 398 trrs.trr_value, relaxed); in thread_reset_pcs_ack_IPI() 408 state.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_rr_wait_if_needed() 471 nstate, relaxed)) { in thread_rr_wait_if_needed() 480 state.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_rr_wait_if_needed() 518 state.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_reset_pcs_ast()
|
| /xnu-11417.121.6/libsyscall/mach/ ! |
| H A D | vm_reclaim.c | 221 tail = os_atomic_load_wide(&indices->tail, relaxed); in mach_vm_reclaim_try_enter() 222 head = os_atomic_load_wide(&indices->head, relaxed); in mach_vm_reclaim_try_enter() 236 os_atomic_inc(&indices->tail, relaxed); in mach_vm_reclaim_try_enter() 239 head = os_atomic_load_wide(&indices->head, relaxed); in mach_vm_reclaim_try_enter() 249 original_tail = os_atomic_load_wide(&indices->tail, relaxed); in mach_vm_reclaim_try_enter() 253 os_atomic_store_wide(&indices->tail, requested_id, relaxed); in mach_vm_reclaim_try_enter() 255 busy = os_atomic_load_wide(&indices->busy, relaxed); in mach_vm_reclaim_try_enter() 258 os_atomic_store_wide(&indices->tail, original_tail, relaxed); in mach_vm_reclaim_try_enter() 272 os_atomic_store_wide(&indices->tail, original_tail, relaxed); in mach_vm_reclaim_try_enter() 313 head = os_atomic_load_wide(&indices->head, relaxed); in mach_vm_reclaim_try_cancel() [all …]
|
| /xnu-11417.121.6/san/memory/ ! |
| H A D | ubsan_log.c | 77 os_atomic_rmw_loop(&ubsan_log_next, i, n, relaxed, { in ubsan_log_append() 90 os_atomic_rmw_loop(&ubsan_log_head, e, n, relaxed, { in ubsan_log_append() 119 head = os_atomic_load(&ubsan_log_head, relaxed); 121 tail = os_atomic_load(&ubsan_log_tail, relaxed); 157 head = os_atomic_load(&ubsan_log_head, relaxed); 159 tail = os_atomic_load(&ubsan_log_tail, relaxed); 172 os_atomic_store(&ubsan_log_tail, head, relaxed);
|
| /xnu-11417.121.6/osfmk/bank/ ! |
| H A D | bank_internal.h | 120 (os_atomic_inc_orig(&(elem)->bt_made, relaxed)) 123 (os_atomic_dec_orig(&(elem)->bt_made, relaxed)) 126 (os_atomic_sub_orig(&(elem)->bt_made, (num), relaxed)) 171 (os_atomic_inc_orig(&(elem)->ba_made, relaxed)) 174 (os_atomic_dec_orig(&(elem)->ba_made, relaxed)) 177 (os_atomic_sub_orig(&(elem)->ba_made, (num), relaxed))
|
| /xnu-11417.121.6/osfmk/arm/ ! |
| H A D | counter.c | 40 os_atomic_add(zpercpu_get(*counter), amount, relaxed); in counter_add() 47 os_atomic_inc(zpercpu_get(*counter), relaxed); in counter_inc() 54 os_atomic_dec(zpercpu_get(*counter), relaxed); in counter_dec()
|
| /xnu-11417.121.6/san/coverage/ ! |
| H A D | kcov.c | 116 os_atomic_add(&kcov_enabled, 1, relaxed); in kcov_enable() 122 os_atomic_sub(&kcov_enabled, 1, relaxed); in kcov_disable() 155 os_atomic_store(&kcov_enabled, 0, relaxed); in kcov_panic_disable() 189 if (__probable(os_atomic_load(&kcov_enabled, relaxed) == 0)) { in trace_pc_guard()
|
| /xnu-11417.121.6/bsd/net/ ! |
| H A D | dlil_ctl.c | 242 idx = os_atomic_inc_orig(cnt, relaxed) % IF_REF_TRACE_HIST_SIZE; in dlil_if_trace() 297 os_atomic_add(&ifp->if_data.ifi_ipackets, s->packets_in, relaxed); in dlil_input_stats_sync() 301 os_atomic_add(&ifp->if_data.ifi_ibytes, s->bytes_in, relaxed); in dlil_input_stats_sync() 305 os_atomic_add(&ifp->if_data.ifi_ierrors, s->errors_in, relaxed); in dlil_input_stats_sync() 310 os_atomic_add(&ifp->if_data.ifi_opackets, s->packets_out, relaxed); in dlil_input_stats_sync() 314 os_atomic_add(&ifp->if_data.ifi_obytes, s->bytes_out, relaxed); in dlil_input_stats_sync() 318 os_atomic_add(&ifp->if_data.ifi_oerrors, s->errors_out, relaxed); in dlil_input_stats_sync() 323 os_atomic_add(&ifp->if_data.ifi_collisions, s->collisions, relaxed); in dlil_input_stats_sync() 327 os_atomic_add(&ifp->if_data.ifi_iqdrops, s->dropped, relaxed); in dlil_input_stats_sync()
|
| /xnu-11417.121.6/libkern/os/ ! |
| H A D | refcnt.c | 76 if (os_atomic_load(rc, relaxed) >= OS_REFCNT_MAX_COUNT) { in os_ref_panic_retain() 362 os_atomic_rmw_loop(rc, cur, next, relaxed, { in __os_ref_retain_try() 624 os_atomic_rmw_loop(rc, cur, next, relaxed, { in os_ref_retain_try_mask_internal() 783 n = os_atomic_load_wide(zpercpu_get_cpu(rc, 0), relaxed); 788 n |= os_atomic_load_wide(zpercpu_get_cpu(rc, cpu), relaxed); 804 v = os_atomic_load_wide(zpercpu_get_cpu(os_pcpu_get(ref), 0), relaxed); in os_pcpu_ref_count() 820 v = os_atomic_load_wide(rcp, relaxed); in __os_pcpu_ref_delta() 828 return os_atomic_add_orig(rcp, OS_PCPU_REF_INC, relaxed); in __os_pcpu_ref_delta() 840 OS_PCPU_REF_INC, relaxed); in __os_pcpu_ref_retain_slow() 845 OS_PCPU_REF_INC, relaxed); in __os_pcpu_ref_retain_slow() [all …]
|
| /xnu-11417.121.6/bsd/skywalk/nexus/flowswitch/flow/ ! |
| H A D | flow_route.c | 459 os_atomic_andnot(&fr->fr_flags, FLOWRTF_DELETED, relaxed); in flow_route_configure() 507 os_atomic_or(&fr->fr_flags, FLOWRTF_GATEWAY, relaxed); in flow_route_configure() 512 os_atomic_or(&fr->fr_flags, FLOWRTF_ONLINK, relaxed); in flow_route_configure() 525 os_atomic_or(&fr->fr_flags, FLOWRTF_STABLE_ADDR, relaxed); in flow_route_configure() 527 os_atomic_andnot(&fr->fr_flags, FLOWRTF_STABLE_ADDR, relaxed); in flow_route_configure() 605 os_atomic_inc(&fr->fr_want_configure, relaxed); in flow_route_find() 688 os_atomic_inc(&fr->fr_want_configure, relaxed); in flow_route_find() 719 os_atomic_or(&fr->fr_flags, FLOWRTF_ATTACHED, relaxed); in flow_route_find() 892 os_atomic_andnot(&fr->fr_flags, FLOWRTF_ATTACHED, relaxed); in flow_route_bucket_purge_common() 1068 os_atomic_inc(&fr->fr_want_configure, relaxed); in flow_route_ev_callback() [all …]
|
| /xnu-11417.121.6/osfmk/arm64/sptm/pmap/ ! |
| H A D | pmap_data.h | 293 assertf(os_atomic_load(&pv_head_table[index], relaxed) & PVH_LOCK_FLAGS, in pvh_assert_locked() 295 &pv_head_table[index], (void*)(os_atomic_load(&pv_head_table[index], relaxed)), index); in pvh_assert_locked() 346 locked_pvh.pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_lock() 380 …const locked_pvh_t locked_pvh = {.pvh = os_atomic_load(&pv_head_table[index], relaxed), .pai = ind… in pvh_lock_nopreempt() 404 locked_pvh.pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_try_lock() 451 const uintptr_t old_pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_lock_enter_sleep_mode() 454 os_atomic_store(&pv_head_table[index], old_pvh | PVH_FLAG_SLEEP, relaxed); in pvh_lock_enter_sleep_mode() 501 const uintptr_t old_pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_unlock() 518 (locked_pvh->pvh & ~PVH_FLAG_SLEEP) | PVH_FLAG_LOCK, relaxed); in pvh_unlock() 522 (locked_pvh->pvh & ~PVH_FLAG_SLEEP) | PVH_FLAG_LOCK, relaxed); in pvh_unlock() [all …]
|