| /xnu-8019.80.24/tools/cocci/ |
| H A D | OSAtomic_rewrite.cocci | 11 + os_atomic_inc_orig(E, relaxed) 14 + os_atomic_inc_orig(E, relaxed) 17 + os_atomic_inc_orig(E, relaxed) 20 + os_atomic_inc_orig(E, relaxed) 23 + os_atomic_inc_orig(E, relaxed) 26 + os_atomic_inc_orig(E, relaxed) 29 + os_atomic_inc_orig(E, relaxed) 32 + os_atomic_inc_orig(E, relaxed) 35 + os_atomic_inc_orig(E, relaxed) 38 + os_atomic_inc_orig(E, relaxed) [all …]
|
| H A D | mcache_atomic_rewrite.cocci | 11 + os_atomic_inc_orig(E, relaxed) 14 + os_atomic_inc(E, relaxed) 17 + os_atomic_inc_orig(E, relaxed) 20 + os_atomic_inc(E, relaxed) 23 + os_atomic_inc_orig(E, relaxed) 26 + os_atomic_inc(E, relaxed) 29 + os_atomic_dec_orig(E, relaxed) 32 + os_atomic_dec(E, relaxed) 35 + os_atomic_dec_orig(E, relaxed) 38 + os_atomic_dec(E, relaxed) [all …]
|
| H A D | hw_atomic_rewrite.cocci | 11 + os_atomic_dec_orig(E, relaxed) 14 + os_atomic_dec(E, relaxed) 17 + os_atomic_sub_orig(E, F, relaxed) 20 + os_atomic_sub(E, F, relaxed) 23 + os_atomic_inc_orig(E, relaxed) 26 + os_atomic_inc(E, relaxed) 29 + os_atomic_add_orig(E, F, relaxed) 32 + os_atomic_add(E, F, relaxed) 39 + os_atomic_inc_orig(E, relaxed) 42 + os_atomic_inc(E, relaxed) [all …]
|
| /xnu-8019.80.24/tests/ |
| H A D | os_atomic.cpp | 16 T_ASSERT_EQ(os_atomic_inc_orig(&i, relaxed), 0, "atomic inc"); 17 T_ASSERT_EQ(os_atomic_cmpxchg(&i, 1, 0, relaxed), true, "os_atomic_cmpxchg"); 18 os_atomic_rmw_loop(&i, a, b, relaxed, { 22 T_ASSERT_EQ(os_atomic_inc_orig(&old_i, relaxed), 0, "atomic inc"); 23 T_ASSERT_EQ(os_atomic_cmpxchg(&old_i, 1, 0, relaxed), true, "os_atomic_cmpxchg"); 24 os_atomic_rmw_loop(&old_i, a, b, relaxed, { 28 T_ASSERT_EQ(os_atomic_inc_orig(&v_i, relaxed), 0, "atomic inc"); 29 T_ASSERT_EQ(os_atomic_cmpxchg(&v_i, 1, 0, relaxed), true, "os_atomic_cmpxchg"); 30 os_atomic_rmw_loop(&v_i, a, b, relaxed, {
|
| /xnu-8019.80.24/libkern/gen/ |
| H A D | OSAtomicOperations.c | 105 return os_atomic_add_orig(address, (SInt8)amount, relaxed); in OSAddAtomic8() 111 return os_atomic_add_orig(address, (SInt16)amount, relaxed); in OSAddAtomic16() 119 return os_atomic_add_orig(address, amount, relaxed); in OSAddAtomic() 129 return os_atomic_add_orig(aligned_address, amount, relaxed); in OSAddAtomic64() 136 return os_atomic_add_orig(address, theAmount, relaxed); in OSAddAtomicLong() 143 return os_atomic_inc_orig(value, relaxed); in OSIncrementAtomic() 150 return os_atomic_dec_orig(value, relaxed); in OSDecrementAtomic() 157 return os_atomic_and_orig(value, mask, relaxed); in OSBitAndAtomic() 164 return os_atomic_or_orig(value, mask, relaxed); in OSBitOrAtomic() 171 return os_atomic_xor_orig(value, mask, relaxed); in OSBitXorAtomic() [all …]
|
| /xnu-8019.80.24/osfmk/kern/ |
| H A D | counter_common.c | 67 uint64_t current_value = os_atomic_load_wide(zpercpu_get(*counter), relaxed); in scalable_counter_static_init() 73 os_atomic_store_wide(zpercpu_get(*counter), current_value, relaxed); in scalable_counter_static_init() 88 os_atomic_store_wide(counter, 0, relaxed); in counter_alloc() 109 os_atomic_add(counter, amount, relaxed); in counter_add() 116 os_atomic_inc(counter, relaxed); in counter_inc() 123 os_atomic_dec(counter, relaxed); in counter_dec() 151 return os_atomic_load_wide(counter, relaxed); in counter_load() 160 value += os_atomic_load_wide(it, relaxed); in counter_load()
|
| H A D | mpsc_queue.c | 47 mpsc_queue_chain_t head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch() 49 os_atomic_store(&last->mpqc_next, head, relaxed); in mpsc_queue_restore_batch() 53 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch() 57 os_atomic_store(&last->mpqc_next, head, relaxed); in mpsc_queue_restore_batch() 60 os_atomic_store(&q->mpqh_head.mpqc_next, first, relaxed); in mpsc_queue_restore_batch() 71 tail = os_atomic_load(&q->mpqh_tail, relaxed); in mpsc_queue_dequeue_batch() 77 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_dequeue_batch() 81 os_atomic_store(&q->mpqh_head.mpqc_next, NULL, relaxed); in mpsc_queue_dequeue_batch() 110 elm = os_atomic_load(&cur->mpqc_next, relaxed); in mpsc_queue_batch_next() 274 os_atomic_andnot(&dq->mpd_state, MPSC_QUEUE_STATE_WAKEUP, relaxed); in _mpsc_daemon_queue_drain() [all …]
|
| H A D | sched_average.c | 193 load_now[TH_BUCKET_RUN] = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in compute_sched_load() 194 load_now[TH_BUCKET_FIXPRI] = os_atomic_load(&sched_run_buckets[TH_BUCKET_FIXPRI], relaxed); in compute_sched_load() 195 load_now[TH_BUCKET_SHARE_FG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_FG], relaxed); in compute_sched_load() 196 load_now[TH_BUCKET_SHARE_DF] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_DF], relaxed); in compute_sched_load() 197 load_now[TH_BUCKET_SHARE_UT] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_UT], relaxed); in compute_sched_load() 198 load_now[TH_BUCKET_SHARE_BG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_BG], relaxed); in compute_sched_load() 291 uint32_t nthreads = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed) - 1; in compute_averages()
|
| H A D | machine.c | 147 os_atomic_inc(&processor_avail_count, relaxed); in processor_up() 149 os_atomic_inc(&processor_avail_count_user, relaxed); in processor_up() 153 os_atomic_inc(&primary_processor_avail_count, relaxed); in processor_up() 155 os_atomic_inc(&primary_processor_avail_count_user, relaxed); in processor_up() 326 os_atomic_dec(&processor_avail_count, relaxed); in processor_doshutdown() 328 os_atomic_dec(&processor_avail_count_user, relaxed); in processor_doshutdown() 331 os_atomic_dec(&primary_processor_avail_count, relaxed); in processor_doshutdown() 333 os_atomic_dec(&primary_processor_avail_count_user, relaxed); in processor_doshutdown() 525 uint32_t const report_phy_read_delay = os_atomic_load(&report_phy_read_delay_to, relaxed); in ml_io_read() 526 uint32_t const trace_phy_read_delay = os_atomic_load(&trace_phy_read_delay_to, relaxed); in ml_io_read() [all …]
|
| H A D | lock_ticket.c | 126 }), relaxed); in hw_lck_ticket_init() 147 }), relaxed); in hw_lck_ticket_init_locked() 176 os_atomic_store(&lck->lck_valid, (uint8_t)0, relaxed); in hw_lck_ticket_destroy_internal() 179 tmp.lck_value = os_atomic_load(&lck->lck_value, relaxed); in hw_lck_ticket_destroy_internal() 200 os_atomic_store(&lck->lck_value, 0U, relaxed); in hw_lck_ticket_destroy_internal() 231 tmp.tcurnext = os_atomic_load(&lck->tcurnext, relaxed); in hw_lck_ticket_held() 257 os_atomic_store(&tlock->lck_owner, (uintptr_t)cthread, relaxed); in tlock_mark_owned() 271 tmp.tcurnext = os_atomic_load(&lck->tcurnext, relaxed); in hw_lck_ticket_timeout_panic() 297 tmp.tcurnext = os_atomic_load(&lck->tcurnext, relaxed); in lck_ticket_timeout_panic() 667 os_atomic_andnot(&lck->lck_value, tmp.lck_value, relaxed); in hw_lck_ticket_invalidate() [all …]
|
| H A D | sched_clutch.c | 872 return (int)os_atomic_load(&sched_clutch_global_bucket_load[bucket], relaxed); in sched_clutch_global_bucket_load_get() 1099 os_atomic_store(&clutch_bucket_group->scbg_timeshare_tick, 0, relaxed); in sched_clutch_bucket_group_init() 1100 os_atomic_store(&clutch_bucket_group->scbg_pri_shift, INT8_MAX, relaxed); in sched_clutch_bucket_group_init() 1101 os_atomic_store(&clutch_bucket_group->scbg_preferred_cluster, pset0.pset_cluster_id, relaxed); in sched_clutch_bucket_group_init() 1108 …pu_data.scbcd_cpu_blocked, (clutch_cpu_data_t)sched_clutch_bucket_group_adjust_threshold, relaxed); in sched_clutch_bucket_group_init() 1135 os_atomic_store(&clutch->sc_thr_count, 0, relaxed); in sched_clutch_init_with_thread_group() 1144 os_atomic_store(&clutch->sc_tg_priority, 0, relaxed); in sched_clutch_init_with_thread_group() 1156 assert(os_atomic_load(&clutch->sc_thr_count, relaxed) == 0); in sched_clutch_destroy() 1177 return os_atomic_load(&clutch_bucket_group->scbg_preferred_cluster, relaxed); in sched_edge_clutch_bucket_group_preferred_cluster() 1246 … os_atomic_inc(&root_clutch->scr_cumulative_run_count[TH_BUCKET_FIXPRI], relaxed); OS_FALLTHROUGH; in sched_edge_cluster_cumulative_count_incr() [all …]
|
| H A D | ast.h | 196 #define thread_ast_set(act, reason) ((void)os_atomic_or(&(act)->ast, (reason), relaxed)) 197 #define thread_ast_clear(act, reason) ((void)os_atomic_andnot(&(act)->ast, (reason), relaxed)) 198 #define thread_ast_peek(act, reason) (os_atomic_load(&(act)->ast, relaxed) & (reason))
|
| H A D | test_lock.c | 61 assert(os_atomic_load(lck, relaxed) == 0); in hw_lock_allow_invalid_test() 68 assert(os_atomic_load(lck, relaxed) == 0x101); in hw_lock_allow_invalid_test() 77 assert(os_atomic_load(lck, relaxed) == 0x100); in hw_lock_allow_invalid_test() 103 tmp.lck_value = os_atomic_load(&lck->lck_value, relaxed); in hw_lck_ticket_test_wait_for_delta()
|
| H A D | cpu_quiesce.c | 200 assert((os_atomic_load(&cpu_quiescing_checkin_state, relaxed) & in cpu_quiescent_counter_join() 225 assert((os_atomic_load(&cpu_quiescing_checkin_state, relaxed) & in cpu_quiescent_counter_ast() 340 checkin_mask_t state = os_atomic_load(&cpu_quiescing_checkin_state, relaxed); in cpu_quiescent_counter_checkin() 378 checkin_mask_t state = os_atomic_load(&cpu_quiescing_checkin_state, relaxed); in cpu_quiescent_counter_assert_ast()
|
| H A D | priority.c | 752 os_atomic_inc(&sched_run_buckets[bucket], relaxed); in sched_incr_bucket() 761 assert(os_atomic_load(&sched_run_buckets[bucket], relaxed) > 0); in sched_decr_bucket() 763 os_atomic_dec(&sched_run_buckets[bucket], relaxed); in sched_decr_bucket() 772 os_atomic_add(&sched_run_buckets[bucket], run_weight, relaxed); in sched_add_bucket() 781 assert(os_atomic_load(&sched_run_buckets[bucket], relaxed) > 0); in sched_sub_bucket() 783 os_atomic_sub(&sched_run_buckets[bucket], run_weight, relaxed); in sched_sub_bucket() 791 uint32_t new_count = os_atomic_inc(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in sched_run_incr() 805 uint32_t new_count = os_atomic_dec(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in sched_run_decr() 818 uint32_t new_count = os_atomic_add(&sched_run_buckets[TH_BUCKET_RUN], run_weight, relaxed); in sched_smt_run_incr() 834 uint32_t new_count = os_atomic_sub(&sched_run_buckets[TH_BUCKET_RUN], run_weight, relaxed); in sched_smt_run_decr()
|
| H A D | hazard.c | 188 os_atomic_store(&guard->hg_val, NULL, relaxed); in hazard_guard_dismiss() 207 os_atomic_store(&guard->hg_val, value, relaxed); in hazard_guard_set() 473 p = os_atomic_load(&hga->hga_array[i].hg_val, relaxed); in hazard_scan_and_reclaim() 567 os_atomic_inc(&hazard_test_outstanding, relaxed); in hazard_test_value_alloc() 575 os_atomic_dec(&hazard_test_outstanding, relaxed); in hazard_test_value_release() 634 start = os_atomic_load(&hazard_test_outstanding, relaxed); in hazard_basic_test() 644 end = os_atomic_load(&hazard_test_outstanding, relaxed); in hazard_basic_test()
|
| /xnu-8019.80.24/san/memory/ |
| H A D | ubsan_log.c | 77 os_atomic_rmw_loop(&ubsan_log_next, i, n, relaxed, { in ubsan_log_append() 90 os_atomic_rmw_loop(&ubsan_log_head, e, n, relaxed, { in ubsan_log_append() 119 head = os_atomic_load(&ubsan_log_head, relaxed); 121 tail = os_atomic_load(&ubsan_log_tail, relaxed); 157 head = os_atomic_load(&ubsan_log_head, relaxed); 159 tail = os_atomic_load(&ubsan_log_tail, relaxed); 172 os_atomic_store(&ubsan_log_tail, head, relaxed);
|
| /xnu-8019.80.24/osfmk/bank/ |
| H A D | bank_internal.h | 116 (os_atomic_inc_orig(&(elem)->bt_made, relaxed)) 119 (os_atomic_dec_orig(&(elem)->bt_made, relaxed)) 122 (os_atomic_sub_orig(&(elem)->bt_made, (num), relaxed)) 166 (os_atomic_inc_orig(&(elem)->ba_made, relaxed)) 169 (os_atomic_dec_orig(&(elem)->ba_made, relaxed)) 172 (os_atomic_sub_orig(&(elem)->ba_made, (num), relaxed))
|
| /xnu-8019.80.24/osfmk/arm/ |
| H A D | cpu_common.c | 228 os_atomic_andnot(&cpu_data_ptr->cpu_signal, SIGPxcall, relaxed); in cpu_handle_xcall() 239 os_atomic_andnot(&cpu_data_ptr->cpu_signal, SIGPxcallImm, relaxed); in cpu_handle_xcall() 293 if (os_atomic_sub(synch, (!self_xcall) ? failsig + 1 : failsig, relaxed) == 0) { in cpu_broadcast_xcall_internal() 329 if (os_atomic_dec(&data->sync, relaxed) == 0) { in cpu_broadcast_xcall_simple_cbk() 575 cpu_signal = os_atomic_or(&cpu_data_ptr->cpu_signal, 0, relaxed); in cpu_signal_handler_internal() 578 os_atomic_or(&cpu_data_ptr->cpu_signal, SIGPdisabled, relaxed); in cpu_signal_handler_internal() 580 os_atomic_andnot(&cpu_data_ptr->cpu_signal, SIGPdisabled, relaxed); in cpu_signal_handler_internal() 585 os_atomic_andnot(&cpu_data_ptr->cpu_signal, SIGPdec, relaxed); in cpu_signal_handler_internal() 592 os_atomic_andnot(&cpu_data_ptr->cpu_signal, SIGPkppet, relaxed); in cpu_signal_handler_internal() 603 os_atomic_andnot(&cpu_data_ptr->cpu_signal, SIGPast, relaxed); in cpu_signal_handler_internal() [all …]
|
| H A D | counter.c | 40 os_atomic_add(zpercpu_get(*counter), amount, relaxed); in counter_add() 47 os_atomic_inc(zpercpu_get(*counter), relaxed); in counter_inc() 54 os_atomic_dec(zpercpu_get(*counter), relaxed); in counter_dec()
|
| H A D | machine_routines.c | 157 os_atomic_store(&LockTimeOut, (uint32_t)lto_abstime, relaxed); in ml_init_lock_timeout() 159 lto_abstime = os_atomic_load(&LockTimeOut, relaxed); in ml_init_lock_timeout() 160 os_atomic_store(&TLockTimeOut, (uint32_t)lto_abstime, relaxed); in ml_init_lock_timeout() 164 os_atomic_store(&LockTimeOutUsec, (uint32_t)(lto_timeout_ns / NSEC_PER_USEC), relaxed); in ml_init_lock_timeout() 173 os_atomic_store(&MutexSpin, (uint32_t)mtx_abstime, relaxed); in ml_init_lock_timeout() 175 mtx_abstime = os_atomic_load(&MutexSpin, relaxed); in ml_init_lock_timeout() 177 low_MutexSpin = os_atomic_load(&MutexSpin, relaxed); in ml_init_lock_timeout() 208 os_atomic_inc(&cluster_type_num_active_cpus[cpu->cluster_type], relaxed); in ml_cpu_up() 210 os_atomic_inc(&machine_info.physical_cpu, relaxed); in ml_cpu_up() 211 os_atomic_inc(&machine_info.logical_cpu, relaxed); in ml_cpu_up() [all …]
|
| /xnu-8019.80.24/san/coverage/ |
| H A D | kcov.c | 116 os_atomic_add(&kcov_enabled, 1, relaxed); in kcov_enable() 122 os_atomic_sub(&kcov_enabled, 1, relaxed); in kcov_disable() 155 os_atomic_store(&kcov_enabled, 0, relaxed); in kcov_panic_disable() 189 if (__probable(os_atomic_load(&kcov_enabled, relaxed) == 0)) { in trace_pc_guard()
|
| H A D | kcov_ksancov.c | 114 if (os_atomic_load(&dev->trace->kt_head, relaxed) >= dev->maxpcs) { in trace_pc_guard_pcs() 118 uint32_t idx = os_atomic_inc_orig(&dev->trace->kt_head, relaxed); in trace_pc_guard_pcs() 131 if (os_atomic_load(&dev->trace->kt_head, relaxed) >= dev->maxpcs) { in trace_pc_guard_pcs_stk() 135 uint32_t idx = os_atomic_inc_orig(&dev->trace->kt_head, relaxed); in trace_pc_guard_pcs_stk() 189 if (os_atomic_load(&dev->hdr->kh_enabled, relaxed) == 0) { in kcov_ksancov_trace_pc() 527 os_atomic_store(&data->ktd_device, d, relaxed); in ksancov_attach() 528 os_atomic_add(&ksancov_enabled, 1, relaxed); in ksancov_attach() 555 os_atomic_store(&data->ktd_device, NULL, relaxed); in ksancov_detach() 564 os_atomic_sub(&ksancov_enabled, 1, relaxed); in ksancov_detach() 594 os_atomic_store(&d->hdr->kh_enabled, 0, relaxed); /* stop tracing */ in ksancov_close()
|
| /xnu-8019.80.24/osfmk/vm/ |
| H A D | analytics.c | 100 e->over_global_limit = os_atomic_load_wide(&vm_add_wire_count_over_global_limit, relaxed); in report_mlock_failures() 101 e->over_user_limit = os_atomic_load_wide(&vm_add_wire_count_over_user_limit, relaxed); in report_mlock_failures() 103 os_atomic_store_wide(&vm_add_wire_count_over_global_limit, 0, relaxed); in report_mlock_failures() 104 os_atomic_store_wide(&vm_add_wire_count_over_user_limit, 0, relaxed); in report_mlock_failures()
|
| /xnu-8019.80.24/libkern/os/ |
| H A D | log_mem.c | 179 os_atomic_inc(&lm->lm_cnt_allocations, relaxed); in logmem_alloc_impl() 182 os_atomic_inc(&lm->lm_cnt_failed_lmoff, relaxed); in logmem_alloc_impl() 187 os_atomic_inc(&lm->lm_cnt_failed_size, relaxed); in logmem_alloc_impl() 197 os_atomic_inc(&lm->lm_cnt_failed_full, relaxed); in logmem_alloc_impl() 202 os_atomic_sub(&lm->lm_cnt_free, (uint32_t)*amount, relaxed); in logmem_alloc_impl() 237 os_atomic_add(&lm->lm_cnt_free, (uint32_t)amount, relaxed); in logmem_free_impl()
|