| /xnu-10063.141.1/tests/ |
| H A D | os_atomic.cpp | 18 os_atomic_rmw_loop(&i, a, b, relaxed, { 24 os_atomic_rmw_loop(&old_i, a, b, relaxed, { 30 os_atomic_rmw_loop(&v_i, a, b, relaxed, {
|
| /xnu-10063.141.1/osfmk/kern/ |
| H A D | kern_apfs_reflock.c | 137 …os_atomic_rmw_loop(&reflock->kern_apfs_rl_data.cond64_data, old_reflock.kern_apfs_rl_data.cond64_d… in kern_apfs_reflock_free() 175 …os_atomic_rmw_loop(&reflock->kern_apfs_rl_data.cond64_data, old_reflock.kern_apfs_rl_data.cond64_d… in kern_apfs_reflock_try_get_ref() 244 …os_atomic_rmw_loop(&reflock->kern_apfs_rl_data.cond64_data, old_reflock.kern_apfs_rl_data.cond64_d… in kern_apfs_reflock_try_put_ref() 316 …os_atomic_rmw_loop(&reflock->kern_apfs_rl_data.cond64_data, old_reflock.kern_apfs_rl_data.cond64_d… in kern_apfs_reflock_try_lock() 371 …os_atomic_rmw_loop(&reflock->kern_apfs_rl_data.cond64_data, old_reflock.kern_apfs_rl_data.cond64_d… in kern_apfs_reflock_wait_for_unlock() 434 …os_atomic_rmw_loop(&reflock->kern_apfs_rl_data.cond64_data, old_reflock.kern_apfs_rl_data.cond64_d… in kern_apfs_reflock_wait_for_unlock() 457 …os_atomic_rmw_loop(&reflock->kern_apfs_rl_data.cond64_data, old_reflock.kern_apfs_rl_data.cond64_d… in kern_apfs_reflock_unlock()
|
| H A D | lock_ticket.c | 729 os_atomic_rmw_loop(&lck->tcurnext, olck.tcurnext, nlck.tcurnext, acquire, { in hw_lck_ticket_lock_try_internal() 769 os_atomic_rmw_loop(&tlock->tu.tcurnext, olck.tcurnext, nlck.tcurnext, acquire, { in lck_ticket_lock_try_internal()
|
| H A D | sched_clutch.c | 1664 …os_atomic_rmw_loop(&clutch_bucket_group->scbg_cpu_data.scbcd_cpu_data_packed, old_cpu_data.scbcd_c… in sched_clutch_bucket_group_cpu_adjust() 2095 …os_atomic_rmw_loop(&clutch_bucket_group->scbg_pending_data.scct_packed, old_pending_data.scct_pack… in sched_clutch_bucket_group_thr_count_inc() 2111 …os_atomic_rmw_loop(&clutch_bucket_group->scbg_pending_data.scct_packed, old_pending_data.scct_pack… in sched_clutch_bucket_group_thr_count_dec() 2131 …os_atomic_rmw_loop(&clutch_bucket_group->scbg_pending_data.scct_packed, old_pending_data.scct_pack… in sched_clutch_bucket_group_pending_ageout() 2175 …bool score_updated = os_atomic_rmw_loop(&clutch_bucket_group->scbg_interactivity_data.scct_packed,… in sched_clutch_bucket_group_interactivity_score_calculate() 2296 …os_atomic_rmw_loop(&clutch_bucket_group->scbg_blocked_data.scct_packed, old_blocked_data.scct_pack… in sched_clutch_bucket_group_run_count_inc() 2324 …os_atomic_rmw_loop(&clutch_bucket_group->scbg_blocked_data.scct_packed, old_blocked_data.scct_pack… in sched_clutch_bucket_group_run_count_dec() 4399 …os_atomic_rmw_loop(&clutch_bucket_group->scbg_amp_rebalance_last_chosen, last_chosen_cluster, new_… in sched_edge_amp_rebalance_pset()
|
| H A D | epoch_sync.c | 320 return os_atomic_rmw_loop(counter, old, new, acq_rel, { in esync_update_epoch()
|
| H A D | kern_cdata.c | 98 os_atomic_rmw_loop(&lw_corpse_obj_cnt, oval, nval, relaxed, { in kcdata_object_throttle_get() 120 os_atomic_rmw_loop(&lw_corpse_obj_cnt, oval, nval, relaxed, { in kcdata_object_throttle_release()
|
| H A D | work_interval.c | 279 os_atomic_rmw_loop(&join_info->status, old_status, new_status, acquire, { in work_interval_auto_join_decrement()
|
| H A D | lock_mtx.c | 504 os_atomic_rmw_loop(&lock->lck_mtx.val, state.val, nstate.val, acquire, { in lck_mtx_ilk_lock_nopreempt()
|
| H A D | locks.c | 351 os_atomic_rmw_loop(&lock->lock_data, oldval, newval, acquire, { in hw_lock_trylock_contended() 378 os_atomic_rmw_loop(target, oldval, newval, acquire, { in hw_lock_trylock_bit()
|
| /xnu-10063.141.1/libkern/os/ |
| H A D | atomic_private_arch.h | 86 #undef os_atomic_rmw_loop 87 #define os_atomic_rmw_loop(p, ov, nv, m, ...) ({ \ macro 181 #undef os_atomic_rmw_loop 182 #define os_atomic_rmw_loop(p, ov, nv, m, ...) ({ \ macro
|
| H A D | refcnt.c | 334 os_atomic_rmw_loop(rc, cur, next, relaxed, { in __os_ref_retain_try() 600 os_atomic_rmw_loop(rc, cur, next, relaxed, { in os_ref_retain_try_mask_internal() 624 os_atomic_rmw_loop(rc, cur, next, acquire, { in os_ref_retain_try_acquire_mask_internal()
|
| H A D | atomic_private.h | 727 #define os_atomic_rmw_loop(p, ov, nv, m, ...) ({ \ macro
|
| /xnu-10063.141.1/san/memory/ |
| H A D | ubsan_log.c | 77 os_atomic_rmw_loop(&ubsan_log_next, i, n, relaxed, { in ubsan_log_append() 90 os_atomic_rmw_loop(&ubsan_log_head, e, n, relaxed, { in ubsan_log_append()
|
| /xnu-10063.141.1/doc/primitives/ |
| H A D | atomics.md | 244 as `os_atomic_rmw_loop` offers a much better alternative for CAS-loops. 246 ### `os_atomic_rmw_loop` 270 `os_atomic_rmw_loop` allows this form: 277 success = os_atomic_rmw_loop(address, old_value, new_value, acquire, { 296 In particular nesting of `os_atomic_rmw_loop` is invalid. 298 Use of `continue` within an `os_atomic_rmw_loop` is also invalid, instead an 308 success = os_atomic_rmw_loop(address, old_value, new_value, acquire, {
|
| /xnu-10063.141.1/bsd/skywalk/mem/ |
| H A D | skmem_cache_var.h | 348 os_atomic_rmw_loop(&bc->bc_usecnt, old, new, relaxed, { in skmem_bufctl_use() 364 os_atomic_rmw_loop(&bc->bc_usecnt, old, new, relaxed, { in skmem_bufctl_unuse()
|
| /xnu-10063.141.1/libkern/firehose/ |
| H A D | chunk_private.h | 105 os_atomic_rmw_loop(&fc->fc_pos.fcp_atomic_pos, in firehose_chunk_tracepoint_try_reserve()
|
| /xnu-10063.141.1/libkern/gen/ |
| H A D | OSAtomicOperations.c | 182 return !os_atomic_rmw_loop(address, oldValue, newValue, relaxed, { in OSTestAndSetClear()
|
| /xnu-10063.141.1/osfmk/prng/ |
| H A D | entropy.c | 185 os_atomic_rmw_loop(&entropy_analysis_sample_count, sample_count, next_sample_count, relaxed, { in entropy_analysis_store()
|
| /xnu-10063.141.1/osfmk/arm64/sptm/pmap/ |
| H A D | pmap_data.h | 1555 os_atomic_rmw_loop(&pp_attr_table[pai], prev_ppattr, new_ppattr, relaxed, { in ppattr_modify_bits()
|
| /xnu-10063.141.1/osfmk/vm/ |
| H A D | vm_reclaim.c | 940 success = os_atomic_rmw_loop(&metadata->vdrm_num_bytes_put_in_buffer, num_bytes_in_buffer, in vm_deferred_reclamation_buffer_update_reclaimable_bytes_internal()
|
| /xnu-10063.141.1/bsd/pthread/ |
| H A D | pthread_workqueue.c | 755 ret = os_atomic_rmw_loop(tg_loc, old_tg, new_tg, relaxed, { in workq_thread_reset_pri() 778 os_atomic_rmw_loop(tg_loc, old_tg, new_tg, relaxed, { in workq_thread_reset_pri() 1731 os_atomic_rmw_loop(&wq->wq_flags, old_flags, new_flags, acquire, { in workq_thread_call_prepost()
|
| /xnu-10063.141.1/bsd/kern/ |
| H A D | kern_aio.c | 372 os_atomic_rmw_loop(&aio_anchor.aio_total_count, old, new, relaxed, { in aio_try_proc_insert_active_locked()
|
| H A D | kern_proc.c | 707 os_atomic_rmw_loop(&p->p_refcount, o_bits, n_bits, release, { in proc_rele() 858 os_atomic_rmw_loop(&p->p_refcount, o_ref, n_ref, relaxed, { in proc_refdrain_will_exec()
|
| H A D | uipc_usrreq.c | 2765 os_atomic_rmw_loop(&fg->fg_flag, oflags, nflags, relaxed, { in unp_mark()
|
| H A D | kern_event.c | 3292 int ret = os_atomic_rmw_loop(&kqwl->kqwl_preadopt_tg, old_tg, new_tg, relaxed, { in kqueue_set_preadopted_thread_group() 5388 int ret = os_atomic_rmw_loop(kqr_preadopt_thread_group_addr(kqr), old_tg, new_tg, relaxed, { in kqueue_threadreq_bind() 5951 int ret = os_atomic_rmw_loop(kqr_preadopt_thread_group_addr(kqr), old_tg, new_tg, relaxed, { in kqworkloop_unbind_locked()
|