Home
last modified time | relevance | path

Searched refs:relaxed (Results 1 – 25 of 304) sorted by relevance

12345678910>>...13

/xnu-12377.81.4/tools/cocci/
H A DOSAtomic_rewrite.cocci11 + os_atomic_inc_orig(E, relaxed)
14 + os_atomic_inc_orig(E, relaxed)
17 + os_atomic_inc_orig(E, relaxed)
20 + os_atomic_inc_orig(E, relaxed)
23 + os_atomic_inc_orig(E, relaxed)
26 + os_atomic_inc_orig(E, relaxed)
29 + os_atomic_inc_orig(E, relaxed)
32 + os_atomic_inc_orig(E, relaxed)
35 + os_atomic_inc_orig(E, relaxed)
38 + os_atomic_inc_orig(E, relaxed)
[all …]
H A Dhw_atomic_rewrite.cocci11 + os_atomic_dec_orig(E, relaxed)
14 + os_atomic_dec(E, relaxed)
17 + os_atomic_sub_orig(E, F, relaxed)
20 + os_atomic_sub(E, F, relaxed)
23 + os_atomic_inc_orig(E, relaxed)
26 + os_atomic_inc(E, relaxed)
29 + os_atomic_add_orig(E, F, relaxed)
32 + os_atomic_add(E, F, relaxed)
39 + os_atomic_inc_orig(E, relaxed)
42 + os_atomic_inc(E, relaxed)
[all …]
/xnu-12377.81.4/tests/
H A Dos_atomic.cpp16 T_ASSERT_EQ(os_atomic_inc_orig(&i, relaxed), 0, "atomic inc");
17 T_ASSERT_EQ(os_atomic_cmpxchg(&i, 1, 0, relaxed), true, "os_atomic_cmpxchg");
18 os_atomic_rmw_loop(&i, a, b, relaxed, {
22 T_ASSERT_EQ(os_atomic_inc_orig(&old_i, relaxed), 0, "atomic inc");
23 T_ASSERT_EQ(os_atomic_cmpxchg(&old_i, 1, 0, relaxed), true, "os_atomic_cmpxchg");
24 os_atomic_rmw_loop(&old_i, a, b, relaxed, {
28 T_ASSERT_EQ(os_atomic_inc_orig(&v_i, relaxed), 0, "atomic inc");
29 T_ASSERT_EQ(os_atomic_cmpxchg(&v_i, 1, 0, relaxed), true, "os_atomic_cmpxchg");
30 os_atomic_rmw_loop(&v_i, a, b, relaxed, {
H A Dos_refcnt.c430 while (os_atomic_load(&pcpu_perf_step, relaxed) == 0) { in worker_ref()
433 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in worker_ref()
439 os_atomic_add(count, n, relaxed); in worker_ref()
447 while (os_atomic_load(&pcpu_perf_step, relaxed) == 0) { in worker_pcpu_ref()
450 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in worker_pcpu_ref()
456 os_atomic_add(count, n, relaxed); in worker_pcpu_ref()
464 os_atomic_store(&pcpu_perf_step, 1, relaxed); in warmup_thread_pool()
468 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in warmup_thread_pool()
473 os_atomic_store(&pcpu_perf_step, 0, relaxed); in warmup_thread_pool()
496 os_atomic_store(&pcpu_perf_step, 1, relaxed);
[all …]
/xnu-12377.81.4/libkern/gen/
H A DOSAtomicOperations.c101 return os_atomic_add_orig(address, (SInt8)amount, relaxed); in OSAddAtomic8()
107 return os_atomic_add_orig(address, (SInt16)amount, relaxed); in OSAddAtomic16()
115 return os_atomic_add_orig(address, amount, relaxed); in OSAddAtomic()
125 return os_atomic_add_orig(aligned_address, amount, relaxed); in OSAddAtomic64()
132 return os_atomic_add_orig(address, theAmount, relaxed); in OSAddAtomicLong()
139 return os_atomic_inc_orig(value, relaxed); in OSIncrementAtomic()
146 return os_atomic_dec_orig(value, relaxed); in OSDecrementAtomic()
153 return os_atomic_and_orig(value, mask, relaxed); in OSBitAndAtomic()
160 return os_atomic_or_orig(value, mask, relaxed); in OSBitOrAtomic()
167 return os_atomic_xor_orig(value, mask, relaxed); in OSBitXorAtomic()
[all …]
/xnu-12377.81.4/libsyscall/mach/
H A Dvm_reclaim.c159 size_t reclaimable_bytes = os_atomic_sub(&ring->reclaimable_bytes, bytes_reclaimed, relaxed); in mach_vm_reclaim_ring_resize()
160 os_atomic_min(&ring->reclaimable_bytes_min, reclaimable_bytes, relaxed); in mach_vm_reclaim_ring_resize()
206 tail = os_atomic_load_wide(&ring->tail, relaxed); in mach_vm_reclaim_try_enter()
207 head = os_atomic_load_wide(&ring->head, relaxed); in mach_vm_reclaim_try_enter()
221 os_atomic_inc(&ring->tail, relaxed); in mach_vm_reclaim_try_enter()
224 head = os_atomic_load_wide(&ring->head, relaxed); in mach_vm_reclaim_try_enter()
234 original_tail = os_atomic_load_wide(&ring->tail, relaxed); in mach_vm_reclaim_try_enter()
238 os_atomic_store_wide(&ring->tail, requested_id, relaxed); in mach_vm_reclaim_try_enter()
240 busy = os_atomic_load_wide(&ring->busy, relaxed); in mach_vm_reclaim_try_enter()
243 os_atomic_store_wide(&ring->tail, original_tail, relaxed); in mach_vm_reclaim_try_enter()
[all …]
/xnu-12377.81.4/osfmk/kern/
H A Dcounter_common.c67 uint64_t current_value = os_atomic_load_wide(zpercpu_get(*counter), relaxed); in scalable_counter_static_init()
73 os_atomic_store_wide(zpercpu_get(*counter), current_value, relaxed); in scalable_counter_static_init()
88 os_atomic_store_wide(counter, 0, relaxed); in counter_alloc()
109 os_atomic_add(counter, amount, relaxed); in counter_add()
116 os_atomic_inc(counter, relaxed); in counter_inc()
123 os_atomic_dec(counter, relaxed); in counter_dec()
151 return os_atomic_load_wide(counter, relaxed); in counter_load()
160 value += os_atomic_load_wide(it, relaxed); in counter_load()
H A Dexclaves_memory.c96 e->pages_alloced = os_atomic_load(&exclaves_allocation_statistics.pages_alloced, relaxed); in exclaves_memory_report_accounting()
97 e->pages_freed = os_atomic_load(&exclaves_allocation_statistics.pages_freed, relaxed); in exclaves_memory_report_accounting()
98 e->time_allocating = os_atomic_load(&exclaves_allocation_statistics.time_allocating, relaxed); in exclaves_memory_report_accounting()
99 e->max_alloc_latency = os_atomic_load(&exclaves_allocation_statistics.max_alloc_latency, relaxed); in exclaves_memory_report_accounting()
100 …ncy_highbit0 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[0], relaxed); in exclaves_memory_report_accounting()
101 …ncy_highbit1 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[1], relaxed); in exclaves_memory_report_accounting()
102 …ncy_highbit2 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[2], relaxed); in exclaves_memory_report_accounting()
103 …ncy_highbit3 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[3], relaxed); in exclaves_memory_report_accounting()
104 …ncy_highbit4 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[4], relaxed); in exclaves_memory_report_accounting()
105 …ncy_highbit5 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit[5], relaxed); in exclaves_memory_report_accounting()
[all …]
H A Dmpsc_queue.c48 mpsc_queue_chain_t head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch()
50 os_atomic_store(&last->mpqc_next, head, relaxed); in mpsc_queue_restore_batch()
54 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch()
58 os_atomic_store(&last->mpqc_next, head, relaxed); in mpsc_queue_restore_batch()
61 os_atomic_store(&q->mpqh_head.mpqc_next, first, relaxed); in mpsc_queue_restore_batch()
72 tail = os_atomic_load(&q->mpqh_tail, relaxed); in mpsc_queue_dequeue_batch()
78 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_dequeue_batch()
82 os_atomic_store(&q->mpqh_head.mpqc_next, NULL, relaxed); in mpsc_queue_dequeue_batch()
111 elm = os_atomic_load(&cur->mpqc_next, relaxed); in mpsc_queue_batch_next()
291 os_atomic_andnot(&dq->mpd_state, MPSC_QUEUE_STATE_WAKEUP, relaxed); in _mpsc_daemon_queue_drain()
[all …]
H A Dexclaves_boot.c147 EXCLAVES_BS_NOT_STARTED, relaxed); in exclaves_check_sk()
150 EXCLAVES_BS_NOT_SUPPORTED, relaxed); in exclaves_check_sk()
160 while (os_atomic_load(&exclaves_boot_status, relaxed) < status) { in exclaves_boot_status_wait()
176 assert3u(status, >, os_atomic_load(&exclaves_boot_status, relaxed)); in exclaves_boot_status_set()
206 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_exclavecore()
258 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_exclavekit()
424 exclaves_boot_status_t boot_status = os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_get_boot_status_string()
449 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_get_boot_stage()
476 status = os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_supported()
H A Dsched_rt.c216 os_atomic_store(&pset_array[src_pset]->sched_rt_edges[dst_pset], edge_config, relaxed); in sched_rt_config_set()
224 return os_atomic_load(&pset_array[src_pset]->sched_rt_edges[dst_pset], relaxed); in sched_rt_config_get()
557 uint64_t nset_deadline = os_atomic_load(&nset->stealable_rt_threads_earliest_deadline, relaxed); in sched_rt_steal_thread()
570 if (os_atomic_load(&pset->stealable_rt_threads_earliest_deadline, relaxed) <= target_deadline) { in sched_rt_steal_thread()
785 …mic_store(&pset->stealable_rt_threads_earliest_deadline, rt_runq_earliest_deadline(pset), relaxed); in pset_update_rt_stealable_state()
787 os_atomic_store(&pset->stealable_rt_threads_earliest_deadline, RT_DEADLINE_NONE, relaxed); in pset_update_rt_stealable_state()
1226 assert(os_atomic_load_wide(&rt_run_queue->earliest_deadline, relaxed) == earliest_deadline); in check_rt_runq_consistency()
1227 assert(os_atomic_load(&rt_run_queue->count, relaxed) == count); in check_rt_runq_consistency()
1228 assert(os_atomic_load(&rt_run_queue->constraint, relaxed) == constraint); in check_rt_runq_consistency()
1229 assert(os_atomic_load(&rt_run_queue->ed_index, relaxed) == ed_index); in check_rt_runq_consistency()
[all …]
H A Dsched_average.c190 load_now[TH_BUCKET_RUN] = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in compute_sched_load()
191 load_now[TH_BUCKET_FIXPRI] = os_atomic_load(&sched_run_buckets[TH_BUCKET_FIXPRI], relaxed); in compute_sched_load()
192 load_now[TH_BUCKET_SHARE_FG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_FG], relaxed); in compute_sched_load()
193 load_now[TH_BUCKET_SHARE_DF] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_DF], relaxed); in compute_sched_load()
194 load_now[TH_BUCKET_SHARE_UT] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_UT], relaxed); in compute_sched_load()
195 load_now[TH_BUCKET_SHARE_BG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_BG], relaxed); in compute_sched_load()
288 uint32_t nthreads = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed) - 1; in compute_averages()
H A Dlock_ptr.c62 hw_lck_ptr_t tmp = os_atomic_load(lck, relaxed); in __hw_lck_ptr_invalid_panic()
113 hw_lck_ptr_t tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_destroy()
128 os_atomic_store(lck, tmp, relaxed); in hw_lck_ptr_destroy()
134 return os_atomic_load(lck, relaxed).lck_ptr_locked; in hw_lck_ptr_held()
147 tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_timeout_panic()
196 os_atomic_store(&pnode->lsm_next, txn.txn_slot, relaxed); in hw_lck_ptr_contended()
235 os_atomic_store(&nnode->lsm_ready, 1, relaxed); in hw_lck_ptr_contended()
271 tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_lock_fastpath()
H A Dsched_clutch.c286 …a_packed = os_atomic_load_wide(&clutch_bucket_group->scbg_cpu_data.scbcd_cpu_data_packed, relaxed); in sched_clutch_thread_group_cpu_time_for_thread()
675 …ri + (int)(os_atomic_load(&prev_clutch_bucket_group->scbg_interactivity_data.scct_count, relaxed)); in sched_clutch_root_unbound_select_aboveui()
1183 return (int)os_atomic_load(&sched_clutch_global_bucket_load[bucket], relaxed); in sched_clutch_global_bucket_load_get()
1404 os_atomic_store(&clutch_bucket_group->scbg_timeshare_tick, 0, relaxed); in sched_clutch_bucket_group_init()
1405 os_atomic_store(&clutch_bucket_group->scbg_pri_shift, INT8_MAX, relaxed); in sched_clutch_bucket_group_init()
1406 …mic_store(&clutch_bucket_group->scbg_preferred_cluster, sched_boot_pset->pset_cluster_id, relaxed); in sched_clutch_bucket_group_init()
1413 …pu_data.scbcd_cpu_blocked, (clutch_cpu_data_t)sched_clutch_bucket_group_adjust_threshold, relaxed); in sched_clutch_bucket_group_init()
1436 os_atomic_store(&clutch->sc_thr_count, 0, relaxed); in sched_clutch_init_with_thread_group()
1456 assert(os_atomic_load(&clutch->sc_thr_count, relaxed) == 0); in sched_clutch_destroy()
1477 return os_atomic_load(&clutch_bucket_group->scbg_preferred_cluster, relaxed); in sched_edge_clutch_bucket_group_preferred_cluster()
[all …]
H A Dast.h195 #define thread_ast_set(act, reason) ((void)os_atomic_or(&(act)->ast, (reason), relaxed))
196 #define thread_ast_clear(act, reason) ((void)os_atomic_andnot(&(act)->ast, (reason), relaxed))
197 #define thread_ast_peek(act, reason) (os_atomic_load(&(act)->ast, relaxed) & (reason))
198 #define thread_ast_get(act) os_atomic_load(&(act)->ast, relaxed)
H A Dlock_group.c276 os_atomic_inc(cnt, relaxed); in lck_grp_reference()
287 os_atomic_dec(cnt, relaxed); in lck_grp_deallocate()
308 os_atomic_or(&lck_debug_state.lds_value, bit, relaxed); in lck_grp_enable_feature()
325 os_atomic_andnot(&lck_debug_state.lds_value, bit, relaxed); in lck_grp_disable_feature()
354 needed = os_atomic_load(&lck_grp_table.cidt_count, relaxed); in host_lockgroup_info()
445 os_atomic_or(&attr->lck_attr_val, LCK_ATTR_DEBUG, relaxed); in lck_attr_setdebug()
451 os_atomic_andnot(&attr->lck_attr_val, LCK_ATTR_DEBUG, relaxed); in lck_attr_cleardebug()
457 os_atomic_or(&attr->lck_attr_val, LCK_ATTR_RW_SHARED_PRIORITY, relaxed); in lck_attr_rw_shared_priority()
496 __unused uint64_t val = os_atomic_inc_orig(&stat->lgs_count, relaxed); in lck_grp_stat_inc()
509 __unused uint64_t val = os_atomic_add_orig(&stat->lgs_count, time, relaxed); in lck_grp_inc_time_stats()
H A Dsmr.c651 s_wr_seq = os_atomic_load(&smr->smr_clock.s_wr_seq, relaxed); in __smr_enter()
657 os_atomic_store(&pcpu->c_rd_seq, s_wr_seq | sleepable, relaxed); in __smr_enter()
724 os_atomic_store(&smrw->sect_waiter, NULL, relaxed); in __smr_wake_oncore_sleepers()
776 os_atomic_store(&pcpu->stall_rd_seq, t->smrt_seq, relaxed); in smr_mark_active_trackers_stalled()
899 os_atomic_rmw_loop(&smr->smr_clock.s_rd_seq, o_seq, rd_seq, relaxed, { in __smr_rd_advance()
1015 if (lock_cmpxchg(&smrw->sect_waiter, NULL, self, relaxed)) { in __smr_wait_for_oncore()
1032 return os_atomic_load(&pcpu->c_rd_seq, relaxed); in __smr_wait_for_oncore()
1067 clk.s_wr_seq, goal, &clk.s_wr_seq, relaxed)) { in __smr_scan()
1104 smr_seq_t seq = os_atomic_load(&pcpu->c_rd_seq, relaxed); in __smr_scan()
1150 smr_seq_t seq = os_atomic_load(&pcpu->stall_rd_seq, relaxed); in __smr_scan()
[all …]
/xnu-12377.81.4/san/memory/
H A Dubsan_log.c77 os_atomic_rmw_loop(&ubsan_log_next, i, n, relaxed, { in ubsan_log_append()
90 os_atomic_rmw_loop(&ubsan_log_head, e, n, relaxed, { in ubsan_log_append()
119 head = os_atomic_load(&ubsan_log_head, relaxed);
121 tail = os_atomic_load(&ubsan_log_tail, relaxed);
157 head = os_atomic_load(&ubsan_log_head, relaxed);
159 tail = os_atomic_load(&ubsan_log_tail, relaxed);
172 os_atomic_store(&ubsan_log_tail, head, relaxed);
/xnu-12377.81.4/bsd/kern/
H A Dmem_acct.c72 allocated = os_atomic_load(&macct->ma_allocated, relaxed); in mem_acct_limited()
126 allocated = os_atomic_add(&macct->ma_allocated, *pcpu, relaxed); in _mem_acct_add()
133 os_atomic_max(&macct->ma_peak, allocated, relaxed); in _mem_acct_add()
299 value = os_atomic_load(&acct->ma_peak, relaxed); in sysctl_subsystem_peak()
305 os_atomic_store(&acct->ma_peak, value, relaxed); in sysctl_subsystem_peak()
321 hardlimit = os_atomic_load(&acct->ma_hardlimit, relaxed); in sysctl_subsystem_soft_limit()
347 value = os_atomic_load(&acct->ma_hardlimit, relaxed); in sysctl_subsystem_hard_limit()
369 value = os_atomic_load(&acct->ma_allocated, relaxed); in sysctl_subsystem_allocated()
484 s->peak = os_atomic_load(&a->ma_peak, relaxed); in memacct_copy_stats()
485 s->allocated = os_atomic_load(&a->ma_allocated, relaxed); in memacct_copy_stats()
/xnu-12377.81.4/osfmk/bank/
H A Dbank_internal.h120 (os_atomic_inc_orig(&(elem)->bt_made, relaxed))
123 (os_atomic_dec_orig(&(elem)->bt_made, relaxed))
126 (os_atomic_sub_orig(&(elem)->bt_made, (num), relaxed))
171 (os_atomic_inc_orig(&(elem)->ba_made, relaxed))
174 (os_atomic_dec_orig(&(elem)->ba_made, relaxed))
177 (os_atomic_sub_orig(&(elem)->ba_made, (num), relaxed))
/xnu-12377.81.4/osfmk/arm/
H A Dcounter.c40 os_atomic_add(zpercpu_get(*counter), amount, relaxed); in counter_add()
47 os_atomic_inc(zpercpu_get(*counter), relaxed); in counter_inc()
54 os_atomic_dec(zpercpu_get(*counter), relaxed); in counter_dec()
/xnu-12377.81.4/libkern/os/
H A Drefcnt.c68 …panic("os_refcnt: overflow (rc=%p, count=%u, max=%u)", rc, os_atomic_load(rc, relaxed), OS_REFCNT_… in os_ref_panic_overflow()
76 if (os_atomic_load(rc, relaxed) >= OS_REFCNT_MAX_COUNT) { in os_ref_panic_retain()
362 os_atomic_rmw_loop(rc, cur, next, relaxed, { in __os_ref_retain_try()
624 os_atomic_rmw_loop(rc, cur, next, relaxed, { in os_ref_retain_try_mask_internal()
783 n = os_atomic_load_wide(zpercpu_get_cpu(rc, 0), relaxed);
788 n |= os_atomic_load_wide(zpercpu_get_cpu(rc, cpu), relaxed);
804 v = os_atomic_load_wide(zpercpu_get_cpu(os_pcpu_get(ref), 0), relaxed); in os_pcpu_ref_count()
820 v = os_atomic_load_wide(rcp, relaxed); in __os_pcpu_ref_delta()
828 return os_atomic_add_orig(rcp, OS_PCPU_REF_INC, relaxed); in __os_pcpu_ref_delta()
840 OS_PCPU_REF_INC, relaxed); in __os_pcpu_ref_retain_slow()
[all …]
/xnu-12377.81.4/bsd/net/
H A Ddlil_ctl.c268 os_atomic_add(&ifp->if_data.ifi_ipackets, s->packets_in, relaxed); in dlil_input_stats_sync()
272 os_atomic_add(&ifp->if_data.ifi_ibytes, s->bytes_in, relaxed); in dlil_input_stats_sync()
276 os_atomic_add(&ifp->if_data.ifi_ierrors, s->errors_in, relaxed); in dlil_input_stats_sync()
281 os_atomic_add(&ifp->if_data.ifi_opackets, s->packets_out, relaxed); in dlil_input_stats_sync()
285 os_atomic_add(&ifp->if_data.ifi_obytes, s->bytes_out, relaxed); in dlil_input_stats_sync()
289 os_atomic_add(&ifp->if_data.ifi_oerrors, s->errors_out, relaxed); in dlil_input_stats_sync()
294 os_atomic_add(&ifp->if_data.ifi_collisions, s->collisions, relaxed); in dlil_input_stats_sync()
298 os_atomic_add(&ifp->if_data.ifi_iqdrops, s->dropped, relaxed); in dlil_input_stats_sync()
/xnu-12377.81.4/san/coverage/
H A Dkcov_ksancov.c221 if (os_atomic_load(&dev->trace->kt_head, relaxed) >= dev->maxpcs) { in trace_pc_guard_pcs()
225 uint32_t idx = os_atomic_inc_orig(&dev->trace->kt_head, relaxed); in trace_pc_guard_pcs()
238 if (os_atomic_load(&dev->trace->kt_head, relaxed) >= dev->maxpcs) { in trace_pc_guard_pcs_stk()
242 uint32_t idx = os_atomic_inc_orig(&dev->trace->kt_head, relaxed); in trace_pc_guard_pcs_stk()
322 if (os_atomic_load(&dev->hdr->kh_enabled, relaxed) == 0) { in kcov_ksancov_trace_pc()
385 if (os_atomic_load(&dev->cmps_hdr, relaxed) == NULL) { in kcov_ksancov_trace_cmp()
388 if (os_atomic_load(&dev->cmps_hdr->kh_enabled, relaxed) == 0) { in kcov_ksancov_trace_cmp()
405 if (os_atomic_load(&dev->cmps_trace->kt_head, relaxed) >= max_entries) { in kcov_ksancov_trace_cmp()
409 uint32_t idx = os_atomic_inc_orig(&dev->cmps_trace->kt_head, relaxed); in kcov_ksancov_trace_cmp()
438 if (os_atomic_load(&dev->cmps_hdr, relaxed) == NULL) { in kcov_ksancov_trace_cmp_func()
[all …]
/xnu-12377.81.4/bsd/skywalk/nexus/flowswitch/flow/
H A Dflow_route.c461 os_atomic_andnot(&fr->fr_flags, FLOWRTF_DELETED, relaxed); in flow_route_configure()
509 os_atomic_or(&fr->fr_flags, FLOWRTF_GATEWAY, relaxed); in flow_route_configure()
514 os_atomic_or(&fr->fr_flags, FLOWRTF_ONLINK, relaxed); in flow_route_configure()
527 os_atomic_or(&fr->fr_flags, FLOWRTF_STABLE_ADDR, relaxed); in flow_route_configure()
529 os_atomic_andnot(&fr->fr_flags, FLOWRTF_STABLE_ADDR, relaxed); in flow_route_configure()
607 os_atomic_inc(&fr->fr_want_configure, relaxed); in flow_route_find()
690 os_atomic_inc(&fr->fr_want_configure, relaxed); in flow_route_find()
721 os_atomic_or(&fr->fr_flags, FLOWRTF_ATTACHED, relaxed); in flow_route_find()
894 os_atomic_andnot(&fr->fr_flags, FLOWRTF_ATTACHED, relaxed); in flow_route_bucket_purge_common()
1070 os_atomic_inc(&fr->fr_want_configure, relaxed); in flow_route_ev_callback()
[all …]

12345678910>>...13