| /xnu-12377.1.9/libsyscall/mach/ |
| H A D | vm_reclaim.c | 206 tail = os_atomic_load_wide(&ring->tail, relaxed); in mach_vm_reclaim_try_enter() 207 head = os_atomic_load_wide(&ring->head, relaxed); in mach_vm_reclaim_try_enter() 224 head = os_atomic_load_wide(&ring->head, relaxed); in mach_vm_reclaim_try_enter() 234 original_tail = os_atomic_load_wide(&ring->tail, relaxed); in mach_vm_reclaim_try_enter() 240 busy = os_atomic_load_wide(&ring->busy, relaxed); in mach_vm_reclaim_try_enter() 304 head = os_atomic_load_wide(&ring->head, relaxed); in mach_vm_reclaim_try_cancel() 327 original_tail = os_atomic_load_wide(&ring->tail, relaxed); in mach_vm_reclaim_try_cancel() 333 busy = os_atomic_load_wide(&ring->busy, relaxed); in mach_vm_reclaim_try_cancel() 380 mach_vm_reclaim_id_t head = os_atomic_load_wide(&ring->head, relaxed); in mach_vm_reclaim_query_state() 395 mach_vm_reclaim_id_t busy = os_atomic_load_wide(&ring->busy, relaxed); in mach_vm_reclaim_query_state() [all …]
|
| /xnu-12377.1.9/osfmk/kern/ |
| H A D | counter_common.c | 67 uint64_t current_value = os_atomic_load_wide(zpercpu_get(*counter), relaxed); in scalable_counter_static_init() 151 return os_atomic_load_wide(counter, relaxed); in counter_load() 160 value += os_atomic_load_wide(it, relaxed); in counter_load()
|
| H A D | sched_common.c | 92 (sched_pset_search_order_t)os_atomic_load_wide(&search_order->spso_packed, relaxed); in sched_iterate_psets_ordered()
|
| H A D | recount.c | 1192 __assert_only uint64_t state_time = os_atomic_load_wide( in recount_processor_idle() 1223 uint64_t state = os_atomic_load_wide(&pr->rpr_state_last_abs_time, relaxed); in recount_processor_run() 1248 uint64_t idle_stamp = os_atomic_load_wide(&pr->rpr_state_last_abs_time, in recount_processor_usage()
|
| H A D | sched_rt.c | 1251 assert(os_atomic_load_wide(&rt_run_queue->earliest_deadline, relaxed) == earliest_deadline); in check_rt_runq_consistency() 1305 if (earliest && (deadline < os_atomic_load_wide(&rt_run_queue->earliest_deadline, relaxed))) { in rt_runq_enqueue() 1325 return os_atomic_load_wide(&pset->rt_runq.earliest_deadline, relaxed); in rt_runq_earliest_deadline()
|
| H A D | sched_clutch.c | 294 …scb_cpu_data.scbcd_cpu_data_packed = os_atomic_load_wide(&clutch_bucket_group->scbg_cpu_data.scbcd… in sched_clutch_thread_group_cpu_time_for_thread() 1692 …scb_cpu_data.scbcd_cpu_data_packed = os_atomic_load_wide(&clutch_bucket_group->scbg_cpu_data.scbcd… in sched_clutch_interactivity_from_cpu_data() 2075 …uint64_t bucket_group_run_count = os_atomic_load_wide(&clutch_bucket_group->scbg_blocked_data.scct… in sched_clutch_bucket_group_pri_shift_update()
|
| H A D | sched_prim.c | 6741 uint64_t load_compute_deadline = os_atomic_load_wide(&sched_load_compute_deadline, relaxed); in sched_timeshare_consider_maintenance()
|
| /xnu-12377.1.9/osfmk/vm/ |
| H A D | analytics.c | 106 e->over_global_limit = os_atomic_load_wide(&vm_add_wire_count_over_global_limit, relaxed); in report_mlock_failures() 107 e->over_user_limit = os_atomic_load_wide(&vm_add_wire_count_over_user_limit, relaxed); in report_mlock_failures()
|
| /xnu-12377.1.9/libkern/os/ |
| H A D | refcnt.c | 783 n = os_atomic_load_wide(zpercpu_get_cpu(rc, 0), relaxed); 788 n |= os_atomic_load_wide(zpercpu_get_cpu(rc, cpu), relaxed); 804 v = os_atomic_load_wide(zpercpu_get_cpu(os_pcpu_get(ref), 0), relaxed); in os_pcpu_ref_count() 820 v = os_atomic_load_wide(rcp, relaxed); in __os_pcpu_ref_delta()
|
| H A D | atomic_private.h | 336 #define os_atomic_load_wide(p, m) ({ \ macro
|
| H A D | log_queue.c | 188 os_atomic_load_wide(_v, dependency); \
|
| /xnu-12377.1.9/bsd/kern/ |
| H A D | counter_test.c | 235 uint64_t value = os_atomic_load_wide(&atomic_counter, relaxed);
|
| H A D | kern_exec.c | 6627 local_experiment_factors = os_atomic_load_wide(&libmalloc_experiment_factors, relaxed); in exec_add_apple_strings() 8342 uint64_t value = os_atomic_load_wide(&libmalloc_experiment_factors, relaxed);
|
| H A D | kern_sysctl.c | 4443 old_value = os_atomic_load_wide(ptr, relaxed);
|
| H A D | kern_event.c | 9572 buf[nknotes] = os_atomic_load_wide(&kn->kn_udata, relaxed);
|
| /xnu-12377.1.9/bsd/dev/ |
| H A D | monotonic.c | 361 uint64_t value = os_atomic_load_wide(&mt_retrograde, relaxed);
|
| /xnu-12377.1.9/osfmk/arm/ |
| H A D | machine_routines_common.c | 381 return os_atomic_load_wide(&perfcontrol_callout_stats[type][stat], relaxed) / in perfcontrol_callout_stat_avg() 382 os_atomic_load_wide(&perfcontrol_callout_count[type], relaxed); in perfcontrol_callout_stat_avg()
|
| /xnu-12377.1.9/osfmk/arm/commpage/ |
| H A D | commpage.c | 953 uint64_t saved_data = os_atomic_load_wide(approx_time_base, relaxed); in commpage_update_mach_approximate_time()
|
| /xnu-12377.1.9/doc/primitives/ |
| H A D | atomics.md | 167 compile to a plain load or store. `os_atomic_load_wide` and
|
| /xnu-12377.1.9/bsd/pthread/ |
| H A D | pthread_workqueue.c | 332 return os_atomic_load_wide(&wq->wq_thactive, relaxed); in _wq_thactive() 1888 uint64_t lastblocked_ts = os_atomic_load_wide(lastblocked_tsp, relaxed); in workq_thread_is_busy()
|