Home
last modified time | relevance | path

Searched refs:os_atomic_load (Results 1 – 25 of 140) sorted by relevance

123456

/xnu-11215.1.10/osfmk/kern/
H A Dexclaves_memory.c79 e->pages_alloced = os_atomic_load(&exclaves_allocation_statistics.pages_alloced, relaxed); in exclaves_memory_report_accounting()
80 e->pages_freed = os_atomic_load(&exclaves_allocation_statistics.pages_freed, relaxed); in exclaves_memory_report_accounting()
81 e->time_allocating = os_atomic_load(&exclaves_allocation_statistics.time_allocating, relaxed); in exclaves_memory_report_accounting()
82 e->max_alloc_latency = os_atomic_load(&exclaves_allocation_statistics.max_alloc_latency, relaxed); in exclaves_memory_report_accounting()
83 …e->alloc_latency_highbit0 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit… in exclaves_memory_report_accounting()
84 …e->alloc_latency_highbit1 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit… in exclaves_memory_report_accounting()
85 …e->alloc_latency_highbit2 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit… in exclaves_memory_report_accounting()
86 …e->alloc_latency_highbit3 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit… in exclaves_memory_report_accounting()
87 …e->alloc_latency_highbit4 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit… in exclaves_memory_report_accounting()
88 …e->alloc_latency_highbit5 = os_atomic_load(&exclaves_allocation_statistics.alloc_latency_byhighbit… in exclaves_memory_report_accounting()
[all …]
H A Dexclaves_boot.c160 while (os_atomic_load(&exclaves_boot_status, relaxed) < status) { in exclaves_boot_status_wait()
176 assert3u(status, >, os_atomic_load(&exclaves_boot_status, relaxed)); in exclaves_boot_status_set()
206 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_stage_2()
258 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_exclavekit()
333 os_atomic_load(&exclaves_boot_status, acquire); in exclaves_boot_wait()
404 exclaves_boot_status_t boot_status = os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_get_boot_status_string()
429 os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_get_boot_stage()
456 status = os_atomic_load(&exclaves_boot_status, relaxed); in exclaves_boot_supported()
H A Dsched_average.c193 load_now[TH_BUCKET_RUN] = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed); in compute_sched_load()
194 load_now[TH_BUCKET_FIXPRI] = os_atomic_load(&sched_run_buckets[TH_BUCKET_FIXPRI], relaxed); in compute_sched_load()
195 load_now[TH_BUCKET_SHARE_FG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_FG], relaxed); in compute_sched_load()
196 load_now[TH_BUCKET_SHARE_DF] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_DF], relaxed); in compute_sched_load()
197 load_now[TH_BUCKET_SHARE_UT] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_UT], relaxed); in compute_sched_load()
198 load_now[TH_BUCKET_SHARE_BG] = os_atomic_load(&sched_run_buckets[TH_BUCKET_SHARE_BG], relaxed); in compute_sched_load()
291 uint32_t nthreads = os_atomic_load(&sched_run_buckets[TH_BUCKET_RUN], relaxed) - 1; in compute_averages()
H A Dlock_ptr.c62 hw_lck_ptr_t tmp = os_atomic_load(lck, relaxed); in __hw_lck_ptr_invalid_panic()
113 hw_lck_ptr_t tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_destroy()
134 return os_atomic_load(lck, relaxed).lck_ptr_locked; in hw_lck_ptr_held()
147 tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_timeout_panic()
271 tmp = os_atomic_load(lck, relaxed); in hw_lck_ptr_lock_fastpath()
413 hw_lck_ptr_t tmp = os_atomic_load(lck, acquire); in hw_lck_ptr_wait_for_value()
H A Dsmr.c651 s_wr_seq = os_atomic_load(&smr->smr_clock.s_wr_seq, relaxed); in __smr_enter()
722 if (__improbable(os_atomic_load(&smrw->sect_waiter, compiler_acq_rel))) { in __smr_wake_oncore_sleepers()
1032 return os_atomic_load(&pcpu->c_rd_seq, relaxed); in __smr_wait_for_oncore()
1104 smr_seq_t seq = os_atomic_load(&pcpu->c_rd_seq, relaxed); in __smr_scan()
1150 smr_seq_t seq = os_atomic_load(&pcpu->stall_rd_seq, relaxed); in __smr_scan()
1188 clk.s_rd_seq = os_atomic_load(&smr->smr_clock.s_rd_seq, acquire); in __smr_poll()
1197 clk.s_wr_seq = os_atomic_load(&smr->smr_clock.s_wr_seq, relaxed); in __smr_poll()
1222 return SMR_SEQ_INC + os_atomic_load(&smr->smr_clock.s_wr_seq, relaxed); in smr_deferred_advance()
1270 clk.s_rd_seq = os_atomic_load(&smr->smr_clock.s_rd_seq, relaxed); in smr_synchronize()
1272 clk.s_wr_seq = os_atomic_load(&smr->smr_clock.s_wr_seq, relaxed); in smr_synchronize()
[all …]
H A Dmpsc_queue.c48 mpsc_queue_chain_t head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch()
54 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_restore_batch()
72 tail = os_atomic_load(&q->mpqh_tail, relaxed); in mpsc_queue_dequeue_batch()
78 head = os_atomic_load(&q->mpqh_head.mpqc_next, relaxed); in mpsc_queue_dequeue_batch()
111 elm = os_atomic_load(&cur->mpqc_next, relaxed); in mpsc_queue_batch_next()
H A Dsocd_client.c87 long available = os_atomic_load(&socd_client_trace_available, relaxed); in socd_client_set_primary_kernelcache_uuid()
108 available = os_atomic_load(&socd_client_trace_available, dependency); in socd_client_trace()
H A Dast.h190 #define thread_ast_peek(act, reason) (os_atomic_load(&(act)->ast, relaxed) & (reason))
191 #define thread_ast_get(act) os_atomic_load(&(act)->ast, relaxed)
H A Dlocks_internal.h97 __auto_type __v = os_atomic_load(p, relaxed); \
104 (os_atomic_load(p, relaxed) == (e))
139 #define lock_load_exclusive(p, m) os_atomic_load(p, relaxed)
H A Dlock_mtx.c716 prev == 0 || (os_atomic_load(astp, relaxed) & AST_URGENT))) { in lck_mtx_lock_adaptive_spin()
737 deadline = ml_get_timebase() + os_atomic_load(&MutexSpin, relaxed); in lck_mtx_lock_adaptive_spin()
769 (os_atomic_load(astp, relaxed) & AST_URGENT) || in lck_mtx_lock_adaptive_spin()
824 state = os_atomic_load(&lock->lck_mtx, relaxed); in lck_mtx_lock_contended()
900 state = os_atomic_load(&lock->lck_mtx, relaxed); in lck_mtx_lock_contended()
1206 data = os_atomic_load(&lock->lck_mtx.data, compiler_acq_rel); in lck_mtx_unlock_contended()
1286 lck_mtx_state_t state = os_atomic_load(&lock->lck_mtx, relaxed); in lck_mtx_assert()
1310 lck_mtx_state_t state = os_atomic_load(&lock->lck_mtx, relaxed); in lck_mtx_convert_spin()
1340 lck_mtx_state_t state = os_atomic_load(&lck->lck_mtx, relaxed); in kdp_lck_mtx_lock_spin_is_acquired()
1358 lck_mtx_state_t state = os_atomic_load(&mutex->lck_mtx, relaxed); in kdp_lck_mtx_find_owner()
H A Dremote_time.c66 if (!os_atomic_load(&bt_init_flag, acquire)) { in mach_bridge_timer_maintenance()
89 assert(os_atomic_load(&bt_init_flag, relaxed)); in mach_bridge_timer_enable()
182 if (os_atomic_load(&bt_init_flag, acquire)) { in bt_params_get_latest()
505 if (!os_atomic_load(&bt_init_flag, acquire)) { in mach_bridge_remote_time()
H A Dexclaves_inspection.c392 assert(!os_atomic_load(&exclaves_inspection_initialized, relaxed)); in exclaves_inspection_init()
492 return os_atomic_load(&exclaves_inspection_initialized, acquire); in exclaves_inspection_is_initialized()
514 …assert((os_atomic_load(&thread->th_exclaves_inspection_state, relaxed) & TH_EXCLAVES_INSPECTION_NO… in exclaves_inspection_check_ast()
521 …while ((os_atomic_load(&thread->th_exclaves_inspection_state, relaxed) & TH_EXCLAVES_INSPECTION_ST… in exclaves_inspection_check_ast()
528 …if ((os_atomic_load(&thread->th_exclaves_inspection_state, relaxed) & TH_EXCLAVES_INSPECTION_KPERF… in exclaves_inspection_check_ast()
H A Drestartable.c390 trrs.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_reset_pcs_ack_IPI()
408 state.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_rr_wait_if_needed()
480 state.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_rr_wait_if_needed()
518 state.trr_value = os_atomic_load(&thread->t_rr_state.trr_value, relaxed); in thread_reset_pcs_ast()
H A Dexclaves_resource.c677 if (os_atomic_load(&resource->r_usecnt, acquire) > 0) { in exclaves_resource_release()
756 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), >, 0); in exclaves_resource_create_port_name()
777 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), >, 1); in exclaves_resource_create_port_name()
815 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), >, 0);
871 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), >, 0); in exclaves_named_buffer_copyin()
935 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), >, 0); in exclaves_named_buffer_copyout()
1147 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), ==, 0); in exclaves_named_buffer_unmap()
1193 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), ==, 0); in exclaves_audio_buffer_delete()
1213 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), >, 0); in exclaves_audio_buffer_copyout()
1568 assert3u(os_atomic_load(&resource->r_usecnt, relaxed), ==, 0); in exclaves_resource_sensor_reset()
[all …]
/xnu-11215.1.10/san/memory/
H A Dubsan_log.c79 if (n == os_atomic_load(&ubsan_log_tail, acquire)) { in ubsan_log_append()
119 head = os_atomic_load(&ubsan_log_head, relaxed);
121 tail = os_atomic_load(&ubsan_log_tail, relaxed);
157 head = os_atomic_load(&ubsan_log_head, relaxed);
159 tail = os_atomic_load(&ubsan_log_tail, relaxed);
/xnu-11215.1.10/osfmk/arm64/
H A Dlock_ticket_pv.c51 const cpumap_t wmask = os_atomic_load(&ticket_waitmask_pv, acquire); in hw_lck_ticket_unlock_kick_pv()
63 const hw_lck_ticket_t *wlck = os_atomic_load(&ltpi->ltpi_lck, in hw_lck_ticket_unlock_kick_pv()
69 const uint8_t wt = os_atomic_load(&ltpi->ltpi_wt, acquire); in hw_lck_ticket_unlock_kick_pv()
123 const uint8_t cticket = os_atomic_load(&lck->cticket, acquire); in hw_lck_ticket_lock_wait_pv()
H A Dmachine_remote_time.c50 if (os_atomic_load(&bt_init_flag, relaxed)) { in mach_bridge_init_timestamp()
69 if (!os_atomic_load(&bt_init_flag, acquire)) { in mach_bridge_recv_timestamps()
91 if (!os_atomic_load(&bt_init_flag, acquire)) { in mach_bridge_set_params()
/xnu-11215.1.10/osfmk/i386/
H A Dlock_ticket_pv.c72 const cpumask_t wmask = os_atomic_load(&ticket_waitmask_pv, acquire); in hw_lck_ticket_unlock_kick_pv()
84 const hw_lck_ticket_t *wlck = os_atomic_load(&ltpi->ltpi_lck, in hw_lck_ticket_unlock_kick_pv()
90 const uint8_t wt = os_atomic_load(&ltpi->ltpi_wt, acquire); in hw_lck_ticket_unlock_kick_pv()
144 const uint8_t cticket = os_atomic_load(&lck->cticket, acquire); in hw_lck_ticket_lock_wait_pv()
/xnu-11215.1.10/bsd/kern/
H A Dkern_memorystatus_policy.c89 status->msh_phantom_cache_pressure = os_atomic_load(&memorystatus_phantom_cache_pressure, acquire); in memorystatus_health_check()
98 …status->msh_compressor_is_low_on_space = os_atomic_load(&memorystatus_compressor_space_shortage, a… in memorystatus_health_check()
99 status->msh_pageout_starved = os_atomic_load(&memorystatus_pageout_starved, acquire); in memorystatus_health_check()
105 status->msh_zone_map_is_exhausted = os_atomic_load(&memorystatus_zone_map_is_exhausted, relaxed); in memorystatus_health_check()
173 …ble_compressor_segments_over_limit && !vm_swapout_thread_running && !os_atomic_load(&vm_swapout_wa… in memorystatus_pick_action()
185 vm_swapout_thread_running, os_atomic_load(&vm_swapout_wake_pending, relaxed)); in memorystatus_pick_action()
H A Dcounter_test.c121 if (!os_atomic_load(&scalable_counter_test_running, seq_cst)) {
139 if (!os_atomic_load(&scalable_counter_test_running, seq_cst)) {
156 if (!os_atomic_load(&scalable_counter_test_running, seq_cst)) {
171 if (!os_atomic_load(&scalable_counter_test_running, seq_cst)) {
H A Dkern_codesigning.c136 if (os_atomic_load(&cs_config_set, acquire) == true) { in code_signing_configuration()
407 assert(os_atomic_load(&cs_config_set, relaxed) == true); in code_signing_configuration()
411 *monitor_type_out = os_atomic_load(&cs_monitor, relaxed); in code_signing_configuration()
416 *config_out = os_atomic_load(&cs_config, relaxed); in code_signing_configuration()
428 if (os_atomic_load(&cs_config_set, acquire) == false) { in disable_code_signing_feature()
503 return os_atomic_load(developer_mode_enabled, relaxed); in developer_mode_state()
/xnu-11215.1.10/osfmk/arm64/sptm/pmap/
H A Dpmap_data.h293 assertf(os_atomic_load(&pv_head_table[index], relaxed) & PVH_LOCK_FLAGS, in pvh_assert_locked()
295 &pv_head_table[index], (void*)(os_atomic_load(&pv_head_table[index], relaxed)), index); in pvh_assert_locked()
346 locked_pvh.pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_lock()
380 …const locked_pvh_t locked_pvh = {.pvh = os_atomic_load(&pv_head_table[index], relaxed), .pai = ind… in pvh_lock_nopreempt()
404 locked_pvh.pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_try_lock()
451 const uintptr_t old_pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_lock_enter_sleep_mode()
501 const uintptr_t old_pvh = os_atomic_load(&pv_head_table[index], relaxed); in pvh_unlock()
2080 os_atomic_load(&pmap_pcpu->retype_epoch.local_seq, relaxed); in pmap_retype_epoch_prepare_drain()
/xnu-11215.1.10/osfmk/tests/
H A Dvfp_state_test.c71 while (os_atomic_load(var, acquire) != num) { in wait_threads()
73 if (os_atomic_load(var, acquire) != num) { in wait_threads()
/xnu-11215.1.10/osfmk/arm/
H A Dpreemption_disable.c323 const uint64_t max_duration = os_atomic_load(&pcpu->pdp_max_mach_duration, relaxed); in _preemption_disable_snap_end()
391 uint64_t const threshold = os_atomic_load(&sched_preemption_disable_threshold_mt, relaxed); in _collect_preemption_disable_measurement()
502 durations[cpu++] = os_atomic_load(&pcpu->pdp_max_mach_duration, relaxed); in preemption_disable_get_max_durations()
/xnu-11215.1.10/tests/
H A Dos_refcnt.c430 while (os_atomic_load(&pcpu_perf_step, relaxed) == 0) { in worker_ref()
433 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in worker_ref()
447 while (os_atomic_load(&pcpu_perf_step, relaxed) == 0) { in worker_pcpu_ref()
450 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in worker_pcpu_ref()
468 while (os_atomic_load(&pcpu_perf_step, relaxed) == 1) { in warmup_thread_pool()

123456