Searched refs:os_atomic_thread_fence (Results 1 – 20 of 20) sorted by relevance
86 os_atomic_thread_fence(release); in ubsan_log_append()120 os_atomic_thread_fence(seq_cst);158 os_atomic_thread_fence(seq_cst);
148 …os_atomic_thread_fence(seq_cst); // tail increment can not be seen before the entry is cleared in … in mach_vm_reclaim_mark_free()183 os_atomic_thread_fence(seq_cst); // Our write to tail must happen before our read of busy in mach_vm_reclaim_mark_used()195 …os_atomic_thread_fence(seq_cst); // tail increment can not be seen before the entry is cleared in … in mach_vm_reclaim_mark_used()
339 os_atomic_thread_fence(seq_cst); /* [R1] */ in __smr_enter()395 os_atomic_thread_fence(seq_cst); /* [S3] */ in __smr_rd_advance()437 os_atomic_thread_fence(seq_cst); /* [S2] */ in __smr_scan()561 os_atomic_thread_fence(seq_cst); in smr_deferred_advance_nopreempt()
227 os_atomic_thread_fence(release); \
298 os_atomic_thread_fence(acquire); in kern_apfs_reflock_try_put_ref()
644 os_atomic_thread_fence(release); in task_restartable_ranges_synchronize()
165 os_atomic_thread_fence(acquire); in _seqlock_shared_lock_slowpath()195 os_atomic_thread_fence(release); in _seqlock_excl_commit()
577 os_atomic_thread_fence(acquire); in hw_lck_ticket_contended()
6994 os_atomic_thread_fence(acquire); in sched_is_in_sleep()
23 + os_atomic_thread_fence(new_m)
86 + os_atomic_thread_fence(seq_cst)
219 os_atomic_thread_fence(acquire); in cpu_handle_xcall()228 os_atomic_thread_fence(acq_rel); in cpu_handle_xcall()239 os_atomic_thread_fence(acq_rel); in cpu_handle_xcall()
1143 os_atomic_thread_fence(release); in ml_set_is_quiescing()1149 os_atomic_thread_fence(acquire); in ml_is_quiescing()
227 #define os_atomic_thread_fence(m) ({ \ macro
536 os_atomic_thread_fence(seq_cst); in reclaim_chunk()
664 os_atomic_thread_fence(release); in compressor_pager_slot_lookup()
154 `os_atomic_thread_fence(mem_order)` provides a memory barrier
2954 os_atomic_thread_fence(release); in pmap_create_options_internal()3181 os_atomic_thread_fence(seq_cst); in pmap_destroy_internal()3723 os_atomic_thread_fence(acquire);3743 os_atomic_thread_fence(acquire);4592 os_atomic_thread_fence(release);9214 os_atomic_thread_fence(seq_cst);
4498 …os_atomic_thread_fence(seq_cst); // fence to make sure all of the writes happen before we update t… in ubc_cs_blob_add()4737 os_atomic_thread_fence(seq_cst); // Fence to prevent reordering here in ubc_cs_blob_add_supplement()
4970 os_atomic_thread_fence(release); in proc_ro_ref_proc()