Searched refs:os_atomic_thread_fence (Results 1 – 21 of 21) sorted by relevance
86 os_atomic_thread_fence(release); in ubsan_log_append()120 os_atomic_thread_fence(seq_cst);158 os_atomic_thread_fence(seq_cst);
148 …os_atomic_thread_fence(seq_cst); // tail increment can not be seen before the entry is cleared in … in mach_vm_reclaim_mark_free()183 os_atomic_thread_fence(seq_cst); // Our write to tail must happen before our read of busy in mach_vm_reclaim_mark_used()195 …os_atomic_thread_fence(seq_cst); // tail increment can not be seen before the entry is cleared in … in mach_vm_reclaim_mark_used()
23 + os_atomic_thread_fence(new_m)
86 + os_atomic_thread_fence(seq_cst)
337 os_atomic_thread_fence(seq_cst); /* [R1] */ in __smr_enter()381 os_atomic_thread_fence(seq_cst); /* [S3] */ in __smr_rd_advance()432 os_atomic_thread_fence(seq_cst); /* [S2] */ in __smr_scan()539 os_atomic_thread_fence(seq_cst); in smr_deferred_advance()582 os_atomic_thread_fence(seq_cst); in smr_synchronize()
404 os_atomic_thread_fence(acquire); in hw_lck_ptr_wait_for_value_contended()
298 os_atomic_thread_fence(acquire); in kern_apfs_reflock_try_put_ref()
644 os_atomic_thread_fence(release); in task_restartable_ranges_synchronize()
170 os_atomic_thread_fence(acquire); in _seqlock_shared_lock_slowpath()200 os_atomic_thread_fence(release); in _seqlock_excl_commit()
167 os_atomic_thread_fence(release); \
577 os_atomic_thread_fence(acquire); in hw_lck_ticket_contended()
7093 os_atomic_thread_fence(acquire); in sched_is_in_sleep()
219 os_atomic_thread_fence(acquire); in cpu_handle_xcall()228 os_atomic_thread_fence(acq_rel); in cpu_handle_xcall()239 os_atomic_thread_fence(acq_rel); in cpu_handle_xcall()
1155 os_atomic_thread_fence(release); in ml_set_is_quiescing()1161 os_atomic_thread_fence(acquire); in ml_is_quiescing()
227 #define os_atomic_thread_fence(m) ({ \ macro
359 os_atomic_thread_fence(release); in checkForPageUnload()
532 os_atomic_thread_fence(seq_cst); in reclaim_chunk()
664 os_atomic_thread_fence(release); in compressor_pager_slot_lookup()
154 `os_atomic_thread_fence(mem_order)` provides a memory barrier
2982 os_atomic_thread_fence(release); in pmap_create_options_internal()3199 os_atomic_thread_fence(seq_cst); in pmap_destroy_internal()3741 os_atomic_thread_fence(acquire);3761 os_atomic_thread_fence(acquire);4610 os_atomic_thread_fence(release);9274 os_atomic_thread_fence(seq_cst);
4935 os_atomic_thread_fence(seq_cst); in ubc_cs_blob_add()5186 os_atomic_thread_fence(seq_cst); // Fence to prevent reordering here in ubc_cs_blob_add_supplement()