Home
last modified time | relevance | path

Searched refs:mcs (Results 1 – 4 of 4) sorted by relevance

/xnu-8796.121.2/osfmk/kern/
H A Dlock_mtx.c132 lck_mtx_mcs_clear(lck_mtx_mcs_t mcs) in lck_mtx_mcs_clear() argument
134 *mcs = (struct lck_mtx_mcs){ }; in lck_mtx_mcs_clear()
333 lck_mtx_mcs_t mcs, in lck_mtx_ilk_lock_cleanup_as_mcs() argument
338 lck_mcs_id_t pidx = (lck_mcs_id_t)mcs->lmm_as_prev; in lck_mtx_ilk_lock_cleanup_as_mcs()
367 mcs, nnode, relaxed); in lck_mtx_ilk_lock_cleanup_as_mcs()
382 while (!hw_spin_wait_until(&mcs->lmm_as_next, nnode, nnode)) { in lck_mtx_ilk_lock_cleanup_as_mcs()
411 lck_mtx_mcs_t mcs, nnode, pnode; in lck_mtx_ilk_lock_contended() local
423 mcs = &lck_mcs_get_current()->mcs_mtx; in lck_mtx_ilk_lock_contended()
428 mcs->lmm_ilk_current = lock; in lck_mtx_ilk_lock_contended()
432 os_atomic_store(&pnode->lmm_ilk_next, mcs, relaxed); in lck_mtx_ilk_lock_contended()
[all …]
H A Dlocks_internal.h321 lck_mcs_t mcs = txn->txn_mcs; in lck_spin_txn_end() local
326 assert(mcs->mcs_spin_rsv == slot + 1); in lck_spin_txn_end()
327 os_atomic_store(&mcs->mcs_spin_rsv, slot, compiler_acq_rel); in lck_spin_txn_end()
/xnu-8796.121.2/tools/lldbmacros/
H A Dmbufs.py30 mcs = Cast(mbuf.mtbl_stats, 'mb_class_stat_t *')
41 print(entry_format.format(mcs.mbcl_cname, mcs.mbcl_total, total,
42 mcs.mbcl_infree, mcs.mbcl_slab_cnt,
43 (mcs.mbcl_total - total - mcs.mbcl_infree),
44 mcs.mbcl_fail_cnt, mbuf.mtbl_cache.mc_waiter_cnt,
45 mcs.mbcl_notified, mcs.mbcl_purge_cnt,
H A Dmemory.py2325 mcs = kern.PERCPU_GET('lck_mcs', cpu).mcs_mtx
2327 if unsigned(mcs.lmm_ilk_current) != unsigned(mtx):
2331 if mcs.lmm_ilk_next:
2332 mcs_ilk_next_map[unsigned(mcs.lmm_ilk_next)] = cpu | 0x4000
2338 mcs = addressof(kern.PERCPU_GET('lck_mcs', idx & 0x3fff).mcs_mtx)
2339 q.append(((idx & 0x3fff), mcs))
2342 idx = unsigned(mcs.lmm_as_prev)
2347 for (cpu, mcs) in q:
2349 cpu, GetCpuDataForCpuID(cpu).cpu_active_thread, mcs)
2355 mcs = addressof(kern.PERCPU_GET('lck_mcs', idx & 0x3fff).mcs_mtx)
[all …]