Searched refs:lq (Results 1 – 5 of 5) sorted by relevance
| /xnu-12377.81.4/libkern/os/ |
| H A D | log_queue.c | 164 log_queue_increment_mem_avail(const log_queue_t lq, size_t idx, size_t size) in log_queue_increment_mem_avail() argument 166 lq->lq_cnt_mem_avail += size; in log_queue_increment_mem_avail() 168 lq->lq_cnt_mem_meta_avail += size; in log_queue_increment_mem_avail() 173 log_queue_decrement_mem_avail(const log_queue_t lq, size_t idx, size_t size) in log_queue_decrement_mem_avail() argument 175 lq->lq_cnt_mem_avail -= size; in log_queue_decrement_mem_avail() 177 lq->lq_cnt_mem_meta_avail -= size; in log_queue_decrement_mem_avail() 200 log_queue_entry_alloc(log_queue_t lq, size_t lqe_size, firehose_stream_t stream_type) in log_queue_entry_alloc() argument 209 if (!LQ_MEM_ENABLED(lq, i)) { in log_queue_entry_alloc() 212 log_queue_entry_t lqe = logmem_alloc(&lq->lq_mem[i], &lqe_size); in log_queue_entry_alloc() 214 assert(lqe_size <= lq->lq_cnt_mem_avail); in log_queue_entry_alloc() [all …]
|
| /xnu-12377.81.4/tools/lldbmacros/ |
| H A D | log.py | 305 def __init__(self, cpu, lq): argument 307 self._lq = lq 480 def show_log_queue(lq, enqueued, dispatched): argument 482 'lq_mem_state_t', lq._lq.lq_mem_state, 'LQ_MEM_STATE_') 484 'lq_req_state_t', lq._lq.lq_req_state, 'LQ_REQ_STATE_') 485 cpu = "N/A" if lq.cpu is None else str(lq.cpu) 486 pp.print(hdr_lq.format(lq=lq, state=state, size=lq.size, reconf=reconf, cpu=cpu, 491 for lq in log_queues: 492 lq_enqueued = list(lq.enqueued) 493 lq_dispatched = list(lq.dispatched) [all …]
|
| /xnu-12377.81.4/osfmk/vm/ |
| H A D | vm_resident.c | 2059 zpercpu_foreach(lq, t_local_q) { in vm_page_init_local_q() 2060 VPL_LOCK_INIT(lq, &vm_page_lck_grp_local, &vm_page_lck_attr); in vm_page_init_local_q() 2061 vm_page_queue_init(&lq->vpl_queue); in vm_page_init_local_q() 6803 struct vpl *lq; in vm_page_reactivate_local() local 6813 lq = zpercpu_get_cpu(vm_page_local_q, lid); in vm_page_reactivate_local() 6816 if (lq->vpl_count < vm_page_local_q_hard_limit && force == FALSE) { in vm_page_reactivate_local() 6824 VPL_LOCK(&lq->vpl_lock); in vm_page_reactivate_local() 6826 if (lq->vpl_count) { in vm_page_reactivate_local() 6830 assert(!vm_page_queue_empty(&lq->vpl_queue)); in vm_page_reactivate_local() 6832 vm_page_queue_iterate(&lq->vpl_queue, m, vmp_pageq) { in vm_page_reactivate_local() [all …]
|
| H A D | vm_fault.c | 3332 struct vpl *lq; in vm_fault_enqueue_page() local 3356 lq = zpercpu_get_cpu(vm_page_local_q, lid); in vm_fault_enqueue_page() 3358 VPL_LOCK(&lq->vpl_lock); in vm_fault_enqueue_page() 3361 vm_page_queue_enter(&lq->vpl_queue, m, vmp_pageq); in vm_fault_enqueue_page() 3364 lq->vpl_count++; in vm_fault_enqueue_page() 3367 lq->vpl_internal_count++; in vm_fault_enqueue_page() 3369 lq->vpl_external_count++; in vm_fault_enqueue_page() 3372 VPL_UNLOCK(&lq->vpl_lock); in vm_fault_enqueue_page() 3374 if (lq->vpl_count > vm_page_local_q_soft_limit) { in vm_fault_enqueue_page()
|
| /xnu-12377.81.4/osfmk/kern/ |
| H A D | host.c | 461 zpercpu_foreach(lq, vm_page_local_q) { in host_statistics() 462 stat32->active_count += VM_STATISTICS_TRUNCATE_TO_32_BIT(lq->vpl_count); in host_statistics() 826 zpercpu_foreach(lq, vm_page_local_q) { in vm_stats() 827 stat->active_count += lq->vpl_count; in vm_stats() 828 local_q_internal_count += lq->vpl_internal_count; in vm_stats() 829 local_q_external_count += lq->vpl_external_count; in vm_stats()
|