| /xnu-11215.61.5/iokit/DriverKit/ |
| H A D | queue_implementation.h | 564 queue_entry_t _tmp_entry = queue_first((head)); \ 652 #define queue_first(q) ((q)->next) macro 704 #define queue_empty(q) queue_end((q), queue_first(q)) 1108 for ((elt) = (type)(void *) queue_first(head); \
|
| /xnu-11215.61.5/osfmk/kern/ |
| H A D | queue.h | 564 queue_entry_t _tmp_entry = queue_first((head)); \ 652 #define queue_first(q) ((q)->next) macro 704 #define queue_empty(q) queue_end((q), queue_first(q)) 1108 for ((elt) = (type)(void *) queue_first(head); \
|
| H A D | bsd_kern.c | 248 thread = (thread_t)(void *)queue_first(&task->threads); in get_firstthread() 277 for (inc = (thread_t)(void *)queue_first(&task->threads); in get_signalact() 327 for (inc = (thread_t)(void *)queue_first(&task->threads); in check_actforsig() 1043 for (inc = (thread_t)(void *)queue_first(&task->threads); in task_act_iterate_wth_args() 1134 for (thact = (thread_t)(void *)queue_first(&task->threads); in fill_taskthreadinfo() 1183 for (thact = (thread_t)(void *)queue_first(&task->threads); in fill_taskthreadlist()
|
| H A D | sched_prim.c | 205 if (iter == queue_first(queue)) { in check_rt_runq_consistency() 4441 if (iter == queue_first(queue)) { in rt_runq_enqueue()
|
| /xnu-11215.61.5/osfmk/vm/ |
| H A D | vm_compressor_backing_store.c | 216 swf = (struct swapfile*) queue_first(&swf_global_queue); in vm_swapfile_for_handle() 677 c_seg = (c_segment_t)queue_first(&c_swappedout_sparse_list_head); in vm_swap_defragment() 1227 c_seg = (c_segment_t)queue_first(swapout_list_head); in vm_swapout_thread() 1527 swf = (struct swapfile*) queue_first(&swf_global_queue); in vm_swap_create_file() 1736 swf = (struct swapfile*) queue_first(&swf_global_queue); in vm_swap_put() 2022 swf = (struct swapfile*) queue_first(&swf_global_queue); in vm_swap_handle_delayed_trims() 2145 swf = (struct swapfile*) queue_first(&swf_global_queue); in vm_swap_reclaim() 2467 swf = (struct swapfile*) queue_first(&swf_global_queue); in vm_swap_max_budget()
|
| H A D | vm_purgeable.c | 695 for (object = (vm_object_t) queue_first(&queue->objq[group]); in vm_purgeable_object_find_and_lock() 1139 for (object = (vm_object_t) queue_first(&queue->objq[group]); in vm_purgeable_stats_helper() 1189 for (object = (vm_object_t) queue_first(&queue->objq[group]); in vm_purgeable_account_volatile_queue() 1230 for (object = (vm_object_t) queue_first(nonvolatile_q); in vm_purgeable_account() 1286 for (object = (vm_object_t) queue_first(&queue->objq[group]); in vm_purgeable_queue_purge_task_owned()
|
| H A D | analytics.c | 175 c_segment_t c_seg = (c_segment_t) queue_first(c_queues[q]); in report_compressor_age()
|
| H A D | vm_compressor.c | 602 c_seg = (c_segment_t) queue_first(&c_age_list_head); in vm_compressor_is_thrashing() 1698 for (c_seg = (c_segment_t) queue_first(&owner_task->task_frozen_cseg_q); in task_disown_frozen_csegs() 1910 c_first = (c_segment_t)queue_first(&c_age_list_head); in c_seg_switch_state() 2687 c_seg = (c_segment_t) queue_first(&c_age_list_head); in vm_compressor_swapout_is_ripe() 3021 c_seg = (c_segment_t)queue_first(&c_minor_list_head); in vm_compressor_do_delayed_compactions() 3247 c_seg = (c_segment_t)queue_first(&c_major_list_head); in vm_compressor_process_major_segments() 3335 c_seg = (c_segment_t)queue_first(special_swappedin_list_head); in vm_compressor_process_special_swapped_in_segments_locked() 3438 c_seg = (c_segment_t)queue_first(&c_regular_swappedin_list_head); in vm_compressor_process_regular_swapped_in_segments() 3511 c_seg = (c_segment_t)queue_first(&c_major_list_head); in vm_compressor_flush() 3734 c_seg = (c_segment_t) queue_first(c_queue); in do_fastwake_warmup() [all …]
|
| H A D | vm_page.h | 786 #define vm_page_queue_first queue_first
|
| H A D | vm_object.c | 1049 next_obj = (vm_object_t)queue_first(&vm_object_cached_list); in vm_object_cache_evict() 9140 for (object = (vm_object_t) queue_first(&task->task_objq); 9182 queue_first(&task->task_objq),
|
| H A D | vm_resident.c | 9837 for (object = (vm_object_t) queue_first(&queue->objq[group]); in vm_page_iterate_purgeable_objects()
|
| /xnu-11215.61.5/iokit/Kernel/ |
| H A D | IOPlatformActions.cpp | 353 next = (typeof(entry))queue_first(&gActionQueues[qidx]); in IORemoveServicePlatformActions()
|
| H A D | IOLib.cpp | 1365 __IGNORE_WCASTALIGN(pa = (typeof(pa))queue_first(&a->list)); in iopa_alloc()
|
| H A D | IOServicePM.cpp | 9105 request = (typeof(request))queue_first(requestQueue); in checkRequestQueue() 9175 __IGNORE_WCASTALIGN(entry = (typeof(entry))queue_first(&fWorkQueue)); in checkForWork() 9309 request = (typeof(request))queue_first(&fQueue); in checkForWork()
|
| H A D | IOUserClient.cpp | 1718 owner = (IOUserClientOwner *)(void *) queue_first(&owners); in noMoreSenders() 1830 owner = (IOUserClientOwner *)(void *) queue_first(taskque); in iokit_task_terminate_phase2()
|
| H A D | IOService.cpp | 335 for (queue_entry_t elt = queue_first(que); \
|
| /xnu-11215.61.5/osfmk/i386/AT386/ |
| H A D | model_dep.c | 1647 for (j = 0, thread = (thread_t) queue_first(&task->threads); j < task->thread_count; in print_tasks_user_threads() 1666 for (j = 0, thread = (thread_t) queue_first(&task->threads); j < task->thread_count; in print_thread_num_that_crashed()
|
| /xnu-11215.61.5/osfmk/arm/ |
| H A D | model_dep.c | 607 for (thread = (thread_t)queue_first(&threads); in do_print_all_backtraces()
|
| /xnu-11215.61.5/osfmk/ipc/ |
| H A D | ipc_importance.c | 859 task_imp = (ipc_importance_task_t)queue_first(queue); in ipc_importance_task_process_updates() 1002 task_imp = (ipc_importance_task_t)queue_first(&ipc_importance_delayed_drop_queue); in ipc_importance_task_delayed_drop_scan()
|
| /xnu-11215.61.5/osfmk/arm/pmap/ |
| H A D | pmap_data.c | 775 pt_desc_t *ptdp = (pt_desc_t *)queue_first(&pt_page_list); in ppr_find_eligible_pt_page()
|