| /xnu-10063.141.1/osfmk/kern/ |
| H A D | queue.h | 209 typedef struct queue_entry *queue_entry_t; typedef 215 extern void __queue_element_linkage_invalid(queue_entry_t e); 221 __QUEUE_ELT_VALIDATE(queue_entry_t elt) in __QUEUE_ELT_VALIDATE() 229 __DEQUEUE_ELT_CLEANUP(queue_entry_t elt) in __DEQUEUE_ELT_CLEANUP() 231 elt->next = elt->prev = (queue_entry_t)NULL; in __DEQUEUE_ELT_CLEANUP() 252 queue_entry_t elt) in enqueue_head() 254 queue_entry_t old_head; in enqueue_head() 256 __QUEUE_ELT_VALIDATE((queue_entry_t)que); in enqueue_head() 267 queue_entry_t elt) in enqueue_tail() 269 queue_entry_t old_tail; in enqueue_tail() [all …]
|
| H A D | circle_queue.h | 46 queue_entry_t head; 55 static inline queue_entry_t 61 static inline queue_entry_t 64 queue_entry_t elt = circle_queue_first(cq); in circle_queue_last() 72 static inline queue_entry_t 73 circle_queue_next(circle_queue_t cq, queue_entry_t elt) in circle_queue_next() 81 queue_entry_t elt = circle_queue_first(cq); in circle_queue_length() 92 circle_enqueue_tail(circle_queue_t cq, queue_entry_t elt) in circle_enqueue_tail() 94 queue_entry_t head = circle_queue_first(cq); in circle_enqueue_tail() 95 queue_entry_t tail = circle_queue_last(cq); in circle_enqueue_tail() [all …]
|
| H A D | sched_grrr.c | 383 enqueue_tail(&tqueue, (queue_entry_t)thread); in sched_grrr_processor_queue_shutdown() 385 enqueue_tail(&bqueue, (queue_entry_t)thread); in sched_grrr_processor_queue_shutdown() 607 group->current_client = (thread_t)(void *)queue_next((queue_entry_t)thread); in grrr_intragroup_schedule() 608 if (queue_end(&group->clients, (queue_entry_t)group->current_client)) { in grrr_intragroup_schedule() 641 if (queue_end(&rq->sorted_group_list, queue_next((queue_entry_t)group))) { in grrr_intergroup_schedule() 645 grrr_group_t nextgroup = (grrr_group_t)queue_next((queue_entry_t)group); in grrr_intergroup_schedule() 731 enqueue_tail(&group->clients, (queue_entry_t)thread); in grrr_enqueue() 738 queue_first(&group->clients) == (queue_entry_t)group->current_client) { in grrr_enqueue() 739 enqueue_head(&group->clients, (queue_entry_t)thread); in grrr_enqueue() 741 insque((queue_entry_t)thread, queue_prev((queue_entry_t)group->current_client)); in grrr_enqueue() [all …]
|
| H A D | host_notify.c | 133 remqueue((queue_entry_t)entry); in host_notify_cancel() 148 queue_entry_t e; in host_notify_all()
|
| H A D | exclaves_inspection.h | 66 exclaves_inspection_queue_add(queue_t queue, queue_entry_t elm) in exclaves_inspection_queue_add()
|
| H A D | mpsc_queue.c | 39 __queue_element_linkage_invalid(queue_entry_t elt) in __queue_element_linkage_invalid() 41 queue_entry_t prev = elt->prev; in __queue_element_linkage_invalid() 42 queue_entry_t next = elt->next; in __queue_element_linkage_invalid()
|
| H A D | bsd_kern.c | 251 if (queue_end(&task->threads, (queue_entry_t)thread)) { in get_firstthread() 279 !queue_end(&task->threads, (queue_entry_t)inc);) { in get_signalact() 329 !queue_end(&task->threads, (queue_entry_t)inc);) { in check_actforsig() 996 !queue_end(&task->threads, (queue_entry_t)inc);) { in task_act_iterate_wth_args_locked() 1096 !queue_end(&task->threads, (queue_entry_t)thact);) { in fill_taskthreadinfo() 1145 !queue_end(&task->threads, (queue_entry_t)thact);) { in fill_taskthreadlist()
|
| H A D | sched_proto.c | 358 thread = (thread_t)queue_next((queue_entry_t)thread); in sched_proto_choose_thread()
|
| H A D | waitq.c | 1830 queue_entry_t elt; in waitq_set_unlink_all_locked() 2028 queue_entry_t elt; in waitq_set_first_prepost()
|
| /xnu-10063.141.1/osfmk/i386/ |
| H A D | pmap_common.c | 349 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_free_prime() 364 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_free_prime() 416 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_replenish() 435 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_replenish()
|
| H A D | pmap_internal.h | 355 pvh_et->qlink.next = (queue_entry_t)pv_hashed_free_list; in PV_HASHED_FREE_LIST() 388 pvh_et->qlink.next = (queue_entry_t)pv_hashed_kern_free_list; in PV_HASHED_KERN_FREE_LIST()
|
| H A D | mp.c | 2001 enqueue_head(&cpu_warm_call_list, (queue_entry_t)&cpu_warm_call_arr[i]); in cpu_prewarm_init() 2029 enqueue_head(&cpu_warm_call_list, (queue_entry_t)call); in free_warm_timer_call()
|
| H A D | pmap_x86_common.c | 1488 pvh_e->qlink.next = (queue_entry_t) pvh_eh; in pmap_remove_range_options() 1941 pvh_e->qlink.next = (queue_entry_t) pvh_eh; in pmap_page_protect_options() 1982 pvh_e->qlink.next = (queue_entry_t) pvh_eh; in pmap_page_protect_options()
|
| /xnu-10063.141.1/bsd/kern/ |
| H A D | kern_malloc.c | 195 enqueue_tail(&OSMalloc_tag_list, (queue_entry_t)OSMTag); in OSMalloc_Tagalloc_external() 227 (void)remque((queue_entry_t)tag); in OSMalloc_Tagrele() 247 (void)remque((queue_entry_t)tag); in OSMalloc_Tagfree_external()
|
| /xnu-10063.141.1/osfmk/vm/ |
| H A D | vm_compressor_backing_store.c | 219 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swapfile_for_handle() 227 if (queue_end(&swf_global_queue, (queue_entry_t) swf)) { in vm_swapfile_for_handle() 1517 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_create_file() 1727 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_put() 1768 assert(queue_end(&swf_global_queue, (queue_entry_t) swf)); in vm_swap_put() 2014 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_handle_delayed_trims() 2139 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_reclaim() 2460 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_max_budget()
|
| H A D | vm_purgeable.c | 708 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_object_find_and_lock() 1159 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_stats_helper() 1209 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_account_volatile_queue() 1250 !queue_end(nonvolatile_q, (queue_entry_t) object); in vm_purgeable_account() 1306 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_queue_purge_task_owned()
|
| H A D | vm_swapfile_pager.c | 713 (queue_entry_t) pager2)) { in swapfile_pager_create() 779 (queue_entry_t) pager)) { in swapfile_pager_setup()
|
| H A D | analytics.c | 175 while (!queue_end(c_queues[q], (queue_entry_t) c_seg)) { in report_compressor_age()
|
| H A D | vm_apple_protect.c | 1113 (queue_entry_t) pager2)) { in apple_protect_pager_create() 1283 (queue_entry_t) pager)) { in apple_protect_pager_setup() 1385 (queue_entry_t) pager); in apple_protect_pager_trim()
|
| H A D | vm_page.h | 112 #define vm_page_queue_entry_t queue_entry_t 1036 #define VM_PAGE_CONVERT_TO_QUEUE_ENTRY(p) ((queue_entry_t)(p))
|
| H A D | vm_fourk_pager.c | 584 (queue_entry_t) pager); in fourk_pager_trim()
|
| H A D | vm_shared_region_pager.c | 1343 !queue_end(&shared_region_pager_queue, (queue_entry_t) pager); in shared_region_pager_trim()
|
| H A D | vm_compressor.c | 1379 if (queue_end(qhead, (queue_entry_t) c_seg_next)) { in c_seg_insert_into_q() 1582 !queue_end(&owner_task->task_frozen_cseg_q, (queue_entry_t) c_seg); in task_disown_frozen_csegs() 2968 if (queue_end(list_head, (queue_entry_t)c_seg_next)) { in vm_compressor_major_compact_cseg() 3402 while (!queue_end(&c_major_list_head, (queue_entry_t)c_seg)) { in vm_compressor_flush()
|
| /xnu-10063.141.1/iokit/Kernel/ |
| H A D | IOKitDebug.cpp | 609 if (queue_end(que, (queue_entry_t) site)) { in IOTrackingAdd() 652 if (queue_end(&site->instances, (queue_entry_t)site->addresses[hashIdx])) { in IOTrackingAdd()
|
| /xnu-10063.141.1/osfmk/arm/ |
| H A D | model_dep.c | 597 PANIC_VALIDATE_PTR(thread) && !queue_end(&threads, (queue_entry_t)thread); in do_print_all_backtraces()
|