| /xnu-8020.101.4/osfmk/kern/ |
| H A D | queue.h | 221 typedef struct queue_entry *queue_entry_t; typedef 239 __QUEUE_ELT_VALIDATE(queue_entry_t elt) in __QUEUE_ELT_VALIDATE() 241 queue_entry_t elt_next, elt_prev; in __QUEUE_ELT_VALIDATE() 243 if (__improbable(elt == (queue_entry_t)NULL)) { in __QUEUE_ELT_VALIDATE() 250 if (__improbable(elt_next == (queue_entry_t)NULL || elt_prev == (queue_entry_t)NULL)) { in __QUEUE_ELT_VALIDATE() 260 __DEQUEUE_ELT_CLEANUP(queue_entry_t elt) in __DEQUEUE_ELT_CLEANUP() 262 (elt)->next = (queue_entry_t)NULL; in __DEQUEUE_ELT_CLEANUP() 263 (elt)->prev = (queue_entry_t)NULL; in __DEQUEUE_ELT_CLEANUP() 273 queue_entry_t elt) in enqueue_head() 275 queue_entry_t old_head; in enqueue_head() [all …]
|
| H A D | circle_queue.h | 46 queue_entry_t head; 55 static inline queue_entry_t 61 static inline queue_entry_t 64 queue_entry_t elt = circle_queue_first(cq); in circle_queue_last() 72 static inline queue_entry_t 73 circle_queue_next(circle_queue_t cq, queue_entry_t elt) in circle_queue_next() 81 queue_entry_t elt = circle_queue_first(cq); in circle_queue_length() 91 circle_enqueue_tail(circle_queue_t cq, queue_entry_t elt) in circle_enqueue_tail() 93 queue_entry_t head = circle_queue_first(cq); in circle_enqueue_tail() 94 queue_entry_t tail = circle_queue_last(cq); in circle_enqueue_tail() [all …]
|
| H A D | sched_grrr.c | 382 enqueue_tail(&tqueue, (queue_entry_t)thread); in sched_grrr_processor_queue_shutdown() 384 enqueue_tail(&bqueue, (queue_entry_t)thread); in sched_grrr_processor_queue_shutdown() 606 group->current_client = (thread_t)(void *)queue_next((queue_entry_t)thread); in grrr_intragroup_schedule() 607 if (queue_end(&group->clients, (queue_entry_t)group->current_client)) { in grrr_intragroup_schedule() 640 if (queue_end(&rq->sorted_group_list, queue_next((queue_entry_t)group))) { in grrr_intergroup_schedule() 644 grrr_group_t nextgroup = (grrr_group_t)queue_next((queue_entry_t)group); in grrr_intergroup_schedule() 730 enqueue_tail(&group->clients, (queue_entry_t)thread); in grrr_enqueue() 737 queue_first(&group->clients) == (queue_entry_t)group->current_client) { in grrr_enqueue() 738 enqueue_head(&group->clients, (queue_entry_t)thread); in grrr_enqueue() 740 insque((queue_entry_t)thread, queue_prev((queue_entry_t)group->current_client)); in grrr_enqueue() [all …]
|
| H A D | host_notify.c | 95 enqueue_tail(&host_notify_queue[notify_type], (queue_entry_t)entry); in host_request_notification() 117 remqueue((queue_entry_t)entry); in host_notify_port_destroy()
|
| H A D | bsd_kern.c | 234 if (queue_end(&task->threads, (queue_entry_t)thread)) { in get_firstthread() 263 !queue_end(&task->threads, (queue_entry_t)inc);) { in get_signalact() 313 !queue_end(&task->threads, (queue_entry_t)inc);) { in check_actforsig() 1004 !queue_end(&task->threads, (queue_entry_t)inc);) { in task_act_iterate_wth_args() 1109 !queue_end(&task->threads, (queue_entry_t)thact);) { in fill_taskthreadinfo() 1158 !queue_end(&task->threads, (queue_entry_t)thact);) { in fill_taskthreadlist()
|
| H A D | sched_traditional.c | 586 queue_entry_t qe; 599 if (qe == (queue_entry_t)thread) {
|
| H A D | sched_proto.c | 358 thread = (thread_t)queue_next((queue_entry_t)thread); in sched_proto_choose_thread()
|
| H A D | waitq.c | 1730 queue_entry_t elt; in waitq_set_unlink_all_locked() 1928 queue_entry_t elt; in waitq_set_first_prepost()
|
| /xnu-8020.101.4/iokit/DriverKit/ |
| H A D | queue_implementation.h | 221 typedef struct queue_entry *queue_entry_t; typedef 239 __QUEUE_ELT_VALIDATE(queue_entry_t elt) in __QUEUE_ELT_VALIDATE() 241 queue_entry_t elt_next, elt_prev; in __QUEUE_ELT_VALIDATE() 243 if (__improbable(elt == (queue_entry_t)NULL)) { in __QUEUE_ELT_VALIDATE() 250 if (__improbable(elt_next == (queue_entry_t)NULL || elt_prev == (queue_entry_t)NULL)) { in __QUEUE_ELT_VALIDATE() 260 __DEQUEUE_ELT_CLEANUP(queue_entry_t elt) in __DEQUEUE_ELT_CLEANUP() 262 (elt)->next = (queue_entry_t)NULL; in __DEQUEUE_ELT_CLEANUP() 263 (elt)->prev = (queue_entry_t)NULL; in __DEQUEUE_ELT_CLEANUP() 273 queue_entry_t elt) in enqueue_head() 275 queue_entry_t old_head; in enqueue_head() [all …]
|
| /xnu-8020.101.4/osfmk/i386/ |
| H A D | pmap_common.c | 349 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_free_prime() 364 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_free_prime() 416 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_replenish() 435 pvh_e->qlink.next = (queue_entry_t)pvh_eh; in mapping_replenish()
|
| H A D | pmap_internal.h | 355 pvh_et->qlink.next = (queue_entry_t)pv_hashed_free_list; in PV_HASHED_FREE_LIST() 388 pvh_et->qlink.next = (queue_entry_t)pv_hashed_kern_free_list; in PV_HASHED_KERN_FREE_LIST()
|
| H A D | pmap_x86_common.c | 1480 pvh_e->qlink.next = (queue_entry_t) pvh_eh; in pmap_remove_range_options() 1919 pvh_e->qlink.next = (queue_entry_t) pvh_eh; in pmap_page_protect_options() 1960 pvh_e->qlink.next = (queue_entry_t) pvh_eh; in pmap_page_protect_options()
|
| H A D | mp.c | 1999 enqueue_head(&cpu_warm_call_list, (queue_entry_t)&cpu_warm_call_arr[i]); in cpu_prewarm_init() 2027 enqueue_head(&cpu_warm_call_list, (queue_entry_t)call); in free_warm_timer_call()
|
| /xnu-8020.101.4/bsd/kern/ |
| H A D | kern_malloc.c | 214 enqueue_tail(&OSMalloc_tag_list, (queue_entry_t)OSMTag); in OSMalloc_Tagalloc_external() 246 (void)remque((queue_entry_t)tag); in OSMalloc_Tagrele() 266 (void)remque((queue_entry_t)tag); in OSMalloc_Tagfree_external()
|
| /xnu-8020.101.4/osfmk/vm/ |
| H A D | vm_compressor_backing_store.c | 204 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swapfile_for_handle() 212 if (queue_end(&swf_global_queue, (queue_entry_t) swf)) { in vm_swapfile_for_handle() 1370 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_create_file() 1580 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_put() 1621 assert(queue_end(&swf_global_queue, (queue_entry_t) swf)); in vm_swap_put() 1866 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_handle_delayed_trims() 1992 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_reclaim() 2311 while (queue_end(&swf_global_queue, (queue_entry_t)swf) == FALSE) { in vm_swap_max_budget()
|
| H A D | vm_purgeable.c | 711 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_object_find_and_lock() 1162 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_stats_helper() 1212 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_account_volatile_queue() 1253 !queue_end(nonvolatile_q, (queue_entry_t) object); in vm_purgeable_account() 1309 !queue_end(&queue->objq[group], (queue_entry_t) object); in vm_purgeable_queue_purge_task_owned()
|
| H A D | vm_swapfile_pager.c | 764 (queue_entry_t) pager2)) { in swapfile_pager_create() 830 (queue_entry_t) pager)) { in swapfile_pager_setup()
|
| H A D | vm_apple_protect.c | 1146 (queue_entry_t) pager2)) { in apple_protect_pager_create() 1316 (queue_entry_t) pager)) { in apple_protect_pager_setup() 1418 (queue_entry_t) pager); in apple_protect_pager_trim()
|
| H A D | vm_page.h | 112 #define vm_page_queue_entry_t queue_entry_t 1021 #define VM_PAGE_CONVERT_TO_QUEUE_ENTRY(p) ((queue_entry_t)(p))
|
| H A D | vm_compressor.c | 1171 if (queue_end(qhead, (queue_entry_t) c_seg_next)) { in c_seg_insert_into_q() 1374 !queue_end(&owner_task->task_frozen_cseg_q, (queue_entry_t) c_seg); in task_disown_frozen_csegs() 2539 while (!queue_end(&c_major_list_head, (queue_entry_t)c_seg)) { in vm_consider_swapping() 2749 while (!queue_end(&c_major_list_head, (queue_entry_t)c_seg)) { in vm_compressor_flush() 3225 if (queue_end(&c_age_list_head, (queue_entry_t)c_seg_next)) { in vm_compressor_compact_and_swap() 3734 while (!queue_end(&c_major_list_head, (queue_entry_t)c_seg)) { in vm_compressor_process_major_segments()
|
| H A D | vm_fourk_pager.c | 617 (queue_entry_t) pager); in fourk_pager_trim()
|
| H A D | vm_shared_region_pager.c | 1375 !queue_end(&shared_region_pager_queue, (queue_entry_t) pager); in shared_region_pager_trim()
|
| H A D | vm_object.c | 1055 …while (!queue_end(&vm_object_cached_list, (queue_entry_t)next_obj) && object_cnt++ < max_objects_t… in vm_object_cache_evict() 6502 object1->cached_list.next = (queue_entry_t) object2; in vm_object_transpose() 6503 object2->cached_list.next = (queue_entry_t) object1; in vm_object_transpose() 8598 !queue_end(&task->task_objq, (queue_entry_t) object); in vm_owned_objects_disown()
|
| /xnu-8020.101.4/iokit/Kernel/ |
| H A D | IOKitDebug.cpp | 608 if (queue_end(que, (queue_entry_t) site)) { in IOTrackingAdd() 642 if (queue_end(&site->instances, (queue_entry_t)site->addresses)) { in IOTrackingAdd()
|
| /xnu-8020.101.4/osfmk/arm/ |
| H A D | model_dep.c | 522 PANIC_VALIDATE_PTR(thread) && !queue_end(&threads, (queue_entry_t)thread); in do_print_all_backtraces()
|