xref: /xnu-8019.80.24/bsd/pthread/pthread_workqueue.c (revision a325d9c4a84054e40bbe985afedcb50ab80993ea)
1*a325d9c4SApple OSS Distributions /*
2*a325d9c4SApple OSS Distributions  * Copyright (c) 2000-2020 Apple Inc. All rights reserved.
3*a325d9c4SApple OSS Distributions  *
4*a325d9c4SApple OSS Distributions  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5*a325d9c4SApple OSS Distributions  *
6*a325d9c4SApple OSS Distributions  * This file contains Original Code and/or Modifications of Original Code
7*a325d9c4SApple OSS Distributions  * as defined in and that are subject to the Apple Public Source License
8*a325d9c4SApple OSS Distributions  * Version 2.0 (the 'License'). You may not use this file except in
9*a325d9c4SApple OSS Distributions  * compliance with the License. The rights granted to you under the License
10*a325d9c4SApple OSS Distributions  * may not be used to create, or enable the creation or redistribution of,
11*a325d9c4SApple OSS Distributions  * unlawful or unlicensed copies of an Apple operating system, or to
12*a325d9c4SApple OSS Distributions  * circumvent, violate, or enable the circumvention or violation of, any
13*a325d9c4SApple OSS Distributions  * terms of an Apple operating system software license agreement.
14*a325d9c4SApple OSS Distributions  *
15*a325d9c4SApple OSS Distributions  * Please obtain a copy of the License at
16*a325d9c4SApple OSS Distributions  * http://www.opensource.apple.com/apsl/ and read it before using this file.
17*a325d9c4SApple OSS Distributions  *
18*a325d9c4SApple OSS Distributions  * The Original Code and all software distributed under the License are
19*a325d9c4SApple OSS Distributions  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20*a325d9c4SApple OSS Distributions  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21*a325d9c4SApple OSS Distributions  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22*a325d9c4SApple OSS Distributions  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23*a325d9c4SApple OSS Distributions  * Please see the License for the specific language governing rights and
24*a325d9c4SApple OSS Distributions  * limitations under the License.
25*a325d9c4SApple OSS Distributions  *
26*a325d9c4SApple OSS Distributions  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27*a325d9c4SApple OSS Distributions  */
28*a325d9c4SApple OSS Distributions /* Copyright (c) 1995-2018 Apple, Inc. All Rights Reserved */
29*a325d9c4SApple OSS Distributions 
30*a325d9c4SApple OSS Distributions #include <sys/cdefs.h>
31*a325d9c4SApple OSS Distributions 
32*a325d9c4SApple OSS Distributions #include <kern/assert.h>
33*a325d9c4SApple OSS Distributions #include <kern/ast.h>
34*a325d9c4SApple OSS Distributions #include <kern/clock.h>
35*a325d9c4SApple OSS Distributions #include <kern/cpu_data.h>
36*a325d9c4SApple OSS Distributions #include <kern/kern_types.h>
37*a325d9c4SApple OSS Distributions #include <kern/policy_internal.h>
38*a325d9c4SApple OSS Distributions #include <kern/processor.h>
39*a325d9c4SApple OSS Distributions #include <kern/sched_prim.h>    /* for thread_exception_return */
40*a325d9c4SApple OSS Distributions #include <kern/task.h>
41*a325d9c4SApple OSS Distributions #include <kern/thread.h>
42*a325d9c4SApple OSS Distributions #include <kern/thread_group.h>
43*a325d9c4SApple OSS Distributions #include <kern/zalloc.h>
44*a325d9c4SApple OSS Distributions #include <mach/kern_return.h>
45*a325d9c4SApple OSS Distributions #include <mach/mach_param.h>
46*a325d9c4SApple OSS Distributions #include <mach/mach_port.h>
47*a325d9c4SApple OSS Distributions #include <mach/mach_types.h>
48*a325d9c4SApple OSS Distributions #include <mach/mach_vm.h>
49*a325d9c4SApple OSS Distributions #include <mach/sync_policy.h>
50*a325d9c4SApple OSS Distributions #include <mach/task.h>
51*a325d9c4SApple OSS Distributions #include <mach/thread_act.h> /* for thread_resume */
52*a325d9c4SApple OSS Distributions #include <mach/thread_policy.h>
53*a325d9c4SApple OSS Distributions #include <mach/thread_status.h>
54*a325d9c4SApple OSS Distributions #include <mach/vm_prot.h>
55*a325d9c4SApple OSS Distributions #include <mach/vm_statistics.h>
56*a325d9c4SApple OSS Distributions #include <machine/atomic.h>
57*a325d9c4SApple OSS Distributions #include <machine/machine_routines.h>
58*a325d9c4SApple OSS Distributions #include <machine/smp.h>
59*a325d9c4SApple OSS Distributions #include <vm/vm_map.h>
60*a325d9c4SApple OSS Distributions #include <vm/vm_protos.h>
61*a325d9c4SApple OSS Distributions 
62*a325d9c4SApple OSS Distributions #include <sys/eventvar.h>
63*a325d9c4SApple OSS Distributions #include <sys/kdebug.h>
64*a325d9c4SApple OSS Distributions #include <sys/kernel.h>
65*a325d9c4SApple OSS Distributions #include <sys/lock.h>
66*a325d9c4SApple OSS Distributions #include <sys/param.h>
67*a325d9c4SApple OSS Distributions #include <sys/proc_info.h>      /* for fill_procworkqueue */
68*a325d9c4SApple OSS Distributions #include <sys/proc_internal.h>
69*a325d9c4SApple OSS Distributions #include <sys/pthread_shims.h>
70*a325d9c4SApple OSS Distributions #include <sys/resourcevar.h>
71*a325d9c4SApple OSS Distributions #include <sys/signalvar.h>
72*a325d9c4SApple OSS Distributions #include <sys/sysctl.h>
73*a325d9c4SApple OSS Distributions #include <sys/sysproto.h>
74*a325d9c4SApple OSS Distributions #include <sys/systm.h>
75*a325d9c4SApple OSS Distributions #include <sys/ulock.h> /* for ulock_owner_value_to_port_name */
76*a325d9c4SApple OSS Distributions 
77*a325d9c4SApple OSS Distributions #include <pthread/bsdthread_private.h>
78*a325d9c4SApple OSS Distributions #include <pthread/workqueue_syscalls.h>
79*a325d9c4SApple OSS Distributions #include <pthread/workqueue_internal.h>
80*a325d9c4SApple OSS Distributions #include <pthread/workqueue_trace.h>
81*a325d9c4SApple OSS Distributions 
82*a325d9c4SApple OSS Distributions #include <os/log.h>
83*a325d9c4SApple OSS Distributions 
84*a325d9c4SApple OSS Distributions static void workq_unpark_continue(void *uth, wait_result_t wr) __dead2;
85*a325d9c4SApple OSS Distributions static void workq_schedule_creator(proc_t p, struct workqueue *wq,
86*a325d9c4SApple OSS Distributions     workq_kern_threadreq_flags_t flags);
87*a325d9c4SApple OSS Distributions 
88*a325d9c4SApple OSS Distributions static bool workq_threadreq_admissible(struct workqueue *wq, struct uthread *uth,
89*a325d9c4SApple OSS Distributions     workq_threadreq_t req);
90*a325d9c4SApple OSS Distributions 
91*a325d9c4SApple OSS Distributions static uint32_t workq_constrained_allowance(struct workqueue *wq,
92*a325d9c4SApple OSS Distributions     thread_qos_t at_qos, struct uthread *uth, bool may_start_timer);
93*a325d9c4SApple OSS Distributions 
94*a325d9c4SApple OSS Distributions static bool _wq_cooperative_queue_refresh_best_req_qos(struct workqueue *wq);
95*a325d9c4SApple OSS Distributions 
96*a325d9c4SApple OSS Distributions static bool workq_thread_is_busy(uint64_t cur_ts,
97*a325d9c4SApple OSS Distributions     _Atomic uint64_t *lastblocked_tsp);
98*a325d9c4SApple OSS Distributions 
99*a325d9c4SApple OSS Distributions static int workq_sysctl_handle_usecs SYSCTL_HANDLER_ARGS;
100*a325d9c4SApple OSS Distributions 
101*a325d9c4SApple OSS Distributions static bool
102*a325d9c4SApple OSS Distributions workq_schedule_delayed_thread_creation(struct workqueue *wq, int flags);
103*a325d9c4SApple OSS Distributions 
104*a325d9c4SApple OSS Distributions static inline void
105*a325d9c4SApple OSS Distributions workq_lock_spin(struct workqueue *wq);
106*a325d9c4SApple OSS Distributions 
107*a325d9c4SApple OSS Distributions static inline void
108*a325d9c4SApple OSS Distributions workq_unlock(struct workqueue *wq);
109*a325d9c4SApple OSS Distributions 
110*a325d9c4SApple OSS Distributions #pragma mark globals
111*a325d9c4SApple OSS Distributions 
112*a325d9c4SApple OSS Distributions struct workq_usec_var {
113*a325d9c4SApple OSS Distributions 	uint32_t usecs;
114*a325d9c4SApple OSS Distributions 	uint64_t abstime;
115*a325d9c4SApple OSS Distributions };
116*a325d9c4SApple OSS Distributions 
117*a325d9c4SApple OSS Distributions #define WORKQ_SYSCTL_USECS(var, init) \
118*a325d9c4SApple OSS Distributions 	        static struct workq_usec_var var = { .usecs = init }; \
119*a325d9c4SApple OSS Distributions 	        SYSCTL_OID(_kern, OID_AUTO, var##_usecs, \
120*a325d9c4SApple OSS Distributions 	                        CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_LOCKED, &var, 0, \
121*a325d9c4SApple OSS Distributions 	                        workq_sysctl_handle_usecs, "I", "")
122*a325d9c4SApple OSS Distributions 
123*a325d9c4SApple OSS Distributions static LCK_GRP_DECLARE(workq_lck_grp, "workq");
124*a325d9c4SApple OSS Distributions os_refgrp_decl(static, workq_refgrp, "workq", NULL);
125*a325d9c4SApple OSS Distributions 
126*a325d9c4SApple OSS Distributions static ZONE_DECLARE(workq_zone_workqueue, "workq.wq",
127*a325d9c4SApple OSS Distributions     sizeof(struct workqueue), ZC_NONE);
128*a325d9c4SApple OSS Distributions static ZONE_DECLARE(workq_zone_threadreq, "workq.threadreq",
129*a325d9c4SApple OSS Distributions     sizeof(struct workq_threadreq_s), ZC_CACHING);
130*a325d9c4SApple OSS Distributions 
131*a325d9c4SApple OSS Distributions static struct mpsc_daemon_queue workq_deallocate_queue;
132*a325d9c4SApple OSS Distributions 
133*a325d9c4SApple OSS Distributions WORKQ_SYSCTL_USECS(wq_stalled_window, WQ_STALLED_WINDOW_USECS);
134*a325d9c4SApple OSS Distributions WORKQ_SYSCTL_USECS(wq_reduce_pool_window, WQ_REDUCE_POOL_WINDOW_USECS);
135*a325d9c4SApple OSS Distributions WORKQ_SYSCTL_USECS(wq_max_timer_interval, WQ_MAX_TIMER_INTERVAL_USECS);
136*a325d9c4SApple OSS Distributions static uint32_t wq_max_threads              = WORKQUEUE_MAXTHREADS;
137*a325d9c4SApple OSS Distributions static uint32_t wq_max_constrained_threads  = WORKQUEUE_MAXTHREADS / 8;
138*a325d9c4SApple OSS Distributions static uint32_t wq_init_constrained_limit   = 1;
139*a325d9c4SApple OSS Distributions static uint16_t wq_death_max_load;
140*a325d9c4SApple OSS Distributions static uint32_t wq_max_parallelism[WORKQ_NUM_QOS_BUCKETS];
141*a325d9c4SApple OSS Distributions 
142*a325d9c4SApple OSS Distributions /*
143*a325d9c4SApple OSS Distributions  * This is not a hard limit but the max size we want to aim to hit across the
144*a325d9c4SApple OSS Distributions  * entire cooperative pool. We can oversubscribe the pool due to non-cooperative
145*a325d9c4SApple OSS Distributions  * workers and the max we will oversubscribe the pool by, is a total of
146*a325d9c4SApple OSS Distributions  * wq_max_cooperative_threads * WORKQ_NUM_QOS_BUCKETS.
147*a325d9c4SApple OSS Distributions  */
148*a325d9c4SApple OSS Distributions static uint32_t wq_max_cooperative_threads;
149*a325d9c4SApple OSS Distributions 
150*a325d9c4SApple OSS Distributions static inline uint32_t
wq_cooperative_queue_max_size(struct workqueue * wq)151*a325d9c4SApple OSS Distributions wq_cooperative_queue_max_size(struct workqueue *wq)
152*a325d9c4SApple OSS Distributions {
153*a325d9c4SApple OSS Distributions 	return wq->wq_cooperative_queue_has_limited_max_size ? 1 : wq_max_cooperative_threads;
154*a325d9c4SApple OSS Distributions }
155*a325d9c4SApple OSS Distributions 
156*a325d9c4SApple OSS Distributions #pragma mark sysctls
157*a325d9c4SApple OSS Distributions 
158*a325d9c4SApple OSS Distributions static int
159*a325d9c4SApple OSS Distributions workq_sysctl_handle_usecs SYSCTL_HANDLER_ARGS
160*a325d9c4SApple OSS Distributions {
161*a325d9c4SApple OSS Distributions #pragma unused(arg2)
162*a325d9c4SApple OSS Distributions 	struct workq_usec_var *v = arg1;
163*a325d9c4SApple OSS Distributions 	int error = sysctl_handle_int(oidp, &v->usecs, 0, req);
164*a325d9c4SApple OSS Distributions 	if (error || !req->newptr) {
165*a325d9c4SApple OSS Distributions 		return error;
166*a325d9c4SApple OSS Distributions 	}
167*a325d9c4SApple OSS Distributions 	clock_interval_to_absolutetime_interval(v->usecs, NSEC_PER_USEC,
168*a325d9c4SApple OSS Distributions 	    &v->abstime);
169*a325d9c4SApple OSS Distributions 	return 0;
170*a325d9c4SApple OSS Distributions }
171*a325d9c4SApple OSS Distributions 
172*a325d9c4SApple OSS Distributions SYSCTL_INT(_kern, OID_AUTO, wq_max_threads, CTLFLAG_RW | CTLFLAG_LOCKED,
173*a325d9c4SApple OSS Distributions     &wq_max_threads, 0, "");
174*a325d9c4SApple OSS Distributions 
175*a325d9c4SApple OSS Distributions SYSCTL_INT(_kern, OID_AUTO, wq_max_constrained_threads, CTLFLAG_RW | CTLFLAG_LOCKED,
176*a325d9c4SApple OSS Distributions     &wq_max_constrained_threads, 0, "");
177*a325d9c4SApple OSS Distributions 
178*a325d9c4SApple OSS Distributions static int
179*a325d9c4SApple OSS Distributions wq_limit_cooperative_threads_for_proc SYSCTL_HANDLER_ARGS
180*a325d9c4SApple OSS Distributions {
181*a325d9c4SApple OSS Distributions #pragma unused(arg1, arg2, oidp)
182*a325d9c4SApple OSS Distributions 	int input_pool_size = 0;
183*a325d9c4SApple OSS Distributions 	int changed;
184*a325d9c4SApple OSS Distributions 	int error = 0;
185*a325d9c4SApple OSS Distributions 
186*a325d9c4SApple OSS Distributions 	error = sysctl_io_number(req, 0, sizeof(int), &input_pool_size, &changed);
187*a325d9c4SApple OSS Distributions 	if (error || !changed) {
188*a325d9c4SApple OSS Distributions 		return error;
189*a325d9c4SApple OSS Distributions 	}
190*a325d9c4SApple OSS Distributions 
191*a325d9c4SApple OSS Distributions #define WQ_COOPERATIVE_POOL_SIZE_DEFAULT 0
192*a325d9c4SApple OSS Distributions #define WQ_COOPERATIVE_POOL_SIZE_STRICT_PER_QOS -1
193*a325d9c4SApple OSS Distributions /* Not available currently, but sysctl interface is designed to allow these
194*a325d9c4SApple OSS Distributions  * extra parameters:
195*a325d9c4SApple OSS Distributions  *		WQ_COOPERATIVE_POOL_SIZE_STRICT : -2 (across all bucket)
196*a325d9c4SApple OSS Distributions  *		WQ_COOPERATIVE_POOL_SIZE_CUSTOM : [1, 512]
197*a325d9c4SApple OSS Distributions  */
198*a325d9c4SApple OSS Distributions 
199*a325d9c4SApple OSS Distributions 	if (input_pool_size != WQ_COOPERATIVE_POOL_SIZE_DEFAULT
200*a325d9c4SApple OSS Distributions 	    && input_pool_size != WQ_COOPERATIVE_POOL_SIZE_STRICT_PER_QOS) {
201*a325d9c4SApple OSS Distributions 		error = EINVAL;
202*a325d9c4SApple OSS Distributions 		goto out;
203*a325d9c4SApple OSS Distributions 	}
204*a325d9c4SApple OSS Distributions 
205*a325d9c4SApple OSS Distributions 	proc_t p = req->p;
206*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
207*a325d9c4SApple OSS Distributions 
208*a325d9c4SApple OSS Distributions 	if (wq != NULL) {
209*a325d9c4SApple OSS Distributions 		workq_lock_spin(wq);
210*a325d9c4SApple OSS Distributions 		if (wq->wq_reqcount > 0 || wq->wq_nthreads > 0) {
211*a325d9c4SApple OSS Distributions 			// Hackily enforce that the workqueue is still new (no requests or
212*a325d9c4SApple OSS Distributions 			// threads)
213*a325d9c4SApple OSS Distributions 			error = ENOTSUP;
214*a325d9c4SApple OSS Distributions 		} else {
215*a325d9c4SApple OSS Distributions 			wq->wq_cooperative_queue_has_limited_max_size = (input_pool_size == WQ_COOPERATIVE_POOL_SIZE_STRICT_PER_QOS);
216*a325d9c4SApple OSS Distributions 		}
217*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
218*a325d9c4SApple OSS Distributions 	} else {
219*a325d9c4SApple OSS Distributions 		/* This process has no workqueue, calling this syctl makes no sense */
220*a325d9c4SApple OSS Distributions 		return ENOTSUP;
221*a325d9c4SApple OSS Distributions 	}
222*a325d9c4SApple OSS Distributions 
223*a325d9c4SApple OSS Distributions out:
224*a325d9c4SApple OSS Distributions 	return error;
225*a325d9c4SApple OSS Distributions }
226*a325d9c4SApple OSS Distributions 
227*a325d9c4SApple OSS Distributions SYSCTL_PROC(_kern, OID_AUTO, wq_limit_cooperative_threads,
228*a325d9c4SApple OSS Distributions     CTLFLAG_ANYBODY | CTLFLAG_MASKED | CTLFLAG_WR | CTLFLAG_LOCKED | CTLTYPE_INT, 0, 0,
229*a325d9c4SApple OSS Distributions     wq_limit_cooperative_threads_for_proc,
230*a325d9c4SApple OSS Distributions     "I", "Modify the max pool size of the cooperative pool");
231*a325d9c4SApple OSS Distributions 
232*a325d9c4SApple OSS Distributions #pragma mark p_wqptr
233*a325d9c4SApple OSS Distributions 
234*a325d9c4SApple OSS Distributions #define WQPTR_IS_INITING_VALUE ((struct workqueue *)~(uintptr_t)0)
235*a325d9c4SApple OSS Distributions 
236*a325d9c4SApple OSS Distributions static struct workqueue *
proc_get_wqptr_fast(struct proc * p)237*a325d9c4SApple OSS Distributions proc_get_wqptr_fast(struct proc *p)
238*a325d9c4SApple OSS Distributions {
239*a325d9c4SApple OSS Distributions 	return os_atomic_load(&p->p_wqptr, relaxed);
240*a325d9c4SApple OSS Distributions }
241*a325d9c4SApple OSS Distributions 
242*a325d9c4SApple OSS Distributions struct workqueue *
proc_get_wqptr(struct proc * p)243*a325d9c4SApple OSS Distributions proc_get_wqptr(struct proc *p)
244*a325d9c4SApple OSS Distributions {
245*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
246*a325d9c4SApple OSS Distributions 	return wq == WQPTR_IS_INITING_VALUE ? NULL : wq;
247*a325d9c4SApple OSS Distributions }
248*a325d9c4SApple OSS Distributions 
249*a325d9c4SApple OSS Distributions static void
proc_set_wqptr(struct proc * p,struct workqueue * wq)250*a325d9c4SApple OSS Distributions proc_set_wqptr(struct proc *p, struct workqueue *wq)
251*a325d9c4SApple OSS Distributions {
252*a325d9c4SApple OSS Distributions 	wq = os_atomic_xchg(&p->p_wqptr, wq, release);
253*a325d9c4SApple OSS Distributions 	if (wq == WQPTR_IS_INITING_VALUE) {
254*a325d9c4SApple OSS Distributions 		proc_lock(p);
255*a325d9c4SApple OSS Distributions 		thread_wakeup(&p->p_wqptr);
256*a325d9c4SApple OSS Distributions 		proc_unlock(p);
257*a325d9c4SApple OSS Distributions 	}
258*a325d9c4SApple OSS Distributions }
259*a325d9c4SApple OSS Distributions 
260*a325d9c4SApple OSS Distributions static bool
proc_init_wqptr_or_wait(struct proc * p)261*a325d9c4SApple OSS Distributions proc_init_wqptr_or_wait(struct proc *p)
262*a325d9c4SApple OSS Distributions {
263*a325d9c4SApple OSS Distributions 	struct workqueue *wq;
264*a325d9c4SApple OSS Distributions 
265*a325d9c4SApple OSS Distributions 	proc_lock(p);
266*a325d9c4SApple OSS Distributions 	wq = os_atomic_load(&p->p_wqptr, relaxed);
267*a325d9c4SApple OSS Distributions 
268*a325d9c4SApple OSS Distributions 	if (wq == NULL) {
269*a325d9c4SApple OSS Distributions 		os_atomic_store(&p->p_wqptr, WQPTR_IS_INITING_VALUE, relaxed);
270*a325d9c4SApple OSS Distributions 		proc_unlock(p);
271*a325d9c4SApple OSS Distributions 		return true;
272*a325d9c4SApple OSS Distributions 	}
273*a325d9c4SApple OSS Distributions 
274*a325d9c4SApple OSS Distributions 	if (wq == WQPTR_IS_INITING_VALUE) {
275*a325d9c4SApple OSS Distributions 		assert_wait(&p->p_wqptr, THREAD_UNINT);
276*a325d9c4SApple OSS Distributions 		proc_unlock(p);
277*a325d9c4SApple OSS Distributions 		thread_block(THREAD_CONTINUE_NULL);
278*a325d9c4SApple OSS Distributions 	} else {
279*a325d9c4SApple OSS Distributions 		proc_unlock(p);
280*a325d9c4SApple OSS Distributions 	}
281*a325d9c4SApple OSS Distributions 	return false;
282*a325d9c4SApple OSS Distributions }
283*a325d9c4SApple OSS Distributions 
284*a325d9c4SApple OSS Distributions static inline event_t
workq_parked_wait_event(struct uthread * uth)285*a325d9c4SApple OSS Distributions workq_parked_wait_event(struct uthread *uth)
286*a325d9c4SApple OSS Distributions {
287*a325d9c4SApple OSS Distributions 	return (event_t)&uth->uu_workq_stackaddr;
288*a325d9c4SApple OSS Distributions }
289*a325d9c4SApple OSS Distributions 
290*a325d9c4SApple OSS Distributions static inline void
workq_thread_wakeup(struct uthread * uth)291*a325d9c4SApple OSS Distributions workq_thread_wakeup(struct uthread *uth)
292*a325d9c4SApple OSS Distributions {
293*a325d9c4SApple OSS Distributions 	thread_wakeup_thread(workq_parked_wait_event(uth), get_machthread(uth));
294*a325d9c4SApple OSS Distributions }
295*a325d9c4SApple OSS Distributions 
296*a325d9c4SApple OSS Distributions #pragma mark wq_thactive
297*a325d9c4SApple OSS Distributions 
298*a325d9c4SApple OSS Distributions #if defined(__LP64__)
299*a325d9c4SApple OSS Distributions // Layout is:
300*a325d9c4SApple OSS Distributions //   127 - 115 : 13 bits of zeroes
301*a325d9c4SApple OSS Distributions //   114 - 112 : best QoS among all pending constrained requests
302*a325d9c4SApple OSS Distributions //   111 -   0 : MGR, AUI, UI, IN, DF, UT, BG+MT buckets every 16 bits
303*a325d9c4SApple OSS Distributions #define WQ_THACTIVE_BUCKET_WIDTH 16
304*a325d9c4SApple OSS Distributions #define WQ_THACTIVE_QOS_SHIFT    (7 * WQ_THACTIVE_BUCKET_WIDTH)
305*a325d9c4SApple OSS Distributions #else
306*a325d9c4SApple OSS Distributions // Layout is:
307*a325d9c4SApple OSS Distributions //   63 - 61 : best QoS among all pending constrained requests
308*a325d9c4SApple OSS Distributions //   60      : Manager bucket (0 or 1)
309*a325d9c4SApple OSS Distributions //   59 -  0 : AUI, UI, IN, DF, UT, BG+MT buckets every 10 bits
310*a325d9c4SApple OSS Distributions #define WQ_THACTIVE_BUCKET_WIDTH 10
311*a325d9c4SApple OSS Distributions #define WQ_THACTIVE_QOS_SHIFT    (6 * WQ_THACTIVE_BUCKET_WIDTH + 1)
312*a325d9c4SApple OSS Distributions #endif
313*a325d9c4SApple OSS Distributions #define WQ_THACTIVE_BUCKET_MASK  ((1U << WQ_THACTIVE_BUCKET_WIDTH) - 1)
314*a325d9c4SApple OSS Distributions #define WQ_THACTIVE_BUCKET_HALF  (1U << (WQ_THACTIVE_BUCKET_WIDTH - 1))
315*a325d9c4SApple OSS Distributions 
316*a325d9c4SApple OSS Distributions static_assert(sizeof(wq_thactive_t) * CHAR_BIT - WQ_THACTIVE_QOS_SHIFT >= 3,
317*a325d9c4SApple OSS Distributions     "Make sure we have space to encode a QoS");
318*a325d9c4SApple OSS Distributions 
319*a325d9c4SApple OSS Distributions static inline wq_thactive_t
_wq_thactive(struct workqueue * wq)320*a325d9c4SApple OSS Distributions _wq_thactive(struct workqueue *wq)
321*a325d9c4SApple OSS Distributions {
322*a325d9c4SApple OSS Distributions 	return os_atomic_load_wide(&wq->wq_thactive, relaxed);
323*a325d9c4SApple OSS Distributions }
324*a325d9c4SApple OSS Distributions 
325*a325d9c4SApple OSS Distributions static inline int
_wq_bucket(thread_qos_t qos)326*a325d9c4SApple OSS Distributions _wq_bucket(thread_qos_t qos)
327*a325d9c4SApple OSS Distributions {
328*a325d9c4SApple OSS Distributions 	// Map both BG and MT to the same bucket by over-shifting down and
329*a325d9c4SApple OSS Distributions 	// clamping MT and BG together.
330*a325d9c4SApple OSS Distributions 	switch (qos) {
331*a325d9c4SApple OSS Distributions 	case THREAD_QOS_MAINTENANCE:
332*a325d9c4SApple OSS Distributions 		return 0;
333*a325d9c4SApple OSS Distributions 	default:
334*a325d9c4SApple OSS Distributions 		return qos - 2;
335*a325d9c4SApple OSS Distributions 	}
336*a325d9c4SApple OSS Distributions }
337*a325d9c4SApple OSS Distributions 
338*a325d9c4SApple OSS Distributions #define WQ_THACTIVE_BEST_CONSTRAINED_REQ_QOS(tha) \
339*a325d9c4SApple OSS Distributions 	        ((thread_qos_t)((tha) >> WQ_THACTIVE_QOS_SHIFT))
340*a325d9c4SApple OSS Distributions 
341*a325d9c4SApple OSS Distributions static inline thread_qos_t
_wq_thactive_best_constrained_req_qos(struct workqueue * wq)342*a325d9c4SApple OSS Distributions _wq_thactive_best_constrained_req_qos(struct workqueue *wq)
343*a325d9c4SApple OSS Distributions {
344*a325d9c4SApple OSS Distributions 	// Avoid expensive atomic operations: the three bits we're loading are in
345*a325d9c4SApple OSS Distributions 	// a single byte, and always updated under the workqueue lock
346*a325d9c4SApple OSS Distributions 	wq_thactive_t v = *(wq_thactive_t *)&wq->wq_thactive;
347*a325d9c4SApple OSS Distributions 	return WQ_THACTIVE_BEST_CONSTRAINED_REQ_QOS(v);
348*a325d9c4SApple OSS Distributions }
349*a325d9c4SApple OSS Distributions 
350*a325d9c4SApple OSS Distributions static void
_wq_thactive_refresh_best_constrained_req_qos(struct workqueue * wq)351*a325d9c4SApple OSS Distributions _wq_thactive_refresh_best_constrained_req_qos(struct workqueue *wq)
352*a325d9c4SApple OSS Distributions {
353*a325d9c4SApple OSS Distributions 	thread_qos_t old_qos, new_qos;
354*a325d9c4SApple OSS Distributions 	workq_threadreq_t req;
355*a325d9c4SApple OSS Distributions 
356*a325d9c4SApple OSS Distributions 	req = priority_queue_max(&wq->wq_constrained_queue,
357*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry);
358*a325d9c4SApple OSS Distributions 	new_qos = req ? req->tr_qos : THREAD_QOS_UNSPECIFIED;
359*a325d9c4SApple OSS Distributions 	old_qos = _wq_thactive_best_constrained_req_qos(wq);
360*a325d9c4SApple OSS Distributions 	if (old_qos != new_qos) {
361*a325d9c4SApple OSS Distributions 		long delta = (long)new_qos - (long)old_qos;
362*a325d9c4SApple OSS Distributions 		wq_thactive_t v = (wq_thactive_t)delta << WQ_THACTIVE_QOS_SHIFT;
363*a325d9c4SApple OSS Distributions 		/*
364*a325d9c4SApple OSS Distributions 		 * We can do an atomic add relative to the initial load because updates
365*a325d9c4SApple OSS Distributions 		 * to this qos are always serialized under the workqueue lock.
366*a325d9c4SApple OSS Distributions 		 */
367*a325d9c4SApple OSS Distributions 		v = os_atomic_add(&wq->wq_thactive, v, relaxed);
368*a325d9c4SApple OSS Distributions #ifdef __LP64__
369*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thactive_update, wq, (uint64_t)v,
370*a325d9c4SApple OSS Distributions 		    (uint64_t)(v >> 64), 0);
371*a325d9c4SApple OSS Distributions #else
372*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thactive_update, wq, v, 0, 0);
373*a325d9c4SApple OSS Distributions #endif
374*a325d9c4SApple OSS Distributions 	}
375*a325d9c4SApple OSS Distributions }
376*a325d9c4SApple OSS Distributions 
377*a325d9c4SApple OSS Distributions static inline wq_thactive_t
_wq_thactive_offset_for_qos(thread_qos_t qos)378*a325d9c4SApple OSS Distributions _wq_thactive_offset_for_qos(thread_qos_t qos)
379*a325d9c4SApple OSS Distributions {
380*a325d9c4SApple OSS Distributions 	return (wq_thactive_t)1 << (_wq_bucket(qos) * WQ_THACTIVE_BUCKET_WIDTH);
381*a325d9c4SApple OSS Distributions }
382*a325d9c4SApple OSS Distributions 
383*a325d9c4SApple OSS Distributions static inline wq_thactive_t
_wq_thactive_inc(struct workqueue * wq,thread_qos_t qos)384*a325d9c4SApple OSS Distributions _wq_thactive_inc(struct workqueue *wq, thread_qos_t qos)
385*a325d9c4SApple OSS Distributions {
386*a325d9c4SApple OSS Distributions 	wq_thactive_t v = _wq_thactive_offset_for_qos(qos);
387*a325d9c4SApple OSS Distributions 	return os_atomic_add_orig(&wq->wq_thactive, v, relaxed);
388*a325d9c4SApple OSS Distributions }
389*a325d9c4SApple OSS Distributions 
390*a325d9c4SApple OSS Distributions static inline wq_thactive_t
_wq_thactive_dec(struct workqueue * wq,thread_qos_t qos)391*a325d9c4SApple OSS Distributions _wq_thactive_dec(struct workqueue *wq, thread_qos_t qos)
392*a325d9c4SApple OSS Distributions {
393*a325d9c4SApple OSS Distributions 	wq_thactive_t v = _wq_thactive_offset_for_qos(qos);
394*a325d9c4SApple OSS Distributions 	return os_atomic_sub_orig(&wq->wq_thactive, v, relaxed);
395*a325d9c4SApple OSS Distributions }
396*a325d9c4SApple OSS Distributions 
397*a325d9c4SApple OSS Distributions static inline void
_wq_thactive_move(struct workqueue * wq,thread_qos_t old_qos,thread_qos_t new_qos)398*a325d9c4SApple OSS Distributions _wq_thactive_move(struct workqueue *wq,
399*a325d9c4SApple OSS Distributions     thread_qos_t old_qos, thread_qos_t new_qos)
400*a325d9c4SApple OSS Distributions {
401*a325d9c4SApple OSS Distributions 	wq_thactive_t v = _wq_thactive_offset_for_qos(new_qos) -
402*a325d9c4SApple OSS Distributions 	    _wq_thactive_offset_for_qos(old_qos);
403*a325d9c4SApple OSS Distributions 	os_atomic_add(&wq->wq_thactive, v, relaxed);
404*a325d9c4SApple OSS Distributions 	wq->wq_thscheduled_count[_wq_bucket(old_qos)]--;
405*a325d9c4SApple OSS Distributions 	wq->wq_thscheduled_count[_wq_bucket(new_qos)]++;
406*a325d9c4SApple OSS Distributions }
407*a325d9c4SApple OSS Distributions 
408*a325d9c4SApple OSS Distributions static inline uint32_t
_wq_thactive_aggregate_downto_qos(struct workqueue * wq,wq_thactive_t v,thread_qos_t qos,uint32_t * busycount,uint32_t * max_busycount)409*a325d9c4SApple OSS Distributions _wq_thactive_aggregate_downto_qos(struct workqueue *wq, wq_thactive_t v,
410*a325d9c4SApple OSS Distributions     thread_qos_t qos, uint32_t *busycount, uint32_t *max_busycount)
411*a325d9c4SApple OSS Distributions {
412*a325d9c4SApple OSS Distributions 	uint32_t count = 0, active;
413*a325d9c4SApple OSS Distributions 	uint64_t curtime;
414*a325d9c4SApple OSS Distributions 
415*a325d9c4SApple OSS Distributions 	assert(WORKQ_THREAD_QOS_MIN <= qos && qos <= WORKQ_THREAD_QOS_MAX);
416*a325d9c4SApple OSS Distributions 
417*a325d9c4SApple OSS Distributions 	if (busycount) {
418*a325d9c4SApple OSS Distributions 		curtime = mach_absolute_time();
419*a325d9c4SApple OSS Distributions 		*busycount = 0;
420*a325d9c4SApple OSS Distributions 	}
421*a325d9c4SApple OSS Distributions 	if (max_busycount) {
422*a325d9c4SApple OSS Distributions 		*max_busycount = THREAD_QOS_LAST - qos;
423*a325d9c4SApple OSS Distributions 	}
424*a325d9c4SApple OSS Distributions 
425*a325d9c4SApple OSS Distributions 	int i = _wq_bucket(qos);
426*a325d9c4SApple OSS Distributions 	v >>= i * WQ_THACTIVE_BUCKET_WIDTH;
427*a325d9c4SApple OSS Distributions 	for (; i < WORKQ_NUM_QOS_BUCKETS; i++, v >>= WQ_THACTIVE_BUCKET_WIDTH) {
428*a325d9c4SApple OSS Distributions 		active = v & WQ_THACTIVE_BUCKET_MASK;
429*a325d9c4SApple OSS Distributions 		count += active;
430*a325d9c4SApple OSS Distributions 
431*a325d9c4SApple OSS Distributions 		if (busycount && wq->wq_thscheduled_count[i] > active) {
432*a325d9c4SApple OSS Distributions 			if (workq_thread_is_busy(curtime, &wq->wq_lastblocked_ts[i])) {
433*a325d9c4SApple OSS Distributions 				/*
434*a325d9c4SApple OSS Distributions 				 * We only consider the last blocked thread for a given bucket
435*a325d9c4SApple OSS Distributions 				 * as busy because we don't want to take the list lock in each
436*a325d9c4SApple OSS Distributions 				 * sched callback. However this is an approximation that could
437*a325d9c4SApple OSS Distributions 				 * contribute to thread creation storms.
438*a325d9c4SApple OSS Distributions 				 */
439*a325d9c4SApple OSS Distributions 				(*busycount)++;
440*a325d9c4SApple OSS Distributions 			}
441*a325d9c4SApple OSS Distributions 		}
442*a325d9c4SApple OSS Distributions 	}
443*a325d9c4SApple OSS Distributions 
444*a325d9c4SApple OSS Distributions 	return count;
445*a325d9c4SApple OSS Distributions }
446*a325d9c4SApple OSS Distributions 
447*a325d9c4SApple OSS Distributions static inline void
_wq_cooperative_queue_scheduled_count_dec(struct workqueue * wq,thread_qos_t qos)448*a325d9c4SApple OSS Distributions _wq_cooperative_queue_scheduled_count_dec(struct workqueue *wq, thread_qos_t qos)
449*a325d9c4SApple OSS Distributions {
450*a325d9c4SApple OSS Distributions 	__assert_only uint8_t old_scheduled_count = wq->wq_cooperative_queue_scheduled_count[_wq_bucket(qos)]--;
451*a325d9c4SApple OSS Distributions 	assert(old_scheduled_count > 0);
452*a325d9c4SApple OSS Distributions }
453*a325d9c4SApple OSS Distributions 
454*a325d9c4SApple OSS Distributions static inline void
_wq_cooperative_queue_scheduled_count_inc(struct workqueue * wq,thread_qos_t qos)455*a325d9c4SApple OSS Distributions _wq_cooperative_queue_scheduled_count_inc(struct workqueue *wq, thread_qos_t qos)
456*a325d9c4SApple OSS Distributions {
457*a325d9c4SApple OSS Distributions 	__assert_only uint8_t old_scheduled_count = wq->wq_cooperative_queue_scheduled_count[_wq_bucket(qos)]++;
458*a325d9c4SApple OSS Distributions 	assert(old_scheduled_count < UINT8_MAX);
459*a325d9c4SApple OSS Distributions }
460*a325d9c4SApple OSS Distributions 
461*a325d9c4SApple OSS Distributions #pragma mark wq_flags
462*a325d9c4SApple OSS Distributions 
463*a325d9c4SApple OSS Distributions static inline uint32_t
_wq_flags(struct workqueue * wq)464*a325d9c4SApple OSS Distributions _wq_flags(struct workqueue *wq)
465*a325d9c4SApple OSS Distributions {
466*a325d9c4SApple OSS Distributions 	return os_atomic_load(&wq->wq_flags, relaxed);
467*a325d9c4SApple OSS Distributions }
468*a325d9c4SApple OSS Distributions 
469*a325d9c4SApple OSS Distributions static inline bool
_wq_exiting(struct workqueue * wq)470*a325d9c4SApple OSS Distributions _wq_exiting(struct workqueue *wq)
471*a325d9c4SApple OSS Distributions {
472*a325d9c4SApple OSS Distributions 	return _wq_flags(wq) & WQ_EXITING;
473*a325d9c4SApple OSS Distributions }
474*a325d9c4SApple OSS Distributions 
475*a325d9c4SApple OSS Distributions bool
workq_is_exiting(struct proc * p)476*a325d9c4SApple OSS Distributions workq_is_exiting(struct proc *p)
477*a325d9c4SApple OSS Distributions {
478*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
479*a325d9c4SApple OSS Distributions 	return !wq || _wq_exiting(wq);
480*a325d9c4SApple OSS Distributions }
481*a325d9c4SApple OSS Distributions 
482*a325d9c4SApple OSS Distributions 
483*a325d9c4SApple OSS Distributions #pragma mark workqueue lock
484*a325d9c4SApple OSS Distributions 
485*a325d9c4SApple OSS Distributions static bool
workq_lock_is_acquired_kdp(struct workqueue * wq)486*a325d9c4SApple OSS Distributions workq_lock_is_acquired_kdp(struct workqueue *wq)
487*a325d9c4SApple OSS Distributions {
488*a325d9c4SApple OSS Distributions 	return kdp_lck_ticket_is_acquired(&wq->wq_lock);
489*a325d9c4SApple OSS Distributions }
490*a325d9c4SApple OSS Distributions 
491*a325d9c4SApple OSS Distributions static inline void
workq_lock_spin(struct workqueue * wq)492*a325d9c4SApple OSS Distributions workq_lock_spin(struct workqueue *wq)
493*a325d9c4SApple OSS Distributions {
494*a325d9c4SApple OSS Distributions 	lck_ticket_lock(&wq->wq_lock, &workq_lck_grp);
495*a325d9c4SApple OSS Distributions }
496*a325d9c4SApple OSS Distributions 
497*a325d9c4SApple OSS Distributions static inline void
workq_lock_held(struct workqueue * wq)498*a325d9c4SApple OSS Distributions workq_lock_held(struct workqueue *wq)
499*a325d9c4SApple OSS Distributions {
500*a325d9c4SApple OSS Distributions 	LCK_TICKET_ASSERT_OWNED(&wq->wq_lock);
501*a325d9c4SApple OSS Distributions }
502*a325d9c4SApple OSS Distributions 
503*a325d9c4SApple OSS Distributions static inline bool
workq_lock_try(struct workqueue * wq)504*a325d9c4SApple OSS Distributions workq_lock_try(struct workqueue *wq)
505*a325d9c4SApple OSS Distributions {
506*a325d9c4SApple OSS Distributions 	return lck_ticket_lock_try(&wq->wq_lock, &workq_lck_grp);
507*a325d9c4SApple OSS Distributions }
508*a325d9c4SApple OSS Distributions 
509*a325d9c4SApple OSS Distributions static inline void
workq_unlock(struct workqueue * wq)510*a325d9c4SApple OSS Distributions workq_unlock(struct workqueue *wq)
511*a325d9c4SApple OSS Distributions {
512*a325d9c4SApple OSS Distributions 	lck_ticket_unlock(&wq->wq_lock);
513*a325d9c4SApple OSS Distributions }
514*a325d9c4SApple OSS Distributions 
515*a325d9c4SApple OSS Distributions #pragma mark idle thread lists
516*a325d9c4SApple OSS Distributions 
517*a325d9c4SApple OSS Distributions #define WORKQ_POLICY_INIT(qos) \
518*a325d9c4SApple OSS Distributions 	        (struct uu_workq_policy){ .qos_req = qos, .qos_bucket = qos }
519*a325d9c4SApple OSS Distributions 
520*a325d9c4SApple OSS Distributions static inline thread_qos_t
workq_pri_bucket(struct uu_workq_policy req)521*a325d9c4SApple OSS Distributions workq_pri_bucket(struct uu_workq_policy req)
522*a325d9c4SApple OSS Distributions {
523*a325d9c4SApple OSS Distributions 	return MAX(MAX(req.qos_req, req.qos_max), req.qos_override);
524*a325d9c4SApple OSS Distributions }
525*a325d9c4SApple OSS Distributions 
526*a325d9c4SApple OSS Distributions static inline thread_qos_t
workq_pri_override(struct uu_workq_policy req)527*a325d9c4SApple OSS Distributions workq_pri_override(struct uu_workq_policy req)
528*a325d9c4SApple OSS Distributions {
529*a325d9c4SApple OSS Distributions 	return MAX(workq_pri_bucket(req), req.qos_bucket);
530*a325d9c4SApple OSS Distributions }
531*a325d9c4SApple OSS Distributions 
532*a325d9c4SApple OSS Distributions static inline bool
workq_thread_needs_params_change(workq_threadreq_t req,struct uthread * uth)533*a325d9c4SApple OSS Distributions workq_thread_needs_params_change(workq_threadreq_t req, struct uthread *uth)
534*a325d9c4SApple OSS Distributions {
535*a325d9c4SApple OSS Distributions 	workq_threadreq_param_t cur_trp, req_trp = { };
536*a325d9c4SApple OSS Distributions 
537*a325d9c4SApple OSS Distributions 	cur_trp.trp_value = uth->uu_save.uus_workq_park_data.workloop_params;
538*a325d9c4SApple OSS Distributions 	if (req->tr_flags & WORKQ_TR_FLAG_WL_PARAMS) {
539*a325d9c4SApple OSS Distributions 		req_trp = kqueue_threadreq_workloop_param(req);
540*a325d9c4SApple OSS Distributions 	}
541*a325d9c4SApple OSS Distributions 
542*a325d9c4SApple OSS Distributions 	/*
543*a325d9c4SApple OSS Distributions 	 * CPU percent flags are handled separately to policy changes, so ignore
544*a325d9c4SApple OSS Distributions 	 * them for all of these checks.
545*a325d9c4SApple OSS Distributions 	 */
546*a325d9c4SApple OSS Distributions 	uint16_t cur_flags = (cur_trp.trp_flags & ~TRP_CPUPERCENT);
547*a325d9c4SApple OSS Distributions 	uint16_t req_flags = (req_trp.trp_flags & ~TRP_CPUPERCENT);
548*a325d9c4SApple OSS Distributions 
549*a325d9c4SApple OSS Distributions 	if (!req_flags && !cur_flags) {
550*a325d9c4SApple OSS Distributions 		return false;
551*a325d9c4SApple OSS Distributions 	}
552*a325d9c4SApple OSS Distributions 
553*a325d9c4SApple OSS Distributions 	if (req_flags != cur_flags) {
554*a325d9c4SApple OSS Distributions 		return true;
555*a325d9c4SApple OSS Distributions 	}
556*a325d9c4SApple OSS Distributions 
557*a325d9c4SApple OSS Distributions 	if ((req_flags & TRP_PRIORITY) && req_trp.trp_pri != cur_trp.trp_pri) {
558*a325d9c4SApple OSS Distributions 		return true;
559*a325d9c4SApple OSS Distributions 	}
560*a325d9c4SApple OSS Distributions 
561*a325d9c4SApple OSS Distributions 	if ((req_flags & TRP_POLICY) && req_trp.trp_pol != cur_trp.trp_pol) {
562*a325d9c4SApple OSS Distributions 		return true;
563*a325d9c4SApple OSS Distributions 	}
564*a325d9c4SApple OSS Distributions 
565*a325d9c4SApple OSS Distributions 	return false;
566*a325d9c4SApple OSS Distributions }
567*a325d9c4SApple OSS Distributions 
568*a325d9c4SApple OSS Distributions static inline bool
workq_thread_needs_priority_change(workq_threadreq_t req,struct uthread * uth)569*a325d9c4SApple OSS Distributions workq_thread_needs_priority_change(workq_threadreq_t req, struct uthread *uth)
570*a325d9c4SApple OSS Distributions {
571*a325d9c4SApple OSS Distributions 	if (workq_thread_needs_params_change(req, uth)) {
572*a325d9c4SApple OSS Distributions 		return true;
573*a325d9c4SApple OSS Distributions 	}
574*a325d9c4SApple OSS Distributions 
575*a325d9c4SApple OSS Distributions 	if (req->tr_qos != workq_pri_override(uth->uu_workq_pri)) {
576*a325d9c4SApple OSS Distributions 		return true;
577*a325d9c4SApple OSS Distributions 	}
578*a325d9c4SApple OSS Distributions 
579*a325d9c4SApple OSS Distributions #if CONFIG_PREADOPT_TG
580*a325d9c4SApple OSS Distributions 	thread_group_qos_t tg = kqr_preadopt_thread_group(req);
581*a325d9c4SApple OSS Distributions 	if (KQWL_HAS_VALID_PREADOPTED_TG(tg)) {
582*a325d9c4SApple OSS Distributions 		/*
583*a325d9c4SApple OSS Distributions 		 * Ideally, we'd add check here to see if thread's preadopt TG is same
584*a325d9c4SApple OSS Distributions 		 * as the thread requests's thread group and short circuit if that is
585*a325d9c4SApple OSS Distributions 		 * the case. But in the interest of keeping the code clean and not
586*a325d9c4SApple OSS Distributions 		 * taking the thread lock here, we're going to skip this. We will
587*a325d9c4SApple OSS Distributions 		 * eventually shortcircuit once we try to set the preadoption thread
588*a325d9c4SApple OSS Distributions 		 * group on the thread.
589*a325d9c4SApple OSS Distributions 		 */
590*a325d9c4SApple OSS Distributions 		return true;
591*a325d9c4SApple OSS Distributions 	}
592*a325d9c4SApple OSS Distributions #endif
593*a325d9c4SApple OSS Distributions 
594*a325d9c4SApple OSS Distributions 	return false;
595*a325d9c4SApple OSS Distributions }
596*a325d9c4SApple OSS Distributions 
597*a325d9c4SApple OSS Distributions static void
workq_thread_update_bucket(proc_t p,struct workqueue * wq,struct uthread * uth,struct uu_workq_policy old_pri,struct uu_workq_policy new_pri,bool force_run)598*a325d9c4SApple OSS Distributions workq_thread_update_bucket(proc_t p, struct workqueue *wq, struct uthread *uth,
599*a325d9c4SApple OSS Distributions     struct uu_workq_policy old_pri, struct uu_workq_policy new_pri,
600*a325d9c4SApple OSS Distributions     bool force_run)
601*a325d9c4SApple OSS Distributions {
602*a325d9c4SApple OSS Distributions 	thread_qos_t old_bucket = old_pri.qos_bucket;
603*a325d9c4SApple OSS Distributions 	thread_qos_t new_bucket = workq_pri_bucket(new_pri);
604*a325d9c4SApple OSS Distributions 
605*a325d9c4SApple OSS Distributions 	if (old_bucket != new_bucket) {
606*a325d9c4SApple OSS Distributions 		_wq_thactive_move(wq, old_bucket, new_bucket);
607*a325d9c4SApple OSS Distributions 	}
608*a325d9c4SApple OSS Distributions 
609*a325d9c4SApple OSS Distributions 	new_pri.qos_bucket = new_bucket;
610*a325d9c4SApple OSS Distributions 	uth->uu_workq_pri = new_pri;
611*a325d9c4SApple OSS Distributions 
612*a325d9c4SApple OSS Distributions 	if (workq_pri_override(old_pri) != new_bucket) {
613*a325d9c4SApple OSS Distributions 		thread_set_workq_override(get_machthread(uth), new_bucket);
614*a325d9c4SApple OSS Distributions 	}
615*a325d9c4SApple OSS Distributions 
616*a325d9c4SApple OSS Distributions 	if (wq->wq_reqcount && (old_bucket > new_bucket || force_run)) {
617*a325d9c4SApple OSS Distributions 		int flags = WORKQ_THREADREQ_CAN_CREATE_THREADS;
618*a325d9c4SApple OSS Distributions 		if (old_bucket > new_bucket) {
619*a325d9c4SApple OSS Distributions 			/*
620*a325d9c4SApple OSS Distributions 			 * When lowering our bucket, we may unblock a thread request,
621*a325d9c4SApple OSS Distributions 			 * but we can't drop our priority before we have evaluated
622*a325d9c4SApple OSS Distributions 			 * whether this is the case, and if we ever drop the workqueue lock
623*a325d9c4SApple OSS Distributions 			 * that would cause a priority inversion.
624*a325d9c4SApple OSS Distributions 			 *
625*a325d9c4SApple OSS Distributions 			 * We hence have to disallow thread creation in that case.
626*a325d9c4SApple OSS Distributions 			 */
627*a325d9c4SApple OSS Distributions 			flags = 0;
628*a325d9c4SApple OSS Distributions 		}
629*a325d9c4SApple OSS Distributions 		workq_schedule_creator(p, wq, flags);
630*a325d9c4SApple OSS Distributions 	}
631*a325d9c4SApple OSS Distributions }
632*a325d9c4SApple OSS Distributions 
633*a325d9c4SApple OSS Distributions /*
634*a325d9c4SApple OSS Distributions  * Sets/resets the cpu percent limits on the current thread. We can't set
635*a325d9c4SApple OSS Distributions  * these limits from outside of the current thread, so this function needs
636*a325d9c4SApple OSS Distributions  * to be called when we're executing on the intended
637*a325d9c4SApple OSS Distributions  */
638*a325d9c4SApple OSS Distributions static void
workq_thread_reset_cpupercent(workq_threadreq_t req,struct uthread * uth)639*a325d9c4SApple OSS Distributions workq_thread_reset_cpupercent(workq_threadreq_t req, struct uthread *uth)
640*a325d9c4SApple OSS Distributions {
641*a325d9c4SApple OSS Distributions 	assert(uth == current_uthread());
642*a325d9c4SApple OSS Distributions 	workq_threadreq_param_t trp = { };
643*a325d9c4SApple OSS Distributions 
644*a325d9c4SApple OSS Distributions 	if (req && (req->tr_flags & WORKQ_TR_FLAG_WL_PARAMS)) {
645*a325d9c4SApple OSS Distributions 		trp = kqueue_threadreq_workloop_param(req);
646*a325d9c4SApple OSS Distributions 	}
647*a325d9c4SApple OSS Distributions 
648*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_CPUPERCENT) {
649*a325d9c4SApple OSS Distributions 		/*
650*a325d9c4SApple OSS Distributions 		 * Going through disable when we have an existing CPU percent limit
651*a325d9c4SApple OSS Distributions 		 * set will force the ledger to refill the token bucket of the current
652*a325d9c4SApple OSS Distributions 		 * thread. Removing any penalty applied by previous thread use.
653*a325d9c4SApple OSS Distributions 		 */
654*a325d9c4SApple OSS Distributions 		thread_set_cpulimit(THREAD_CPULIMIT_DISABLE, 0, 0);
655*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags &= ~UT_WORKQ_CPUPERCENT;
656*a325d9c4SApple OSS Distributions 	}
657*a325d9c4SApple OSS Distributions 
658*a325d9c4SApple OSS Distributions 	if (trp.trp_flags & TRP_CPUPERCENT) {
659*a325d9c4SApple OSS Distributions 		thread_set_cpulimit(THREAD_CPULIMIT_BLOCK, trp.trp_cpupercent,
660*a325d9c4SApple OSS Distributions 		    (uint64_t)trp.trp_refillms * NSEC_PER_SEC);
661*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags |= UT_WORKQ_CPUPERCENT;
662*a325d9c4SApple OSS Distributions 	}
663*a325d9c4SApple OSS Distributions }
664*a325d9c4SApple OSS Distributions 
665*a325d9c4SApple OSS Distributions /* Called with the workq lock held */
666*a325d9c4SApple OSS Distributions static void
workq_thread_reset_pri(struct workqueue * wq,struct uthread * uth,workq_threadreq_t req,bool unpark)667*a325d9c4SApple OSS Distributions workq_thread_reset_pri(struct workqueue *wq, struct uthread *uth,
668*a325d9c4SApple OSS Distributions     workq_threadreq_t req, bool unpark)
669*a325d9c4SApple OSS Distributions {
670*a325d9c4SApple OSS Distributions 	thread_t th = get_machthread(uth);
671*a325d9c4SApple OSS Distributions 	thread_qos_t qos = req ? req->tr_qos : WORKQ_THREAD_QOS_CLEANUP;
672*a325d9c4SApple OSS Distributions 	workq_threadreq_param_t trp = { };
673*a325d9c4SApple OSS Distributions 	int priority = 31;
674*a325d9c4SApple OSS Distributions 	int policy = POLICY_TIMESHARE;
675*a325d9c4SApple OSS Distributions 
676*a325d9c4SApple OSS Distributions 	if (req && (req->tr_flags & WORKQ_TR_FLAG_WL_PARAMS)) {
677*a325d9c4SApple OSS Distributions 		trp = kqueue_threadreq_workloop_param(req);
678*a325d9c4SApple OSS Distributions 	}
679*a325d9c4SApple OSS Distributions 
680*a325d9c4SApple OSS Distributions 	uth->uu_workq_pri = WORKQ_POLICY_INIT(qos);
681*a325d9c4SApple OSS Distributions 	uth->uu_workq_flags &= ~UT_WORKQ_OUTSIDE_QOS;
682*a325d9c4SApple OSS Distributions 
683*a325d9c4SApple OSS Distributions 	if (unpark) {
684*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.workloop_params = trp.trp_value;
685*a325d9c4SApple OSS Distributions 		// qos sent out to userspace (may differ from uu_workq_pri on param threads)
686*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.qos = qos;
687*a325d9c4SApple OSS Distributions 	}
688*a325d9c4SApple OSS Distributions 
689*a325d9c4SApple OSS Distributions 	if (qos == WORKQ_THREAD_QOS_MANAGER) {
690*a325d9c4SApple OSS Distributions 		uint32_t mgr_pri = wq->wq_event_manager_priority;
691*a325d9c4SApple OSS Distributions 		assert(trp.trp_value == 0); // manager qos and thread policy don't mix
692*a325d9c4SApple OSS Distributions 
693*a325d9c4SApple OSS Distributions 		if (mgr_pri & _PTHREAD_PRIORITY_SCHED_PRI_FLAG) {
694*a325d9c4SApple OSS Distributions 			mgr_pri &= _PTHREAD_PRIORITY_SCHED_PRI_MASK;
695*a325d9c4SApple OSS Distributions 			thread_set_workq_pri(th, THREAD_QOS_UNSPECIFIED, mgr_pri,
696*a325d9c4SApple OSS Distributions 			    POLICY_TIMESHARE);
697*a325d9c4SApple OSS Distributions 			return;
698*a325d9c4SApple OSS Distributions 		}
699*a325d9c4SApple OSS Distributions 
700*a325d9c4SApple OSS Distributions 		qos = _pthread_priority_thread_qos(mgr_pri);
701*a325d9c4SApple OSS Distributions 	} else {
702*a325d9c4SApple OSS Distributions 		if (trp.trp_flags & TRP_PRIORITY) {
703*a325d9c4SApple OSS Distributions 			qos = THREAD_QOS_UNSPECIFIED;
704*a325d9c4SApple OSS Distributions 			priority = trp.trp_pri;
705*a325d9c4SApple OSS Distributions 			uth->uu_workq_flags |= UT_WORKQ_OUTSIDE_QOS;
706*a325d9c4SApple OSS Distributions 		}
707*a325d9c4SApple OSS Distributions 
708*a325d9c4SApple OSS Distributions 		if (trp.trp_flags & TRP_POLICY) {
709*a325d9c4SApple OSS Distributions 			policy = trp.trp_pol;
710*a325d9c4SApple OSS Distributions 		}
711*a325d9c4SApple OSS Distributions 	}
712*a325d9c4SApple OSS Distributions 
713*a325d9c4SApple OSS Distributions #if CONFIG_PREADOPT_TG
714*a325d9c4SApple OSS Distributions 	if (req && (req->tr_flags & WORKQ_TR_FLAG_WORKLOOP)) {
715*a325d9c4SApple OSS Distributions 		/*
716*a325d9c4SApple OSS Distributions 		 * We cannot safely read and borrow the reference from the kqwl since it
717*a325d9c4SApple OSS Distributions 		 * can disappear from under us at any time due to the max-ing logic in
718*a325d9c4SApple OSS Distributions 		 * kqueue_set_preadopted_thread_group.
719*a325d9c4SApple OSS Distributions 		 *
720*a325d9c4SApple OSS Distributions 		 * As such, we do the following dance:
721*a325d9c4SApple OSS Distributions 		 *
722*a325d9c4SApple OSS Distributions 		 * 1) cmpxchng and steal the kqwl's preadopt thread group and leave
723*a325d9c4SApple OSS Distributions 		 * behind with (NULL + QoS). At this point, we have the reference
724*a325d9c4SApple OSS Distributions 		 * to the thread group from the kqwl.
725*a325d9c4SApple OSS Distributions 		 * 2) Have the thread set the preadoption thread group on itself.
726*a325d9c4SApple OSS Distributions 		 * 3) cmpxchng from (NULL + QoS) which we set earlier in (1), back to
727*a325d9c4SApple OSS Distributions 		 * thread_group + QoS. ie we try to give the reference back to the kqwl.
728*a325d9c4SApple OSS Distributions 		 * If we fail, that's because a higher QoS thread group was set on the
729*a325d9c4SApple OSS Distributions 		 * kqwl in kqueue_set_preadopted_thread_group in which case, we need to
730*a325d9c4SApple OSS Distributions 		 * go back to (1).
731*a325d9c4SApple OSS Distributions 		 */
732*a325d9c4SApple OSS Distributions 
733*a325d9c4SApple OSS Distributions 		_Atomic(struct thread_group *) * tg_loc = kqr_preadopt_thread_group_addr(req);
734*a325d9c4SApple OSS Distributions 
735*a325d9c4SApple OSS Distributions 		thread_group_qos_t old_tg, new_tg;
736*a325d9c4SApple OSS Distributions 		int ret = 0;
737*a325d9c4SApple OSS Distributions again:
738*a325d9c4SApple OSS Distributions 		ret = os_atomic_rmw_loop(tg_loc, old_tg, new_tg, relaxed, {
739*a325d9c4SApple OSS Distributions 			if (!KQWL_HAS_VALID_PREADOPTED_TG(old_tg)) {
740*a325d9c4SApple OSS Distributions 			        os_atomic_rmw_loop_give_up(break);
741*a325d9c4SApple OSS Distributions 			}
742*a325d9c4SApple OSS Distributions 
743*a325d9c4SApple OSS Distributions 			/*
744*a325d9c4SApple OSS Distributions 			 * Leave the QoS behind - kqueue_set_preadopted_thread_group will
745*a325d9c4SApple OSS Distributions 			 * only modify it if there is a higher QoS thread group to attach
746*a325d9c4SApple OSS Distributions 			 */
747*a325d9c4SApple OSS Distributions 			new_tg = (thread_group_qos_t) ((uintptr_t) old_tg & KQWL_PREADOPT_TG_QOS_MASK);
748*a325d9c4SApple OSS Distributions 		});
749*a325d9c4SApple OSS Distributions 
750*a325d9c4SApple OSS Distributions 		if (ret) {
751*a325d9c4SApple OSS Distributions 			/*
752*a325d9c4SApple OSS Distributions 			 * We successfully took the ref from the kqwl so set it on the
753*a325d9c4SApple OSS Distributions 			 * thread now
754*a325d9c4SApple OSS Distributions 			 */
755*a325d9c4SApple OSS Distributions 			thread_set_preadopt_thread_group(th, KQWL_GET_PREADOPTED_TG(old_tg));
756*a325d9c4SApple OSS Distributions 
757*a325d9c4SApple OSS Distributions 			thread_group_qos_t thread_group_to_expect = new_tg;
758*a325d9c4SApple OSS Distributions 			thread_group_qos_t thread_group_to_set = old_tg;
759*a325d9c4SApple OSS Distributions 
760*a325d9c4SApple OSS Distributions 			os_atomic_rmw_loop(tg_loc, old_tg, new_tg, relaxed, {
761*a325d9c4SApple OSS Distributions 				if (old_tg != thread_group_to_expect) {
762*a325d9c4SApple OSS Distributions 				        /*
763*a325d9c4SApple OSS Distributions 				         * There was an intervening write to the kqwl_preadopt_tg,
764*a325d9c4SApple OSS Distributions 				         * and it has a higher QoS than what we are working with
765*a325d9c4SApple OSS Distributions 				         * here. Abandon our current adopted thread group and redo
766*a325d9c4SApple OSS Distributions 				         * the full dance
767*a325d9c4SApple OSS Distributions 				         */
768*a325d9c4SApple OSS Distributions 				        thread_group_deallocate_safe(KQWL_GET_PREADOPTED_TG(thread_group_to_set));
769*a325d9c4SApple OSS Distributions 				        os_atomic_rmw_loop_give_up(goto again);
770*a325d9c4SApple OSS Distributions 				}
771*a325d9c4SApple OSS Distributions 
772*a325d9c4SApple OSS Distributions 				new_tg = thread_group_to_set;
773*a325d9c4SApple OSS Distributions 			});
774*a325d9c4SApple OSS Distributions 		} else {
775*a325d9c4SApple OSS Distributions 			/* Nothing valid on the kqwl, just clear what's on the thread */
776*a325d9c4SApple OSS Distributions 			thread_set_preadopt_thread_group(th, NULL);
777*a325d9c4SApple OSS Distributions 		}
778*a325d9c4SApple OSS Distributions 	} else {
779*a325d9c4SApple OSS Distributions 		/* Not even a kqwl, clear what's on the thread */
780*a325d9c4SApple OSS Distributions 		thread_set_preadopt_thread_group(th, NULL);
781*a325d9c4SApple OSS Distributions 	}
782*a325d9c4SApple OSS Distributions #endif
783*a325d9c4SApple OSS Distributions 	thread_set_workq_pri(th, qos, priority, policy);
784*a325d9c4SApple OSS Distributions }
785*a325d9c4SApple OSS Distributions 
786*a325d9c4SApple OSS Distributions /*
787*a325d9c4SApple OSS Distributions  * Called by kevent with the NOTE_WL_THREAD_REQUEST knote lock held,
788*a325d9c4SApple OSS Distributions  * every time a servicer is being told about a new max QoS.
789*a325d9c4SApple OSS Distributions  */
790*a325d9c4SApple OSS Distributions void
workq_thread_set_max_qos(struct proc * p,workq_threadreq_t kqr)791*a325d9c4SApple OSS Distributions workq_thread_set_max_qos(struct proc *p, workq_threadreq_t kqr)
792*a325d9c4SApple OSS Distributions {
793*a325d9c4SApple OSS Distributions 	struct uu_workq_policy old_pri, new_pri;
794*a325d9c4SApple OSS Distributions 	struct uthread *uth = current_uthread();
795*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
796*a325d9c4SApple OSS Distributions 	thread_qos_t qos = kqr->tr_kq_qos_index;
797*a325d9c4SApple OSS Distributions 
798*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_pri.qos_max == qos) {
799*a325d9c4SApple OSS Distributions 		return;
800*a325d9c4SApple OSS Distributions 	}
801*a325d9c4SApple OSS Distributions 
802*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
803*a325d9c4SApple OSS Distributions 	old_pri = new_pri = uth->uu_workq_pri;
804*a325d9c4SApple OSS Distributions 	new_pri.qos_max = qos;
805*a325d9c4SApple OSS Distributions 	workq_thread_update_bucket(p, wq, uth, old_pri, new_pri, false);
806*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
807*a325d9c4SApple OSS Distributions }
808*a325d9c4SApple OSS Distributions 
809*a325d9c4SApple OSS Distributions #pragma mark idle threads accounting and handling
810*a325d9c4SApple OSS Distributions 
811*a325d9c4SApple OSS Distributions static inline struct uthread *
workq_oldest_killable_idle_thread(struct workqueue * wq)812*a325d9c4SApple OSS Distributions workq_oldest_killable_idle_thread(struct workqueue *wq)
813*a325d9c4SApple OSS Distributions {
814*a325d9c4SApple OSS Distributions 	struct uthread *uth = TAILQ_LAST(&wq->wq_thidlelist, workq_uthread_head);
815*a325d9c4SApple OSS Distributions 
816*a325d9c4SApple OSS Distributions 	if (uth && !uth->uu_save.uus_workq_park_data.has_stack) {
817*a325d9c4SApple OSS Distributions 		uth = TAILQ_PREV(uth, workq_uthread_head, uu_workq_entry);
818*a325d9c4SApple OSS Distributions 		if (uth) {
819*a325d9c4SApple OSS Distributions 			assert(uth->uu_save.uus_workq_park_data.has_stack);
820*a325d9c4SApple OSS Distributions 		}
821*a325d9c4SApple OSS Distributions 	}
822*a325d9c4SApple OSS Distributions 	return uth;
823*a325d9c4SApple OSS Distributions }
824*a325d9c4SApple OSS Distributions 
825*a325d9c4SApple OSS Distributions static inline uint64_t
workq_kill_delay_for_idle_thread(struct workqueue * wq)826*a325d9c4SApple OSS Distributions workq_kill_delay_for_idle_thread(struct workqueue *wq)
827*a325d9c4SApple OSS Distributions {
828*a325d9c4SApple OSS Distributions 	uint64_t delay = wq_reduce_pool_window.abstime;
829*a325d9c4SApple OSS Distributions 	uint16_t idle = wq->wq_thidlecount;
830*a325d9c4SApple OSS Distributions 
831*a325d9c4SApple OSS Distributions 	/*
832*a325d9c4SApple OSS Distributions 	 * If we have less than wq_death_max_load threads, have a 5s timer.
833*a325d9c4SApple OSS Distributions 	 *
834*a325d9c4SApple OSS Distributions 	 * For the next wq_max_constrained_threads ones, decay linearly from
835*a325d9c4SApple OSS Distributions 	 * from 5s to 50ms.
836*a325d9c4SApple OSS Distributions 	 */
837*a325d9c4SApple OSS Distributions 	if (idle <= wq_death_max_load) {
838*a325d9c4SApple OSS Distributions 		return delay;
839*a325d9c4SApple OSS Distributions 	}
840*a325d9c4SApple OSS Distributions 
841*a325d9c4SApple OSS Distributions 	if (wq_max_constrained_threads > idle - wq_death_max_load) {
842*a325d9c4SApple OSS Distributions 		delay *= (wq_max_constrained_threads - (idle - wq_death_max_load));
843*a325d9c4SApple OSS Distributions 	}
844*a325d9c4SApple OSS Distributions 	return delay / wq_max_constrained_threads;
845*a325d9c4SApple OSS Distributions }
846*a325d9c4SApple OSS Distributions 
847*a325d9c4SApple OSS Distributions static inline bool
workq_should_kill_idle_thread(struct workqueue * wq,struct uthread * uth,uint64_t now)848*a325d9c4SApple OSS Distributions workq_should_kill_idle_thread(struct workqueue *wq, struct uthread *uth,
849*a325d9c4SApple OSS Distributions     uint64_t now)
850*a325d9c4SApple OSS Distributions {
851*a325d9c4SApple OSS Distributions 	uint64_t delay = workq_kill_delay_for_idle_thread(wq);
852*a325d9c4SApple OSS Distributions 	return now - uth->uu_save.uus_workq_park_data.idle_stamp > delay;
853*a325d9c4SApple OSS Distributions }
854*a325d9c4SApple OSS Distributions 
855*a325d9c4SApple OSS Distributions static void
workq_death_call_schedule(struct workqueue * wq,uint64_t deadline)856*a325d9c4SApple OSS Distributions workq_death_call_schedule(struct workqueue *wq, uint64_t deadline)
857*a325d9c4SApple OSS Distributions {
858*a325d9c4SApple OSS Distributions 	uint32_t wq_flags = os_atomic_load(&wq->wq_flags, relaxed);
859*a325d9c4SApple OSS Distributions 
860*a325d9c4SApple OSS Distributions 	if (wq_flags & (WQ_EXITING | WQ_DEATH_CALL_SCHEDULED)) {
861*a325d9c4SApple OSS Distributions 		return;
862*a325d9c4SApple OSS Distributions 	}
863*a325d9c4SApple OSS Distributions 	os_atomic_or(&wq->wq_flags, WQ_DEATH_CALL_SCHEDULED, relaxed);
864*a325d9c4SApple OSS Distributions 
865*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_death_call | DBG_FUNC_NONE, wq, 1, 0, 0);
866*a325d9c4SApple OSS Distributions 
867*a325d9c4SApple OSS Distributions 	/*
868*a325d9c4SApple OSS Distributions 	 * <rdar://problem/13139182> Due to how long term timers work, the leeway
869*a325d9c4SApple OSS Distributions 	 * can't be too short, so use 500ms which is long enough that we will not
870*a325d9c4SApple OSS Distributions 	 * wake up the CPU for killing threads, but short enough that it doesn't
871*a325d9c4SApple OSS Distributions 	 * fall into long-term timer list shenanigans.
872*a325d9c4SApple OSS Distributions 	 */
873*a325d9c4SApple OSS Distributions 	thread_call_enter_delayed_with_leeway(wq->wq_death_call, NULL, deadline,
874*a325d9c4SApple OSS Distributions 	    wq_reduce_pool_window.abstime / 10,
875*a325d9c4SApple OSS Distributions 	    THREAD_CALL_DELAY_LEEWAY | THREAD_CALL_DELAY_USER_BACKGROUND);
876*a325d9c4SApple OSS Distributions }
877*a325d9c4SApple OSS Distributions 
878*a325d9c4SApple OSS Distributions /*
879*a325d9c4SApple OSS Distributions  * `decrement` is set to the number of threads that are no longer dying:
880*a325d9c4SApple OSS Distributions  * - because they have been resuscitated just in time (workq_pop_idle_thread)
881*a325d9c4SApple OSS Distributions  * - or have been killed (workq_thread_terminate).
882*a325d9c4SApple OSS Distributions  */
883*a325d9c4SApple OSS Distributions static void
workq_death_policy_evaluate(struct workqueue * wq,uint16_t decrement)884*a325d9c4SApple OSS Distributions workq_death_policy_evaluate(struct workqueue *wq, uint16_t decrement)
885*a325d9c4SApple OSS Distributions {
886*a325d9c4SApple OSS Distributions 	struct uthread *uth;
887*a325d9c4SApple OSS Distributions 
888*a325d9c4SApple OSS Distributions 	assert(wq->wq_thdying_count >= decrement);
889*a325d9c4SApple OSS Distributions 	if ((wq->wq_thdying_count -= decrement) > 0) {
890*a325d9c4SApple OSS Distributions 		return;
891*a325d9c4SApple OSS Distributions 	}
892*a325d9c4SApple OSS Distributions 
893*a325d9c4SApple OSS Distributions 	if (wq->wq_thidlecount <= 1) {
894*a325d9c4SApple OSS Distributions 		return;
895*a325d9c4SApple OSS Distributions 	}
896*a325d9c4SApple OSS Distributions 
897*a325d9c4SApple OSS Distributions 	if ((uth = workq_oldest_killable_idle_thread(wq)) == NULL) {
898*a325d9c4SApple OSS Distributions 		return;
899*a325d9c4SApple OSS Distributions 	}
900*a325d9c4SApple OSS Distributions 
901*a325d9c4SApple OSS Distributions 	uint64_t now = mach_absolute_time();
902*a325d9c4SApple OSS Distributions 	uint64_t delay = workq_kill_delay_for_idle_thread(wq);
903*a325d9c4SApple OSS Distributions 
904*a325d9c4SApple OSS Distributions 	if (now - uth->uu_save.uus_workq_park_data.idle_stamp > delay) {
905*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thread_terminate | DBG_FUNC_START,
906*a325d9c4SApple OSS Distributions 		    wq, wq->wq_thidlecount, 0, 0);
907*a325d9c4SApple OSS Distributions 		wq->wq_thdying_count++;
908*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags |= UT_WORKQ_DYING;
909*a325d9c4SApple OSS Distributions 		if ((uth->uu_workq_flags & UT_WORKQ_IDLE_CLEANUP) == 0) {
910*a325d9c4SApple OSS Distributions 			workq_thread_wakeup(uth);
911*a325d9c4SApple OSS Distributions 		}
912*a325d9c4SApple OSS Distributions 		return;
913*a325d9c4SApple OSS Distributions 	}
914*a325d9c4SApple OSS Distributions 
915*a325d9c4SApple OSS Distributions 	workq_death_call_schedule(wq,
916*a325d9c4SApple OSS Distributions 	    uth->uu_save.uus_workq_park_data.idle_stamp + delay);
917*a325d9c4SApple OSS Distributions }
918*a325d9c4SApple OSS Distributions 
919*a325d9c4SApple OSS Distributions void
workq_thread_terminate(struct proc * p,struct uthread * uth)920*a325d9c4SApple OSS Distributions workq_thread_terminate(struct proc *p, struct uthread *uth)
921*a325d9c4SApple OSS Distributions {
922*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
923*a325d9c4SApple OSS Distributions 
924*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
925*a325d9c4SApple OSS Distributions 	TAILQ_REMOVE(&wq->wq_thrunlist, uth, uu_workq_entry);
926*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_DYING) {
927*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thread_terminate | DBG_FUNC_END,
928*a325d9c4SApple OSS Distributions 		    wq, wq->wq_thidlecount, 0, 0);
929*a325d9c4SApple OSS Distributions 		workq_death_policy_evaluate(wq, 1);
930*a325d9c4SApple OSS Distributions 	}
931*a325d9c4SApple OSS Distributions 	if (wq->wq_nthreads-- == wq_max_threads) {
932*a325d9c4SApple OSS Distributions 		/*
933*a325d9c4SApple OSS Distributions 		 * We got under the thread limit again, which may have prevented
934*a325d9c4SApple OSS Distributions 		 * thread creation from happening, redrive if there are pending requests
935*a325d9c4SApple OSS Distributions 		 */
936*a325d9c4SApple OSS Distributions 		if (wq->wq_reqcount) {
937*a325d9c4SApple OSS Distributions 			workq_schedule_creator(p, wq, WORKQ_THREADREQ_CAN_CREATE_THREADS);
938*a325d9c4SApple OSS Distributions 		}
939*a325d9c4SApple OSS Distributions 	}
940*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
941*a325d9c4SApple OSS Distributions 
942*a325d9c4SApple OSS Distributions 	thread_deallocate(get_machthread(uth));
943*a325d9c4SApple OSS Distributions }
944*a325d9c4SApple OSS Distributions 
945*a325d9c4SApple OSS Distributions static void
workq_kill_old_threads_call(void * param0,void * param1 __unused)946*a325d9c4SApple OSS Distributions workq_kill_old_threads_call(void *param0, void *param1 __unused)
947*a325d9c4SApple OSS Distributions {
948*a325d9c4SApple OSS Distributions 	struct workqueue *wq = param0;
949*a325d9c4SApple OSS Distributions 
950*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
951*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_death_call | DBG_FUNC_START, wq, 0, 0, 0);
952*a325d9c4SApple OSS Distributions 	os_atomic_andnot(&wq->wq_flags, WQ_DEATH_CALL_SCHEDULED, relaxed);
953*a325d9c4SApple OSS Distributions 	workq_death_policy_evaluate(wq, 0);
954*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_death_call | DBG_FUNC_END, wq, 0, 0, 0);
955*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
956*a325d9c4SApple OSS Distributions }
957*a325d9c4SApple OSS Distributions 
958*a325d9c4SApple OSS Distributions static struct uthread *
workq_pop_idle_thread(struct workqueue * wq,uint16_t uu_flags,bool * needs_wakeup)959*a325d9c4SApple OSS Distributions workq_pop_idle_thread(struct workqueue *wq, uint16_t uu_flags,
960*a325d9c4SApple OSS Distributions     bool *needs_wakeup)
961*a325d9c4SApple OSS Distributions {
962*a325d9c4SApple OSS Distributions 	struct uthread *uth;
963*a325d9c4SApple OSS Distributions 
964*a325d9c4SApple OSS Distributions 	if ((uth = TAILQ_FIRST(&wq->wq_thidlelist))) {
965*a325d9c4SApple OSS Distributions 		TAILQ_REMOVE(&wq->wq_thidlelist, uth, uu_workq_entry);
966*a325d9c4SApple OSS Distributions 	} else {
967*a325d9c4SApple OSS Distributions 		uth = TAILQ_FIRST(&wq->wq_thnewlist);
968*a325d9c4SApple OSS Distributions 		TAILQ_REMOVE(&wq->wq_thnewlist, uth, uu_workq_entry);
969*a325d9c4SApple OSS Distributions 	}
970*a325d9c4SApple OSS Distributions 	TAILQ_INSERT_TAIL(&wq->wq_thrunlist, uth, uu_workq_entry);
971*a325d9c4SApple OSS Distributions 
972*a325d9c4SApple OSS Distributions 	assert((uth->uu_workq_flags & UT_WORKQ_RUNNING) == 0);
973*a325d9c4SApple OSS Distributions 	uth->uu_workq_flags |= UT_WORKQ_RUNNING | uu_flags;
974*a325d9c4SApple OSS Distributions 
975*a325d9c4SApple OSS Distributions 	/* A thread is never woken up as part of the cooperative pool */
976*a325d9c4SApple OSS Distributions 	assert((uu_flags & UT_WORKQ_COOPERATIVE) == 0);
977*a325d9c4SApple OSS Distributions 
978*a325d9c4SApple OSS Distributions 	if ((uu_flags & UT_WORKQ_OVERCOMMIT) == 0) {
979*a325d9c4SApple OSS Distributions 		wq->wq_constrained_threads_scheduled++;
980*a325d9c4SApple OSS Distributions 	}
981*a325d9c4SApple OSS Distributions 	wq->wq_threads_scheduled++;
982*a325d9c4SApple OSS Distributions 	wq->wq_thidlecount--;
983*a325d9c4SApple OSS Distributions 
984*a325d9c4SApple OSS Distributions 	if (__improbable(uth->uu_workq_flags & UT_WORKQ_DYING)) {
985*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags ^= UT_WORKQ_DYING;
986*a325d9c4SApple OSS Distributions 		workq_death_policy_evaluate(wq, 1);
987*a325d9c4SApple OSS Distributions 		*needs_wakeup = false;
988*a325d9c4SApple OSS Distributions 	} else if (uth->uu_workq_flags & UT_WORKQ_IDLE_CLEANUP) {
989*a325d9c4SApple OSS Distributions 		*needs_wakeup = false;
990*a325d9c4SApple OSS Distributions 	} else {
991*a325d9c4SApple OSS Distributions 		*needs_wakeup = true;
992*a325d9c4SApple OSS Distributions 	}
993*a325d9c4SApple OSS Distributions 	return uth;
994*a325d9c4SApple OSS Distributions }
995*a325d9c4SApple OSS Distributions 
996*a325d9c4SApple OSS Distributions /*
997*a325d9c4SApple OSS Distributions  * Called by thread_create_workq_waiting() during thread initialization, before
998*a325d9c4SApple OSS Distributions  * assert_wait, before the thread has been started.
999*a325d9c4SApple OSS Distributions  */
1000*a325d9c4SApple OSS Distributions event_t
workq_thread_init_and_wq_lock(task_t task,thread_t th)1001*a325d9c4SApple OSS Distributions workq_thread_init_and_wq_lock(task_t task, thread_t th)
1002*a325d9c4SApple OSS Distributions {
1003*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(th);
1004*a325d9c4SApple OSS Distributions 
1005*a325d9c4SApple OSS Distributions 	uth->uu_workq_flags = UT_WORKQ_NEW;
1006*a325d9c4SApple OSS Distributions 	uth->uu_workq_pri = WORKQ_POLICY_INIT(THREAD_QOS_LEGACY);
1007*a325d9c4SApple OSS Distributions 	uth->uu_workq_thport = MACH_PORT_NULL;
1008*a325d9c4SApple OSS Distributions 	uth->uu_workq_stackaddr = 0;
1009*a325d9c4SApple OSS Distributions 	uth->uu_workq_pthread_kill_allowed = 0;
1010*a325d9c4SApple OSS Distributions 
1011*a325d9c4SApple OSS Distributions 	thread_set_tag(th, THREAD_TAG_PTHREAD | THREAD_TAG_WORKQUEUE);
1012*a325d9c4SApple OSS Distributions 	thread_reset_workq_qos(th, THREAD_QOS_LEGACY);
1013*a325d9c4SApple OSS Distributions 
1014*a325d9c4SApple OSS Distributions 	workq_lock_spin(proc_get_wqptr_fast(get_bsdtask_info(task)));
1015*a325d9c4SApple OSS Distributions 	return workq_parked_wait_event(uth);
1016*a325d9c4SApple OSS Distributions }
1017*a325d9c4SApple OSS Distributions 
1018*a325d9c4SApple OSS Distributions /**
1019*a325d9c4SApple OSS Distributions  * Try to add a new workqueue thread.
1020*a325d9c4SApple OSS Distributions  *
1021*a325d9c4SApple OSS Distributions  * - called with workq lock held
1022*a325d9c4SApple OSS Distributions  * - dropped and retaken around thread creation
1023*a325d9c4SApple OSS Distributions  * - return with workq lock held
1024*a325d9c4SApple OSS Distributions  */
1025*a325d9c4SApple OSS Distributions static bool
workq_add_new_idle_thread(proc_t p,struct workqueue * wq)1026*a325d9c4SApple OSS Distributions workq_add_new_idle_thread(proc_t p, struct workqueue *wq)
1027*a325d9c4SApple OSS Distributions {
1028*a325d9c4SApple OSS Distributions 	mach_vm_offset_t th_stackaddr;
1029*a325d9c4SApple OSS Distributions 	kern_return_t kret;
1030*a325d9c4SApple OSS Distributions 	thread_t th;
1031*a325d9c4SApple OSS Distributions 
1032*a325d9c4SApple OSS Distributions 	wq->wq_nthreads++;
1033*a325d9c4SApple OSS Distributions 
1034*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
1035*a325d9c4SApple OSS Distributions 
1036*a325d9c4SApple OSS Distributions 	vm_map_t vmap = get_task_map(p->task);
1037*a325d9c4SApple OSS Distributions 
1038*a325d9c4SApple OSS Distributions 	kret = pthread_functions->workq_create_threadstack(p, vmap, &th_stackaddr);
1039*a325d9c4SApple OSS Distributions 	if (kret != KERN_SUCCESS) {
1040*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thread_create_failed | DBG_FUNC_NONE, wq,
1041*a325d9c4SApple OSS Distributions 		    kret, 1, 0);
1042*a325d9c4SApple OSS Distributions 		goto out;
1043*a325d9c4SApple OSS Distributions 	}
1044*a325d9c4SApple OSS Distributions 
1045*a325d9c4SApple OSS Distributions 	kret = thread_create_workq_waiting(p->task, workq_unpark_continue, &th);
1046*a325d9c4SApple OSS Distributions 	if (kret != KERN_SUCCESS) {
1047*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thread_create_failed | DBG_FUNC_NONE, wq,
1048*a325d9c4SApple OSS Distributions 		    kret, 0, 0);
1049*a325d9c4SApple OSS Distributions 		pthread_functions->workq_destroy_threadstack(p, vmap, th_stackaddr);
1050*a325d9c4SApple OSS Distributions 		goto out;
1051*a325d9c4SApple OSS Distributions 	}
1052*a325d9c4SApple OSS Distributions 
1053*a325d9c4SApple OSS Distributions 	// thread_create_workq_waiting() will return with the wq lock held
1054*a325d9c4SApple OSS Distributions 	// on success, because it calls workq_thread_init_and_wq_lock() above
1055*a325d9c4SApple OSS Distributions 
1056*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(th);
1057*a325d9c4SApple OSS Distributions 
1058*a325d9c4SApple OSS Distributions 	wq->wq_creations++;
1059*a325d9c4SApple OSS Distributions 	wq->wq_thidlecount++;
1060*a325d9c4SApple OSS Distributions 	uth->uu_workq_stackaddr = (user_addr_t)th_stackaddr;
1061*a325d9c4SApple OSS Distributions 	TAILQ_INSERT_TAIL(&wq->wq_thnewlist, uth, uu_workq_entry);
1062*a325d9c4SApple OSS Distributions 
1063*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_thread_create | DBG_FUNC_NONE, wq, 0, 0, 0);
1064*a325d9c4SApple OSS Distributions 	return true;
1065*a325d9c4SApple OSS Distributions 
1066*a325d9c4SApple OSS Distributions out:
1067*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
1068*a325d9c4SApple OSS Distributions 	/*
1069*a325d9c4SApple OSS Distributions 	 * Do not redrive here if we went under wq_max_threads again,
1070*a325d9c4SApple OSS Distributions 	 * it is the responsibility of the callers of this function
1071*a325d9c4SApple OSS Distributions 	 * to do so when it fails.
1072*a325d9c4SApple OSS Distributions 	 */
1073*a325d9c4SApple OSS Distributions 	wq->wq_nthreads--;
1074*a325d9c4SApple OSS Distributions 	return false;
1075*a325d9c4SApple OSS Distributions }
1076*a325d9c4SApple OSS Distributions 
1077*a325d9c4SApple OSS Distributions static inline bool
workq_thread_is_overcommit(struct uthread * uth)1078*a325d9c4SApple OSS Distributions workq_thread_is_overcommit(struct uthread *uth)
1079*a325d9c4SApple OSS Distributions {
1080*a325d9c4SApple OSS Distributions 	return (uth->uu_workq_flags & UT_WORKQ_OVERCOMMIT) != 0;
1081*a325d9c4SApple OSS Distributions }
1082*a325d9c4SApple OSS Distributions 
1083*a325d9c4SApple OSS Distributions static inline bool
workq_thread_is_nonovercommit(struct uthread * uth)1084*a325d9c4SApple OSS Distributions workq_thread_is_nonovercommit(struct uthread *uth)
1085*a325d9c4SApple OSS Distributions {
1086*a325d9c4SApple OSS Distributions 	return (uth->uu_workq_flags & (UT_WORKQ_OVERCOMMIT | UT_WORKQ_COOPERATIVE)) == 0;
1087*a325d9c4SApple OSS Distributions }
1088*a325d9c4SApple OSS Distributions 
1089*a325d9c4SApple OSS Distributions static inline bool
workq_thread_is_cooperative(struct uthread * uth)1090*a325d9c4SApple OSS Distributions workq_thread_is_cooperative(struct uthread *uth)
1091*a325d9c4SApple OSS Distributions {
1092*a325d9c4SApple OSS Distributions 	return (uth->uu_workq_flags & UT_WORKQ_COOPERATIVE) != 0;
1093*a325d9c4SApple OSS Distributions }
1094*a325d9c4SApple OSS Distributions 
1095*a325d9c4SApple OSS Distributions static inline void
workq_thread_set_type(struct uthread * uth,uint16_t flags)1096*a325d9c4SApple OSS Distributions workq_thread_set_type(struct uthread *uth, uint16_t flags)
1097*a325d9c4SApple OSS Distributions {
1098*a325d9c4SApple OSS Distributions 	uth->uu_workq_flags &= ~(UT_WORKQ_OVERCOMMIT | UT_WORKQ_COOPERATIVE);
1099*a325d9c4SApple OSS Distributions 	uth->uu_workq_flags |= flags;
1100*a325d9c4SApple OSS Distributions }
1101*a325d9c4SApple OSS Distributions 
1102*a325d9c4SApple OSS Distributions 
1103*a325d9c4SApple OSS Distributions #define WORKQ_UNPARK_FOR_DEATH_WAS_IDLE 0x1
1104*a325d9c4SApple OSS Distributions 
1105*a325d9c4SApple OSS Distributions __attribute__((noreturn, noinline))
1106*a325d9c4SApple OSS Distributions static void
workq_unpark_for_death_and_unlock(proc_t p,struct workqueue * wq,struct uthread * uth,uint32_t death_flags,uint32_t setup_flags)1107*a325d9c4SApple OSS Distributions workq_unpark_for_death_and_unlock(proc_t p, struct workqueue *wq,
1108*a325d9c4SApple OSS Distributions     struct uthread *uth, uint32_t death_flags, uint32_t setup_flags)
1109*a325d9c4SApple OSS Distributions {
1110*a325d9c4SApple OSS Distributions 	thread_qos_t qos = workq_pri_override(uth->uu_workq_pri);
1111*a325d9c4SApple OSS Distributions 	bool first_use = uth->uu_workq_flags & UT_WORKQ_NEW;
1112*a325d9c4SApple OSS Distributions 
1113*a325d9c4SApple OSS Distributions 	if (qos > WORKQ_THREAD_QOS_CLEANUP) {
1114*a325d9c4SApple OSS Distributions 		workq_thread_reset_pri(wq, uth, NULL, /*unpark*/ true);
1115*a325d9c4SApple OSS Distributions 		qos = WORKQ_THREAD_QOS_CLEANUP;
1116*a325d9c4SApple OSS Distributions 	}
1117*a325d9c4SApple OSS Distributions 
1118*a325d9c4SApple OSS Distributions 	workq_thread_reset_cpupercent(NULL, uth);
1119*a325d9c4SApple OSS Distributions 
1120*a325d9c4SApple OSS Distributions 	if (death_flags & WORKQ_UNPARK_FOR_DEATH_WAS_IDLE) {
1121*a325d9c4SApple OSS Distributions 		wq->wq_thidlecount--;
1122*a325d9c4SApple OSS Distributions 		if (first_use) {
1123*a325d9c4SApple OSS Distributions 			TAILQ_REMOVE(&wq->wq_thnewlist, uth, uu_workq_entry);
1124*a325d9c4SApple OSS Distributions 		} else {
1125*a325d9c4SApple OSS Distributions 			TAILQ_REMOVE(&wq->wq_thidlelist, uth, uu_workq_entry);
1126*a325d9c4SApple OSS Distributions 		}
1127*a325d9c4SApple OSS Distributions 	}
1128*a325d9c4SApple OSS Distributions 	TAILQ_INSERT_TAIL(&wq->wq_thrunlist, uth, uu_workq_entry);
1129*a325d9c4SApple OSS Distributions 
1130*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
1131*a325d9c4SApple OSS Distributions 
1132*a325d9c4SApple OSS Distributions 	if (setup_flags & WQ_SETUP_CLEAR_VOUCHER) {
1133*a325d9c4SApple OSS Distributions 		__assert_only kern_return_t kr;
1134*a325d9c4SApple OSS Distributions 		kr = thread_set_voucher_name(MACH_PORT_NULL);
1135*a325d9c4SApple OSS Distributions 		assert(kr == KERN_SUCCESS);
1136*a325d9c4SApple OSS Distributions 	}
1137*a325d9c4SApple OSS Distributions 
1138*a325d9c4SApple OSS Distributions 	uint32_t flags = WQ_FLAG_THREAD_NEWSPI | qos | WQ_FLAG_THREAD_PRIO_QOS;
1139*a325d9c4SApple OSS Distributions 	thread_t th = get_machthread(uth);
1140*a325d9c4SApple OSS Distributions 	vm_map_t vmap = get_task_map(p->task);
1141*a325d9c4SApple OSS Distributions 
1142*a325d9c4SApple OSS Distributions 	if (!first_use) {
1143*a325d9c4SApple OSS Distributions 		flags |= WQ_FLAG_THREAD_REUSE;
1144*a325d9c4SApple OSS Distributions 	}
1145*a325d9c4SApple OSS Distributions 
1146*a325d9c4SApple OSS Distributions 	pthread_functions->workq_setup_thread(p, th, vmap, uth->uu_workq_stackaddr,
1147*a325d9c4SApple OSS Distributions 	    uth->uu_workq_thport, 0, WQ_SETUP_EXIT_THREAD, flags);
1148*a325d9c4SApple OSS Distributions 	__builtin_unreachable();
1149*a325d9c4SApple OSS Distributions }
1150*a325d9c4SApple OSS Distributions 
1151*a325d9c4SApple OSS Distributions bool
workq_is_current_thread_updating_turnstile(struct workqueue * wq)1152*a325d9c4SApple OSS Distributions workq_is_current_thread_updating_turnstile(struct workqueue *wq)
1153*a325d9c4SApple OSS Distributions {
1154*a325d9c4SApple OSS Distributions 	return wq->wq_turnstile_updater == current_thread();
1155*a325d9c4SApple OSS Distributions }
1156*a325d9c4SApple OSS Distributions 
1157*a325d9c4SApple OSS Distributions __attribute__((always_inline))
1158*a325d9c4SApple OSS Distributions static inline void
1159*a325d9c4SApple OSS Distributions workq_perform_turnstile_operation_locked(struct workqueue *wq,
1160*a325d9c4SApple OSS Distributions     void (^operation)(void))
1161*a325d9c4SApple OSS Distributions {
1162*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
1163*a325d9c4SApple OSS Distributions 	wq->wq_turnstile_updater = current_thread();
1164*a325d9c4SApple OSS Distributions 	operation();
1165*a325d9c4SApple OSS Distributions 	wq->wq_turnstile_updater = THREAD_NULL;
1166*a325d9c4SApple OSS Distributions }
1167*a325d9c4SApple OSS Distributions 
1168*a325d9c4SApple OSS Distributions static void
workq_turnstile_update_inheritor(struct workqueue * wq,turnstile_inheritor_t inheritor,turnstile_update_flags_t flags)1169*a325d9c4SApple OSS Distributions workq_turnstile_update_inheritor(struct workqueue *wq,
1170*a325d9c4SApple OSS Distributions     turnstile_inheritor_t inheritor,
1171*a325d9c4SApple OSS Distributions     turnstile_update_flags_t flags)
1172*a325d9c4SApple OSS Distributions {
1173*a325d9c4SApple OSS Distributions 	if (wq->wq_inheritor == inheritor) {
1174*a325d9c4SApple OSS Distributions 		return;
1175*a325d9c4SApple OSS Distributions 	}
1176*a325d9c4SApple OSS Distributions 	wq->wq_inheritor = inheritor;
1177*a325d9c4SApple OSS Distributions 	workq_perform_turnstile_operation_locked(wq, ^{
1178*a325d9c4SApple OSS Distributions 		turnstile_update_inheritor(wq->wq_turnstile, inheritor,
1179*a325d9c4SApple OSS Distributions 		flags | TURNSTILE_IMMEDIATE_UPDATE);
1180*a325d9c4SApple OSS Distributions 		turnstile_update_inheritor_complete(wq->wq_turnstile,
1181*a325d9c4SApple OSS Distributions 		TURNSTILE_INTERLOCK_HELD);
1182*a325d9c4SApple OSS Distributions 	});
1183*a325d9c4SApple OSS Distributions }
1184*a325d9c4SApple OSS Distributions 
1185*a325d9c4SApple OSS Distributions static void
workq_push_idle_thread(proc_t p,struct workqueue * wq,struct uthread * uth,uint32_t setup_flags)1186*a325d9c4SApple OSS Distributions workq_push_idle_thread(proc_t p, struct workqueue *wq, struct uthread *uth,
1187*a325d9c4SApple OSS Distributions     uint32_t setup_flags)
1188*a325d9c4SApple OSS Distributions {
1189*a325d9c4SApple OSS Distributions 	uint64_t now = mach_absolute_time();
1190*a325d9c4SApple OSS Distributions 	bool is_creator = (uth == wq->wq_creator);
1191*a325d9c4SApple OSS Distributions 
1192*a325d9c4SApple OSS Distributions 	if (workq_thread_is_cooperative(uth)) {
1193*a325d9c4SApple OSS Distributions 		assert(!is_creator);
1194*a325d9c4SApple OSS Distributions 
1195*a325d9c4SApple OSS Distributions 		thread_qos_t thread_qos = uth->uu_workq_pri.qos_bucket;
1196*a325d9c4SApple OSS Distributions 		_wq_cooperative_queue_scheduled_count_dec(wq, thread_qos);
1197*a325d9c4SApple OSS Distributions 
1198*a325d9c4SApple OSS Distributions 		/* Before we get here, we always go through
1199*a325d9c4SApple OSS Distributions 		 * workq_select_threadreq_or_park_and_unlock. If we got here, it means
1200*a325d9c4SApple OSS Distributions 		 * that we went through the logic in workq_threadreq_select which
1201*a325d9c4SApple OSS Distributions 		 * did the refresh for the next best cooperative qos while
1202*a325d9c4SApple OSS Distributions 		 * excluding the current thread - we shouldn't need to do it again.
1203*a325d9c4SApple OSS Distributions 		 */
1204*a325d9c4SApple OSS Distributions 		assert(_wq_cooperative_queue_refresh_best_req_qos(wq) == false);
1205*a325d9c4SApple OSS Distributions 	} else if (workq_thread_is_nonovercommit(uth)) {
1206*a325d9c4SApple OSS Distributions 		assert(!is_creator);
1207*a325d9c4SApple OSS Distributions 
1208*a325d9c4SApple OSS Distributions 		wq->wq_constrained_threads_scheduled--;
1209*a325d9c4SApple OSS Distributions 	}
1210*a325d9c4SApple OSS Distributions 
1211*a325d9c4SApple OSS Distributions 	uth->uu_workq_flags &= ~(UT_WORKQ_RUNNING | UT_WORKQ_OVERCOMMIT | UT_WORKQ_COOPERATIVE);
1212*a325d9c4SApple OSS Distributions 	TAILQ_REMOVE(&wq->wq_thrunlist, uth, uu_workq_entry);
1213*a325d9c4SApple OSS Distributions 	wq->wq_threads_scheduled--;
1214*a325d9c4SApple OSS Distributions 
1215*a325d9c4SApple OSS Distributions 	if (is_creator) {
1216*a325d9c4SApple OSS Distributions 		wq->wq_creator = NULL;
1217*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_creator_select, wq, 3, 0,
1218*a325d9c4SApple OSS Distributions 		    uth->uu_save.uus_workq_park_data.yields);
1219*a325d9c4SApple OSS Distributions 	}
1220*a325d9c4SApple OSS Distributions 
1221*a325d9c4SApple OSS Distributions 	if (wq->wq_inheritor == get_machthread(uth)) {
1222*a325d9c4SApple OSS Distributions 		assert(wq->wq_creator == NULL);
1223*a325d9c4SApple OSS Distributions 		if (wq->wq_reqcount) {
1224*a325d9c4SApple OSS Distributions 			workq_turnstile_update_inheritor(wq, wq, TURNSTILE_INHERITOR_WORKQ);
1225*a325d9c4SApple OSS Distributions 		} else {
1226*a325d9c4SApple OSS Distributions 			workq_turnstile_update_inheritor(wq, TURNSTILE_INHERITOR_NULL, 0);
1227*a325d9c4SApple OSS Distributions 		}
1228*a325d9c4SApple OSS Distributions 	}
1229*a325d9c4SApple OSS Distributions 
1230*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_NEW) {
1231*a325d9c4SApple OSS Distributions 		assert(is_creator || (_wq_flags(wq) & WQ_EXITING));
1232*a325d9c4SApple OSS Distributions 		TAILQ_INSERT_TAIL(&wq->wq_thnewlist, uth, uu_workq_entry);
1233*a325d9c4SApple OSS Distributions 		wq->wq_thidlecount++;
1234*a325d9c4SApple OSS Distributions 		return;
1235*a325d9c4SApple OSS Distributions 	}
1236*a325d9c4SApple OSS Distributions 
1237*a325d9c4SApple OSS Distributions 	if (!is_creator) {
1238*a325d9c4SApple OSS Distributions 		_wq_thactive_dec(wq, uth->uu_workq_pri.qos_bucket);
1239*a325d9c4SApple OSS Distributions 		wq->wq_thscheduled_count[_wq_bucket(uth->uu_workq_pri.qos_bucket)]--;
1240*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags |= UT_WORKQ_IDLE_CLEANUP;
1241*a325d9c4SApple OSS Distributions 	}
1242*a325d9c4SApple OSS Distributions 
1243*a325d9c4SApple OSS Distributions 	uth->uu_save.uus_workq_park_data.idle_stamp = now;
1244*a325d9c4SApple OSS Distributions 
1245*a325d9c4SApple OSS Distributions 	struct uthread *oldest = workq_oldest_killable_idle_thread(wq);
1246*a325d9c4SApple OSS Distributions 	uint16_t cur_idle = wq->wq_thidlecount;
1247*a325d9c4SApple OSS Distributions 
1248*a325d9c4SApple OSS Distributions 	if (cur_idle >= wq_max_constrained_threads ||
1249*a325d9c4SApple OSS Distributions 	    (wq->wq_thdying_count == 0 && oldest &&
1250*a325d9c4SApple OSS Distributions 	    workq_should_kill_idle_thread(wq, oldest, now))) {
1251*a325d9c4SApple OSS Distributions 		/*
1252*a325d9c4SApple OSS Distributions 		 * Immediately kill threads if we have too may of them.
1253*a325d9c4SApple OSS Distributions 		 *
1254*a325d9c4SApple OSS Distributions 		 * And swap "place" with the oldest one we'd have woken up.
1255*a325d9c4SApple OSS Distributions 		 * This is a relatively desperate situation where we really
1256*a325d9c4SApple OSS Distributions 		 * need to kill threads quickly and it's best to kill
1257*a325d9c4SApple OSS Distributions 		 * the one that's currently on core than context switching.
1258*a325d9c4SApple OSS Distributions 		 */
1259*a325d9c4SApple OSS Distributions 		if (oldest) {
1260*a325d9c4SApple OSS Distributions 			oldest->uu_save.uus_workq_park_data.idle_stamp = now;
1261*a325d9c4SApple OSS Distributions 			TAILQ_REMOVE(&wq->wq_thidlelist, oldest, uu_workq_entry);
1262*a325d9c4SApple OSS Distributions 			TAILQ_INSERT_HEAD(&wq->wq_thidlelist, oldest, uu_workq_entry);
1263*a325d9c4SApple OSS Distributions 		}
1264*a325d9c4SApple OSS Distributions 
1265*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thread_terminate | DBG_FUNC_START,
1266*a325d9c4SApple OSS Distributions 		    wq, cur_idle, 0, 0);
1267*a325d9c4SApple OSS Distributions 		wq->wq_thdying_count++;
1268*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags |= UT_WORKQ_DYING;
1269*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags &= ~UT_WORKQ_IDLE_CLEANUP;
1270*a325d9c4SApple OSS Distributions 		workq_unpark_for_death_and_unlock(p, wq, uth, 0, setup_flags);
1271*a325d9c4SApple OSS Distributions 		__builtin_unreachable();
1272*a325d9c4SApple OSS Distributions 	}
1273*a325d9c4SApple OSS Distributions 
1274*a325d9c4SApple OSS Distributions 	struct uthread *tail = TAILQ_LAST(&wq->wq_thidlelist, workq_uthread_head);
1275*a325d9c4SApple OSS Distributions 
1276*a325d9c4SApple OSS Distributions 	cur_idle += 1;
1277*a325d9c4SApple OSS Distributions 	wq->wq_thidlecount = cur_idle;
1278*a325d9c4SApple OSS Distributions 
1279*a325d9c4SApple OSS Distributions 	if (cur_idle >= wq_death_max_load && tail &&
1280*a325d9c4SApple OSS Distributions 	    tail->uu_save.uus_workq_park_data.has_stack) {
1281*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.has_stack = false;
1282*a325d9c4SApple OSS Distributions 		TAILQ_INSERT_TAIL(&wq->wq_thidlelist, uth, uu_workq_entry);
1283*a325d9c4SApple OSS Distributions 	} else {
1284*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.has_stack = true;
1285*a325d9c4SApple OSS Distributions 		TAILQ_INSERT_HEAD(&wq->wq_thidlelist, uth, uu_workq_entry);
1286*a325d9c4SApple OSS Distributions 	}
1287*a325d9c4SApple OSS Distributions 
1288*a325d9c4SApple OSS Distributions 	if (!tail) {
1289*a325d9c4SApple OSS Distributions 		uint64_t delay = workq_kill_delay_for_idle_thread(wq);
1290*a325d9c4SApple OSS Distributions 		workq_death_call_schedule(wq, now + delay);
1291*a325d9c4SApple OSS Distributions 	}
1292*a325d9c4SApple OSS Distributions }
1293*a325d9c4SApple OSS Distributions 
1294*a325d9c4SApple OSS Distributions #pragma mark thread requests
1295*a325d9c4SApple OSS Distributions 
1296*a325d9c4SApple OSS Distributions static inline bool
workq_tr_is_overcommit(workq_tr_flags_t tr_flags)1297*a325d9c4SApple OSS Distributions workq_tr_is_overcommit(workq_tr_flags_t tr_flags)
1298*a325d9c4SApple OSS Distributions {
1299*a325d9c4SApple OSS Distributions 	return (tr_flags & WORKQ_TR_FLAG_OVERCOMMIT) != 0;
1300*a325d9c4SApple OSS Distributions }
1301*a325d9c4SApple OSS Distributions 
1302*a325d9c4SApple OSS Distributions static inline bool
workq_tr_is_nonovercommit(workq_tr_flags_t tr_flags)1303*a325d9c4SApple OSS Distributions workq_tr_is_nonovercommit(workq_tr_flags_t tr_flags)
1304*a325d9c4SApple OSS Distributions {
1305*a325d9c4SApple OSS Distributions 	return (tr_flags & (WORKQ_TR_FLAG_OVERCOMMIT | WORKQ_TR_FLAG_COOPERATIVE)) == 0;
1306*a325d9c4SApple OSS Distributions }
1307*a325d9c4SApple OSS Distributions 
1308*a325d9c4SApple OSS Distributions static inline bool
workq_tr_is_cooperative(workq_tr_flags_t tr_flags)1309*a325d9c4SApple OSS Distributions workq_tr_is_cooperative(workq_tr_flags_t tr_flags)
1310*a325d9c4SApple OSS Distributions {
1311*a325d9c4SApple OSS Distributions 	return (tr_flags & WORKQ_TR_FLAG_COOPERATIVE) != 0;
1312*a325d9c4SApple OSS Distributions }
1313*a325d9c4SApple OSS Distributions 
1314*a325d9c4SApple OSS Distributions #define workq_threadreq_is_overcommit(req) workq_tr_is_overcommit((req)->tr_flags)
1315*a325d9c4SApple OSS Distributions #define workq_threadreq_is_nonovercommit(req) workq_tr_is_nonovercommit((req)->tr_flags)
1316*a325d9c4SApple OSS Distributions #define workq_threadreq_is_cooperative(req) workq_tr_is_cooperative((req)->tr_flags)
1317*a325d9c4SApple OSS Distributions 
1318*a325d9c4SApple OSS Distributions static inline int
workq_priority_for_req(workq_threadreq_t req)1319*a325d9c4SApple OSS Distributions workq_priority_for_req(workq_threadreq_t req)
1320*a325d9c4SApple OSS Distributions {
1321*a325d9c4SApple OSS Distributions 	thread_qos_t qos = req->tr_qos;
1322*a325d9c4SApple OSS Distributions 
1323*a325d9c4SApple OSS Distributions 	if (req->tr_flags & WORKQ_TR_FLAG_WL_OUTSIDE_QOS) {
1324*a325d9c4SApple OSS Distributions 		workq_threadreq_param_t trp = kqueue_threadreq_workloop_param(req);
1325*a325d9c4SApple OSS Distributions 		assert(trp.trp_flags & TRP_PRIORITY);
1326*a325d9c4SApple OSS Distributions 		return trp.trp_pri;
1327*a325d9c4SApple OSS Distributions 	}
1328*a325d9c4SApple OSS Distributions 	return thread_workq_pri_for_qos(qos);
1329*a325d9c4SApple OSS Distributions }
1330*a325d9c4SApple OSS Distributions 
1331*a325d9c4SApple OSS Distributions static inline struct priority_queue_sched_max *
workq_priority_queue_for_req(struct workqueue * wq,workq_threadreq_t req)1332*a325d9c4SApple OSS Distributions workq_priority_queue_for_req(struct workqueue *wq, workq_threadreq_t req)
1333*a325d9c4SApple OSS Distributions {
1334*a325d9c4SApple OSS Distributions 	assert(!workq_tr_is_cooperative(req->tr_flags));
1335*a325d9c4SApple OSS Distributions 
1336*a325d9c4SApple OSS Distributions 	if (req->tr_flags & WORKQ_TR_FLAG_WL_OUTSIDE_QOS) {
1337*a325d9c4SApple OSS Distributions 		return &wq->wq_special_queue;
1338*a325d9c4SApple OSS Distributions 	} else if (workq_tr_is_overcommit(req->tr_flags)) {
1339*a325d9c4SApple OSS Distributions 		return &wq->wq_overcommit_queue;
1340*a325d9c4SApple OSS Distributions 	} else {
1341*a325d9c4SApple OSS Distributions 		return &wq->wq_constrained_queue;
1342*a325d9c4SApple OSS Distributions 	}
1343*a325d9c4SApple OSS Distributions }
1344*a325d9c4SApple OSS Distributions 
1345*a325d9c4SApple OSS Distributions 
1346*a325d9c4SApple OSS Distributions /* Calculates the number of threads scheduled >= the input QoS */
1347*a325d9c4SApple OSS Distributions static uint64_t
workq_num_cooperative_threads_scheduled_to_qos(struct workqueue * wq,thread_qos_t qos)1348*a325d9c4SApple OSS Distributions workq_num_cooperative_threads_scheduled_to_qos(struct workqueue *wq, thread_qos_t qos)
1349*a325d9c4SApple OSS Distributions {
1350*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
1351*a325d9c4SApple OSS Distributions 
1352*a325d9c4SApple OSS Distributions 	uint64_t num_cooperative_threads = 0;
1353*a325d9c4SApple OSS Distributions 
1354*a325d9c4SApple OSS Distributions 	for (thread_qos_t cur_qos = WORKQ_THREAD_QOS_MAX; cur_qos >= qos; cur_qos--) {
1355*a325d9c4SApple OSS Distributions 		int bucket = _wq_bucket(cur_qos);
1356*a325d9c4SApple OSS Distributions 		num_cooperative_threads += wq->wq_cooperative_queue_scheduled_count[bucket];
1357*a325d9c4SApple OSS Distributions 	}
1358*a325d9c4SApple OSS Distributions 
1359*a325d9c4SApple OSS Distributions 	return num_cooperative_threads;
1360*a325d9c4SApple OSS Distributions }
1361*a325d9c4SApple OSS Distributions 
1362*a325d9c4SApple OSS Distributions static uint64_t
workq_num_cooperative_threads_scheduled_total(struct workqueue * wq)1363*a325d9c4SApple OSS Distributions workq_num_cooperative_threads_scheduled_total(struct workqueue *wq)
1364*a325d9c4SApple OSS Distributions {
1365*a325d9c4SApple OSS Distributions 	return workq_num_cooperative_threads_scheduled_to_qos(wq, WORKQ_THREAD_QOS_MIN);
1366*a325d9c4SApple OSS Distributions }
1367*a325d9c4SApple OSS Distributions 
1368*a325d9c4SApple OSS Distributions #if DEBUG || DEVELOPMENT
1369*a325d9c4SApple OSS Distributions static bool
workq_has_cooperative_thread_requests(struct workqueue * wq)1370*a325d9c4SApple OSS Distributions workq_has_cooperative_thread_requests(struct workqueue *wq)
1371*a325d9c4SApple OSS Distributions {
1372*a325d9c4SApple OSS Distributions 	for (thread_qos_t qos = WORKQ_THREAD_QOS_MAX; qos >= WORKQ_THREAD_QOS_MIN; qos--) {
1373*a325d9c4SApple OSS Distributions 		int bucket = _wq_bucket(qos);
1374*a325d9c4SApple OSS Distributions 		if (!STAILQ_EMPTY(&wq->wq_cooperative_queue[bucket])) {
1375*a325d9c4SApple OSS Distributions 			return true;
1376*a325d9c4SApple OSS Distributions 		}
1377*a325d9c4SApple OSS Distributions 	}
1378*a325d9c4SApple OSS Distributions 
1379*a325d9c4SApple OSS Distributions 	return false;
1380*a325d9c4SApple OSS Distributions }
1381*a325d9c4SApple OSS Distributions #endif
1382*a325d9c4SApple OSS Distributions 
1383*a325d9c4SApple OSS Distributions /*
1384*a325d9c4SApple OSS Distributions  * Determines the next QoS bucket we should service next in the cooperative
1385*a325d9c4SApple OSS Distributions  * pool. This function will always return a QoS for cooperative pool as long as
1386*a325d9c4SApple OSS Distributions  * there are requests to be serviced.
1387*a325d9c4SApple OSS Distributions  *
1388*a325d9c4SApple OSS Distributions  * Unlike the other thread pools, for the cooperative thread pool the schedule
1389*a325d9c4SApple OSS Distributions  * counts for the various buckets in the pool affect the next best request for
1390*a325d9c4SApple OSS Distributions  * it.
1391*a325d9c4SApple OSS Distributions  *
1392*a325d9c4SApple OSS Distributions  * This function is called in the following contexts:
1393*a325d9c4SApple OSS Distributions  *
1394*a325d9c4SApple OSS Distributions  * a) When determining the best thread QoS for cooperative bucket for the
1395*a325d9c4SApple OSS Distributions  * creator/thread reuse
1396*a325d9c4SApple OSS Distributions  *
1397*a325d9c4SApple OSS Distributions  * b) Once (a) has happened and thread has bound to a thread request, figuring
1398*a325d9c4SApple OSS Distributions  * out whether the next best request for this pool has changed so that creator
1399*a325d9c4SApple OSS Distributions  * can be scheduled.
1400*a325d9c4SApple OSS Distributions  *
1401*a325d9c4SApple OSS Distributions  * Returns true if the cooperative queue's best qos changed from previous
1402*a325d9c4SApple OSS Distributions  * value.
1403*a325d9c4SApple OSS Distributions  */
1404*a325d9c4SApple OSS Distributions static bool
_wq_cooperative_queue_refresh_best_req_qos(struct workqueue * wq)1405*a325d9c4SApple OSS Distributions _wq_cooperative_queue_refresh_best_req_qos(struct workqueue *wq)
1406*a325d9c4SApple OSS Distributions {
1407*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
1408*a325d9c4SApple OSS Distributions 
1409*a325d9c4SApple OSS Distributions 	thread_qos_t old_best_req_qos = wq->wq_cooperative_queue_best_req_qos;
1410*a325d9c4SApple OSS Distributions 
1411*a325d9c4SApple OSS Distributions 	/* We determine the next best cooperative thread request based on the
1412*a325d9c4SApple OSS Distributions 	 * following:
1413*a325d9c4SApple OSS Distributions 	 *
1414*a325d9c4SApple OSS Distributions 	 * 1. Take the MAX of the following:
1415*a325d9c4SApple OSS Distributions 	 *		a) Highest qos with pending TRs such that number of scheduled
1416*a325d9c4SApple OSS Distributions 	 *		threads so far with >= qos is < wq_max_cooperative_threads
1417*a325d9c4SApple OSS Distributions 	 *		b) Highest qos bucket with pending TRs but no scheduled threads for that bucket
1418*a325d9c4SApple OSS Distributions 	 *
1419*a325d9c4SApple OSS Distributions 	 * 2. If the result of (1) is UN, then we pick the highest priority amongst
1420*a325d9c4SApple OSS Distributions 	 * pending thread requests in the pool.
1421*a325d9c4SApple OSS Distributions 	 *
1422*a325d9c4SApple OSS Distributions 	 */
1423*a325d9c4SApple OSS Distributions 	thread_qos_t highest_qos_with_no_scheduled = THREAD_QOS_UNSPECIFIED;
1424*a325d9c4SApple OSS Distributions 	thread_qos_t highest_qos_req_with_width = THREAD_QOS_UNSPECIFIED;
1425*a325d9c4SApple OSS Distributions 
1426*a325d9c4SApple OSS Distributions 	thread_qos_t highest_qos_req = THREAD_QOS_UNSPECIFIED;
1427*a325d9c4SApple OSS Distributions 
1428*a325d9c4SApple OSS Distributions 	int scheduled_count_till_qos = 0;
1429*a325d9c4SApple OSS Distributions 
1430*a325d9c4SApple OSS Distributions 	for (thread_qos_t qos = WORKQ_THREAD_QOS_MAX; qos >= WORKQ_THREAD_QOS_MIN; qos--) {
1431*a325d9c4SApple OSS Distributions 		int bucket = _wq_bucket(qos);
1432*a325d9c4SApple OSS Distributions 		uint8_t scheduled_count_for_bucket = wq->wq_cooperative_queue_scheduled_count[bucket];
1433*a325d9c4SApple OSS Distributions 		scheduled_count_till_qos += scheduled_count_for_bucket;
1434*a325d9c4SApple OSS Distributions 
1435*a325d9c4SApple OSS Distributions 		if (!STAILQ_EMPTY(&wq->wq_cooperative_queue[bucket])) {
1436*a325d9c4SApple OSS Distributions 			if (qos > highest_qos_req) {
1437*a325d9c4SApple OSS Distributions 				highest_qos_req = qos;
1438*a325d9c4SApple OSS Distributions 			}
1439*a325d9c4SApple OSS Distributions 			/*
1440*a325d9c4SApple OSS Distributions 			 * The pool isn't saturated for threads at and above this QoS, and
1441*a325d9c4SApple OSS Distributions 			 * this qos bucket has pending requests
1442*a325d9c4SApple OSS Distributions 			 */
1443*a325d9c4SApple OSS Distributions 			if (scheduled_count_till_qos < wq_cooperative_queue_max_size(wq)) {
1444*a325d9c4SApple OSS Distributions 				if (qos > highest_qos_req_with_width) {
1445*a325d9c4SApple OSS Distributions 					highest_qos_req_with_width = qos;
1446*a325d9c4SApple OSS Distributions 				}
1447*a325d9c4SApple OSS Distributions 			}
1448*a325d9c4SApple OSS Distributions 
1449*a325d9c4SApple OSS Distributions 			/*
1450*a325d9c4SApple OSS Distributions 			 * There are no threads scheduled for this bucket but there
1451*a325d9c4SApple OSS Distributions 			 * is work pending, give it at least 1 thread
1452*a325d9c4SApple OSS Distributions 			 */
1453*a325d9c4SApple OSS Distributions 			if (scheduled_count_for_bucket == 0) {
1454*a325d9c4SApple OSS Distributions 				if (qos > highest_qos_with_no_scheduled) {
1455*a325d9c4SApple OSS Distributions 					highest_qos_with_no_scheduled = qos;
1456*a325d9c4SApple OSS Distributions 				}
1457*a325d9c4SApple OSS Distributions 			}
1458*a325d9c4SApple OSS Distributions 		}
1459*a325d9c4SApple OSS Distributions 	}
1460*a325d9c4SApple OSS Distributions 
1461*a325d9c4SApple OSS Distributions 	wq->wq_cooperative_queue_best_req_qos = MAX(highest_qos_with_no_scheduled, highest_qos_req_with_width);
1462*a325d9c4SApple OSS Distributions 	if (wq->wq_cooperative_queue_best_req_qos == THREAD_QOS_UNSPECIFIED) {
1463*a325d9c4SApple OSS Distributions 		wq->wq_cooperative_queue_best_req_qos = highest_qos_req;
1464*a325d9c4SApple OSS Distributions 	}
1465*a325d9c4SApple OSS Distributions 
1466*a325d9c4SApple OSS Distributions #if DEBUG || DEVELOPMENT
1467*a325d9c4SApple OSS Distributions 	/* Assert that if we are showing up the next best req as UN, then there
1468*a325d9c4SApple OSS Distributions 	 * actually is no thread request in the cooperative pool buckets */
1469*a325d9c4SApple OSS Distributions 	if (wq->wq_cooperative_queue_best_req_qos == THREAD_QOS_UNSPECIFIED) {
1470*a325d9c4SApple OSS Distributions 		assert(!workq_has_cooperative_thread_requests(wq));
1471*a325d9c4SApple OSS Distributions 	}
1472*a325d9c4SApple OSS Distributions #endif
1473*a325d9c4SApple OSS Distributions 
1474*a325d9c4SApple OSS Distributions 	return old_best_req_qos != wq->wq_cooperative_queue_best_req_qos;
1475*a325d9c4SApple OSS Distributions }
1476*a325d9c4SApple OSS Distributions 
1477*a325d9c4SApple OSS Distributions /*
1478*a325d9c4SApple OSS Distributions  * Returns whether or not the input thread (or creator thread if uth is NULL)
1479*a325d9c4SApple OSS Distributions  * should be allowed to work as part of the cooperative pool for the <input qos>
1480*a325d9c4SApple OSS Distributions  * bucket.
1481*a325d9c4SApple OSS Distributions  *
1482*a325d9c4SApple OSS Distributions  * This function is called in a bunch of places:
1483*a325d9c4SApple OSS Distributions  *		a) Quantum expires for a thread and it is part of the cooperative pool
1484*a325d9c4SApple OSS Distributions  *		b) When trying to pick a thread request for the creator thread to
1485*a325d9c4SApple OSS Distributions  *		represent.
1486*a325d9c4SApple OSS Distributions  *		c) When a thread is trying to pick a thread request to actually bind to
1487*a325d9c4SApple OSS Distributions  *		and service.
1488*a325d9c4SApple OSS Distributions  *
1489*a325d9c4SApple OSS Distributions  * Called with workq lock held.
1490*a325d9c4SApple OSS Distributions  */
1491*a325d9c4SApple OSS Distributions 
1492*a325d9c4SApple OSS Distributions #define WQ_COOPERATIVE_POOL_UNSATURATED 1
1493*a325d9c4SApple OSS Distributions #define WQ_COOPERATIVE_BUCKET_UNSERVICED 2
1494*a325d9c4SApple OSS Distributions #define WQ_COOPERATIVE_POOL_SATURATED_UP_TO_QOS 3
1495*a325d9c4SApple OSS Distributions 
1496*a325d9c4SApple OSS Distributions static bool
workq_cooperative_allowance(struct workqueue * wq,thread_qos_t qos,struct uthread * uth,bool may_start_timer)1497*a325d9c4SApple OSS Distributions workq_cooperative_allowance(struct workqueue *wq, thread_qos_t qos, struct uthread *uth,
1498*a325d9c4SApple OSS Distributions     bool may_start_timer)
1499*a325d9c4SApple OSS Distributions {
1500*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
1501*a325d9c4SApple OSS Distributions 
1502*a325d9c4SApple OSS Distributions 	bool exclude_thread_as_scheduled = false;
1503*a325d9c4SApple OSS Distributions 	bool passed_admissions = false;
1504*a325d9c4SApple OSS Distributions 	int bucket = _wq_bucket(qos);
1505*a325d9c4SApple OSS Distributions 
1506*a325d9c4SApple OSS Distributions 	if (uth && workq_thread_is_cooperative(uth)) {
1507*a325d9c4SApple OSS Distributions 		exclude_thread_as_scheduled = true;
1508*a325d9c4SApple OSS Distributions 		_wq_cooperative_queue_scheduled_count_dec(wq, uth->uu_workq_pri.qos_bucket);
1509*a325d9c4SApple OSS Distributions 	}
1510*a325d9c4SApple OSS Distributions 
1511*a325d9c4SApple OSS Distributions 	/*
1512*a325d9c4SApple OSS Distributions 	 * We have not saturated the pool yet, let this thread continue
1513*a325d9c4SApple OSS Distributions 	 */
1514*a325d9c4SApple OSS Distributions 	uint64_t total_cooperative_threads;
1515*a325d9c4SApple OSS Distributions 	total_cooperative_threads = workq_num_cooperative_threads_scheduled_total(wq);
1516*a325d9c4SApple OSS Distributions 	if (total_cooperative_threads < wq_cooperative_queue_max_size(wq)) {
1517*a325d9c4SApple OSS Distributions 		passed_admissions = true;
1518*a325d9c4SApple OSS Distributions 		WQ_TRACE(TRACE_wq_cooperative_admission | DBG_FUNC_NONE,
1519*a325d9c4SApple OSS Distributions 		    total_cooperative_threads, qos, passed_admissions,
1520*a325d9c4SApple OSS Distributions 		    WQ_COOPERATIVE_POOL_UNSATURATED);
1521*a325d9c4SApple OSS Distributions 		goto out;
1522*a325d9c4SApple OSS Distributions 	}
1523*a325d9c4SApple OSS Distributions 
1524*a325d9c4SApple OSS Distributions 	/*
1525*a325d9c4SApple OSS Distributions 	 * Without this thread, nothing is servicing the bucket which has pending
1526*a325d9c4SApple OSS Distributions 	 * work
1527*a325d9c4SApple OSS Distributions 	 */
1528*a325d9c4SApple OSS Distributions 	uint64_t bucket_scheduled = wq->wq_cooperative_queue_scheduled_count[bucket];
1529*a325d9c4SApple OSS Distributions 	if (bucket_scheduled == 0 &&
1530*a325d9c4SApple OSS Distributions 	    !STAILQ_EMPTY(&wq->wq_cooperative_queue[bucket])) {
1531*a325d9c4SApple OSS Distributions 		passed_admissions = true;
1532*a325d9c4SApple OSS Distributions 		WQ_TRACE(TRACE_wq_cooperative_admission | DBG_FUNC_NONE,
1533*a325d9c4SApple OSS Distributions 		    total_cooperative_threads, qos, passed_admissions,
1534*a325d9c4SApple OSS Distributions 		    WQ_COOPERATIVE_BUCKET_UNSERVICED);
1535*a325d9c4SApple OSS Distributions 		goto out;
1536*a325d9c4SApple OSS Distributions 	}
1537*a325d9c4SApple OSS Distributions 
1538*a325d9c4SApple OSS Distributions 	/*
1539*a325d9c4SApple OSS Distributions 	 * If number of threads at the QoS bucket >= input QoS exceeds the max we want
1540*a325d9c4SApple OSS Distributions 	 * for the pool, deny this thread
1541*a325d9c4SApple OSS Distributions 	 */
1542*a325d9c4SApple OSS Distributions 	uint64_t aggregate_down_to_qos = workq_num_cooperative_threads_scheduled_to_qos(wq, qos);
1543*a325d9c4SApple OSS Distributions 	passed_admissions = (aggregate_down_to_qos < wq_cooperative_queue_max_size(wq));
1544*a325d9c4SApple OSS Distributions 	WQ_TRACE(TRACE_wq_cooperative_admission | DBG_FUNC_NONE, aggregate_down_to_qos,
1545*a325d9c4SApple OSS Distributions 	    qos, passed_admissions, WQ_COOPERATIVE_POOL_SATURATED_UP_TO_QOS);
1546*a325d9c4SApple OSS Distributions 
1547*a325d9c4SApple OSS Distributions 	if (!passed_admissions && may_start_timer) {
1548*a325d9c4SApple OSS Distributions 		workq_schedule_delayed_thread_creation(wq, 0);
1549*a325d9c4SApple OSS Distributions 	}
1550*a325d9c4SApple OSS Distributions 
1551*a325d9c4SApple OSS Distributions out:
1552*a325d9c4SApple OSS Distributions 	if (exclude_thread_as_scheduled) {
1553*a325d9c4SApple OSS Distributions 		_wq_cooperative_queue_scheduled_count_inc(wq, uth->uu_workq_pri.qos_bucket);
1554*a325d9c4SApple OSS Distributions 	}
1555*a325d9c4SApple OSS Distributions 	return passed_admissions;
1556*a325d9c4SApple OSS Distributions }
1557*a325d9c4SApple OSS Distributions 
1558*a325d9c4SApple OSS Distributions /*
1559*a325d9c4SApple OSS Distributions  * returns true if the best request for the pool changed as a result of
1560*a325d9c4SApple OSS Distributions  * enqueuing this thread request.
1561*a325d9c4SApple OSS Distributions  */
1562*a325d9c4SApple OSS Distributions static bool
workq_threadreq_enqueue(struct workqueue * wq,workq_threadreq_t req)1563*a325d9c4SApple OSS Distributions workq_threadreq_enqueue(struct workqueue *wq, workq_threadreq_t req)
1564*a325d9c4SApple OSS Distributions {
1565*a325d9c4SApple OSS Distributions 	assert(req->tr_state == WORKQ_TR_STATE_NEW);
1566*a325d9c4SApple OSS Distributions 
1567*a325d9c4SApple OSS Distributions 	req->tr_state = WORKQ_TR_STATE_QUEUED;
1568*a325d9c4SApple OSS Distributions 	wq->wq_reqcount += req->tr_count;
1569*a325d9c4SApple OSS Distributions 
1570*a325d9c4SApple OSS Distributions 	if (req->tr_qos == WORKQ_THREAD_QOS_MANAGER) {
1571*a325d9c4SApple OSS Distributions 		assert(wq->wq_event_manager_threadreq == NULL);
1572*a325d9c4SApple OSS Distributions 		assert(req->tr_flags & WORKQ_TR_FLAG_KEVENT);
1573*a325d9c4SApple OSS Distributions 		assert(req->tr_count == 1);
1574*a325d9c4SApple OSS Distributions 		wq->wq_event_manager_threadreq = req;
1575*a325d9c4SApple OSS Distributions 		return true;
1576*a325d9c4SApple OSS Distributions 	}
1577*a325d9c4SApple OSS Distributions 
1578*a325d9c4SApple OSS Distributions 	if (workq_threadreq_is_cooperative(req)) {
1579*a325d9c4SApple OSS Distributions 		assert(req->tr_qos != WORKQ_THREAD_QOS_MANAGER);
1580*a325d9c4SApple OSS Distributions 		assert(req->tr_qos != WORKQ_THREAD_QOS_ABOVEUI);
1581*a325d9c4SApple OSS Distributions 
1582*a325d9c4SApple OSS Distributions 		struct workq_threadreq_tailq *bucket = &wq->wq_cooperative_queue[_wq_bucket(req->tr_qos)];
1583*a325d9c4SApple OSS Distributions 		STAILQ_INSERT_TAIL(bucket, req, tr_link);
1584*a325d9c4SApple OSS Distributions 
1585*a325d9c4SApple OSS Distributions 		return _wq_cooperative_queue_refresh_best_req_qos(wq);
1586*a325d9c4SApple OSS Distributions 	}
1587*a325d9c4SApple OSS Distributions 
1588*a325d9c4SApple OSS Distributions 	struct priority_queue_sched_max *q = workq_priority_queue_for_req(wq, req);
1589*a325d9c4SApple OSS Distributions 
1590*a325d9c4SApple OSS Distributions 	priority_queue_entry_set_sched_pri(q, &req->tr_entry,
1591*a325d9c4SApple OSS Distributions 	    workq_priority_for_req(req), false);
1592*a325d9c4SApple OSS Distributions 
1593*a325d9c4SApple OSS Distributions 	if (priority_queue_insert(q, &req->tr_entry)) {
1594*a325d9c4SApple OSS Distributions 		if (workq_threadreq_is_nonovercommit(req)) {
1595*a325d9c4SApple OSS Distributions 			_wq_thactive_refresh_best_constrained_req_qos(wq);
1596*a325d9c4SApple OSS Distributions 		}
1597*a325d9c4SApple OSS Distributions 		return true;
1598*a325d9c4SApple OSS Distributions 	}
1599*a325d9c4SApple OSS Distributions 	return false;
1600*a325d9c4SApple OSS Distributions }
1601*a325d9c4SApple OSS Distributions 
1602*a325d9c4SApple OSS Distributions /*
1603*a325d9c4SApple OSS Distributions  * returns true if one of the following is true (so as to update creator if
1604*a325d9c4SApple OSS Distributions  * needed):
1605*a325d9c4SApple OSS Distributions  *
1606*a325d9c4SApple OSS Distributions  * (a) the next highest request of the pool we dequeued the request from changed
1607*a325d9c4SApple OSS Distributions  * (b) the next highest requests of the pool the current thread used to be a
1608*a325d9c4SApple OSS Distributions  * part of, changed
1609*a325d9c4SApple OSS Distributions  *
1610*a325d9c4SApple OSS Distributions  * For overcommit, special and constrained pools, the next highest QoS for each
1611*a325d9c4SApple OSS Distributions  * pool just a MAX of pending requests so tracking (a) is sufficient.
1612*a325d9c4SApple OSS Distributions  *
1613*a325d9c4SApple OSS Distributions  * But for cooperative thread pool, the next highest QoS for the pool depends on
1614*a325d9c4SApple OSS Distributions  * schedule counts in the pool as well. So if the current thread used to be
1615*a325d9c4SApple OSS Distributions  * cooperative in it's previous logical run ie (b), then that can also affect
1616*a325d9c4SApple OSS Distributions  * cooperative pool's next best QoS requests.
1617*a325d9c4SApple OSS Distributions  */
1618*a325d9c4SApple OSS Distributions static bool
workq_threadreq_dequeue(struct workqueue * wq,workq_threadreq_t req,bool cooperative_sched_count_changed)1619*a325d9c4SApple OSS Distributions workq_threadreq_dequeue(struct workqueue *wq, workq_threadreq_t req,
1620*a325d9c4SApple OSS Distributions     bool cooperative_sched_count_changed)
1621*a325d9c4SApple OSS Distributions {
1622*a325d9c4SApple OSS Distributions 	wq->wq_reqcount--;
1623*a325d9c4SApple OSS Distributions 
1624*a325d9c4SApple OSS Distributions 	bool next_highest_request_changed = false;
1625*a325d9c4SApple OSS Distributions 
1626*a325d9c4SApple OSS Distributions 	if (--req->tr_count == 0) {
1627*a325d9c4SApple OSS Distributions 		if (req->tr_qos == WORKQ_THREAD_QOS_MANAGER) {
1628*a325d9c4SApple OSS Distributions 			assert(wq->wq_event_manager_threadreq == req);
1629*a325d9c4SApple OSS Distributions 			assert(req->tr_count == 0);
1630*a325d9c4SApple OSS Distributions 			wq->wq_event_manager_threadreq = NULL;
1631*a325d9c4SApple OSS Distributions 
1632*a325d9c4SApple OSS Distributions 			/* If a cooperative thread was the one which picked up the manager
1633*a325d9c4SApple OSS Distributions 			 * thread request, we need to reevaluate the cooperative pool
1634*a325d9c4SApple OSS Distributions 			 * anyways.
1635*a325d9c4SApple OSS Distributions 			 */
1636*a325d9c4SApple OSS Distributions 			if (cooperative_sched_count_changed) {
1637*a325d9c4SApple OSS Distributions 				_wq_cooperative_queue_refresh_best_req_qos(wq);
1638*a325d9c4SApple OSS Distributions 			}
1639*a325d9c4SApple OSS Distributions 			return true;
1640*a325d9c4SApple OSS Distributions 		}
1641*a325d9c4SApple OSS Distributions 
1642*a325d9c4SApple OSS Distributions 		if (workq_threadreq_is_cooperative(req)) {
1643*a325d9c4SApple OSS Distributions 			assert(req->tr_qos != WORKQ_THREAD_QOS_MANAGER);
1644*a325d9c4SApple OSS Distributions 			assert(req->tr_qos != WORKQ_THREAD_QOS_ABOVEUI);
1645*a325d9c4SApple OSS Distributions 			assert(req->tr_qos == wq->wq_cooperative_queue_best_req_qos);
1646*a325d9c4SApple OSS Distributions 
1647*a325d9c4SApple OSS Distributions 			struct workq_threadreq_tailq *bucket = &wq->wq_cooperative_queue[_wq_bucket(req->tr_qos)];
1648*a325d9c4SApple OSS Distributions 			__assert_only workq_threadreq_t head = STAILQ_FIRST(bucket);
1649*a325d9c4SApple OSS Distributions 
1650*a325d9c4SApple OSS Distributions 			assert(head == req);
1651*a325d9c4SApple OSS Distributions 			STAILQ_REMOVE_HEAD(bucket, tr_link);
1652*a325d9c4SApple OSS Distributions 
1653*a325d9c4SApple OSS Distributions 			/*
1654*a325d9c4SApple OSS Distributions 			 * If the request we're dequeueing is cooperative, then the sched
1655*a325d9c4SApple OSS Distributions 			 * counts definitely changed.
1656*a325d9c4SApple OSS Distributions 			 */
1657*a325d9c4SApple OSS Distributions 			assert(cooperative_sched_count_changed);
1658*a325d9c4SApple OSS Distributions 		}
1659*a325d9c4SApple OSS Distributions 
1660*a325d9c4SApple OSS Distributions 		/*
1661*a325d9c4SApple OSS Distributions 		 * We want to do the cooperative pool refresh after dequeueing a
1662*a325d9c4SApple OSS Distributions 		 * cooperative thread request if any (to combine both effects into 1
1663*a325d9c4SApple OSS Distributions 		 * refresh operation)
1664*a325d9c4SApple OSS Distributions 		 */
1665*a325d9c4SApple OSS Distributions 		if (cooperative_sched_count_changed) {
1666*a325d9c4SApple OSS Distributions 			next_highest_request_changed = _wq_cooperative_queue_refresh_best_req_qos(wq);
1667*a325d9c4SApple OSS Distributions 		}
1668*a325d9c4SApple OSS Distributions 
1669*a325d9c4SApple OSS Distributions 		if (!workq_threadreq_is_cooperative(req)) {
1670*a325d9c4SApple OSS Distributions 			/*
1671*a325d9c4SApple OSS Distributions 			 * All other types of requests are enqueued in priority queues
1672*a325d9c4SApple OSS Distributions 			 */
1673*a325d9c4SApple OSS Distributions 
1674*a325d9c4SApple OSS Distributions 			if (priority_queue_remove(workq_priority_queue_for_req(wq, req),
1675*a325d9c4SApple OSS Distributions 			    &req->tr_entry)) {
1676*a325d9c4SApple OSS Distributions 				next_highest_request_changed |= true;
1677*a325d9c4SApple OSS Distributions 				if (workq_threadreq_is_nonovercommit(req)) {
1678*a325d9c4SApple OSS Distributions 					_wq_thactive_refresh_best_constrained_req_qos(wq);
1679*a325d9c4SApple OSS Distributions 				}
1680*a325d9c4SApple OSS Distributions 			}
1681*a325d9c4SApple OSS Distributions 		}
1682*a325d9c4SApple OSS Distributions 	}
1683*a325d9c4SApple OSS Distributions 
1684*a325d9c4SApple OSS Distributions 	return next_highest_request_changed;
1685*a325d9c4SApple OSS Distributions }
1686*a325d9c4SApple OSS Distributions 
1687*a325d9c4SApple OSS Distributions static void
workq_threadreq_destroy(proc_t p,workq_threadreq_t req)1688*a325d9c4SApple OSS Distributions workq_threadreq_destroy(proc_t p, workq_threadreq_t req)
1689*a325d9c4SApple OSS Distributions {
1690*a325d9c4SApple OSS Distributions 	req->tr_state = WORKQ_TR_STATE_CANCELED;
1691*a325d9c4SApple OSS Distributions 	if (req->tr_flags & (WORKQ_TR_FLAG_WORKLOOP | WORKQ_TR_FLAG_KEVENT)) {
1692*a325d9c4SApple OSS Distributions 		kqueue_threadreq_cancel(p, req);
1693*a325d9c4SApple OSS Distributions 	} else {
1694*a325d9c4SApple OSS Distributions 		zfree(workq_zone_threadreq, req);
1695*a325d9c4SApple OSS Distributions 	}
1696*a325d9c4SApple OSS Distributions }
1697*a325d9c4SApple OSS Distributions 
1698*a325d9c4SApple OSS Distributions #pragma mark workqueue thread creation thread calls
1699*a325d9c4SApple OSS Distributions 
1700*a325d9c4SApple OSS Distributions static inline bool
workq_thread_call_prepost(struct workqueue * wq,uint32_t sched,uint32_t pend,uint32_t fail_mask)1701*a325d9c4SApple OSS Distributions workq_thread_call_prepost(struct workqueue *wq, uint32_t sched, uint32_t pend,
1702*a325d9c4SApple OSS Distributions     uint32_t fail_mask)
1703*a325d9c4SApple OSS Distributions {
1704*a325d9c4SApple OSS Distributions 	uint32_t old_flags, new_flags;
1705*a325d9c4SApple OSS Distributions 
1706*a325d9c4SApple OSS Distributions 	os_atomic_rmw_loop(&wq->wq_flags, old_flags, new_flags, acquire, {
1707*a325d9c4SApple OSS Distributions 		if (__improbable(old_flags & (WQ_EXITING | sched | pend | fail_mask))) {
1708*a325d9c4SApple OSS Distributions 		        os_atomic_rmw_loop_give_up(return false);
1709*a325d9c4SApple OSS Distributions 		}
1710*a325d9c4SApple OSS Distributions 		if (__improbable(old_flags & WQ_PROC_SUSPENDED)) {
1711*a325d9c4SApple OSS Distributions 		        new_flags = old_flags | pend;
1712*a325d9c4SApple OSS Distributions 		} else {
1713*a325d9c4SApple OSS Distributions 		        new_flags = old_flags | sched;
1714*a325d9c4SApple OSS Distributions 		}
1715*a325d9c4SApple OSS Distributions 	});
1716*a325d9c4SApple OSS Distributions 
1717*a325d9c4SApple OSS Distributions 	return (old_flags & WQ_PROC_SUSPENDED) == 0;
1718*a325d9c4SApple OSS Distributions }
1719*a325d9c4SApple OSS Distributions 
1720*a325d9c4SApple OSS Distributions #define WORKQ_SCHEDULE_DELAYED_THREAD_CREATION_RESTART 0x1
1721*a325d9c4SApple OSS Distributions 
1722*a325d9c4SApple OSS Distributions static bool
workq_schedule_delayed_thread_creation(struct workqueue * wq,int flags)1723*a325d9c4SApple OSS Distributions workq_schedule_delayed_thread_creation(struct workqueue *wq, int flags)
1724*a325d9c4SApple OSS Distributions {
1725*a325d9c4SApple OSS Distributions 	assert(!preemption_enabled());
1726*a325d9c4SApple OSS Distributions 
1727*a325d9c4SApple OSS Distributions 	if (!workq_thread_call_prepost(wq, WQ_DELAYED_CALL_SCHEDULED,
1728*a325d9c4SApple OSS Distributions 	    WQ_DELAYED_CALL_PENDED, WQ_IMMEDIATE_CALL_PENDED |
1729*a325d9c4SApple OSS Distributions 	    WQ_IMMEDIATE_CALL_SCHEDULED)) {
1730*a325d9c4SApple OSS Distributions 		return false;
1731*a325d9c4SApple OSS Distributions 	}
1732*a325d9c4SApple OSS Distributions 
1733*a325d9c4SApple OSS Distributions 	uint64_t now = mach_absolute_time();
1734*a325d9c4SApple OSS Distributions 
1735*a325d9c4SApple OSS Distributions 	if (flags & WORKQ_SCHEDULE_DELAYED_THREAD_CREATION_RESTART) {
1736*a325d9c4SApple OSS Distributions 		/* do not change the window */
1737*a325d9c4SApple OSS Distributions 	} else if (now - wq->wq_thread_call_last_run <= wq->wq_timer_interval) {
1738*a325d9c4SApple OSS Distributions 		wq->wq_timer_interval *= 2;
1739*a325d9c4SApple OSS Distributions 		if (wq->wq_timer_interval > wq_max_timer_interval.abstime) {
1740*a325d9c4SApple OSS Distributions 			wq->wq_timer_interval = (uint32_t)wq_max_timer_interval.abstime;
1741*a325d9c4SApple OSS Distributions 		}
1742*a325d9c4SApple OSS Distributions 	} else if (now - wq->wq_thread_call_last_run > 2 * wq->wq_timer_interval) {
1743*a325d9c4SApple OSS Distributions 		wq->wq_timer_interval /= 2;
1744*a325d9c4SApple OSS Distributions 		if (wq->wq_timer_interval < wq_stalled_window.abstime) {
1745*a325d9c4SApple OSS Distributions 			wq->wq_timer_interval = (uint32_t)wq_stalled_window.abstime;
1746*a325d9c4SApple OSS Distributions 		}
1747*a325d9c4SApple OSS Distributions 	}
1748*a325d9c4SApple OSS Distributions 
1749*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_start_add_timer, wq, wq->wq_reqcount,
1750*a325d9c4SApple OSS Distributions 	    _wq_flags(wq), wq->wq_timer_interval);
1751*a325d9c4SApple OSS Distributions 
1752*a325d9c4SApple OSS Distributions 	thread_call_t call = wq->wq_delayed_call;
1753*a325d9c4SApple OSS Distributions 	uintptr_t arg = WQ_DELAYED_CALL_SCHEDULED;
1754*a325d9c4SApple OSS Distributions 	uint64_t deadline = now + wq->wq_timer_interval;
1755*a325d9c4SApple OSS Distributions 	if (thread_call_enter1_delayed(call, (void *)arg, deadline)) {
1756*a325d9c4SApple OSS Distributions 		panic("delayed_call was already enqueued");
1757*a325d9c4SApple OSS Distributions 	}
1758*a325d9c4SApple OSS Distributions 	return true;
1759*a325d9c4SApple OSS Distributions }
1760*a325d9c4SApple OSS Distributions 
1761*a325d9c4SApple OSS Distributions static void
workq_schedule_immediate_thread_creation(struct workqueue * wq)1762*a325d9c4SApple OSS Distributions workq_schedule_immediate_thread_creation(struct workqueue *wq)
1763*a325d9c4SApple OSS Distributions {
1764*a325d9c4SApple OSS Distributions 	assert(!preemption_enabled());
1765*a325d9c4SApple OSS Distributions 
1766*a325d9c4SApple OSS Distributions 	if (workq_thread_call_prepost(wq, WQ_IMMEDIATE_CALL_SCHEDULED,
1767*a325d9c4SApple OSS Distributions 	    WQ_IMMEDIATE_CALL_PENDED, 0)) {
1768*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_start_add_timer, wq, wq->wq_reqcount,
1769*a325d9c4SApple OSS Distributions 		    _wq_flags(wq), 0);
1770*a325d9c4SApple OSS Distributions 
1771*a325d9c4SApple OSS Distributions 		uintptr_t arg = WQ_IMMEDIATE_CALL_SCHEDULED;
1772*a325d9c4SApple OSS Distributions 		if (thread_call_enter1(wq->wq_immediate_call, (void *)arg)) {
1773*a325d9c4SApple OSS Distributions 			panic("immediate_call was already enqueued");
1774*a325d9c4SApple OSS Distributions 		}
1775*a325d9c4SApple OSS Distributions 	}
1776*a325d9c4SApple OSS Distributions }
1777*a325d9c4SApple OSS Distributions 
1778*a325d9c4SApple OSS Distributions void
workq_proc_suspended(struct proc * p)1779*a325d9c4SApple OSS Distributions workq_proc_suspended(struct proc *p)
1780*a325d9c4SApple OSS Distributions {
1781*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
1782*a325d9c4SApple OSS Distributions 
1783*a325d9c4SApple OSS Distributions 	if (wq) {
1784*a325d9c4SApple OSS Distributions 		os_atomic_or(&wq->wq_flags, WQ_PROC_SUSPENDED, relaxed);
1785*a325d9c4SApple OSS Distributions 	}
1786*a325d9c4SApple OSS Distributions }
1787*a325d9c4SApple OSS Distributions 
1788*a325d9c4SApple OSS Distributions void
workq_proc_resumed(struct proc * p)1789*a325d9c4SApple OSS Distributions workq_proc_resumed(struct proc *p)
1790*a325d9c4SApple OSS Distributions {
1791*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
1792*a325d9c4SApple OSS Distributions 	uint32_t wq_flags;
1793*a325d9c4SApple OSS Distributions 
1794*a325d9c4SApple OSS Distributions 	if (!wq) {
1795*a325d9c4SApple OSS Distributions 		return;
1796*a325d9c4SApple OSS Distributions 	}
1797*a325d9c4SApple OSS Distributions 
1798*a325d9c4SApple OSS Distributions 	wq_flags = os_atomic_andnot_orig(&wq->wq_flags, WQ_PROC_SUSPENDED |
1799*a325d9c4SApple OSS Distributions 	    WQ_DELAYED_CALL_PENDED | WQ_IMMEDIATE_CALL_PENDED, relaxed);
1800*a325d9c4SApple OSS Distributions 	if ((wq_flags & WQ_EXITING) == 0) {
1801*a325d9c4SApple OSS Distributions 		disable_preemption();
1802*a325d9c4SApple OSS Distributions 		if (wq_flags & WQ_IMMEDIATE_CALL_PENDED) {
1803*a325d9c4SApple OSS Distributions 			workq_schedule_immediate_thread_creation(wq);
1804*a325d9c4SApple OSS Distributions 		} else if (wq_flags & WQ_DELAYED_CALL_PENDED) {
1805*a325d9c4SApple OSS Distributions 			workq_schedule_delayed_thread_creation(wq,
1806*a325d9c4SApple OSS Distributions 			    WORKQ_SCHEDULE_DELAYED_THREAD_CREATION_RESTART);
1807*a325d9c4SApple OSS Distributions 		}
1808*a325d9c4SApple OSS Distributions 		enable_preemption();
1809*a325d9c4SApple OSS Distributions 	}
1810*a325d9c4SApple OSS Distributions }
1811*a325d9c4SApple OSS Distributions 
1812*a325d9c4SApple OSS Distributions /**
1813*a325d9c4SApple OSS Distributions  * returns whether lastblocked_tsp is within wq_stalled_window usecs of now
1814*a325d9c4SApple OSS Distributions  */
1815*a325d9c4SApple OSS Distributions static bool
workq_thread_is_busy(uint64_t now,_Atomic uint64_t * lastblocked_tsp)1816*a325d9c4SApple OSS Distributions workq_thread_is_busy(uint64_t now, _Atomic uint64_t *lastblocked_tsp)
1817*a325d9c4SApple OSS Distributions {
1818*a325d9c4SApple OSS Distributions 	uint64_t lastblocked_ts = os_atomic_load_wide(lastblocked_tsp, relaxed);
1819*a325d9c4SApple OSS Distributions 	if (now <= lastblocked_ts) {
1820*a325d9c4SApple OSS Distributions 		/*
1821*a325d9c4SApple OSS Distributions 		 * Because the update of the timestamp when a thread blocks
1822*a325d9c4SApple OSS Distributions 		 * isn't serialized against us looking at it (i.e. we don't hold
1823*a325d9c4SApple OSS Distributions 		 * the workq lock), it's possible to have a timestamp that matches
1824*a325d9c4SApple OSS Distributions 		 * the current time or that even looks to be in the future relative
1825*a325d9c4SApple OSS Distributions 		 * to when we grabbed the current time...
1826*a325d9c4SApple OSS Distributions 		 *
1827*a325d9c4SApple OSS Distributions 		 * Just treat this as a busy thread since it must have just blocked.
1828*a325d9c4SApple OSS Distributions 		 */
1829*a325d9c4SApple OSS Distributions 		return true;
1830*a325d9c4SApple OSS Distributions 	}
1831*a325d9c4SApple OSS Distributions 	return (now - lastblocked_ts) < wq_stalled_window.abstime;
1832*a325d9c4SApple OSS Distributions }
1833*a325d9c4SApple OSS Distributions 
1834*a325d9c4SApple OSS Distributions static void
workq_add_new_threads_call(void * _p,void * flags)1835*a325d9c4SApple OSS Distributions workq_add_new_threads_call(void *_p, void *flags)
1836*a325d9c4SApple OSS Distributions {
1837*a325d9c4SApple OSS Distributions 	proc_t p = _p;
1838*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
1839*a325d9c4SApple OSS Distributions 	uint32_t my_flag = (uint32_t)(uintptr_t)flags;
1840*a325d9c4SApple OSS Distributions 
1841*a325d9c4SApple OSS Distributions 	/*
1842*a325d9c4SApple OSS Distributions 	 * workq_exit() will set the workqueue to NULL before
1843*a325d9c4SApple OSS Distributions 	 * it cancels thread calls.
1844*a325d9c4SApple OSS Distributions 	 */
1845*a325d9c4SApple OSS Distributions 	if (!wq) {
1846*a325d9c4SApple OSS Distributions 		return;
1847*a325d9c4SApple OSS Distributions 	}
1848*a325d9c4SApple OSS Distributions 
1849*a325d9c4SApple OSS Distributions 	assert((my_flag == WQ_DELAYED_CALL_SCHEDULED) ||
1850*a325d9c4SApple OSS Distributions 	    (my_flag == WQ_IMMEDIATE_CALL_SCHEDULED));
1851*a325d9c4SApple OSS Distributions 
1852*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_add_timer | DBG_FUNC_START, wq, _wq_flags(wq),
1853*a325d9c4SApple OSS Distributions 	    wq->wq_nthreads, wq->wq_thidlecount);
1854*a325d9c4SApple OSS Distributions 
1855*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
1856*a325d9c4SApple OSS Distributions 
1857*a325d9c4SApple OSS Distributions 	wq->wq_thread_call_last_run = mach_absolute_time();
1858*a325d9c4SApple OSS Distributions 	os_atomic_andnot(&wq->wq_flags, my_flag, release);
1859*a325d9c4SApple OSS Distributions 
1860*a325d9c4SApple OSS Distributions 	/* This can drop the workqueue lock, and take it again */
1861*a325d9c4SApple OSS Distributions 	workq_schedule_creator(p, wq, WORKQ_THREADREQ_CAN_CREATE_THREADS);
1862*a325d9c4SApple OSS Distributions 
1863*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
1864*a325d9c4SApple OSS Distributions 
1865*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_add_timer | DBG_FUNC_END, wq, 0,
1866*a325d9c4SApple OSS Distributions 	    wq->wq_nthreads, wq->wq_thidlecount);
1867*a325d9c4SApple OSS Distributions }
1868*a325d9c4SApple OSS Distributions 
1869*a325d9c4SApple OSS Distributions #pragma mark thread state tracking
1870*a325d9c4SApple OSS Distributions 
1871*a325d9c4SApple OSS Distributions static void
workq_sched_callback(int type,thread_t thread)1872*a325d9c4SApple OSS Distributions workq_sched_callback(int type, thread_t thread)
1873*a325d9c4SApple OSS Distributions {
1874*a325d9c4SApple OSS Distributions 	thread_ro_t tro = get_thread_ro(thread);
1875*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(thread);
1876*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(tro->tro_proc);
1877*a325d9c4SApple OSS Distributions 	thread_qos_t req_qos, qos = uth->uu_workq_pri.qos_bucket;
1878*a325d9c4SApple OSS Distributions 	wq_thactive_t old_thactive;
1879*a325d9c4SApple OSS Distributions 	bool start_timer = false;
1880*a325d9c4SApple OSS Distributions 
1881*a325d9c4SApple OSS Distributions 	if (qos == WORKQ_THREAD_QOS_MANAGER) {
1882*a325d9c4SApple OSS Distributions 		return;
1883*a325d9c4SApple OSS Distributions 	}
1884*a325d9c4SApple OSS Distributions 
1885*a325d9c4SApple OSS Distributions 	switch (type) {
1886*a325d9c4SApple OSS Distributions 	case SCHED_CALL_BLOCK:
1887*a325d9c4SApple OSS Distributions 		old_thactive = _wq_thactive_dec(wq, qos);
1888*a325d9c4SApple OSS Distributions 		req_qos = WQ_THACTIVE_BEST_CONSTRAINED_REQ_QOS(old_thactive);
1889*a325d9c4SApple OSS Distributions 
1890*a325d9c4SApple OSS Distributions 		/*
1891*a325d9c4SApple OSS Distributions 		 * Remember the timestamp of the last thread that blocked in this
1892*a325d9c4SApple OSS Distributions 		 * bucket, it used used by admission checks to ignore one thread
1893*a325d9c4SApple OSS Distributions 		 * being inactive if this timestamp is recent enough.
1894*a325d9c4SApple OSS Distributions 		 *
1895*a325d9c4SApple OSS Distributions 		 * If we collide with another thread trying to update the
1896*a325d9c4SApple OSS Distributions 		 * last_blocked (really unlikely since another thread would have to
1897*a325d9c4SApple OSS Distributions 		 * get scheduled and then block after we start down this path), it's
1898*a325d9c4SApple OSS Distributions 		 * not a problem.  Either timestamp is adequate, so no need to retry
1899*a325d9c4SApple OSS Distributions 		 */
1900*a325d9c4SApple OSS Distributions 		os_atomic_store_wide(&wq->wq_lastblocked_ts[_wq_bucket(qos)],
1901*a325d9c4SApple OSS Distributions 		    thread_last_run_time(thread), relaxed);
1902*a325d9c4SApple OSS Distributions 
1903*a325d9c4SApple OSS Distributions 		if (req_qos == THREAD_QOS_UNSPECIFIED) {
1904*a325d9c4SApple OSS Distributions 			/*
1905*a325d9c4SApple OSS Distributions 			 * No pending request at the moment we could unblock, move on.
1906*a325d9c4SApple OSS Distributions 			 */
1907*a325d9c4SApple OSS Distributions 		} else if (qos < req_qos) {
1908*a325d9c4SApple OSS Distributions 			/*
1909*a325d9c4SApple OSS Distributions 			 * The blocking thread is at a lower QoS than the highest currently
1910*a325d9c4SApple OSS Distributions 			 * pending constrained request, nothing has to be redriven
1911*a325d9c4SApple OSS Distributions 			 */
1912*a325d9c4SApple OSS Distributions 		} else {
1913*a325d9c4SApple OSS Distributions 			uint32_t max_busycount, old_req_count;
1914*a325d9c4SApple OSS Distributions 			old_req_count = _wq_thactive_aggregate_downto_qos(wq, old_thactive,
1915*a325d9c4SApple OSS Distributions 			    req_qos, NULL, &max_busycount);
1916*a325d9c4SApple OSS Distributions 			/*
1917*a325d9c4SApple OSS Distributions 			 * If it is possible that may_start_constrained_thread had refused
1918*a325d9c4SApple OSS Distributions 			 * admission due to being over the max concurrency, we may need to
1919*a325d9c4SApple OSS Distributions 			 * spin up a new thread.
1920*a325d9c4SApple OSS Distributions 			 *
1921*a325d9c4SApple OSS Distributions 			 * We take into account the maximum number of busy threads
1922*a325d9c4SApple OSS Distributions 			 * that can affect may_start_constrained_thread as looking at the
1923*a325d9c4SApple OSS Distributions 			 * actual number may_start_constrained_thread will see is racy.
1924*a325d9c4SApple OSS Distributions 			 *
1925*a325d9c4SApple OSS Distributions 			 * IOW at NCPU = 4, for IN (req_qos = 1), if the old req count is
1926*a325d9c4SApple OSS Distributions 			 * between NCPU (4) and NCPU - 2 (2) we need to redrive.
1927*a325d9c4SApple OSS Distributions 			 */
1928*a325d9c4SApple OSS Distributions 			uint32_t conc = wq_max_parallelism[_wq_bucket(qos)];
1929*a325d9c4SApple OSS Distributions 			if (old_req_count <= conc && conc <= old_req_count + max_busycount) {
1930*a325d9c4SApple OSS Distributions 				start_timer = workq_schedule_delayed_thread_creation(wq, 0);
1931*a325d9c4SApple OSS Distributions 			}
1932*a325d9c4SApple OSS Distributions 		}
1933*a325d9c4SApple OSS Distributions 		if (__improbable(kdebug_enable)) {
1934*a325d9c4SApple OSS Distributions 			__unused uint32_t old = _wq_thactive_aggregate_downto_qos(wq,
1935*a325d9c4SApple OSS Distributions 			    old_thactive, qos, NULL, NULL);
1936*a325d9c4SApple OSS Distributions 			WQ_TRACE_WQ(TRACE_wq_thread_block | DBG_FUNC_START, wq,
1937*a325d9c4SApple OSS Distributions 			    old - 1, qos | (req_qos << 8),
1938*a325d9c4SApple OSS Distributions 			    wq->wq_reqcount << 1 | start_timer);
1939*a325d9c4SApple OSS Distributions 		}
1940*a325d9c4SApple OSS Distributions 		break;
1941*a325d9c4SApple OSS Distributions 
1942*a325d9c4SApple OSS Distributions 	case SCHED_CALL_UNBLOCK:
1943*a325d9c4SApple OSS Distributions 		/*
1944*a325d9c4SApple OSS Distributions 		 * we cannot take the workqueue_lock here...
1945*a325d9c4SApple OSS Distributions 		 * an UNBLOCK can occur from a timer event which
1946*a325d9c4SApple OSS Distributions 		 * is run from an interrupt context... if the workqueue_lock
1947*a325d9c4SApple OSS Distributions 		 * is already held by this processor, we'll deadlock...
1948*a325d9c4SApple OSS Distributions 		 * the thread lock for the thread being UNBLOCKED
1949*a325d9c4SApple OSS Distributions 		 * is also held
1950*a325d9c4SApple OSS Distributions 		 */
1951*a325d9c4SApple OSS Distributions 		old_thactive = _wq_thactive_inc(wq, qos);
1952*a325d9c4SApple OSS Distributions 		if (__improbable(kdebug_enable)) {
1953*a325d9c4SApple OSS Distributions 			__unused uint32_t old = _wq_thactive_aggregate_downto_qos(wq,
1954*a325d9c4SApple OSS Distributions 			    old_thactive, qos, NULL, NULL);
1955*a325d9c4SApple OSS Distributions 			req_qos = WQ_THACTIVE_BEST_CONSTRAINED_REQ_QOS(old_thactive);
1956*a325d9c4SApple OSS Distributions 			WQ_TRACE_WQ(TRACE_wq_thread_block | DBG_FUNC_END, wq,
1957*a325d9c4SApple OSS Distributions 			    old + 1, qos | (req_qos << 8),
1958*a325d9c4SApple OSS Distributions 			    wq->wq_threads_scheduled);
1959*a325d9c4SApple OSS Distributions 		}
1960*a325d9c4SApple OSS Distributions 		break;
1961*a325d9c4SApple OSS Distributions 	}
1962*a325d9c4SApple OSS Distributions }
1963*a325d9c4SApple OSS Distributions 
1964*a325d9c4SApple OSS Distributions #pragma mark workq lifecycle
1965*a325d9c4SApple OSS Distributions 
1966*a325d9c4SApple OSS Distributions void
workq_reference(struct workqueue * wq)1967*a325d9c4SApple OSS Distributions workq_reference(struct workqueue *wq)
1968*a325d9c4SApple OSS Distributions {
1969*a325d9c4SApple OSS Distributions 	os_ref_retain(&wq->wq_refcnt);
1970*a325d9c4SApple OSS Distributions }
1971*a325d9c4SApple OSS Distributions 
1972*a325d9c4SApple OSS Distributions static void
workq_deallocate_queue_invoke(mpsc_queue_chain_t e,__assert_only mpsc_daemon_queue_t dq)1973*a325d9c4SApple OSS Distributions workq_deallocate_queue_invoke(mpsc_queue_chain_t e,
1974*a325d9c4SApple OSS Distributions     __assert_only mpsc_daemon_queue_t dq)
1975*a325d9c4SApple OSS Distributions {
1976*a325d9c4SApple OSS Distributions 	struct workqueue *wq;
1977*a325d9c4SApple OSS Distributions 	struct turnstile *ts;
1978*a325d9c4SApple OSS Distributions 
1979*a325d9c4SApple OSS Distributions 	wq = mpsc_queue_element(e, struct workqueue, wq_destroy_link);
1980*a325d9c4SApple OSS Distributions 	assert(dq == &workq_deallocate_queue);
1981*a325d9c4SApple OSS Distributions 
1982*a325d9c4SApple OSS Distributions 	turnstile_complete((uintptr_t)wq, &wq->wq_turnstile, &ts, TURNSTILE_WORKQS);
1983*a325d9c4SApple OSS Distributions 	assert(ts);
1984*a325d9c4SApple OSS Distributions 	turnstile_cleanup();
1985*a325d9c4SApple OSS Distributions 	turnstile_deallocate(ts);
1986*a325d9c4SApple OSS Distributions 
1987*a325d9c4SApple OSS Distributions 	lck_ticket_destroy(&wq->wq_lock, &workq_lck_grp);
1988*a325d9c4SApple OSS Distributions 	zfree(workq_zone_workqueue, wq);
1989*a325d9c4SApple OSS Distributions }
1990*a325d9c4SApple OSS Distributions 
1991*a325d9c4SApple OSS Distributions static void
workq_deallocate(struct workqueue * wq)1992*a325d9c4SApple OSS Distributions workq_deallocate(struct workqueue *wq)
1993*a325d9c4SApple OSS Distributions {
1994*a325d9c4SApple OSS Distributions 	if (os_ref_release_relaxed(&wq->wq_refcnt) == 0) {
1995*a325d9c4SApple OSS Distributions 		workq_deallocate_queue_invoke(&wq->wq_destroy_link,
1996*a325d9c4SApple OSS Distributions 		    &workq_deallocate_queue);
1997*a325d9c4SApple OSS Distributions 	}
1998*a325d9c4SApple OSS Distributions }
1999*a325d9c4SApple OSS Distributions 
2000*a325d9c4SApple OSS Distributions void
workq_deallocate_safe(struct workqueue * wq)2001*a325d9c4SApple OSS Distributions workq_deallocate_safe(struct workqueue *wq)
2002*a325d9c4SApple OSS Distributions {
2003*a325d9c4SApple OSS Distributions 	if (__improbable(os_ref_release_relaxed(&wq->wq_refcnt) == 0)) {
2004*a325d9c4SApple OSS Distributions 		mpsc_daemon_enqueue(&workq_deallocate_queue, &wq->wq_destroy_link,
2005*a325d9c4SApple OSS Distributions 		    MPSC_QUEUE_DISABLE_PREEMPTION);
2006*a325d9c4SApple OSS Distributions 	}
2007*a325d9c4SApple OSS Distributions }
2008*a325d9c4SApple OSS Distributions 
2009*a325d9c4SApple OSS Distributions /**
2010*a325d9c4SApple OSS Distributions  * Setup per-process state for the workqueue.
2011*a325d9c4SApple OSS Distributions  */
2012*a325d9c4SApple OSS Distributions int
workq_open(struct proc * p,__unused struct workq_open_args * uap,__unused int32_t * retval)2013*a325d9c4SApple OSS Distributions workq_open(struct proc *p, __unused struct workq_open_args *uap,
2014*a325d9c4SApple OSS Distributions     __unused int32_t *retval)
2015*a325d9c4SApple OSS Distributions {
2016*a325d9c4SApple OSS Distributions 	struct workqueue *wq;
2017*a325d9c4SApple OSS Distributions 	int error = 0;
2018*a325d9c4SApple OSS Distributions 
2019*a325d9c4SApple OSS Distributions 	if ((p->p_lflag & P_LREGISTER) == 0) {
2020*a325d9c4SApple OSS Distributions 		return EINVAL;
2021*a325d9c4SApple OSS Distributions 	}
2022*a325d9c4SApple OSS Distributions 
2023*a325d9c4SApple OSS Distributions 	if (wq_init_constrained_limit) {
2024*a325d9c4SApple OSS Distributions 		uint32_t limit, num_cpus = ml_wait_max_cpus();
2025*a325d9c4SApple OSS Distributions 
2026*a325d9c4SApple OSS Distributions 		/*
2027*a325d9c4SApple OSS Distributions 		 * set up the limit for the constrained pool
2028*a325d9c4SApple OSS Distributions 		 * this is a virtual pool in that we don't
2029*a325d9c4SApple OSS Distributions 		 * maintain it on a separate idle and run list
2030*a325d9c4SApple OSS Distributions 		 */
2031*a325d9c4SApple OSS Distributions 		limit = num_cpus * WORKQUEUE_CONSTRAINED_FACTOR;
2032*a325d9c4SApple OSS Distributions 
2033*a325d9c4SApple OSS Distributions 		if (limit > wq_max_constrained_threads) {
2034*a325d9c4SApple OSS Distributions 			wq_max_constrained_threads = limit;
2035*a325d9c4SApple OSS Distributions 		}
2036*a325d9c4SApple OSS Distributions 
2037*a325d9c4SApple OSS Distributions 		if (wq_max_threads > WQ_THACTIVE_BUCKET_HALF) {
2038*a325d9c4SApple OSS Distributions 			wq_max_threads = WQ_THACTIVE_BUCKET_HALF;
2039*a325d9c4SApple OSS Distributions 		}
2040*a325d9c4SApple OSS Distributions 		if (wq_max_threads > CONFIG_THREAD_MAX - 20) {
2041*a325d9c4SApple OSS Distributions 			wq_max_threads = CONFIG_THREAD_MAX - 20;
2042*a325d9c4SApple OSS Distributions 		}
2043*a325d9c4SApple OSS Distributions 
2044*a325d9c4SApple OSS Distributions 		wq_death_max_load = (uint16_t)fls(num_cpus) + 1;
2045*a325d9c4SApple OSS Distributions 
2046*a325d9c4SApple OSS Distributions 		for (thread_qos_t qos = WORKQ_THREAD_QOS_MIN; qos <= WORKQ_THREAD_QOS_MAX; qos++) {
2047*a325d9c4SApple OSS Distributions 			wq_max_parallelism[_wq_bucket(qos)] =
2048*a325d9c4SApple OSS Distributions 			    qos_max_parallelism(qos, QOS_PARALLELISM_COUNT_LOGICAL);
2049*a325d9c4SApple OSS Distributions 		}
2050*a325d9c4SApple OSS Distributions 
2051*a325d9c4SApple OSS Distributions 		wq_max_cooperative_threads = num_cpus;
2052*a325d9c4SApple OSS Distributions 
2053*a325d9c4SApple OSS Distributions 		wq_init_constrained_limit = 0;
2054*a325d9c4SApple OSS Distributions 	}
2055*a325d9c4SApple OSS Distributions 
2056*a325d9c4SApple OSS Distributions 	if (proc_get_wqptr(p) == NULL) {
2057*a325d9c4SApple OSS Distributions 		if (proc_init_wqptr_or_wait(p) == FALSE) {
2058*a325d9c4SApple OSS Distributions 			assert(proc_get_wqptr(p) != NULL);
2059*a325d9c4SApple OSS Distributions 			goto out;
2060*a325d9c4SApple OSS Distributions 		}
2061*a325d9c4SApple OSS Distributions 
2062*a325d9c4SApple OSS Distributions 		wq = zalloc_flags(workq_zone_workqueue, Z_WAITOK | Z_ZERO);
2063*a325d9c4SApple OSS Distributions 
2064*a325d9c4SApple OSS Distributions 		os_ref_init_count(&wq->wq_refcnt, &workq_refgrp, 1);
2065*a325d9c4SApple OSS Distributions 
2066*a325d9c4SApple OSS Distributions 		// Start the event manager at the priority hinted at by the policy engine
2067*a325d9c4SApple OSS Distributions 		thread_qos_t mgr_priority_hint = task_get_default_manager_qos(current_task());
2068*a325d9c4SApple OSS Distributions 		pthread_priority_t pp = _pthread_priority_make_from_thread_qos(mgr_priority_hint, 0, 0);
2069*a325d9c4SApple OSS Distributions 		wq->wq_event_manager_priority = (uint32_t)pp;
2070*a325d9c4SApple OSS Distributions 		wq->wq_timer_interval = (uint32_t)wq_stalled_window.abstime;
2071*a325d9c4SApple OSS Distributions 		wq->wq_proc = p;
2072*a325d9c4SApple OSS Distributions 		turnstile_prepare((uintptr_t)wq, &wq->wq_turnstile, turnstile_alloc(),
2073*a325d9c4SApple OSS Distributions 		    TURNSTILE_WORKQS);
2074*a325d9c4SApple OSS Distributions 
2075*a325d9c4SApple OSS Distributions 		TAILQ_INIT(&wq->wq_thrunlist);
2076*a325d9c4SApple OSS Distributions 		TAILQ_INIT(&wq->wq_thnewlist);
2077*a325d9c4SApple OSS Distributions 		TAILQ_INIT(&wq->wq_thidlelist);
2078*a325d9c4SApple OSS Distributions 		priority_queue_init(&wq->wq_overcommit_queue);
2079*a325d9c4SApple OSS Distributions 		priority_queue_init(&wq->wq_constrained_queue);
2080*a325d9c4SApple OSS Distributions 		priority_queue_init(&wq->wq_special_queue);
2081*a325d9c4SApple OSS Distributions 		for (int bucket = 0; bucket < WORKQ_NUM_QOS_BUCKETS; bucket++) {
2082*a325d9c4SApple OSS Distributions 			STAILQ_INIT(&wq->wq_cooperative_queue[bucket]);
2083*a325d9c4SApple OSS Distributions 		}
2084*a325d9c4SApple OSS Distributions 
2085*a325d9c4SApple OSS Distributions 		/* We are only using the delayed thread call for the constrained pool
2086*a325d9c4SApple OSS Distributions 		 * which can't have work at >= UI QoS and so we can be fine with a
2087*a325d9c4SApple OSS Distributions 		 * UI QoS thread call.
2088*a325d9c4SApple OSS Distributions 		 */
2089*a325d9c4SApple OSS Distributions 		wq->wq_delayed_call = thread_call_allocate_with_qos(
2090*a325d9c4SApple OSS Distributions 			workq_add_new_threads_call, p, THREAD_QOS_USER_INTERACTIVE,
2091*a325d9c4SApple OSS Distributions 			THREAD_CALL_OPTIONS_ONCE);
2092*a325d9c4SApple OSS Distributions 		wq->wq_immediate_call = thread_call_allocate_with_options(
2093*a325d9c4SApple OSS Distributions 			workq_add_new_threads_call, p, THREAD_CALL_PRIORITY_KERNEL,
2094*a325d9c4SApple OSS Distributions 			THREAD_CALL_OPTIONS_ONCE);
2095*a325d9c4SApple OSS Distributions 		wq->wq_death_call = thread_call_allocate_with_options(
2096*a325d9c4SApple OSS Distributions 			workq_kill_old_threads_call, wq,
2097*a325d9c4SApple OSS Distributions 			THREAD_CALL_PRIORITY_USER, THREAD_CALL_OPTIONS_ONCE);
2098*a325d9c4SApple OSS Distributions 
2099*a325d9c4SApple OSS Distributions 		lck_ticket_init(&wq->wq_lock, &workq_lck_grp);
2100*a325d9c4SApple OSS Distributions 
2101*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_create | DBG_FUNC_NONE, wq,
2102*a325d9c4SApple OSS Distributions 		    VM_KERNEL_ADDRHIDE(wq), 0, 0);
2103*a325d9c4SApple OSS Distributions 		proc_set_wqptr(p, wq);
2104*a325d9c4SApple OSS Distributions 	}
2105*a325d9c4SApple OSS Distributions out:
2106*a325d9c4SApple OSS Distributions 
2107*a325d9c4SApple OSS Distributions 	return error;
2108*a325d9c4SApple OSS Distributions }
2109*a325d9c4SApple OSS Distributions 
2110*a325d9c4SApple OSS Distributions /*
2111*a325d9c4SApple OSS Distributions  * Routine:	workq_mark_exiting
2112*a325d9c4SApple OSS Distributions  *
2113*a325d9c4SApple OSS Distributions  * Function:	Mark the work queue such that new threads will not be added to the
2114*a325d9c4SApple OSS Distributions  *		work queue after we return.
2115*a325d9c4SApple OSS Distributions  *
2116*a325d9c4SApple OSS Distributions  * Conditions:	Called against the current process.
2117*a325d9c4SApple OSS Distributions  */
2118*a325d9c4SApple OSS Distributions void
workq_mark_exiting(struct proc * p)2119*a325d9c4SApple OSS Distributions workq_mark_exiting(struct proc *p)
2120*a325d9c4SApple OSS Distributions {
2121*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
2122*a325d9c4SApple OSS Distributions 	uint32_t wq_flags;
2123*a325d9c4SApple OSS Distributions 	workq_threadreq_t mgr_req;
2124*a325d9c4SApple OSS Distributions 
2125*a325d9c4SApple OSS Distributions 	if (!wq) {
2126*a325d9c4SApple OSS Distributions 		return;
2127*a325d9c4SApple OSS Distributions 	}
2128*a325d9c4SApple OSS Distributions 
2129*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_pthread_exit | DBG_FUNC_START, wq, 0, 0, 0);
2130*a325d9c4SApple OSS Distributions 
2131*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
2132*a325d9c4SApple OSS Distributions 
2133*a325d9c4SApple OSS Distributions 	wq_flags = os_atomic_or_orig(&wq->wq_flags, WQ_EXITING, relaxed);
2134*a325d9c4SApple OSS Distributions 	if (__improbable(wq_flags & WQ_EXITING)) {
2135*a325d9c4SApple OSS Distributions 		panic("workq_mark_exiting called twice");
2136*a325d9c4SApple OSS Distributions 	}
2137*a325d9c4SApple OSS Distributions 
2138*a325d9c4SApple OSS Distributions 	/*
2139*a325d9c4SApple OSS Distributions 	 * Opportunistically try to cancel thread calls that are likely in flight.
2140*a325d9c4SApple OSS Distributions 	 * workq_exit() will do the proper cleanup.
2141*a325d9c4SApple OSS Distributions 	 */
2142*a325d9c4SApple OSS Distributions 	if (wq_flags & WQ_IMMEDIATE_CALL_SCHEDULED) {
2143*a325d9c4SApple OSS Distributions 		thread_call_cancel(wq->wq_immediate_call);
2144*a325d9c4SApple OSS Distributions 	}
2145*a325d9c4SApple OSS Distributions 	if (wq_flags & WQ_DELAYED_CALL_SCHEDULED) {
2146*a325d9c4SApple OSS Distributions 		thread_call_cancel(wq->wq_delayed_call);
2147*a325d9c4SApple OSS Distributions 	}
2148*a325d9c4SApple OSS Distributions 	if (wq_flags & WQ_DEATH_CALL_SCHEDULED) {
2149*a325d9c4SApple OSS Distributions 		thread_call_cancel(wq->wq_death_call);
2150*a325d9c4SApple OSS Distributions 	}
2151*a325d9c4SApple OSS Distributions 
2152*a325d9c4SApple OSS Distributions 	mgr_req = wq->wq_event_manager_threadreq;
2153*a325d9c4SApple OSS Distributions 	wq->wq_event_manager_threadreq = NULL;
2154*a325d9c4SApple OSS Distributions 	wq->wq_reqcount = 0; /* workq_schedule_creator must not look at queues */
2155*a325d9c4SApple OSS Distributions 	wq->wq_creator = NULL;
2156*a325d9c4SApple OSS Distributions 	workq_turnstile_update_inheritor(wq, TURNSTILE_INHERITOR_NULL, 0);
2157*a325d9c4SApple OSS Distributions 
2158*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
2159*a325d9c4SApple OSS Distributions 
2160*a325d9c4SApple OSS Distributions 	if (mgr_req) {
2161*a325d9c4SApple OSS Distributions 		kqueue_threadreq_cancel(p, mgr_req);
2162*a325d9c4SApple OSS Distributions 	}
2163*a325d9c4SApple OSS Distributions 	/*
2164*a325d9c4SApple OSS Distributions 	 * No one touches the priority queues once WQ_EXITING is set.
2165*a325d9c4SApple OSS Distributions 	 * It is hence safe to do the tear down without holding any lock.
2166*a325d9c4SApple OSS Distributions 	 */
2167*a325d9c4SApple OSS Distributions 	priority_queue_destroy(&wq->wq_overcommit_queue,
2168*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry, ^(workq_threadreq_t e){
2169*a325d9c4SApple OSS Distributions 		workq_threadreq_destroy(p, e);
2170*a325d9c4SApple OSS Distributions 	});
2171*a325d9c4SApple OSS Distributions 	priority_queue_destroy(&wq->wq_constrained_queue,
2172*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry, ^(workq_threadreq_t e){
2173*a325d9c4SApple OSS Distributions 		workq_threadreq_destroy(p, e);
2174*a325d9c4SApple OSS Distributions 	});
2175*a325d9c4SApple OSS Distributions 	priority_queue_destroy(&wq->wq_special_queue,
2176*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry, ^(workq_threadreq_t e){
2177*a325d9c4SApple OSS Distributions 		workq_threadreq_destroy(p, e);
2178*a325d9c4SApple OSS Distributions 	});
2179*a325d9c4SApple OSS Distributions 
2180*a325d9c4SApple OSS Distributions 	WQ_TRACE(TRACE_wq_pthread_exit | DBG_FUNC_END, 0, 0, 0, 0);
2181*a325d9c4SApple OSS Distributions }
2182*a325d9c4SApple OSS Distributions 
2183*a325d9c4SApple OSS Distributions /*
2184*a325d9c4SApple OSS Distributions  * Routine:	workq_exit
2185*a325d9c4SApple OSS Distributions  *
2186*a325d9c4SApple OSS Distributions  * Function:	clean up the work queue structure(s) now that there are no threads
2187*a325d9c4SApple OSS Distributions  *		left running inside the work queue (except possibly current_thread).
2188*a325d9c4SApple OSS Distributions  *
2189*a325d9c4SApple OSS Distributions  * Conditions:	Called by the last thread in the process.
2190*a325d9c4SApple OSS Distributions  *		Called against current process.
2191*a325d9c4SApple OSS Distributions  */
2192*a325d9c4SApple OSS Distributions void
workq_exit(struct proc * p)2193*a325d9c4SApple OSS Distributions workq_exit(struct proc *p)
2194*a325d9c4SApple OSS Distributions {
2195*a325d9c4SApple OSS Distributions 	struct workqueue *wq;
2196*a325d9c4SApple OSS Distributions 	struct uthread *uth, *tmp;
2197*a325d9c4SApple OSS Distributions 
2198*a325d9c4SApple OSS Distributions 	wq = os_atomic_xchg(&p->p_wqptr, NULL, relaxed);
2199*a325d9c4SApple OSS Distributions 	if (wq != NULL) {
2200*a325d9c4SApple OSS Distributions 		thread_t th = current_thread();
2201*a325d9c4SApple OSS Distributions 
2202*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_workqueue_exit | DBG_FUNC_START, wq, 0, 0, 0);
2203*a325d9c4SApple OSS Distributions 
2204*a325d9c4SApple OSS Distributions 		if (thread_get_tag(th) & THREAD_TAG_WORKQUEUE) {
2205*a325d9c4SApple OSS Distributions 			/*
2206*a325d9c4SApple OSS Distributions 			 * <rdar://problem/40111515> Make sure we will no longer call the
2207*a325d9c4SApple OSS Distributions 			 * sched call, if we ever block this thread, which the cancel_wait
2208*a325d9c4SApple OSS Distributions 			 * below can do.
2209*a325d9c4SApple OSS Distributions 			 */
2210*a325d9c4SApple OSS Distributions 			thread_sched_call(th, NULL);
2211*a325d9c4SApple OSS Distributions 		}
2212*a325d9c4SApple OSS Distributions 
2213*a325d9c4SApple OSS Distributions 		/*
2214*a325d9c4SApple OSS Distributions 		 * Thread calls are always scheduled by the proc itself or under the
2215*a325d9c4SApple OSS Distributions 		 * workqueue spinlock if WQ_EXITING is not yet set.
2216*a325d9c4SApple OSS Distributions 		 *
2217*a325d9c4SApple OSS Distributions 		 * Either way, when this runs, the proc has no threads left beside
2218*a325d9c4SApple OSS Distributions 		 * the one running this very code, so we know no thread call can be
2219*a325d9c4SApple OSS Distributions 		 * dispatched anymore.
2220*a325d9c4SApple OSS Distributions 		 */
2221*a325d9c4SApple OSS Distributions 		thread_call_cancel_wait(wq->wq_delayed_call);
2222*a325d9c4SApple OSS Distributions 		thread_call_cancel_wait(wq->wq_immediate_call);
2223*a325d9c4SApple OSS Distributions 		thread_call_cancel_wait(wq->wq_death_call);
2224*a325d9c4SApple OSS Distributions 		thread_call_free(wq->wq_delayed_call);
2225*a325d9c4SApple OSS Distributions 		thread_call_free(wq->wq_immediate_call);
2226*a325d9c4SApple OSS Distributions 		thread_call_free(wq->wq_death_call);
2227*a325d9c4SApple OSS Distributions 
2228*a325d9c4SApple OSS Distributions 		/*
2229*a325d9c4SApple OSS Distributions 		 * Clean up workqueue data structures for threads that exited and
2230*a325d9c4SApple OSS Distributions 		 * didn't get a chance to clean up after themselves.
2231*a325d9c4SApple OSS Distributions 		 *
2232*a325d9c4SApple OSS Distributions 		 * idle/new threads should have been interrupted and died on their own
2233*a325d9c4SApple OSS Distributions 		 */
2234*a325d9c4SApple OSS Distributions 		TAILQ_FOREACH_SAFE(uth, &wq->wq_thrunlist, uu_workq_entry, tmp) {
2235*a325d9c4SApple OSS Distributions 			thread_t mth = get_machthread(uth);
2236*a325d9c4SApple OSS Distributions 			thread_sched_call(mth, NULL);
2237*a325d9c4SApple OSS Distributions 			thread_deallocate(mth);
2238*a325d9c4SApple OSS Distributions 		}
2239*a325d9c4SApple OSS Distributions 		assert(TAILQ_EMPTY(&wq->wq_thnewlist));
2240*a325d9c4SApple OSS Distributions 		assert(TAILQ_EMPTY(&wq->wq_thidlelist));
2241*a325d9c4SApple OSS Distributions 
2242*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_destroy | DBG_FUNC_END, wq,
2243*a325d9c4SApple OSS Distributions 		    VM_KERNEL_ADDRHIDE(wq), 0, 0);
2244*a325d9c4SApple OSS Distributions 
2245*a325d9c4SApple OSS Distributions 		workq_deallocate(wq);
2246*a325d9c4SApple OSS Distributions 
2247*a325d9c4SApple OSS Distributions 		WQ_TRACE(TRACE_wq_workqueue_exit | DBG_FUNC_END, 0, 0, 0, 0);
2248*a325d9c4SApple OSS Distributions 	}
2249*a325d9c4SApple OSS Distributions }
2250*a325d9c4SApple OSS Distributions 
2251*a325d9c4SApple OSS Distributions 
2252*a325d9c4SApple OSS Distributions #pragma mark bsd thread control
2253*a325d9c4SApple OSS Distributions 
2254*a325d9c4SApple OSS Distributions bool
bsdthread_part_of_cooperative_workqueue(struct uthread * uth)2255*a325d9c4SApple OSS Distributions bsdthread_part_of_cooperative_workqueue(struct uthread *uth)
2256*a325d9c4SApple OSS Distributions {
2257*a325d9c4SApple OSS Distributions 	return (workq_thread_is_cooperative(uth) || workq_thread_is_nonovercommit(uth)) &&
2258*a325d9c4SApple OSS Distributions 	       (uth->uu_workq_pri.qos_bucket != WORKQ_THREAD_QOS_MANAGER);
2259*a325d9c4SApple OSS Distributions }
2260*a325d9c4SApple OSS Distributions 
2261*a325d9c4SApple OSS Distributions static bool
_pthread_priority_to_policy(pthread_priority_t priority,thread_qos_policy_data_t * data)2262*a325d9c4SApple OSS Distributions _pthread_priority_to_policy(pthread_priority_t priority,
2263*a325d9c4SApple OSS Distributions     thread_qos_policy_data_t *data)
2264*a325d9c4SApple OSS Distributions {
2265*a325d9c4SApple OSS Distributions 	data->qos_tier = _pthread_priority_thread_qos(priority);
2266*a325d9c4SApple OSS Distributions 	data->tier_importance = _pthread_priority_relpri(priority);
2267*a325d9c4SApple OSS Distributions 	if (data->qos_tier == THREAD_QOS_UNSPECIFIED || data->tier_importance > 0 ||
2268*a325d9c4SApple OSS Distributions 	    data->tier_importance < THREAD_QOS_MIN_TIER_IMPORTANCE) {
2269*a325d9c4SApple OSS Distributions 		return false;
2270*a325d9c4SApple OSS Distributions 	}
2271*a325d9c4SApple OSS Distributions 	return true;
2272*a325d9c4SApple OSS Distributions }
2273*a325d9c4SApple OSS Distributions 
2274*a325d9c4SApple OSS Distributions static int
bsdthread_set_self(proc_t p,thread_t th,pthread_priority_t priority,mach_port_name_t voucher,enum workq_set_self_flags flags)2275*a325d9c4SApple OSS Distributions bsdthread_set_self(proc_t p, thread_t th, pthread_priority_t priority,
2276*a325d9c4SApple OSS Distributions     mach_port_name_t voucher, enum workq_set_self_flags flags)
2277*a325d9c4SApple OSS Distributions {
2278*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(th);
2279*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
2280*a325d9c4SApple OSS Distributions 
2281*a325d9c4SApple OSS Distributions 	kern_return_t kr;
2282*a325d9c4SApple OSS Distributions 	int unbind_rv = 0, qos_rv = 0, voucher_rv = 0, fixedpri_rv = 0;
2283*a325d9c4SApple OSS Distributions 	bool is_wq_thread = (thread_get_tag(th) & THREAD_TAG_WORKQUEUE);
2284*a325d9c4SApple OSS Distributions 
2285*a325d9c4SApple OSS Distributions 	if (flags & WORKQ_SET_SELF_WQ_KEVENT_UNBIND) {
2286*a325d9c4SApple OSS Distributions 		if (!is_wq_thread) {
2287*a325d9c4SApple OSS Distributions 			unbind_rv = EINVAL;
2288*a325d9c4SApple OSS Distributions 			goto qos;
2289*a325d9c4SApple OSS Distributions 		}
2290*a325d9c4SApple OSS Distributions 
2291*a325d9c4SApple OSS Distributions 		if (uth->uu_workq_pri.qos_bucket == WORKQ_THREAD_QOS_MANAGER) {
2292*a325d9c4SApple OSS Distributions 			unbind_rv = EINVAL;
2293*a325d9c4SApple OSS Distributions 			goto qos;
2294*a325d9c4SApple OSS Distributions 		}
2295*a325d9c4SApple OSS Distributions 
2296*a325d9c4SApple OSS Distributions 		workq_threadreq_t kqr = uth->uu_kqr_bound;
2297*a325d9c4SApple OSS Distributions 		if (kqr == NULL) {
2298*a325d9c4SApple OSS Distributions 			unbind_rv = EALREADY;
2299*a325d9c4SApple OSS Distributions 			goto qos;
2300*a325d9c4SApple OSS Distributions 		}
2301*a325d9c4SApple OSS Distributions 
2302*a325d9c4SApple OSS Distributions 		if (kqr->tr_flags & WORKQ_TR_FLAG_WORKLOOP) {
2303*a325d9c4SApple OSS Distributions 			unbind_rv = EINVAL;
2304*a325d9c4SApple OSS Distributions 			goto qos;
2305*a325d9c4SApple OSS Distributions 		}
2306*a325d9c4SApple OSS Distributions 
2307*a325d9c4SApple OSS Distributions 		kqueue_threadreq_unbind(p, kqr);
2308*a325d9c4SApple OSS Distributions 	}
2309*a325d9c4SApple OSS Distributions 
2310*a325d9c4SApple OSS Distributions qos:
2311*a325d9c4SApple OSS Distributions 	if (flags & WORKQ_SET_SELF_QOS_FLAG) {
2312*a325d9c4SApple OSS Distributions 		thread_qos_policy_data_t new_policy;
2313*a325d9c4SApple OSS Distributions 
2314*a325d9c4SApple OSS Distributions 		if (!_pthread_priority_to_policy(priority, &new_policy)) {
2315*a325d9c4SApple OSS Distributions 			qos_rv = EINVAL;
2316*a325d9c4SApple OSS Distributions 			goto voucher;
2317*a325d9c4SApple OSS Distributions 		}
2318*a325d9c4SApple OSS Distributions 
2319*a325d9c4SApple OSS Distributions 		if (!is_wq_thread) {
2320*a325d9c4SApple OSS Distributions 			/*
2321*a325d9c4SApple OSS Distributions 			 * Threads opted out of QoS can't change QoS
2322*a325d9c4SApple OSS Distributions 			 */
2323*a325d9c4SApple OSS Distributions 			if (!thread_has_qos_policy(th)) {
2324*a325d9c4SApple OSS Distributions 				qos_rv = EPERM;
2325*a325d9c4SApple OSS Distributions 				goto voucher;
2326*a325d9c4SApple OSS Distributions 			}
2327*a325d9c4SApple OSS Distributions 		} else if (uth->uu_workq_pri.qos_bucket == WORKQ_THREAD_QOS_MANAGER ||
2328*a325d9c4SApple OSS Distributions 		    uth->uu_workq_pri.qos_bucket == WORKQ_THREAD_QOS_ABOVEUI) {
2329*a325d9c4SApple OSS Distributions 			/*
2330*a325d9c4SApple OSS Distributions 			 * Workqueue manager threads or threads above UI can't change QoS
2331*a325d9c4SApple OSS Distributions 			 */
2332*a325d9c4SApple OSS Distributions 			qos_rv = EINVAL;
2333*a325d9c4SApple OSS Distributions 			goto voucher;
2334*a325d9c4SApple OSS Distributions 		} else {
2335*a325d9c4SApple OSS Distributions 			/*
2336*a325d9c4SApple OSS Distributions 			 * For workqueue threads, possibly adjust buckets and redrive thread
2337*a325d9c4SApple OSS Distributions 			 * requests.
2338*a325d9c4SApple OSS Distributions 			 *
2339*a325d9c4SApple OSS Distributions 			 * Transitions allowed:
2340*a325d9c4SApple OSS Distributions 			 *
2341*a325d9c4SApple OSS Distributions 			 * overcommit --> non-overcommit
2342*a325d9c4SApple OSS Distributions 			 * overcommit --> overcommit
2343*a325d9c4SApple OSS Distributions 			 * non-overcommit --> non-overcommit
2344*a325d9c4SApple OSS Distributions 			 * non-overcommit --> overcommit (to be deprecated later)
2345*a325d9c4SApple OSS Distributions 			 * cooperative --> cooperative
2346*a325d9c4SApple OSS Distributions 			 *
2347*a325d9c4SApple OSS Distributions 			 * All other transitions aren't allowed so reject them.
2348*a325d9c4SApple OSS Distributions 			 */
2349*a325d9c4SApple OSS Distributions 			if (workq_thread_is_overcommit(uth) && _pthread_priority_is_cooperative(priority)) {
2350*a325d9c4SApple OSS Distributions 				qos_rv = EINVAL;
2351*a325d9c4SApple OSS Distributions 				goto voucher;
2352*a325d9c4SApple OSS Distributions 			} else if (workq_thread_is_cooperative(uth) && !_pthread_priority_is_cooperative(priority)) {
2353*a325d9c4SApple OSS Distributions 				qos_rv = EINVAL;
2354*a325d9c4SApple OSS Distributions 				goto voucher;
2355*a325d9c4SApple OSS Distributions 			} else if (workq_thread_is_nonovercommit(uth) && _pthread_priority_is_cooperative(priority)) {
2356*a325d9c4SApple OSS Distributions 				qos_rv = EINVAL;
2357*a325d9c4SApple OSS Distributions 				goto voucher;
2358*a325d9c4SApple OSS Distributions 			}
2359*a325d9c4SApple OSS Distributions 
2360*a325d9c4SApple OSS Distributions 			struct uu_workq_policy old_pri, new_pri;
2361*a325d9c4SApple OSS Distributions 			bool force_run = false;
2362*a325d9c4SApple OSS Distributions 
2363*a325d9c4SApple OSS Distributions 			workq_lock_spin(wq);
2364*a325d9c4SApple OSS Distributions 
2365*a325d9c4SApple OSS Distributions 			old_pri = new_pri = uth->uu_workq_pri;
2366*a325d9c4SApple OSS Distributions 			new_pri.qos_req = (thread_qos_t)new_policy.qos_tier;
2367*a325d9c4SApple OSS Distributions 
2368*a325d9c4SApple OSS Distributions 			/* Adjust schedule counts for various types of transitions */
2369*a325d9c4SApple OSS Distributions 
2370*a325d9c4SApple OSS Distributions 			/* overcommit -> non-overcommit */
2371*a325d9c4SApple OSS Distributions 			if (workq_thread_is_overcommit(uth) && _pthread_priority_is_nonovercommit(priority)) {
2372*a325d9c4SApple OSS Distributions 				workq_thread_set_type(uth, 0);
2373*a325d9c4SApple OSS Distributions 				wq->wq_constrained_threads_scheduled++;
2374*a325d9c4SApple OSS Distributions 
2375*a325d9c4SApple OSS Distributions 				/* non-overcommit -> overcommit */
2376*a325d9c4SApple OSS Distributions 			} else if (workq_thread_is_nonovercommit(uth) && _pthread_priority_is_overcommit(priority)) {
2377*a325d9c4SApple OSS Distributions 				workq_thread_set_type(uth, UT_WORKQ_OVERCOMMIT);
2378*a325d9c4SApple OSS Distributions 				force_run = (wq->wq_constrained_threads_scheduled-- == wq_max_constrained_threads);
2379*a325d9c4SApple OSS Distributions 
2380*a325d9c4SApple OSS Distributions 				/* cooperative -> cooperative */
2381*a325d9c4SApple OSS Distributions 			} else if (workq_thread_is_cooperative(uth)) {
2382*a325d9c4SApple OSS Distributions 				_wq_cooperative_queue_scheduled_count_dec(wq, old_pri.qos_bucket);
2383*a325d9c4SApple OSS Distributions 				_wq_cooperative_queue_scheduled_count_inc(wq, workq_pri_bucket(new_pri));
2384*a325d9c4SApple OSS Distributions 
2385*a325d9c4SApple OSS Distributions 				/* We're changing schedule counts within cooperative pool, we
2386*a325d9c4SApple OSS Distributions 				 * need to refresh best cooperative QoS logic again */
2387*a325d9c4SApple OSS Distributions 				force_run = _wq_cooperative_queue_refresh_best_req_qos(wq);
2388*a325d9c4SApple OSS Distributions 			}
2389*a325d9c4SApple OSS Distributions 
2390*a325d9c4SApple OSS Distributions 			/* This will also call schedule_creator if needed */
2391*a325d9c4SApple OSS Distributions 			workq_thread_update_bucket(p, wq, uth, old_pri, new_pri, force_run);
2392*a325d9c4SApple OSS Distributions 			workq_unlock(wq);
2393*a325d9c4SApple OSS Distributions 
2394*a325d9c4SApple OSS Distributions 			if (workq_thread_is_overcommit(uth)) {
2395*a325d9c4SApple OSS Distributions 				thread_disarm_workqueue_quantum(th);
2396*a325d9c4SApple OSS Distributions 			} else {
2397*a325d9c4SApple OSS Distributions 				/* If the thread changed QoS buckets, the quantum duration
2398*a325d9c4SApple OSS Distributions 				 * may have changed too */
2399*a325d9c4SApple OSS Distributions 				thread_arm_workqueue_quantum(th);
2400*a325d9c4SApple OSS Distributions 			}
2401*a325d9c4SApple OSS Distributions 		}
2402*a325d9c4SApple OSS Distributions 
2403*a325d9c4SApple OSS Distributions 		kr = thread_policy_set_internal(th, THREAD_QOS_POLICY,
2404*a325d9c4SApple OSS Distributions 		    (thread_policy_t)&new_policy, THREAD_QOS_POLICY_COUNT);
2405*a325d9c4SApple OSS Distributions 		if (kr != KERN_SUCCESS) {
2406*a325d9c4SApple OSS Distributions 			qos_rv = EINVAL;
2407*a325d9c4SApple OSS Distributions 		}
2408*a325d9c4SApple OSS Distributions 	}
2409*a325d9c4SApple OSS Distributions 
2410*a325d9c4SApple OSS Distributions voucher:
2411*a325d9c4SApple OSS Distributions 	if (flags & WORKQ_SET_SELF_VOUCHER_FLAG) {
2412*a325d9c4SApple OSS Distributions 		kr = thread_set_voucher_name(voucher);
2413*a325d9c4SApple OSS Distributions 		if (kr != KERN_SUCCESS) {
2414*a325d9c4SApple OSS Distributions 			voucher_rv = ENOENT;
2415*a325d9c4SApple OSS Distributions 			goto fixedpri;
2416*a325d9c4SApple OSS Distributions 		}
2417*a325d9c4SApple OSS Distributions 	}
2418*a325d9c4SApple OSS Distributions 
2419*a325d9c4SApple OSS Distributions fixedpri:
2420*a325d9c4SApple OSS Distributions 	if (qos_rv) {
2421*a325d9c4SApple OSS Distributions 		goto done;
2422*a325d9c4SApple OSS Distributions 	}
2423*a325d9c4SApple OSS Distributions 	if (flags & WORKQ_SET_SELF_FIXEDPRIORITY_FLAG) {
2424*a325d9c4SApple OSS Distributions 		thread_extended_policy_data_t extpol = {.timeshare = 0};
2425*a325d9c4SApple OSS Distributions 
2426*a325d9c4SApple OSS Distributions 		if (is_wq_thread) {
2427*a325d9c4SApple OSS Distributions 			/* Not allowed on workqueue threads */
2428*a325d9c4SApple OSS Distributions 			fixedpri_rv = ENOTSUP;
2429*a325d9c4SApple OSS Distributions 			goto done;
2430*a325d9c4SApple OSS Distributions 		}
2431*a325d9c4SApple OSS Distributions 
2432*a325d9c4SApple OSS Distributions 		kr = thread_policy_set_internal(th, THREAD_EXTENDED_POLICY,
2433*a325d9c4SApple OSS Distributions 		    (thread_policy_t)&extpol, THREAD_EXTENDED_POLICY_COUNT);
2434*a325d9c4SApple OSS Distributions 		if (kr != KERN_SUCCESS) {
2435*a325d9c4SApple OSS Distributions 			fixedpri_rv = EINVAL;
2436*a325d9c4SApple OSS Distributions 			goto done;
2437*a325d9c4SApple OSS Distributions 		}
2438*a325d9c4SApple OSS Distributions 	} else if (flags & WORKQ_SET_SELF_TIMESHARE_FLAG) {
2439*a325d9c4SApple OSS Distributions 		thread_extended_policy_data_t extpol = {.timeshare = 1};
2440*a325d9c4SApple OSS Distributions 
2441*a325d9c4SApple OSS Distributions 		if (is_wq_thread) {
2442*a325d9c4SApple OSS Distributions 			/* Not allowed on workqueue threads */
2443*a325d9c4SApple OSS Distributions 			fixedpri_rv = ENOTSUP;
2444*a325d9c4SApple OSS Distributions 			goto done;
2445*a325d9c4SApple OSS Distributions 		}
2446*a325d9c4SApple OSS Distributions 
2447*a325d9c4SApple OSS Distributions 		kr = thread_policy_set_internal(th, THREAD_EXTENDED_POLICY,
2448*a325d9c4SApple OSS Distributions 		    (thread_policy_t)&extpol, THREAD_EXTENDED_POLICY_COUNT);
2449*a325d9c4SApple OSS Distributions 		if (kr != KERN_SUCCESS) {
2450*a325d9c4SApple OSS Distributions 			fixedpri_rv = EINVAL;
2451*a325d9c4SApple OSS Distributions 			goto done;
2452*a325d9c4SApple OSS Distributions 		}
2453*a325d9c4SApple OSS Distributions 	}
2454*a325d9c4SApple OSS Distributions 
2455*a325d9c4SApple OSS Distributions done:
2456*a325d9c4SApple OSS Distributions 	if (qos_rv && voucher_rv) {
2457*a325d9c4SApple OSS Distributions 		/* Both failed, give that a unique error. */
2458*a325d9c4SApple OSS Distributions 		return EBADMSG;
2459*a325d9c4SApple OSS Distributions 	}
2460*a325d9c4SApple OSS Distributions 
2461*a325d9c4SApple OSS Distributions 	if (unbind_rv) {
2462*a325d9c4SApple OSS Distributions 		return unbind_rv;
2463*a325d9c4SApple OSS Distributions 	}
2464*a325d9c4SApple OSS Distributions 
2465*a325d9c4SApple OSS Distributions 	if (qos_rv) {
2466*a325d9c4SApple OSS Distributions 		return qos_rv;
2467*a325d9c4SApple OSS Distributions 	}
2468*a325d9c4SApple OSS Distributions 
2469*a325d9c4SApple OSS Distributions 	if (voucher_rv) {
2470*a325d9c4SApple OSS Distributions 		return voucher_rv;
2471*a325d9c4SApple OSS Distributions 	}
2472*a325d9c4SApple OSS Distributions 
2473*a325d9c4SApple OSS Distributions 	if (fixedpri_rv) {
2474*a325d9c4SApple OSS Distributions 		return fixedpri_rv;
2475*a325d9c4SApple OSS Distributions 	}
2476*a325d9c4SApple OSS Distributions 
2477*a325d9c4SApple OSS Distributions 
2478*a325d9c4SApple OSS Distributions 	return 0;
2479*a325d9c4SApple OSS Distributions }
2480*a325d9c4SApple OSS Distributions 
2481*a325d9c4SApple OSS Distributions static int
bsdthread_add_explicit_override(proc_t p,mach_port_name_t kport,pthread_priority_t pp,user_addr_t resource)2482*a325d9c4SApple OSS Distributions bsdthread_add_explicit_override(proc_t p, mach_port_name_t kport,
2483*a325d9c4SApple OSS Distributions     pthread_priority_t pp, user_addr_t resource)
2484*a325d9c4SApple OSS Distributions {
2485*a325d9c4SApple OSS Distributions 	thread_qos_t qos = _pthread_priority_thread_qos(pp);
2486*a325d9c4SApple OSS Distributions 	if (qos == THREAD_QOS_UNSPECIFIED) {
2487*a325d9c4SApple OSS Distributions 		return EINVAL;
2488*a325d9c4SApple OSS Distributions 	}
2489*a325d9c4SApple OSS Distributions 
2490*a325d9c4SApple OSS Distributions 	thread_t th = port_name_to_thread(kport,
2491*a325d9c4SApple OSS Distributions 	    PORT_INTRANS_THREAD_IN_CURRENT_TASK);
2492*a325d9c4SApple OSS Distributions 	if (th == THREAD_NULL) {
2493*a325d9c4SApple OSS Distributions 		return ESRCH;
2494*a325d9c4SApple OSS Distributions 	}
2495*a325d9c4SApple OSS Distributions 
2496*a325d9c4SApple OSS Distributions 	int rv = proc_thread_qos_add_override(p->task, th, 0, qos, TRUE,
2497*a325d9c4SApple OSS Distributions 	    resource, THREAD_QOS_OVERRIDE_TYPE_PTHREAD_EXPLICIT_OVERRIDE);
2498*a325d9c4SApple OSS Distributions 
2499*a325d9c4SApple OSS Distributions 	thread_deallocate(th);
2500*a325d9c4SApple OSS Distributions 	return rv;
2501*a325d9c4SApple OSS Distributions }
2502*a325d9c4SApple OSS Distributions 
2503*a325d9c4SApple OSS Distributions static int
bsdthread_remove_explicit_override(proc_t p,mach_port_name_t kport,user_addr_t resource)2504*a325d9c4SApple OSS Distributions bsdthread_remove_explicit_override(proc_t p, mach_port_name_t kport,
2505*a325d9c4SApple OSS Distributions     user_addr_t resource)
2506*a325d9c4SApple OSS Distributions {
2507*a325d9c4SApple OSS Distributions 	thread_t th = port_name_to_thread(kport,
2508*a325d9c4SApple OSS Distributions 	    PORT_INTRANS_THREAD_IN_CURRENT_TASK);
2509*a325d9c4SApple OSS Distributions 	if (th == THREAD_NULL) {
2510*a325d9c4SApple OSS Distributions 		return ESRCH;
2511*a325d9c4SApple OSS Distributions 	}
2512*a325d9c4SApple OSS Distributions 
2513*a325d9c4SApple OSS Distributions 	int rv = proc_thread_qos_remove_override(p->task, th, 0, resource,
2514*a325d9c4SApple OSS Distributions 	    THREAD_QOS_OVERRIDE_TYPE_PTHREAD_EXPLICIT_OVERRIDE);
2515*a325d9c4SApple OSS Distributions 
2516*a325d9c4SApple OSS Distributions 	thread_deallocate(th);
2517*a325d9c4SApple OSS Distributions 	return rv;
2518*a325d9c4SApple OSS Distributions }
2519*a325d9c4SApple OSS Distributions 
2520*a325d9c4SApple OSS Distributions static int
workq_thread_add_dispatch_override(proc_t p,mach_port_name_t kport,pthread_priority_t pp,user_addr_t ulock_addr)2521*a325d9c4SApple OSS Distributions workq_thread_add_dispatch_override(proc_t p, mach_port_name_t kport,
2522*a325d9c4SApple OSS Distributions     pthread_priority_t pp, user_addr_t ulock_addr)
2523*a325d9c4SApple OSS Distributions {
2524*a325d9c4SApple OSS Distributions 	struct uu_workq_policy old_pri, new_pri;
2525*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
2526*a325d9c4SApple OSS Distributions 
2527*a325d9c4SApple OSS Distributions 	thread_qos_t qos_override = _pthread_priority_thread_qos(pp);
2528*a325d9c4SApple OSS Distributions 	if (qos_override == THREAD_QOS_UNSPECIFIED) {
2529*a325d9c4SApple OSS Distributions 		return EINVAL;
2530*a325d9c4SApple OSS Distributions 	}
2531*a325d9c4SApple OSS Distributions 
2532*a325d9c4SApple OSS Distributions 	thread_t thread = port_name_to_thread(kport,
2533*a325d9c4SApple OSS Distributions 	    PORT_INTRANS_THREAD_IN_CURRENT_TASK);
2534*a325d9c4SApple OSS Distributions 	if (thread == THREAD_NULL) {
2535*a325d9c4SApple OSS Distributions 		return ESRCH;
2536*a325d9c4SApple OSS Distributions 	}
2537*a325d9c4SApple OSS Distributions 
2538*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(thread);
2539*a325d9c4SApple OSS Distributions 	if ((thread_get_tag(thread) & THREAD_TAG_WORKQUEUE) == 0) {
2540*a325d9c4SApple OSS Distributions 		thread_deallocate(thread);
2541*a325d9c4SApple OSS Distributions 		return EPERM;
2542*a325d9c4SApple OSS Distributions 	}
2543*a325d9c4SApple OSS Distributions 
2544*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_override_dispatch | DBG_FUNC_NONE,
2545*a325d9c4SApple OSS Distributions 	    wq, thread_tid(thread), 1, pp);
2546*a325d9c4SApple OSS Distributions 
2547*a325d9c4SApple OSS Distributions 	thread_mtx_lock(thread);
2548*a325d9c4SApple OSS Distributions 
2549*a325d9c4SApple OSS Distributions 	if (ulock_addr) {
2550*a325d9c4SApple OSS Distributions 		uint32_t val;
2551*a325d9c4SApple OSS Distributions 		int rc;
2552*a325d9c4SApple OSS Distributions 		/*
2553*a325d9c4SApple OSS Distributions 		 * Workaround lack of explicit support for 'no-fault copyin'
2554*a325d9c4SApple OSS Distributions 		 * <rdar://problem/24999882>, as disabling preemption prevents paging in
2555*a325d9c4SApple OSS Distributions 		 */
2556*a325d9c4SApple OSS Distributions 		disable_preemption();
2557*a325d9c4SApple OSS Distributions 		rc = copyin_atomic32(ulock_addr, &val);
2558*a325d9c4SApple OSS Distributions 		enable_preemption();
2559*a325d9c4SApple OSS Distributions 		if (rc == 0 && ulock_owner_value_to_port_name(val) != kport) {
2560*a325d9c4SApple OSS Distributions 			goto out;
2561*a325d9c4SApple OSS Distributions 		}
2562*a325d9c4SApple OSS Distributions 	}
2563*a325d9c4SApple OSS Distributions 
2564*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
2565*a325d9c4SApple OSS Distributions 
2566*a325d9c4SApple OSS Distributions 	old_pri = uth->uu_workq_pri;
2567*a325d9c4SApple OSS Distributions 	if (old_pri.qos_override >= qos_override) {
2568*a325d9c4SApple OSS Distributions 		/* Nothing to do */
2569*a325d9c4SApple OSS Distributions 	} else if (thread == current_thread()) {
2570*a325d9c4SApple OSS Distributions 		new_pri = old_pri;
2571*a325d9c4SApple OSS Distributions 		new_pri.qos_override = qos_override;
2572*a325d9c4SApple OSS Distributions 		workq_thread_update_bucket(p, wq, uth, old_pri, new_pri, false);
2573*a325d9c4SApple OSS Distributions 	} else {
2574*a325d9c4SApple OSS Distributions 		uth->uu_workq_pri.qos_override = qos_override;
2575*a325d9c4SApple OSS Distributions 		if (qos_override > workq_pri_override(old_pri)) {
2576*a325d9c4SApple OSS Distributions 			thread_set_workq_override(thread, qos_override);
2577*a325d9c4SApple OSS Distributions 		}
2578*a325d9c4SApple OSS Distributions 	}
2579*a325d9c4SApple OSS Distributions 
2580*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
2581*a325d9c4SApple OSS Distributions 
2582*a325d9c4SApple OSS Distributions out:
2583*a325d9c4SApple OSS Distributions 	thread_mtx_unlock(thread);
2584*a325d9c4SApple OSS Distributions 	thread_deallocate(thread);
2585*a325d9c4SApple OSS Distributions 	return 0;
2586*a325d9c4SApple OSS Distributions }
2587*a325d9c4SApple OSS Distributions 
2588*a325d9c4SApple OSS Distributions static int
workq_thread_reset_dispatch_override(proc_t p,thread_t thread)2589*a325d9c4SApple OSS Distributions workq_thread_reset_dispatch_override(proc_t p, thread_t thread)
2590*a325d9c4SApple OSS Distributions {
2591*a325d9c4SApple OSS Distributions 	struct uu_workq_policy old_pri, new_pri;
2592*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
2593*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(thread);
2594*a325d9c4SApple OSS Distributions 
2595*a325d9c4SApple OSS Distributions 	if ((thread_get_tag(thread) & THREAD_TAG_WORKQUEUE) == 0) {
2596*a325d9c4SApple OSS Distributions 		return EPERM;
2597*a325d9c4SApple OSS Distributions 	}
2598*a325d9c4SApple OSS Distributions 
2599*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_override_reset | DBG_FUNC_NONE, wq, 0, 0, 0);
2600*a325d9c4SApple OSS Distributions 
2601*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
2602*a325d9c4SApple OSS Distributions 	old_pri = new_pri = uth->uu_workq_pri;
2603*a325d9c4SApple OSS Distributions 	new_pri.qos_override = THREAD_QOS_UNSPECIFIED;
2604*a325d9c4SApple OSS Distributions 	workq_thread_update_bucket(p, wq, uth, old_pri, new_pri, false);
2605*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
2606*a325d9c4SApple OSS Distributions 	return 0;
2607*a325d9c4SApple OSS Distributions }
2608*a325d9c4SApple OSS Distributions 
2609*a325d9c4SApple OSS Distributions static int
workq_thread_allow_kill(__unused proc_t p,thread_t thread,bool enable)2610*a325d9c4SApple OSS Distributions workq_thread_allow_kill(__unused proc_t p, thread_t thread, bool enable)
2611*a325d9c4SApple OSS Distributions {
2612*a325d9c4SApple OSS Distributions 	if (!(thread_get_tag(thread) & THREAD_TAG_WORKQUEUE)) {
2613*a325d9c4SApple OSS Distributions 		// If the thread isn't a workqueue thread, don't set the
2614*a325d9c4SApple OSS Distributions 		// kill_allowed bit; however, we still need to return 0
2615*a325d9c4SApple OSS Distributions 		// instead of an error code since this code is executed
2616*a325d9c4SApple OSS Distributions 		// on the abort path which needs to not depend on the
2617*a325d9c4SApple OSS Distributions 		// pthread_t (returning an error depends on pthread_t via
2618*a325d9c4SApple OSS Distributions 		// cerror_nocancel)
2619*a325d9c4SApple OSS Distributions 		return 0;
2620*a325d9c4SApple OSS Distributions 	}
2621*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(thread);
2622*a325d9c4SApple OSS Distributions 	uth->uu_workq_pthread_kill_allowed = enable;
2623*a325d9c4SApple OSS Distributions 	return 0;
2624*a325d9c4SApple OSS Distributions }
2625*a325d9c4SApple OSS Distributions 
2626*a325d9c4SApple OSS Distributions static int
bsdthread_get_max_parallelism(thread_qos_t qos,unsigned long flags,int * retval)2627*a325d9c4SApple OSS Distributions bsdthread_get_max_parallelism(thread_qos_t qos, unsigned long flags,
2628*a325d9c4SApple OSS Distributions     int *retval)
2629*a325d9c4SApple OSS Distributions {
2630*a325d9c4SApple OSS Distributions 	static_assert(QOS_PARALLELISM_COUNT_LOGICAL ==
2631*a325d9c4SApple OSS Distributions 	    _PTHREAD_QOS_PARALLELISM_COUNT_LOGICAL, "logical");
2632*a325d9c4SApple OSS Distributions 	static_assert(QOS_PARALLELISM_REALTIME ==
2633*a325d9c4SApple OSS Distributions 	    _PTHREAD_QOS_PARALLELISM_REALTIME, "realtime");
2634*a325d9c4SApple OSS Distributions 	static_assert(QOS_PARALLELISM_CLUSTER_SHARED_RESOURCE ==
2635*a325d9c4SApple OSS Distributions 	    _PTHREAD_QOS_PARALLELISM_CLUSTER_SHARED_RSRC, "cluster shared resource");
2636*a325d9c4SApple OSS Distributions 
2637*a325d9c4SApple OSS Distributions 	if (flags & ~(QOS_PARALLELISM_REALTIME | QOS_PARALLELISM_COUNT_LOGICAL | QOS_PARALLELISM_CLUSTER_SHARED_RESOURCE)) {
2638*a325d9c4SApple OSS Distributions 		return EINVAL;
2639*a325d9c4SApple OSS Distributions 	}
2640*a325d9c4SApple OSS Distributions 
2641*a325d9c4SApple OSS Distributions 	/* No units are present */
2642*a325d9c4SApple OSS Distributions 	if (flags & QOS_PARALLELISM_CLUSTER_SHARED_RESOURCE) {
2643*a325d9c4SApple OSS Distributions 		return ENOTSUP;
2644*a325d9c4SApple OSS Distributions 	}
2645*a325d9c4SApple OSS Distributions 
2646*a325d9c4SApple OSS Distributions 	if (flags & QOS_PARALLELISM_REALTIME) {
2647*a325d9c4SApple OSS Distributions 		if (qos) {
2648*a325d9c4SApple OSS Distributions 			return EINVAL;
2649*a325d9c4SApple OSS Distributions 		}
2650*a325d9c4SApple OSS Distributions 	} else if (qos == THREAD_QOS_UNSPECIFIED || qos >= THREAD_QOS_LAST) {
2651*a325d9c4SApple OSS Distributions 		return EINVAL;
2652*a325d9c4SApple OSS Distributions 	}
2653*a325d9c4SApple OSS Distributions 
2654*a325d9c4SApple OSS Distributions 	*retval = qos_max_parallelism(qos, flags);
2655*a325d9c4SApple OSS Distributions 	return 0;
2656*a325d9c4SApple OSS Distributions }
2657*a325d9c4SApple OSS Distributions 
2658*a325d9c4SApple OSS Distributions static int
bsdthread_dispatch_apply_attr(__unused struct proc * p,thread_t thread,unsigned long flags,uint64_t value1,__unused uint64_t value2)2659*a325d9c4SApple OSS Distributions bsdthread_dispatch_apply_attr(__unused struct proc *p, thread_t thread,
2660*a325d9c4SApple OSS Distributions     unsigned long flags, uint64_t value1, __unused uint64_t value2)
2661*a325d9c4SApple OSS Distributions {
2662*a325d9c4SApple OSS Distributions 	uint32_t apply_worker_index;
2663*a325d9c4SApple OSS Distributions 	kern_return_t kr;
2664*a325d9c4SApple OSS Distributions 
2665*a325d9c4SApple OSS Distributions 	switch (flags) {
2666*a325d9c4SApple OSS Distributions 	case _PTHREAD_DISPATCH_APPLY_ATTR_CLUSTER_SHARED_RSRC_SET:
2667*a325d9c4SApple OSS Distributions 		apply_worker_index = (uint32_t)value1;
2668*a325d9c4SApple OSS Distributions 		kr = thread_shared_rsrc_policy_set(thread, apply_worker_index, CLUSTER_SHARED_RSRC_TYPE_RR, SHARED_RSRC_POLICY_AGENT_DISPATCH);
2669*a325d9c4SApple OSS Distributions 		/*
2670*a325d9c4SApple OSS Distributions 		 * KERN_INVALID_POLICY indicates that the thread was trying to bind to a
2671*a325d9c4SApple OSS Distributions 		 * cluster which it was not eligible to execute on.
2672*a325d9c4SApple OSS Distributions 		 */
2673*a325d9c4SApple OSS Distributions 		return (kr == KERN_SUCCESS) ? 0 : ((kr == KERN_INVALID_POLICY) ? ENOTSUP : EINVAL);
2674*a325d9c4SApple OSS Distributions 	case _PTHREAD_DISPATCH_APPLY_ATTR_CLUSTER_SHARED_RSRC_CLEAR:
2675*a325d9c4SApple OSS Distributions 		kr = thread_shared_rsrc_policy_clear(thread, CLUSTER_SHARED_RSRC_TYPE_RR, SHARED_RSRC_POLICY_AGENT_DISPATCH);
2676*a325d9c4SApple OSS Distributions 		return (kr == KERN_SUCCESS) ? 0 : EINVAL;
2677*a325d9c4SApple OSS Distributions 	default:
2678*a325d9c4SApple OSS Distributions 		return EINVAL;
2679*a325d9c4SApple OSS Distributions 	}
2680*a325d9c4SApple OSS Distributions }
2681*a325d9c4SApple OSS Distributions 
2682*a325d9c4SApple OSS Distributions #define ENSURE_UNUSED(arg) \
2683*a325d9c4SApple OSS Distributions 	        ({ if ((arg) != 0) { return EINVAL; } })
2684*a325d9c4SApple OSS Distributions 
2685*a325d9c4SApple OSS Distributions int
bsdthread_ctl(struct proc * p,struct bsdthread_ctl_args * uap,int * retval)2686*a325d9c4SApple OSS Distributions bsdthread_ctl(struct proc *p, struct bsdthread_ctl_args *uap, int *retval)
2687*a325d9c4SApple OSS Distributions {
2688*a325d9c4SApple OSS Distributions 	switch (uap->cmd) {
2689*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_QOS_OVERRIDE_START:
2690*a325d9c4SApple OSS Distributions 		return bsdthread_add_explicit_override(p, (mach_port_name_t)uap->arg1,
2691*a325d9c4SApple OSS Distributions 		           (pthread_priority_t)uap->arg2, uap->arg3);
2692*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_QOS_OVERRIDE_END:
2693*a325d9c4SApple OSS Distributions 		ENSURE_UNUSED(uap->arg3);
2694*a325d9c4SApple OSS Distributions 		return bsdthread_remove_explicit_override(p, (mach_port_name_t)uap->arg1,
2695*a325d9c4SApple OSS Distributions 		           (user_addr_t)uap->arg2);
2696*a325d9c4SApple OSS Distributions 
2697*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_QOS_OVERRIDE_DISPATCH:
2698*a325d9c4SApple OSS Distributions 		return workq_thread_add_dispatch_override(p, (mach_port_name_t)uap->arg1,
2699*a325d9c4SApple OSS Distributions 		           (pthread_priority_t)uap->arg2, uap->arg3);
2700*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_QOS_OVERRIDE_RESET:
2701*a325d9c4SApple OSS Distributions 		return workq_thread_reset_dispatch_override(p, current_thread());
2702*a325d9c4SApple OSS Distributions 
2703*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_SET_SELF:
2704*a325d9c4SApple OSS Distributions 		return bsdthread_set_self(p, current_thread(),
2705*a325d9c4SApple OSS Distributions 		           (pthread_priority_t)uap->arg1, (mach_port_name_t)uap->arg2,
2706*a325d9c4SApple OSS Distributions 		           (enum workq_set_self_flags)uap->arg3);
2707*a325d9c4SApple OSS Distributions 
2708*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_QOS_MAX_PARALLELISM:
2709*a325d9c4SApple OSS Distributions 		ENSURE_UNUSED(uap->arg3);
2710*a325d9c4SApple OSS Distributions 		return bsdthread_get_max_parallelism((thread_qos_t)uap->arg1,
2711*a325d9c4SApple OSS Distributions 		           (unsigned long)uap->arg2, retval);
2712*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_WORKQ_ALLOW_KILL:
2713*a325d9c4SApple OSS Distributions 		ENSURE_UNUSED(uap->arg2);
2714*a325d9c4SApple OSS Distributions 		ENSURE_UNUSED(uap->arg3);
2715*a325d9c4SApple OSS Distributions 		return workq_thread_allow_kill(p, current_thread(), (bool)uap->arg1);
2716*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_DISPATCH_APPLY_ATTR:
2717*a325d9c4SApple OSS Distributions 		return bsdthread_dispatch_apply_attr(p, current_thread(),
2718*a325d9c4SApple OSS Distributions 		           (unsigned long)uap->arg1, (uint64_t)uap->arg2,
2719*a325d9c4SApple OSS Distributions 		           (uint64_t)uap->arg3);
2720*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_SET_QOS:
2721*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_QOS_DISPATCH_ASYNCHRONOUS_OVERRIDE_ADD:
2722*a325d9c4SApple OSS Distributions 	case BSDTHREAD_CTL_QOS_DISPATCH_ASYNCHRONOUS_OVERRIDE_RESET:
2723*a325d9c4SApple OSS Distributions 		/* no longer supported */
2724*a325d9c4SApple OSS Distributions 		return ENOTSUP;
2725*a325d9c4SApple OSS Distributions 
2726*a325d9c4SApple OSS Distributions 	default:
2727*a325d9c4SApple OSS Distributions 		return EINVAL;
2728*a325d9c4SApple OSS Distributions 	}
2729*a325d9c4SApple OSS Distributions }
2730*a325d9c4SApple OSS Distributions 
2731*a325d9c4SApple OSS Distributions #pragma mark workqueue thread manipulation
2732*a325d9c4SApple OSS Distributions 
2733*a325d9c4SApple OSS Distributions static void __dead2
2734*a325d9c4SApple OSS Distributions workq_unpark_select_threadreq_or_park_and_unlock(proc_t p, struct workqueue *wq,
2735*a325d9c4SApple OSS Distributions     struct uthread *uth, uint32_t setup_flags);
2736*a325d9c4SApple OSS Distributions 
2737*a325d9c4SApple OSS Distributions static void __dead2
2738*a325d9c4SApple OSS Distributions workq_select_threadreq_or_park_and_unlock(proc_t p, struct workqueue *wq,
2739*a325d9c4SApple OSS Distributions     struct uthread *uth, uint32_t setup_flags);
2740*a325d9c4SApple OSS Distributions 
2741*a325d9c4SApple OSS Distributions static void workq_setup_and_run(proc_t p, struct uthread *uth, int flags) __dead2;
2742*a325d9c4SApple OSS Distributions 
2743*a325d9c4SApple OSS Distributions #if KDEBUG_LEVEL >= KDEBUG_LEVEL_STANDARD
2744*a325d9c4SApple OSS Distributions static inline uint64_t
workq_trace_req_id(workq_threadreq_t req)2745*a325d9c4SApple OSS Distributions workq_trace_req_id(workq_threadreq_t req)
2746*a325d9c4SApple OSS Distributions {
2747*a325d9c4SApple OSS Distributions 	struct kqworkloop *kqwl;
2748*a325d9c4SApple OSS Distributions 	if (req->tr_flags & WORKQ_TR_FLAG_WORKLOOP) {
2749*a325d9c4SApple OSS Distributions 		kqwl = __container_of(req, struct kqworkloop, kqwl_request);
2750*a325d9c4SApple OSS Distributions 		return kqwl->kqwl_dynamicid;
2751*a325d9c4SApple OSS Distributions 	}
2752*a325d9c4SApple OSS Distributions 
2753*a325d9c4SApple OSS Distributions 	return VM_KERNEL_ADDRHIDE(req);
2754*a325d9c4SApple OSS Distributions }
2755*a325d9c4SApple OSS Distributions #endif
2756*a325d9c4SApple OSS Distributions 
2757*a325d9c4SApple OSS Distributions /**
2758*a325d9c4SApple OSS Distributions  * Entry point for libdispatch to ask for threads
2759*a325d9c4SApple OSS Distributions  */
2760*a325d9c4SApple OSS Distributions static int
workq_reqthreads(struct proc * p,uint32_t reqcount,pthread_priority_t pp,bool cooperative)2761*a325d9c4SApple OSS Distributions workq_reqthreads(struct proc *p, uint32_t reqcount, pthread_priority_t pp, bool cooperative)
2762*a325d9c4SApple OSS Distributions {
2763*a325d9c4SApple OSS Distributions 	thread_qos_t qos = _pthread_priority_thread_qos(pp);
2764*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
2765*a325d9c4SApple OSS Distributions 	uint32_t unpaced, upcall_flags = WQ_FLAG_THREAD_NEWSPI;
2766*a325d9c4SApple OSS Distributions 	int ret = 0;
2767*a325d9c4SApple OSS Distributions 
2768*a325d9c4SApple OSS Distributions 	if (wq == NULL || reqcount <= 0 || reqcount > UINT16_MAX ||
2769*a325d9c4SApple OSS Distributions 	    qos == THREAD_QOS_UNSPECIFIED) {
2770*a325d9c4SApple OSS Distributions 		ret = EINVAL;
2771*a325d9c4SApple OSS Distributions 		goto exit;
2772*a325d9c4SApple OSS Distributions 	}
2773*a325d9c4SApple OSS Distributions 
2774*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_wqops_reqthreads | DBG_FUNC_NONE,
2775*a325d9c4SApple OSS Distributions 	    wq, reqcount, pp, cooperative);
2776*a325d9c4SApple OSS Distributions 
2777*a325d9c4SApple OSS Distributions 	workq_threadreq_t req = zalloc(workq_zone_threadreq);
2778*a325d9c4SApple OSS Distributions 	priority_queue_entry_init(&req->tr_entry);
2779*a325d9c4SApple OSS Distributions 	req->tr_state = WORKQ_TR_STATE_NEW;
2780*a325d9c4SApple OSS Distributions 	req->tr_qos   = qos;
2781*a325d9c4SApple OSS Distributions 	workq_tr_flags_t tr_flags = 0;
2782*a325d9c4SApple OSS Distributions 
2783*a325d9c4SApple OSS Distributions 	if (pp & _PTHREAD_PRIORITY_OVERCOMMIT_FLAG) {
2784*a325d9c4SApple OSS Distributions 		tr_flags |= WORKQ_TR_FLAG_OVERCOMMIT;
2785*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_OVERCOMMIT;
2786*a325d9c4SApple OSS Distributions 	}
2787*a325d9c4SApple OSS Distributions 
2788*a325d9c4SApple OSS Distributions 	if (cooperative) {
2789*a325d9c4SApple OSS Distributions 		tr_flags |= WORKQ_TR_FLAG_COOPERATIVE;
2790*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_COOPERATIVE;
2791*a325d9c4SApple OSS Distributions 
2792*a325d9c4SApple OSS Distributions 		if (reqcount > 1) {
2793*a325d9c4SApple OSS Distributions 			ret = ENOTSUP;
2794*a325d9c4SApple OSS Distributions 			goto free_and_exit;
2795*a325d9c4SApple OSS Distributions 		}
2796*a325d9c4SApple OSS Distributions 	}
2797*a325d9c4SApple OSS Distributions 
2798*a325d9c4SApple OSS Distributions 	/* A thread request cannot be both overcommit and cooperative */
2799*a325d9c4SApple OSS Distributions 	if (workq_tr_is_cooperative(tr_flags) &&
2800*a325d9c4SApple OSS Distributions 	    workq_tr_is_overcommit(tr_flags)) {
2801*a325d9c4SApple OSS Distributions 		ret = EINVAL;
2802*a325d9c4SApple OSS Distributions 		goto free_and_exit;
2803*a325d9c4SApple OSS Distributions 	}
2804*a325d9c4SApple OSS Distributions 	req->tr_flags = tr_flags;
2805*a325d9c4SApple OSS Distributions 
2806*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_thread_request_initiate | DBG_FUNC_NONE,
2807*a325d9c4SApple OSS Distributions 	    wq, workq_trace_req_id(req), req->tr_qos, reqcount);
2808*a325d9c4SApple OSS Distributions 
2809*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
2810*a325d9c4SApple OSS Distributions 	do {
2811*a325d9c4SApple OSS Distributions 		if (_wq_exiting(wq)) {
2812*a325d9c4SApple OSS Distributions 			goto unlock_and_exit;
2813*a325d9c4SApple OSS Distributions 		}
2814*a325d9c4SApple OSS Distributions 
2815*a325d9c4SApple OSS Distributions 		/*
2816*a325d9c4SApple OSS Distributions 		 * When userspace is asking for parallelism, wakeup up to (reqcount - 1)
2817*a325d9c4SApple OSS Distributions 		 * threads without pacing, to inform the scheduler of that workload.
2818*a325d9c4SApple OSS Distributions 		 *
2819*a325d9c4SApple OSS Distributions 		 * The last requests, or the ones that failed the admission checks are
2820*a325d9c4SApple OSS Distributions 		 * enqueued and go through the regular creator codepath.
2821*a325d9c4SApple OSS Distributions 		 *
2822*a325d9c4SApple OSS Distributions 		 * If there aren't enough threads, add one, but re-evaluate everything
2823*a325d9c4SApple OSS Distributions 		 * as conditions may now have changed.
2824*a325d9c4SApple OSS Distributions 		 */
2825*a325d9c4SApple OSS Distributions 		unpaced = reqcount - 1;
2826*a325d9c4SApple OSS Distributions 
2827*a325d9c4SApple OSS Distributions 		if (reqcount > 1) {
2828*a325d9c4SApple OSS Distributions 			/* We don't handle asking for parallelism on the cooperative
2829*a325d9c4SApple OSS Distributions 			 * workqueue just yet */
2830*a325d9c4SApple OSS Distributions 			assert(!workq_threadreq_is_cooperative(req));
2831*a325d9c4SApple OSS Distributions 
2832*a325d9c4SApple OSS Distributions 			if (workq_threadreq_is_nonovercommit(req)) {
2833*a325d9c4SApple OSS Distributions 				unpaced = workq_constrained_allowance(wq, qos, NULL, false);
2834*a325d9c4SApple OSS Distributions 				if (unpaced >= reqcount - 1) {
2835*a325d9c4SApple OSS Distributions 					unpaced = reqcount - 1;
2836*a325d9c4SApple OSS Distributions 				}
2837*a325d9c4SApple OSS Distributions 			}
2838*a325d9c4SApple OSS Distributions 		}
2839*a325d9c4SApple OSS Distributions 
2840*a325d9c4SApple OSS Distributions 		/*
2841*a325d9c4SApple OSS Distributions 		 * This path does not currently handle custom workloop parameters
2842*a325d9c4SApple OSS Distributions 		 * when creating threads for parallelism.
2843*a325d9c4SApple OSS Distributions 		 */
2844*a325d9c4SApple OSS Distributions 		assert(!(req->tr_flags & WORKQ_TR_FLAG_WL_PARAMS));
2845*a325d9c4SApple OSS Distributions 
2846*a325d9c4SApple OSS Distributions 		/*
2847*a325d9c4SApple OSS Distributions 		 * This is a trimmed down version of workq_threadreq_bind_and_unlock()
2848*a325d9c4SApple OSS Distributions 		 */
2849*a325d9c4SApple OSS Distributions 		while (unpaced > 0 && wq->wq_thidlecount) {
2850*a325d9c4SApple OSS Distributions 			struct uthread *uth;
2851*a325d9c4SApple OSS Distributions 			bool needs_wakeup;
2852*a325d9c4SApple OSS Distributions 			uint8_t uu_flags = UT_WORKQ_EARLY_BOUND;
2853*a325d9c4SApple OSS Distributions 
2854*a325d9c4SApple OSS Distributions 			if (workq_tr_is_overcommit(req->tr_flags)) {
2855*a325d9c4SApple OSS Distributions 				uu_flags |= UT_WORKQ_OVERCOMMIT;
2856*a325d9c4SApple OSS Distributions 			}
2857*a325d9c4SApple OSS Distributions 
2858*a325d9c4SApple OSS Distributions 			uth = workq_pop_idle_thread(wq, uu_flags, &needs_wakeup);
2859*a325d9c4SApple OSS Distributions 
2860*a325d9c4SApple OSS Distributions 			_wq_thactive_inc(wq, qos);
2861*a325d9c4SApple OSS Distributions 			wq->wq_thscheduled_count[_wq_bucket(qos)]++;
2862*a325d9c4SApple OSS Distributions 			workq_thread_reset_pri(wq, uth, req, /*unpark*/ true);
2863*a325d9c4SApple OSS Distributions 			wq->wq_fulfilled++;
2864*a325d9c4SApple OSS Distributions 
2865*a325d9c4SApple OSS Distributions 			uth->uu_save.uus_workq_park_data.upcall_flags = upcall_flags;
2866*a325d9c4SApple OSS Distributions 			uth->uu_save.uus_workq_park_data.thread_request = req;
2867*a325d9c4SApple OSS Distributions 			if (needs_wakeup) {
2868*a325d9c4SApple OSS Distributions 				workq_thread_wakeup(uth);
2869*a325d9c4SApple OSS Distributions 			}
2870*a325d9c4SApple OSS Distributions 			unpaced--;
2871*a325d9c4SApple OSS Distributions 			reqcount--;
2872*a325d9c4SApple OSS Distributions 		}
2873*a325d9c4SApple OSS Distributions 	} while (unpaced && wq->wq_nthreads < wq_max_threads &&
2874*a325d9c4SApple OSS Distributions 	    workq_add_new_idle_thread(p, wq));
2875*a325d9c4SApple OSS Distributions 
2876*a325d9c4SApple OSS Distributions 	if (_wq_exiting(wq)) {
2877*a325d9c4SApple OSS Distributions 		goto unlock_and_exit;
2878*a325d9c4SApple OSS Distributions 	}
2879*a325d9c4SApple OSS Distributions 
2880*a325d9c4SApple OSS Distributions 	req->tr_count = (uint16_t)reqcount;
2881*a325d9c4SApple OSS Distributions 	if (workq_threadreq_enqueue(wq, req)) {
2882*a325d9c4SApple OSS Distributions 		/* This can drop the workqueue lock, and take it again */
2883*a325d9c4SApple OSS Distributions 		workq_schedule_creator(p, wq, WORKQ_THREADREQ_CAN_CREATE_THREADS);
2884*a325d9c4SApple OSS Distributions 	}
2885*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
2886*a325d9c4SApple OSS Distributions 	return 0;
2887*a325d9c4SApple OSS Distributions 
2888*a325d9c4SApple OSS Distributions unlock_and_exit:
2889*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
2890*a325d9c4SApple OSS Distributions free_and_exit:
2891*a325d9c4SApple OSS Distributions 	zfree(workq_zone_threadreq, req);
2892*a325d9c4SApple OSS Distributions exit:
2893*a325d9c4SApple OSS Distributions 	return ret;
2894*a325d9c4SApple OSS Distributions }
2895*a325d9c4SApple OSS Distributions 
2896*a325d9c4SApple OSS Distributions bool
workq_kern_threadreq_initiate(struct proc * p,workq_threadreq_t req,struct turnstile * workloop_ts,thread_qos_t qos,workq_kern_threadreq_flags_t flags)2897*a325d9c4SApple OSS Distributions workq_kern_threadreq_initiate(struct proc *p, workq_threadreq_t req,
2898*a325d9c4SApple OSS Distributions     struct turnstile *workloop_ts, thread_qos_t qos,
2899*a325d9c4SApple OSS Distributions     workq_kern_threadreq_flags_t flags)
2900*a325d9c4SApple OSS Distributions {
2901*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
2902*a325d9c4SApple OSS Distributions 	struct uthread *uth = NULL;
2903*a325d9c4SApple OSS Distributions 
2904*a325d9c4SApple OSS Distributions 	assert(req->tr_flags & (WORKQ_TR_FLAG_WORKLOOP | WORKQ_TR_FLAG_KEVENT));
2905*a325d9c4SApple OSS Distributions 
2906*a325d9c4SApple OSS Distributions 	if (req->tr_flags & WORKQ_TR_FLAG_WL_OUTSIDE_QOS) {
2907*a325d9c4SApple OSS Distributions 		workq_threadreq_param_t trp = kqueue_threadreq_workloop_param(req);
2908*a325d9c4SApple OSS Distributions 		qos = thread_workq_qos_for_pri(trp.trp_pri);
2909*a325d9c4SApple OSS Distributions 		if (qos == THREAD_QOS_UNSPECIFIED) {
2910*a325d9c4SApple OSS Distributions 			qos = WORKQ_THREAD_QOS_ABOVEUI;
2911*a325d9c4SApple OSS Distributions 		}
2912*a325d9c4SApple OSS Distributions 	}
2913*a325d9c4SApple OSS Distributions 
2914*a325d9c4SApple OSS Distributions 	assert(req->tr_state == WORKQ_TR_STATE_IDLE);
2915*a325d9c4SApple OSS Distributions 	priority_queue_entry_init(&req->tr_entry);
2916*a325d9c4SApple OSS Distributions 	req->tr_count = 1;
2917*a325d9c4SApple OSS Distributions 	req->tr_state = WORKQ_TR_STATE_NEW;
2918*a325d9c4SApple OSS Distributions 	req->tr_qos   = qos;
2919*a325d9c4SApple OSS Distributions 
2920*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_thread_request_initiate | DBG_FUNC_NONE, wq,
2921*a325d9c4SApple OSS Distributions 	    workq_trace_req_id(req), qos, 1);
2922*a325d9c4SApple OSS Distributions 
2923*a325d9c4SApple OSS Distributions 	if (flags & WORKQ_THREADREQ_ATTEMPT_REBIND) {
2924*a325d9c4SApple OSS Distributions 		/*
2925*a325d9c4SApple OSS Distributions 		 * we're called back synchronously from the context of
2926*a325d9c4SApple OSS Distributions 		 * kqueue_threadreq_unbind from within workq_thread_return()
2927*a325d9c4SApple OSS Distributions 		 * we can try to match up this thread with this request !
2928*a325d9c4SApple OSS Distributions 		 */
2929*a325d9c4SApple OSS Distributions 		uth = current_uthread();
2930*a325d9c4SApple OSS Distributions 		assert(uth->uu_kqr_bound == NULL);
2931*a325d9c4SApple OSS Distributions 	}
2932*a325d9c4SApple OSS Distributions 
2933*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
2934*a325d9c4SApple OSS Distributions 	if (_wq_exiting(wq)) {
2935*a325d9c4SApple OSS Distributions 		req->tr_state = WORKQ_TR_STATE_IDLE;
2936*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
2937*a325d9c4SApple OSS Distributions 		return false;
2938*a325d9c4SApple OSS Distributions 	}
2939*a325d9c4SApple OSS Distributions 
2940*a325d9c4SApple OSS Distributions 	if (uth && workq_threadreq_admissible(wq, uth, req)) {
2941*a325d9c4SApple OSS Distributions 		/* This is the case of the rebind - we were about to park and unbind
2942*a325d9c4SApple OSS Distributions 		 * when more events came so keep the binding.
2943*a325d9c4SApple OSS Distributions 		 */
2944*a325d9c4SApple OSS Distributions 		assert(uth != wq->wq_creator);
2945*a325d9c4SApple OSS Distributions 
2946*a325d9c4SApple OSS Distributions 		if (uth->uu_workq_pri.qos_bucket != req->tr_qos) {
2947*a325d9c4SApple OSS Distributions 			_wq_thactive_move(wq, uth->uu_workq_pri.qos_bucket, req->tr_qos);
2948*a325d9c4SApple OSS Distributions 			workq_thread_reset_pri(wq, uth, req, /*unpark*/ false);
2949*a325d9c4SApple OSS Distributions 		}
2950*a325d9c4SApple OSS Distributions 		/*
2951*a325d9c4SApple OSS Distributions 		 * We're called from workq_kern_threadreq_initiate()
2952*a325d9c4SApple OSS Distributions 		 * due to an unbind, with the kq req held.
2953*a325d9c4SApple OSS Distributions 		 */
2954*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thread_logical_run | DBG_FUNC_START, wq,
2955*a325d9c4SApple OSS Distributions 		    workq_trace_req_id(req), req->tr_flags, 0);
2956*a325d9c4SApple OSS Distributions 		wq->wq_fulfilled++;
2957*a325d9c4SApple OSS Distributions 
2958*a325d9c4SApple OSS Distributions 		kqueue_threadreq_bind(p, req, get_machthread(uth), 0);
2959*a325d9c4SApple OSS Distributions 	} else {
2960*a325d9c4SApple OSS Distributions 		if (workloop_ts) {
2961*a325d9c4SApple OSS Distributions 			workq_perform_turnstile_operation_locked(wq, ^{
2962*a325d9c4SApple OSS Distributions 				turnstile_update_inheritor(workloop_ts, wq->wq_turnstile,
2963*a325d9c4SApple OSS Distributions 				TURNSTILE_IMMEDIATE_UPDATE | TURNSTILE_INHERITOR_TURNSTILE);
2964*a325d9c4SApple OSS Distributions 				turnstile_update_inheritor_complete(workloop_ts,
2965*a325d9c4SApple OSS Distributions 				TURNSTILE_INTERLOCK_HELD);
2966*a325d9c4SApple OSS Distributions 			});
2967*a325d9c4SApple OSS Distributions 		}
2968*a325d9c4SApple OSS Distributions 
2969*a325d9c4SApple OSS Distributions 		bool reevaluate_creator_thread_group = false;
2970*a325d9c4SApple OSS Distributions #if CONFIG_PREADOPT_TG
2971*a325d9c4SApple OSS Distributions 		reevaluate_creator_thread_group = (flags & WORKQ_THREADREQ_REEVALUATE_PREADOPT_TG);
2972*a325d9c4SApple OSS Distributions #endif
2973*a325d9c4SApple OSS Distributions 		/* We enqueued the highest priority item or we may need to reevaluate if
2974*a325d9c4SApple OSS Distributions 		 * the creator needs a thread group pre-adoption */
2975*a325d9c4SApple OSS Distributions 		if (workq_threadreq_enqueue(wq, req) || reevaluate_creator_thread_group) {
2976*a325d9c4SApple OSS Distributions 			workq_schedule_creator(p, wq, flags);
2977*a325d9c4SApple OSS Distributions 		}
2978*a325d9c4SApple OSS Distributions 	}
2979*a325d9c4SApple OSS Distributions 
2980*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
2981*a325d9c4SApple OSS Distributions 
2982*a325d9c4SApple OSS Distributions 	return true;
2983*a325d9c4SApple OSS Distributions }
2984*a325d9c4SApple OSS Distributions 
2985*a325d9c4SApple OSS Distributions void
workq_kern_threadreq_modify(struct proc * p,workq_threadreq_t req,thread_qos_t qos,workq_kern_threadreq_flags_t flags)2986*a325d9c4SApple OSS Distributions workq_kern_threadreq_modify(struct proc *p, workq_threadreq_t req,
2987*a325d9c4SApple OSS Distributions     thread_qos_t qos, workq_kern_threadreq_flags_t flags)
2988*a325d9c4SApple OSS Distributions {
2989*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
2990*a325d9c4SApple OSS Distributions 	bool make_overcommit = false;
2991*a325d9c4SApple OSS Distributions 
2992*a325d9c4SApple OSS Distributions 	if (req->tr_flags & WORKQ_TR_FLAG_WL_OUTSIDE_QOS) {
2993*a325d9c4SApple OSS Distributions 		/* Requests outside-of-QoS shouldn't accept modify operations */
2994*a325d9c4SApple OSS Distributions 		return;
2995*a325d9c4SApple OSS Distributions 	}
2996*a325d9c4SApple OSS Distributions 
2997*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
2998*a325d9c4SApple OSS Distributions 
2999*a325d9c4SApple OSS Distributions 	assert(req->tr_qos != WORKQ_THREAD_QOS_MANAGER);
3000*a325d9c4SApple OSS Distributions 	assert(req->tr_flags & (WORKQ_TR_FLAG_KEVENT | WORKQ_TR_FLAG_WORKLOOP));
3001*a325d9c4SApple OSS Distributions 
3002*a325d9c4SApple OSS Distributions 	if (req->tr_state == WORKQ_TR_STATE_BINDING) {
3003*a325d9c4SApple OSS Distributions 		kqueue_threadreq_bind(p, req, req->tr_thread, 0);
3004*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
3005*a325d9c4SApple OSS Distributions 		return;
3006*a325d9c4SApple OSS Distributions 	}
3007*a325d9c4SApple OSS Distributions 
3008*a325d9c4SApple OSS Distributions 	if (flags & WORKQ_THREADREQ_MAKE_OVERCOMMIT) {
3009*a325d9c4SApple OSS Distributions 		/* TODO (rokhinip): We come into this code path for kqwl thread
3010*a325d9c4SApple OSS Distributions 		 * requests. kqwl requests cannot be cooperative.
3011*a325d9c4SApple OSS Distributions 		 */
3012*a325d9c4SApple OSS Distributions 		assert(!workq_threadreq_is_cooperative(req));
3013*a325d9c4SApple OSS Distributions 
3014*a325d9c4SApple OSS Distributions 		make_overcommit = workq_threadreq_is_nonovercommit(req);
3015*a325d9c4SApple OSS Distributions 	}
3016*a325d9c4SApple OSS Distributions 
3017*a325d9c4SApple OSS Distributions 	if (_wq_exiting(wq) || (req->tr_qos == qos && !make_overcommit)) {
3018*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
3019*a325d9c4SApple OSS Distributions 		return;
3020*a325d9c4SApple OSS Distributions 	}
3021*a325d9c4SApple OSS Distributions 
3022*a325d9c4SApple OSS Distributions 	assert(req->tr_count == 1);
3023*a325d9c4SApple OSS Distributions 	if (req->tr_state != WORKQ_TR_STATE_QUEUED) {
3024*a325d9c4SApple OSS Distributions 		panic("Invalid thread request (%p) state %d", req, req->tr_state);
3025*a325d9c4SApple OSS Distributions 	}
3026*a325d9c4SApple OSS Distributions 
3027*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_thread_request_modify | DBG_FUNC_NONE, wq,
3028*a325d9c4SApple OSS Distributions 	    workq_trace_req_id(req), qos, 0);
3029*a325d9c4SApple OSS Distributions 
3030*a325d9c4SApple OSS Distributions 	struct priority_queue_sched_max *pq = workq_priority_queue_for_req(wq, req);
3031*a325d9c4SApple OSS Distributions 	workq_threadreq_t req_max;
3032*a325d9c4SApple OSS Distributions 
3033*a325d9c4SApple OSS Distributions 	/*
3034*a325d9c4SApple OSS Distributions 	 * Stage 1: Dequeue the request from its priority queue.
3035*a325d9c4SApple OSS Distributions 	 *
3036*a325d9c4SApple OSS Distributions 	 * If we dequeue the root item of the constrained priority queue,
3037*a325d9c4SApple OSS Distributions 	 * maintain the best constrained request qos invariant.
3038*a325d9c4SApple OSS Distributions 	 */
3039*a325d9c4SApple OSS Distributions 	if (priority_queue_remove(pq, &req->tr_entry)) {
3040*a325d9c4SApple OSS Distributions 		if (workq_threadreq_is_nonovercommit(req)) {
3041*a325d9c4SApple OSS Distributions 			_wq_thactive_refresh_best_constrained_req_qos(wq);
3042*a325d9c4SApple OSS Distributions 		}
3043*a325d9c4SApple OSS Distributions 	}
3044*a325d9c4SApple OSS Distributions 
3045*a325d9c4SApple OSS Distributions 	/*
3046*a325d9c4SApple OSS Distributions 	 * Stage 2: Apply changes to the thread request
3047*a325d9c4SApple OSS Distributions 	 *
3048*a325d9c4SApple OSS Distributions 	 * If the item will not become the root of the priority queue it belongs to,
3049*a325d9c4SApple OSS Distributions 	 * then we need to wait in line, just enqueue and return quickly.
3050*a325d9c4SApple OSS Distributions 	 */
3051*a325d9c4SApple OSS Distributions 	if (__improbable(make_overcommit)) {
3052*a325d9c4SApple OSS Distributions 		req->tr_flags ^= WORKQ_TR_FLAG_OVERCOMMIT;
3053*a325d9c4SApple OSS Distributions 		pq = workq_priority_queue_for_req(wq, req);
3054*a325d9c4SApple OSS Distributions 	}
3055*a325d9c4SApple OSS Distributions 	req->tr_qos = qos;
3056*a325d9c4SApple OSS Distributions 
3057*a325d9c4SApple OSS Distributions 	req_max = priority_queue_max(pq, struct workq_threadreq_s, tr_entry);
3058*a325d9c4SApple OSS Distributions 	if (req_max && req_max->tr_qos >= qos) {
3059*a325d9c4SApple OSS Distributions 		priority_queue_entry_set_sched_pri(pq, &req->tr_entry,
3060*a325d9c4SApple OSS Distributions 		    workq_priority_for_req(req), false);
3061*a325d9c4SApple OSS Distributions 		priority_queue_insert(pq, &req->tr_entry);
3062*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
3063*a325d9c4SApple OSS Distributions 		return;
3064*a325d9c4SApple OSS Distributions 	}
3065*a325d9c4SApple OSS Distributions 
3066*a325d9c4SApple OSS Distributions 	/*
3067*a325d9c4SApple OSS Distributions 	 * Stage 3: Reevaluate whether we should run the thread request.
3068*a325d9c4SApple OSS Distributions 	 *
3069*a325d9c4SApple OSS Distributions 	 * Pretend the thread request is new again:
3070*a325d9c4SApple OSS Distributions 	 * - adjust wq_reqcount to not count it anymore.
3071*a325d9c4SApple OSS Distributions 	 * - make its state WORKQ_TR_STATE_NEW (so that workq_threadreq_bind_and_unlock
3072*a325d9c4SApple OSS Distributions 	 *   properly attempts a synchronous bind)
3073*a325d9c4SApple OSS Distributions 	 */
3074*a325d9c4SApple OSS Distributions 	wq->wq_reqcount--;
3075*a325d9c4SApple OSS Distributions 	req->tr_state = WORKQ_TR_STATE_NEW;
3076*a325d9c4SApple OSS Distributions 
3077*a325d9c4SApple OSS Distributions 	/* We enqueued the highest priority item or we may need to reevaluate if
3078*a325d9c4SApple OSS Distributions 	 * the creator needs a thread group pre-adoption if the request got a new TG */
3079*a325d9c4SApple OSS Distributions 	bool reevaluate_creator_tg = false;
3080*a325d9c4SApple OSS Distributions 
3081*a325d9c4SApple OSS Distributions #if CONFIG_PREADOPT_TG
3082*a325d9c4SApple OSS Distributions 	reevaluate_creator_tg = (flags & WORKQ_THREADREQ_REEVALUATE_PREADOPT_TG);
3083*a325d9c4SApple OSS Distributions #endif
3084*a325d9c4SApple OSS Distributions 
3085*a325d9c4SApple OSS Distributions 	if (workq_threadreq_enqueue(wq, req) || reevaluate_creator_tg) {
3086*a325d9c4SApple OSS Distributions 		workq_schedule_creator(p, wq, flags);
3087*a325d9c4SApple OSS Distributions 	}
3088*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
3089*a325d9c4SApple OSS Distributions }
3090*a325d9c4SApple OSS Distributions 
3091*a325d9c4SApple OSS Distributions void
workq_kern_threadreq_lock(struct proc * p)3092*a325d9c4SApple OSS Distributions workq_kern_threadreq_lock(struct proc *p)
3093*a325d9c4SApple OSS Distributions {
3094*a325d9c4SApple OSS Distributions 	workq_lock_spin(proc_get_wqptr_fast(p));
3095*a325d9c4SApple OSS Distributions }
3096*a325d9c4SApple OSS Distributions 
3097*a325d9c4SApple OSS Distributions void
workq_kern_threadreq_unlock(struct proc * p)3098*a325d9c4SApple OSS Distributions workq_kern_threadreq_unlock(struct proc *p)
3099*a325d9c4SApple OSS Distributions {
3100*a325d9c4SApple OSS Distributions 	workq_unlock(proc_get_wqptr_fast(p));
3101*a325d9c4SApple OSS Distributions }
3102*a325d9c4SApple OSS Distributions 
3103*a325d9c4SApple OSS Distributions void
workq_kern_threadreq_update_inheritor(struct proc * p,workq_threadreq_t req,thread_t owner,struct turnstile * wl_ts,turnstile_update_flags_t flags)3104*a325d9c4SApple OSS Distributions workq_kern_threadreq_update_inheritor(struct proc *p, workq_threadreq_t req,
3105*a325d9c4SApple OSS Distributions     thread_t owner, struct turnstile *wl_ts,
3106*a325d9c4SApple OSS Distributions     turnstile_update_flags_t flags)
3107*a325d9c4SApple OSS Distributions {
3108*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
3109*a325d9c4SApple OSS Distributions 	turnstile_inheritor_t inheritor;
3110*a325d9c4SApple OSS Distributions 
3111*a325d9c4SApple OSS Distributions 	assert(req->tr_qos != WORKQ_THREAD_QOS_MANAGER);
3112*a325d9c4SApple OSS Distributions 	assert(req->tr_flags & WORKQ_TR_FLAG_WORKLOOP);
3113*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
3114*a325d9c4SApple OSS Distributions 
3115*a325d9c4SApple OSS Distributions 	if (req->tr_state == WORKQ_TR_STATE_BINDING) {
3116*a325d9c4SApple OSS Distributions 		kqueue_threadreq_bind(p, req, req->tr_thread,
3117*a325d9c4SApple OSS Distributions 		    KQUEUE_THREADERQ_BIND_NO_INHERITOR_UPDATE);
3118*a325d9c4SApple OSS Distributions 		return;
3119*a325d9c4SApple OSS Distributions 	}
3120*a325d9c4SApple OSS Distributions 
3121*a325d9c4SApple OSS Distributions 	if (_wq_exiting(wq)) {
3122*a325d9c4SApple OSS Distributions 		inheritor = TURNSTILE_INHERITOR_NULL;
3123*a325d9c4SApple OSS Distributions 	} else {
3124*a325d9c4SApple OSS Distributions 		if (req->tr_state != WORKQ_TR_STATE_QUEUED) {
3125*a325d9c4SApple OSS Distributions 			panic("Invalid thread request (%p) state %d", req, req->tr_state);
3126*a325d9c4SApple OSS Distributions 		}
3127*a325d9c4SApple OSS Distributions 
3128*a325d9c4SApple OSS Distributions 		if (owner) {
3129*a325d9c4SApple OSS Distributions 			inheritor = owner;
3130*a325d9c4SApple OSS Distributions 			flags |= TURNSTILE_INHERITOR_THREAD;
3131*a325d9c4SApple OSS Distributions 		} else {
3132*a325d9c4SApple OSS Distributions 			inheritor = wq->wq_turnstile;
3133*a325d9c4SApple OSS Distributions 			flags |= TURNSTILE_INHERITOR_TURNSTILE;
3134*a325d9c4SApple OSS Distributions 		}
3135*a325d9c4SApple OSS Distributions 	}
3136*a325d9c4SApple OSS Distributions 
3137*a325d9c4SApple OSS Distributions 	workq_perform_turnstile_operation_locked(wq, ^{
3138*a325d9c4SApple OSS Distributions 		turnstile_update_inheritor(wl_ts, inheritor, flags);
3139*a325d9c4SApple OSS Distributions 	});
3140*a325d9c4SApple OSS Distributions }
3141*a325d9c4SApple OSS Distributions 
3142*a325d9c4SApple OSS Distributions void
workq_kern_threadreq_redrive(struct proc * p,workq_kern_threadreq_flags_t flags)3143*a325d9c4SApple OSS Distributions workq_kern_threadreq_redrive(struct proc *p, workq_kern_threadreq_flags_t flags)
3144*a325d9c4SApple OSS Distributions {
3145*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
3146*a325d9c4SApple OSS Distributions 
3147*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
3148*a325d9c4SApple OSS Distributions 	workq_schedule_creator(p, wq, flags);
3149*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
3150*a325d9c4SApple OSS Distributions }
3151*a325d9c4SApple OSS Distributions 
3152*a325d9c4SApple OSS Distributions /*
3153*a325d9c4SApple OSS Distributions  * Always called at AST by the thread on itself
3154*a325d9c4SApple OSS Distributions  *
3155*a325d9c4SApple OSS Distributions  * Upon quantum expiry, the workqueue subsystem evaluates its state and decides
3156*a325d9c4SApple OSS Distributions  * on what the thread should do next. The TSD value is always set by the thread
3157*a325d9c4SApple OSS Distributions  * on itself in the kernel and cleared either by userspace when it acks the TSD
3158*a325d9c4SApple OSS Distributions  * value and takes action, or by the thread in the kernel when the quantum
3159*a325d9c4SApple OSS Distributions  * expires again.
3160*a325d9c4SApple OSS Distributions  */
3161*a325d9c4SApple OSS Distributions void
workq_kern_quantum_expiry_reevaluate(proc_t proc,thread_t thread)3162*a325d9c4SApple OSS Distributions workq_kern_quantum_expiry_reevaluate(proc_t proc, thread_t thread)
3163*a325d9c4SApple OSS Distributions {
3164*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(thread);
3165*a325d9c4SApple OSS Distributions 
3166*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_DYING) {
3167*a325d9c4SApple OSS Distributions 		return;
3168*a325d9c4SApple OSS Distributions 	}
3169*a325d9c4SApple OSS Distributions 
3170*a325d9c4SApple OSS Distributions 	if (!thread_supports_cooperative_workqueue(thread)) {
3171*a325d9c4SApple OSS Distributions 		panic("Quantum expired for thread that doesn't support cooperative workqueue");
3172*a325d9c4SApple OSS Distributions 	}
3173*a325d9c4SApple OSS Distributions 
3174*a325d9c4SApple OSS Distributions 	thread_qos_t qos = uth->uu_workq_pri.qos_bucket;
3175*a325d9c4SApple OSS Distributions 	if (qos == THREAD_QOS_UNSPECIFIED) {
3176*a325d9c4SApple OSS Distributions 		panic("Thread should not have workq bucket of QoS UN");
3177*a325d9c4SApple OSS Distributions 	}
3178*a325d9c4SApple OSS Distributions 
3179*a325d9c4SApple OSS Distributions 	assert(thread_has_expired_workqueue_quantum(thread, false));
3180*a325d9c4SApple OSS Distributions 
3181*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(proc);
3182*a325d9c4SApple OSS Distributions 	assert(wq != NULL);
3183*a325d9c4SApple OSS Distributions 
3184*a325d9c4SApple OSS Distributions 	/*
3185*a325d9c4SApple OSS Distributions 	 * For starters, we're just going to evaluate and see if we need to narrow
3186*a325d9c4SApple OSS Distributions 	 * the pool and tell this thread to park if needed. In the future, we'll
3187*a325d9c4SApple OSS Distributions 	 * evaluate and convey other workqueue state information like needing to
3188*a325d9c4SApple OSS Distributions 	 * pump kevents, etc.
3189*a325d9c4SApple OSS Distributions 	 */
3190*a325d9c4SApple OSS Distributions 	uint64_t flags = 0;
3191*a325d9c4SApple OSS Distributions 
3192*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
3193*a325d9c4SApple OSS Distributions 
3194*a325d9c4SApple OSS Distributions 	if (workq_thread_is_cooperative(uth)) {
3195*a325d9c4SApple OSS Distributions 		if (!workq_cooperative_allowance(wq, qos, uth, false)) {
3196*a325d9c4SApple OSS Distributions 			flags |= PTHREAD_WQ_QUANTUM_EXPIRY_NARROW;
3197*a325d9c4SApple OSS Distributions 		} else {
3198*a325d9c4SApple OSS Distributions 			/* In the future, when we have kevent hookups for the cooperative
3199*a325d9c4SApple OSS Distributions 			 * pool, we need fancier logic for what userspace should do. But
3200*a325d9c4SApple OSS Distributions 			 * right now, only userspace thread requests exist - so we'll just
3201*a325d9c4SApple OSS Distributions 			 * tell userspace to shuffle work items */
3202*a325d9c4SApple OSS Distributions 			flags |= PTHREAD_WQ_QUANTUM_EXPIRY_SHUFFLE;
3203*a325d9c4SApple OSS Distributions 		}
3204*a325d9c4SApple OSS Distributions 	} else if (workq_thread_is_nonovercommit(uth)) {
3205*a325d9c4SApple OSS Distributions 		if (!workq_constrained_allowance(wq, qos, uth, false)) {
3206*a325d9c4SApple OSS Distributions 			flags |= PTHREAD_WQ_QUANTUM_EXPIRY_NARROW;
3207*a325d9c4SApple OSS Distributions 		}
3208*a325d9c4SApple OSS Distributions 	}
3209*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
3210*a325d9c4SApple OSS Distributions 
3211*a325d9c4SApple OSS Distributions 	WQ_TRACE(TRACE_wq_quantum_expiry_reevaluate, flags, 0, 0, 0);
3212*a325d9c4SApple OSS Distributions 
3213*a325d9c4SApple OSS Distributions 	kevent_set_workq_quantum_expiry_user_tsd(proc, thread, flags);
3214*a325d9c4SApple OSS Distributions 
3215*a325d9c4SApple OSS Distributions 	/* We have conveyed to userspace about what it needs to do upon quantum
3216*a325d9c4SApple OSS Distributions 	 * expiry, now rearm the workqueue quantum again */
3217*a325d9c4SApple OSS Distributions 	thread_arm_workqueue_quantum(get_machthread(uth));
3218*a325d9c4SApple OSS Distributions }
3219*a325d9c4SApple OSS Distributions 
3220*a325d9c4SApple OSS Distributions void
workq_schedule_creator_turnstile_redrive(struct workqueue * wq,bool locked)3221*a325d9c4SApple OSS Distributions workq_schedule_creator_turnstile_redrive(struct workqueue *wq, bool locked)
3222*a325d9c4SApple OSS Distributions {
3223*a325d9c4SApple OSS Distributions 	if (locked) {
3224*a325d9c4SApple OSS Distributions 		workq_schedule_creator(NULL, wq, WORKQ_THREADREQ_NONE);
3225*a325d9c4SApple OSS Distributions 	} else {
3226*a325d9c4SApple OSS Distributions 		workq_schedule_immediate_thread_creation(wq);
3227*a325d9c4SApple OSS Distributions 	}
3228*a325d9c4SApple OSS Distributions }
3229*a325d9c4SApple OSS Distributions 
3230*a325d9c4SApple OSS Distributions static int
workq_thread_return(struct proc * p,struct workq_kernreturn_args * uap,struct workqueue * wq)3231*a325d9c4SApple OSS Distributions workq_thread_return(struct proc *p, struct workq_kernreturn_args *uap,
3232*a325d9c4SApple OSS Distributions     struct workqueue *wq)
3233*a325d9c4SApple OSS Distributions {
3234*a325d9c4SApple OSS Distributions 	thread_t th = current_thread();
3235*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(th);
3236*a325d9c4SApple OSS Distributions 	workq_threadreq_t kqr = uth->uu_kqr_bound;
3237*a325d9c4SApple OSS Distributions 	workq_threadreq_param_t trp = { };
3238*a325d9c4SApple OSS Distributions 	int nevents = uap->affinity, error;
3239*a325d9c4SApple OSS Distributions 	user_addr_t eventlist = uap->item;
3240*a325d9c4SApple OSS Distributions 
3241*a325d9c4SApple OSS Distributions 	if (((thread_get_tag(th) & THREAD_TAG_WORKQUEUE) == 0) ||
3242*a325d9c4SApple OSS Distributions 	    (uth->uu_workq_flags & UT_WORKQ_DYING)) {
3243*a325d9c4SApple OSS Distributions 		return EINVAL;
3244*a325d9c4SApple OSS Distributions 	}
3245*a325d9c4SApple OSS Distributions 
3246*a325d9c4SApple OSS Distributions 	if (eventlist && nevents && kqr == NULL) {
3247*a325d9c4SApple OSS Distributions 		return EINVAL;
3248*a325d9c4SApple OSS Distributions 	}
3249*a325d9c4SApple OSS Distributions 
3250*a325d9c4SApple OSS Distributions 	/* reset signal mask on the workqueue thread to default state */
3251*a325d9c4SApple OSS Distributions 	if (uth->uu_sigmask != (sigset_t)(~workq_threadmask)) {
3252*a325d9c4SApple OSS Distributions 		proc_lock(p);
3253*a325d9c4SApple OSS Distributions 		uth->uu_sigmask = ~workq_threadmask;
3254*a325d9c4SApple OSS Distributions 		proc_unlock(p);
3255*a325d9c4SApple OSS Distributions 	}
3256*a325d9c4SApple OSS Distributions 
3257*a325d9c4SApple OSS Distributions 	if (kqr && kqr->tr_flags & WORKQ_TR_FLAG_WL_PARAMS) {
3258*a325d9c4SApple OSS Distributions 		/*
3259*a325d9c4SApple OSS Distributions 		 * Ensure we store the threadreq param before unbinding
3260*a325d9c4SApple OSS Distributions 		 * the kqr from this thread.
3261*a325d9c4SApple OSS Distributions 		 */
3262*a325d9c4SApple OSS Distributions 		trp = kqueue_threadreq_workloop_param(kqr);
3263*a325d9c4SApple OSS Distributions 	}
3264*a325d9c4SApple OSS Distributions 
3265*a325d9c4SApple OSS Distributions 	/*
3266*a325d9c4SApple OSS Distributions 	 * Freeze the base pri while we decide the fate of this thread.
3267*a325d9c4SApple OSS Distributions 	 *
3268*a325d9c4SApple OSS Distributions 	 * Either:
3269*a325d9c4SApple OSS Distributions 	 * - we return to user and kevent_cleanup will have unfrozen the base pri,
3270*a325d9c4SApple OSS Distributions 	 * - or we proceed to workq_select_threadreq_or_park_and_unlock() who will.
3271*a325d9c4SApple OSS Distributions 	 */
3272*a325d9c4SApple OSS Distributions 	thread_freeze_base_pri(th);
3273*a325d9c4SApple OSS Distributions 
3274*a325d9c4SApple OSS Distributions 	if (kqr) {
3275*a325d9c4SApple OSS Distributions 		uint32_t upcall_flags = WQ_FLAG_THREAD_NEWSPI | WQ_FLAG_THREAD_REUSE;
3276*a325d9c4SApple OSS Distributions 		if (kqr->tr_flags & WORKQ_TR_FLAG_WORKLOOP) {
3277*a325d9c4SApple OSS Distributions 			upcall_flags |= WQ_FLAG_THREAD_WORKLOOP | WQ_FLAG_THREAD_KEVENT;
3278*a325d9c4SApple OSS Distributions 		} else {
3279*a325d9c4SApple OSS Distributions 			upcall_flags |= WQ_FLAG_THREAD_KEVENT;
3280*a325d9c4SApple OSS Distributions 		}
3281*a325d9c4SApple OSS Distributions 		if (uth->uu_workq_pri.qos_bucket == WORKQ_THREAD_QOS_MANAGER) {
3282*a325d9c4SApple OSS Distributions 			upcall_flags |= WQ_FLAG_THREAD_EVENT_MANAGER;
3283*a325d9c4SApple OSS Distributions 		} else {
3284*a325d9c4SApple OSS Distributions 			if (workq_thread_is_overcommit(uth)) {
3285*a325d9c4SApple OSS Distributions 				upcall_flags |= WQ_FLAG_THREAD_OVERCOMMIT;
3286*a325d9c4SApple OSS Distributions 			}
3287*a325d9c4SApple OSS Distributions 			if (uth->uu_workq_flags & UT_WORKQ_OUTSIDE_QOS) {
3288*a325d9c4SApple OSS Distributions 				upcall_flags |= WQ_FLAG_THREAD_OUTSIDEQOS;
3289*a325d9c4SApple OSS Distributions 			} else {
3290*a325d9c4SApple OSS Distributions 				upcall_flags |= uth->uu_workq_pri.qos_req |
3291*a325d9c4SApple OSS Distributions 				    WQ_FLAG_THREAD_PRIO_QOS;
3292*a325d9c4SApple OSS Distributions 			}
3293*a325d9c4SApple OSS Distributions 		}
3294*a325d9c4SApple OSS Distributions 		error = pthread_functions->workq_handle_stack_events(p, th,
3295*a325d9c4SApple OSS Distributions 		    get_task_map(p->task), uth->uu_workq_stackaddr,
3296*a325d9c4SApple OSS Distributions 		    uth->uu_workq_thport, eventlist, nevents, upcall_flags);
3297*a325d9c4SApple OSS Distributions 		if (error) {
3298*a325d9c4SApple OSS Distributions 			assert(uth->uu_kqr_bound == kqr);
3299*a325d9c4SApple OSS Distributions 			return error;
3300*a325d9c4SApple OSS Distributions 		}
3301*a325d9c4SApple OSS Distributions 
3302*a325d9c4SApple OSS Distributions 		// pthread is supposed to pass KEVENT_FLAG_PARKING here
3303*a325d9c4SApple OSS Distributions 		// which should cause the above call to either:
3304*a325d9c4SApple OSS Distributions 		// - not return
3305*a325d9c4SApple OSS Distributions 		// - return an error
3306*a325d9c4SApple OSS Distributions 		// - return 0 and have unbound properly
3307*a325d9c4SApple OSS Distributions 		assert(uth->uu_kqr_bound == NULL);
3308*a325d9c4SApple OSS Distributions 	}
3309*a325d9c4SApple OSS Distributions 
3310*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_runthread | DBG_FUNC_END, wq, uap->options, 0, 0);
3311*a325d9c4SApple OSS Distributions 
3312*a325d9c4SApple OSS Distributions 	thread_sched_call(th, NULL);
3313*a325d9c4SApple OSS Distributions 	thread_will_park_or_terminate(th);
3314*a325d9c4SApple OSS Distributions #if CONFIG_WORKLOOP_DEBUG
3315*a325d9c4SApple OSS Distributions 	UU_KEVENT_HISTORY_WRITE_ENTRY(uth, { .uu_error = -1, });
3316*a325d9c4SApple OSS Distributions #endif
3317*a325d9c4SApple OSS Distributions 
3318*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
3319*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_thread_logical_run | DBG_FUNC_END, wq, 0, 0, 0);
3320*a325d9c4SApple OSS Distributions 	uth->uu_save.uus_workq_park_data.workloop_params = trp.trp_value;
3321*a325d9c4SApple OSS Distributions 	workq_select_threadreq_or_park_and_unlock(p, wq, uth,
3322*a325d9c4SApple OSS Distributions 	    WQ_SETUP_CLEAR_VOUCHER);
3323*a325d9c4SApple OSS Distributions 	__builtin_unreachable();
3324*a325d9c4SApple OSS Distributions }
3325*a325d9c4SApple OSS Distributions 
3326*a325d9c4SApple OSS Distributions /**
3327*a325d9c4SApple OSS Distributions  * Multiplexed call to interact with the workqueue mechanism
3328*a325d9c4SApple OSS Distributions  */
3329*a325d9c4SApple OSS Distributions int
workq_kernreturn(struct proc * p,struct workq_kernreturn_args * uap,int32_t * retval)3330*a325d9c4SApple OSS Distributions workq_kernreturn(struct proc *p, struct workq_kernreturn_args *uap, int32_t *retval)
3331*a325d9c4SApple OSS Distributions {
3332*a325d9c4SApple OSS Distributions 	int options = uap->options;
3333*a325d9c4SApple OSS Distributions 	int arg2 = uap->affinity;
3334*a325d9c4SApple OSS Distributions 	int arg3 = uap->prio;
3335*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
3336*a325d9c4SApple OSS Distributions 	int error = 0;
3337*a325d9c4SApple OSS Distributions 
3338*a325d9c4SApple OSS Distributions 	if ((p->p_lflag & P_LREGISTER) == 0) {
3339*a325d9c4SApple OSS Distributions 		return EINVAL;
3340*a325d9c4SApple OSS Distributions 	}
3341*a325d9c4SApple OSS Distributions 
3342*a325d9c4SApple OSS Distributions 	switch (options) {
3343*a325d9c4SApple OSS Distributions 	case WQOPS_QUEUE_NEWSPISUPP: {
3344*a325d9c4SApple OSS Distributions 		/*
3345*a325d9c4SApple OSS Distributions 		 * arg2 = offset of serialno into dispatch queue
3346*a325d9c4SApple OSS Distributions 		 * arg3 = kevent support
3347*a325d9c4SApple OSS Distributions 		 */
3348*a325d9c4SApple OSS Distributions 		int offset = arg2;
3349*a325d9c4SApple OSS Distributions 		if (arg3 & 0x01) {
3350*a325d9c4SApple OSS Distributions 			// If we get here, then userspace has indicated support for kevent delivery.
3351*a325d9c4SApple OSS Distributions 		}
3352*a325d9c4SApple OSS Distributions 
3353*a325d9c4SApple OSS Distributions 		p->p_dispatchqueue_serialno_offset = (uint64_t)offset;
3354*a325d9c4SApple OSS Distributions 		break;
3355*a325d9c4SApple OSS Distributions 	}
3356*a325d9c4SApple OSS Distributions 	case WQOPS_QUEUE_REQTHREADS: {
3357*a325d9c4SApple OSS Distributions 		/*
3358*a325d9c4SApple OSS Distributions 		 * arg2 = number of threads to start
3359*a325d9c4SApple OSS Distributions 		 * arg3 = priority
3360*a325d9c4SApple OSS Distributions 		 */
3361*a325d9c4SApple OSS Distributions 		error = workq_reqthreads(p, arg2, arg3, false);
3362*a325d9c4SApple OSS Distributions 		break;
3363*a325d9c4SApple OSS Distributions 	}
3364*a325d9c4SApple OSS Distributions 	/* For requesting threads for the cooperative pool */
3365*a325d9c4SApple OSS Distributions 	case WQOPS_QUEUE_REQTHREADS2: {
3366*a325d9c4SApple OSS Distributions 		/*
3367*a325d9c4SApple OSS Distributions 		 * arg2 = number of threads to start
3368*a325d9c4SApple OSS Distributions 		 * arg3 = priority
3369*a325d9c4SApple OSS Distributions 		 */
3370*a325d9c4SApple OSS Distributions 		error = workq_reqthreads(p, arg2, arg3, true);
3371*a325d9c4SApple OSS Distributions 		break;
3372*a325d9c4SApple OSS Distributions 	}
3373*a325d9c4SApple OSS Distributions 	case WQOPS_SET_EVENT_MANAGER_PRIORITY: {
3374*a325d9c4SApple OSS Distributions 		/*
3375*a325d9c4SApple OSS Distributions 		 * arg2 = priority for the manager thread
3376*a325d9c4SApple OSS Distributions 		 *
3377*a325d9c4SApple OSS Distributions 		 * if _PTHREAD_PRIORITY_SCHED_PRI_FLAG is set,
3378*a325d9c4SApple OSS Distributions 		 * the low bits of the value contains a scheduling priority
3379*a325d9c4SApple OSS Distributions 		 * instead of a QOS value
3380*a325d9c4SApple OSS Distributions 		 */
3381*a325d9c4SApple OSS Distributions 		pthread_priority_t pri = arg2;
3382*a325d9c4SApple OSS Distributions 
3383*a325d9c4SApple OSS Distributions 		if (wq == NULL) {
3384*a325d9c4SApple OSS Distributions 			error = EINVAL;
3385*a325d9c4SApple OSS Distributions 			break;
3386*a325d9c4SApple OSS Distributions 		}
3387*a325d9c4SApple OSS Distributions 
3388*a325d9c4SApple OSS Distributions 		/*
3389*a325d9c4SApple OSS Distributions 		 * Normalize the incoming priority so that it is ordered numerically.
3390*a325d9c4SApple OSS Distributions 		 */
3391*a325d9c4SApple OSS Distributions 		if (pri & _PTHREAD_PRIORITY_SCHED_PRI_FLAG) {
3392*a325d9c4SApple OSS Distributions 			pri &= (_PTHREAD_PRIORITY_SCHED_PRI_MASK |
3393*a325d9c4SApple OSS Distributions 			    _PTHREAD_PRIORITY_SCHED_PRI_FLAG);
3394*a325d9c4SApple OSS Distributions 		} else {
3395*a325d9c4SApple OSS Distributions 			thread_qos_t qos = _pthread_priority_thread_qos(pri);
3396*a325d9c4SApple OSS Distributions 			int relpri = _pthread_priority_relpri(pri);
3397*a325d9c4SApple OSS Distributions 			if (relpri > 0 || relpri < THREAD_QOS_MIN_TIER_IMPORTANCE ||
3398*a325d9c4SApple OSS Distributions 			    qos == THREAD_QOS_UNSPECIFIED) {
3399*a325d9c4SApple OSS Distributions 				error = EINVAL;
3400*a325d9c4SApple OSS Distributions 				break;
3401*a325d9c4SApple OSS Distributions 			}
3402*a325d9c4SApple OSS Distributions 			pri &= ~_PTHREAD_PRIORITY_FLAGS_MASK;
3403*a325d9c4SApple OSS Distributions 		}
3404*a325d9c4SApple OSS Distributions 
3405*a325d9c4SApple OSS Distributions 		/*
3406*a325d9c4SApple OSS Distributions 		 * If userspace passes a scheduling priority, that wins over any QoS.
3407*a325d9c4SApple OSS Distributions 		 * Userspace should takes care not to lower the priority this way.
3408*a325d9c4SApple OSS Distributions 		 */
3409*a325d9c4SApple OSS Distributions 		workq_lock_spin(wq);
3410*a325d9c4SApple OSS Distributions 		if (wq->wq_event_manager_priority < (uint32_t)pri) {
3411*a325d9c4SApple OSS Distributions 			wq->wq_event_manager_priority = (uint32_t)pri;
3412*a325d9c4SApple OSS Distributions 		}
3413*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
3414*a325d9c4SApple OSS Distributions 		break;
3415*a325d9c4SApple OSS Distributions 	}
3416*a325d9c4SApple OSS Distributions 	case WQOPS_THREAD_KEVENT_RETURN:
3417*a325d9c4SApple OSS Distributions 	case WQOPS_THREAD_WORKLOOP_RETURN:
3418*a325d9c4SApple OSS Distributions 	case WQOPS_THREAD_RETURN: {
3419*a325d9c4SApple OSS Distributions 		error = workq_thread_return(p, uap, wq);
3420*a325d9c4SApple OSS Distributions 		break;
3421*a325d9c4SApple OSS Distributions 	}
3422*a325d9c4SApple OSS Distributions 
3423*a325d9c4SApple OSS Distributions 	case WQOPS_SHOULD_NARROW: {
3424*a325d9c4SApple OSS Distributions 		/*
3425*a325d9c4SApple OSS Distributions 		 * arg2 = priority to test
3426*a325d9c4SApple OSS Distributions 		 * arg3 = unused
3427*a325d9c4SApple OSS Distributions 		 */
3428*a325d9c4SApple OSS Distributions 		thread_t th = current_thread();
3429*a325d9c4SApple OSS Distributions 		struct uthread *uth = get_bsdthread_info(th);
3430*a325d9c4SApple OSS Distributions 		if (((thread_get_tag(th) & THREAD_TAG_WORKQUEUE) == 0) ||
3431*a325d9c4SApple OSS Distributions 		    (uth->uu_workq_flags & (UT_WORKQ_DYING | UT_WORKQ_OVERCOMMIT))) {
3432*a325d9c4SApple OSS Distributions 			error = EINVAL;
3433*a325d9c4SApple OSS Distributions 			break;
3434*a325d9c4SApple OSS Distributions 		}
3435*a325d9c4SApple OSS Distributions 
3436*a325d9c4SApple OSS Distributions 		thread_qos_t qos = _pthread_priority_thread_qos(arg2);
3437*a325d9c4SApple OSS Distributions 		if (qos == THREAD_QOS_UNSPECIFIED) {
3438*a325d9c4SApple OSS Distributions 			error = EINVAL;
3439*a325d9c4SApple OSS Distributions 			break;
3440*a325d9c4SApple OSS Distributions 		}
3441*a325d9c4SApple OSS Distributions 		workq_lock_spin(wq);
3442*a325d9c4SApple OSS Distributions 		bool should_narrow = !workq_constrained_allowance(wq, qos, uth, false);
3443*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
3444*a325d9c4SApple OSS Distributions 
3445*a325d9c4SApple OSS Distributions 		*retval = should_narrow;
3446*a325d9c4SApple OSS Distributions 		break;
3447*a325d9c4SApple OSS Distributions 	}
3448*a325d9c4SApple OSS Distributions 	case WQOPS_SETUP_DISPATCH: {
3449*a325d9c4SApple OSS Distributions 		/*
3450*a325d9c4SApple OSS Distributions 		 * item = pointer to workq_dispatch_config structure
3451*a325d9c4SApple OSS Distributions 		 * arg2 = sizeof(item)
3452*a325d9c4SApple OSS Distributions 		 */
3453*a325d9c4SApple OSS Distributions 		struct workq_dispatch_config cfg;
3454*a325d9c4SApple OSS Distributions 		bzero(&cfg, sizeof(cfg));
3455*a325d9c4SApple OSS Distributions 
3456*a325d9c4SApple OSS Distributions 		error = copyin(uap->item, &cfg, MIN(sizeof(cfg), (unsigned long) arg2));
3457*a325d9c4SApple OSS Distributions 		if (error) {
3458*a325d9c4SApple OSS Distributions 			break;
3459*a325d9c4SApple OSS Distributions 		}
3460*a325d9c4SApple OSS Distributions 
3461*a325d9c4SApple OSS Distributions 		if (cfg.wdc_flags & ~WORKQ_DISPATCH_SUPPORTED_FLAGS ||
3462*a325d9c4SApple OSS Distributions 		    cfg.wdc_version < WORKQ_DISPATCH_MIN_SUPPORTED_VERSION) {
3463*a325d9c4SApple OSS Distributions 			error = ENOTSUP;
3464*a325d9c4SApple OSS Distributions 			break;
3465*a325d9c4SApple OSS Distributions 		}
3466*a325d9c4SApple OSS Distributions 
3467*a325d9c4SApple OSS Distributions 		/* Load fields from version 1 */
3468*a325d9c4SApple OSS Distributions 		p->p_dispatchqueue_serialno_offset = cfg.wdc_queue_serialno_offs;
3469*a325d9c4SApple OSS Distributions 
3470*a325d9c4SApple OSS Distributions 		/* Load fields from version 2 */
3471*a325d9c4SApple OSS Distributions 		if (cfg.wdc_version >= 2) {
3472*a325d9c4SApple OSS Distributions 			p->p_dispatchqueue_label_offset = cfg.wdc_queue_label_offs;
3473*a325d9c4SApple OSS Distributions 		}
3474*a325d9c4SApple OSS Distributions 
3475*a325d9c4SApple OSS Distributions 		break;
3476*a325d9c4SApple OSS Distributions 	}
3477*a325d9c4SApple OSS Distributions 	default:
3478*a325d9c4SApple OSS Distributions 		error = EINVAL;
3479*a325d9c4SApple OSS Distributions 		break;
3480*a325d9c4SApple OSS Distributions 	}
3481*a325d9c4SApple OSS Distributions 
3482*a325d9c4SApple OSS Distributions 	return error;
3483*a325d9c4SApple OSS Distributions }
3484*a325d9c4SApple OSS Distributions 
3485*a325d9c4SApple OSS Distributions /*
3486*a325d9c4SApple OSS Distributions  * We have no work to do, park ourselves on the idle list.
3487*a325d9c4SApple OSS Distributions  *
3488*a325d9c4SApple OSS Distributions  * Consumes the workqueue lock and does not return.
3489*a325d9c4SApple OSS Distributions  */
3490*a325d9c4SApple OSS Distributions __attribute__((noreturn, noinline))
3491*a325d9c4SApple OSS Distributions static void
workq_park_and_unlock(proc_t p,struct workqueue * wq,struct uthread * uth,uint32_t setup_flags)3492*a325d9c4SApple OSS Distributions workq_park_and_unlock(proc_t p, struct workqueue *wq, struct uthread *uth,
3493*a325d9c4SApple OSS Distributions     uint32_t setup_flags)
3494*a325d9c4SApple OSS Distributions {
3495*a325d9c4SApple OSS Distributions 	assert(uth == current_uthread());
3496*a325d9c4SApple OSS Distributions 	assert(uth->uu_kqr_bound == NULL);
3497*a325d9c4SApple OSS Distributions 	workq_push_idle_thread(p, wq, uth, setup_flags); // may not return
3498*a325d9c4SApple OSS Distributions 
3499*a325d9c4SApple OSS Distributions 	workq_thread_reset_cpupercent(NULL, uth);
3500*a325d9c4SApple OSS Distributions 
3501*a325d9c4SApple OSS Distributions #if CONFIG_PREADOPT_TG
3502*a325d9c4SApple OSS Distributions 	/* Clear the preadoption thread group on the thread.
3503*a325d9c4SApple OSS Distributions 	 *
3504*a325d9c4SApple OSS Distributions 	 * Case 1:
3505*a325d9c4SApple OSS Distributions 	 *		Creator thread which never picked up a thread request. We set a
3506*a325d9c4SApple OSS Distributions 	 *		preadoption thread group on creator threads but if it never picked
3507*a325d9c4SApple OSS Distributions 	 *		up a thread request and didn't go to userspace, then the thread will
3508*a325d9c4SApple OSS Distributions 	 *		park with a preadoption thread group but no explicitly adopted
3509*a325d9c4SApple OSS Distributions 	 *		voucher or work interval.
3510*a325d9c4SApple OSS Distributions 	 *
3511*a325d9c4SApple OSS Distributions 	 *		We drop the preadoption thread group here before proceeding to park.
3512*a325d9c4SApple OSS Distributions 	 *		Note - we may get preempted when we drop the workq lock below.
3513*a325d9c4SApple OSS Distributions 	 *
3514*a325d9c4SApple OSS Distributions 	 * Case 2:
3515*a325d9c4SApple OSS Distributions 	 *		Thread picked up a thread request and bound to it and returned back
3516*a325d9c4SApple OSS Distributions 	 *		from userspace and is parking. At this point, preadoption thread
3517*a325d9c4SApple OSS Distributions 	 *		group should be NULL since the thread has unbound from the thread
3518*a325d9c4SApple OSS Distributions 	 *		request. So this operation should be a no-op.
3519*a325d9c4SApple OSS Distributions 	 */
3520*a325d9c4SApple OSS Distributions 	thread_set_preadopt_thread_group(get_machthread(uth), NULL);
3521*a325d9c4SApple OSS Distributions #endif
3522*a325d9c4SApple OSS Distributions 
3523*a325d9c4SApple OSS Distributions 	if ((uth->uu_workq_flags & UT_WORKQ_IDLE_CLEANUP) &&
3524*a325d9c4SApple OSS Distributions 	    !(uth->uu_workq_flags & UT_WORKQ_DYING)) {
3525*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
3526*a325d9c4SApple OSS Distributions 
3527*a325d9c4SApple OSS Distributions 		/*
3528*a325d9c4SApple OSS Distributions 		 * workq_push_idle_thread() will unset `has_stack`
3529*a325d9c4SApple OSS Distributions 		 * if it wants us to free the stack before parking.
3530*a325d9c4SApple OSS Distributions 		 */
3531*a325d9c4SApple OSS Distributions 		if (!uth->uu_save.uus_workq_park_data.has_stack) {
3532*a325d9c4SApple OSS Distributions 			pthread_functions->workq_markfree_threadstack(p,
3533*a325d9c4SApple OSS Distributions 			    get_machthread(uth), get_task_map(p->task),
3534*a325d9c4SApple OSS Distributions 			    uth->uu_workq_stackaddr);
3535*a325d9c4SApple OSS Distributions 		}
3536*a325d9c4SApple OSS Distributions 
3537*a325d9c4SApple OSS Distributions 		/*
3538*a325d9c4SApple OSS Distributions 		 * When we remove the voucher from the thread, we may lose our importance
3539*a325d9c4SApple OSS Distributions 		 * causing us to get preempted, so we do this after putting the thread on
3540*a325d9c4SApple OSS Distributions 		 * the idle list.  Then, when we get our importance back we'll be able to
3541*a325d9c4SApple OSS Distributions 		 * use this thread from e.g. the kevent call out to deliver a boosting
3542*a325d9c4SApple OSS Distributions 		 * message.
3543*a325d9c4SApple OSS Distributions 		 *
3544*a325d9c4SApple OSS Distributions 		 * Note that setting the voucher to NULL will not clear the preadoption
3545*a325d9c4SApple OSS Distributions 		 * thread since this thread could have become the creator again and
3546*a325d9c4SApple OSS Distributions 		 * perhaps acquired a preadoption thread group.
3547*a325d9c4SApple OSS Distributions 		 */
3548*a325d9c4SApple OSS Distributions 		__assert_only kern_return_t kr;
3549*a325d9c4SApple OSS Distributions 		kr = thread_set_voucher_name(MACH_PORT_NULL);
3550*a325d9c4SApple OSS Distributions 		assert(kr == KERN_SUCCESS);
3551*a325d9c4SApple OSS Distributions 
3552*a325d9c4SApple OSS Distributions 		workq_lock_spin(wq);
3553*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags &= ~UT_WORKQ_IDLE_CLEANUP;
3554*a325d9c4SApple OSS Distributions 		setup_flags &= ~WQ_SETUP_CLEAR_VOUCHER;
3555*a325d9c4SApple OSS Distributions 	}
3556*a325d9c4SApple OSS Distributions 
3557*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_thread_logical_run | DBG_FUNC_END, wq, 0, 0, 0);
3558*a325d9c4SApple OSS Distributions 
3559*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_RUNNING) {
3560*a325d9c4SApple OSS Distributions 		/*
3561*a325d9c4SApple OSS Distributions 		 * While we'd dropped the lock to unset our voucher, someone came
3562*a325d9c4SApple OSS Distributions 		 * around and made us runnable.  But because we weren't waiting on the
3563*a325d9c4SApple OSS Distributions 		 * event their thread_wakeup() was ineffectual.  To correct for that,
3564*a325d9c4SApple OSS Distributions 		 * we just run the continuation ourselves.
3565*a325d9c4SApple OSS Distributions 		 */
3566*a325d9c4SApple OSS Distributions 		workq_unpark_select_threadreq_or_park_and_unlock(p, wq, uth, setup_flags);
3567*a325d9c4SApple OSS Distributions 		__builtin_unreachable();
3568*a325d9c4SApple OSS Distributions 	}
3569*a325d9c4SApple OSS Distributions 
3570*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_DYING) {
3571*a325d9c4SApple OSS Distributions 		workq_unpark_for_death_and_unlock(p, wq, uth,
3572*a325d9c4SApple OSS Distributions 		    WORKQ_UNPARK_FOR_DEATH_WAS_IDLE, setup_flags);
3573*a325d9c4SApple OSS Distributions 		__builtin_unreachable();
3574*a325d9c4SApple OSS Distributions 	}
3575*a325d9c4SApple OSS Distributions 
3576*a325d9c4SApple OSS Distributions 	/* Disarm the workqueue quantum since the thread is now idle */
3577*a325d9c4SApple OSS Distributions 	thread_disarm_workqueue_quantum(get_machthread(uth));
3578*a325d9c4SApple OSS Distributions 
3579*a325d9c4SApple OSS Distributions 	thread_set_pending_block_hint(get_machthread(uth), kThreadWaitParkedWorkQueue);
3580*a325d9c4SApple OSS Distributions 	assert_wait(workq_parked_wait_event(uth), THREAD_INTERRUPTIBLE);
3581*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
3582*a325d9c4SApple OSS Distributions 	thread_block(workq_unpark_continue);
3583*a325d9c4SApple OSS Distributions 	__builtin_unreachable();
3584*a325d9c4SApple OSS Distributions }
3585*a325d9c4SApple OSS Distributions 
3586*a325d9c4SApple OSS Distributions static inline bool
workq_may_start_event_mgr_thread(struct workqueue * wq,struct uthread * uth)3587*a325d9c4SApple OSS Distributions workq_may_start_event_mgr_thread(struct workqueue *wq, struct uthread *uth)
3588*a325d9c4SApple OSS Distributions {
3589*a325d9c4SApple OSS Distributions 	/*
3590*a325d9c4SApple OSS Distributions 	 * There's an event manager request and either:
3591*a325d9c4SApple OSS Distributions 	 * - no event manager currently running
3592*a325d9c4SApple OSS Distributions 	 * - we are re-using the event manager
3593*a325d9c4SApple OSS Distributions 	 */
3594*a325d9c4SApple OSS Distributions 	return wq->wq_thscheduled_count[_wq_bucket(WORKQ_THREAD_QOS_MANAGER)] == 0 ||
3595*a325d9c4SApple OSS Distributions 	       (uth && uth->uu_workq_pri.qos_bucket == WORKQ_THREAD_QOS_MANAGER);
3596*a325d9c4SApple OSS Distributions }
3597*a325d9c4SApple OSS Distributions 
3598*a325d9c4SApple OSS Distributions static uint32_t
workq_constrained_allowance(struct workqueue * wq,thread_qos_t at_qos,struct uthread * uth,bool may_start_timer)3599*a325d9c4SApple OSS Distributions workq_constrained_allowance(struct workqueue *wq, thread_qos_t at_qos,
3600*a325d9c4SApple OSS Distributions     struct uthread *uth, bool may_start_timer)
3601*a325d9c4SApple OSS Distributions {
3602*a325d9c4SApple OSS Distributions 	assert(at_qos != WORKQ_THREAD_QOS_MANAGER);
3603*a325d9c4SApple OSS Distributions 	uint32_t count = 0;
3604*a325d9c4SApple OSS Distributions 
3605*a325d9c4SApple OSS Distributions 	uint32_t max_count = wq->wq_constrained_threads_scheduled;
3606*a325d9c4SApple OSS Distributions 	if (uth && workq_thread_is_nonovercommit(uth)) {
3607*a325d9c4SApple OSS Distributions 		/*
3608*a325d9c4SApple OSS Distributions 		 * don't count the current thread as scheduled
3609*a325d9c4SApple OSS Distributions 		 */
3610*a325d9c4SApple OSS Distributions 		assert(max_count > 0);
3611*a325d9c4SApple OSS Distributions 		max_count--;
3612*a325d9c4SApple OSS Distributions 	}
3613*a325d9c4SApple OSS Distributions 	if (max_count >= wq_max_constrained_threads) {
3614*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_constrained_admission | DBG_FUNC_NONE, wq, 1,
3615*a325d9c4SApple OSS Distributions 		    wq->wq_constrained_threads_scheduled,
3616*a325d9c4SApple OSS Distributions 		    wq_max_constrained_threads);
3617*a325d9c4SApple OSS Distributions 		/*
3618*a325d9c4SApple OSS Distributions 		 * we need 1 or more constrained threads to return to the kernel before
3619*a325d9c4SApple OSS Distributions 		 * we can dispatch additional work
3620*a325d9c4SApple OSS Distributions 		 */
3621*a325d9c4SApple OSS Distributions 		return 0;
3622*a325d9c4SApple OSS Distributions 	}
3623*a325d9c4SApple OSS Distributions 	max_count -= wq_max_constrained_threads;
3624*a325d9c4SApple OSS Distributions 
3625*a325d9c4SApple OSS Distributions 	/*
3626*a325d9c4SApple OSS Distributions 	 * Compute a metric for many how many threads are active.  We find the
3627*a325d9c4SApple OSS Distributions 	 * highest priority request outstanding and then add up the number of active
3628*a325d9c4SApple OSS Distributions 	 * threads in that and all higher-priority buckets.  We'll also add any
3629*a325d9c4SApple OSS Distributions 	 * "busy" threads which are not currently active but blocked recently enough
3630*a325d9c4SApple OSS Distributions 	 * that we can't be sure that they won't be unblocked soon and start
3631*a325d9c4SApple OSS Distributions 	 * being active again.
3632*a325d9c4SApple OSS Distributions 	 *
3633*a325d9c4SApple OSS Distributions 	 * We'll then compare this metric to our max concurrency to decide whether
3634*a325d9c4SApple OSS Distributions 	 * to add a new thread.
3635*a325d9c4SApple OSS Distributions 	 */
3636*a325d9c4SApple OSS Distributions 
3637*a325d9c4SApple OSS Distributions 	uint32_t busycount, thactive_count;
3638*a325d9c4SApple OSS Distributions 
3639*a325d9c4SApple OSS Distributions 	thactive_count = _wq_thactive_aggregate_downto_qos(wq, _wq_thactive(wq),
3640*a325d9c4SApple OSS Distributions 	    at_qos, &busycount, NULL);
3641*a325d9c4SApple OSS Distributions 
3642*a325d9c4SApple OSS Distributions 	if (uth && uth->uu_workq_pri.qos_bucket != WORKQ_THREAD_QOS_MANAGER &&
3643*a325d9c4SApple OSS Distributions 	    at_qos <= uth->uu_workq_pri.qos_bucket) {
3644*a325d9c4SApple OSS Distributions 		/*
3645*a325d9c4SApple OSS Distributions 		 * Don't count this thread as currently active, but only if it's not
3646*a325d9c4SApple OSS Distributions 		 * a manager thread, as _wq_thactive_aggregate_downto_qos ignores active
3647*a325d9c4SApple OSS Distributions 		 * managers.
3648*a325d9c4SApple OSS Distributions 		 */
3649*a325d9c4SApple OSS Distributions 		assert(thactive_count > 0);
3650*a325d9c4SApple OSS Distributions 		thactive_count--;
3651*a325d9c4SApple OSS Distributions 	}
3652*a325d9c4SApple OSS Distributions 
3653*a325d9c4SApple OSS Distributions 	count = wq_max_parallelism[_wq_bucket(at_qos)];
3654*a325d9c4SApple OSS Distributions 	if (count > thactive_count + busycount) {
3655*a325d9c4SApple OSS Distributions 		count -= thactive_count + busycount;
3656*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_constrained_admission | DBG_FUNC_NONE, wq, 2,
3657*a325d9c4SApple OSS Distributions 		    thactive_count, busycount);
3658*a325d9c4SApple OSS Distributions 		return MIN(count, max_count);
3659*a325d9c4SApple OSS Distributions 	} else {
3660*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_constrained_admission | DBG_FUNC_NONE, wq, 3,
3661*a325d9c4SApple OSS Distributions 		    thactive_count, busycount);
3662*a325d9c4SApple OSS Distributions 	}
3663*a325d9c4SApple OSS Distributions 
3664*a325d9c4SApple OSS Distributions 	if (may_start_timer) {
3665*a325d9c4SApple OSS Distributions 		/*
3666*a325d9c4SApple OSS Distributions 		 * If this is called from the add timer, we won't have another timer
3667*a325d9c4SApple OSS Distributions 		 * fire when the thread exits the "busy" state, so rearm the timer.
3668*a325d9c4SApple OSS Distributions 		 */
3669*a325d9c4SApple OSS Distributions 		workq_schedule_delayed_thread_creation(wq, 0);
3670*a325d9c4SApple OSS Distributions 	}
3671*a325d9c4SApple OSS Distributions 
3672*a325d9c4SApple OSS Distributions 	return 0;
3673*a325d9c4SApple OSS Distributions }
3674*a325d9c4SApple OSS Distributions 
3675*a325d9c4SApple OSS Distributions static bool
workq_threadreq_admissible(struct workqueue * wq,struct uthread * uth,workq_threadreq_t req)3676*a325d9c4SApple OSS Distributions workq_threadreq_admissible(struct workqueue *wq, struct uthread *uth,
3677*a325d9c4SApple OSS Distributions     workq_threadreq_t req)
3678*a325d9c4SApple OSS Distributions {
3679*a325d9c4SApple OSS Distributions 	if (req->tr_qos == WORKQ_THREAD_QOS_MANAGER) {
3680*a325d9c4SApple OSS Distributions 		return workq_may_start_event_mgr_thread(wq, uth);
3681*a325d9c4SApple OSS Distributions 	}
3682*a325d9c4SApple OSS Distributions 	if (workq_threadreq_is_cooperative(req)) {
3683*a325d9c4SApple OSS Distributions 		return workq_cooperative_allowance(wq, req->tr_qos, uth, true);
3684*a325d9c4SApple OSS Distributions 	}
3685*a325d9c4SApple OSS Distributions 	if (workq_threadreq_is_nonovercommit(req)) {
3686*a325d9c4SApple OSS Distributions 		return workq_constrained_allowance(wq, req->tr_qos, uth, true);
3687*a325d9c4SApple OSS Distributions 	}
3688*a325d9c4SApple OSS Distributions 
3689*a325d9c4SApple OSS Distributions 	return true;
3690*a325d9c4SApple OSS Distributions }
3691*a325d9c4SApple OSS Distributions 
3692*a325d9c4SApple OSS Distributions /*
3693*a325d9c4SApple OSS Distributions  * Called from the context of selecting thread requests for threads returning
3694*a325d9c4SApple OSS Distributions  * from userspace or creator thread
3695*a325d9c4SApple OSS Distributions  */
3696*a325d9c4SApple OSS Distributions static workq_threadreq_t
workq_cooperative_queue_best_req(struct workqueue * wq,struct uthread * uth)3697*a325d9c4SApple OSS Distributions workq_cooperative_queue_best_req(struct workqueue *wq, struct uthread *uth)
3698*a325d9c4SApple OSS Distributions {
3699*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
3700*a325d9c4SApple OSS Distributions 
3701*a325d9c4SApple OSS Distributions 	/*
3702*a325d9c4SApple OSS Distributions 	 * If the current thread is cooperative, we need to exclude it as part of
3703*a325d9c4SApple OSS Distributions 	 * cooperative schedule count since this thread is looking for a new
3704*a325d9c4SApple OSS Distributions 	 * request. Change in the schedule count for cooperative pool therefore
3705*a325d9c4SApple OSS Distributions 	 * requires us to reeevaluate the next best request for it.
3706*a325d9c4SApple OSS Distributions 	 */
3707*a325d9c4SApple OSS Distributions 	if (uth && workq_thread_is_cooperative(uth)) {
3708*a325d9c4SApple OSS Distributions 		_wq_cooperative_queue_scheduled_count_dec(wq, uth->uu_workq_pri.qos_bucket);
3709*a325d9c4SApple OSS Distributions 
3710*a325d9c4SApple OSS Distributions 		(void) _wq_cooperative_queue_refresh_best_req_qos(wq);
3711*a325d9c4SApple OSS Distributions 
3712*a325d9c4SApple OSS Distributions 		_wq_cooperative_queue_scheduled_count_inc(wq, uth->uu_workq_pri.qos_bucket);
3713*a325d9c4SApple OSS Distributions 	} else {
3714*a325d9c4SApple OSS Distributions 		/*
3715*a325d9c4SApple OSS Distributions 		 * The old value that was already precomputed should be safe to use -
3716*a325d9c4SApple OSS Distributions 		 * add an assert that asserts that the best req QoS doesn't change in
3717*a325d9c4SApple OSS Distributions 		 * this case
3718*a325d9c4SApple OSS Distributions 		 */
3719*a325d9c4SApple OSS Distributions 		assert(_wq_cooperative_queue_refresh_best_req_qos(wq) == false);
3720*a325d9c4SApple OSS Distributions 	}
3721*a325d9c4SApple OSS Distributions 
3722*a325d9c4SApple OSS Distributions 	thread_qos_t qos = wq->wq_cooperative_queue_best_req_qos;
3723*a325d9c4SApple OSS Distributions 
3724*a325d9c4SApple OSS Distributions 	/* There are no eligible requests in the cooperative pool */
3725*a325d9c4SApple OSS Distributions 	if (qos == THREAD_QOS_UNSPECIFIED) {
3726*a325d9c4SApple OSS Distributions 		return NULL;
3727*a325d9c4SApple OSS Distributions 	}
3728*a325d9c4SApple OSS Distributions 	assert(qos != WORKQ_THREAD_QOS_ABOVEUI);
3729*a325d9c4SApple OSS Distributions 	assert(qos != WORKQ_THREAD_QOS_MANAGER);
3730*a325d9c4SApple OSS Distributions 
3731*a325d9c4SApple OSS Distributions 	int bucket = _wq_bucket(qos);
3732*a325d9c4SApple OSS Distributions 	assert(!STAILQ_EMPTY(&wq->wq_cooperative_queue[bucket]));
3733*a325d9c4SApple OSS Distributions 
3734*a325d9c4SApple OSS Distributions 	return STAILQ_FIRST(&wq->wq_cooperative_queue[bucket]);
3735*a325d9c4SApple OSS Distributions }
3736*a325d9c4SApple OSS Distributions 
3737*a325d9c4SApple OSS Distributions static workq_threadreq_t
workq_threadreq_select_for_creator(struct workqueue * wq)3738*a325d9c4SApple OSS Distributions workq_threadreq_select_for_creator(struct workqueue *wq)
3739*a325d9c4SApple OSS Distributions {
3740*a325d9c4SApple OSS Distributions 	workq_threadreq_t req_qos, req_pri, req_tmp, req_mgr;
3741*a325d9c4SApple OSS Distributions 	thread_qos_t qos = THREAD_QOS_UNSPECIFIED;
3742*a325d9c4SApple OSS Distributions 	uint8_t pri = 0;
3743*a325d9c4SApple OSS Distributions 
3744*a325d9c4SApple OSS Distributions 	/*
3745*a325d9c4SApple OSS Distributions 	 * Compute the best priority request, and ignore the turnstile for now
3746*a325d9c4SApple OSS Distributions 	 */
3747*a325d9c4SApple OSS Distributions 
3748*a325d9c4SApple OSS Distributions 	req_pri = priority_queue_max(&wq->wq_special_queue,
3749*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry);
3750*a325d9c4SApple OSS Distributions 	if (req_pri) {
3751*a325d9c4SApple OSS Distributions 		pri = (uint8_t)priority_queue_entry_sched_pri(&wq->wq_special_queue,
3752*a325d9c4SApple OSS Distributions 		    &req_pri->tr_entry);
3753*a325d9c4SApple OSS Distributions 	}
3754*a325d9c4SApple OSS Distributions 
3755*a325d9c4SApple OSS Distributions 	/*
3756*a325d9c4SApple OSS Distributions 	 * Handle the manager thread request. The special queue might yield
3757*a325d9c4SApple OSS Distributions 	 * a higher priority, but the manager always beats the QoS world.
3758*a325d9c4SApple OSS Distributions 	 */
3759*a325d9c4SApple OSS Distributions 
3760*a325d9c4SApple OSS Distributions 	req_mgr = wq->wq_event_manager_threadreq;
3761*a325d9c4SApple OSS Distributions 	if (req_mgr && workq_may_start_event_mgr_thread(wq, NULL)) {
3762*a325d9c4SApple OSS Distributions 		uint32_t mgr_pri = wq->wq_event_manager_priority;
3763*a325d9c4SApple OSS Distributions 
3764*a325d9c4SApple OSS Distributions 		if (mgr_pri & _PTHREAD_PRIORITY_SCHED_PRI_FLAG) {
3765*a325d9c4SApple OSS Distributions 			mgr_pri &= _PTHREAD_PRIORITY_SCHED_PRI_MASK;
3766*a325d9c4SApple OSS Distributions 		} else {
3767*a325d9c4SApple OSS Distributions 			mgr_pri = thread_workq_pri_for_qos(
3768*a325d9c4SApple OSS Distributions 				_pthread_priority_thread_qos(mgr_pri));
3769*a325d9c4SApple OSS Distributions 		}
3770*a325d9c4SApple OSS Distributions 
3771*a325d9c4SApple OSS Distributions 		return mgr_pri >= pri ? req_mgr : req_pri;
3772*a325d9c4SApple OSS Distributions 	}
3773*a325d9c4SApple OSS Distributions 
3774*a325d9c4SApple OSS Distributions 	/*
3775*a325d9c4SApple OSS Distributions 	 * Compute the best QoS Request, and check whether it beats the "pri" one
3776*a325d9c4SApple OSS Distributions 	 *
3777*a325d9c4SApple OSS Distributions 	 * Start by comparing the overcommit and the cooperative pool
3778*a325d9c4SApple OSS Distributions 	 */
3779*a325d9c4SApple OSS Distributions 	req_qos = priority_queue_max(&wq->wq_overcommit_queue,
3780*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry);
3781*a325d9c4SApple OSS Distributions 	if (req_qos) {
3782*a325d9c4SApple OSS Distributions 		qos = req_qos->tr_qos;
3783*a325d9c4SApple OSS Distributions 	}
3784*a325d9c4SApple OSS Distributions 
3785*a325d9c4SApple OSS Distributions 	req_tmp = workq_cooperative_queue_best_req(wq, NULL);
3786*a325d9c4SApple OSS Distributions 	if (req_tmp && qos <= req_tmp->tr_qos) {
3787*a325d9c4SApple OSS Distributions 		/*
3788*a325d9c4SApple OSS Distributions 		 * Cooperative TR is better between overcommit and cooperative.  Note
3789*a325d9c4SApple OSS Distributions 		 * that if qos is same between overcommit and cooperative, we choose
3790*a325d9c4SApple OSS Distributions 		 * cooperative.
3791*a325d9c4SApple OSS Distributions 		 *
3792*a325d9c4SApple OSS Distributions 		 * Pick cooperative pool if it passes the admissions check
3793*a325d9c4SApple OSS Distributions 		 */
3794*a325d9c4SApple OSS Distributions 		if (workq_cooperative_allowance(wq, req_tmp->tr_qos, NULL, true)) {
3795*a325d9c4SApple OSS Distributions 			req_qos = req_tmp;
3796*a325d9c4SApple OSS Distributions 			qos = req_qos->tr_qos;
3797*a325d9c4SApple OSS Distributions 		}
3798*a325d9c4SApple OSS Distributions 	}
3799*a325d9c4SApple OSS Distributions 
3800*a325d9c4SApple OSS Distributions 	/*
3801*a325d9c4SApple OSS Distributions 	 * Compare the best QoS so far - either from overcommit or from cooperative
3802*a325d9c4SApple OSS Distributions 	 * pool - and compare it with the constrained pool
3803*a325d9c4SApple OSS Distributions 	 */
3804*a325d9c4SApple OSS Distributions 	req_tmp = priority_queue_max(&wq->wq_constrained_queue,
3805*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry);
3806*a325d9c4SApple OSS Distributions 
3807*a325d9c4SApple OSS Distributions 	if (req_tmp && qos < req_tmp->tr_qos) {
3808*a325d9c4SApple OSS Distributions 		/*
3809*a325d9c4SApple OSS Distributions 		 * Constrained pool is best in QoS between overcommit, cooperative
3810*a325d9c4SApple OSS Distributions 		 * and constrained. Now check how it fairs against the priority case
3811*a325d9c4SApple OSS Distributions 		 */
3812*a325d9c4SApple OSS Distributions 		if (pri && pri >= thread_workq_pri_for_qos(req_tmp->tr_qos)) {
3813*a325d9c4SApple OSS Distributions 			return req_pri;
3814*a325d9c4SApple OSS Distributions 		}
3815*a325d9c4SApple OSS Distributions 
3816*a325d9c4SApple OSS Distributions 		if (workq_constrained_allowance(wq, req_tmp->tr_qos, NULL, true)) {
3817*a325d9c4SApple OSS Distributions 			/*
3818*a325d9c4SApple OSS Distributions 			 * If the constrained thread request is the best one and passes
3819*a325d9c4SApple OSS Distributions 			 * the admission check, pick it.
3820*a325d9c4SApple OSS Distributions 			 */
3821*a325d9c4SApple OSS Distributions 			return req_tmp;
3822*a325d9c4SApple OSS Distributions 		}
3823*a325d9c4SApple OSS Distributions 	}
3824*a325d9c4SApple OSS Distributions 
3825*a325d9c4SApple OSS Distributions 	/*
3826*a325d9c4SApple OSS Distributions 	 * Compare the best of the QoS world with the priority
3827*a325d9c4SApple OSS Distributions 	 */
3828*a325d9c4SApple OSS Distributions 	if (pri && (!qos || pri >= thread_workq_pri_for_qos(qos))) {
3829*a325d9c4SApple OSS Distributions 		return req_pri;
3830*a325d9c4SApple OSS Distributions 	}
3831*a325d9c4SApple OSS Distributions 
3832*a325d9c4SApple OSS Distributions 	if (req_qos) {
3833*a325d9c4SApple OSS Distributions 		return req_qos;
3834*a325d9c4SApple OSS Distributions 	}
3835*a325d9c4SApple OSS Distributions 
3836*a325d9c4SApple OSS Distributions 	/*
3837*a325d9c4SApple OSS Distributions 	 * If we had no eligible request but we have a turnstile push,
3838*a325d9c4SApple OSS Distributions 	 * it must be a non overcommit thread request that failed
3839*a325d9c4SApple OSS Distributions 	 * the admission check.
3840*a325d9c4SApple OSS Distributions 	 *
3841*a325d9c4SApple OSS Distributions 	 * Just fake a BG thread request so that if the push stops the creator
3842*a325d9c4SApple OSS Distributions 	 * priority just drops to 4.
3843*a325d9c4SApple OSS Distributions 	 */
3844*a325d9c4SApple OSS Distributions 	if (turnstile_workq_proprietor_of_max_turnstile(wq->wq_turnstile, NULL)) {
3845*a325d9c4SApple OSS Distributions 		static struct workq_threadreq_s workq_sync_push_fake_req = {
3846*a325d9c4SApple OSS Distributions 			.tr_qos = THREAD_QOS_BACKGROUND,
3847*a325d9c4SApple OSS Distributions 		};
3848*a325d9c4SApple OSS Distributions 
3849*a325d9c4SApple OSS Distributions 		return &workq_sync_push_fake_req;
3850*a325d9c4SApple OSS Distributions 	}
3851*a325d9c4SApple OSS Distributions 
3852*a325d9c4SApple OSS Distributions 	return NULL;
3853*a325d9c4SApple OSS Distributions }
3854*a325d9c4SApple OSS Distributions 
3855*a325d9c4SApple OSS Distributions /*
3856*a325d9c4SApple OSS Distributions  * Returns true if this caused a change in the schedule counts of the
3857*a325d9c4SApple OSS Distributions  * cooperative pool
3858*a325d9c4SApple OSS Distributions  */
3859*a325d9c4SApple OSS Distributions static bool
workq_adjust_cooperative_constrained_schedule_counts(struct workqueue * wq,struct uthread * uth,thread_qos_t old_thread_qos,workq_tr_flags_t tr_flags)3860*a325d9c4SApple OSS Distributions workq_adjust_cooperative_constrained_schedule_counts(struct workqueue *wq,
3861*a325d9c4SApple OSS Distributions     struct uthread *uth, thread_qos_t old_thread_qos, workq_tr_flags_t tr_flags)
3862*a325d9c4SApple OSS Distributions {
3863*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
3864*a325d9c4SApple OSS Distributions 
3865*a325d9c4SApple OSS Distributions 	/*
3866*a325d9c4SApple OSS Distributions 	 * Row: thread type
3867*a325d9c4SApple OSS Distributions 	 * Column: Request type
3868*a325d9c4SApple OSS Distributions 	 *
3869*a325d9c4SApple OSS Distributions 	 *					overcommit		non-overcommit		cooperative
3870*a325d9c4SApple OSS Distributions 	 * overcommit			X				case 1				case 2
3871*a325d9c4SApple OSS Distributions 	 * cooperative		case 3				case 4				case 5
3872*a325d9c4SApple OSS Distributions 	 * non-overcommit	case 6					X				case 7
3873*a325d9c4SApple OSS Distributions 	 *
3874*a325d9c4SApple OSS Distributions 	 * Move the thread to the right bucket depending on what state it currently
3875*a325d9c4SApple OSS Distributions 	 * has and what state the thread req it picks, is going to have.
3876*a325d9c4SApple OSS Distributions 	 *
3877*a325d9c4SApple OSS Distributions 	 * Note that the creator thread is an overcommit thread.
3878*a325d9c4SApple OSS Distributions 	 */
3879*a325d9c4SApple OSS Distributions 	thread_qos_t new_thread_qos = uth->uu_workq_pri.qos_bucket;
3880*a325d9c4SApple OSS Distributions 
3881*a325d9c4SApple OSS Distributions 	/*
3882*a325d9c4SApple OSS Distributions 	 * Anytime a cooperative bucket's schedule count changes, we need to
3883*a325d9c4SApple OSS Distributions 	 * potentially refresh the next best QoS for that pool when we determine
3884*a325d9c4SApple OSS Distributions 	 * the next request for the creator
3885*a325d9c4SApple OSS Distributions 	 */
3886*a325d9c4SApple OSS Distributions 	bool cooperative_pool_sched_count_changed = false;
3887*a325d9c4SApple OSS Distributions 
3888*a325d9c4SApple OSS Distributions 	if (workq_thread_is_overcommit(uth)) {
3889*a325d9c4SApple OSS Distributions 		if (workq_tr_is_nonovercommit(tr_flags)) {
3890*a325d9c4SApple OSS Distributions 			// Case 1: thread is overcommit, req is non-overcommit
3891*a325d9c4SApple OSS Distributions 			wq->wq_constrained_threads_scheduled++;
3892*a325d9c4SApple OSS Distributions 		} else if (workq_tr_is_cooperative(tr_flags)) {
3893*a325d9c4SApple OSS Distributions 			// Case 2: thread is overcommit, req is cooperative
3894*a325d9c4SApple OSS Distributions 			_wq_cooperative_queue_scheduled_count_inc(wq, new_thread_qos);
3895*a325d9c4SApple OSS Distributions 			cooperative_pool_sched_count_changed = true;
3896*a325d9c4SApple OSS Distributions 		}
3897*a325d9c4SApple OSS Distributions 	} else if (workq_thread_is_cooperative(uth)) {
3898*a325d9c4SApple OSS Distributions 		if (workq_tr_is_overcommit(tr_flags)) {
3899*a325d9c4SApple OSS Distributions 			// Case 3: thread is cooperative, req is overcommit
3900*a325d9c4SApple OSS Distributions 			_wq_cooperative_queue_scheduled_count_dec(wq, old_thread_qos);
3901*a325d9c4SApple OSS Distributions 		} else if (workq_tr_is_nonovercommit(tr_flags)) {
3902*a325d9c4SApple OSS Distributions 			// Case 4: thread is cooperative, req is non-overcommit
3903*a325d9c4SApple OSS Distributions 			_wq_cooperative_queue_scheduled_count_dec(wq, old_thread_qos);
3904*a325d9c4SApple OSS Distributions 			wq->wq_constrained_threads_scheduled++;
3905*a325d9c4SApple OSS Distributions 		} else {
3906*a325d9c4SApple OSS Distributions 			// Case 5: thread is cooperative, req is also cooperative
3907*a325d9c4SApple OSS Distributions 			assert(workq_tr_is_cooperative(tr_flags));
3908*a325d9c4SApple OSS Distributions 			_wq_cooperative_queue_scheduled_count_dec(wq, old_thread_qos);
3909*a325d9c4SApple OSS Distributions 			_wq_cooperative_queue_scheduled_count_inc(wq, new_thread_qos);
3910*a325d9c4SApple OSS Distributions 		}
3911*a325d9c4SApple OSS Distributions 		cooperative_pool_sched_count_changed = true;
3912*a325d9c4SApple OSS Distributions 	} else {
3913*a325d9c4SApple OSS Distributions 		if (workq_tr_is_overcommit(tr_flags)) {
3914*a325d9c4SApple OSS Distributions 			// Case 6: Thread is non-overcommit, req is overcommit
3915*a325d9c4SApple OSS Distributions 			wq->wq_constrained_threads_scheduled--;
3916*a325d9c4SApple OSS Distributions 		} else if (workq_tr_is_cooperative(tr_flags)) {
3917*a325d9c4SApple OSS Distributions 			// Case 7: Thread is non-overcommit, req is cooperative
3918*a325d9c4SApple OSS Distributions 			wq->wq_constrained_threads_scheduled--;
3919*a325d9c4SApple OSS Distributions 			_wq_cooperative_queue_scheduled_count_inc(wq, new_thread_qos);
3920*a325d9c4SApple OSS Distributions 			cooperative_pool_sched_count_changed = true;
3921*a325d9c4SApple OSS Distributions 		}
3922*a325d9c4SApple OSS Distributions 	}
3923*a325d9c4SApple OSS Distributions 
3924*a325d9c4SApple OSS Distributions 	return cooperative_pool_sched_count_changed;
3925*a325d9c4SApple OSS Distributions }
3926*a325d9c4SApple OSS Distributions 
3927*a325d9c4SApple OSS Distributions static workq_threadreq_t
workq_threadreq_select(struct workqueue * wq,struct uthread * uth)3928*a325d9c4SApple OSS Distributions workq_threadreq_select(struct workqueue *wq, struct uthread *uth)
3929*a325d9c4SApple OSS Distributions {
3930*a325d9c4SApple OSS Distributions 	workq_threadreq_t req_qos, req_pri, req_tmp, req_mgr;
3931*a325d9c4SApple OSS Distributions 	uintptr_t proprietor;
3932*a325d9c4SApple OSS Distributions 	thread_qos_t qos = THREAD_QOS_UNSPECIFIED;
3933*a325d9c4SApple OSS Distributions 	uint8_t pri = 0;
3934*a325d9c4SApple OSS Distributions 
3935*a325d9c4SApple OSS Distributions 	if (uth == wq->wq_creator) {
3936*a325d9c4SApple OSS Distributions 		uth = NULL;
3937*a325d9c4SApple OSS Distributions 	}
3938*a325d9c4SApple OSS Distributions 
3939*a325d9c4SApple OSS Distributions 	/*
3940*a325d9c4SApple OSS Distributions 	 * Compute the best priority request (special or turnstile)
3941*a325d9c4SApple OSS Distributions 	 */
3942*a325d9c4SApple OSS Distributions 
3943*a325d9c4SApple OSS Distributions 	pri = (uint8_t)turnstile_workq_proprietor_of_max_turnstile(wq->wq_turnstile,
3944*a325d9c4SApple OSS Distributions 	    &proprietor);
3945*a325d9c4SApple OSS Distributions 	if (pri) {
3946*a325d9c4SApple OSS Distributions 		struct kqworkloop *kqwl = (struct kqworkloop *)proprietor;
3947*a325d9c4SApple OSS Distributions 		req_pri = &kqwl->kqwl_request;
3948*a325d9c4SApple OSS Distributions 		if (req_pri->tr_state != WORKQ_TR_STATE_QUEUED) {
3949*a325d9c4SApple OSS Distributions 			panic("Invalid thread request (%p) state %d",
3950*a325d9c4SApple OSS Distributions 			    req_pri, req_pri->tr_state);
3951*a325d9c4SApple OSS Distributions 		}
3952*a325d9c4SApple OSS Distributions 	} else {
3953*a325d9c4SApple OSS Distributions 		req_pri = NULL;
3954*a325d9c4SApple OSS Distributions 	}
3955*a325d9c4SApple OSS Distributions 
3956*a325d9c4SApple OSS Distributions 	req_tmp = priority_queue_max(&wq->wq_special_queue,
3957*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry);
3958*a325d9c4SApple OSS Distributions 	if (req_tmp && pri < priority_queue_entry_sched_pri(&wq->wq_special_queue,
3959*a325d9c4SApple OSS Distributions 	    &req_tmp->tr_entry)) {
3960*a325d9c4SApple OSS Distributions 		req_pri = req_tmp;
3961*a325d9c4SApple OSS Distributions 		pri = (uint8_t)priority_queue_entry_sched_pri(&wq->wq_special_queue,
3962*a325d9c4SApple OSS Distributions 		    &req_tmp->tr_entry);
3963*a325d9c4SApple OSS Distributions 	}
3964*a325d9c4SApple OSS Distributions 
3965*a325d9c4SApple OSS Distributions 	/*
3966*a325d9c4SApple OSS Distributions 	 * Handle the manager thread request. The special queue might yield
3967*a325d9c4SApple OSS Distributions 	 * a higher priority, but the manager always beats the QoS world.
3968*a325d9c4SApple OSS Distributions 	 */
3969*a325d9c4SApple OSS Distributions 
3970*a325d9c4SApple OSS Distributions 	req_mgr = wq->wq_event_manager_threadreq;
3971*a325d9c4SApple OSS Distributions 	if (req_mgr && workq_may_start_event_mgr_thread(wq, uth)) {
3972*a325d9c4SApple OSS Distributions 		uint32_t mgr_pri = wq->wq_event_manager_priority;
3973*a325d9c4SApple OSS Distributions 
3974*a325d9c4SApple OSS Distributions 		if (mgr_pri & _PTHREAD_PRIORITY_SCHED_PRI_FLAG) {
3975*a325d9c4SApple OSS Distributions 			mgr_pri &= _PTHREAD_PRIORITY_SCHED_PRI_MASK;
3976*a325d9c4SApple OSS Distributions 		} else {
3977*a325d9c4SApple OSS Distributions 			mgr_pri = thread_workq_pri_for_qos(
3978*a325d9c4SApple OSS Distributions 				_pthread_priority_thread_qos(mgr_pri));
3979*a325d9c4SApple OSS Distributions 		}
3980*a325d9c4SApple OSS Distributions 
3981*a325d9c4SApple OSS Distributions 		return mgr_pri >= pri ? req_mgr : req_pri;
3982*a325d9c4SApple OSS Distributions 	}
3983*a325d9c4SApple OSS Distributions 
3984*a325d9c4SApple OSS Distributions 	/*
3985*a325d9c4SApple OSS Distributions 	 * Compute the best QoS Request, and check whether it beats the "pri" one
3986*a325d9c4SApple OSS Distributions 	 */
3987*a325d9c4SApple OSS Distributions 
3988*a325d9c4SApple OSS Distributions 	req_qos = priority_queue_max(&wq->wq_overcommit_queue,
3989*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry);
3990*a325d9c4SApple OSS Distributions 	if (req_qos) {
3991*a325d9c4SApple OSS Distributions 		qos = req_qos->tr_qos;
3992*a325d9c4SApple OSS Distributions 	}
3993*a325d9c4SApple OSS Distributions 
3994*a325d9c4SApple OSS Distributions 	req_tmp = workq_cooperative_queue_best_req(wq, uth);
3995*a325d9c4SApple OSS Distributions 	if (req_tmp && qos <= req_tmp->tr_qos) {
3996*a325d9c4SApple OSS Distributions 		/*
3997*a325d9c4SApple OSS Distributions 		 * Cooperative TR is better between overcommit and cooperative.  Note
3998*a325d9c4SApple OSS Distributions 		 * that if qos is same between overcommit and cooperative, we choose
3999*a325d9c4SApple OSS Distributions 		 * cooperative.
4000*a325d9c4SApple OSS Distributions 		 *
4001*a325d9c4SApple OSS Distributions 		 * Pick cooperative pool if it passes the admissions check
4002*a325d9c4SApple OSS Distributions 		 */
4003*a325d9c4SApple OSS Distributions 		if (workq_cooperative_allowance(wq, req_tmp->tr_qos, uth, true)) {
4004*a325d9c4SApple OSS Distributions 			req_qos = req_tmp;
4005*a325d9c4SApple OSS Distributions 			qos = req_qos->tr_qos;
4006*a325d9c4SApple OSS Distributions 		}
4007*a325d9c4SApple OSS Distributions 	}
4008*a325d9c4SApple OSS Distributions 
4009*a325d9c4SApple OSS Distributions 	/*
4010*a325d9c4SApple OSS Distributions 	 * Compare the best QoS so far - either from overcommit or from cooperative
4011*a325d9c4SApple OSS Distributions 	 * pool - and compare it with the constrained pool
4012*a325d9c4SApple OSS Distributions 	 */
4013*a325d9c4SApple OSS Distributions 	req_tmp = priority_queue_max(&wq->wq_constrained_queue,
4014*a325d9c4SApple OSS Distributions 	    struct workq_threadreq_s, tr_entry);
4015*a325d9c4SApple OSS Distributions 
4016*a325d9c4SApple OSS Distributions 	if (req_tmp && qos < req_tmp->tr_qos) {
4017*a325d9c4SApple OSS Distributions 		/*
4018*a325d9c4SApple OSS Distributions 		 * Constrained pool is best in QoS between overcommit, cooperative
4019*a325d9c4SApple OSS Distributions 		 * and constrained. Now check how it fairs against the priority case
4020*a325d9c4SApple OSS Distributions 		 */
4021*a325d9c4SApple OSS Distributions 		if (pri && pri >= thread_workq_pri_for_qos(req_tmp->tr_qos)) {
4022*a325d9c4SApple OSS Distributions 			return req_pri;
4023*a325d9c4SApple OSS Distributions 		}
4024*a325d9c4SApple OSS Distributions 
4025*a325d9c4SApple OSS Distributions 		if (workq_constrained_allowance(wq, req_tmp->tr_qos, uth, true)) {
4026*a325d9c4SApple OSS Distributions 			/*
4027*a325d9c4SApple OSS Distributions 			 * If the constrained thread request is the best one and passes
4028*a325d9c4SApple OSS Distributions 			 * the admission check, pick it.
4029*a325d9c4SApple OSS Distributions 			 */
4030*a325d9c4SApple OSS Distributions 			return req_tmp;
4031*a325d9c4SApple OSS Distributions 		}
4032*a325d9c4SApple OSS Distributions 	}
4033*a325d9c4SApple OSS Distributions 
4034*a325d9c4SApple OSS Distributions 	if (req_pri && (!qos || pri >= thread_workq_pri_for_qos(qos))) {
4035*a325d9c4SApple OSS Distributions 		return req_pri;
4036*a325d9c4SApple OSS Distributions 	}
4037*a325d9c4SApple OSS Distributions 
4038*a325d9c4SApple OSS Distributions 	return req_qos;
4039*a325d9c4SApple OSS Distributions }
4040*a325d9c4SApple OSS Distributions 
4041*a325d9c4SApple OSS Distributions /*
4042*a325d9c4SApple OSS Distributions  * The creator is an anonymous thread that is counted as scheduled,
4043*a325d9c4SApple OSS Distributions  * but otherwise without its scheduler callback set or tracked as active
4044*a325d9c4SApple OSS Distributions  * that is used to make other threads.
4045*a325d9c4SApple OSS Distributions  *
4046*a325d9c4SApple OSS Distributions  * When more requests are added or an existing one is hurried along,
4047*a325d9c4SApple OSS Distributions  * a creator is elected and setup, or the existing one overridden accordingly.
4048*a325d9c4SApple OSS Distributions  *
4049*a325d9c4SApple OSS Distributions  * While this creator is in flight, because no request has been dequeued,
4050*a325d9c4SApple OSS Distributions  * already running threads have a chance at stealing thread requests avoiding
4051*a325d9c4SApple OSS Distributions  * useless context switches, and the creator once scheduled may not find any
4052*a325d9c4SApple OSS Distributions  * work to do and will then just park again.
4053*a325d9c4SApple OSS Distributions  *
4054*a325d9c4SApple OSS Distributions  * The creator serves the dual purpose of informing the scheduler of work that
4055*a325d9c4SApple OSS Distributions  * hasn't be materialized as threads yet, and also as a natural pacing mechanism
4056*a325d9c4SApple OSS Distributions  * for thread creation.
4057*a325d9c4SApple OSS Distributions  *
4058*a325d9c4SApple OSS Distributions  * By being anonymous (and not bound to anything) it means that thread requests
4059*a325d9c4SApple OSS Distributions  * can be stolen from this creator by threads already on core yielding more
4060*a325d9c4SApple OSS Distributions  * efficient scheduling and reduced context switches.
4061*a325d9c4SApple OSS Distributions  */
4062*a325d9c4SApple OSS Distributions static void
workq_schedule_creator(proc_t p,struct workqueue * wq,workq_kern_threadreq_flags_t flags)4063*a325d9c4SApple OSS Distributions workq_schedule_creator(proc_t p, struct workqueue *wq,
4064*a325d9c4SApple OSS Distributions     workq_kern_threadreq_flags_t flags)
4065*a325d9c4SApple OSS Distributions {
4066*a325d9c4SApple OSS Distributions 	workq_threadreq_t req;
4067*a325d9c4SApple OSS Distributions 	struct uthread *uth;
4068*a325d9c4SApple OSS Distributions 	bool needs_wakeup;
4069*a325d9c4SApple OSS Distributions 
4070*a325d9c4SApple OSS Distributions 	workq_lock_held(wq);
4071*a325d9c4SApple OSS Distributions 	assert(p || (flags & WORKQ_THREADREQ_CAN_CREATE_THREADS) == 0);
4072*a325d9c4SApple OSS Distributions 
4073*a325d9c4SApple OSS Distributions again:
4074*a325d9c4SApple OSS Distributions 	uth = wq->wq_creator;
4075*a325d9c4SApple OSS Distributions 
4076*a325d9c4SApple OSS Distributions 	if (!wq->wq_reqcount) {
4077*a325d9c4SApple OSS Distributions 		/*
4078*a325d9c4SApple OSS Distributions 		 * There is no thread request left.
4079*a325d9c4SApple OSS Distributions 		 *
4080*a325d9c4SApple OSS Distributions 		 * If there is a creator, leave everything in place, so that it cleans
4081*a325d9c4SApple OSS Distributions 		 * up itself in workq_push_idle_thread().
4082*a325d9c4SApple OSS Distributions 		 *
4083*a325d9c4SApple OSS Distributions 		 * Else, make sure the turnstile state is reset to no inheritor.
4084*a325d9c4SApple OSS Distributions 		 */
4085*a325d9c4SApple OSS Distributions 		if (uth == NULL) {
4086*a325d9c4SApple OSS Distributions 			workq_turnstile_update_inheritor(wq, TURNSTILE_INHERITOR_NULL, 0);
4087*a325d9c4SApple OSS Distributions 		}
4088*a325d9c4SApple OSS Distributions 		return;
4089*a325d9c4SApple OSS Distributions 	}
4090*a325d9c4SApple OSS Distributions 
4091*a325d9c4SApple OSS Distributions 	req = workq_threadreq_select_for_creator(wq);
4092*a325d9c4SApple OSS Distributions 	if (req == NULL) {
4093*a325d9c4SApple OSS Distributions 		/*
4094*a325d9c4SApple OSS Distributions 		 * There isn't a thread request that passes the admission check.
4095*a325d9c4SApple OSS Distributions 		 *
4096*a325d9c4SApple OSS Distributions 		 * If there is a creator, do not touch anything, the creator will sort
4097*a325d9c4SApple OSS Distributions 		 * it out when it runs.
4098*a325d9c4SApple OSS Distributions 		 *
4099*a325d9c4SApple OSS Distributions 		 * Else, set the inheritor to "WORKQ" so that the turnstile propagation
4100*a325d9c4SApple OSS Distributions 		 * code calls us if anything changes.
4101*a325d9c4SApple OSS Distributions 		 */
4102*a325d9c4SApple OSS Distributions 		if (uth == NULL) {
4103*a325d9c4SApple OSS Distributions 			workq_turnstile_update_inheritor(wq, wq, TURNSTILE_INHERITOR_WORKQ);
4104*a325d9c4SApple OSS Distributions 		}
4105*a325d9c4SApple OSS Distributions 		return;
4106*a325d9c4SApple OSS Distributions 	}
4107*a325d9c4SApple OSS Distributions 
4108*a325d9c4SApple OSS Distributions 
4109*a325d9c4SApple OSS Distributions 	if (uth) {
4110*a325d9c4SApple OSS Distributions 		/*
4111*a325d9c4SApple OSS Distributions 		 * We need to maybe override the creator we already have
4112*a325d9c4SApple OSS Distributions 		 */
4113*a325d9c4SApple OSS Distributions 		if (workq_thread_needs_priority_change(req, uth)) {
4114*a325d9c4SApple OSS Distributions 			WQ_TRACE_WQ(TRACE_wq_creator_select | DBG_FUNC_NONE,
4115*a325d9c4SApple OSS Distributions 			    wq, 1, uthread_tid(uth), req->tr_qos);
4116*a325d9c4SApple OSS Distributions 			workq_thread_reset_pri(wq, uth, req, /*unpark*/ true);
4117*a325d9c4SApple OSS Distributions 		}
4118*a325d9c4SApple OSS Distributions 		assert(wq->wq_inheritor == get_machthread(uth));
4119*a325d9c4SApple OSS Distributions 	} else if (wq->wq_thidlecount) {
4120*a325d9c4SApple OSS Distributions 		/*
4121*a325d9c4SApple OSS Distributions 		 * We need to unpark a creator thread
4122*a325d9c4SApple OSS Distributions 		 */
4123*a325d9c4SApple OSS Distributions 		wq->wq_creator = uth = workq_pop_idle_thread(wq, UT_WORKQ_OVERCOMMIT,
4124*a325d9c4SApple OSS Distributions 		    &needs_wakeup);
4125*a325d9c4SApple OSS Distributions 		/* Always reset the priorities on the newly chosen creator */
4126*a325d9c4SApple OSS Distributions 		workq_thread_reset_pri(wq, uth, req, /*unpark*/ true);
4127*a325d9c4SApple OSS Distributions 		workq_turnstile_update_inheritor(wq, get_machthread(uth),
4128*a325d9c4SApple OSS Distributions 		    TURNSTILE_INHERITOR_THREAD);
4129*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_creator_select | DBG_FUNC_NONE,
4130*a325d9c4SApple OSS Distributions 		    wq, 2, uthread_tid(uth), req->tr_qos);
4131*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.fulfilled_snapshot = wq->wq_fulfilled;
4132*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.yields = 0;
4133*a325d9c4SApple OSS Distributions 		if (needs_wakeup) {
4134*a325d9c4SApple OSS Distributions 			workq_thread_wakeup(uth);
4135*a325d9c4SApple OSS Distributions 		}
4136*a325d9c4SApple OSS Distributions 	} else {
4137*a325d9c4SApple OSS Distributions 		/*
4138*a325d9c4SApple OSS Distributions 		 * We need to allocate a thread...
4139*a325d9c4SApple OSS Distributions 		 */
4140*a325d9c4SApple OSS Distributions 		if (__improbable(wq->wq_nthreads >= wq_max_threads)) {
4141*a325d9c4SApple OSS Distributions 			/* out of threads, just go away */
4142*a325d9c4SApple OSS Distributions 			flags = WORKQ_THREADREQ_NONE;
4143*a325d9c4SApple OSS Distributions 		} else if (flags & WORKQ_THREADREQ_SET_AST_ON_FAILURE) {
4144*a325d9c4SApple OSS Distributions 			act_set_astkevent(current_thread(), AST_KEVENT_REDRIVE_THREADREQ);
4145*a325d9c4SApple OSS Distributions 		} else if (!(flags & WORKQ_THREADREQ_CAN_CREATE_THREADS)) {
4146*a325d9c4SApple OSS Distributions 			/* This can drop the workqueue lock, and take it again */
4147*a325d9c4SApple OSS Distributions 			workq_schedule_immediate_thread_creation(wq);
4148*a325d9c4SApple OSS Distributions 		} else if (workq_add_new_idle_thread(p, wq)) {
4149*a325d9c4SApple OSS Distributions 			goto again;
4150*a325d9c4SApple OSS Distributions 		} else {
4151*a325d9c4SApple OSS Distributions 			workq_schedule_delayed_thread_creation(wq, 0);
4152*a325d9c4SApple OSS Distributions 		}
4153*a325d9c4SApple OSS Distributions 
4154*a325d9c4SApple OSS Distributions 		/*
4155*a325d9c4SApple OSS Distributions 		 * If the current thread is the inheritor:
4156*a325d9c4SApple OSS Distributions 		 *
4157*a325d9c4SApple OSS Distributions 		 * If we set the AST, then the thread will stay the inheritor until
4158*a325d9c4SApple OSS Distributions 		 * either the AST calls workq_kern_threadreq_redrive(), or it parks
4159*a325d9c4SApple OSS Distributions 		 * and calls workq_push_idle_thread().
4160*a325d9c4SApple OSS Distributions 		 *
4161*a325d9c4SApple OSS Distributions 		 * Else, the responsibility of the thread creation is with a thread-call
4162*a325d9c4SApple OSS Distributions 		 * and we need to clear the inheritor.
4163*a325d9c4SApple OSS Distributions 		 */
4164*a325d9c4SApple OSS Distributions 		if ((flags & WORKQ_THREADREQ_SET_AST_ON_FAILURE) == 0 &&
4165*a325d9c4SApple OSS Distributions 		    wq->wq_inheritor == current_thread()) {
4166*a325d9c4SApple OSS Distributions 			workq_turnstile_update_inheritor(wq, TURNSTILE_INHERITOR_NULL, 0);
4167*a325d9c4SApple OSS Distributions 		}
4168*a325d9c4SApple OSS Distributions 	}
4169*a325d9c4SApple OSS Distributions }
4170*a325d9c4SApple OSS Distributions 
4171*a325d9c4SApple OSS Distributions /**
4172*a325d9c4SApple OSS Distributions  * Same as workq_unpark_select_threadreq_or_park_and_unlock,
4173*a325d9c4SApple OSS Distributions  * but do not allow early binds.
4174*a325d9c4SApple OSS Distributions  *
4175*a325d9c4SApple OSS Distributions  * Called with the base pri frozen, will unfreeze it.
4176*a325d9c4SApple OSS Distributions  */
4177*a325d9c4SApple OSS Distributions __attribute__((noreturn, noinline))
4178*a325d9c4SApple OSS Distributions static void
workq_select_threadreq_or_park_and_unlock(proc_t p,struct workqueue * wq,struct uthread * uth,uint32_t setup_flags)4179*a325d9c4SApple OSS Distributions workq_select_threadreq_or_park_and_unlock(proc_t p, struct workqueue *wq,
4180*a325d9c4SApple OSS Distributions     struct uthread *uth, uint32_t setup_flags)
4181*a325d9c4SApple OSS Distributions {
4182*a325d9c4SApple OSS Distributions 	workq_threadreq_t req = NULL;
4183*a325d9c4SApple OSS Distributions 	bool is_creator = (wq->wq_creator == uth);
4184*a325d9c4SApple OSS Distributions 	bool schedule_creator = false;
4185*a325d9c4SApple OSS Distributions 
4186*a325d9c4SApple OSS Distributions 	if (__improbable(_wq_exiting(wq))) {
4187*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_select_threadreq | DBG_FUNC_NONE, wq, 0, 0, 0);
4188*a325d9c4SApple OSS Distributions 		goto park;
4189*a325d9c4SApple OSS Distributions 	}
4190*a325d9c4SApple OSS Distributions 
4191*a325d9c4SApple OSS Distributions 	if (wq->wq_reqcount == 0) {
4192*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_select_threadreq | DBG_FUNC_NONE, wq, 1, 0, 0);
4193*a325d9c4SApple OSS Distributions 		goto park;
4194*a325d9c4SApple OSS Distributions 	}
4195*a325d9c4SApple OSS Distributions 
4196*a325d9c4SApple OSS Distributions 	req = workq_threadreq_select(wq, uth);
4197*a325d9c4SApple OSS Distributions 	if (__improbable(req == NULL)) {
4198*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_select_threadreq | DBG_FUNC_NONE, wq, 2, 0, 0);
4199*a325d9c4SApple OSS Distributions 		goto park;
4200*a325d9c4SApple OSS Distributions 	}
4201*a325d9c4SApple OSS Distributions 
4202*a325d9c4SApple OSS Distributions 	thread_qos_t old_thread_bucket = uth->uu_workq_pri.qos_bucket;
4203*a325d9c4SApple OSS Distributions 	uint8_t tr_flags = req->tr_flags;
4204*a325d9c4SApple OSS Distributions 	struct turnstile *req_ts = kqueue_threadreq_get_turnstile(req);
4205*a325d9c4SApple OSS Distributions 
4206*a325d9c4SApple OSS Distributions 	/*
4207*a325d9c4SApple OSS Distributions 	 * Attempt to setup ourselves as the new thing to run, moving all priority
4208*a325d9c4SApple OSS Distributions 	 * pushes to ourselves.
4209*a325d9c4SApple OSS Distributions 	 *
4210*a325d9c4SApple OSS Distributions 	 * If the current thread is the creator, then the fact that we are presently
4211*a325d9c4SApple OSS Distributions 	 * running is proof that we'll do something useful, so keep going.
4212*a325d9c4SApple OSS Distributions 	 *
4213*a325d9c4SApple OSS Distributions 	 * For other cases, peek at the AST to know whether the scheduler wants
4214*a325d9c4SApple OSS Distributions 	 * to preempt us, if yes, park instead, and move the thread request
4215*a325d9c4SApple OSS Distributions 	 * turnstile back to the workqueue.
4216*a325d9c4SApple OSS Distributions 	 */
4217*a325d9c4SApple OSS Distributions 	if (req_ts) {
4218*a325d9c4SApple OSS Distributions 		workq_perform_turnstile_operation_locked(wq, ^{
4219*a325d9c4SApple OSS Distributions 			turnstile_update_inheritor(req_ts, get_machthread(uth),
4220*a325d9c4SApple OSS Distributions 			TURNSTILE_IMMEDIATE_UPDATE | TURNSTILE_INHERITOR_THREAD);
4221*a325d9c4SApple OSS Distributions 			turnstile_update_inheritor_complete(req_ts,
4222*a325d9c4SApple OSS Distributions 			TURNSTILE_INTERLOCK_HELD);
4223*a325d9c4SApple OSS Distributions 		});
4224*a325d9c4SApple OSS Distributions 	}
4225*a325d9c4SApple OSS Distributions 
4226*a325d9c4SApple OSS Distributions 	/* accounting changes of aggregate thscheduled_count and thactive which has
4227*a325d9c4SApple OSS Distributions 	 * to be paired with the workq_thread_reset_pri below so that we have
4228*a325d9c4SApple OSS Distributions 	 * uth->uu_workq_pri match with thactive.
4229*a325d9c4SApple OSS Distributions 	 *
4230*a325d9c4SApple OSS Distributions 	 * This is undone when the thread parks */
4231*a325d9c4SApple OSS Distributions 	if (is_creator) {
4232*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_creator_select, wq, 4, 0,
4233*a325d9c4SApple OSS Distributions 		    uth->uu_save.uus_workq_park_data.yields);
4234*a325d9c4SApple OSS Distributions 		wq->wq_creator = NULL;
4235*a325d9c4SApple OSS Distributions 		_wq_thactive_inc(wq, req->tr_qos);
4236*a325d9c4SApple OSS Distributions 		wq->wq_thscheduled_count[_wq_bucket(req->tr_qos)]++;
4237*a325d9c4SApple OSS Distributions 	} else if (old_thread_bucket != req->tr_qos) {
4238*a325d9c4SApple OSS Distributions 		_wq_thactive_move(wq, old_thread_bucket, req->tr_qos);
4239*a325d9c4SApple OSS Distributions 	}
4240*a325d9c4SApple OSS Distributions 	workq_thread_reset_pri(wq, uth, req, /*unpark*/ true);
4241*a325d9c4SApple OSS Distributions 
4242*a325d9c4SApple OSS Distributions 	/*
4243*a325d9c4SApple OSS Distributions 	 * Make relevant accounting changes for pool specific counts.
4244*a325d9c4SApple OSS Distributions 	 *
4245*a325d9c4SApple OSS Distributions 	 * The schedule counts changing can affect what the next best request
4246*a325d9c4SApple OSS Distributions 	 * for cooperative thread pool is if this request is dequeued.
4247*a325d9c4SApple OSS Distributions 	 */
4248*a325d9c4SApple OSS Distributions 	bool cooperative_sched_count_changed =
4249*a325d9c4SApple OSS Distributions 	    workq_adjust_cooperative_constrained_schedule_counts(wq, uth,
4250*a325d9c4SApple OSS Distributions 	    old_thread_bucket, tr_flags);
4251*a325d9c4SApple OSS Distributions 
4252*a325d9c4SApple OSS Distributions 	if (workq_tr_is_overcommit(tr_flags)) {
4253*a325d9c4SApple OSS Distributions 		workq_thread_set_type(uth, UT_WORKQ_OVERCOMMIT);
4254*a325d9c4SApple OSS Distributions 	} else if (workq_tr_is_cooperative(tr_flags)) {
4255*a325d9c4SApple OSS Distributions 		workq_thread_set_type(uth, UT_WORKQ_COOPERATIVE);
4256*a325d9c4SApple OSS Distributions 	} else {
4257*a325d9c4SApple OSS Distributions 		workq_thread_set_type(uth, 0);
4258*a325d9c4SApple OSS Distributions 	}
4259*a325d9c4SApple OSS Distributions 
4260*a325d9c4SApple OSS Distributions 	if (__improbable(thread_unfreeze_base_pri(get_machthread(uth)) && !is_creator)) {
4261*a325d9c4SApple OSS Distributions 		if (req_ts) {
4262*a325d9c4SApple OSS Distributions 			workq_perform_turnstile_operation_locked(wq, ^{
4263*a325d9c4SApple OSS Distributions 				turnstile_update_inheritor(req_ts, wq->wq_turnstile,
4264*a325d9c4SApple OSS Distributions 				TURNSTILE_IMMEDIATE_UPDATE | TURNSTILE_INHERITOR_TURNSTILE);
4265*a325d9c4SApple OSS Distributions 				turnstile_update_inheritor_complete(req_ts,
4266*a325d9c4SApple OSS Distributions 				TURNSTILE_INTERLOCK_HELD);
4267*a325d9c4SApple OSS Distributions 			});
4268*a325d9c4SApple OSS Distributions 		}
4269*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_select_threadreq | DBG_FUNC_NONE, wq, 3, 0, 0);
4270*a325d9c4SApple OSS Distributions 		goto park_thawed;
4271*a325d9c4SApple OSS Distributions 	}
4272*a325d9c4SApple OSS Distributions 
4273*a325d9c4SApple OSS Distributions 	/*
4274*a325d9c4SApple OSS Distributions 	 * We passed all checks, dequeue the request, bind to it, and set it up
4275*a325d9c4SApple OSS Distributions 	 * to return to user.
4276*a325d9c4SApple OSS Distributions 	 */
4277*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_thread_logical_run | DBG_FUNC_START, wq,
4278*a325d9c4SApple OSS Distributions 	    workq_trace_req_id(req), tr_flags, 0);
4279*a325d9c4SApple OSS Distributions 	wq->wq_fulfilled++;
4280*a325d9c4SApple OSS Distributions 	schedule_creator = workq_threadreq_dequeue(wq, req,
4281*a325d9c4SApple OSS Distributions 	    cooperative_sched_count_changed);
4282*a325d9c4SApple OSS Distributions 
4283*a325d9c4SApple OSS Distributions 	workq_thread_reset_cpupercent(req, uth);
4284*a325d9c4SApple OSS Distributions 
4285*a325d9c4SApple OSS Distributions 	if (tr_flags & (WORKQ_TR_FLAG_KEVENT | WORKQ_TR_FLAG_WORKLOOP)) {
4286*a325d9c4SApple OSS Distributions 		kqueue_threadreq_bind_prepost(p, req, uth);
4287*a325d9c4SApple OSS Distributions 		req = NULL;
4288*a325d9c4SApple OSS Distributions 	} else if (req->tr_count > 0) {
4289*a325d9c4SApple OSS Distributions 		req = NULL;
4290*a325d9c4SApple OSS Distributions 	}
4291*a325d9c4SApple OSS Distributions 
4292*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_NEW) {
4293*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags ^= UT_WORKQ_NEW;
4294*a325d9c4SApple OSS Distributions 		setup_flags |= WQ_SETUP_FIRST_USE;
4295*a325d9c4SApple OSS Distributions 	}
4296*a325d9c4SApple OSS Distributions 
4297*a325d9c4SApple OSS Distributions 	/* If one of the following is true, call workq_schedule_creator (which also
4298*a325d9c4SApple OSS Distributions 	 * adjusts priority of existing creator):
4299*a325d9c4SApple OSS Distributions 	 *
4300*a325d9c4SApple OSS Distributions 	 *	  - We are the creator currently so the wq may need a new creator
4301*a325d9c4SApple OSS Distributions 	 *	  - The request we're binding to is the highest priority one, existing
4302*a325d9c4SApple OSS Distributions 	 *	  creator's priority might need to be adjusted to reflect the next
4303*a325d9c4SApple OSS Distributions 	 *	  highest TR
4304*a325d9c4SApple OSS Distributions 	 */
4305*a325d9c4SApple OSS Distributions 	if (is_creator || schedule_creator) {
4306*a325d9c4SApple OSS Distributions 		/* This can drop the workqueue lock, and take it again */
4307*a325d9c4SApple OSS Distributions 		workq_schedule_creator(p, wq, WORKQ_THREADREQ_CAN_CREATE_THREADS);
4308*a325d9c4SApple OSS Distributions 	}
4309*a325d9c4SApple OSS Distributions 
4310*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
4311*a325d9c4SApple OSS Distributions 
4312*a325d9c4SApple OSS Distributions 	if (req) {
4313*a325d9c4SApple OSS Distributions 		zfree(workq_zone_threadreq, req);
4314*a325d9c4SApple OSS Distributions 	}
4315*a325d9c4SApple OSS Distributions 
4316*a325d9c4SApple OSS Distributions 	/*
4317*a325d9c4SApple OSS Distributions 	 * Run Thread, Run!
4318*a325d9c4SApple OSS Distributions 	 */
4319*a325d9c4SApple OSS Distributions 	uint32_t upcall_flags = WQ_FLAG_THREAD_NEWSPI;
4320*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_pri.qos_bucket == WORKQ_THREAD_QOS_MANAGER) {
4321*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_EVENT_MANAGER;
4322*a325d9c4SApple OSS Distributions 	} else if (workq_tr_is_overcommit(tr_flags)) {
4323*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_OVERCOMMIT;
4324*a325d9c4SApple OSS Distributions 	} else if (workq_tr_is_cooperative(tr_flags)) {
4325*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_COOPERATIVE;
4326*a325d9c4SApple OSS Distributions 	}
4327*a325d9c4SApple OSS Distributions 	if (tr_flags & WORKQ_TR_FLAG_KEVENT) {
4328*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_KEVENT;
4329*a325d9c4SApple OSS Distributions 		assert((upcall_flags & WQ_FLAG_THREAD_COOPERATIVE) == 0);
4330*a325d9c4SApple OSS Distributions 	}
4331*a325d9c4SApple OSS Distributions 
4332*a325d9c4SApple OSS Distributions 	if (tr_flags & WORKQ_TR_FLAG_WORKLOOP) {
4333*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_WORKLOOP | WQ_FLAG_THREAD_KEVENT;
4334*a325d9c4SApple OSS Distributions 	}
4335*a325d9c4SApple OSS Distributions 	uth->uu_save.uus_workq_park_data.upcall_flags = upcall_flags;
4336*a325d9c4SApple OSS Distributions 
4337*a325d9c4SApple OSS Distributions 	if (tr_flags & (WORKQ_TR_FLAG_KEVENT | WORKQ_TR_FLAG_WORKLOOP)) {
4338*a325d9c4SApple OSS Distributions 		kqueue_threadreq_bind_commit(p, get_machthread(uth));
4339*a325d9c4SApple OSS Distributions 	} else {
4340*a325d9c4SApple OSS Distributions #if CONFIG_PREADOPT_TG
4341*a325d9c4SApple OSS Distributions 		/*
4342*a325d9c4SApple OSS Distributions 		 * The thread may have a preadopt thread group on it already because it
4343*a325d9c4SApple OSS Distributions 		 * got tagged with it as a creator thread. So we need to make sure to
4344*a325d9c4SApple OSS Distributions 		 * clear that since we don't have preadoption for anonymous thread
4345*a325d9c4SApple OSS Distributions 		 * requests
4346*a325d9c4SApple OSS Distributions 		 */
4347*a325d9c4SApple OSS Distributions 		thread_set_preadopt_thread_group(get_machthread(uth), NULL);
4348*a325d9c4SApple OSS Distributions #endif
4349*a325d9c4SApple OSS Distributions 	}
4350*a325d9c4SApple OSS Distributions 
4351*a325d9c4SApple OSS Distributions 	workq_setup_and_run(p, uth, setup_flags);
4352*a325d9c4SApple OSS Distributions 	__builtin_unreachable();
4353*a325d9c4SApple OSS Distributions 
4354*a325d9c4SApple OSS Distributions park:
4355*a325d9c4SApple OSS Distributions 	thread_unfreeze_base_pri(get_machthread(uth));
4356*a325d9c4SApple OSS Distributions park_thawed:
4357*a325d9c4SApple OSS Distributions 	workq_park_and_unlock(p, wq, uth, setup_flags);
4358*a325d9c4SApple OSS Distributions }
4359*a325d9c4SApple OSS Distributions 
4360*a325d9c4SApple OSS Distributions /**
4361*a325d9c4SApple OSS Distributions  * Runs a thread request on a thread
4362*a325d9c4SApple OSS Distributions  *
4363*a325d9c4SApple OSS Distributions  * - if thread is THREAD_NULL, will find a thread and run the request there.
4364*a325d9c4SApple OSS Distributions  *   Otherwise, the thread must be the current thread.
4365*a325d9c4SApple OSS Distributions  *
4366*a325d9c4SApple OSS Distributions  * - if req is NULL, will find the highest priority request and run that.  If
4367*a325d9c4SApple OSS Distributions  *   it is not NULL, it must be a threadreq object in state NEW.  If it can not
4368*a325d9c4SApple OSS Distributions  *   be run immediately, it will be enqueued and moved to state QUEUED.
4369*a325d9c4SApple OSS Distributions  *
4370*a325d9c4SApple OSS Distributions  *   Either way, the thread request object serviced will be moved to state
4371*a325d9c4SApple OSS Distributions  *   BINDING and attached to the uthread.
4372*a325d9c4SApple OSS Distributions  *
4373*a325d9c4SApple OSS Distributions  * Should be called with the workqueue lock held.  Will drop it.
4374*a325d9c4SApple OSS Distributions  * Should be called with the base pri not frozen.
4375*a325d9c4SApple OSS Distributions  */
4376*a325d9c4SApple OSS Distributions __attribute__((noreturn, noinline))
4377*a325d9c4SApple OSS Distributions static void
workq_unpark_select_threadreq_or_park_and_unlock(proc_t p,struct workqueue * wq,struct uthread * uth,uint32_t setup_flags)4378*a325d9c4SApple OSS Distributions workq_unpark_select_threadreq_or_park_and_unlock(proc_t p, struct workqueue *wq,
4379*a325d9c4SApple OSS Distributions     struct uthread *uth, uint32_t setup_flags)
4380*a325d9c4SApple OSS Distributions {
4381*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_EARLY_BOUND) {
4382*a325d9c4SApple OSS Distributions 		if (uth->uu_workq_flags & UT_WORKQ_NEW) {
4383*a325d9c4SApple OSS Distributions 			setup_flags |= WQ_SETUP_FIRST_USE;
4384*a325d9c4SApple OSS Distributions 		}
4385*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags &= ~(UT_WORKQ_NEW | UT_WORKQ_EARLY_BOUND);
4386*a325d9c4SApple OSS Distributions 		/*
4387*a325d9c4SApple OSS Distributions 		 * This pointer is possibly freed and only used for tracing purposes.
4388*a325d9c4SApple OSS Distributions 		 */
4389*a325d9c4SApple OSS Distributions 		workq_threadreq_t req = uth->uu_save.uus_workq_park_data.thread_request;
4390*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
4391*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_thread_logical_run | DBG_FUNC_START, wq,
4392*a325d9c4SApple OSS Distributions 		    VM_KERNEL_ADDRHIDE(req), 0, 0);
4393*a325d9c4SApple OSS Distributions 		(void)req;
4394*a325d9c4SApple OSS Distributions 
4395*a325d9c4SApple OSS Distributions 		workq_setup_and_run(p, uth, setup_flags);
4396*a325d9c4SApple OSS Distributions 		__builtin_unreachable();
4397*a325d9c4SApple OSS Distributions 	}
4398*a325d9c4SApple OSS Distributions 
4399*a325d9c4SApple OSS Distributions 	thread_freeze_base_pri(get_machthread(uth));
4400*a325d9c4SApple OSS Distributions 	workq_select_threadreq_or_park_and_unlock(p, wq, uth, setup_flags);
4401*a325d9c4SApple OSS Distributions }
4402*a325d9c4SApple OSS Distributions 
4403*a325d9c4SApple OSS Distributions static bool
workq_creator_should_yield(struct workqueue * wq,struct uthread * uth)4404*a325d9c4SApple OSS Distributions workq_creator_should_yield(struct workqueue *wq, struct uthread *uth)
4405*a325d9c4SApple OSS Distributions {
4406*a325d9c4SApple OSS Distributions 	thread_qos_t qos = workq_pri_override(uth->uu_workq_pri);
4407*a325d9c4SApple OSS Distributions 
4408*a325d9c4SApple OSS Distributions 	if (qos >= THREAD_QOS_USER_INTERACTIVE) {
4409*a325d9c4SApple OSS Distributions 		return false;
4410*a325d9c4SApple OSS Distributions 	}
4411*a325d9c4SApple OSS Distributions 
4412*a325d9c4SApple OSS Distributions 	uint32_t snapshot = uth->uu_save.uus_workq_park_data.fulfilled_snapshot;
4413*a325d9c4SApple OSS Distributions 	if (wq->wq_fulfilled == snapshot) {
4414*a325d9c4SApple OSS Distributions 		return false;
4415*a325d9c4SApple OSS Distributions 	}
4416*a325d9c4SApple OSS Distributions 
4417*a325d9c4SApple OSS Distributions 	uint32_t cnt = 0, conc = wq_max_parallelism[_wq_bucket(qos)];
4418*a325d9c4SApple OSS Distributions 	if (wq->wq_fulfilled - snapshot > conc) {
4419*a325d9c4SApple OSS Distributions 		/* we fulfilled more than NCPU requests since being dispatched */
4420*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_creator_yield, wq, 1,
4421*a325d9c4SApple OSS Distributions 		    wq->wq_fulfilled, snapshot);
4422*a325d9c4SApple OSS Distributions 		return true;
4423*a325d9c4SApple OSS Distributions 	}
4424*a325d9c4SApple OSS Distributions 
4425*a325d9c4SApple OSS Distributions 	for (int i = _wq_bucket(qos); i < WORKQ_NUM_QOS_BUCKETS; i++) {
4426*a325d9c4SApple OSS Distributions 		cnt += wq->wq_thscheduled_count[i];
4427*a325d9c4SApple OSS Distributions 	}
4428*a325d9c4SApple OSS Distributions 	if (conc <= cnt) {
4429*a325d9c4SApple OSS Distributions 		/* We fulfilled requests and have more than NCPU scheduled threads */
4430*a325d9c4SApple OSS Distributions 		WQ_TRACE_WQ(TRACE_wq_creator_yield, wq, 2,
4431*a325d9c4SApple OSS Distributions 		    wq->wq_fulfilled, snapshot);
4432*a325d9c4SApple OSS Distributions 		return true;
4433*a325d9c4SApple OSS Distributions 	}
4434*a325d9c4SApple OSS Distributions 
4435*a325d9c4SApple OSS Distributions 	return false;
4436*a325d9c4SApple OSS Distributions }
4437*a325d9c4SApple OSS Distributions 
4438*a325d9c4SApple OSS Distributions /**
4439*a325d9c4SApple OSS Distributions  * parked thread wakes up
4440*a325d9c4SApple OSS Distributions  */
4441*a325d9c4SApple OSS Distributions __attribute__((noreturn, noinline))
4442*a325d9c4SApple OSS Distributions static void
workq_unpark_continue(void * parameter __unused,wait_result_t wr __unused)4443*a325d9c4SApple OSS Distributions workq_unpark_continue(void *parameter __unused, wait_result_t wr __unused)
4444*a325d9c4SApple OSS Distributions {
4445*a325d9c4SApple OSS Distributions 	thread_t th = current_thread();
4446*a325d9c4SApple OSS Distributions 	struct uthread *uth = get_bsdthread_info(th);
4447*a325d9c4SApple OSS Distributions 	proc_t p = current_proc();
4448*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr_fast(p);
4449*a325d9c4SApple OSS Distributions 
4450*a325d9c4SApple OSS Distributions 	workq_lock_spin(wq);
4451*a325d9c4SApple OSS Distributions 
4452*a325d9c4SApple OSS Distributions 	if (wq->wq_creator == uth && workq_creator_should_yield(wq, uth)) {
4453*a325d9c4SApple OSS Distributions 		/*
4454*a325d9c4SApple OSS Distributions 		 * If the number of threads we have out are able to keep up with the
4455*a325d9c4SApple OSS Distributions 		 * demand, then we should avoid sending this creator thread to
4456*a325d9c4SApple OSS Distributions 		 * userspace.
4457*a325d9c4SApple OSS Distributions 		 */
4458*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.fulfilled_snapshot = wq->wq_fulfilled;
4459*a325d9c4SApple OSS Distributions 		uth->uu_save.uus_workq_park_data.yields++;
4460*a325d9c4SApple OSS Distributions 		workq_unlock(wq);
4461*a325d9c4SApple OSS Distributions 		thread_yield_with_continuation(workq_unpark_continue, NULL);
4462*a325d9c4SApple OSS Distributions 		__builtin_unreachable();
4463*a325d9c4SApple OSS Distributions 	}
4464*a325d9c4SApple OSS Distributions 
4465*a325d9c4SApple OSS Distributions 	if (__probable(uth->uu_workq_flags & UT_WORKQ_RUNNING)) {
4466*a325d9c4SApple OSS Distributions 		workq_unpark_select_threadreq_or_park_and_unlock(p, wq, uth, WQ_SETUP_NONE);
4467*a325d9c4SApple OSS Distributions 		__builtin_unreachable();
4468*a325d9c4SApple OSS Distributions 	}
4469*a325d9c4SApple OSS Distributions 
4470*a325d9c4SApple OSS Distributions 	if (__probable(wr == THREAD_AWAKENED)) {
4471*a325d9c4SApple OSS Distributions 		/*
4472*a325d9c4SApple OSS Distributions 		 * We were set running, but for the purposes of dying.
4473*a325d9c4SApple OSS Distributions 		 */
4474*a325d9c4SApple OSS Distributions 		assert(uth->uu_workq_flags & UT_WORKQ_DYING);
4475*a325d9c4SApple OSS Distributions 		assert((uth->uu_workq_flags & UT_WORKQ_NEW) == 0);
4476*a325d9c4SApple OSS Distributions 	} else {
4477*a325d9c4SApple OSS Distributions 		/*
4478*a325d9c4SApple OSS Distributions 		 * workaround for <rdar://problem/38647347>,
4479*a325d9c4SApple OSS Distributions 		 * in case we do hit userspace, make sure calling
4480*a325d9c4SApple OSS Distributions 		 * workq_thread_terminate() does the right thing here,
4481*a325d9c4SApple OSS Distributions 		 * and if we never call it, that workq_exit() will too because it sees
4482*a325d9c4SApple OSS Distributions 		 * this thread on the runlist.
4483*a325d9c4SApple OSS Distributions 		 */
4484*a325d9c4SApple OSS Distributions 		assert(wr == THREAD_INTERRUPTED);
4485*a325d9c4SApple OSS Distributions 		wq->wq_thdying_count++;
4486*a325d9c4SApple OSS Distributions 		uth->uu_workq_flags |= UT_WORKQ_DYING;
4487*a325d9c4SApple OSS Distributions 	}
4488*a325d9c4SApple OSS Distributions 
4489*a325d9c4SApple OSS Distributions 	workq_unpark_for_death_and_unlock(p, wq, uth,
4490*a325d9c4SApple OSS Distributions 	    WORKQ_UNPARK_FOR_DEATH_WAS_IDLE, WQ_SETUP_NONE);
4491*a325d9c4SApple OSS Distributions 	__builtin_unreachable();
4492*a325d9c4SApple OSS Distributions }
4493*a325d9c4SApple OSS Distributions 
4494*a325d9c4SApple OSS Distributions __attribute__((noreturn, noinline))
4495*a325d9c4SApple OSS Distributions static void
workq_setup_and_run(proc_t p,struct uthread * uth,int setup_flags)4496*a325d9c4SApple OSS Distributions workq_setup_and_run(proc_t p, struct uthread *uth, int setup_flags)
4497*a325d9c4SApple OSS Distributions {
4498*a325d9c4SApple OSS Distributions 	thread_t th = get_machthread(uth);
4499*a325d9c4SApple OSS Distributions 	vm_map_t vmap = get_task_map(p->task);
4500*a325d9c4SApple OSS Distributions 
4501*a325d9c4SApple OSS Distributions 	if (setup_flags & WQ_SETUP_CLEAR_VOUCHER) {
4502*a325d9c4SApple OSS Distributions 		/*
4503*a325d9c4SApple OSS Distributions 		 * For preemption reasons, we want to reset the voucher as late as
4504*a325d9c4SApple OSS Distributions 		 * possible, so we do it in two places:
4505*a325d9c4SApple OSS Distributions 		 *   - Just before parking (i.e. in workq_park_and_unlock())
4506*a325d9c4SApple OSS Distributions 		 *   - Prior to doing the setup for the next workitem (i.e. here)
4507*a325d9c4SApple OSS Distributions 		 *
4508*a325d9c4SApple OSS Distributions 		 * Those two places are sufficient to ensure we always reset it before
4509*a325d9c4SApple OSS Distributions 		 * it goes back out to user space, but be careful to not break that
4510*a325d9c4SApple OSS Distributions 		 * guarantee.
4511*a325d9c4SApple OSS Distributions 		 *
4512*a325d9c4SApple OSS Distributions 		 * Note that setting the voucher to NULL will not clear the preadoption
4513*a325d9c4SApple OSS Distributions 		 * thread group on this thread
4514*a325d9c4SApple OSS Distributions 		 */
4515*a325d9c4SApple OSS Distributions 		__assert_only kern_return_t kr;
4516*a325d9c4SApple OSS Distributions 		kr = thread_set_voucher_name(MACH_PORT_NULL);
4517*a325d9c4SApple OSS Distributions 		assert(kr == KERN_SUCCESS);
4518*a325d9c4SApple OSS Distributions 	}
4519*a325d9c4SApple OSS Distributions 
4520*a325d9c4SApple OSS Distributions 	uint32_t upcall_flags = uth->uu_save.uus_workq_park_data.upcall_flags;
4521*a325d9c4SApple OSS Distributions 	if (!(setup_flags & WQ_SETUP_FIRST_USE)) {
4522*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_REUSE;
4523*a325d9c4SApple OSS Distributions 	}
4524*a325d9c4SApple OSS Distributions 
4525*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_flags & UT_WORKQ_OUTSIDE_QOS) {
4526*a325d9c4SApple OSS Distributions 		/*
4527*a325d9c4SApple OSS Distributions 		 * For threads that have an outside-of-QoS thread priority, indicate
4528*a325d9c4SApple OSS Distributions 		 * to userspace that setting QoS should only affect the TSD and not
4529*a325d9c4SApple OSS Distributions 		 * change QOS in the kernel.
4530*a325d9c4SApple OSS Distributions 		 */
4531*a325d9c4SApple OSS Distributions 		upcall_flags |= WQ_FLAG_THREAD_OUTSIDEQOS;
4532*a325d9c4SApple OSS Distributions 	} else {
4533*a325d9c4SApple OSS Distributions 		/*
4534*a325d9c4SApple OSS Distributions 		 * Put the QoS class value into the lower bits of the reuse_thread
4535*a325d9c4SApple OSS Distributions 		 * register, this is where the thread priority used to be stored
4536*a325d9c4SApple OSS Distributions 		 * anyway.
4537*a325d9c4SApple OSS Distributions 		 */
4538*a325d9c4SApple OSS Distributions 		upcall_flags |= uth->uu_save.uus_workq_park_data.qos |
4539*a325d9c4SApple OSS Distributions 		    WQ_FLAG_THREAD_PRIO_QOS;
4540*a325d9c4SApple OSS Distributions 	}
4541*a325d9c4SApple OSS Distributions 
4542*a325d9c4SApple OSS Distributions 	if (uth->uu_workq_thport == MACH_PORT_NULL) {
4543*a325d9c4SApple OSS Distributions 		/* convert_thread_to_port_pinned() consumes a reference */
4544*a325d9c4SApple OSS Distributions 		thread_reference(th);
4545*a325d9c4SApple OSS Distributions 		/* Convert to immovable/pinned thread port, but port is not pinned yet */
4546*a325d9c4SApple OSS Distributions 		ipc_port_t port = convert_thread_to_port_pinned(th);
4547*a325d9c4SApple OSS Distributions 		/* Atomically, pin and copy out the port */
4548*a325d9c4SApple OSS Distributions 		uth->uu_workq_thport = ipc_port_copyout_send_pinned(port, get_task_ipcspace(p->task));
4549*a325d9c4SApple OSS Distributions 	}
4550*a325d9c4SApple OSS Distributions 
4551*a325d9c4SApple OSS Distributions 	/* Thread has been set up to run, arm its next workqueue quantum or disarm
4552*a325d9c4SApple OSS Distributions 	 * if it is no longer supporting that */
4553*a325d9c4SApple OSS Distributions 	if (thread_supports_cooperative_workqueue(th)) {
4554*a325d9c4SApple OSS Distributions 		thread_arm_workqueue_quantum(th);
4555*a325d9c4SApple OSS Distributions 	} else {
4556*a325d9c4SApple OSS Distributions 		thread_disarm_workqueue_quantum(th);
4557*a325d9c4SApple OSS Distributions 	}
4558*a325d9c4SApple OSS Distributions 
4559*a325d9c4SApple OSS Distributions 	/*
4560*a325d9c4SApple OSS Distributions 	 * Call out to pthread, this sets up the thread, pulls in kevent structs
4561*a325d9c4SApple OSS Distributions 	 * onto the stack, sets up the thread state and then returns to userspace.
4562*a325d9c4SApple OSS Distributions 	 */
4563*a325d9c4SApple OSS Distributions 	WQ_TRACE_WQ(TRACE_wq_runthread | DBG_FUNC_START,
4564*a325d9c4SApple OSS Distributions 	    proc_get_wqptr_fast(p), 0, 0, 0);
4565*a325d9c4SApple OSS Distributions 
4566*a325d9c4SApple OSS Distributions 	if (workq_thread_is_cooperative(uth)) {
4567*a325d9c4SApple OSS Distributions 		thread_sched_call(th, NULL);
4568*a325d9c4SApple OSS Distributions 	} else {
4569*a325d9c4SApple OSS Distributions 		thread_sched_call(th, workq_sched_callback);
4570*a325d9c4SApple OSS Distributions 	}
4571*a325d9c4SApple OSS Distributions 
4572*a325d9c4SApple OSS Distributions 	pthread_functions->workq_setup_thread(p, th, vmap, uth->uu_workq_stackaddr,
4573*a325d9c4SApple OSS Distributions 	    uth->uu_workq_thport, 0, setup_flags, upcall_flags);
4574*a325d9c4SApple OSS Distributions 
4575*a325d9c4SApple OSS Distributions 	__builtin_unreachable();
4576*a325d9c4SApple OSS Distributions }
4577*a325d9c4SApple OSS Distributions 
4578*a325d9c4SApple OSS Distributions #pragma mark misc
4579*a325d9c4SApple OSS Distributions 
4580*a325d9c4SApple OSS Distributions int
fill_procworkqueue(proc_t p,struct proc_workqueueinfo * pwqinfo)4581*a325d9c4SApple OSS Distributions fill_procworkqueue(proc_t p, struct proc_workqueueinfo * pwqinfo)
4582*a325d9c4SApple OSS Distributions {
4583*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
4584*a325d9c4SApple OSS Distributions 	int error = 0;
4585*a325d9c4SApple OSS Distributions 	int     activecount;
4586*a325d9c4SApple OSS Distributions 
4587*a325d9c4SApple OSS Distributions 	if (wq == NULL) {
4588*a325d9c4SApple OSS Distributions 		return EINVAL;
4589*a325d9c4SApple OSS Distributions 	}
4590*a325d9c4SApple OSS Distributions 
4591*a325d9c4SApple OSS Distributions 	/*
4592*a325d9c4SApple OSS Distributions 	 * This is sometimes called from interrupt context by the kperf sampler.
4593*a325d9c4SApple OSS Distributions 	 * In that case, it's not safe to spin trying to take the lock since we
4594*a325d9c4SApple OSS Distributions 	 * might already hold it.  So, we just try-lock it and error out if it's
4595*a325d9c4SApple OSS Distributions 	 * already held.  Since this is just a debugging aid, and all our callers
4596*a325d9c4SApple OSS Distributions 	 * are able to handle an error, that's fine.
4597*a325d9c4SApple OSS Distributions 	 */
4598*a325d9c4SApple OSS Distributions 	bool locked = workq_lock_try(wq);
4599*a325d9c4SApple OSS Distributions 	if (!locked) {
4600*a325d9c4SApple OSS Distributions 		return EBUSY;
4601*a325d9c4SApple OSS Distributions 	}
4602*a325d9c4SApple OSS Distributions 
4603*a325d9c4SApple OSS Distributions 	wq_thactive_t act = _wq_thactive(wq);
4604*a325d9c4SApple OSS Distributions 	activecount = _wq_thactive_aggregate_downto_qos(wq, act,
4605*a325d9c4SApple OSS Distributions 	    WORKQ_THREAD_QOS_MIN, NULL, NULL);
4606*a325d9c4SApple OSS Distributions 	if (act & _wq_thactive_offset_for_qos(WORKQ_THREAD_QOS_MANAGER)) {
4607*a325d9c4SApple OSS Distributions 		activecount++;
4608*a325d9c4SApple OSS Distributions 	}
4609*a325d9c4SApple OSS Distributions 	pwqinfo->pwq_nthreads = wq->wq_nthreads;
4610*a325d9c4SApple OSS Distributions 	pwqinfo->pwq_runthreads = activecount;
4611*a325d9c4SApple OSS Distributions 	pwqinfo->pwq_blockedthreads = wq->wq_threads_scheduled - activecount;
4612*a325d9c4SApple OSS Distributions 	pwqinfo->pwq_state = 0;
4613*a325d9c4SApple OSS Distributions 
4614*a325d9c4SApple OSS Distributions 	if (wq->wq_constrained_threads_scheduled >= wq_max_constrained_threads) {
4615*a325d9c4SApple OSS Distributions 		pwqinfo->pwq_state |= WQ_EXCEEDED_CONSTRAINED_THREAD_LIMIT;
4616*a325d9c4SApple OSS Distributions 	}
4617*a325d9c4SApple OSS Distributions 
4618*a325d9c4SApple OSS Distributions 	if (wq->wq_nthreads >= wq_max_threads) {
4619*a325d9c4SApple OSS Distributions 		pwqinfo->pwq_state |= WQ_EXCEEDED_TOTAL_THREAD_LIMIT;
4620*a325d9c4SApple OSS Distributions 	}
4621*a325d9c4SApple OSS Distributions 
4622*a325d9c4SApple OSS Distributions 	workq_unlock(wq);
4623*a325d9c4SApple OSS Distributions 	return error;
4624*a325d9c4SApple OSS Distributions }
4625*a325d9c4SApple OSS Distributions 
4626*a325d9c4SApple OSS Distributions boolean_t
workqueue_get_pwq_exceeded(void * v,boolean_t * exceeded_total,boolean_t * exceeded_constrained)4627*a325d9c4SApple OSS Distributions workqueue_get_pwq_exceeded(void *v, boolean_t *exceeded_total,
4628*a325d9c4SApple OSS Distributions     boolean_t *exceeded_constrained)
4629*a325d9c4SApple OSS Distributions {
4630*a325d9c4SApple OSS Distributions 	proc_t p = v;
4631*a325d9c4SApple OSS Distributions 	struct proc_workqueueinfo pwqinfo;
4632*a325d9c4SApple OSS Distributions 	int err;
4633*a325d9c4SApple OSS Distributions 
4634*a325d9c4SApple OSS Distributions 	assert(p != NULL);
4635*a325d9c4SApple OSS Distributions 	assert(exceeded_total != NULL);
4636*a325d9c4SApple OSS Distributions 	assert(exceeded_constrained != NULL);
4637*a325d9c4SApple OSS Distributions 
4638*a325d9c4SApple OSS Distributions 	err = fill_procworkqueue(p, &pwqinfo);
4639*a325d9c4SApple OSS Distributions 	if (err) {
4640*a325d9c4SApple OSS Distributions 		return FALSE;
4641*a325d9c4SApple OSS Distributions 	}
4642*a325d9c4SApple OSS Distributions 	if (!(pwqinfo.pwq_state & WQ_FLAGS_AVAILABLE)) {
4643*a325d9c4SApple OSS Distributions 		return FALSE;
4644*a325d9c4SApple OSS Distributions 	}
4645*a325d9c4SApple OSS Distributions 
4646*a325d9c4SApple OSS Distributions 	*exceeded_total = (pwqinfo.pwq_state & WQ_EXCEEDED_TOTAL_THREAD_LIMIT);
4647*a325d9c4SApple OSS Distributions 	*exceeded_constrained = (pwqinfo.pwq_state & WQ_EXCEEDED_CONSTRAINED_THREAD_LIMIT);
4648*a325d9c4SApple OSS Distributions 
4649*a325d9c4SApple OSS Distributions 	return TRUE;
4650*a325d9c4SApple OSS Distributions }
4651*a325d9c4SApple OSS Distributions 
4652*a325d9c4SApple OSS Distributions uint32_t
workqueue_get_pwq_state_kdp(void * v)4653*a325d9c4SApple OSS Distributions workqueue_get_pwq_state_kdp(void * v)
4654*a325d9c4SApple OSS Distributions {
4655*a325d9c4SApple OSS Distributions 	static_assert((WQ_EXCEEDED_CONSTRAINED_THREAD_LIMIT << 17) ==
4656*a325d9c4SApple OSS Distributions 	    kTaskWqExceededConstrainedThreadLimit);
4657*a325d9c4SApple OSS Distributions 	static_assert((WQ_EXCEEDED_TOTAL_THREAD_LIMIT << 17) ==
4658*a325d9c4SApple OSS Distributions 	    kTaskWqExceededTotalThreadLimit);
4659*a325d9c4SApple OSS Distributions 	static_assert((WQ_FLAGS_AVAILABLE << 17) == kTaskWqFlagsAvailable);
4660*a325d9c4SApple OSS Distributions 	static_assert((WQ_FLAGS_AVAILABLE | WQ_EXCEEDED_TOTAL_THREAD_LIMIT |
4661*a325d9c4SApple OSS Distributions 	    WQ_EXCEEDED_CONSTRAINED_THREAD_LIMIT) == 0x7);
4662*a325d9c4SApple OSS Distributions 
4663*a325d9c4SApple OSS Distributions 	if (v == NULL) {
4664*a325d9c4SApple OSS Distributions 		return 0;
4665*a325d9c4SApple OSS Distributions 	}
4666*a325d9c4SApple OSS Distributions 
4667*a325d9c4SApple OSS Distributions 	proc_t p = v;
4668*a325d9c4SApple OSS Distributions 	struct workqueue *wq = proc_get_wqptr(p);
4669*a325d9c4SApple OSS Distributions 
4670*a325d9c4SApple OSS Distributions 	if (wq == NULL || workq_lock_is_acquired_kdp(wq)) {
4671*a325d9c4SApple OSS Distributions 		return 0;
4672*a325d9c4SApple OSS Distributions 	}
4673*a325d9c4SApple OSS Distributions 
4674*a325d9c4SApple OSS Distributions 	uint32_t pwq_state = WQ_FLAGS_AVAILABLE;
4675*a325d9c4SApple OSS Distributions 
4676*a325d9c4SApple OSS Distributions 	if (wq->wq_constrained_threads_scheduled >= wq_max_constrained_threads) {
4677*a325d9c4SApple OSS Distributions 		pwq_state |= WQ_EXCEEDED_CONSTRAINED_THREAD_LIMIT;
4678*a325d9c4SApple OSS Distributions 	}
4679*a325d9c4SApple OSS Distributions 
4680*a325d9c4SApple OSS Distributions 	if (wq->wq_nthreads >= wq_max_threads) {
4681*a325d9c4SApple OSS Distributions 		pwq_state |= WQ_EXCEEDED_TOTAL_THREAD_LIMIT;
4682*a325d9c4SApple OSS Distributions 	}
4683*a325d9c4SApple OSS Distributions 
4684*a325d9c4SApple OSS Distributions 	return pwq_state;
4685*a325d9c4SApple OSS Distributions }
4686*a325d9c4SApple OSS Distributions 
4687*a325d9c4SApple OSS Distributions void
workq_init(void)4688*a325d9c4SApple OSS Distributions workq_init(void)
4689*a325d9c4SApple OSS Distributions {
4690*a325d9c4SApple OSS Distributions 	clock_interval_to_absolutetime_interval(wq_stalled_window.usecs,
4691*a325d9c4SApple OSS Distributions 	    NSEC_PER_USEC, &wq_stalled_window.abstime);
4692*a325d9c4SApple OSS Distributions 	clock_interval_to_absolutetime_interval(wq_reduce_pool_window.usecs,
4693*a325d9c4SApple OSS Distributions 	    NSEC_PER_USEC, &wq_reduce_pool_window.abstime);
4694*a325d9c4SApple OSS Distributions 	clock_interval_to_absolutetime_interval(wq_max_timer_interval.usecs,
4695*a325d9c4SApple OSS Distributions 	    NSEC_PER_USEC, &wq_max_timer_interval.abstime);
4696*a325d9c4SApple OSS Distributions 
4697*a325d9c4SApple OSS Distributions 	thread_deallocate_daemon_register_queue(&workq_deallocate_queue,
4698*a325d9c4SApple OSS Distributions 	    workq_deallocate_queue_invoke);
4699*a325d9c4SApple OSS Distributions }
4700