xref: /xnu-10063.101.15/osfmk/ipc/ipc_pset.c (revision 94d3b452840153a99b38a3a9659680b2a006908e)
1 /*
2  * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved.
3  *
4  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5  *
6  * This file contains Original Code and/or Modifications of Original Code
7  * as defined in and that are subject to the Apple Public Source License
8  * Version 2.0 (the 'License'). You may not use this file except in
9  * compliance with the License. The rights granted to you under the License
10  * may not be used to create, or enable the creation or redistribution of,
11  * unlawful or unlicensed copies of an Apple operating system, or to
12  * circumvent, violate, or enable the circumvention or violation of, any
13  * terms of an Apple operating system software license agreement.
14  *
15  * Please obtain a copy of the License at
16  * http://www.opensource.apple.com/apsl/ and read it before using this file.
17  *
18  * The Original Code and all software distributed under the License are
19  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23  * Please see the License for the specific language governing rights and
24  * limitations under the License.
25  *
26  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27  */
28 /*
29  * @OSF_COPYRIGHT@
30  */
31 /*
32  * Mach Operating System
33  * Copyright (c) 1991,1990,1989 Carnegie Mellon University
34  * All Rights Reserved.
35  *
36  * Permission to use, copy, modify and distribute this software and its
37  * documentation is hereby granted, provided that both the copyright
38  * notice and this permission notice appear in all copies of the
39  * software, derivative works or modified versions, and any portions
40  * thereof, and that both notices appear in supporting documentation.
41  *
42  * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
43  * CONDITION.  CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR
44  * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
45  *
46  * Carnegie Mellon requests users of this software to return to
47  *
48  *  Software Distribution Coordinator  or  [email protected]
49  *  School of Computer Science
50  *  Carnegie Mellon University
51  *  Pittsburgh PA 15213-3890
52  *
53  * any improvements or extensions that they make and grant Carnegie Mellon
54  * the rights to redistribute these changes.
55  */
56 /*
57  */
58 /*
59  *	File:	ipc/ipc_pset.c
60  *	Author:	Rich Draves
61  *	Date:	1989
62  *
63  *	Functions to manipulate IPC port sets.
64  */
65 
66 #include <mach/port.h>
67 #include <mach/kern_return.h>
68 #include <mach/message.h>
69 #include <ipc/ipc_mqueue.h>
70 #include <ipc/ipc_object.h>
71 #include <ipc/ipc_pset.h>
72 #include <ipc/ipc_right.h>
73 #include <ipc/ipc_space.h>
74 #include <ipc/ipc_port.h>
75 #include <ipc/ipc_kmsg.h>
76 #include <kern/policy_internal.h>
77 
78 #include <kern/kern_types.h>
79 
80 #include <vm/vm_map.h>
81 #include <libkern/section_keywords.h>
82 #include <pthread/priority_private.h>
83 
84 /* processor_set stole ipc_pset_init */
85 static void
ipc_port_set_init(ipc_pset_t pset,mach_port_name_t name,int policy)86 ipc_port_set_init(ipc_pset_t pset, mach_port_name_t name, int policy)
87 {
88 	waitq_init(&pset->ips_wqset, WQT_PORT_SET, policy | SYNC_POLICY_FIFO);
89 	klist_init(&pset->ips_klist);
90 	pset->ips_wqset.wqset_index = MACH_PORT_INDEX(name);
91 }
92 
93 /*
94  *	Routine:	ipc_pset_alloc
95  *	Purpose:
96  *		Allocate a port set.
97  *	Conditions:
98  *		Nothing locked.  If successful, the port set is returned
99  *		locked.  (The caller doesn't have a reference.)
100  *	Returns:
101  *		KERN_SUCCESS		The port set is allocated.
102  *		KERN_INVALID_TASK	The space is dead.
103  *		KERN_NO_SPACE		No room for an entry in the space.
104  */
105 
106 kern_return_t
ipc_pset_alloc(ipc_space_t space,mach_port_name_t * namep,ipc_pset_t * psetp)107 ipc_pset_alloc(
108 	ipc_space_t             space,
109 	mach_port_name_t        *namep,
110 	ipc_pset_t              *psetp)
111 {
112 	ipc_pset_t pset;
113 	mach_port_name_t name;
114 	kern_return_t kr;
115 
116 	kr = ipc_object_alloc(space, IOT_PORT_SET,
117 	    MACH_PORT_TYPE_PORT_SET, 0,
118 	    &name, (ipc_object_t *) &pset);
119 	if (kr != KERN_SUCCESS) {
120 		return kr;
121 	}
122 	/* space is locked */
123 
124 	ipc_port_set_init(pset, name, SYNC_POLICY_INIT_LOCKED);
125 	/* port set is locked */
126 
127 	is_write_unlock(space);
128 
129 	*namep = name;
130 	*psetp = pset;
131 	return KERN_SUCCESS;
132 }
133 
134 /*
135  *	Routine:	ipc_pset_alloc_name
136  *	Purpose:
137  *		Allocate a port set, with a specific name.
138  *	Conditions:
139  *		Nothing locked.  If successful, the port set is returned
140  *		locked.  (The caller doesn't have a reference.)
141  *	Returns:
142  *		KERN_SUCCESS		The port set is allocated.
143  *		KERN_INVALID_TASK	The space is dead.
144  *		KERN_NAME_EXISTS	The name already denotes a right.
145  */
146 
147 kern_return_t
ipc_pset_alloc_name(ipc_space_t space,mach_port_name_t name,ipc_pset_t * psetp)148 ipc_pset_alloc_name(
149 	ipc_space_t             space,
150 	mach_port_name_t        name,
151 	ipc_pset_t              *psetp)
152 {
153 	return ipc_object_alloc_name(space, IOT_PORT_SET,
154 	           MACH_PORT_TYPE_PORT_SET, 0,
155 	           name, (ipc_object_t *)psetp, ^(ipc_object_t object){
156 		ipc_port_set_init(ips_object_to_pset(object), name,
157 		SYNC_POLICY_INIT_LOCKED);
158 	});
159 }
160 
161 
162 /*
163  *	Routine:	ipc_pset_alloc_special
164  *	Purpose:
165  *		Allocate a port set in a special space.
166  *		The new port set is returned with one ref.
167  *		If unsuccessful, IPS_NULL is returned.
168  *	Conditions:
169  *		Nothing locked.
170  */
171 ipc_pset_t
ipc_pset_alloc_special(__assert_only ipc_space_t space)172 ipc_pset_alloc_special(
173 	__assert_only ipc_space_t space)
174 {
175 	ipc_pset_t pset;
176 
177 	assert(space != IS_NULL);
178 	assert(!is_active(space));
179 
180 	pset = ips_object_to_pset(io_alloc(IOT_PORT_SET, Z_WAITOK | Z_ZERO));
181 	if (pset == IPS_NULL) {
182 		return IPS_NULL;
183 	}
184 
185 	os_atomic_init(&pset->ips_object.io_bits, io_makebits(IOT_PORT_SET));
186 	os_atomic_init(&pset->ips_object.io_references, 1);
187 
188 	ipc_port_set_init(pset, MACH_PORT_SPECIAL_DEFAULT, 0);
189 
190 	return pset;
191 }
192 
193 
194 /*
195  *	Routine:	ipc_pset_destroy
196  *	Purpose:
197  *		Destroys a port_set.
198  *	Conditions:
199  *		The port_set is locked and alive.
200  *		The caller has a reference, which is consumed.
201  *		Afterwards, the port_set is unlocked and dead.
202  */
203 
204 void
ipc_pset_destroy(ipc_space_t space,ipc_pset_t pset)205 ipc_pset_destroy(
206 	ipc_space_t     space,
207 	ipc_pset_t      pset)
208 {
209 	waitq_link_list_t free_l = { };
210 
211 	assert(ips_active(pset));
212 
213 	io_bits_andnot(ips_to_object(pset), IO_BITS_ACTIVE);
214 
215 	/*
216 	 * Set all waiters on the portset running to
217 	 * discover the change.
218 	 *
219 	 * Then under the same lock hold, deinit the waitq-set,
220 	 * which will remove all the member message queues,
221 	 * linkages and clean up preposts.
222 	 */
223 	ipc_mqueue_changed(space, &pset->ips_wqset);
224 	waitq_invalidate(&pset->ips_wqset);
225 	waitq_set_unlink_all_locked(&pset->ips_wqset, &free_l);
226 
227 	ips_mq_unlock(pset);
228 
229 	ips_release(pset);       /* consume the ref our caller gave us */
230 
231 	waitq_link_free_list(WQT_PORT_SET, &free_l);
232 }
233 
234 /*
235  *	Routine:	ipc_pset_finalize
236  *	Purpose:
237  *		Called on last reference deallocate to
238  *		free any remaining data associated with the pset.
239  *	Conditions:
240  *		Nothing locked.
241  */
242 void
ipc_pset_finalize(ipc_pset_t pset)243 ipc_pset_finalize(
244 	ipc_pset_t              pset)
245 {
246 	waitq_deinit(&pset->ips_wqset);
247 }
248 
249 
250 #pragma mark - kevent support
251 
252 /*
253  * Kqueue EVFILT_MACHPORT support
254  *
255  * - kn_ipc_{port,pset} points to the monitored ipc port or pset. If the knote
256  *   is using a kqwl, it is eligible to participate in sync IPC overrides.
257  *
258  *   For the first such sync IPC message in the port, we set up the port's
259  *   turnstile to directly push on the kqwl's turnstile (which is in turn set up
260  *   during filt_machportattach). If userspace responds to the message, the
261  *   turnstile push is severed the point of reply. If userspace returns without
262  *   responding to the message, we sever the turnstile push at the
263  *   point of reenabling the knote to deliver the next message. This is why the
264  *   knote needs to remember the port. For more details, see also
265  *   filt_machport_turnstile_complete.
266  *
267  *   If there are multiple other sync IPC messages in the port, messages 2 to n
268  *   redirect their turnstile push to the kqwl through an intermediatry "knote"
269  *   turnstile which in turn, pushes on the kqwl turnstile. This knote turnstile
270  *   is stored in the kn_hook. See also filt_machport_turnstile_prepare_lazily.
271  *
272  * - (in/out) ext[0] holds a mach_vm_address_t to a userspace buffer
273  *   that can be used to direct-deliver messages when
274  *   MACH_RCV_MSG is set in kn_sfflags
275  *
276  * - (in/out) ext[1] holds a mach_msg_size_t representing the size
277  *   of the userspace buffer held in ext[0].
278  *
279  * - (out)    ext[2] is used to deliver qos information
280  *   about the send queue to userspace.
281  *
282  * - (abused) ext[3] is used in kernel to hold a reference to the first port
283  *   with a turnstile that participate to sync IPC override. For more details,
284  *   see filt_machport_stash_port
285  *
286  * - kn_hook is optionally a "knote" turnstile. It is used as the inheritor
287  *   of turnstiles for rights copied out as part of direct message delivery
288  *   when they can participate to sync IPC override.
289  *
290  *   It is used to atomically neuter the sync IPC override when the knote is
291  *   re-enabled.
292  *
293  */
294 
295 #include <sys/event.h>
296 #include <sys/errno.h>
297 
298 static int
filt_pset_filter_result(ipc_pset_t pset)299 filt_pset_filter_result(ipc_pset_t pset)
300 {
301 	ips_mq_lock_held(pset);
302 
303 	if (!waitq_is_valid(&pset->ips_wqset)) {
304 		return 0;
305 	}
306 
307 	return waitq_set_first_prepost(&pset->ips_wqset, WQS_PREPOST_PEEK) ?
308 	       FILTER_ACTIVE : 0;
309 }
310 
311 static int
filt_port_filter_result(struct knote * kn,ipc_port_t port)312 filt_port_filter_result(struct knote *kn, ipc_port_t port)
313 {
314 	struct kqueue *kqwl = knote_get_kq(kn);
315 	ipc_kmsg_t first;
316 	int result = 0;
317 
318 	ip_mq_lock_held(port);
319 
320 	if (kn->kn_sfflags & MACH_RCV_MSG) {
321 		result = FILTER_RESET_EVENT_QOS;
322 	}
323 
324 	if (!waitq_is_valid(&port->ip_waitq)) {
325 		return result;
326 	}
327 
328 	if (port->ip_kernel_iotier_override != kqueue_get_iotier_override(kqwl)) {
329 		kqueue_set_iotier_override(kqwl, port->ip_kernel_iotier_override);
330 		result |= FILTER_ADJUST_EVENT_IOTIER_BIT;
331 	}
332 
333 	first = ipc_kmsg_queue_first(&port->ip_messages.imq_messages);
334 	if (!first) {
335 		return result;
336 	}
337 
338 	result = FILTER_ACTIVE;
339 	if (kn->kn_sfflags & MACH_RCV_MSG) {
340 		result |= FILTER_ADJUST_EVENT_QOS(first->ikm_qos_override);
341 	}
342 
343 #if CONFIG_PREADOPT_TG
344 	struct thread_group *tg = ipc_kmsg_get_thread_group(first);
345 	if (tg) {
346 		struct kqueue *kq = knote_get_kq(kn);
347 		kqueue_set_preadopted_thread_group(kq, tg,
348 		    first->ikm_qos_override);
349 	}
350 #endif
351 
352 	return result;
353 }
354 
355 struct turnstile *
filt_ipc_kqueue_turnstile(struct knote * kn)356 filt_ipc_kqueue_turnstile(struct knote *kn)
357 {
358 	assert(kn->kn_filter == EVFILT_MACHPORT || kn->kn_filter == EVFILT_WORKLOOP);
359 	return kqueue_turnstile(knote_get_kq(kn));
360 }
361 
362 bool
filt_machport_kqueue_has_turnstile(struct knote * kn)363 filt_machport_kqueue_has_turnstile(struct knote *kn)
364 {
365 	assert(kn->kn_filter == EVFILT_MACHPORT);
366 	return ((kn->kn_sfflags & MACH_RCV_MSG) || (kn->kn_sfflags & MACH_RCV_SYNC_PEEK))
367 	       && (kn->kn_flags & EV_DISPATCH);
368 }
369 
370 /*
371  * Stashes a port that participate to sync IPC override on the knote until the
372  * knote is re-enabled.
373  *
374  * It returns:
375  * - the turnstile to use as an inheritor for the stashed port
376  * - the kind of stash that happened as PORT_SYNC_* value among:
377  *   o not stashed (no sync IPC support)
378  *   o stashed in the knote (in kn_ext[3])
379  *   o to be hooked to the kn_hook knote
380  */
381 struct turnstile *
filt_machport_stash_port(struct knote * kn,ipc_port_t port,int * link)382 filt_machport_stash_port(struct knote *kn, ipc_port_t port, int *link)
383 {
384 	struct turnstile *ts = TURNSTILE_NULL;
385 
386 	if (kn->kn_filter == EVFILT_WORKLOOP) {
387 		assert(kn->kn_ipc_port == NULL);
388 		kn->kn_ipc_port = port;
389 		ip_reference(port);
390 		if (link) {
391 			*link = PORT_SYNC_LINK_WORKLOOP_KNOTE;
392 		}
393 		ts = filt_ipc_kqueue_turnstile(kn);
394 	} else if (!filt_machport_kqueue_has_turnstile(kn)) {
395 		if (link) {
396 			*link = PORT_SYNC_LINK_NO_LINKAGE;
397 		}
398 	} else if (kn->kn_ext[3] == 0) {
399 		ip_reference(port);
400 		kn->kn_ext[3] = (uintptr_t)port;
401 		ts = filt_ipc_kqueue_turnstile(kn);
402 		if (link) {
403 			*link = PORT_SYNC_LINK_WORKLOOP_KNOTE;
404 		}
405 	} else {
406 		ts = (struct turnstile *)knote_kn_hook_get_raw(kn);
407 		if (link) {
408 			*link = PORT_SYNC_LINK_WORKLOOP_STASH;
409 		}
410 	}
411 
412 	return ts;
413 }
414 
415 /*
416  * Lazily prepare a turnstile so that filt_machport_stash_port()
417  * can be called with the mqueue lock held.
418  *
419  * It will allocate a turnstile in kn_hook if:
420  * - the knote supports sync IPC override,
421  * - we already stashed a port in kn_ext[3],
422  * - the object that will be copied out has a chance to ask to be stashed.
423  *
424  * It is setup so that its inheritor is the workloop turnstile that has been
425  * allocated when this knote was attached.
426  */
427 void
filt_machport_turnstile_prepare_lazily(struct knote * kn,mach_msg_type_name_t msgt_name,ipc_port_t port)428 filt_machport_turnstile_prepare_lazily(
429 	struct knote *kn,
430 	mach_msg_type_name_t msgt_name,
431 	ipc_port_t port)
432 {
433 	/* This is called from within filt_machportprocess */
434 	assert((kn->kn_status & KN_SUPPRESSED) && (kn->kn_status & KN_LOCKED));
435 
436 	if (!filt_machport_kqueue_has_turnstile(kn)) {
437 		return;
438 	}
439 
440 	if (kn->kn_ext[3] == 0 || knote_kn_hook_get_raw(kn)) {
441 		return;
442 	}
443 
444 	struct turnstile *ts = filt_ipc_kqueue_turnstile(kn);
445 	if ((msgt_name == MACH_MSG_TYPE_PORT_SEND_ONCE && port->ip_specialreply) ||
446 	    (msgt_name == MACH_MSG_TYPE_PORT_RECEIVE)) {
447 		struct turnstile *kn_ts = turnstile_alloc();
448 		struct turnstile *ts_store;
449 		kn_ts = turnstile_prepare((uintptr_t)kn, &ts_store, kn_ts, TURNSTILE_KNOTE);
450 		knote_kn_hook_set_raw(kn, ts_store);
451 
452 		turnstile_update_inheritor(kn_ts, ts,
453 		    TURNSTILE_IMMEDIATE_UPDATE | TURNSTILE_INHERITOR_TURNSTILE);
454 		turnstile_cleanup();
455 	}
456 }
457 
458 static void
filt_machport_turnstile_complete_port(struct knote * kn,ipc_port_t port)459 filt_machport_turnstile_complete_port(struct knote *kn, ipc_port_t port)
460 {
461 	struct turnstile *ts = TURNSTILE_NULL;
462 
463 	ip_mq_lock(port);
464 	if (port->ip_specialreply) {
465 		/*
466 		 * If the reply has been sent to the special reply port already,
467 		 * then the special reply port may already be reused to do something
468 		 * entirely different.
469 		 *
470 		 * However, the only reason for it to still point to this knote is
471 		 * that it's still waiting for a reply, so when this is the case,
472 		 * neuter the linkage.
473 		 */
474 		if (port->ip_sync_link_state == PORT_SYNC_LINK_WORKLOOP_KNOTE &&
475 		    port->ip_sync_inheritor_knote == kn) {
476 			ipc_port_adjust_special_reply_port_locked(port, NULL,
477 			    (IPC_PORT_ADJUST_SR_NONE | IPC_PORT_ADJUST_SR_ENABLE_EVENT), FALSE);
478 			/* port unlocked */
479 		} else {
480 			ip_mq_unlock(port);
481 		}
482 	} else {
483 		/*
484 		 * For receive rights, if their IMQ_KNOTE() is still this
485 		 * knote, then sever the link.
486 		 */
487 		if (port->ip_sync_link_state == PORT_SYNC_LINK_WORKLOOP_KNOTE &&
488 		    port->ip_messages.imq_inheritor_knote == kn) {
489 			ipc_port_adjust_sync_link_state_locked(port, PORT_SYNC_LINK_ANY, NULL);
490 			ts = port_send_turnstile(port);
491 		}
492 		if (ts) {
493 			turnstile_reference(ts);
494 			turnstile_update_inheritor(ts, TURNSTILE_INHERITOR_NULL,
495 			    TURNSTILE_IMMEDIATE_UPDATE);
496 		}
497 		ip_mq_unlock(port);
498 
499 		if (ts) {
500 			turnstile_update_inheritor_complete(ts,
501 			    TURNSTILE_INTERLOCK_NOT_HELD);
502 			turnstile_deallocate(ts);
503 		}
504 	}
505 
506 	ip_release(port);
507 }
508 
509 void
filt_wldetach_sync_ipc(struct knote * kn)510 filt_wldetach_sync_ipc(struct knote *kn)
511 {
512 	ipc_port_t port = kn->kn_ipc_port;
513 	filt_machport_turnstile_complete_port(kn, port);
514 	kn->kn_ipc_port = IP_NULL;
515 }
516 
517 /*
518  * Other half of filt_machport_turnstile_prepare_lazily()
519  *
520  * This is serialized by the knote state machine.
521  */
522 static void
filt_machport_turnstile_complete(struct knote * kn)523 filt_machport_turnstile_complete(struct knote *kn)
524 {
525 	if (kn->kn_ext[3]) {
526 		ipc_port_t port = (ipc_port_t)kn->kn_ext[3];
527 		filt_machport_turnstile_complete_port(kn, port);
528 		kn->kn_ext[3] = 0;
529 	}
530 
531 	struct turnstile *ts = knote_kn_hook_get_raw(kn);
532 	if (ts) {
533 		turnstile_update_inheritor(ts, TURNSTILE_INHERITOR_NULL,
534 		    TURNSTILE_IMMEDIATE_UPDATE);
535 		turnstile_update_inheritor_complete(ts, TURNSTILE_INTERLOCK_HELD);
536 
537 		struct turnstile *ts_store = ts;
538 		turnstile_complete((uintptr_t)kn, (struct turnstile **)&ts_store, &ts, TURNSTILE_KNOTE);
539 		knote_kn_hook_set_raw(kn, ts_store);
540 
541 		turnstile_cleanup();
542 
543 		assert(ts);
544 		turnstile_deallocate(ts);
545 	}
546 }
547 
548 static void
filt_machport_link(struct klist * klist,struct knote * kn)549 filt_machport_link(struct klist *klist, struct knote *kn)
550 {
551 	struct knote *hd = SLIST_FIRST(klist);
552 
553 	if (hd && filt_machport_kqueue_has_turnstile(kn)) {
554 		SLIST_INSERT_AFTER(hd, kn, kn_selnext);
555 	} else {
556 		SLIST_INSERT_HEAD(klist, kn, kn_selnext);
557 	}
558 }
559 
560 static void
filt_machport_unlink(struct klist * klist,struct knote * kn)561 filt_machport_unlink(struct klist *klist, struct knote *kn)
562 {
563 	struct knote **knprev;
564 
565 	KNOTE_DETACH(klist, kn);
566 
567 	/* make sure the first knote is a knote we can push on */
568 	SLIST_FOREACH_PREVPTR(kn, knprev, klist, kn_selnext) {
569 		if (filt_machport_kqueue_has_turnstile(kn)) {
570 			*knprev = SLIST_NEXT(kn, kn_selnext);
571 			SLIST_INSERT_HEAD(klist, kn, kn_selnext);
572 			break;
573 		}
574 	}
575 }
576 
577 int
filt_wlattach_sync_ipc(struct knote * kn)578 filt_wlattach_sync_ipc(struct knote *kn)
579 {
580 	mach_port_name_t name = (mach_port_name_t)kn->kn_id;
581 	ipc_space_t space = current_space();
582 	ipc_entry_bits_t bits;
583 	ipc_object_t object;
584 	ipc_port_t port = IP_NULL;
585 	int error = 0;
586 
587 	if (ipc_right_lookup_read(space, name, &bits, &object) != KERN_SUCCESS) {
588 		return ENOENT;
589 	}
590 	/* object is locked and active */
591 
592 	if (bits & MACH_PORT_TYPE_RECEIVE) {
593 		port = ip_object_to_port(object);
594 		if (port->ip_specialreply) {
595 			error = ENOENT;
596 		}
597 	} else if (bits & MACH_PORT_TYPE_SEND_ONCE) {
598 		port = ip_object_to_port(object);
599 		if (!port->ip_specialreply) {
600 			error = ENOENT;
601 		}
602 	} else {
603 		error = ENOENT;
604 	}
605 	if (error) {
606 		io_unlock(object);
607 		return error;
608 	}
609 
610 	if (port->ip_sync_link_state == PORT_SYNC_LINK_ANY) {
611 		io_unlock(object);
612 		/*
613 		 * We cannot start a sync IPC inheritance chain, only further one
614 		 * Note: this can also happen if the inheritance chain broke
615 		 * because the original requestor died.
616 		 */
617 		return ENOENT;
618 	}
619 
620 	if (port->ip_specialreply) {
621 		ipc_port_adjust_special_reply_port_locked(port, kn,
622 		    IPC_PORT_ADJUST_SR_LINK_WORKLOOP, FALSE);
623 	} else {
624 		ipc_port_adjust_port_locked(port, kn, FALSE);
625 	}
626 
627 	/* make sure the port was stashed */
628 	assert(kn->kn_ipc_port == port);
629 
630 	/* port has been unlocked by ipc_port_adjust_* */
631 
632 	return 0;
633 }
634 
635 static int
filt_psetattach(struct knote * kn,ipc_pset_t pset)636 filt_psetattach(struct knote *kn, ipc_pset_t pset)
637 {
638 	int result = 0;
639 
640 	ips_reference(pset);
641 	kn->kn_ipc_pset = pset;
642 
643 	filt_machport_link(&pset->ips_klist, kn);
644 	result = filt_pset_filter_result(pset);
645 	ips_mq_unlock(pset);
646 
647 	return result;
648 }
649 
650 static int
filt_portattach(struct knote * kn,ipc_port_t port)651 filt_portattach(struct knote *kn, ipc_port_t port)
652 {
653 	struct turnstile *send_turnstile = TURNSTILE_NULL;
654 	int result = 0;
655 
656 	if (port->ip_specialreply) {
657 		/*
658 		 * Registering for kevents on special reply ports
659 		 * isn't supported for two reasons:
660 		 *
661 		 * 1. it really makes very little sense for a port that
662 		 *    is supposed to be used synchronously
663 		 *
664 		 * 2. their ports's ip_klist field will be used to
665 		 *    store the receive turnstile, so we can't possibly
666 		 *    attach them anyway.
667 		 */
668 		ip_mq_unlock(port);
669 		knote_set_error(kn, ENOTSUP);
670 		return 0;
671 	}
672 
673 	ip_reference(port);
674 	kn->kn_ipc_port = port;
675 	if (port->ip_sync_link_state != PORT_SYNC_LINK_ANY) {
676 		/*
677 		 * We're attaching a port that used to have an IMQ_KNOTE,
678 		 * clobber this state, we'll fixup its turnstile inheritor below.
679 		 */
680 		ipc_port_adjust_sync_link_state_locked(port, PORT_SYNC_LINK_ANY, NULL);
681 	}
682 
683 	filt_machport_link(&port->ip_klist, kn);
684 	result = filt_port_filter_result(kn, port);
685 
686 	/*
687 	 * Update the port's turnstile inheritor
688 	 *
689 	 * Unlike filt_machportdetach(), we don't have to care about races for
690 	 * turnstile_workloop_pusher_info(): filt_machport_link() doesn't affect
691 	 * already pushing knotes, and if the current one becomes the new
692 	 * pusher, it'll only be visible when turnstile_workloop_pusher_info()
693 	 * returns.
694 	 */
695 	send_turnstile = port_send_turnstile(port);
696 	if (send_turnstile) {
697 		turnstile_reference(send_turnstile);
698 		ipc_port_send_update_inheritor(port, send_turnstile,
699 		    TURNSTILE_IMMEDIATE_UPDATE);
700 
701 		/*
702 		 * rdar://problem/48861190
703 		 *
704 		 * When a listener connection resumes a peer,
705 		 * updating the inheritor above has moved the push
706 		 * from the current thread to the workloop.
707 		 *
708 		 * However, we haven't told the workloop yet
709 		 * that it needs a thread request, and we risk
710 		 * to be preeempted as soon as we drop the space
711 		 * lock below.
712 		 *
713 		 * To avoid this disable preemption and let kevent
714 		 * reenable it after it takes the kqlock.
715 		 */
716 		disable_preemption();
717 		result |= FILTER_THREADREQ_NODEFEER;
718 	}
719 
720 	ip_mq_unlock(port);
721 
722 	if (send_turnstile) {
723 		turnstile_update_inheritor_complete(send_turnstile,
724 		    TURNSTILE_INTERLOCK_NOT_HELD);
725 		turnstile_deallocate_safe(send_turnstile);
726 	}
727 
728 	return result;
729 }
730 
731 static int
filt_machportattach(struct knote * kn,__unused struct kevent_qos_s * kev)732 filt_machportattach(struct knote *kn, __unused struct kevent_qos_s *kev)
733 {
734 	mach_port_name_t name = (mach_port_name_t)kn->kn_id;
735 	ipc_space_t space = current_space();
736 	ipc_entry_bits_t bits;
737 	ipc_object_t object;
738 	kern_return_t kr;
739 
740 	kn->kn_flags &= ~EV_EOF;
741 	kn->kn_ext[3] = 0;
742 
743 	if (filt_machport_kqueue_has_turnstile(kn)) {
744 		/*
745 		 * If the filter is likely to support sync IPC override,
746 		 * and it happens to be attaching to a workloop,
747 		 * make sure the workloop has an allocated turnstile.
748 		 */
749 		kqueue_alloc_turnstile(knote_get_kq(kn));
750 	}
751 
752 	kr = ipc_right_lookup_read(space, name, &bits, &object);
753 
754 	if (kr != KERN_SUCCESS) {
755 		knote_set_error(kn, ENOENT);
756 		return 0;
757 	}
758 	/* object is locked and active */
759 
760 	if (bits & MACH_PORT_TYPE_PORT_SET) {
761 		kn->kn_filtid = EVFILTID_MACH_PORT_SET;
762 		return filt_psetattach(kn, ips_object_to_pset(object));
763 	}
764 
765 	if (bits & MACH_PORT_TYPE_RECEIVE) {
766 		kn->kn_filtid = EVFILTID_MACH_PORT;
767 		return filt_portattach(kn, ip_object_to_port(object));
768 	}
769 
770 	io_unlock(object);
771 	knote_set_error(kn, ENOTSUP);
772 	return 0;
773 }
774 
775 static void
filt_psetdetach(struct knote * kn)776 filt_psetdetach(struct knote *kn)
777 {
778 	ipc_pset_t pset = kn->kn_ipc_pset;
779 
780 	filt_machport_turnstile_complete(kn);
781 
782 	ips_mq_lock(pset);
783 
784 	if ((kn->kn_status & KN_VANISHED) || (kn->kn_flags & EV_EOF)) {
785 		/*
786 		 * ipc_mqueue_changed() already unhooked this knote from the waitq,
787 		 */
788 	} else {
789 		filt_machport_unlink(&pset->ips_klist, kn);
790 	}
791 
792 	kn->kn_ipc_pset = IPS_NULL;
793 	ips_mq_unlock(pset);
794 	ips_release(pset);
795 }
796 
797 static void
filt_portdetach(struct knote * kn)798 filt_portdetach(struct knote *kn)
799 {
800 	ipc_port_t port = kn->kn_ipc_port;
801 	struct turnstile *send_turnstile = TURNSTILE_NULL;
802 
803 	filt_machport_turnstile_complete(kn);
804 
805 	ip_mq_lock(port);
806 	if ((kn->kn_status & KN_VANISHED) || (kn->kn_flags & EV_EOF)) {
807 		/*
808 		 * ipc_mqueue_changed() already unhooked this knote from the waitq,
809 		 */
810 	} else {
811 		/*
812 		 * When the knote being detached is the first one in the list,
813 		 * then unlinking the knote *and* updating the turnstile inheritor
814 		 * need to happen atomically with respect to the callers of
815 		 * turnstile_workloop_pusher_info().
816 		 *
817 		 * The caller of turnstile_workloop_pusher_info() will use the kq req
818 		 * lock (and hence the kqlock), so we just need to hold the kqlock too.
819 		 */
820 		assert(port->ip_sync_link_state == PORT_SYNC_LINK_ANY);
821 		if (kn == SLIST_FIRST(&port->ip_klist)) {
822 			send_turnstile = port_send_turnstile(port);
823 		}
824 		filt_machport_unlink(&port->ip_klist, kn);
825 		struct kqueue *kq = knote_get_kq(kn);
826 		kqueue_set_iotier_override(kq, THROTTLE_LEVEL_END);
827 	}
828 
829 	if (send_turnstile) {
830 		turnstile_reference(send_turnstile);
831 		ipc_port_send_update_inheritor(port, send_turnstile,
832 		    TURNSTILE_IMMEDIATE_UPDATE);
833 	}
834 
835 	/* Clear the knote pointer once the knote has been removed from turnstile */
836 	kn->kn_ipc_port = IP_NULL;
837 	ip_mq_unlock(port);
838 
839 	if (send_turnstile) {
840 		turnstile_update_inheritor_complete(send_turnstile,
841 		    TURNSTILE_INTERLOCK_NOT_HELD);
842 		turnstile_deallocate(send_turnstile);
843 	}
844 
845 	ip_release(port);
846 }
847 
848 /*
849  * filt_{pset,port}event - deliver events into the mach port filter
850  *
851  * Mach port message arrival events are currently only posted via the
852  * kqueue filter routine for ports.
853  *
854  * If there is a message at the head of the queue,
855  * we indicate that the knote should go active.  If
856  * the message is to be direct-received, we adjust the
857  * QoS of the knote according the requested and override
858  * QoS of that first message.
859  *
860  * When the knote is for a port-set, the hint is non 0
861  * and is the waitq which is posting.
862  */
863 static int
filt_psetevent(struct knote * kn __unused,long hint __assert_only)864 filt_psetevent(struct knote *kn __unused, long hint __assert_only)
865 {
866 	/*
867 	 * When called for a port-set,
868 	 * the posting port waitq is locked.
869 	 *
870 	 * waitq_set_first_prepost()
871 	 * in filt_machport_filter_result()
872 	 * would try to lock it and be very sad.
873 	 *
874 	 * Just trust what we know to be true.
875 	 */
876 	assert(hint != 0);
877 	return FILTER_ACTIVE;
878 }
879 
880 static int
filt_portevent(struct knote * kn,long hint __assert_only)881 filt_portevent(struct knote *kn, long hint __assert_only)
882 {
883 	assert(hint == 0);
884 	return filt_port_filter_result(kn, kn->kn_ipc_port);
885 }
886 
887 void
ipc_pset_prepost(struct waitq_set * wqs,struct waitq * waitq)888 ipc_pset_prepost(struct waitq_set *wqs, struct waitq *waitq)
889 {
890 	KNOTE(&ips_from_waitq(wqs)->ips_klist, (long)waitq);
891 }
892 
893 static void
filt_machporttouch(struct knote * kn,struct kevent_qos_s * kev)894 filt_machporttouch(struct knote *kn, struct kevent_qos_s *kev)
895 {
896 	/*
897 	 * Specificying MACH_RCV_MSG or MACH_RCV_SYNC_PEEK during attach results in
898 	 * allocation of a turnstile. Modifying the filter flags to include these
899 	 * flags later, without a turnstile being allocated, leads to
900 	 * inconsistencies.
901 	 */
902 	if ((kn->kn_sfflags ^ kev->fflags) & (MACH_RCV_MSG | MACH_RCV_SYNC_PEEK)) {
903 		kev->flags |= EV_ERROR;
904 		kev->data = EINVAL;
905 		return;
906 	}
907 
908 	/* copy in new settings and save off new input fflags */
909 	kn->kn_sfflags = kev->fflags;
910 	kn->kn_ext[0] = kev->ext[0];
911 	kn->kn_ext[1] = kev->ext[1];
912 
913 	if (kev->flags & EV_ENABLE) {
914 		/*
915 		 * If the knote is being enabled, make sure there's no lingering
916 		 * IPC overrides from the previous message delivery.
917 		 */
918 		filt_machport_turnstile_complete(kn);
919 	}
920 }
921 
922 static int
filt_psettouch(struct knote * kn,struct kevent_qos_s * kev)923 filt_psettouch(struct knote *kn, struct kevent_qos_s *kev)
924 {
925 	ipc_pset_t pset = kn->kn_ipc_pset;
926 	int result = 0;
927 
928 	filt_machporttouch(kn, kev);
929 	if (kev->flags & EV_ERROR) {
930 		return 0;
931 	}
932 
933 	ips_mq_lock(pset);
934 	result = filt_pset_filter_result(pset);
935 	ips_mq_unlock(pset);
936 
937 	return result;
938 }
939 
940 static int
filt_porttouch(struct knote * kn,struct kevent_qos_s * kev)941 filt_porttouch(struct knote *kn, struct kevent_qos_s *kev)
942 {
943 	ipc_port_t port = kn->kn_ipc_port;
944 	int result = 0;
945 
946 	filt_machporttouch(kn, kev);
947 	if (kev->flags & EV_ERROR) {
948 		return 0;
949 	}
950 
951 	ip_mq_lock(port);
952 	result = filt_port_filter_result(kn, port);
953 	ip_mq_unlock(port);
954 
955 	return result;
956 }
957 
958 static int
filt_machportprocess(struct knote * kn,struct kevent_qos_s * kev,ipc_object_t object,ipc_object_type_t otype)959 filt_machportprocess(
960 	struct knote           *kn,
961 	struct kevent_qos_s    *kev,
962 	ipc_object_t            object,
963 	ipc_object_type_t       otype)
964 {
965 	thread_t self = current_thread();
966 	kevent_ctx_t kectx = NULL;
967 
968 	wait_result_t wresult;
969 	mach_msg_option64_t option64;
970 	mach_vm_address_t msg_addr;
971 	mach_msg_size_t max_msg_size, cpout_aux_size, cpout_msg_size;
972 	uint32_t ppri;
973 	mach_msg_qos_t oqos;
974 
975 	int result = FILTER_ACTIVE;
976 
977 	/* Capture current state */
978 	knote_fill_kevent(kn, kev, MACH_PORT_NULL);
979 
980 	/* Clear port reference, use ext3 as size of msg aux data */
981 	kev->ext[3] = 0;
982 
983 	/* If already deallocated/moved return one last EOF event */
984 	if (kev->flags & EV_EOF) {
985 		return FILTER_ACTIVE | FILTER_RESET_EVENT_QOS;
986 	}
987 
988 	/*
989 	 * Only honor supported receive options. If no options are
990 	 * provided, just force a MACH_RCV_LARGE to detect the
991 	 * name of the port and sizeof the waiting message.
992 	 *
993 	 * Extend kn_sfflags to 64 bits.
994 	 */
995 	option64 = (mach_msg_option64_t)kn->kn_sfflags & (MACH_RCV_MSG |
996 	    MACH_RCV_LARGE | MACH_RCV_LARGE_IDENTITY |
997 	    MACH_RCV_TRAILER_MASK | MACH_RCV_VOUCHER | MACH_MSG_STRICT_REPLY);
998 
999 	if (option64 & MACH_RCV_MSG) {
1000 		msg_addr = (mach_vm_address_t) kn->kn_ext[0];
1001 		max_msg_size = (mach_msg_size_t) kn->kn_ext[1];
1002 
1003 		/*
1004 		 * Copy out the incoming message as vector, and append aux data
1005 		 * immediately after the message proper (if any) and report its
1006 		 * size on ext3.
1007 		 */
1008 		option64 |= (MACH64_MSG_VECTOR | MACH64_RCV_LINEAR_VECTOR);
1009 
1010 		/*
1011 		 * If the kevent didn't specify a buffer and length, carve a buffer
1012 		 * from the filter processing data according to the flags.
1013 		 */
1014 		if (max_msg_size == 0) {
1015 			kectx = kevent_get_context(self);
1016 			msg_addr  = (mach_vm_address_t)kectx->kec_data_out;
1017 			max_msg_size  = (mach_msg_size_t)kectx->kec_data_resid;
1018 			option64 |= (MACH_RCV_LARGE | MACH_RCV_LARGE_IDENTITY);
1019 			/* Receive vector linearly onto stack */
1020 			if (kectx->kec_process_flags & KEVENT_FLAG_STACK_DATA) {
1021 				option64 |= MACH64_RCV_STACK;
1022 			}
1023 		}
1024 	} else {
1025 		/* just detect the port name (if a set) and size of the first message */
1026 		option64 = MACH_RCV_LARGE;
1027 		msg_addr = 0;
1028 		max_msg_size = 0;
1029 	}
1030 
1031 	/*
1032 	 * Set up to receive a message or the notification of a
1033 	 * too large message.  But never allow this call to wait.
1034 	 * If the user provided aditional options, like trailer
1035 	 * options, pass those through here.  But we don't support
1036 	 * scatter lists through this interface.
1037 	 *
1038 	 * Note: while in filt_machportprocess(),
1039 	 *       the knote has a reference on `object` that we can borrow.
1040 	 */
1041 	self->ith_object = object;
1042 
1043 	/* Using msg_addr as combined buffer for message proper and aux */
1044 	self->ith_msg_addr = msg_addr;
1045 	self->ith_max_msize = max_msg_size;
1046 	self->ith_msize = 0;
1047 
1048 	self->ith_aux_addr = 0;
1049 	self->ith_max_asize = 0;
1050 	self->ith_asize = 0;
1051 
1052 	self->ith_option = option64;
1053 	self->ith_receiver_name = MACH_PORT_NULL;
1054 	option64 |= MACH_RCV_TIMEOUT; // never wait
1055 	self->ith_state = MACH_RCV_IN_PROGRESS;
1056 	self->ith_knote = kn;
1057 
1058 	ipc_object_lock(object, otype);
1059 
1060 	wresult = ipc_mqueue_receive_on_thread_and_unlock(
1061 		io_waitq(object),
1062 		option64,
1063 		self->ith_max_msize,       /* max msg suze */
1064 		0,                         /* max aux size 0, using combined buffer */
1065 		0,                         /* immediate timeout */
1066 		THREAD_INTERRUPTIBLE,
1067 		self);
1068 	/* port unlocked */
1069 
1070 	/* If we timed out, or the process is exiting, just zero.  */
1071 	if (wresult == THREAD_RESTART || self->ith_state == MACH_RCV_TIMED_OUT) {
1072 		assert(self->turnstile != TURNSTILE_NULL);
1073 		self->ith_knote = ITH_KNOTE_NULL;
1074 		return 0;
1075 	}
1076 
1077 	assert(wresult == THREAD_NOT_WAITING);
1078 	assert(self->ith_state != MACH_RCV_IN_PROGRESS);
1079 
1080 	/*
1081 	 * If we weren't attempting to receive a message
1082 	 * directly, we need to return the port name in
1083 	 * the kevent structure.
1084 	 */
1085 	if ((option64 & MACH_RCV_MSG) != MACH_RCV_MSG) {
1086 		assert(self->ith_state == MACH_RCV_TOO_LARGE);
1087 		assert(self->ith_kmsg == IKM_NULL);
1088 		kev->data = self->ith_receiver_name;
1089 		self->ith_knote = ITH_KNOTE_NULL;
1090 		return result;
1091 	}
1092 
1093 #if CONFIG_PREADOPT_TG
1094 	/* If we're the first EVFILT_MACHPORT knote that is being processed for this
1095 	 * kqwl, then make sure to preadopt the thread group from the kmsg we're
1096 	 * about to receive. This is to make sure that we fix up the preadoption
1097 	 * thread group correctly on the receive side for the first message.
1098 	 */
1099 	struct kqueue *kq = knote_get_kq(kn);
1100 
1101 	if (self->ith_kmsg) {
1102 		struct thread_group *tg = ipc_kmsg_get_thread_group(self->ith_kmsg);
1103 
1104 		kqueue_process_preadopt_thread_group(self, kq, tg);
1105 	}
1106 #endif
1107 	if (otype == IOT_PORT) {
1108 		ipc_port_t port = ip_object_to_port(object);
1109 		struct kqueue *kqwl = knote_get_kq(kn);
1110 		if (port->ip_kernel_iotier_override != kqueue_get_iotier_override(kqwl)) {
1111 			/*
1112 			 * Lock the port to make sure port->ip_kernel_iotier_override does
1113 			 * not change while updating the kqueue override, else kqueue could
1114 			 * have old iotier value.
1115 			 */
1116 			ip_mq_lock(port);
1117 			kqueue_set_iotier_override(kqwl, port->ip_kernel_iotier_override);
1118 			ip_mq_unlock(port);
1119 			result |= FILTER_ADJUST_EVENT_IOTIER_BIT;
1120 		}
1121 	}
1122 
1123 	/*
1124 	 * Attempt to receive the message directly, returning
1125 	 * the results in the fflags field.
1126 	 */
1127 	io_reference(object);
1128 	kev->fflags = mach_msg_receive_results_kevent(&cpout_msg_size,
1129 	    &cpout_aux_size, &ppri, &oqos);
1130 
1131 	/* kmsg and object reference consumed */
1132 
1133 	/*
1134 	 * if the user asked for the identity of ports containing a
1135 	 * a too-large message, return it in the data field (as we
1136 	 * do for messages we didn't try to receive).
1137 	 */
1138 	if (kev->fflags == MACH_RCV_TOO_LARGE) {
1139 		kev->ext[1] = self->ith_msize;
1140 		kev->ext[3] = self->ith_asize;  /* Only lower 32 bits of ext3 are used */
1141 		if (option64 & MACH_RCV_LARGE_IDENTITY) {
1142 			kev->data = self->ith_receiver_name;
1143 		} else {
1144 			kev->data = MACH_PORT_NULL;
1145 		}
1146 	} else {
1147 		kev->ext[1] = cpout_msg_size;
1148 		kev->ext[3] = cpout_aux_size; /* Only lower 32 bits of ext3 are used */
1149 		kev->data = MACH_PORT_NULL;
1150 	}
1151 
1152 	/*
1153 	 * If we used a data buffer carved out from the filt_process data,
1154 	 * store the address used in the knote and adjust the residual and
1155 	 * other parameters for future use.
1156 	 */
1157 	if (kectx) {
1158 		assert(kectx->kec_data_resid >= cpout_msg_size + cpout_aux_size);
1159 		kectx->kec_data_resid -= cpout_msg_size + cpout_aux_size;
1160 		if ((kectx->kec_process_flags & KEVENT_FLAG_STACK_DATA) == 0) {
1161 			kev->ext[0] = kectx->kec_data_out;
1162 			kectx->kec_data_out += cpout_msg_size + cpout_aux_size;
1163 		} else {
1164 			assert(option64 & MACH64_RCV_STACK);
1165 			kev->ext[0] = kectx->kec_data_out + kectx->kec_data_resid;
1166 		}
1167 	}
1168 
1169 	/*
1170 	 * Apply message-based QoS values to output kevent as prescribed.
1171 	 * The kev->ext[2] field gets (msg-qos << 32) | (override-qos).
1172 	 */
1173 	if (kev->fflags == MACH_MSG_SUCCESS) {
1174 		kev->ext[2] = ((uint64_t)ppri << 32) |
1175 		    _pthread_priority_make_from_thread_qos(oqos, 0, 0);
1176 	}
1177 
1178 	self->ith_knote = ITH_KNOTE_NULL;
1179 	return result;
1180 }
1181 
1182 static int
filt_psetprocess(struct knote * kn,struct kevent_qos_s * kev)1183 filt_psetprocess(struct knote *kn, struct kevent_qos_s *kev)
1184 {
1185 	ipc_object_t io = ips_to_object(kn->kn_ipc_pset);
1186 
1187 	return filt_machportprocess(kn, kev, io, IOT_PORT_SET);
1188 }
1189 
1190 static int
filt_portprocess(struct knote * kn,struct kevent_qos_s * kev)1191 filt_portprocess(struct knote *kn, struct kevent_qos_s *kev)
1192 {
1193 	ipc_object_t io = ip_to_object(kn->kn_ipc_port);
1194 
1195 	return filt_machportprocess(kn, kev, io, IOT_PORT);
1196 }
1197 
1198 static void
filt_machportsanitizedcopyout(struct knote * kn,struct kevent_qos_s * kev)1199 filt_machportsanitizedcopyout(struct knote *kn, struct kevent_qos_s *kev)
1200 {
1201 	*kev = *(struct kevent_qos_s *)&kn->kn_kevent;
1202 
1203 	// We may have stashed the address to the port that is pushing on the sync
1204 	// IPC so clear it out.
1205 	kev->ext[3] = 0;
1206 }
1207 
1208 const struct filterops machport_attach_filtops = {
1209 	.f_adjusts_qos = true,
1210 	.f_extended_codes = true,
1211 	.f_attach = filt_machportattach,
1212 	.f_sanitized_copyout = filt_machportsanitizedcopyout,
1213 };
1214 
1215 const struct filterops mach_port_filtops = {
1216 	.f_adjusts_qos = true,
1217 	.f_extended_codes = true,
1218 	.f_detach = filt_portdetach,
1219 	.f_event = filt_portevent,
1220 	.f_touch = filt_porttouch,
1221 	.f_process = filt_portprocess,
1222 	.f_sanitized_copyout = filt_machportsanitizedcopyout,
1223 };
1224 
1225 const struct filterops mach_port_set_filtops = {
1226 	.f_adjusts_qos = true,
1227 	.f_extended_codes = true,
1228 	.f_detach = filt_psetdetach,
1229 	.f_event = filt_psetevent,
1230 	.f_touch = filt_psettouch,
1231 	.f_process = filt_psetprocess,
1232 	.f_sanitized_copyout = filt_machportsanitizedcopyout,
1233 };
1234