xref: /xnu-10002.61.3/osfmk/device/iokit_rpc.c (revision 0f4c859e951fba394238ab619495c4e1d54d0f34)
1 /*
2  * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved.
3  *
4  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5  *
6  * This file contains Original Code and/or Modifications of Original Code
7  * as defined in and that are subject to the Apple Public Source License
8  * Version 2.0 (the 'License'). You may not use this file except in
9  * compliance with the License. The rights granted to you under the License
10  * may not be used to create, or enable the creation or redistribution of,
11  * unlawful or unlicensed copies of an Apple operating system, or to
12  * circumvent, violate, or enable the circumvention or violation of, any
13  * terms of an Apple operating system software license agreement.
14  *
15  * Please obtain a copy of the License at
16  * http://www.opensource.apple.com/apsl/ and read it before using this file.
17  *
18  * The Original Code and all software distributed under the License are
19  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23  * Please see the License for the specific language governing rights and
24  * limitations under the License.
25  *
26  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27  */
28 #include <mach/boolean.h>
29 #include <mach/kern_return.h>
30 #include <mach/mig_errors.h>
31 #include <mach/port.h>
32 #include <mach/vm_param.h>
33 #include <mach/notify.h>
34 //#include <mach/mach_host_server.h>
35 #include <mach/mach_types.h>
36 
37 #include <machine/machparam.h>          /* spl definitions */
38 
39 #include <ipc/ipc_port.h>
40 #include <ipc/ipc_space.h>
41 
42 #include <kern/clock.h>
43 #include <kern/spl.h>
44 #include <kern/queue.h>
45 #include <kern/zalloc.h>
46 #include <kern/thread.h>
47 #include <kern/task.h>
48 #include <kern/sched_prim.h>
49 #include <kern/misc_protos.h>
50 
51 #include <vm/pmap.h>
52 #include <vm/vm_map.h>
53 #include <vm/vm_kern.h>
54 
55 #include <device/device_types.h>
56 #include <device/device_port.h>
57 #include <device/device_server.h>
58 
59 #include <machine/machparam.h>
60 
61 #if defined(__i386__) || defined(__x86_64__)
62 #include <i386/pmap.h>
63 #endif
64 #if defined(__arm64__)
65 #include <arm/pmap.h>
66 #endif
67 #include <IOKit/IOKitServer.h>
68 
69 #define EXTERN
70 #define MIGEXTERN
71 
72 static void
73 iokit_no_senders( ipc_port_t port, mach_port_mscount_t mscount );
74 
75 /*
76  * Lifetime:
77  * - non lazy port with no-more senders
78  * - can be destroyed by iokit_destroy_object_port
79  *
80  */
81 IPC_KOBJECT_DEFINE(IKOT_IOKIT_IDENT,
82     .iko_op_no_senders = iokit_no_senders);
83 IPC_KOBJECT_DEFINE(IKOT_IOKIT_OBJECT,
84     .iko_op_no_senders = iokit_no_senders);
85 IPC_KOBJECT_DEFINE(IKOT_IOKIT_CONNECT,
86     .iko_op_no_senders = iokit_no_senders);
87 IPC_KOBJECT_DEFINE(IKOT_UEXT_OBJECT,
88     .iko_op_no_senders = iokit_no_senders);
89 
90 /*
91  * Lookup a device by its port.
92  * Doesn't consume the naked send right; produces a device reference.
93  */
94 io_object_t
iokit_lookup_io_object(ipc_port_t port,ipc_kobject_type_t type)95 iokit_lookup_io_object(ipc_port_t port, ipc_kobject_type_t type)
96 {
97 	io_object_t  obj = NULL;
98 	io_kobject_t kobj = NULL;
99 
100 	if (!IP_VALID(port)) {
101 		return NULL;
102 	}
103 
104 	ip_mq_lock(port);
105 	if (ip_active(port)) {
106 		kobj = ipc_kobject_get_locked(port, type);
107 		if (kobj) {
108 			iokit_kobject_retain(kobj);
109 		}
110 	}
111 	ip_mq_unlock(port);
112 	if (kobj) {
113 		obj = iokit_copy_object_for_consumed_kobject(kobj, type);
114 	}
115 
116 	return obj;
117 }
118 
119 MIGEXTERN io_object_t
iokit_lookup_object_port(ipc_port_t port)120 iokit_lookup_object_port(
121 	ipc_port_t      port)
122 {
123 	return iokit_lookup_io_object(port, IKOT_IOKIT_OBJECT);
124 }
125 
126 MIGEXTERN io_object_t
iokit_lookup_connect_port(ipc_port_t port)127 iokit_lookup_connect_port(
128 	ipc_port_t      port)
129 {
130 	return iokit_lookup_io_object(port, IKOT_IOKIT_CONNECT);
131 }
132 
133 MIGEXTERN io_object_t
iokit_lookup_ident_port(ipc_port_t port)134 iokit_lookup_ident_port(
135 	ipc_port_t      port)
136 {
137 	return iokit_lookup_io_object(port, IKOT_IOKIT_IDENT);
138 }
139 
140 MIGEXTERN io_object_t
iokit_lookup_uext_object_port(ipc_port_t port)141 iokit_lookup_uext_object_port(
142 	ipc_port_t      port)
143 {
144 	return iokit_lookup_io_object(port, IKOT_UEXT_OBJECT);
145 }
146 
147 static io_object_t
iokit_lookup_object_in_space_with_port_name(mach_port_name_t name,ipc_kobject_type_t type,ipc_space_t space)148 iokit_lookup_object_in_space_with_port_name(mach_port_name_t name, ipc_kobject_type_t type, ipc_space_t space)
149 {
150 	io_object_t obj = NULL;
151 	io_kobject_t kobj;
152 
153 	if (name && MACH_PORT_VALID(name)) {
154 		ipc_port_t port;
155 		kern_return_t kr;
156 
157 		kr = ipc_port_translate_send(space, name, &port);
158 
159 		if (kr == KERN_SUCCESS) {
160 			assert(IP_VALID(port));
161 			assert(ip_active(port));
162 			kobj = ipc_kobject_get_locked(port, type);
163 			if (kobj) {
164 				iokit_kobject_retain(kobj);
165 			}
166 			ip_mq_unlock(port);
167 			if (kobj) {
168 				obj = iokit_copy_object_for_consumed_kobject(kobj, type);
169 			}
170 		}
171 	}
172 
173 	return obj;
174 }
175 
176 EXTERN io_object_t
iokit_lookup_object_with_port_name(mach_port_name_t name,ipc_kobject_type_t type,task_t task)177 iokit_lookup_object_with_port_name(mach_port_name_t name, ipc_kobject_type_t type, task_t task)
178 {
179 	return iokit_lookup_object_in_space_with_port_name(name, type, task->itk_space);
180 }
181 
182 EXTERN io_object_t
iokit_lookup_connect_ref_current_task(mach_port_name_t name)183 iokit_lookup_connect_ref_current_task(mach_port_name_t name)
184 {
185 	return iokit_lookup_object_in_space_with_port_name(name, IKOT_IOKIT_CONNECT, current_space());
186 }
187 
188 EXTERN io_object_t
iokit_lookup_uext_ref_current_task(mach_port_name_t name)189 iokit_lookup_uext_ref_current_task(mach_port_name_t name)
190 {
191 	return iokit_lookup_object_in_space_with_port_name(name, IKOT_UEXT_OBJECT, current_space());
192 }
193 
194 /*
195  * Look up a port given a port name.
196  * This returns the port unlocked with a +1 send right.
197  * Release with iokit_release_port_send()
198  */
199 EXTERN ipc_port_t
iokit_lookup_raw_current_task(mach_port_name_t name,ipc_kobject_type_t * type)200 iokit_lookup_raw_current_task(mach_port_name_t name, ipc_kobject_type_t * type)
201 {
202 	ipc_port_t port = NULL;
203 	if (name && MACH_PORT_VALID(name)) {
204 		kern_return_t kr = ipc_object_copyin(current_space(), name, MACH_MSG_TYPE_COPY_SEND, (ipc_object_t *)&port, 0, NULL, IPC_OBJECT_COPYIN_FLAGS_NONE);
205 		if (kr == KERN_SUCCESS) {
206 			assert(IP_VALID(port));
207 			assert(ip_active(port));
208 			if (type != NULL) {
209 				*type = ip_kotype(port);
210 			}
211 		}
212 	}
213 
214 	return port;
215 }
216 
217 EXTERN void
iokit_retain_port(ipc_port_t port)218 iokit_retain_port( ipc_port_t port )
219 {
220 	ipc_port_reference( port );
221 }
222 
223 EXTERN void
iokit_release_port(ipc_port_t port)224 iokit_release_port( ipc_port_t port )
225 {
226 	ipc_port_release( port );
227 }
228 
229 EXTERN void
iokit_release_port_send(ipc_port_t port)230 iokit_release_port_send( ipc_port_t port )
231 {
232 	ipc_port_release_send( port );
233 }
234 
235 /*
236  * Get the port for a device.
237  * Consumes a device reference; produces a naked send right.
238  */
239 
240 static ipc_port_t
iokit_make_port_of_type(io_object_t obj,ipc_kobject_type_t type)241 iokit_make_port_of_type(io_object_t obj, ipc_kobject_type_t type)
242 {
243 	ipc_port_t  port;
244 	ipc_port_t  sendPort;
245 	ipc_kobject_t kobj;
246 
247 	if (obj == NULL) {
248 		return IP_NULL;
249 	}
250 
251 	port = iokit_port_for_object(obj, type, &kobj);
252 	if (port) {
253 		sendPort = ipc_kobject_make_send( port, kobj, type );
254 		iokit_release_port( port );
255 	} else {
256 		sendPort = IP_NULL;
257 	}
258 
259 	iokit_remove_reference( obj );
260 
261 	return sendPort;
262 }
263 
264 MIGEXTERN ipc_port_t
iokit_make_object_port(io_object_t obj)265 iokit_make_object_port(
266 	io_object_t     obj )
267 {
268 	return iokit_make_port_of_type(obj, IKOT_IOKIT_OBJECT);
269 }
270 
271 MIGEXTERN ipc_port_t
iokit_make_connect_port(io_object_t obj)272 iokit_make_connect_port(
273 	io_object_t     obj )
274 {
275 	return iokit_make_port_of_type(obj, IKOT_IOKIT_CONNECT);
276 }
277 
278 MIGEXTERN ipc_port_t
iokit_make_ident_port(io_object_t obj)279 iokit_make_ident_port(
280 	io_object_t     obj )
281 {
282 	return iokit_make_port_of_type(obj, IKOT_IOKIT_IDENT);
283 }
284 
285 EXTERN ipc_port_t
iokit_alloc_object_port(io_kobject_t obj,ipc_kobject_type_t type)286 iokit_alloc_object_port( io_kobject_t obj, ipc_kobject_type_t type )
287 {
288 	/* Allocate port, keeping a reference for it. */
289 	ipc_kobject_alloc_options_t options = IPC_KOBJECT_ALLOC_NSREQUEST;
290 	if (type == IKOT_IOKIT_CONNECT) {
291 		options |= IPC_KOBJECT_ALLOC_IMMOVABLE_SEND;
292 	}
293 	if (type == IKOT_UEXT_OBJECT) {
294 		ipc_label_t label = IPC_LABEL_DEXT;
295 		return ipc_kobject_alloc_labeled_port((ipc_kobject_t) obj, type, label, options);
296 	} else {
297 		return ipc_kobject_alloc_port((ipc_kobject_t) obj, type, options);
298 	}
299 }
300 
301 EXTERN void
iokit_remove_object_port(ipc_port_t port,ipc_kobject_type_t type)302 iokit_remove_object_port( ipc_port_t port, ipc_kobject_type_t type )
303 {
304 	ipc_kobject_disable(port, type);
305 }
306 
307 EXTERN kern_return_t
iokit_destroy_object_port(ipc_port_t port,ipc_kobject_type_t type)308 iokit_destroy_object_port( ipc_port_t port, ipc_kobject_type_t type )
309 {
310 	ipc_kobject_dealloc_port(port, 0, type);
311 	return KERN_SUCCESS;
312 }
313 
314 EXTERN ipc_kobject_type_t
iokit_port_type(ipc_port_t port)315 iokit_port_type(ipc_port_t port)
316 {
317 	return ip_kotype(port);
318 }
319 
320 EXTERN mach_port_name_t
iokit_make_send_right(task_t task,io_object_t obj,ipc_kobject_type_t type)321 iokit_make_send_right( task_t task, io_object_t obj, ipc_kobject_type_t type )
322 {
323 	ipc_port_t          port;
324 	ipc_port_t          sendPort;
325 	mach_port_name_t    name = 0;
326 	ipc_kobject_t       kobj;
327 
328 	if (obj == NULL) {
329 		return MACH_PORT_NULL;
330 	}
331 
332 	port = iokit_port_for_object( obj, type, &kobj );
333 	if (port) {
334 		sendPort = ipc_kobject_make_send( port, kobj, type );
335 		iokit_release_port( port );
336 	} else {
337 		sendPort = IP_NULL;
338 	}
339 
340 	if (IP_VALID( sendPort )) {
341 		kern_return_t   kr;
342 		// Remove once <rdar://problem/45522961> is fixed.
343 		// We need to make ith_knote NULL as ipc_object_copyout() uses
344 		// thread-argument-passing and its value should not be garbage
345 		current_thread()->ith_knote = ITH_KNOTE_NULL;
346 		kr = ipc_object_copyout( task->itk_space, ip_to_object(sendPort),
347 		    MACH_MSG_TYPE_PORT_SEND, IPC_OBJECT_COPYOUT_FLAGS_NONE, NULL, NULL, &name);
348 		if (kr != KERN_SUCCESS) {
349 			name = MACH_PORT_NULL;
350 		}
351 	} else if (sendPort == IP_NULL) {
352 		name = MACH_PORT_NULL;
353 	} else if (sendPort == IP_DEAD) {
354 		name = MACH_PORT_DEAD;
355 	}
356 
357 	return name;
358 }
359 
360 EXTERN kern_return_t
iokit_mod_send_right(task_t task,mach_port_name_t name,mach_port_delta_t delta)361 iokit_mod_send_right( task_t task, mach_port_name_t name, mach_port_delta_t delta )
362 {
363 	return mach_port_mod_refs( task->itk_space, name, MACH_PORT_RIGHT_SEND, delta );
364 }
365 
366 /*
367  * Handle the No-More_Senders notification generated from a device port destroy.
368  * Since there are no longer any tasks which hold a send right to this device
369  * port a NMS notification has been generated.
370  */
371 
372 static void
iokit_no_senders(ipc_port_t port,mach_port_mscount_t mscount)373 iokit_no_senders( ipc_port_t port, mach_port_mscount_t mscount )
374 {
375 	io_object_t         obj = NULL;
376 	io_kobject_t        kobj = NULL;
377 	ipc_kobject_type_t  type = IKOT_NONE;
378 
379 	// convert a port to io_object_t.
380 	if (IP_VALID(port)) {
381 		ip_mq_lock(port);
382 		if (ip_active(port)) {
383 			type = ip_kotype( port );
384 			assert((IKOT_IOKIT_OBJECT == type)
385 			    || (IKOT_IOKIT_CONNECT == type)
386 			    || (IKOT_IOKIT_IDENT == type)
387 			    || (IKOT_UEXT_OBJECT == type));
388 			kobj = ipc_kobject_get_locked(port, type);
389 			if (kobj) {
390 				iokit_kobject_retain(kobj);
391 			}
392 		}
393 		ip_mq_unlock(port);
394 		if (kobj) {
395 			// IKOT_IOKIT_OBJECT since iokit_remove_reference() follows
396 			obj = iokit_copy_object_for_consumed_kobject(kobj, IKOT_IOKIT_OBJECT);
397 		}
398 	}
399 
400 	if (obj) {
401 		while (iokit_client_died( obj, port, type, &mscount ) != KERN_SUCCESS) {
402 			kern_return_t kr;
403 
404 			/* Re-request no-senders notifications on the port (if still active) */
405 			kr = ipc_kobject_nsrequest(port, mscount + 1, &mscount);
406 			if (kr != KERN_FAILURE) {
407 				break;
408 			}
409 			/*
410 			 * port has no outstanding rights or pending make-sends,
411 			 * and the notification would fire recursively, try again.
412 			 */
413 		}
414 
415 		iokit_remove_reference( obj );
416 	}
417 }
418 
419 
420 kern_return_t
iokit_label_dext_task(task_t task)421 iokit_label_dext_task(task_t task)
422 {
423 	return ipc_space_add_label(task->itk_space, IPC_LABEL_DEXT);
424 }
425 
426 /*
427  *	Routine:	iokit_clear_registered_ports
428  *	Purpose:
429  *		Clean up a task's registered IOKit kobject ports.
430  *	Conditions:
431  *		Nothing locked.
432  */
433 void
iokit_clear_registered_ports(task_t task)434 iokit_clear_registered_ports(
435 	task_t task)
436 {
437 	mach_port_t port;
438 	ipc_kobject_type_t type;
439 
440 	itk_lock(task);
441 	for (int i = 0; i < TASK_PORT_REGISTER_MAX; i++) {
442 		port = task->itk_registered[i];
443 		if (!IP_VALID(port)) {
444 			continue;
445 		}
446 		type = ip_kotype( port );
447 		if ((IKOT_IOKIT_OBJECT == type)
448 		    || (IKOT_IOKIT_CONNECT == type)
449 		    || (IKOT_IOKIT_IDENT == type)
450 		    || (IKOT_UEXT_OBJECT == type)) {
451 			ipc_port_release_send(port);
452 			task->itk_registered[i] = IP_NULL;
453 		}
454 	}
455 	itk_unlock(task);
456 }
457 
458 /* need to create a pmap function to generalize */
459 unsigned int
IODefaultCacheBits(addr64_t pa)460 IODefaultCacheBits(addr64_t pa)
461 {
462 	return pmap_cache_attributes((ppnum_t)(pa >> PAGE_SHIFT));
463 }
464 
465 kern_return_t
IOMapPages(vm_map_t map,mach_vm_address_t va,mach_vm_address_t pa,mach_vm_size_t length,unsigned int options)466 IOMapPages(vm_map_t map, mach_vm_address_t va, mach_vm_address_t pa,
467     mach_vm_size_t length, unsigned int options)
468 {
469 	vm_prot_t    prot;
470 	unsigned int flags;
471 	ppnum_t      pagenum;
472 	pmap_t       pmap = map->pmap;
473 
474 	prot = (options & kIOMapReadOnly)
475 	    ? VM_PROT_READ : (VM_PROT_READ | VM_PROT_WRITE);
476 
477 	pagenum = (ppnum_t)atop_64(pa);
478 
479 	switch (options & kIOMapCacheMask) {                    /* What cache mode do we need? */
480 	case kIOMapDefaultCache:
481 	default:
482 		flags = IODefaultCacheBits(pa);
483 		break;
484 
485 	case kIOMapInhibitCache:
486 		flags = VM_WIMG_IO;
487 		break;
488 
489 	case kIOMapWriteThruCache:
490 		flags = VM_WIMG_WTHRU;
491 		break;
492 
493 	case kIOMapWriteCombineCache:
494 		flags = VM_WIMG_WCOMB;
495 		break;
496 
497 	case kIOMapCopybackCache:
498 		flags = VM_WIMG_COPYBACK;
499 		break;
500 
501 	case kIOMapCopybackInnerCache:
502 		flags = VM_WIMG_INNERWBACK;
503 		break;
504 
505 	case kIOMapPostedWrite:
506 		flags = VM_WIMG_POSTED;
507 		break;
508 
509 	case kIOMapRealTimeCache:
510 		flags = VM_WIMG_RT;
511 		break;
512 	}
513 
514 	pmap_set_cache_attributes(pagenum, flags);
515 
516 	vm_map_set_cache_attr(map, (vm_map_offset_t)va);
517 
518 
519 	// Set up a block mapped area
520 	return pmap_map_block(pmap, va, pagenum, (uint32_t) atop_64(round_page_64(length)), prot, 0, 0);
521 }
522 
523 kern_return_t
IOUnmapPages(vm_map_t map,mach_vm_address_t va,mach_vm_size_t length)524 IOUnmapPages(vm_map_t map, mach_vm_address_t va, mach_vm_size_t length)
525 {
526 	pmap_t      pmap = map->pmap;
527 
528 	pmap_remove(pmap, trunc_page_64(va), round_page_64(va + length));
529 
530 	return KERN_SUCCESS;
531 }
532 
533 kern_return_t
IOProtectCacheMode(vm_map_t __unused map,mach_vm_address_t __unused va,mach_vm_size_t __unused length,unsigned int __unused options)534 IOProtectCacheMode(vm_map_t __unused map, mach_vm_address_t __unused va,
535     mach_vm_size_t __unused length, unsigned int __unused options)
536 {
537 	mach_vm_size_t off;
538 	vm_prot_t      prot;
539 	unsigned int   flags;
540 	pmap_t         pmap = map->pmap;
541 	pmap_flush_context  pmap_flush_context_storage;
542 	boolean_t           delayed_pmap_flush = FALSE;
543 
544 	prot = (options & kIOMapReadOnly)
545 	    ? VM_PROT_READ : (VM_PROT_READ | VM_PROT_WRITE);
546 
547 	switch (options & kIOMapCacheMask) {
548 	// what cache mode do we need?
549 	case kIOMapDefaultCache:
550 	default:
551 		return KERN_INVALID_ARGUMENT;
552 
553 	case kIOMapInhibitCache:
554 		flags = VM_WIMG_IO;
555 		break;
556 
557 	case kIOMapWriteThruCache:
558 		flags = VM_WIMG_WTHRU;
559 		break;
560 
561 	case kIOMapWriteCombineCache:
562 		flags = VM_WIMG_WCOMB;
563 		break;
564 
565 	case kIOMapCopybackCache:
566 		flags = VM_WIMG_COPYBACK;
567 		break;
568 
569 	case kIOMapCopybackInnerCache:
570 		flags = VM_WIMG_INNERWBACK;
571 		break;
572 
573 	case kIOMapPostedWrite:
574 		flags = VM_WIMG_POSTED;
575 		break;
576 
577 	case kIOMapRealTimeCache:
578 		flags = VM_WIMG_RT;
579 		break;
580 	}
581 
582 	pmap_flush_context_init(&pmap_flush_context_storage);
583 	delayed_pmap_flush = FALSE;
584 
585 	//  enter each page's physical address in the target map
586 	for (off = 0; off < length; off += page_size) {
587 		ppnum_t ppnum = pmap_find_phys(pmap, va + off);
588 		if (ppnum) {
589 			pmap_enter_options(pmap, va + off, ppnum, prot, VM_PROT_NONE, flags, TRUE,
590 			    PMAP_OPTIONS_NOFLUSH, (void *)&pmap_flush_context_storage, PMAP_MAPPING_TYPE_INFER);
591 			delayed_pmap_flush = TRUE;
592 		}
593 	}
594 	if (delayed_pmap_flush == TRUE) {
595 		pmap_flush(&pmap_flush_context_storage);
596 	}
597 
598 	return KERN_SUCCESS;
599 }
600 
601 ppnum_t
IOGetLastPageNumber(void)602 IOGetLastPageNumber(void)
603 {
604 #if __i386__ || __x86_64__
605 	ppnum_t  lastPage, highest = 0;
606 	unsigned int idx;
607 
608 	for (idx = 0; idx < pmap_memory_region_count; idx++) {
609 		lastPage = pmap_memory_regions[idx].end - 1;
610 		if (lastPage > highest) {
611 			highest = lastPage;
612 		}
613 	}
614 	return highest;
615 #elif __arm64__
616 	return 0;
617 #else
618 #error unknown arch
619 #endif
620 }
621 
622 
623 void IOGetTime( mach_timespec_t * clock_time);
624 void
IOGetTime(mach_timespec_t * clock_time)625 IOGetTime( mach_timespec_t * clock_time)
626 {
627 	clock_sec_t sec;
628 	clock_nsec_t nsec;
629 	clock_get_system_nanotime(&sec, &nsec);
630 	clock_time->tv_sec = (typeof(clock_time->tv_sec))sec;
631 	clock_time->tv_nsec = nsec;
632 }
633