/* * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in * compliance with the License. The rights granted to you under the License * may not be used to create, or enable the creation or redistribution of, * unlawful or unlicensed copies of an Apple operating system, or to * circumvent, violate, or enable the circumvention or violation of, any * terms of an Apple operating system software license agreement. * * Please obtain a copy of the License at * http://www.opensource.apple.com/apsl/ and read it before using this file. * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. * Please see the License for the specific language governing rights and * limitations under the License. * * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ #include #include #include #include #include #include //#include #include #include /* spl definitions */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #if defined(__i386__) || defined(__x86_64__) #include #endif #if defined(__arm64__) #include #endif #include #define EXTERN #define MIGEXTERN static void iokit_no_senders( ipc_port_t port, mach_port_mscount_t mscount ); /* * Lifetime: * - non lazy port with no-more senders * - can be destroyed by iokit_destroy_object_port * */ IPC_KOBJECT_DEFINE(IKOT_IOKIT_IDENT, .iko_op_no_senders = iokit_no_senders); IPC_KOBJECT_DEFINE(IKOT_IOKIT_OBJECT, .iko_op_no_senders = iokit_no_senders); IPC_KOBJECT_DEFINE(IKOT_IOKIT_CONNECT, .iko_op_no_senders = iokit_no_senders); IPC_KOBJECT_DEFINE(IKOT_UEXT_OBJECT, .iko_op_no_senders = iokit_no_senders); /* * Lookup a device by its port. * Doesn't consume the naked send right; produces a device reference. */ io_object_t iokit_lookup_io_object(ipc_port_t port, ipc_kobject_type_t type) { io_object_t obj = NULL; io_kobject_t kobj = NULL; if (!IP_VALID(port)) { return NULL; } ip_mq_lock(port); if (ip_active(port)) { kobj = ipc_kobject_get_locked(port, type); if (kobj) { iokit_kobject_retain(kobj); } } ip_mq_unlock(port); if (kobj) { obj = iokit_copy_object_for_consumed_kobject(kobj, type); } return obj; } MIGEXTERN io_object_t iokit_lookup_object_port( ipc_port_t port) { return iokit_lookup_io_object(port, IKOT_IOKIT_OBJECT); } MIGEXTERN io_object_t iokit_lookup_connect_port( ipc_port_t port) { return iokit_lookup_io_object(port, IKOT_IOKIT_CONNECT); } MIGEXTERN io_object_t iokit_lookup_ident_port( ipc_port_t port) { return iokit_lookup_io_object(port, IKOT_IOKIT_IDENT); } MIGEXTERN io_object_t iokit_lookup_uext_object_port( ipc_port_t port) { return iokit_lookup_io_object(port, IKOT_UEXT_OBJECT); } static io_object_t iokit_lookup_object_in_space_with_port_name(mach_port_name_t name, ipc_kobject_type_t type, ipc_space_t space) { io_object_t obj = NULL; io_kobject_t kobj; if (name && MACH_PORT_VALID(name)) { ipc_port_t port; kern_return_t kr; kr = ipc_port_translate_send(space, name, &port); if (kr == KERN_SUCCESS) { assert(IP_VALID(port)); assert(ip_active(port)); kobj = ipc_kobject_get_locked(port, type); if (kobj) { iokit_kobject_retain(kobj); } ip_mq_unlock(port); if (kobj) { obj = iokit_copy_object_for_consumed_kobject(kobj, type); } } } return obj; } EXTERN io_object_t iokit_lookup_object_with_port_name(mach_port_name_t name, ipc_kobject_type_t type, task_t task) { return iokit_lookup_object_in_space_with_port_name(name, type, task->itk_space); } EXTERN io_object_t iokit_lookup_connect_ref_current_task(mach_port_name_t name) { return iokit_lookup_object_in_space_with_port_name(name, IKOT_IOKIT_CONNECT, current_space()); } EXTERN io_object_t iokit_lookup_uext_ref_current_task(mach_port_name_t name) { return iokit_lookup_object_in_space_with_port_name(name, IKOT_UEXT_OBJECT, current_space()); } /* * Look up a port given a port name. * This returns the port unlocked with a +1 send right. * Release with iokit_release_port_send() */ EXTERN ipc_port_t iokit_lookup_raw_current_task(mach_port_name_t name, ipc_kobject_type_t * type) { ipc_port_t port = NULL; if (name && MACH_PORT_VALID(name)) { kern_return_t kr = ipc_object_copyin(current_space(), name, MACH_MSG_TYPE_COPY_SEND, (ipc_object_t *)&port, 0, NULL, IPC_OBJECT_COPYIN_FLAGS_NONE); if (kr == KERN_SUCCESS) { assert(IP_VALID(port)); assert(ip_active(port)); if (type != NULL) { *type = ip_kotype(port); } } } return port; } EXTERN void iokit_retain_port( ipc_port_t port ) { ipc_port_reference( port ); } EXTERN void iokit_release_port( ipc_port_t port ) { ipc_port_release( port ); } EXTERN void iokit_release_port_send( ipc_port_t port ) { ipc_port_release_send( port ); } /* * Get the port for a device. * Consumes a device reference; produces a naked send right. */ static ipc_port_t iokit_make_port_of_type(io_object_t obj, ipc_kobject_type_t type) { ipc_port_t port; ipc_port_t sendPort; ipc_kobject_t kobj; if (obj == NULL) { return IP_NULL; } port = iokit_port_for_object(obj, type, &kobj); if (port) { sendPort = ipc_kobject_make_send( port, kobj, type ); iokit_release_port( port ); } else { sendPort = IP_NULL; } iokit_remove_reference( obj ); return sendPort; } MIGEXTERN ipc_port_t iokit_make_object_port( io_object_t obj ) { return iokit_make_port_of_type(obj, IKOT_IOKIT_OBJECT); } MIGEXTERN ipc_port_t iokit_make_connect_port( io_object_t obj ) { return iokit_make_port_of_type(obj, IKOT_IOKIT_CONNECT); } MIGEXTERN ipc_port_t iokit_make_ident_port( io_object_t obj ) { return iokit_make_port_of_type(obj, IKOT_IOKIT_IDENT); } EXTERN ipc_port_t iokit_alloc_object_port( io_kobject_t obj, ipc_kobject_type_t type ) { /* Allocate port, keeping a reference for it. */ ipc_kobject_alloc_options_t options = IPC_KOBJECT_ALLOC_NSREQUEST; if (type == IKOT_IOKIT_CONNECT) { options |= IPC_KOBJECT_ALLOC_IMMOVABLE_SEND; } if (type == IKOT_UEXT_OBJECT) { ipc_label_t label = IPC_LABEL_DEXT; return ipc_kobject_alloc_labeled_port((ipc_kobject_t) obj, type, label, options); } else { return ipc_kobject_alloc_port((ipc_kobject_t) obj, type, options); } } EXTERN void iokit_remove_object_port( ipc_port_t port, ipc_kobject_type_t type ) { ipc_kobject_disable(port, type); } EXTERN kern_return_t iokit_destroy_object_port( ipc_port_t port, ipc_kobject_type_t type ) { ipc_kobject_dealloc_port(port, 0, type); return KERN_SUCCESS; } EXTERN ipc_kobject_type_t iokit_port_type(ipc_port_t port) { return ip_kotype(port); } EXTERN mach_port_name_t iokit_make_send_right( task_t task, io_object_t obj, ipc_kobject_type_t type ) { ipc_port_t port; ipc_port_t sendPort; mach_port_name_t name = 0; ipc_kobject_t kobj; if (obj == NULL) { return MACH_PORT_NULL; } port = iokit_port_for_object( obj, type, &kobj ); if (port) { sendPort = ipc_kobject_make_send( port, kobj, type ); iokit_release_port( port ); } else { sendPort = IP_NULL; } if (IP_VALID( sendPort )) { kern_return_t kr; // Remove once is fixed. // We need to make ith_knote NULL as ipc_object_copyout() uses // thread-argument-passing and its value should not be garbage current_thread()->ith_knote = ITH_KNOTE_NULL; kr = ipc_object_copyout( task->itk_space, ip_to_object(sendPort), MACH_MSG_TYPE_PORT_SEND, IPC_OBJECT_COPYOUT_FLAGS_NONE, NULL, NULL, &name); if (kr != KERN_SUCCESS) { name = MACH_PORT_NULL; } } else if (sendPort == IP_NULL) { name = MACH_PORT_NULL; } else if (sendPort == IP_DEAD) { name = MACH_PORT_DEAD; } return name; } EXTERN kern_return_t iokit_mod_send_right( task_t task, mach_port_name_t name, mach_port_delta_t delta ) { return mach_port_mod_refs( task->itk_space, name, MACH_PORT_RIGHT_SEND, delta ); } /* * Handle the No-More_Senders notification generated from a device port destroy. * Since there are no longer any tasks which hold a send right to this device * port a NMS notification has been generated. */ static void iokit_no_senders( ipc_port_t port, mach_port_mscount_t mscount ) { io_object_t obj = NULL; io_kobject_t kobj = NULL; ipc_kobject_type_t type = IKOT_NONE; // convert a port to io_object_t. if (IP_VALID(port)) { ip_mq_lock(port); if (ip_active(port)) { type = ip_kotype( port ); assert((IKOT_IOKIT_OBJECT == type) || (IKOT_IOKIT_CONNECT == type) || (IKOT_IOKIT_IDENT == type) || (IKOT_UEXT_OBJECT == type)); kobj = ipc_kobject_get_locked(port, type); if (kobj) { iokit_kobject_retain(kobj); } } ip_mq_unlock(port); if (kobj) { // IKOT_IOKIT_OBJECT since iokit_remove_reference() follows obj = iokit_copy_object_for_consumed_kobject(kobj, IKOT_IOKIT_OBJECT); } } if (obj) { while (iokit_client_died( obj, port, type, &mscount ) != KERN_SUCCESS) { kern_return_t kr; /* Re-request no-senders notifications on the port (if still active) */ kr = ipc_kobject_nsrequest(port, mscount + 1, &mscount); if (kr != KERN_FAILURE) { break; } /* * port has no outstanding rights or pending make-sends, * and the notification would fire recursively, try again. */ } iokit_remove_reference( obj ); } } kern_return_t iokit_label_dext_task(task_t task) { return ipc_space_add_label(task->itk_space, IPC_LABEL_DEXT); } /* * Routine: iokit_clear_registered_ports * Purpose: * Clean up a task's registered IOKit kobject ports. * Conditions: * Nothing locked. */ void iokit_clear_registered_ports( task_t task) { mach_port_t port; ipc_kobject_type_t type; itk_lock(task); for (int i = 0; i < TASK_PORT_REGISTER_MAX; i++) { port = task->itk_registered[i]; if (!IP_VALID(port)) { continue; } type = ip_kotype( port ); if ((IKOT_IOKIT_OBJECT == type) || (IKOT_IOKIT_CONNECT == type) || (IKOT_IOKIT_IDENT == type) || (IKOT_UEXT_OBJECT == type)) { ipc_port_release_send(port); task->itk_registered[i] = IP_NULL; } } itk_unlock(task); } /* need to create a pmap function to generalize */ unsigned int IODefaultCacheBits(addr64_t pa) { return pmap_cache_attributes((ppnum_t)(pa >> PAGE_SHIFT)); } kern_return_t IOMapPages(vm_map_t map, mach_vm_address_t va, mach_vm_address_t pa, mach_vm_size_t length, unsigned int options) { vm_prot_t prot; unsigned int flags; ppnum_t pagenum; pmap_t pmap = map->pmap; prot = (options & kIOMapReadOnly) ? VM_PROT_READ : (VM_PROT_READ | VM_PROT_WRITE); pagenum = (ppnum_t)atop_64(pa); switch (options & kIOMapCacheMask) { /* What cache mode do we need? */ case kIOMapDefaultCache: default: flags = IODefaultCacheBits(pa); break; case kIOMapInhibitCache: flags = VM_WIMG_IO; break; case kIOMapWriteThruCache: flags = VM_WIMG_WTHRU; break; case kIOMapWriteCombineCache: flags = VM_WIMG_WCOMB; break; case kIOMapCopybackCache: flags = VM_WIMG_COPYBACK; break; case kIOMapCopybackInnerCache: flags = VM_WIMG_INNERWBACK; break; case kIOMapPostedWrite: flags = VM_WIMG_POSTED; break; case kIOMapRealTimeCache: flags = VM_WIMG_RT; break; } pmap_set_cache_attributes(pagenum, flags); vm_map_set_cache_attr(map, (vm_map_offset_t)va); // Set up a block mapped area return pmap_map_block(pmap, va, pagenum, (uint32_t) atop_64(round_page_64(length)), prot, 0, 0); } kern_return_t IOUnmapPages(vm_map_t map, mach_vm_address_t va, mach_vm_size_t length) { pmap_t pmap = map->pmap; pmap_remove(pmap, trunc_page_64(va), round_page_64(va + length)); return KERN_SUCCESS; } kern_return_t IOProtectCacheMode(vm_map_t __unused map, mach_vm_address_t __unused va, mach_vm_size_t __unused length, unsigned int __unused options) { mach_vm_size_t off; vm_prot_t prot; unsigned int flags; pmap_t pmap = map->pmap; pmap_flush_context pmap_flush_context_storage; boolean_t delayed_pmap_flush = FALSE; prot = (options & kIOMapReadOnly) ? VM_PROT_READ : (VM_PROT_READ | VM_PROT_WRITE); switch (options & kIOMapCacheMask) { // what cache mode do we need? case kIOMapDefaultCache: default: return KERN_INVALID_ARGUMENT; case kIOMapInhibitCache: flags = VM_WIMG_IO; break; case kIOMapWriteThruCache: flags = VM_WIMG_WTHRU; break; case kIOMapWriteCombineCache: flags = VM_WIMG_WCOMB; break; case kIOMapCopybackCache: flags = VM_WIMG_COPYBACK; break; case kIOMapCopybackInnerCache: flags = VM_WIMG_INNERWBACK; break; case kIOMapPostedWrite: flags = VM_WIMG_POSTED; break; case kIOMapRealTimeCache: flags = VM_WIMG_RT; break; } pmap_flush_context_init(&pmap_flush_context_storage); delayed_pmap_flush = FALSE; // enter each page's physical address in the target map for (off = 0; off < length; off += page_size) { ppnum_t ppnum = pmap_find_phys(pmap, va + off); if (ppnum) { pmap_enter_options(pmap, va + off, ppnum, prot, VM_PROT_NONE, flags, TRUE, PMAP_OPTIONS_NOFLUSH, (void *)&pmap_flush_context_storage); delayed_pmap_flush = TRUE; } } if (delayed_pmap_flush == TRUE) { pmap_flush(&pmap_flush_context_storage); } return KERN_SUCCESS; } ppnum_t IOGetLastPageNumber(void) { #if __i386__ || __x86_64__ ppnum_t lastPage, highest = 0; unsigned int idx; for (idx = 0; idx < pmap_memory_region_count; idx++) { lastPage = pmap_memory_regions[idx].end - 1; if (lastPage > highest) { highest = lastPage; } } return highest; #elif __arm64__ return 0; #else #error unknown arch #endif } void IOGetTime( mach_timespec_t * clock_time); void IOGetTime( mach_timespec_t * clock_time) { clock_sec_t sec; clock_nsec_t nsec; clock_get_system_nanotime(&sec, &nsec); clock_time->tv_sec = (typeof(clock_time->tv_sec))sec; clock_time->tv_nsec = nsec; }