xref: /xnu-8020.140.41/iokit/Kernel/IOCPU.cpp (revision 27b03b360a988dfd3dfdf34262bb0042026747cc)
1 /*
2  * Copyright (c) 1999-2016 Apple Inc.  All rights reserved.
3  *
4  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5  *
6  * This file contains Original Code and/or Modifications of Original Code
7  * as defined in and that are subject to the Apple Public Source License
8  * Version 2.0 (the 'License'). You may not use this file except in
9  * compliance with the License. The rights granted to you under the License
10  * may not be used to create, or enable the creation or redistribution of,
11  * unlawful or unlicensed copies of an Apple operating system, or to
12  * circumvent, violate, or enable the circumvention or violation of, any
13  * terms of an Apple operating system software license agreement.
14  *
15  * Please obtain a copy of the License at
16  * http://www.opensource.apple.com/apsl/ and read it before using this file.
17  *
18  * The Original Code and all software distributed under the License are
19  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23  * Please see the License for the specific language governing rights and
24  * limitations under the License.
25  *
26  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27  */
28 
29 #define IOKIT_ENABLE_SHARED_PTR
30 
31 extern "C" {
32 #include <pexpert/pexpert.h>
33 #include <kern/cpu_number.h>
34 extern void kperf_kernel_configure(char *);
35 }
36 
37 #include <machine/machine_routines.h>
38 #include <IOKit/IOLib.h>
39 #include <IOKit/IOPlatformExpert.h>
40 #include <IOKit/pwr_mgt/RootDomain.h>
41 #include <IOKit/pwr_mgt/IOPMPrivate.h>
42 #include <libkern/c++/OSSharedPtr.h>
43 #include <IOKit/IOUserClient.h>
44 #include <IOKit/IOKitKeysPrivate.h>
45 #include <IOKit/IOCPU.h>
46 #include "IOKitKernelInternal.h"
47 
48 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
49 
50 #include <kern/queue.h>
51 #include <kern/sched_prim.h>
52 
53 extern "C" void console_suspend();
54 extern "C" void console_resume();
55 extern "C" void sched_override_recommended_cores_for_sleep(void);
56 extern "C" void sched_restore_recommended_cores_after_sleep(void);
57 
58 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
59 
60 static IOLock *gIOCPUsLock;
61 static OSSharedPtr<OSArray> gIOCPUs;
62 static OSSharedPtr<const OSSymbol> gIOCPUStateKey;
63 static OSSharedPtr<OSString> gIOCPUStateNames[kIOCPUStateCount];
64 
65 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
66 
67 #if !USE_APPLEARMSMP
68 
69 void
IOCPUInitialize(void)70 IOCPUInitialize(void)
71 {
72 	gIOCPUsLock = IOLockAlloc();
73 	gIOCPUs     = OSArray::withCapacity(1);
74 
75 	gIOCPUStateKey = OSSymbol::withCStringNoCopy("IOCPUState");
76 
77 	gIOCPUStateNames[kIOCPUStateUnregistered] =
78 	    OSString::withCStringNoCopy("Unregistered");
79 	gIOCPUStateNames[kIOCPUStateUninitalized] =
80 	    OSString::withCStringNoCopy("Uninitalized");
81 	gIOCPUStateNames[kIOCPUStateStopped] =
82 	    OSString::withCStringNoCopy("Stopped");
83 	gIOCPUStateNames[kIOCPUStateRunning] =
84 	    OSString::withCStringNoCopy("Running");
85 }
86 
87 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
88 
89 kern_return_t
PE_cpu_start(cpu_id_t target,vm_offset_t start_paddr,vm_offset_t arg_paddr)90 PE_cpu_start(cpu_id_t target,
91     vm_offset_t start_paddr, vm_offset_t arg_paddr)
92 {
93 	IOCPU *targetCPU = (IOCPU *)target;
94 
95 	if (targetCPU == NULL) {
96 		return KERN_FAILURE;
97 	}
98 	return targetCPU->startCPU(start_paddr, arg_paddr);
99 }
100 
101 void
PE_cpu_halt(cpu_id_t target)102 PE_cpu_halt(cpu_id_t target)
103 {
104 	IOCPU *targetCPU = (IOCPU *)target;
105 
106 	targetCPU->haltCPU();
107 }
108 
109 void
PE_cpu_signal(cpu_id_t source,cpu_id_t target)110 PE_cpu_signal(cpu_id_t source, cpu_id_t target)
111 {
112 	IOCPU *sourceCPU = (IOCPU *)source;
113 	IOCPU *targetCPU = (IOCPU *)target;
114 
115 	sourceCPU->signalCPU(targetCPU);
116 }
117 
118 void
PE_cpu_signal_deferred(cpu_id_t source,cpu_id_t target)119 PE_cpu_signal_deferred(cpu_id_t source, cpu_id_t target)
120 {
121 	IOCPU *sourceCPU = (IOCPU *)source;
122 	IOCPU *targetCPU = (IOCPU *)target;
123 
124 	sourceCPU->signalCPUDeferred(targetCPU);
125 }
126 
127 void
PE_cpu_signal_cancel(cpu_id_t source,cpu_id_t target)128 PE_cpu_signal_cancel(cpu_id_t source, cpu_id_t target)
129 {
130 	IOCPU *sourceCPU = (IOCPU *)source;
131 	IOCPU *targetCPU = (IOCPU *)target;
132 
133 	sourceCPU->signalCPUCancel(targetCPU);
134 }
135 
136 void
PE_cpu_machine_init(cpu_id_t target,boolean_t bootb)137 PE_cpu_machine_init(cpu_id_t target, boolean_t bootb)
138 {
139 	IOCPU *targetCPU = OSDynamicCast(IOCPU, (OSObject *)target);
140 
141 	if (targetCPU == NULL) {
142 		panic("%s: invalid target CPU %p", __func__, target);
143 	}
144 
145 	targetCPU->initCPU(bootb);
146 #if defined(__arm__) || defined(__arm64__)
147 	if (!bootb && (targetCPU->getCPUNumber() == (UInt32)master_cpu)) {
148 		ml_set_is_quiescing(false);
149 	}
150 #endif /* defined(__arm__) || defined(__arm64__) */
151 }
152 
153 void
PE_cpu_machine_quiesce(cpu_id_t target)154 PE_cpu_machine_quiesce(cpu_id_t target)
155 {
156 	IOCPU *targetCPU = (IOCPU*)target;
157 #if defined(__arm__) || defined(__arm64__)
158 	if (targetCPU->getCPUNumber() == (UInt32)master_cpu) {
159 		ml_set_is_quiescing(true);
160 	}
161 #endif /* defined(__arm__) || defined(__arm64__) */
162 	targetCPU->quiesceCPU();
163 }
164 
165 #if defined(__arm__) || defined(__arm64__)
166 static perfmon_interrupt_handler_func pmi_handler = NULL;
167 
168 kern_return_t
PE_cpu_perfmon_interrupt_install_handler(perfmon_interrupt_handler_func handler)169 PE_cpu_perfmon_interrupt_install_handler(perfmon_interrupt_handler_func handler)
170 {
171 	pmi_handler = handler;
172 
173 	return KERN_SUCCESS;
174 }
175 
176 void
PE_cpu_perfmon_interrupt_enable(cpu_id_t target,boolean_t enable)177 PE_cpu_perfmon_interrupt_enable(cpu_id_t target, boolean_t enable)
178 {
179 	IOCPU *targetCPU = (IOCPU*)target;
180 
181 	if (targetCPU == nullptr) {
182 		return;
183 	}
184 
185 	if (enable) {
186 		targetCPU->getProvider()->registerInterrupt(1, targetCPU, (IOInterruptAction)pmi_handler, NULL);
187 		targetCPU->getProvider()->enableInterrupt(1);
188 	} else {
189 		targetCPU->getProvider()->disableInterrupt(1);
190 	}
191 }
192 #endif
193 
194 #endif /* !USE_APPLEARMSMP */
195 
196 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
197 
198 #define super IOService
199 
200 OSDefineMetaClassAndAbstractStructors(IOCPU, IOService);
201 OSMetaClassDefineReservedUnused(IOCPU, 0);
202 OSMetaClassDefineReservedUnused(IOCPU, 1);
203 OSMetaClassDefineReservedUnused(IOCPU, 2);
204 OSMetaClassDefineReservedUnused(IOCPU, 3);
205 OSMetaClassDefineReservedUnused(IOCPU, 4);
206 OSMetaClassDefineReservedUnused(IOCPU, 5);
207 OSMetaClassDefineReservedUnused(IOCPU, 6);
208 OSMetaClassDefineReservedUnused(IOCPU, 7);
209 
210 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
211 
212 #if !USE_APPLEARMSMP
213 void
IOCPUSleepKernel(void)214 IOCPUSleepKernel(void)
215 {
216 #if defined(__x86_64__)
217 	extern IOCPU *currentShutdownTarget;
218 #endif
219 	unsigned int cnt, numCPUs;
220 	IOCPU *target;
221 	IOCPU *bootCPU = NULL;
222 	IOPMrootDomain  *rootDomain = IOService::getPMRootDomain();
223 
224 	printf("IOCPUSleepKernel enter\n");
225 #if defined(__arm64__)
226 	sched_override_recommended_cores_for_sleep();
227 #endif
228 
229 	rootDomain->tracePoint( kIOPMTracePointSleepPlatformActions );
230 	IOPlatformActionsPreSleep();
231 	rootDomain->tracePoint( kIOPMTracePointSleepCPUs );
232 
233 	numCPUs = gIOCPUs->getCount();
234 #if defined(__x86_64__)
235 	currentShutdownTarget = NULL;
236 #endif
237 
238 	integer_t old_pri;
239 	thread_t self = current_thread();
240 
241 	/*
242 	 * We need to boost this thread's priority to the maximum kernel priority to
243 	 * ensure we can urgently preempt ANY thread currently executing on the
244 	 * target CPU.  Note that realtime threads have their own mechanism to eventually
245 	 * demote their priority below MAXPRI_KERNEL if they hog the CPU for too long.
246 	 */
247 	old_pri = thread_kern_get_pri(self);
248 	thread_kern_set_pri(self, thread_kern_get_kernel_maxpri());
249 
250 	// Sleep the CPUs.
251 	ml_set_is_quiescing(true);
252 	cnt = numCPUs;
253 	while (cnt--) {
254 		target = OSDynamicCast(IOCPU, gIOCPUs->getObject(cnt));
255 
256 		// We make certain that the bootCPU is the last to sleep
257 		// We'll skip it for now, and halt it after finishing the
258 		// non-boot CPU's.
259 		if (target->getCPUNumber() == (UInt32)master_cpu) {
260 			bootCPU = target;
261 		} else if (target->getCPUState() == kIOCPUStateRunning) {
262 #if defined(__x86_64__)
263 			currentShutdownTarget = target;
264 #endif
265 			target->haltCPU();
266 		}
267 	}
268 
269 	assert(bootCPU != NULL);
270 	assert(cpu_number() == master_cpu);
271 
272 	console_suspend();
273 
274 	rootDomain->tracePoint( kIOPMTracePointSleepPlatformDriver );
275 	rootDomain->stop_watchdog_timer();
276 
277 	/*
278 	 * Now sleep the boot CPU, including calling the kQueueQuiesce actions.
279 	 * The system sleeps here.
280 	 */
281 
282 	bootCPU->haltCPU();
283 	ml_set_is_quiescing(false);
284 
285 	/*
286 	 * The system is now coming back from sleep on the boot CPU.
287 	 * The kQueueActive actions have already been called.
288 	 */
289 
290 	rootDomain->start_watchdog_timer();
291 
292 #if defined(XNU_TARGET_OS_OSX)
293 	console_resume();
294 
295 	rootDomain->tracePoint( kIOPMTracePointWakeCPUs );
296 
297 	// Wake the other CPUs.
298 	for (cnt = 0; cnt < numCPUs; cnt++) {
299 		target = OSDynamicCast(IOCPU, gIOCPUs->getObject(cnt));
300 
301 		// Skip the already-woken boot CPU.
302 		if (target->getCPUNumber() != (UInt32)master_cpu) {
303 			if (target->getCPUState() == kIOCPUStateRunning) {
304 				panic("Spurious wakeup of cpu %u", (unsigned int)(target->getCPUNumber()));
305 			}
306 
307 			if (target->getCPUState() == kIOCPUStateStopped) {
308 				processor_start(target->getMachProcessor());
309 			}
310 		}
311 	}
312 
313 #if defined(__arm64__)
314 	sched_restore_recommended_cores_after_sleep();
315 #endif
316 
317 	rootDomain->tracePoint( kIOPMTracePointWakePlatformActions );
318 	IOPlatformActionsPostResume();
319 
320 #else /* defined(!XNU_TARGET_OS_OSX) */
321 	// Keep the old ordering around for iOS temporarily - rdar://88891040
322 
323 	rootDomain->tracePoint( kIOPMTracePointWakePlatformActions );
324 
325 	console_resume();
326 
327 	IOPlatformActionsPostResume();
328 	rootDomain->tracePoint( kIOPMTracePointWakeCPUs );
329 
330 	// Wake the other CPUs.
331 	for (cnt = 0; cnt < numCPUs; cnt++) {
332 		target = OSDynamicCast(IOCPU, gIOCPUs->getObject(cnt));
333 
334 		// Skip the already-woken boot CPU.
335 		if (target->getCPUNumber() != (UInt32)master_cpu) {
336 			if (target->getCPUState() == kIOCPUStateRunning) {
337 				panic("Spurious wakeup of cpu %u", (unsigned int)(target->getCPUNumber()));
338 			}
339 
340 			if (target->getCPUState() == kIOCPUStateStopped) {
341 				processor_start(target->getMachProcessor());
342 			}
343 		}
344 	}
345 
346 #if defined(__arm64__)
347 	sched_restore_recommended_cores_after_sleep();
348 #endif
349 
350 #endif /* defined(XNU_TARGET_OS_OSX) */
351 
352 	thread_kern_set_pri(self, old_pri);
353 	printf("IOCPUSleepKernel exit\n");
354 }
355 
356 static bool
is_IOCPU_disabled(void)357 is_IOCPU_disabled(void)
358 {
359 	return false;
360 }
361 #else /* !USE_APPLEARMSMP */
362 static bool
is_IOCPU_disabled(void)363 is_IOCPU_disabled(void)
364 {
365 	return true;
366 }
367 #endif /* !USE_APPLEARMSMP */
368 
369 bool
start(IOService * provider)370 IOCPU::start(IOService *provider)
371 {
372 	if (is_IOCPU_disabled()) {
373 		return false;
374 	}
375 
376 	if (!super::start(provider)) {
377 		return false;
378 	}
379 
380 	_cpuGroup = gIOCPUs;
381 	cpuNub = provider;
382 
383 	IOLockLock(gIOCPUsLock);
384 	gIOCPUs->setObject(this);
385 	IOLockUnlock(gIOCPUsLock);
386 
387 	// Correct the bus, cpu and timebase frequencies in the device tree.
388 	if (gPEClockFrequencyInfo.bus_frequency_hz < 0x100000000ULL) {
389 		OSSharedPtr<OSData> busFrequency = OSData::withBytesNoCopy((void *)&gPEClockFrequencyInfo.bus_clock_rate_hz, 4);
390 		provider->setProperty("bus-frequency", busFrequency.get());
391 	} else {
392 		OSSharedPtr<OSData> busFrequency = OSData::withBytesNoCopy((void *)&gPEClockFrequencyInfo.bus_frequency_hz, 8);
393 		provider->setProperty("bus-frequency", busFrequency.get());
394 	}
395 
396 	if (gPEClockFrequencyInfo.cpu_frequency_hz < 0x100000000ULL) {
397 		OSSharedPtr<OSData> cpuFrequency = OSData::withBytesNoCopy((void *)&gPEClockFrequencyInfo.cpu_clock_rate_hz, 4);
398 		provider->setProperty("clock-frequency", cpuFrequency.get());
399 	} else {
400 		OSSharedPtr<OSData> cpuFrequency = OSData::withBytesNoCopy((void *)&gPEClockFrequencyInfo.cpu_frequency_hz, 8);
401 		provider->setProperty("clock-frequency", cpuFrequency.get());
402 	}
403 
404 	OSSharedPtr<OSData> timebaseFrequency = OSData::withBytesNoCopy((void *)&gPEClockFrequencyInfo.timebase_frequency_hz, 4);
405 	provider->setProperty("timebase-frequency", timebaseFrequency.get());
406 
407 	super::setProperty("IOCPUID", getRegistryEntryID(), sizeof(uint64_t) * 8);
408 
409 	setCPUNumber(0);
410 	setCPUState(kIOCPUStateUnregistered);
411 
412 	return true;
413 }
414 
415 void
detach(IOService * provider)416 IOCPU::detach(IOService *provider)
417 {
418 	if (is_IOCPU_disabled()) {
419 		return;
420 	}
421 
422 	super::detach(provider);
423 	IOLockLock(gIOCPUsLock);
424 	unsigned int index = gIOCPUs->getNextIndexOfObject(this, 0);
425 	if (index != (unsigned int)-1) {
426 		gIOCPUs->removeObject(index);
427 	}
428 	IOLockUnlock(gIOCPUsLock);
429 }
430 
431 OSObject *
getProperty(const OSSymbol * aKey) const432 IOCPU::getProperty(const OSSymbol *aKey) const
433 {
434 	if (aKey == gIOCPUStateKey) {
435 		return gIOCPUStateNames[_cpuState].get();
436 	}
437 #pragma clang diagnostic push
438 #pragma clang diagnostic ignored "-Wdeprecated-declarations"
439 	return super::getProperty(aKey);
440 #pragma clang diagnostic pop
441 }
442 
443 bool
setProperty(const OSSymbol * aKey,OSObject * anObject)444 IOCPU::setProperty(const OSSymbol *aKey, OSObject *anObject)
445 {
446 	if (aKey == gIOCPUStateKey) {
447 		return false;
448 	}
449 
450 	return super::setProperty(aKey, anObject);
451 }
452 
453 bool
serializeProperties(OSSerialize * serialize) const454 IOCPU::serializeProperties(OSSerialize *serialize) const
455 {
456 	bool result;
457 	OSSharedPtr<OSDictionary> dict = dictionaryWithProperties();
458 	if (!dict) {
459 		return false;
460 	}
461 	dict->setObject(gIOCPUStateKey.get(), gIOCPUStateNames[_cpuState].get());
462 	result = dict->serialize(serialize);
463 	return result;
464 }
465 
466 IOReturn
setProperties(OSObject * properties)467 IOCPU::setProperties(OSObject *properties)
468 {
469 	OSDictionary *dict = OSDynamicCast(OSDictionary, properties);
470 	OSString     *stateStr;
471 	IOReturn     result;
472 
473 	if (dict == NULL) {
474 		return kIOReturnUnsupported;
475 	}
476 
477 	stateStr = OSDynamicCast(OSString, dict->getObject(gIOCPUStateKey.get()));
478 	if (stateStr != NULL) {
479 		result = IOUserClient::clientHasPrivilege(current_task(), kIOClientPrivilegeAdministrator);
480 		if (result != kIOReturnSuccess) {
481 			return result;
482 		}
483 
484 		if (setProperty(gIOCPUStateKey.get(), stateStr)) {
485 			return kIOReturnSuccess;
486 		}
487 
488 		return kIOReturnUnsupported;
489 	}
490 
491 	return kIOReturnUnsupported;
492 }
493 
494 void
signalCPU(IOCPU *)495 IOCPU::signalCPU(IOCPU */*target*/)
496 {
497 }
498 
499 void
signalCPUDeferred(IOCPU * target)500 IOCPU::signalCPUDeferred(IOCPU *target)
501 {
502 	// Our CPU may not support deferred IPIs,
503 	// so send a regular IPI by default
504 	signalCPU(target);
505 }
506 
507 void
signalCPUCancel(IOCPU *)508 IOCPU::signalCPUCancel(IOCPU */*target*/)
509 {
510 	// Meant to cancel signals sent by
511 	// signalCPUDeferred; unsupported
512 	// by default
513 }
514 
515 void
enableCPUTimeBase(bool)516 IOCPU::enableCPUTimeBase(bool /*enable*/)
517 {
518 }
519 
520 UInt32
getCPUNumber(void)521 IOCPU::getCPUNumber(void)
522 {
523 	return _cpuNumber;
524 }
525 
526 void
setCPUNumber(UInt32 cpuNumber)527 IOCPU::setCPUNumber(UInt32 cpuNumber)
528 {
529 	_cpuNumber = cpuNumber;
530 	super::setProperty("IOCPUNumber", _cpuNumber, 32);
531 }
532 
533 UInt32
getCPUState(void)534 IOCPU::getCPUState(void)
535 {
536 	return _cpuState;
537 }
538 
539 void
setCPUState(UInt32 cpuState)540 IOCPU::setCPUState(UInt32 cpuState)
541 {
542 	if (cpuState < kIOCPUStateCount) {
543 		_cpuState = cpuState;
544 	}
545 }
546 
547 OSArray *
getCPUGroup(void)548 IOCPU::getCPUGroup(void)
549 {
550 	return _cpuGroup.get();
551 }
552 
553 UInt32
getCPUGroupSize(void)554 IOCPU::getCPUGroupSize(void)
555 {
556 	return _cpuGroup->getCount();
557 }
558 
559 processor_t
getMachProcessor(void)560 IOCPU::getMachProcessor(void)
561 {
562 	return machProcessor;
563 }
564 
565 
566 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
567 
568 #undef super
569 #define super IOInterruptController
570 
571 OSDefineMetaClassAndStructors(IOCPUInterruptController, IOInterruptController);
572 
573 OSMetaClassDefineReservedUnused(IOCPUInterruptController, 1);
574 OSMetaClassDefineReservedUnused(IOCPUInterruptController, 2);
575 OSMetaClassDefineReservedUnused(IOCPUInterruptController, 3);
576 OSMetaClassDefineReservedUnused(IOCPUInterruptController, 4);
577 OSMetaClassDefineReservedUnused(IOCPUInterruptController, 5);
578 
579 
580 
581 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
582 
583 IOReturn
initCPUInterruptController(int sources)584 IOCPUInterruptController::initCPUInterruptController(int sources)
585 {
586 	return initCPUInterruptController(sources, sources);
587 }
588 
589 IOReturn
initCPUInterruptController(int sources,int cpus)590 IOCPUInterruptController::initCPUInterruptController(int sources, int cpus)
591 {
592 	int cnt;
593 
594 	if (!super::init()) {
595 		return kIOReturnInvalid;
596 	}
597 
598 	numSources = sources;
599 	numCPUs = cpus;
600 
601 	vectors = (IOInterruptVector *)zalloc_permanent(numSources *
602 	    sizeof(IOInterruptVector), ZALIGN(IOInterruptVector));
603 
604 	// Allocate a lock for each vector
605 	for (cnt = 0; cnt < numSources; cnt++) {
606 		vectors[cnt].interruptLock = IOLockAlloc();
607 		if (vectors[cnt].interruptLock == NULL) {
608 			for (cnt = 0; cnt < numSources; cnt++) {
609 				if (vectors[cnt].interruptLock != NULL) {
610 					IOLockFree(vectors[cnt].interruptLock);
611 				}
612 			}
613 			return kIOReturnNoResources;
614 		}
615 	}
616 
617 	ml_set_max_cpus(numSources);
618 	return kIOReturnSuccess;
619 }
620 
621 void
registerCPUInterruptController(void)622 IOCPUInterruptController::registerCPUInterruptController(void)
623 {
624 	setProperty(gPlatformInterruptControllerName, kOSBooleanTrue);
625 	registerService();
626 
627 	getPlatform()->registerInterruptController(gPlatformInterruptControllerName,
628 	    this);
629 }
630 
631 void
setCPUInterruptProperties(IOService * service)632 IOCPUInterruptController::setCPUInterruptProperties(IOService *service)
633 {
634 	int          cnt;
635 	OSSharedPtr<OSArray> specifier;
636 	OSSharedPtr<OSArray> controller;
637 	long         tmpLong;
638 
639 	if ((service->propertyExists(gIOInterruptControllersKey)) &&
640 	    (service->propertyExists(gIOInterruptSpecifiersKey))) {
641 		return;
642 	}
643 
644 	// Create the interrupt specifer array.
645 	specifier = OSArray::withCapacity(numSources);
646 	for (cnt = 0; cnt < numSources; cnt++) {
647 		tmpLong = cnt;
648 		OSSharedPtr<OSData> tmpData = OSData::withValue(tmpLong);
649 		specifier->setObject(tmpData.get());
650 	}
651 
652 	// Create the interrupt controller array.
653 	controller = OSArray::withCapacity(numSources);
654 	for (cnt = 0; cnt < numSources; cnt++) {
655 		controller->setObject(gPlatformInterruptControllerName);
656 	}
657 
658 	// Put the two arrays into the property table.
659 	service->setProperty(gIOInterruptControllersKey, controller.get());
660 	service->setProperty(gIOInterruptSpecifiersKey, specifier.get());
661 }
662 
663 void
enableCPUInterrupt(IOCPU * cpu)664 IOCPUInterruptController::enableCPUInterrupt(IOCPU *cpu)
665 {
666 	IOInterruptHandler handler = OSMemberFunctionCast(
667 		IOInterruptHandler, this, &IOCPUInterruptController::handleInterrupt);
668 
669 	assert(numCPUs > 0);
670 
671 	ml_install_interrupt_handler(cpu, cpu->getCPUNumber(), this, handler, NULL);
672 
673 	IOTakeLock(vectors[0].interruptLock);
674 	++enabledCPUs;
675 
676 	if (enabledCPUs == numCPUs) {
677 		IOService::cpusRunning();
678 		thread_wakeup(this);
679 	}
680 	IOUnlock(vectors[0].interruptLock);
681 }
682 
683 IOReturn
registerInterrupt(IOService * nub,int source,void * target,IOInterruptHandler handler,void * refCon)684 IOCPUInterruptController::registerInterrupt(IOService *nub,
685     int source,
686     void *target,
687     IOInterruptHandler handler,
688     void *refCon)
689 {
690 	IOInterruptVector *vector;
691 
692 	// Interrupts must be enabled, as this can allocate memory.
693 	assert(ml_get_interrupts_enabled() == TRUE);
694 
695 	if (source >= numSources) {
696 		return kIOReturnNoResources;
697 	}
698 
699 	vector = &vectors[source];
700 
701 	// Get the lock for this vector.
702 	IOTakeLock(vector->interruptLock);
703 
704 	// Make sure the vector is not in use.
705 	if (vector->interruptRegistered) {
706 		IOUnlock(vector->interruptLock);
707 		return kIOReturnNoResources;
708 	}
709 
710 	// Fill in vector with the client's info.
711 	vector->handler = handler;
712 	vector->nub     = nub;
713 	vector->source  = source;
714 	vector->target  = target;
715 	vector->refCon  = refCon;
716 
717 	// Get the vector ready.  It starts hard disabled.
718 	vector->interruptDisabledHard = 1;
719 	vector->interruptDisabledSoft = 1;
720 	vector->interruptRegistered   = 1;
721 
722 	IOUnlock(vector->interruptLock);
723 
724 	IOTakeLock(vectors[0].interruptLock);
725 	if (enabledCPUs != numCPUs) {
726 		assert_wait(this, THREAD_UNINT);
727 		IOUnlock(vectors[0].interruptLock);
728 		thread_block(THREAD_CONTINUE_NULL);
729 	} else {
730 		IOUnlock(vectors[0].interruptLock);
731 	}
732 
733 	return kIOReturnSuccess;
734 }
735 
736 IOReturn
getInterruptType(IOService *,int,int * interruptType)737 IOCPUInterruptController::getInterruptType(IOService */*nub*/,
738     int /*source*/,
739     int *interruptType)
740 {
741 	if (interruptType == NULL) {
742 		return kIOReturnBadArgument;
743 	}
744 
745 	*interruptType = kIOInterruptTypeLevel;
746 
747 	return kIOReturnSuccess;
748 }
749 
750 IOReturn
enableInterrupt(IOService *,int)751 IOCPUInterruptController::enableInterrupt(IOService */*nub*/,
752     int /*source*/)
753 {
754 //  ml_set_interrupts_enabled(true);
755 	return kIOReturnSuccess;
756 }
757 
758 IOReturn
disableInterrupt(IOService *,int)759 IOCPUInterruptController::disableInterrupt(IOService */*nub*/,
760     int /*source*/)
761 {
762 //  ml_set_interrupts_enabled(false);
763 	return kIOReturnSuccess;
764 }
765 
766 IOReturn
causeInterrupt(IOService *,int)767 IOCPUInterruptController::causeInterrupt(IOService */*nub*/,
768     int /*source*/)
769 {
770 	ml_cause_interrupt();
771 	return kIOReturnSuccess;
772 }
773 
774 IOReturn
handleInterrupt(void *,IOService *,int source)775 IOCPUInterruptController::handleInterrupt(void */*refCon*/,
776     IOService */*nub*/,
777     int source)
778 {
779 	IOInterruptVector *vector;
780 
781 	vector = &vectors[source];
782 
783 	if (!vector->interruptRegistered) {
784 		return kIOReturnInvalid;
785 	}
786 
787 	vector->handler(vector->target, vector->refCon,
788 	    vector->nub, vector->source);
789 
790 	return kIOReturnSuccess;
791 }
792 
793 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
794