1 /*
2 * Copyright (c) 2007-2010 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28
29 #include <mach/machine.h>
30 #include <mach/processor.h>
31 #include <kern/kalloc.h>
32 #include <i386/cpu_affinity.h>
33 #include <i386/cpu_topology.h>
34 #include <i386/cpu_threads.h>
35 #include <i386/machine_cpu.h>
36 #include <i386/bit_routines.h>
37 #include <i386/cpu_data.h>
38 #include <i386/lapic.h>
39 #include <i386/machine_routines.h>
40 #include <stddef.h>
41
42 __private_extern__ void qsort(
43 void * array,
44 size_t nmembers,
45 size_t member_size,
46 int (*)(const void *, const void *));
47
48 static int lapicid_cmp(const void *x, const void *y);
49 static x86_affinity_set_t *find_cache_affinity(x86_cpu_cache_t *L2_cachep);
50
51 x86_affinity_set_t *x86_affinities = NULL;
52 static int x86_affinity_count = 0;
53
54 extern cpu_data_t cpshadows[];
55
56 #if DEVELOPMENT || DEBUG
57 void traptrace_init(void);
58 #endif /* DEVELOPMENT || DEBUG */
59
60
61 /* Re-sort double-mapped CPU data shadows after topology discovery sorts the
62 * primary CPU data structures by physical/APIC CPU ID.
63 */
64 static void
cpu_shadow_sort(int ncpus)65 cpu_shadow_sort(int ncpus)
66 {
67 for (int i = 0; i < ncpus; i++) {
68 cpu_data_t *cpup = cpu_datap(i);
69 ptrdiff_t coff = cpup - cpu_datap(0);
70
71 cpup->cd_shadow = &cpshadows[coff];
72 }
73 }
74
75 /*
76 * cpu_topology_sort() is called after all processors have been registered but
77 * before any non-boot processor is started. We establish canonical logical
78 * processor numbering - logical cpus must be contiguous, zero-based and
79 * assigned in physical (local apic id) order. This step is required because
80 * the discovery/registration order is non-deterministic - cores are registered
81 * in differing orders over boots. Enforcing canonical numbering simplifies
82 * identification of processors.
83 */
84 void
cpu_topology_sort(int ncpus)85 cpu_topology_sort(int ncpus)
86 {
87 int i;
88 boolean_t istate;
89 processor_t lprim = NULL;
90
91 assert(machine_info.physical_cpu == 1);
92 assert(machine_info.logical_cpu == 1);
93 assert(master_cpu == 0);
94 assert(cpu_number() == 0);
95 assert(cpu_datap(0)->cpu_number == 0);
96
97 uint32_t cpus_per_pset = 0;
98
99 #if DEVELOPMENT || DEBUG
100 PE_parse_boot_argn("cpus_per_pset", &cpus_per_pset, sizeof(cpus_per_pset));
101 #endif
102
103 /* Lights out for this */
104 istate = ml_set_interrupts_enabled(FALSE);
105
106 if (topo_dbg) {
107 TOPO_DBG("cpu_topology_start() %d cpu%s registered\n",
108 ncpus, (ncpus > 1) ? "s" : "");
109 for (i = 0; i < ncpus; i++) {
110 cpu_data_t *cpup = cpu_datap(i);
111 TOPO_DBG("\tcpu_data[%d]:%p local apic 0x%x\n",
112 i, (void *) cpup, cpup->cpu_phys_number);
113 }
114 }
115
116 /*
117 * Re-order the cpu_data_ptr vector sorting by physical id.
118 * Skip the boot processor, it's required to be correct.
119 */
120 if (ncpus > 1) {
121 qsort((void *) &cpu_data_ptr[1],
122 ncpus - 1,
123 sizeof(cpu_data_t *),
124 lapicid_cmp);
125 }
126 if (topo_dbg) {
127 TOPO_DBG("cpu_topology_start() after sorting:\n");
128 for (i = 0; i < ncpus; i++) {
129 cpu_data_t *cpup = cpu_datap(i);
130 TOPO_DBG("\tcpu_data[%d]:%p local apic 0x%x\n",
131 i, (void *) cpup, cpup->cpu_phys_number);
132 }
133 }
134
135 /*
136 * Finalize logical numbers and map kept by the lapic code.
137 */
138 for (i = 0; i < ncpus; i++) {
139 cpu_data_t *cpup = cpu_datap(i);
140
141 if (cpup->cpu_number != i) {
142 kprintf("cpu_datap(%d):%p local apic id 0x%x "
143 "remapped from %d\n",
144 i, cpup, cpup->cpu_phys_number,
145 cpup->cpu_number);
146 }
147 cpup->cpu_number = i;
148 lapic_cpu_map(cpup->cpu_phys_number, i);
149 x86_set_logical_topology(&cpup->lcpu, cpup->cpu_phys_number, i);
150 }
151
152 cpu_shadow_sort(ncpus);
153 x86_validate_topology();
154
155 ml_set_interrupts_enabled(istate);
156 TOPO_DBG("cpu_topology_start() LLC is L%d\n", topoParms.LLCDepth + 1);
157
158 #if DEVELOPMENT || DEBUG
159 traptrace_init();
160 #endif /* DEVELOPMENT || DEBUG */
161
162 /*
163 * Let the CPU Power Management know that the topology is stable.
164 */
165 topoParms.stable = TRUE;
166 pmCPUStateInit();
167
168 /*
169 * Iterate over all logical cpus finding or creating the affinity set
170 * for their LLC cache. Each affinity set possesses a processor set
171 * into which each logical processor is added.
172 */
173 TOPO_DBG("cpu_topology_start() creating affinity sets:ncpus=%d max_cpus=%d\n", ncpus, machine_info.max_cpus);
174
175 uint32_t pset_cluster_id = 0;
176 for (i = 0; i < machine_info.max_cpus; i++) {
177 cpu_data_t *cpup = cpu_datap(i);
178 x86_lcpu_t *lcpup = cpu_to_lcpu(i);
179 x86_cpu_cache_t *LLC_cachep;
180 x86_affinity_set_t *aset;
181
182 LLC_cachep = lcpup->caches[topoParms.LLCDepth];
183 assert(LLC_cachep->type == CPU_CACHE_TYPE_UNIF);
184 aset = find_cache_affinity(LLC_cachep);
185 if ((aset == NULL) || ((cpus_per_pset != 0) && (i % cpus_per_pset) == 0)) {
186 aset = kalloc_type(x86_affinity_set_t, Z_WAITOK | Z_NOFAIL);
187 aset->next = x86_affinities;
188 x86_affinities = aset;
189 aset->num = x86_affinity_count++;
190 aset->cache = LLC_cachep;
191 if (i == master_cpu) {
192 aset->pset = processor_pset(master_processor);
193 } else {
194 pset_cluster_id++;
195 aset->pset = pset_create(CLUSTER_TYPE_SMP, pset_cluster_id, pset_cluster_id);
196 if (aset->pset == PROCESSOR_SET_NULL) {
197 panic("cpu_topology_start: pset_create");
198 }
199 }
200 TOPO_DBG("\tnew set %p(%d) pset %p for cache %p\n",
201 aset, aset->num, aset->pset, aset->cache);
202 }
203
204 TOPO_DBG("\tprocessor_init set %p(%d) lcpup %p(%d) cpu %p processor %p\n",
205 aset, aset->num, lcpup, lcpup->cpu_num, cpup, cpup->cpu_processor);
206
207 if (i != master_cpu) {
208 processor_init(cpup->cpu_processor, i, aset->pset);
209 }
210
211 if (lcpup->core->num_lcpus > 1) {
212 if (lcpup->lnum == 0) {
213 lprim = cpup->cpu_processor;
214 }
215
216 #if CONFIG_SCHED_SMT
217 processor_set_primary(cpup->cpu_processor, lprim);
218 #endif /* CONFIG_SCHED_SMT */
219 }
220 }
221
222 if (machine_info.max_cpus < machine_info.logical_cpu_max) {
223 /* boot-args cpus=n is set, so adjust max numbers to match */
224 int logical_max = machine_info.max_cpus;
225 int physical_max = logical_max;
226 if (machine_info.logical_cpu_max != machine_info.physical_cpu_max) {
227 physical_max = (logical_max + 1) / 2;
228 }
229 machine_info.logical_cpu_max = logical_max;
230 machine_info.physical_cpu_max = physical_max;
231 }
232 }
233
234 /* We got a request to start a CPU. Check that this CPU is within the
235 * max cpu limit set before we do.
236 */
237 kern_return_t
cpu_topology_start_cpu(int cpunum)238 cpu_topology_start_cpu( int cpunum )
239 {
240 int ncpus = machine_info.max_cpus;
241 int i = cpunum;
242
243 /* Decide whether to start a CPU, and actually start it */
244 TOPO_DBG("cpu_topology_start() processor_start():\n");
245 if (i < ncpus) {
246 TOPO_DBG("\tlcpu %d\n", cpu_datap(i)->cpu_number);
247 processor_boot(cpu_datap(i)->cpu_processor);
248 return KERN_SUCCESS;
249 } else {
250 return KERN_FAILURE;
251 }
252 }
253
254 static int
lapicid_cmp(const void * x,const void * y)255 lapicid_cmp(const void *x, const void *y)
256 {
257 cpu_data_t *cpu_x = *((cpu_data_t **)(uintptr_t)x);
258 cpu_data_t *cpu_y = *((cpu_data_t **)(uintptr_t)y);
259
260 TOPO_DBG("lapicid_cmp(%p,%p) (%d,%d)\n",
261 x, y, cpu_x->cpu_phys_number, cpu_y->cpu_phys_number);
262 if (cpu_x->cpu_phys_number < cpu_y->cpu_phys_number) {
263 return -1;
264 }
265 if (cpu_x->cpu_phys_number == cpu_y->cpu_phys_number) {
266 return 0;
267 }
268 return 1;
269 }
270
271 static x86_affinity_set_t *
find_cache_affinity(x86_cpu_cache_t * l2_cachep)272 find_cache_affinity(x86_cpu_cache_t *l2_cachep)
273 {
274 x86_affinity_set_t *aset;
275
276 for (aset = x86_affinities; aset != NULL; aset = aset->next) {
277 if (l2_cachep == aset->cache) {
278 break;
279 }
280 }
281 return aset;
282 }
283
284 int
ml_get_max_affinity_sets(void)285 ml_get_max_affinity_sets(void)
286 {
287 return x86_affinity_count;
288 }
289
290 processor_set_t
ml_affinity_to_pset(uint32_t affinity_num)291 ml_affinity_to_pset(uint32_t affinity_num)
292 {
293 x86_affinity_set_t *aset;
294
295 for (aset = x86_affinities; aset != NULL; aset = aset->next) {
296 if (affinity_num == aset->num) {
297 break;
298 }
299 }
300 return (aset == NULL) ? PROCESSOR_SET_NULL : aset->pset;
301 }
302
303 uint64_t
ml_cpu_cache_size(unsigned int level)304 ml_cpu_cache_size(unsigned int level)
305 {
306 x86_cpu_cache_t *cachep;
307
308 if (level == 0) {
309 return machine_info.max_mem;
310 } else if (1 <= level && level <= MAX_CACHE_DEPTH) {
311 cachep = current_cpu_datap()->lcpu.caches[level - 1];
312 return cachep ? cachep->cache_size : 0;
313 } else {
314 return 0;
315 }
316 }
317
318 unsigned int
ml_cpu_cache_sharing(unsigned int level,cluster_type_t cluster_type __unused,bool include_all_cpu_types __unused)319 ml_cpu_cache_sharing(unsigned int level, cluster_type_t cluster_type __unused, bool include_all_cpu_types __unused)
320 {
321 x86_cpu_cache_t *cachep;
322
323 if (level == 0) {
324 return machine_info.max_cpus;
325 } else if (1 <= level && level <= MAX_CACHE_DEPTH) {
326 cachep = current_cpu_datap()->lcpu.caches[level - 1];
327 return cachep ? cachep->nlcpus : 0;
328 } else {
329 return 0;
330 }
331 }
332
333 #if DEVELOPMENT || DEBUG
334
335 volatile int traptrace_enabled = 1;
336 uint32_t traptrace_entries_per_cpu = 0;
337 uint32_t PERCPU_DATA(traptrace_next);
338 traptrace_entry_t *PERCPU_DATA(traptrace_ring);
339
340 static void
init_traptrace_bufs(int entries_per_cpu)341 init_traptrace_bufs(int entries_per_cpu)
342 {
343 size_t size = entries_per_cpu * sizeof(traptrace_entry_t);
344
345 percpu_foreach(ring, traptrace_ring) {
346 *ring = zalloc_permanent_tag(size, 63, VM_KERN_MEMORY_DIAG);
347 };
348
349 traptrace_entries_per_cpu = entries_per_cpu;
350 }
351
352 static void
gentrace_configure_from_bootargs(const char * ena_prop,int * ena_valp,const char * epc_prop,int * epcp,int max_epc,int def_epc,int override)353 gentrace_configure_from_bootargs(const char *ena_prop, int *ena_valp, const char *epc_prop,
354 int *epcp, int max_epc, int def_epc, int override)
355 {
356 if (kern_feature_override(override)) {
357 *ena_valp = 0;
358 }
359
360 (void) PE_parse_boot_argn(ena_prop, ena_valp, sizeof(*ena_valp));
361
362 if (*ena_valp == 0) {
363 return;
364 }
365
366 if (PE_parse_boot_argn(epc_prop, epcp, sizeof(*epcp)) &&
367 (*epcp < 1 || *epcp > max_epc)) {
368 *epcp = def_epc;
369 }
370 }
371
372 void
traptrace_init(void)373 traptrace_init(void)
374 {
375 int entries_per_cpu = DEFAULT_TRAPTRACE_ENTRIES_PER_CPU;
376 int enable = traptrace_enabled;
377
378 gentrace_configure_from_bootargs("traptrace", &enable, "traptrace_epc", &entries_per_cpu,
379 TRAPTRACE_MAX_ENTRIES_PER_CPU, DEFAULT_TRAPTRACE_ENTRIES_PER_CPU, KF_TRAPTRACE_OVRD);
380
381 traptrace_enabled = enable;
382
383 if (traptrace_enabled) {
384 init_traptrace_bufs(entries_per_cpu);
385 }
386 }
387
388 #endif /* DEVELOPMENT || DEBUG */
389