1*e3723e1fSApple OSS Distributions /*
2*e3723e1fSApple OSS Distributions * Copyright (c) 2019 Apple Computer, Inc. All rights reserved.
3*e3723e1fSApple OSS Distributions *
4*e3723e1fSApple OSS Distributions * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5*e3723e1fSApple OSS Distributions *
6*e3723e1fSApple OSS Distributions * This file contains Original Code and/or Modifications of Original Code
7*e3723e1fSApple OSS Distributions * as defined in and that are subject to the Apple Public Source License
8*e3723e1fSApple OSS Distributions * Version 2.0 (the 'License'). You may not use this file except in
9*e3723e1fSApple OSS Distributions * compliance with the License. The rights granted to you under the License
10*e3723e1fSApple OSS Distributions * may not be used to create, or enable the creation or redistribution of,
11*e3723e1fSApple OSS Distributions * unlawful or unlicensed copies of an Apple operating system, or to
12*e3723e1fSApple OSS Distributions * circumvent, violate, or enable the circumvention or violation of, any
13*e3723e1fSApple OSS Distributions * terms of an Apple operating system software license agreement.
14*e3723e1fSApple OSS Distributions *
15*e3723e1fSApple OSS Distributions * Please obtain a copy of the License at
16*e3723e1fSApple OSS Distributions * http://www.opensource.apple.com/apsl/ and read it before using this file.
17*e3723e1fSApple OSS Distributions *
18*e3723e1fSApple OSS Distributions * The Original Code and all software distributed under the License are
19*e3723e1fSApple OSS Distributions * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20*e3723e1fSApple OSS Distributions * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21*e3723e1fSApple OSS Distributions * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22*e3723e1fSApple OSS Distributions * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23*e3723e1fSApple OSS Distributions * Please see the License for the specific language governing rights and
24*e3723e1fSApple OSS Distributions * limitations under the License.
25*e3723e1fSApple OSS Distributions *
26*e3723e1fSApple OSS Distributions * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27*e3723e1fSApple OSS Distributions */
28*e3723e1fSApple OSS Distributions /**
29*e3723e1fSApple OSS Distributions * On devices that support it, this test ensures that a mach exception is
30*e3723e1fSApple OSS Distributions * generated when a matrix-math exception is triggered, and that the
31*e3723e1fSApple OSS Distributions * matrix register file is correctly preserved or zeroed on context switch.
32*e3723e1fSApple OSS Distributions */
33*e3723e1fSApple OSS Distributions
34*e3723e1fSApple OSS Distributions /*
35*e3723e1fSApple OSS Distributions * IMPLEMENTATION NOTE:
36*e3723e1fSApple OSS Distributions *
37*e3723e1fSApple OSS Distributions * This test code goes to some unusual lengths to avoid calling out to libc or
38*e3723e1fSApple OSS Distributions * libdarwintest while the CPU is in streaming SVE mode (i.e., between
39*e3723e1fSApple OSS Distributions * ops->start() and ops->stop()). Both of these libraries are built with SIMD
40*e3723e1fSApple OSS Distributions * instructions that will cause the test executable to crash while in streaming
41*e3723e1fSApple OSS Distributions * SVE mode.
42*e3723e1fSApple OSS Distributions *
43*e3723e1fSApple OSS Distributions * Ordinarily this is the wrong way to solve this problem. Functions that use
44*e3723e1fSApple OSS Distributions * streaming SVE mode should have annotations telling the compiler so, and the
45*e3723e1fSApple OSS Distributions * compiler will automatically generate appropriate interworking code. However
46*e3723e1fSApple OSS Distributions * this interworking code will stash SME state to memory and temporarily exit
47*e3723e1fSApple OSS Distributions * streaming SVE mode. We're specifically testing how xnu manages live SME
48*e3723e1fSApple OSS Distributions * register state, so we can't let the compiler stash and disable this state
49*e3723e1fSApple OSS Distributions * behind our backs.
50*e3723e1fSApple OSS Distributions */
51*e3723e1fSApple OSS Distributions
52*e3723e1fSApple OSS Distributions #ifdef __arm64__
53*e3723e1fSApple OSS Distributions #include <mach/error.h>
54*e3723e1fSApple OSS Distributions #endif /* __arm64__ */
55*e3723e1fSApple OSS Distributions
56*e3723e1fSApple OSS Distributions #include <darwintest.h>
57*e3723e1fSApple OSS Distributions #include <pthread.h>
58*e3723e1fSApple OSS Distributions #include <stdlib.h>
59*e3723e1fSApple OSS Distributions #include <mach/mach.h>
60*e3723e1fSApple OSS Distributions #include <mach/thread_act.h>
61*e3723e1fSApple OSS Distributions #include <mach/thread_status.h>
62*e3723e1fSApple OSS Distributions #include <mach/exception.h>
63*e3723e1fSApple OSS Distributions #include <machine/cpu_capabilities.h>
64*e3723e1fSApple OSS Distributions #include <sys/types.h>
65*e3723e1fSApple OSS Distributions #include <sys/sysctl.h>
66*e3723e1fSApple OSS Distributions
67*e3723e1fSApple OSS Distributions #include "arm_matrix.h"
68*e3723e1fSApple OSS Distributions #include "exc_helpers.h"
69*e3723e1fSApple OSS Distributions #include "test_utils.h"
70*e3723e1fSApple OSS Distributions
71*e3723e1fSApple OSS Distributions T_GLOBAL_META(
72*e3723e1fSApple OSS Distributions T_META_NAMESPACE("xnu.arm"),
73*e3723e1fSApple OSS Distributions T_META_RADAR_COMPONENT_NAME("xnu"),
74*e3723e1fSApple OSS Distributions T_META_RADAR_COMPONENT_VERSION("arm"),
75*e3723e1fSApple OSS Distributions T_META_OWNER("ghackmann"),
76*e3723e1fSApple OSS Distributions T_META_RUN_CONCURRENTLY(true)
77*e3723e1fSApple OSS Distributions );
78*e3723e1fSApple OSS Distributions
79*e3723e1fSApple OSS Distributions #ifdef __arm64__
80*e3723e1fSApple OSS Distributions
81*e3723e1fSApple OSS Distributions #ifndef EXC_ARM_SME_DISALLOWED
82*e3723e1fSApple OSS Distributions #define EXC_ARM_SME_DISALLOWED 2
83*e3723e1fSApple OSS Distributions #endif
84*e3723e1fSApple OSS Distributions
85*e3723e1fSApple OSS Distributions /* Whether we caught the EXC_BAD_INSTRUCTION mach exception or not. */
86*e3723e1fSApple OSS Distributions static volatile bool mach_exc_caught = false;
87*e3723e1fSApple OSS Distributions
88*e3723e1fSApple OSS Distributions static size_t
bad_instruction_exception_handler(__unused mach_port_t task,__unused mach_port_t thread,exception_type_t type,mach_exception_data_t codes)89*e3723e1fSApple OSS Distributions bad_instruction_exception_handler(
90*e3723e1fSApple OSS Distributions __unused mach_port_t task,
91*e3723e1fSApple OSS Distributions __unused mach_port_t thread,
92*e3723e1fSApple OSS Distributions exception_type_t type,
93*e3723e1fSApple OSS Distributions mach_exception_data_t codes)
94*e3723e1fSApple OSS Distributions {
95*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_EQ(type, EXC_BAD_INSTRUCTION, "Caught an EXC_BAD_INSTRUCTION exception");
96*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_EQ(codes[0], (uint64_t)EXC_ARM_UNDEFINED, "The subcode is EXC_ARM_UNDEFINED");
97*e3723e1fSApple OSS Distributions
98*e3723e1fSApple OSS Distributions mach_exc_caught = true;
99*e3723e1fSApple OSS Distributions return 4;
100*e3723e1fSApple OSS Distributions }
101*e3723e1fSApple OSS Distributions #endif
102*e3723e1fSApple OSS Distributions
103*e3723e1fSApple OSS Distributions
104*e3723e1fSApple OSS Distributions #ifdef __arm64__
105*e3723e1fSApple OSS Distributions static void
test_matrix_not_started(const struct arm_matrix_operations * ops)106*e3723e1fSApple OSS Distributions test_matrix_not_started(const struct arm_matrix_operations *ops)
107*e3723e1fSApple OSS Distributions {
108*e3723e1fSApple OSS Distributions if (!ops->is_available()) {
109*e3723e1fSApple OSS Distributions T_SKIP("Running on non-%s target, skipping...", ops->name);
110*e3723e1fSApple OSS Distributions }
111*e3723e1fSApple OSS Distributions
112*e3723e1fSApple OSS Distributions mach_port_t exc_port = create_exception_port(EXC_MASK_BAD_INSTRUCTION);
113*e3723e1fSApple OSS Distributions
114*e3723e1fSApple OSS Distributions size_t size = ops->data_size();
115*e3723e1fSApple OSS Distributions uint8_t *d = ops->alloc_data();
116*e3723e1fSApple OSS Distributions bzero(d, size);
117*e3723e1fSApple OSS Distributions
118*e3723e1fSApple OSS Distributions ops->start();
119*e3723e1fSApple OSS Distributions ops->load_one_vector(d);
120*e3723e1fSApple OSS Distributions ops->stop();
121*e3723e1fSApple OSS Distributions T_PASS("%s instruction after start instruction should not cause an exception", ops->name);
122*e3723e1fSApple OSS Distributions
123*e3723e1fSApple OSS Distributions mach_exc_caught = false;
124*e3723e1fSApple OSS Distributions run_exception_handler(exc_port, bad_instruction_exception_handler);
125*e3723e1fSApple OSS Distributions ops->load_one_vector(d);
126*e3723e1fSApple OSS Distributions T_EXPECT_TRUE(mach_exc_caught, "%s instruction before start instruction should cause an exception", ops->name);
127*e3723e1fSApple OSS Distributions
128*e3723e1fSApple OSS Distributions free(d);
129*e3723e1fSApple OSS Distributions }
130*e3723e1fSApple OSS Distributions #endif
131*e3723e1fSApple OSS Distributions
132*e3723e1fSApple OSS Distributions
133*e3723e1fSApple OSS Distributions T_DECL(sme_not_started,
134*e3723e1fSApple OSS Distributions "Test that SME instructions before smstart generate mach exceptions.", T_META_TAG_VM_NOT_ELIGIBLE)
135*e3723e1fSApple OSS Distributions {
136*e3723e1fSApple OSS Distributions #ifndef __arm64__
137*e3723e1fSApple OSS Distributions T_SKIP("Running on non-arm64 target, skipping...");
138*e3723e1fSApple OSS Distributions #else
139*e3723e1fSApple OSS Distributions test_matrix_not_started(&sme_operations);
140*e3723e1fSApple OSS Distributions #endif
141*e3723e1fSApple OSS Distributions }
142*e3723e1fSApple OSS Distributions
143*e3723e1fSApple OSS Distributions #ifdef __arm64__
144*e3723e1fSApple OSS Distributions struct test_thread;
145*e3723e1fSApple OSS Distributions typedef bool (*thread_fn_t)(struct test_thread const* thread);
146*e3723e1fSApple OSS Distributions
147*e3723e1fSApple OSS Distributions struct test_thread {
148*e3723e1fSApple OSS Distributions pthread_t thread;
149*e3723e1fSApple OSS Distributions pthread_t companion_thread;
150*e3723e1fSApple OSS Distributions thread_fn_t thread_fn;
151*e3723e1fSApple OSS Distributions uint32_t cpuid;
152*e3723e1fSApple OSS Distributions uint32_t thread_id;
153*e3723e1fSApple OSS Distributions const struct arm_matrix_operations *ops;
154*e3723e1fSApple OSS Distributions };
155*e3723e1fSApple OSS Distributions
156*e3723e1fSApple OSS Distributions static uint32_t barrier;
157*e3723e1fSApple OSS Distributions static pthread_cond_t barrier_cond = PTHREAD_COND_INITIALIZER;
158*e3723e1fSApple OSS Distributions static pthread_mutex_t barrier_lock = PTHREAD_MUTEX_INITIALIZER;
159*e3723e1fSApple OSS Distributions
160*e3723e1fSApple OSS Distributions static uint32_t end_barrier;
161*e3723e1fSApple OSS Distributions static pthread_cond_t end_barrier_cond = PTHREAD_COND_INITIALIZER;
162*e3723e1fSApple OSS Distributions static pthread_mutex_t end_barrier_lock = PTHREAD_MUTEX_INITIALIZER;
163*e3723e1fSApple OSS Distributions
164*e3723e1fSApple OSS Distributions static void
test_thread_barrier(void)165*e3723e1fSApple OSS Distributions test_thread_barrier(void)
166*e3723e1fSApple OSS Distributions {
167*e3723e1fSApple OSS Distributions /* Wait for all threads to reach this barrier */
168*e3723e1fSApple OSS Distributions pthread_mutex_lock(&barrier_lock);
169*e3723e1fSApple OSS Distributions barrier--;
170*e3723e1fSApple OSS Distributions if (barrier) {
171*e3723e1fSApple OSS Distributions while (barrier) {
172*e3723e1fSApple OSS Distributions pthread_cond_wait(&barrier_cond, &barrier_lock);
173*e3723e1fSApple OSS Distributions }
174*e3723e1fSApple OSS Distributions } else {
175*e3723e1fSApple OSS Distributions pthread_cond_broadcast(&barrier_cond);
176*e3723e1fSApple OSS Distributions }
177*e3723e1fSApple OSS Distributions pthread_mutex_unlock(&barrier_lock);
178*e3723e1fSApple OSS Distributions }
179*e3723e1fSApple OSS Distributions
180*e3723e1fSApple OSS Distributions static void
test_thread_notify_exited(void)181*e3723e1fSApple OSS Distributions test_thread_notify_exited(void)
182*e3723e1fSApple OSS Distributions {
183*e3723e1fSApple OSS Distributions pthread_mutex_lock(&end_barrier_lock);
184*e3723e1fSApple OSS Distributions if (0 == --end_barrier) {
185*e3723e1fSApple OSS Distributions pthread_cond_signal(&end_barrier_cond);
186*e3723e1fSApple OSS Distributions }
187*e3723e1fSApple OSS Distributions pthread_mutex_unlock(&end_barrier_lock);
188*e3723e1fSApple OSS Distributions }
189*e3723e1fSApple OSS Distributions
190*e3723e1fSApple OSS Distributions static void
wait_for_test_threads(void)191*e3723e1fSApple OSS Distributions wait_for_test_threads(void)
192*e3723e1fSApple OSS Distributions {
193*e3723e1fSApple OSS Distributions pthread_mutex_lock(&end_barrier_lock);
194*e3723e1fSApple OSS Distributions while (end_barrier) {
195*e3723e1fSApple OSS Distributions pthread_cond_wait(&end_barrier_cond, &end_barrier_lock);
196*e3723e1fSApple OSS Distributions }
197*e3723e1fSApple OSS Distributions pthread_mutex_unlock(&end_barrier_lock);
198*e3723e1fSApple OSS Distributions }
199*e3723e1fSApple OSS Distributions
200*e3723e1fSApple OSS Distributions static uint32_t
ncpus(void)201*e3723e1fSApple OSS Distributions ncpus(void)
202*e3723e1fSApple OSS Distributions {
203*e3723e1fSApple OSS Distributions uint32_t ncpu;
204*e3723e1fSApple OSS Distributions size_t ncpu_size = sizeof(ncpu);
205*e3723e1fSApple OSS Distributions int err = sysctlbyname("hw.ncpu", &ncpu, &ncpu_size, NULL, 0);
206*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_POSIX_ZERO(err, "Retrieved CPU count");
207*e3723e1fSApple OSS Distributions
208*e3723e1fSApple OSS Distributions return ncpu;
209*e3723e1fSApple OSS Distributions }
210*e3723e1fSApple OSS Distributions
211*e3723e1fSApple OSS Distributions static int
thread_bind_cpu_unchecked(uint32_t cpuid)212*e3723e1fSApple OSS Distributions thread_bind_cpu_unchecked(uint32_t cpuid)
213*e3723e1fSApple OSS Distributions {
214*e3723e1fSApple OSS Distributions /*
215*e3723e1fSApple OSS Distributions * libc's sysctl() implementation calls strlen(name), which is
216*e3723e1fSApple OSS Distributions * SIMD-accelerated. Avoid this by directly invoking the libsyscall
217*e3723e1fSApple OSS Distributions * wrapper with namelen computed at compile time.
218*e3723e1fSApple OSS Distributions */
219*e3723e1fSApple OSS Distributions #define THREAD_BIND_CPU "kern.sched_thread_bind_cpu"
220*e3723e1fSApple OSS Distributions extern int __sysctlbyname(const char *name, size_t namelen, void *oldp, size_t *oldlenp, void *newp, size_t newlen);
221*e3723e1fSApple OSS Distributions const char *name = THREAD_BIND_CPU;
222*e3723e1fSApple OSS Distributions size_t namelen = sizeof(THREAD_BIND_CPU) - 1;
223*e3723e1fSApple OSS Distributions return __sysctlbyname(name, namelen, NULL, 0, &cpuid, sizeof(cpuid));
224*e3723e1fSApple OSS Distributions }
225*e3723e1fSApple OSS Distributions
226*e3723e1fSApple OSS Distributions static void
thread_bind_cpu(uint32_t cpuid)227*e3723e1fSApple OSS Distributions thread_bind_cpu(uint32_t cpuid)
228*e3723e1fSApple OSS Distributions {
229*e3723e1fSApple OSS Distributions int err = thread_bind_cpu_unchecked(cpuid);
230*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_POSIX_ZERO(err, "Bound thread to CPU %u", cpuid);
231*e3723e1fSApple OSS Distributions }
232*e3723e1fSApple OSS Distributions
233*e3723e1fSApple OSS Distributions static void *
test_thread_shim(void * arg)234*e3723e1fSApple OSS Distributions test_thread_shim(void *arg)
235*e3723e1fSApple OSS Distributions {
236*e3723e1fSApple OSS Distributions struct test_thread const *thread = arg;
237*e3723e1fSApple OSS Distributions
238*e3723e1fSApple OSS Distributions thread_bind_cpu(thread->cpuid);
239*e3723e1fSApple OSS Distributions bool const ret = thread->thread_fn(thread);
240*e3723e1fSApple OSS Distributions test_thread_notify_exited();
241*e3723e1fSApple OSS Distributions return (void *)(uintptr_t)ret;
242*e3723e1fSApple OSS Distributions }
243*e3723e1fSApple OSS Distributions
244*e3723e1fSApple OSS Distributions static void
test_on_each_cpu(thread_fn_t thread_fn,const struct arm_matrix_operations * ops,const char * desc)245*e3723e1fSApple OSS Distributions test_on_each_cpu(thread_fn_t thread_fn, const struct arm_matrix_operations *ops, const char *desc)
246*e3723e1fSApple OSS Distributions {
247*e3723e1fSApple OSS Distributions uint32_t ncpu = ncpus();
248*e3723e1fSApple OSS Distributions uint32_t nthreads = ncpu * 2;
249*e3723e1fSApple OSS Distributions barrier = 1 /* This thread */ + nthreads;
250*e3723e1fSApple OSS Distributions end_barrier = nthreads;
251*e3723e1fSApple OSS Distributions struct test_thread *threads = calloc(nthreads, sizeof(threads[0]));
252*e3723e1fSApple OSS Distributions
253*e3723e1fSApple OSS Distributions for (uint32_t i = 0; i < nthreads; i++) {
254*e3723e1fSApple OSS Distributions threads[i].thread_fn = thread_fn;
255*e3723e1fSApple OSS Distributions threads[i].cpuid = i % ncpu;
256*e3723e1fSApple OSS Distributions threads[i].thread_id = i;
257*e3723e1fSApple OSS Distributions threads[i].ops = ops;
258*e3723e1fSApple OSS Distributions
259*e3723e1fSApple OSS Distributions int const err = pthread_create(&threads[i].thread, NULL, test_thread_shim, &threads[i]);
260*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_EQ(err, 0, "%s: created thread #%u", desc, i);
261*e3723e1fSApple OSS Distributions
262*e3723e1fSApple OSS Distributions // The other of two threads under test pinned to the same CPU.
263*e3723e1fSApple OSS Distributions threads[(ncpu + i) % nthreads].companion_thread = threads[i].thread;
264*e3723e1fSApple OSS Distributions }
265*e3723e1fSApple OSS Distributions
266*e3723e1fSApple OSS Distributions // Wait for all companion_threads to be set.
267*e3723e1fSApple OSS Distributions test_thread_barrier();
268*e3723e1fSApple OSS Distributions
269*e3723e1fSApple OSS Distributions // like pthread_join()ing all threads, but without the priority boosting shenanigans.
270*e3723e1fSApple OSS Distributions wait_for_test_threads();
271*e3723e1fSApple OSS Distributions
272*e3723e1fSApple OSS Distributions for (uint32_t i = 0; i < nthreads; i++) {
273*e3723e1fSApple OSS Distributions void *thread_ret_ptr;
274*e3723e1fSApple OSS Distributions int err = pthread_join(threads[i].thread, &thread_ret_ptr);
275*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_EQ(err, 0, "%s: joined thread #%u", desc, i);
276*e3723e1fSApple OSS Distributions
277*e3723e1fSApple OSS Distributions bool thread_ret = (uintptr_t)thread_ret_ptr;
278*e3723e1fSApple OSS Distributions if (thread_ret) {
279*e3723e1fSApple OSS Distributions T_PASS("%s: thread #%u passed", desc, i);
280*e3723e1fSApple OSS Distributions } else {
281*e3723e1fSApple OSS Distributions T_FAIL("%s: thread #%u failed", desc, i);
282*e3723e1fSApple OSS Distributions }
283*e3723e1fSApple OSS Distributions }
284*e3723e1fSApple OSS Distributions
285*e3723e1fSApple OSS Distributions free(threads);
286*e3723e1fSApple OSS Distributions }
287*e3723e1fSApple OSS Distributions
288*e3723e1fSApple OSS Distributions static bool
active_context_switch_thread(struct test_thread const * thread)289*e3723e1fSApple OSS Distributions active_context_switch_thread(struct test_thread const* thread)
290*e3723e1fSApple OSS Distributions {
291*e3723e1fSApple OSS Distributions const struct arm_matrix_operations *ops = thread->ops;
292*e3723e1fSApple OSS Distributions const uint32_t thread_id = thread->thread_id;
293*e3723e1fSApple OSS Distributions size_t size = ops->data_size();
294*e3723e1fSApple OSS Distributions uint8_t *d1 = ops->alloc_data();
295*e3723e1fSApple OSS Distributions memset(d1, (char)thread_id, size);
296*e3723e1fSApple OSS Distributions
297*e3723e1fSApple OSS Distributions uint8_t *d2 = ops->alloc_data();
298*e3723e1fSApple OSS Distributions
299*e3723e1fSApple OSS Distributions test_thread_barrier();
300*e3723e1fSApple OSS Distributions
301*e3723e1fSApple OSS Distributions // companion_thread will be valid only after the barrier.
302*e3723e1fSApple OSS Distributions thread_t const companion_thread = pthread_mach_thread_np(thread->companion_thread);
303*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_NE(companion_thread, THREAD_NULL, "pthread_mach_thread_np");
304*e3723e1fSApple OSS Distributions
305*e3723e1fSApple OSS Distributions bool ok = true;
306*e3723e1fSApple OSS Distributions for (unsigned int i = 0; i < 100000 && ok; i++) {
307*e3723e1fSApple OSS Distributions ops->start();
308*e3723e1fSApple OSS Distributions ops->load_data(d1);
309*e3723e1fSApple OSS Distributions
310*e3723e1fSApple OSS Distributions /*
311*e3723e1fSApple OSS Distributions * Rescheduling with the matrix registers active must preserve
312*e3723e1fSApple OSS Distributions * state, even after a context switch.
313*e3723e1fSApple OSS Distributions */
314*e3723e1fSApple OSS Distributions thread_switch(companion_thread, SWITCH_OPTION_NONE, 0);
315*e3723e1fSApple OSS Distributions
316*e3723e1fSApple OSS Distributions ops->store_data(d2);
317*e3723e1fSApple OSS Distributions ops->stop();
318*e3723e1fSApple OSS Distributions
319*e3723e1fSApple OSS Distributions if (memcmp(d1, d2, size)) {
320*e3723e1fSApple OSS Distributions ok = false;
321*e3723e1fSApple OSS Distributions }
322*e3723e1fSApple OSS Distributions }
323*e3723e1fSApple OSS Distributions
324*e3723e1fSApple OSS Distributions free(d2);
325*e3723e1fSApple OSS Distributions free(d1);
326*e3723e1fSApple OSS Distributions return ok;
327*e3723e1fSApple OSS Distributions }
328*e3723e1fSApple OSS Distributions
329*e3723e1fSApple OSS Distributions static bool
inactive_context_switch_thread(struct test_thread const * thread)330*e3723e1fSApple OSS Distributions inactive_context_switch_thread(struct test_thread const* thread)
331*e3723e1fSApple OSS Distributions {
332*e3723e1fSApple OSS Distributions const struct arm_matrix_operations *ops = thread->ops;
333*e3723e1fSApple OSS Distributions const uint32_t thread_id = thread->thread_id;
334*e3723e1fSApple OSS Distributions size_t size = ops->data_size();
335*e3723e1fSApple OSS Distributions uint8_t *d1 = ops->alloc_data();
336*e3723e1fSApple OSS Distributions memset(d1, (char)thread_id, size);
337*e3723e1fSApple OSS Distributions
338*e3723e1fSApple OSS Distributions uint8_t *d2 = ops->alloc_data();
339*e3723e1fSApple OSS Distributions
340*e3723e1fSApple OSS Distributions test_thread_barrier();
341*e3723e1fSApple OSS Distributions
342*e3723e1fSApple OSS Distributions // companion_thread will be valid only after the barrier.
343*e3723e1fSApple OSS Distributions thread_t const companion_thread = pthread_mach_thread_np(thread->companion_thread);
344*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_NE(companion_thread, THREAD_NULL, "pthread_mach_thread_np");
345*e3723e1fSApple OSS Distributions
346*e3723e1fSApple OSS Distributions bool ok = true;
347*e3723e1fSApple OSS Distributions for (unsigned int i = 0; i < 100000 && ok; i++) {
348*e3723e1fSApple OSS Distributions ops->start();
349*e3723e1fSApple OSS Distributions ops->load_data(d1);
350*e3723e1fSApple OSS Distributions ops->stop();
351*e3723e1fSApple OSS Distributions
352*e3723e1fSApple OSS Distributions /*
353*e3723e1fSApple OSS Distributions * Rescheduling with the matrix registers inactive may preserve
354*e3723e1fSApple OSS Distributions * state or may zero it out.
355*e3723e1fSApple OSS Distributions */
356*e3723e1fSApple OSS Distributions thread_switch(companion_thread, SWITCH_OPTION_NONE, 0);
357*e3723e1fSApple OSS Distributions
358*e3723e1fSApple OSS Distributions ops->start();
359*e3723e1fSApple OSS Distributions ops->store_data(d2);
360*e3723e1fSApple OSS Distributions ops->stop();
361*e3723e1fSApple OSS Distributions
362*e3723e1fSApple OSS Distributions for (size_t j = 0; j < size; j++) {
363*e3723e1fSApple OSS Distributions if (d1[j] != d2[j] && d2[j] != 0) {
364*e3723e1fSApple OSS Distributions ok = false;
365*e3723e1fSApple OSS Distributions }
366*e3723e1fSApple OSS Distributions }
367*e3723e1fSApple OSS Distributions }
368*e3723e1fSApple OSS Distributions
369*e3723e1fSApple OSS Distributions free(d2);
370*e3723e1fSApple OSS Distributions free(d1);
371*e3723e1fSApple OSS Distributions return ok;
372*e3723e1fSApple OSS Distributions }
373*e3723e1fSApple OSS Distributions
374*e3723e1fSApple OSS Distributions static void
test_thread_migration(const struct arm_matrix_operations * ops)375*e3723e1fSApple OSS Distributions test_thread_migration(const struct arm_matrix_operations *ops)
376*e3723e1fSApple OSS Distributions {
377*e3723e1fSApple OSS Distributions size_t size = ops->data_size();
378*e3723e1fSApple OSS Distributions uint8_t *d = ops->alloc_data();
379*e3723e1fSApple OSS Distributions arc4random_buf(d, size);
380*e3723e1fSApple OSS Distributions
381*e3723e1fSApple OSS Distributions uint32_t ncpu = ncpus();
382*e3723e1fSApple OSS Distributions uint8_t *cpu_d[ncpu];
383*e3723e1fSApple OSS Distributions for (uint32_t cpuid = 0; cpuid < ncpu; cpuid++) {
384*e3723e1fSApple OSS Distributions cpu_d[cpuid] = ops->alloc_data();
385*e3723e1fSApple OSS Distributions memset(cpu_d[cpuid], 0, size);
386*e3723e1fSApple OSS Distributions }
387*e3723e1fSApple OSS Distributions
388*e3723e1fSApple OSS Distributions ops->start();
389*e3723e1fSApple OSS Distributions ops->load_data(d);
390*e3723e1fSApple OSS Distributions for (uint32_t cpuid = 0; cpuid < ncpu; cpuid++) {
391*e3723e1fSApple OSS Distributions int err = thread_bind_cpu_unchecked(cpuid);
392*e3723e1fSApple OSS Distributions if (err) {
393*e3723e1fSApple OSS Distributions ops->stop();
394*e3723e1fSApple OSS Distributions T_ASSERT_POSIX_ZERO(err, "Bound thread to CPU %u", cpuid);
395*e3723e1fSApple OSS Distributions }
396*e3723e1fSApple OSS Distributions ops->store_data(cpu_d[cpuid]);
397*e3723e1fSApple OSS Distributions }
398*e3723e1fSApple OSS Distributions ops->stop();
399*e3723e1fSApple OSS Distributions
400*e3723e1fSApple OSS Distributions for (uint32_t cpuid = 0; cpuid < ncpu; cpuid++) {
401*e3723e1fSApple OSS Distributions int cmp = memcmp(d, cpu_d[cpuid], size);
402*e3723e1fSApple OSS Distributions T_EXPECT_EQ(cmp, 0, "Matrix state migrated to CPU %u", cpuid);
403*e3723e1fSApple OSS Distributions free(cpu_d[cpuid]);
404*e3723e1fSApple OSS Distributions }
405*e3723e1fSApple OSS Distributions free(d);
406*e3723e1fSApple OSS Distributions }
407*e3723e1fSApple OSS Distributions #endif
408*e3723e1fSApple OSS Distributions
409*e3723e1fSApple OSS Distributions
410*e3723e1fSApple OSS Distributions T_DECL(sme_context_switch,
411*e3723e1fSApple OSS Distributions "Test that SME contexts are migrated during context switch and do not leak between process contexts.",
412*e3723e1fSApple OSS Distributions T_META_BOOTARGS_SET("enable_skstb=1"),
413*e3723e1fSApple OSS Distributions T_META_REQUIRES_SYSCTL_EQ("hw.optional.arm.FEAT_SME2", 1),
414*e3723e1fSApple OSS Distributions XNU_T_META_SOC_SPECIFIC, T_META_TAG_VM_NOT_ELIGIBLE)
415*e3723e1fSApple OSS Distributions {
416*e3723e1fSApple OSS Distributions #ifndef __arm64__
417*e3723e1fSApple OSS Distributions T_SKIP("Running on non-arm64 target, skipping...");
418*e3723e1fSApple OSS Distributions #else
419*e3723e1fSApple OSS Distributions if (!sme_operations.is_available()) {
420*e3723e1fSApple OSS Distributions T_SKIP("Running on non-SME target, skipping...");
421*e3723e1fSApple OSS Distributions }
422*e3723e1fSApple OSS Distributions
423*e3723e1fSApple OSS Distributions test_thread_migration(&sme_operations);
424*e3723e1fSApple OSS Distributions test_on_each_cpu(active_context_switch_thread, &sme_operations, "SME context migrates when active");
425*e3723e1fSApple OSS Distributions test_on_each_cpu(inactive_context_switch_thread, &sme_operations, "SME context does not leak across processes");
426*e3723e1fSApple OSS Distributions #endif
427*e3723e1fSApple OSS Distributions }
428*e3723e1fSApple OSS Distributions
429*e3723e1fSApple OSS Distributions
430*e3723e1fSApple OSS Distributions #if __arm64__
431*e3723e1fSApple OSS Distributions /*
432*e3723e1fSApple OSS Distributions * Sequence of events in thread_{get,set}_state test:
433*e3723e1fSApple OSS Distributions *
434*e3723e1fSApple OSS Distributions * 1. Parent creates child thread.
435*e3723e1fSApple OSS Distributions * 2. Child thread signals parent thread to proceed.
436*e3723e1fSApple OSS Distributions * 3. Parent populates child's matrix state registers via thread_set_state(),
437*e3723e1fSApple OSS Distributions * and signals child thread to proceed.
438*e3723e1fSApple OSS Distributions * 4. Child arbitrarily updates each byte in its local matrix register state
439*e3723e1fSApple OSS Distributions * by adding 1, and signals parent thread to proceed.
440*e3723e1fSApple OSS Distributions * 5. Parent reads back the child's updated matrix state with
441*e3723e1fSApple OSS Distributions * thread_get_state(), and confirms that every byte has been modified as
442*e3723e1fSApple OSS Distributions * expected.
443*e3723e1fSApple OSS Distributions */
444*e3723e1fSApple OSS Distributions static enum thread_state_test_state {
445*e3723e1fSApple OSS Distributions INIT,
446*e3723e1fSApple OSS Distributions CHILD_READY,
447*e3723e1fSApple OSS Distributions PARENT_POPULATED_MATRIX_STATE,
448*e3723e1fSApple OSS Distributions CHILD_UPDATED_MATRIX_STATE,
449*e3723e1fSApple OSS Distributions DONE
450*e3723e1fSApple OSS Distributions } thread_state_test_state;
451*e3723e1fSApple OSS Distributions
452*e3723e1fSApple OSS Distributions static pthread_cond_t thread_state_test_cond = PTHREAD_COND_INITIALIZER;
453*e3723e1fSApple OSS Distributions static pthread_mutex_t thread_state_test_lock = PTHREAD_MUTEX_INITIALIZER;
454*e3723e1fSApple OSS Distributions
455*e3723e1fSApple OSS Distributions static void
wait_for_thread_state_test_state(enum thread_state_test_state state)456*e3723e1fSApple OSS Distributions wait_for_thread_state_test_state(enum thread_state_test_state state)
457*e3723e1fSApple OSS Distributions {
458*e3723e1fSApple OSS Distributions pthread_mutex_lock(&thread_state_test_lock);
459*e3723e1fSApple OSS Distributions while (thread_state_test_state != state) {
460*e3723e1fSApple OSS Distributions pthread_cond_wait(&thread_state_test_cond, &thread_state_test_lock);
461*e3723e1fSApple OSS Distributions }
462*e3723e1fSApple OSS Distributions pthread_mutex_unlock(&thread_state_test_lock);
463*e3723e1fSApple OSS Distributions }
464*e3723e1fSApple OSS Distributions
465*e3723e1fSApple OSS Distributions static void
thread_set_state_test_state(enum thread_state_test_state state)466*e3723e1fSApple OSS Distributions thread_set_state_test_state(enum thread_state_test_state state)
467*e3723e1fSApple OSS Distributions {
468*e3723e1fSApple OSS Distributions pthread_mutex_lock(&thread_state_test_lock);
469*e3723e1fSApple OSS Distributions thread_state_test_state = state;
470*e3723e1fSApple OSS Distributions pthread_cond_broadcast(&thread_state_test_cond);
471*e3723e1fSApple OSS Distributions pthread_mutex_unlock(&thread_state_test_lock);
472*e3723e1fSApple OSS Distributions }
473*e3723e1fSApple OSS Distributions
474*e3723e1fSApple OSS Distributions static void *
test_matrix_thread_state_child(void * arg __unused)475*e3723e1fSApple OSS Distributions test_matrix_thread_state_child(void *arg __unused)
476*e3723e1fSApple OSS Distributions {
477*e3723e1fSApple OSS Distributions const struct arm_matrix_operations *ops = arg;
478*e3723e1fSApple OSS Distributions
479*e3723e1fSApple OSS Distributions size_t size = ops->data_size();
480*e3723e1fSApple OSS Distributions uint8_t *d = ops->alloc_data();
481*e3723e1fSApple OSS Distributions
482*e3723e1fSApple OSS Distributions
483*e3723e1fSApple OSS Distributions thread_set_state_test_state(CHILD_READY);
484*e3723e1fSApple OSS Distributions wait_for_thread_state_test_state(PARENT_POPULATED_MATRIX_STATE);
485*e3723e1fSApple OSS Distributions ops->store_data(d);
486*e3723e1fSApple OSS Distributions for (size_t i = 0; i < size; i++) {
487*e3723e1fSApple OSS Distributions d[i]++;
488*e3723e1fSApple OSS Distributions }
489*e3723e1fSApple OSS Distributions ops->load_data(d);
490*e3723e1fSApple OSS Distributions thread_set_state_test_state(CHILD_UPDATED_MATRIX_STATE);
491*e3723e1fSApple OSS Distributions
492*e3723e1fSApple OSS Distributions wait_for_thread_state_test_state(DONE);
493*e3723e1fSApple OSS Distributions ops->stop();
494*e3723e1fSApple OSS Distributions return NULL;
495*e3723e1fSApple OSS Distributions }
496*e3723e1fSApple OSS Distributions
497*e3723e1fSApple OSS Distributions static void
test_matrix_thread_state(const struct arm_matrix_operations * ops)498*e3723e1fSApple OSS Distributions test_matrix_thread_state(const struct arm_matrix_operations *ops)
499*e3723e1fSApple OSS Distributions {
500*e3723e1fSApple OSS Distributions if (!ops->is_available()) {
501*e3723e1fSApple OSS Distributions T_SKIP("Running on non-%s target, skipping...", ops->name);
502*e3723e1fSApple OSS Distributions }
503*e3723e1fSApple OSS Distributions
504*e3723e1fSApple OSS Distributions size_t size = ops->data_size();
505*e3723e1fSApple OSS Distributions uint8_t *d = ops->alloc_data();
506*e3723e1fSApple OSS Distributions arc4random_buf(d, size);
507*e3723e1fSApple OSS Distributions
508*e3723e1fSApple OSS Distributions thread_state_test_state = INIT;
509*e3723e1fSApple OSS Distributions
510*e3723e1fSApple OSS Distributions pthread_t thread;
511*e3723e1fSApple OSS Distributions #pragma clang diagnostic push
512*e3723e1fSApple OSS Distributions #pragma clang diagnostic ignored "-Wincompatible-pointer-types-discards-qualifiers"
513*e3723e1fSApple OSS Distributions void *arg = ops;
514*e3723e1fSApple OSS Distributions #pragma clang diagnostic pop
515*e3723e1fSApple OSS Distributions int err = pthread_create(&thread, NULL, test_matrix_thread_state_child, arg);
516*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_EQ(err, 0, "pthread_create()");
517*e3723e1fSApple OSS Distributions
518*e3723e1fSApple OSS Distributions mach_port_t mach_thread = pthread_mach_thread_np(thread);
519*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_NE(mach_thread, MACH_PORT_NULL, "pthread_mach_thread_np()");
520*e3723e1fSApple OSS Distributions
521*e3723e1fSApple OSS Distributions wait_for_thread_state_test_state(CHILD_READY);
522*e3723e1fSApple OSS Distributions kern_return_t kr = ops->thread_set_state(mach_thread, d);
523*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_EQ(kr, KERN_SUCCESS, "%s thread_set_state()", ops->name);
524*e3723e1fSApple OSS Distributions thread_set_state_test_state(PARENT_POPULATED_MATRIX_STATE);
525*e3723e1fSApple OSS Distributions
526*e3723e1fSApple OSS Distributions wait_for_thread_state_test_state(CHILD_UPDATED_MATRIX_STATE);
527*e3723e1fSApple OSS Distributions uint8_t *thread_d = ops->alloc_data();
528*e3723e1fSApple OSS Distributions kr = ops->thread_get_state(mach_thread, thread_d);
529*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_EQ(kr, KERN_SUCCESS, "%s thread_get_state()", ops->name);
530*e3723e1fSApple OSS Distributions for (size_t i = 0; i < size; i++) {
531*e3723e1fSApple OSS Distributions d[i]++;
532*e3723e1fSApple OSS Distributions }
533*e3723e1fSApple OSS Distributions T_EXPECT_EQ(memcmp(d, thread_d, size), 0, "thread_get_state() read expected %s data from child thread", ops->name);
534*e3723e1fSApple OSS Distributions
535*e3723e1fSApple OSS Distributions thread_set_state_test_state(DONE);
536*e3723e1fSApple OSS Distributions free(thread_d);
537*e3723e1fSApple OSS Distributions free(d);
538*e3723e1fSApple OSS Distributions pthread_join(thread, NULL);
539*e3723e1fSApple OSS Distributions }
540*e3723e1fSApple OSS Distributions
541*e3723e1fSApple OSS Distributions #endif
542*e3723e1fSApple OSS Distributions
543*e3723e1fSApple OSS Distributions #ifdef __arm64__
544*e3723e1fSApple OSS Distributions
545*e3723e1fSApple OSS Distributions T_DECL(sme_thread_state,
546*e3723e1fSApple OSS Distributions "Test thread_{get,set}_state with SME thread state.",
547*e3723e1fSApple OSS Distributions XNU_T_META_SOC_SPECIFIC)
548*e3723e1fSApple OSS Distributions {
549*e3723e1fSApple OSS Distributions test_matrix_thread_state(&sme_operations);
550*e3723e1fSApple OSS Distributions }
551*e3723e1fSApple OSS Distributions
552*e3723e1fSApple OSS Distributions T_DECL(sme_exception_ports,
553*e3723e1fSApple OSS Distributions "Test that thread_set_exception_ports rejects SME thread-state flavors.",
554*e3723e1fSApple OSS Distributions XNU_T_META_SOC_SPECIFIC)
555*e3723e1fSApple OSS Distributions {
556*e3723e1fSApple OSS Distributions mach_port_t exc_port;
557*e3723e1fSApple OSS Distributions mach_port_t task = mach_task_self();
558*e3723e1fSApple OSS Distributions mach_port_t thread = mach_thread_self();
559*e3723e1fSApple OSS Distributions
560*e3723e1fSApple OSS Distributions kern_return_t kr = mach_port_allocate(task, MACH_PORT_RIGHT_RECEIVE, &exc_port);
561*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_MACH_SUCCESS(kr, "Allocated mach exception port");
562*e3723e1fSApple OSS Distributions kr = mach_port_insert_right(task, exc_port, exc_port, MACH_MSG_TYPE_MAKE_SEND);
563*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_MACH_SUCCESS(kr, "Inserted a SEND right into the exception port");
564*e3723e1fSApple OSS Distributions
565*e3723e1fSApple OSS Distributions kr = thread_set_exception_ports(thread, EXC_MASK_ALL, exc_port, EXCEPTION_STATE, ARM_THREAD_STATE64);
566*e3723e1fSApple OSS Distributions T_EXPECT_MACH_SUCCESS(kr, "thread_set_exception_ports accepts flavor %u", (unsigned int)ARM_THREAD_STATE64);
567*e3723e1fSApple OSS Distributions
568*e3723e1fSApple OSS Distributions for (thread_state_flavor_t flavor = ARM_SME_STATE; flavor <= ARM_SME2_STATE; flavor++) {
569*e3723e1fSApple OSS Distributions kr = thread_set_exception_ports(thread, EXC_MASK_ALL, exc_port, EXCEPTION_STATE, flavor);
570*e3723e1fSApple OSS Distributions T_EXPECT_MACH_ERROR(kr, KERN_INVALID_ARGUMENT, "thread_set_exception_ports rejects flavor %u", (unsigned int)flavor);
571*e3723e1fSApple OSS Distributions }
572*e3723e1fSApple OSS Distributions }
573*e3723e1fSApple OSS Distributions
574*e3723e1fSApple OSS Distributions T_DECL(sme_max_svl_b_sysctl,
575*e3723e1fSApple OSS Distributions "Test the hw.optional.arm.sme_max_svl_b sysctl",
576*e3723e1fSApple OSS Distributions XNU_T_META_SOC_SPECIFIC)
577*e3723e1fSApple OSS Distributions {
578*e3723e1fSApple OSS Distributions unsigned int max_svl_b;
579*e3723e1fSApple OSS Distributions size_t max_svl_b_size = sizeof(max_svl_b);
580*e3723e1fSApple OSS Distributions
581*e3723e1fSApple OSS Distributions int err = sysctlbyname("hw.optional.arm.sme_max_svl_b", &max_svl_b, &max_svl_b_size, NULL, 0);
582*e3723e1fSApple OSS Distributions T_QUIET; T_ASSERT_POSIX_SUCCESS(err, "sysctlbyname(hw.optional.arm.sme_max_svl_b)");
583*e3723e1fSApple OSS Distributions if (sme_operations.is_available()) {
584*e3723e1fSApple OSS Distributions /* Architecturally SVL must be a power-of-two between 128 and 2048 bits */
585*e3723e1fSApple OSS Distributions const unsigned int ARCH_MIN_SVL_B = 128 / 8;
586*e3723e1fSApple OSS Distributions const unsigned int ARCH_MAX_SVL_B = 2048 / 8;
587*e3723e1fSApple OSS Distributions
588*e3723e1fSApple OSS Distributions T_EXPECT_EQ(__builtin_popcount(max_svl_b), 1, "Maximum SVL_B is a power of 2");
589*e3723e1fSApple OSS Distributions T_EXPECT_GE(max_svl_b, ARCH_MIN_SVL_B, "Maximum SVL_B >= architectural minimum");
590*e3723e1fSApple OSS Distributions T_EXPECT_LE(max_svl_b, ARCH_MAX_SVL_B, "Maximum SVL_B <= architectural maximum");
591*e3723e1fSApple OSS Distributions } else {
592*e3723e1fSApple OSS Distributions T_EXPECT_EQ(max_svl_b, 0, "Maximum SVL_B is 0 when SME is unavailable");
593*e3723e1fSApple OSS Distributions }
594*e3723e1fSApple OSS Distributions }
595*e3723e1fSApple OSS Distributions #endif /* __arm64__ */
596