| /xnu-12377.81.4/osfmk/arm64/ |
| H A D | start.s | 51 msr VBAR_EL1, x0 57 mov x0, x1 72 msr TTBR1_EL1, x0 87 msr SCTLR_EL1, x0 135 adrp x0, EXT(LowExceptionVectorBase)@page 136 add x0, x0, EXT(LowExceptionVectorBase)@pageoff 137 msr VBAR_EL1, x0 147 and x0, x15, #0xFFFF // CPU number in Affinity0, cluster ID in Affinity1 149 and x0, x15, #0xFF // CPU number is in MPIDR Affinity Level 0 157 cmp x0, x2 // Compare cpu data phys cpu and MPIDR_EL1 phys cpu [all …]
|
| H A D | machine_routines_asm.s | 111 mov x2, x0 112 MOV64 x0, VMAPPLE_PAC_SET_EL0_DIVERSIFIER_AT_EL1 115 cbnz x0, . 116 LOAD_CPU_JOP_KEY x0, x1 129 MOV64 x0, VMAPPLE_PAC_SET_EL0_DIVERSIFIER_AT_EL1 132 cbnz x0, . 190 orr x0, x4, x5 191 and x1, x1, x0 // Be paranoid, and clear bits we expect to 196 orr x0, x4, x5 197 and x2, x2, x0 // Be paranoid, and clear bits we expect to [all …]
|
| H A D | locore.s | 107 mov x0, #0 132 mrs x0, TPIDR_EL1 138 cbz x0, Lbegin_panic_lockdown_real_\@ 139 ldr x1, [x0, TH_EXPECTED_FAULT_HANDLER] 144 ldr x1, [x0, TH_EXPECTED_FAULT_PC] 155 ldr x1, [x0, TH_EXPECTED_FAULT_ADDR] 170 mov x0, #0 // not a simulated lockdown 179 mov x0, #1 // this is a simulated lockdown! 283 mrs x0, SP_EL0 // Get SP_EL0 296 cmp x0, x2 // if (SP_EL0 >= kstack top) [all …]
|
| H A D | caches_asm.s | 67 and x2, x0, x9 68 bic x0, x0, x9 // Cached aligned 73 ic ivau, x0 // Invalidate icache line 74 add x0, x0, #1<<MMU_I_CLINE // Get next cache aligned addr 168 cmp x1, x0 222 mrs x0, CLIDR_EL1 223 ubfx x0, x0, #24, #3 // extract CLIDR_EL1.LoC 242 mrs x0, CLIDR_EL1 243 ubfx x0, x0, #21, 3 // extract CLIDR_EL1.LoUIS 261 and x2, x0, x9 [all …]
|
| H A D | exception_asm.h | 173 stp x2, x3, [x0, SS64_X2] 175 stp x4, x5, [x0, SS64_X4] 176 stp x6, x7, [x0, SS64_X6] 177 stp x8, x9, [x0, SS64_X8] 178 stp x10, x11, [x0, SS64_X10] 179 stp x12, x13, [x0, SS64_X12] 180 stp x14, x15, [x0, SS64_X14] 181 stp x16, x17, [x0, SS64_X16] 182 stp x18, x19, [x0, SS64_X18] 183 stp x20, x21, [x0, SS64_X20] [all …]
|
| H A D | memcmp_zero.s | 85 mov x2, x0 // copy the original addr 86 add x0, x0, #64 87 and x0, x0, #-64 // aligned addr 90 sub x2, x0, x2 // bytes between original and aligned addr 96 ldp q0, q1, [x0] 97 ldp q2, q3, [x0, #32] 102 add x0, x0, #64 // advance pointer 111 add x0, x0, x1 112 ldp q0, q1, [x0] 113 ldp q2, q3, [x0, #32] [all …]
|
| H A D | platform_tests_asm.s | 75 ldr x0, [x0] 118 braaz x0 124 ldraa x0, [x0] 132 autiza x0 157 mov x0, sp 161 str x0, [sp, #-16]! 170 mrs x0, SP_EL0 171 ldr x1, [x0], #16 176 mov sp, x0 178 mov x0, #1 [all …]
|
| H A D | strnlen.s | 82 and x2, x0, #-16 96 and x4, x0, #0xf 118 sub x0, x2, x0 120 add x0, x0, x1 124 mov x0, #0 136 sub x0, x2, x0 // index of vector in string 139 add x0, x0, x1 150 and x1, x0, #-16 164 and x2, x0, #0xf 201 sub x0, x1, x0 [all …]
|
| H A D | strncmp.s | 77 0: tst x0, #(kVectorSize-1) 79 ldrb w4, [x0],#1 // load byte from src1 90 mov x0, x3 101 ldrb w4, [x0] 139 1: ldr q0, [x0],#(kVectorSize) 153 2: ldrb w4, [x0],#1 // load byte from src1 168 tst x0, #(kVectorSize-1) 180 ldrb w4, [x0, #-1] 182 sub x0, x4, x5 194 add x4, x0, x3 // save the addresses of the last vectors [all …]
|
| H A D | arm64_hypercall.c | 54 hvc_5(uint64_t *x0, uint64_t *x1, uint64_t *x2, uint64_t *x3, uint64_t *x4) in hvc_5() argument 68 : [o0] "=m" (*x0), in hvc_5() 73 : [i0] "r" (*x0), in hvc_5() 80 return *(int64_t *)x0 >= 0; in hvc_5() 84 hvc_2(uint64_t *x0, uint64_t *x1) in hvc_2() argument 87 return hvc_5(x0, x1, &x, &x, &x); in hvc_2() 91 hvc_1(uint64_t *x0) in hvc_1() argument 94 return hvc_5(x0, &x, &x, &x, &x); in hvc_1() 193 uint64_t x0 = VMAPPLE_PAC_NOP; in hvg_is_hcall_available() local 194 (void) hvc_1(&x0); in hvg_is_hcall_available() [all …]
|
| H A D | bzero.s | 57 mov x3, x0 69 stp x1, x1, [x0] 70 stp x1, x1, [x0, #16] 71 stp x1, x1, [x0, #32] 72 stp x1, x1, [x0, #48] 73 add x3, x0, #64 75 add x2, x2, x0 // end of buffer 123 mov x3, x0 135 stp x1, x1, [x0] 136 add x3, x0, #16 [all …]
|
| H A D | cswitch.s | 290 set_thread_registers x0, x1, x2 291 LOAD_KERN_STACK_TOP dst=x1, src=x0, tmp=x2 // Get top of kernel stack 293 set_process_dependent_keys_and_sync_context x0, x1, x2, x3, w4 294 mov x0, #0 // Clear argument to thread_continue 320 mov x20, x0 //continuation 325 mov x0, #1 329 mov x0, x21 // Set the first parameter 337 mrs x0, TPIDR_EL1 // Get the current thread pointer 353 LOAD_KERN_STACK_TOP dst=x3, src=x0, tmp=x4 // Get the old kernel stack top 390 mrs x0, TPIDR_EL1 // Get thread pointer [all …]
|
| H A D | bcopy.s | 74 mov x3, x0 75 mov x0, x1 96 sub x3, x0, x1 99 mov x3, x0 // copy destination pointer 117 sub x5, x3, x0 // bytes between original dst and aligned dst 141 stp x12,x13,[x0] // initial unaligned store 142 stp x14,x15,[x0, #16] // initial unaligned store 233 add x4, x0, x2 273 stp x12,x13,[x0, #16] // In the forward copy, we need to compute the 274 stp x14,x15,[x0] // address of these stores, but here we already
|
| H A D | pac_asm.h | 111 MOV64 x0, VMAPPLE_PAC_SET_EL0_DIVERSIFIER 114 cbnz x0, . 132 MOV64 x0, VMAPPLE_PAC_SET_B_KEYS 135 cbnz x0, . 158 mov x0, #VMAPPLE_PAC_SET_INITIAL_STATE 160 cbnz x0, .
|
| H A D | pinst.s | 63 msr TTBR1_EL1, x0 67 msr VBAR_EL1, x0 71 msr TCR_EL1, x0 76 msr SCTLR_EL1, x0
|
| H A D | iofilter_asm.s | 46 at s1e1w, x0 // Get PA of the addr passed in for comparison. 51 bfxil x10, x0, #0, #14 // Copy the page offset from the VA to assemble the PA. 53 csel x0, x10, xzr, eq // If translation was successful return PA, else 0.
|
| H A D | machine_routines_asm.h | 60 ldr w2, [x0, SS64_CPSR] 67 ldr x1, [x0, SS64_PC] 68 ldp x16, x17, [x0, SS64_X16] 90 ldr x3, [x0, SS64_LR] 114 ldr lr, [x0, SS64_LR]
|
| H A D | smccc_asm.h | 43 stp x0, x1, [sp, #- 16]! 55 ldp x0, x1, [sp], #16
|
| /xnu-12377.81.4/san/memory/ |
| H A D | kasan-test-arm64.s | 24 stp x19, x20, [x0, JMP_r19_20] 25 stp x21, x22, [x0, JMP_r21_22] 26 stp x23, x24, [x0, JMP_r23_24] 27 stp x25, x26, [x0, JMP_r25_26] 28 stp x27, x28, [x0, JMP_r27_28] 29 stp x29, lr, [x0, JMP_r29_lr] 30 stp fp, x1, [x0, JMP_fp_sp] 31 stp d8, d9, [x0, JMP_d8_d9] 32 stp d10, d11, [x0, JMP_d10_d11] 33 stp d12, d13, [x0, JMP_d12_d13] [all …]
|
| /xnu-12377.81.4/osfmk/arm/commpage/ |
| H A D | commpage_asm.s | 150 stp x0, x1, [sp, #-16]! 154 mov x0, \lock_addr 159 ldp x0, x1, [sp], #16 170 stp x0, xzr, [sp, #-16]! // Save x0 since it'll be clobbered by return value 173 mov \result, x0 175 ldp x0, xzr, [sp], #16 // Restore saved registers 185 stp x0, xzr, [sp, #-16]! // Save x0 since it'll be clobbered by return value 188 mov \result, x0 190 ldp x0, xzr, [sp], #16 // Restore saved registers 279 mov x0, #-1 // Failed [all …]
|
| /xnu-12377.81.4/doc/debugging/ |
| H A D | task_ref.md | 71 0xffffff801ace9250 task_kernel 68 367663 367595 0x0 72 0xffffff801ace9288 task_internal 974 4953 3979 0x0 73 0xffffff801ace92c0 task_mig 0 3670 3670 0x0 74 0xffffff801ace9218 task_external 35 108 73 0x0 75 0xffffff9369dc7b20 task_com.apple.iokit.IOAcceleratorFamily2 29 77 48 0x0 76 0xffffff936a3f0a20 task_com.apple.iokit.CoreAnalyticsFamily 1 1 0 0x0 77 0xffffff936a22cb20 task_com.apple.iokit.EndpointSecurity 0 1 1 0x0 78 0xffffff936a283f60 task_com.apple.iokit.IOSurface 5 5 0 0x0 79 0xffffff936a3f08a0 task_com.apple.security.sandbox 0 24 24 0x0 88 0xffffff936a4b9200 task_local_kernel 1 6 5 0x0 [all …]
|
| /xnu-12377.81.4/osfmk/arm64/sptm/ |
| H A D | start_sptm.s | 79 cmp x0, x8 91 mov x0, x1 111 cmp x0, x20 142 mov x0, x26 167 mov x0, x26 199 and x0, x15, #(MPIDR_AFF0_MASK | MPIDR_AFF1_MASK) 201 and x0, x15, #(MPIDR_AFF0_MASK) 225 cmp x0, x2 247 mov x0, x21 278 MOV64 x0, 0xDEADB001 [all …]
|
| /xnu-12377.81.4/bsd/dev/arm64/ |
| H A D | cpu_copy_in_cksum.s | 69 #define src x0 328 and x0, sum, #0xffff 329 add x0, x0, sum, lsr #16 332 add x0, x0, partial, lsr #16 334 add x0, x0, partial 337 and t, x0, #0xffff 338 add x0, t, x0, lsr #16 344 and t, x0, #0xffff 345 add x0, t, x0, lsr #16
|
| H A D | cpu_in_cksum.s | 99 #define m x0 110 #define ptr_m x0 157 mov x0, x3 429 and x0, x4, x3, lsr #48 452 adrp x0, Lin_cksum_whoops_str@page 453 add x0, x0, Lin_cksum_whoops_str@pageoff 455 mov x0, #-1
|
| /xnu-12377.81.4/libsyscall/wrappers/ |
| H A D | __get_cpu_capabilities.s | 69 ldr x0, Lcommpage_cc_addr 70 ldr x0, [x0]
|