| /xnu-10063.101.15/osfmk/arm64/ |
| H A D | bcopy.s | 96 cmp x3, x2 99 cmp x2, #(kSmallCopy) 139 sub x2, x2, x5 // update length 142 subs x2, x2, #64 162 subs x2, x2, #32 173 add x1, x1, x2 178 add x3, x3, x2 199 subs x2, x2, #8 201 adds x2, x2, #8 205 subs x2, x2, #1 [all …]
|
| H A D | bzero.s | 55 mov x2, x1 58 cmp x2, #128 75 add x2, x2, x0 // end of buffer 77 subs x2, x2, x4 // if the end of the buffer comes first, jump 81 subs x2, x2, #64 // decrement length 83 1: add x3, x3, x2 // back up pointer to (end of buffer) - 64. 108 cmp x2, #64 122 add x2, x2, x0 // end of buffer 124 subs x2, x2, x4 // if the end of the buffer comes first, jump 131 subs x2, x2, #64 [all …]
|
| H A D | caches_asm.s | 66 and x2, x0, x9 68 add x1, x1, x2 148 CACHE_AT_LEVEL x2, x1, x3 149 cbz x2, 5f // No cache at this level, all higher levels may be skipped 150 cmp x2, #2 152 mov x2, x1 153 GET_CACHE_CONFIG x2, x9, x10, x11 154 lsl x2, x1, #1 // level field for cisw/csw, bits 1:3 157 dc $0, x2 // clean dcache line by way/set 158 add x2, x2, x9 // increment set index [all …]
|
| H A D | machine_routines_asm.s | 97 mov x2, x0 113 mov x2, x1 176 mrs x2, FPCR // Grab FPCR 180 and x2, x2, x0 // Be paranoid, and clear bits we expect to 182 orr x0, x1, x2 // OR them to get FPSCR equivalent state 202 orr x2, x4, x5 203 and x2, x2, x0 // Clear the bits that don't apply to FPCR 205 msr FPCR, x2 // Write FPSR 229 mrs x2, MDSCR_EL1 230 bic x2, x2, x0 [all …]
|
| H A D | strncmp.s | 75 cbz x2, L_scalarDone 84 subs x2, x2, #1 // decrement length 119 0: cmp x2, x7 121 sub x6, x2, x7 134 mov x2, x6 143 subs x2, x2, #1 // decrement length 158 subs x3, x2, #(kVectorSize) 162 mov x2, x3 // length -= kVectorSize 172 subs x2, x2, #(kVectorSize) 211 subs x2, x2, #1 // decrement length
|
| H A D | memcmp_zero.s | 85 mov x2, x0 // copy the original addr 88 ldp q4, q5, [x2] 89 ldp q6, q7, [x2, #32] 90 sub x2, x0, x2 // bytes between original and aligned addr 91 sub x1, x1, x2 // update length 132 0: ldr x2, [x0],#8 133 orr x3, x3, x2 // use orr to keep non-zero bytes
|
| H A D | strnlen.s | 80 and x2, x0, #-16 81 ldr q0, [x2] 107 0: ldr q0, [x2, #16]! 116 sub x0, x2, x0 134 sub x0, x2, x0 // index of vector in string 162 and x2, x0, #0xf 163 sub x3, x3, x2 200 add x0, x0, x2
|
| H A D | cswitch.s | 272 set_thread_registers x0, x1, x2 273 LOAD_KERN_STACK_TOP dst=x1, src=x0, tmp=x2 // Get top of kernel stack 275 set_process_dependent_keys_and_sync_context x0, x1, x2, x3, w4 303 mov x22, x2 //wait result 336 set_thread_registers x2, x3, x4 337 LOAD_KERN_STACK_TOP dst=x3, src=x2, tmp=x4 339 set_process_dependent_keys_and_sync_context x2, x3, x4, x5, w6 369 LOAD_KERN_STACK_TOP dst=x1, src=x0, tmp=x2 // Get the top of the kernel stack 371 LOAD_INT_STACK dst=x2, src=x0, tmp=x1 372 mov sp, x2 [all …]
|
| H A D | pinst.s | 87 check_instruction x2, x3, __pinst_set_ttbr1, 0xd65f03c0d5182020 92 check_instruction x2, x3, __pinst_set_vbar, 0xd65f03c0d518c000 97 check_instruction x2, x3, __pinst_set_tcr, 0xd65f03c0d5182040 102 check_instruction x2, x3, __pinst_set_sctlr, 0xd65f03c0d5181000 123 check_instruction x2, x3, __pinst_spsel_1, 0xd65f03c0d50041bf
|
| H A D | start.s | 157 cmp x0, x2 // Compare cpu data phys cpu and MPIDR_EL1 phys cpu 173 adrp x2, EXT(resume_idle_cpu)@page 174 add x2, x2, EXT(resume_idle_cpu)@pageoff 175 cmp x0, x2 177 adrp x2, EXT(start_cpu)@page 178 add x2, x2, EXT(start_cpu)@pageoff 179 cmp x0, x2 529 mov x2, #(TTE_PGENTRIES) // Load number of entries per page 530 lsl x2, x2, #2 // Shift by 2 for num entries on 4 pages 534 subs x2, x2, #1 // entries-- [all …]
|
| H A D | locore.s | 205 stp x2, x3, [sp, #-16]! // Save {x2-x3} 207 mov x2, #(ESR_EC_SP_ALIGN << ESR_EC_SHIFT) 208 cmp x1, x2 // If we have a stack alignment exception 210 mov x2, #(ESR_EC_DABORT_EL1 << ESR_EC_SHIFT) 211 cmp x1, x2 // If we have a data abort, we need to 216 LOAD_KERN_STACK_TOP dst=x2, src=x1, tmp=x3 // Get top of kernel stack 217 sub x3, x2, KERNEL_STACK_SIZE // Find bottom of kernel stack 218 cmp x0, x2 // if (SP_EL0 >= kstack top) 224 ldr x2, [x1, CPU_INTSTACK_TOP] // Get top of istack 225 sub x3, x2, INTSTACK_SIZE_NUM // Find bottom of istack [all …]
|
| H A D | smccc_asm.h | 44 stp x2, x3, [sp, #- 16]! 54 ldp x2, x3, [sp], #16
|
| H A D | arm64_hypercall.c | 54 hvc_5(uint64_t *x0, uint64_t *x1, uint64_t *x2, uint64_t *x3, uint64_t *x4) in hvc_5() argument 70 [o2] "=m" (*x2), in hvc_5() 75 [i2] "r" (*x2), in hvc_5()
|
| H A D | machine_routines_asm.h | 87 mov \tmp2, x2 93 mov x2, \tmp2
|
| H A D | lz4_encode_arm64.s | 47 #define src_ptr x2 86 ldr x11, [x2] // src 386 str x11, [x2]
|
| /xnu-10063.101.15/osfmk/arm64/sptm/ |
| H A D | start_sptm.s | 137 mov x27, x2 205 cmp x0, x2 234 adrp x2, EXT(arm_init_cpu)@page 235 add x2, x2, EXT(arm_init_cpu)@pageoff 236 cmp x1, x2 239 adrp x2, EXT(arm_init_idle_cpu)@page 240 add x2, x2, EXT(arm_init_idle_cpu)@pageoff 241 cmp x1, x2
|
| /xnu-10063.101.15/osfmk/corecrypto/ |
| H A D | ccmode_gcm_gf_mult.c | 47 cc_dunit x1, x2, x3, x4, x5; in bmul64() local 59 x2 = x & m2; in bmul64() 68 z = (x1 * y1) ^ (x2 * y5) ^ (x3 * y4) ^ (x4 * y3) ^ (x5 * y2); in bmul64() 70 z = (x1 * y2) ^ (x2 * y1) ^ (x3 * y5) ^ (x4 * y4) ^ (x5 * y3); in bmul64() 72 z = (x1 * y3) ^ (x2 * y2) ^ (x3 * y1) ^ (x4 * y5) ^ (x5 * y4); in bmul64() 74 z = (x1 * y4) ^ (x2 * y3) ^ (x3 * y2) ^ (x4 * y1) ^ (x5 * y5); in bmul64() 76 z = (x1 * y5) ^ (x2 * y4) ^ (x3 * y3) ^ (x4 * y2) ^ (x5 * y1); in bmul64() 123 uint32_t x0, x1, x2, x3; in bmul32() local 134 x2 = x & m4; in bmul32() 141 z0 = ((uint64_t)x0 * y0) ^ ((uint64_t)x1 * y3) ^ ((uint64_t)x2 * y2) ^ ((uint64_t)x3 * y1); in bmul32() [all …]
|
| /xnu-10063.101.15/osfmk/arm/commpage/ |
| H A D | commpage_asm.s | 151 stp x2, x9, [sp, #-16]! 158 ldp x2, x9, [sp], #16 291 str x1, [x10, x2] // Set old tail -> offset = new elem 326 casa w10, w11, [x2] // Atomic CAS with acquire barrier 349 stlr wzr, [x2] // Drop spin lock with release barrier (pairs with acquire in casa) 374 str xzr, [x1, x2] // Zero the forward link in the new element 431 add x2, x0, #16 // address of lock = x2 = x0 + 16 442 ldxr w9, [x2] // arm the monitor for the lock address 454 BACKOFF x2
|
| /xnu-10063.101.15/libsyscall/wrappers/ |
| H A D | varargs_wrappers.s | 42 ldp x2, x3, [fp, #16] 65 ldr x2, [fp, #16] 81 ldp x1, x2, [fp, #16] 98 ldp x1, x2, [fp, #16] 132 ldp x1, x2, [fp, #16]
|
| H A D | mach_absolute_time.s | 231 ldr x2, [x3] // Load the offset 232 cmp x1, x2 // Compare our offset values... 258 cmp x2, #USER_TIMEBASE_NONE // Are userspace reads supported? 260 cmp x2, #USER_TIMEBASE_NOSPEC 262 cmp x2, #USER_TIMEBASE_NOSPEC_APPLE
|
| /xnu-10063.101.15/osfmk/kern/ |
| H A D | arithmetic_128.h | 71 uint32_t x2 = (uint32_t)x; in mul64x64() local 83 prod->low = (uint64_t)x2 * (uint64_t)y2; in mul64x64() 88 add.low = (uint64_t)x2 * (uint64_t)y1; in mul64x64()
|
| /xnu-10063.101.15/san/memory/ |
| H A D | kasan-test-arm64.s | 46 ldp fp, x2, [x0, JMP_fp_sp] 51 add sp, x2, #0
|
| /xnu-10063.101.15/libsyscall/custom/ |
| H A D | __syscall.s | 65 ldp x1, x2, [sp]
|
| H A D | __gettimeofday.s | 114 movz x2, #0x0
|
| /xnu-10063.101.15/bsd/kern/code_signing/ |
| H A D | txm.c | 243 registers->x2 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup() 253 registers->x2 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup() 262 registers->x2 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup() 270 registers->x2 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup() 277 registers->x2 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup() 283 registers->x2 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup()
|