| /xnu-12377.81.4/osfmk/arm64/ |
| H A D | bzero.s | 55 mov x2, x1 58 cmp x2, #128 75 add x2, x2, x0 // end of buffer 77 subs x2, x2, x4 // if the end of the buffer comes first, jump 81 subs x2, x2, #64 // decrement length 83 1: add x3, x3, x2 // back up pointer to (end of buffer) - 64. 97 subs x2, x2, #8 99 adds x2, x2, #8 102 subs x2, x2, #1 124 cmp x2, #64 [all …]
|
| H A D | bcopy.s | 97 cmp x3, x2 100 cmp x2, #(kSmallCopy) 140 sub x2, x2, x5 // update length 143 subs x2, x2, #64 163 subs x2, x2, #32 174 add x1, x1, x2 179 add x3, x3, x2 200 subs x2, x2, #8 202 adds x2, x2, #8 206 subs x2, x2, #1 [all …]
|
| H A D | caches_asm.s | 67 and x2, x0, x9 69 add x1, x1, x2 149 CACHE_AT_LEVEL x2, x1, x3 150 cbz x2, 5f // No cache at this level, all higher levels may be skipped 151 cmp x2, #2 153 mov x2, x1 154 GET_CACHE_CONFIG x2, x9, x10, x11 155 lsl x2, x1, #1 // level field for cisw/csw, bits 1:3 158 dc $0, x2 // clean dcache line by way/set 159 add x2, x2, x9 // increment set index [all …]
|
| H A D | locore.s | 109 mov x2, #0 145 mrs x2, ELR_EL1 148 and x5, x2, x3 156 mrs x2, FAR_EL1 158 and x5, x2, x3 275 stp x2, x3, [sp, #-16]! // Save {x2-x3} 277 mov x2, #(ESR_EC_SP_ALIGN << ESR_EC_SHIFT) 278 cmp x1, x2 // If we have a stack alignment exception 280 mov x2, #(ESR_EC_DABORT_EL1 << ESR_EC_SHIFT) 281 cmp x1, x2 // If we have a data abort, we need to [all …]
|
| H A D | strncmp.s | 75 cbz x2, L_scalarDone 84 subs x2, x2, #1 // decrement length 134 0: cmp x2, x7 136 sub x6, x2, x7 149 mov x2, x6 162 subs x2, x2, #1 // decrement length 192 subs x3, x2, #(kVectorSize) 196 mov x2, x3 // length -= kVectorSize 206 subs x2, x2, #(kVectorSize) 257 subs x2, x2, #1 // decrement length
|
| H A D | machine_routines_asm.s | 111 mov x2, x0 128 mov x2, x1 193 mrs x2, FPCR // Grab FPCR 197 and x2, x2, x0 // Be paranoid, and clear bits we expect to 199 orr x0, x1, x2 // OR them to get FPSCR equivalent state 220 orr x2, x4, x5 221 and x2, x2, x0 // Clear the bits that don't apply to FPCR 223 msr FPCR, x2 // Write FPSR 322 LOAD_PMAP_CPU_DATA x0, x1, x2 412 mrs x2, DAIF // Load current DAIF [all …]
|
| H A D | memcmp_zero.s | 85 mov x2, x0 // copy the original addr 88 ldp q4, q5, [x2] 89 ldp q6, q7, [x2, #32] 90 sub x2, x0, x2 // bytes between original and aligned addr 91 sub x1, x1, x2 // update length 132 0: ldr x2, [x0],#8 133 orr x3, x3, x2 // use orr to keep non-zero bytes
|
| H A D | strnlen.s | 82 and x2, x0, #-16 83 ldr q0, [x2] 109 0: ldr q0, [x2, #16]! 118 sub x0, x2, x0 136 sub x0, x2, x0 // index of vector in string 164 and x2, x0, #0xf 165 sub x3, x3, x2 202 add x0, x0, x2
|
| H A D | cswitch.s | 290 set_thread_registers x0, x1, x2 291 LOAD_KERN_STACK_TOP dst=x1, src=x0, tmp=x2 // Get top of kernel stack 293 set_process_dependent_keys_and_sync_context x0, x1, x2, x3, w4 322 mov x22, x2 //wait result 356 set_thread_registers x2, x3, x4 357 LOAD_KERN_STACK_TOP dst=x3, src=x2, tmp=x4 359 set_process_dependent_keys_and_sync_context x2, x3, x4, x5, w6 391 LOAD_KERN_STACK_TOP dst=x1, src=x0, tmp=x2 // Get the top of the kernel stack 393 LOAD_INT_STACK_THREAD dst=x2, src=x0, tmp=x1 394 mov sp, x2 [all …]
|
| H A D | pinst.s | 89 check_instruction x2, x3, __pinst_set_ttbr1, 0xd65f03c0d5182020 95 check_instruction x2, x3, __pinst_set_vbar, 0xd65f03c0d518c000 101 check_instruction x2, x3, __pinst_set_tcr, 0xd65f03c0d5182040 107 check_instruction x2, x3, __pinst_set_sctlr, 0xd65f03c0d5181000 129 check_instruction x2, x3, __pinst_spsel_1, 0xd65f03c0d50041bf
|
| H A D | start.s | 157 cmp x0, x2 // Compare cpu data phys cpu and MPIDR_EL1 phys cpu 171 SET_PIO_ONLY_REGISTERS x21, x2, x3, x4, x5, x6 180 adrp x2, EXT(resume_idle_cpu)@page 181 add x2, x2, EXT(resume_idle_cpu)@pageoff 182 cmp x0, x2 184 adrp x2, EXT(start_cpu)@page 185 add x2, x2, EXT(start_cpu)@pageoff 186 cmp x0, x2 538 mov x2, #(TTE_PGENTRIES) // Load number of entries per page 539 lsl x2, x2, #2 // Shift by 2 for num entries on 4 pages [all …]
|
| H A D | smccc_asm.h | 44 stp x2, x3, [sp, #- 16]! 54 ldp x2, x3, [sp], #16
|
| H A D | arm64_hypercall.c | 54 hvc_5(uint64_t *x0, uint64_t *x1, uint64_t *x2, uint64_t *x3, uint64_t *x4) in hvc_5() argument 70 [o2] "=m" (*x2), in hvc_5() 75 [i2] "r" (*x2), in hvc_5()
|
| H A D | machine_routines_asm.h | 89 mov \tmp2, x2 95 mov x2, \tmp2
|
| H A D | lz4_encode_arm64.s | 47 #define src_ptr x2 86 ldr x11, [x2] // src 386 str x11, [x2]
|
| /xnu-12377.81.4/osfmk/corecrypto/ |
| H A D | ccmode_gcm_gf_mult.c | 47 cc_dunit x1, x2, x3, x4, x5; in bmul64() local 59 x2 = x & m2; in bmul64() 68 z = (x1 * y1) ^ (x2 * y5) ^ (x3 * y4) ^ (x4 * y3) ^ (x5 * y2); in bmul64() 70 z = (x1 * y2) ^ (x2 * y1) ^ (x3 * y5) ^ (x4 * y4) ^ (x5 * y3); in bmul64() 72 z = (x1 * y3) ^ (x2 * y2) ^ (x3 * y1) ^ (x4 * y5) ^ (x5 * y4); in bmul64() 74 z = (x1 * y4) ^ (x2 * y3) ^ (x3 * y2) ^ (x4 * y1) ^ (x5 * y5); in bmul64() 76 z = (x1 * y5) ^ (x2 * y4) ^ (x3 * y3) ^ (x4 * y2) ^ (x5 * y1); in bmul64() 123 uint32_t x0, x1, x2, x3; in bmul32() local 134 x2 = x & m4; in bmul32() 141 z0 = ((uint64_t)x0 * y0) ^ ((uint64_t)x1 * y3) ^ ((uint64_t)x2 * y2) ^ ((uint64_t)x3 * y1); in bmul32() [all …]
|
| /xnu-12377.81.4/osfmk/arm/commpage/ |
| H A D | commpage_asm.s | 151 stp x2, x9, [sp, #-16]! 158 ldp x2, x9, [sp], #16 291 str x1, [x10, x2] // Set old tail -> offset = new elem 326 casa w10, w11, [x2] // Atomic CAS with acquire barrier 349 stlr wzr, [x2] // Drop spin lock with release barrier (pairs with acquire in casa) 374 str xzr, [x1, x2] // Zero the forward link in the new element 431 add x2, x0, #16 // address of lock = x2 = x0 + 16 442 ldxr w9, [x2] // arm the monitor for the lock address 454 BACKOFF x2
|
| /xnu-12377.81.4/libsyscall/wrappers/ |
| H A D | varargs_wrappers.s | 42 ldp x2, x3, [fp, #16] 65 ldr x2, [fp, #16] 81 ldp x1, x2, [fp, #16] 98 ldp x1, x2, [fp, #16] 132 ldp x1, x2, [fp, #16]
|
| H A D | mach_absolute_time.s | 231 ldr x2, [x3] // Load the offset 232 cmp x1, x2 // Compare our offset values... 258 cmp x2, #USER_TIMEBASE_NONE // Are userspace reads supported? 260 cmp x2, #USER_TIMEBASE_NOSPEC 262 cmp x2, #USER_TIMEBASE_NOSPEC_APPLE
|
| /xnu-12377.81.4/osfmk/arm64/sptm/ |
| H A D | start_sptm.s | 137 mov x27, x2 225 cmp x0, x2 251 ldr x2, [x21, CPU_RESET_HANDLER] 252 cbz x2, Lskip_cpu_reset_handler 257 cmp x2, x3 262 cmp x2, x3
|
| /xnu-12377.81.4/osfmk/kern/ |
| H A D | arithmetic_128.h | 71 uint32_t x2 = (uint32_t)x; in mul64x64() local 83 prod->low = (uint64_t)x2 * (uint64_t)y2; in mul64x64() 88 add.low = (uint64_t)x2 * (uint64_t)y1; in mul64x64()
|
| /xnu-12377.81.4/san/memory/ |
| H A D | kasan-test-arm64.s | 48 ldp fp, x2, [x0, JMP_fp_sp] 53 add sp, x2, #0
|
| /xnu-12377.81.4/libsyscall/custom/ |
| H A D | __syscall.s | 65 ldp x1, x2, [sp]
|
| H A D | __gettimeofday.s | 114 movz x2, #0x0
|
| /xnu-12377.81.4/bsd/dev/arm64/ |
| H A D | cpu_in_cksum.s | 101 #define off x2 431 and x2, x4, x3, lsr #16
|