| /xnu-10063.101.15/osfmk/arm64/ |
| H A D | strncmp.s | 81 subs x3, x4, x5 // if the are not equal 140 subs x3, x4, x5 // if the are not equal 161 add x5, x1, x3 177 mov x1, x5 202 sub x0, x4, x5 208 subs x3, x4, x5 // if the are not equal
|
| H A D | bcopy.s | 116 sub x5, x3, x0 // bytes between original dst and aligned dst 117 add x1, x1, x5 // update src pointer 139 sub x2, x2, x5 // update length 245 sub x5, x4, x3 // buffer. Here we use dst-1 & -32 instead, 246 sub x1, x1, x5 // because we are going backwards. 247 sub x2, x2, x5
|
| H A D | machine_routines_asm.s | 172 mov x5, #(FPSR_MASK & 0xFFFF0000) 173 orr x0, x4, x5 178 mov x5, #(FPCR_MASK & 0xFFFF0000) 179 orr x0, x4, x5 197 mov x5, #(FPSR_MASK & 0xFFFF0000) 198 orr x1, x4, x5 201 mov x5, #(FPCR_MASK & 0xFFFF0000) 202 orr x2, x4, x5 718 cbz x5, Lcopyinstr_done // If this byte is null, we're done 722 mov x5, #ENAMETOOLONG // Set current byte to error code for later return [all …]
|
| H A D | machine_routines_asm.h | 74 mov \tmp5, x5 90 mov x5, x17 108 mov x5, \tmp5
|
| H A D | lz4_encode_arm64.s | 50 #define hash_table x5 106 add x20, x5, x20, lsl #3 // hash_table_entry ptr (hash + 8*index) 115 add x21, x5, x21, lsl #3 // hash_table_entry ptr (hash + 8*index) 136 add x20, x5, x20, lsl #3 // hash_table_entry ptr (hash + 8*index) 157 add x20, x5, x20, lsl #3 // hash_table_entry ptr (hash + 8*index) 178 add x20, x5, x20, lsl #3 // hash_table_entry ptr (hash + 8*index)
|
| H A D | cswitch.s | 295 LOAD_KERN_STACK_TOP dst=x5, src=x4, tmp=x6 296 mov sp, x5 // Set stack pointer 299 set_process_dependent_keys_and_sync_context x4, x5, x6, x7, w20 339 set_process_dependent_keys_and_sync_context x2, x3, x4, x5, w6
|
| H A D | locore.s | 643 mov x5, #0 825 mrs x5, FAR_EL1 834 mov x11, x5 841 ccmp x5, x11, #0b0100 /* Z/EQ */, HS 1469 mrs x5, FPCR 1470 CMSR FPCR, x5, x4, 1 1497 ldp x4, x5, [x0, SS64_X4] 1645 GET_PMAP_CPU_DATA x5, x6, x7 1646 ldr x6, [x5, PMAP_CPU_DATA_KERN_SAVED_SP] 1668 GET_PMAP_CPU_DATA x5, x6, x7 [all …]
|
| H A D | exception_asm.h | 150 stp x4, x5, [x0, SS64_X4] 216 mov x5, x17
|
| H A D | start.s | 606 lsr x5, x24, #(ARM_TT_L2_SHIFT) 608 create_bootstrap_mapping x0, x4, x5, x1, x2, x6, x10, x11, x12, x13 611 lsr x5, x24, #(ARM_TT_L2_SHIFT) 612 create_bootstrap_mapping x14, x15, x5, x3, x2, x9, x10, x11, x12, x13
|
| H A D | lz4_decode_arm64.s | 58 #define n_matches x5
|
| H A D | WKdmDecompress_16k.s | 110 #define rcx x5
|
| H A D | WKdmDecompress_4k.s | 110 #define rcx x5
|
| H A D | WKdmCompress_4k.s | 185 #define dictionary x5
|
| H A D | WKdmCompress_16k.s | 187 #define dictionary x5
|
| /xnu-10063.101.15/osfmk/corecrypto/ |
| H A D | ccmode_gcm_gf_mult.c | 47 cc_dunit x1, x2, x3, x4, x5; in bmul64() local 65 x5 = x & m5; in bmul64() 68 z = (x1 * y1) ^ (x2 * y5) ^ (x3 * y4) ^ (x4 * y3) ^ (x5 * y2); in bmul64() 70 z = (x1 * y2) ^ (x2 * y1) ^ (x3 * y5) ^ (x4 * y4) ^ (x5 * y3); in bmul64() 72 z = (x1 * y3) ^ (x2 * y2) ^ (x3 * y1) ^ (x4 * y5) ^ (x5 * y4); in bmul64() 74 z = (x1 * y4) ^ (x2 * y3) ^ (x3 * y2) ^ (x4 * y1) ^ (x5 * y5); in bmul64() 76 z = (x1 * y5) ^ (x2 * y4) ^ (x3 * y3) ^ (x4 * y2) ^ (x5 * y1); in bmul64()
|
| /xnu-10063.101.15/libsyscall/custom/ |
| H A D | __syscall.s | 67 ldp x5, x6, [sp, #32]
|
| /xnu-10063.101.15/bsd/kern/code_signing/ |
| H A D | txm.c | 246 registers->x5 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup() 256 registers->x5 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup() 265 registers->x5 = va_arg(args, uintptr_t); in txm_kernel_call_registers_setup()
|
| /xnu-10063.101.15/bsd/dev/arm64/ |
| H A D | cpu_in_cksum.s | 102 #define started_on_odd x5
|
| H A D | cpu_copy_in_cksum.s | 69 #define need_swap x5
|
| /xnu-10063.101.15/libsyscall/mach/ |
| H A D | err_libkern.sub | 67 "(libkern/kext) invalid argument", /* 0x5 */
|
| /xnu-10063.101.15/tools/lldbmacros/core/ |
| H A D | operating_system.py | 73 self.x5 = 0 109 return struct.pack('34QII', self.x0, self.x1, self.x2, self.x3, self.x4, self.x5, 127 self.x5 = saved_state.GetChildMemberWithName('x').GetChildAtIndex(5).GetValueAsUnsigned()
|