| /xnu-11215.61.5/tests/ |
| H A D | avx.c | 163 __asm__ volatile ("vmovaps %%ymm0, %0" :"=m" (vec256array[i])); in store_ymm() 164 i++; __asm__ volatile ("vmovaps %%ymm1, %0" :"=m" (vec256array[i])); in store_ymm() 165 i++; __asm__ volatile ("vmovaps %%ymm2, %0" :"=m" (vec256array[i])); in store_ymm() 166 i++; __asm__ volatile ("vmovaps %%ymm3, %0" :"=m" (vec256array[i])); in store_ymm() 167 i++; __asm__ volatile ("vmovaps %%ymm4, %0" :"=m" (vec256array[i])); in store_ymm() 168 i++; __asm__ volatile ("vmovaps %%ymm5, %0" :"=m" (vec256array[i])); in store_ymm() 169 i++; __asm__ volatile ("vmovaps %%ymm6, %0" :"=m" (vec256array[i])); in store_ymm() 170 i++; __asm__ volatile ("vmovaps %%ymm7, %0" :"=m" (vec256array[i])); in store_ymm() 172 i++; __asm__ volatile ("vmovaps %%ymm8, %0" :"=m" (vec256array[i])); in store_ymm() 173 i++; __asm__ volatile ("vmovaps %%ymm9, %0" :"=m" (vec256array[i])); in store_ymm() [all …]
|
| H A D | restartable_ranges.c | 25 __asm__(" .align 4\n" 36 __asm__(" .align 4\n" 55 __asm__(" .align 4\n" 68 __asm__(" .align 4\n" 90 __asm__(" .align 4\n"
|
| /xnu-11215.61.5/osfmk/i386/ |
| H A D | proc_reg.h | 217 __asm__ volatile ("mov %%es, %0" : "=r" (es)); in get_es() 224 __asm__ volatile ("mov %0, %%es" : : "r" (es)); in set_es() 231 __asm__ volatile ("mov %%ds, %0" : "=r" (ds)); in get_ds() 238 __asm__ volatile ("mov %0, %%ds" : : "r" (ds)); in set_ds() 245 __asm__ volatile ("mov %%fs, %0" : "=r" (fs)); in get_fs() 252 __asm__ volatile ("mov %0, %%fs" : : "r" (fs)); in set_fs() 259 __asm__ volatile ("mov %%gs, %0" : "=r" (gs)); in get_gs() 266 __asm__ volatile ("mov %0, %%gs" : : "r" (gs)); in set_gs() 273 __asm__ volatile ("mov %%ss, %0" : "=r" (ss)); in get_ss() 280 __asm__ volatile ("mov %0, %%ss" : : "r" (ss)); in set_ss() [all …]
|
| H A D | trap.h | 122 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"((void *)((unsigned long long)code))); in ml_recoverable_trap() 130 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"((void *)((unsigned long long)code))); in ml_fatal_trap() 139 register unsigned long long _value __asm__("rax") = value; in ml_fatal_trap_with_value() 140 __asm__ volatile ("ud1l %[_code](%%eax), %%eax" in ml_fatal_trap_with_value() 148 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"(code)) 150 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"(code)); \ 154 register unsigned long long _value __asm__("rax") = (value); \ 155 __asm__ volatile ("ud1l %[_code](%%eax), %%eax" \ 168 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"((void *)((unsigned long long)code))); in ml_trap()
|
| H A D | bit_routines.h | 75 __asm__ volatile(" jmp 1f \n \ 86 __asm__ volatile(" lock \n \ 98 __asm__ volatile(" lock \n \ 104 __asm__ volatile(" lock \n \ 112 __asm__ volatile (" lock \n \ in atomic_incl() 120 __asm__ volatile (" lock \n \ in atomic_decl() 129 __asm__ volatile ( in atomic_decl_and_test()
|
| H A D | pal_native.h | 40 #define pal_hlt() __asm__ volatile ("sti; hlt") 41 #define pal_sti() __asm__ volatile ("sti") 42 #define pal_cli() __asm__ volatile ("cli") 48 __asm__ volatile ( "cli"); in pal_stop_cpu() 50 __asm__ volatile ( "wbinvd; hlt"); in pal_stop_cpu() 64 #define pal_pmc_swi() __asm__ __volatile__("int %0"::"i"(LAPIC_PMC_SWI_VECTOR):"memory")
|
| H A D | trap.c | 506 __asm__ volatile ("mov %0,%%dr7" : : "r" (dr7)); in reset_dr7() 1096 __asm__ volatile ("mov %%db6, %0" : "=r" (dr6)); in user_trap() 1097 __asm__ volatile ("mov %0, %%db6" : : "r" (clear)); in user_trap() 1644 __asm__ volatile ("movq %%rbx, %0" : "=m" (iks->k_rbx)); in sync_iss_to_iks() 1645 __asm__ volatile ("movq %%rsp, %0" : "=m" (iks->k_rsp)); in sync_iss_to_iks() 1646 __asm__ volatile ("movq %%rbp, %0" : "=m" (iks->k_rbp)); in sync_iss_to_iks() 1647 __asm__ volatile ("movq %%r12, %0" : "=m" (iks->k_r12)); in sync_iss_to_iks() 1648 __asm__ volatile ("movq %%r13, %0" : "=m" (iks->k_r13)); in sync_iss_to_iks() 1649 __asm__ volatile ("movq %%r14, %0" : "=m" (iks->k_r14)); in sync_iss_to_iks() 1650 __asm__ volatile ("movq %%r15, %0" : "=m" (iks->k_r15)); in sync_iss_to_iks() [all …]
|
| H A D | x86_hypercall.h | 94 __asm__ __volatile__ goto (\ 111 __asm__ __volatile__ ("movq %12, %%r8 \n\t" in hvg_hypercall6() 185 __asm__ __volatile__ ( in kvmcompat_hypercall2()
|
| H A D | fpu.c | 107 __asm__ volatile("fninit") 110 __asm__("fnstcw %0" : "=m" (*(unsigned short *)(control))) 113 __asm__ volatile("fldcw %0" : : "m" (*(unsigned short *) &(control)) ) 116 __asm__ volatile("fnclex") 119 __asm__ volatile("fnsave %0" : "=m" (*state)) 122 __asm__ volatile("frstor %0" : : "m" (state)) 125 __asm__("fwait"); 130 __asm__ __volatile__ ("fxrstor %0" :: "m" (*a)); in fxrstor() 136 __asm__ __volatile__ ("fxsave %0" : "=m" (*a)); in fxsave() 142 __asm__ __volatile__ ("fxrstor64 %0" :: "m" (*a)); in fxrstor64() [all …]
|
| /xnu-11215.61.5/osfmk/corecrypto/ |
| H A D | ccaes_vng_gcm.h | 69 extern void gcm_init(void *Htable, void *H) __asm__("_gcm_init"); 70 extern void gcm_gmult(const void *X, const void *Htable, void *out) __asm__("_gcm_gmult"); 71 extern void gcm_ghash(void *X, void *Htable, const void *in, size_t len) __asm__("_gcm_ghash"); 73 …_SupplementalSSE3(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmEncrypt_Supp… 74 …SupplementalSSE3(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmDecrypt_Supp… 75 extern void gcmEncrypt_avx1(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmEnc… 76 extern void gcmDecrypt_avx1(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmDec… 78 extern void gcmEncrypt(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmEncrypt"… 79 extern void gcmDecrypt(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmDecrypt"…
|
| H A D | cc_internal.h | 102 __asm__ __volatile__ ( in cc_enable_dit() 224 __asm__ ("roll %%cl,%0" in CC_ROL() 233 __asm__ ("rorl %%cl,%0" in CC_ROR() 242 __asm__ __volatile__ ("roll %2,%0" \ 251 __asm__ __volatile__ ("rorl %2,%0" \ 286 __asm__("rolq %%cl,%0" in CC_ROL64() 295 __asm__("rorq %%cl,%0" in CC_ROR64() 305 __asm__("rolq %2,%0" \ 314 __asm__("rorq %2,%0" \
|
| /xnu-11215.61.5/osfmk/kern/ |
| H A D | exclaves_stackshot.h | 36 extern void * exclaves_enter_start_label __asm__("EXCLAVES_ENTRY_START"); 37 extern void * exclaves_enter_end_label __asm__("EXCLAVES_ENTRY_END"); 39 extern void * exclaves_upcall_start_label __asm__("EXCLAVES_UPCALL_START"); 40 extern void * exclaves_upcall_end_label __asm__("EXCLAVES_UPCALL_END"); 42 extern void * exclaves_scheduler_request_start_label __asm__("EXCLAVES_SCHEDULER_REQUEST_START"); 43 extern void * exclaves_scheduler_request_end_label __asm__("EXCLAVES_SCHEDULER_REQUEST_END");
|
| /xnu-11215.61.5/osfmk/arm/ |
| H A D | trap.h | 94 __asm__ volatile ("brk #%0" : : "i"(code)); in ml_recoverable_trap() 102 __asm__ volatile ("brk #%0" : : "i"(code)); in ml_fatal_trap() 112 register unsigned long long _value __asm__("x8") = (value); in ml_fatal_trap_with_value() 114 register unsigned long _value __asm__("r8") = (value); in ml_fatal_trap_with_value() 116 __asm__ volatile ("brk #%[_code]" in ml_fatal_trap_with_value() 124 __asm__ volatile ("brk #%0" : : "i"(code)) 126 __asm__ volatile ("brk #%0" : : "i"(code)); \ 135 __asm__ volatile ("brk #%0" : : "i"(code)); \
|
| H A D | machine_cpuid.c | 41 __asm__ volatile ("mrs %0, MIDR_EL1" : "=r" (midr)); in machine_read_midr() 50 __asm__ volatile ("mrs %0, CLIDR_EL1" : "=r" (clidr)); in machine_read_clidr() 59 __asm__ volatile ("mrs %0, CCSIDR_EL1" : "=r" (ccsidr)); in machine_read_ccsidr() 68 __asm__ volatile ("msr CSSELR_EL1, %0" : : "r" (csselr)); in machine_write_csselr() 79 __asm__ volatile ("mrs %0, ID_AA64DFR0_EL1" : "=r"(id_dfr0.value)); in machine_do_debugid()
|
| H A D | cpu_data.h | 63 __asm__ ("mrs %0, TPIDR_EL1" : "=r" (result)); in current_thread_fast() 98 __asm__ volatile ("mrs %0, TPIDR_EL1" : "=r" (result)); in current_thread_volatile() 112 __asm__ volatile ( in exception_stack_pointer()
|
| H A D | machdep_call.c | 52 __asm__ volatile ("mrs %0, TPIDRRO_EL0" : "=r" (uthread)); in get_tpidrro() 59 __asm__ volatile ("msr TPIDRRO_EL0, %0" : : "r" (uthread)); in set_tpidrro()
|
| /xnu-11215.61.5/EXTERNAL_HEADERS/architecture/i386/ |
| H A D | pio.h | 68 __asm__ volatile("inl %w1, %0" : "=a" (datum) : "Nd" (port)); in inl() 76 __asm__ volatile("inw %w1, %w0" : "=a" (datum) : "Nd" (port)); in inw() 84 __asm__ volatile("inb %w1, %b0" : "=a" (datum) : "Nd" (port)); in inb() 92 __asm__ volatile("outl %0, %w1" : : "a" (datum), "Nd" (port)); in outl() 99 __asm__ volatile("outw %w0, %w1" : : "a" (datum), "Nd" (port)); in outw() 106 __asm__ volatile("outb %b0, %w1" : : "a" (datum), "Nd" (port)); in outb()
|
| /xnu-11215.61.5/libsyscall/os/ |
| H A D | tsd.h | 76 __asm__ __volatile__ ("mrc p15, 0, %[p], c13, c0, 3" : [p] "=&r" (p)); in _os_cpu_number() 80 __asm__ __volatile__ ("mrs %0, TPIDR_EL0" : "=r" (p)); in _os_cpu_number() 84 __asm__ __volatile__ ("sidt %[p]" : [p] "=&m" (p)); in _os_cpu_number() 101 __asm__ __volatile__ ("mrs %0, TPIDR_EL0" : "=r" (p)); in _os_cpu_cluster_number() 126 __asm__("mov %%gs:%1, %0" : "=r" (ret) : "m" (*(void **)(slot * sizeof(void *)))); in _os_tsd_get_direct() 135 __asm__("movl %1, %%gs:%0" : "=m" (*(void **)(slot * sizeof(void *))) : "rn" (val)); in _os_tsd_set_direct() 137 __asm__("movl %1, %%gs:%0" : "=m" (*(void **)(slot * sizeof(void *))) : "ri" (val)); in _os_tsd_set_direct() 139 __asm__("movq %1, %%gs:%0" : "=m" (*(void **)(slot * sizeof(void *))) : "rn" (val)); in _os_tsd_set_direct() 153 __asm__("mrc p15, 0, %0, c13, c0, 3\n" in _os_tsd_get_base() 163 __asm__ ("mrs %0, TPIDRRO_EL0" : "=r" (tsd)); in _os_tsd_get_base()
|
| /xnu-11215.61.5/libkern/os/ |
| H A D | trace.h | 69 __asm__(""); /* avoid tailcall */ \ 86 __asm__(""); /* avoid tailcall */ \ 107 __asm__(""); /* avoid tailcall */ \ 131 __asm__(""); /* avoid tailcall */ \ 158 __asm__(""); /* avoid tailcall */ \ 188 __asm__(""); /* avoid tailcall */ \ 221 __asm__(""); /* avoid tailcall */ \ 257 __asm__(""); /* avoid tailcall */ \ 264 __asm__(""); /* avoid tailcall */ \ 281 __asm__(""); /* avoid tailcall */ \ [all …]
|
| H A D | log.h | 44 #define OS_LOG_NOTAILCALL_MARKER __asm__("") 384 __asm__(""); /* avoid tailcall */ \ 419 __asm__(""); /* avoid tailcall */ \ 453 __asm__(""); /* avoid tailcall */ \
|
| H A D | atomic_private_arch.h | 59 __asm__ __volatile__("and %[_dep], %[_v], #0" \ 134 __asm__ __volatile__("and %w[_dep], %w[_v], wzr" \ 143 __asm__ __volatile__("and %[_dep], %[_v], xzr" \
|
| /xnu-11215.61.5/libkern/libkern/i386/ |
| H A D | _OSByteOrder.h | 66 __asm__ ("bswap %0" : "+r" (_data)); in _OSSwapInt32() 88 __asm__ ("bswap %%eax\n\t" in _OSSwapInt64() 101 __asm__ ("bswap %0" : "+r" (_data)); in _OSSwapInt64()
|
| /xnu-11215.61.5/bsd/net/ |
| H A D | flowhash.c | 261 __asm__ ( "rol $31, %[k1]\n\t" :[k1] "+r" (k1) : :); in net_flowhash_mh3_x64_128() 263 __asm__ ( "ror %[k1], %[k1], #(64-31)\n\t" :[k1] "+r" (k1) : :); in net_flowhash_mh3_x64_128() 271 __asm__ ( "rol $27, %[h1]\n\t" :[h1] "+r" (h1) : :); in net_flowhash_mh3_x64_128() 273 __asm__ ( "ror %[h1], %[h1], #(64-27)\n\t" :[h1] "+r" (h1) : :); in net_flowhash_mh3_x64_128() 282 __asm__ ( "rol $33, %[k2]\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128() 284 __asm__ ( "ror %[k2], %[k2], #(64-33)\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128() 292 __asm__ ( "rol $31, %[h2]\n\t" :[h2] "+r" (h2) : :); in net_flowhash_mh3_x64_128() 294 __asm__ ( "ror %[h2], %[h2], #(64-31)\n\t" :[h2] "+r" (h2) : :); in net_flowhash_mh3_x64_128() 330 __asm__ ( "rol $33, %[k2]\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128() 332 __asm__ ( "ror %[k2], %[k2], #(64-33)\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128() [all …]
|
| /xnu-11215.61.5/libsa/ |
| H A D | lastkernelconstructor.c | 39 __asm__(".globl _last_kernel_symbol"); 40 __asm__(".zerofill __LAST, __last, _last_kernel_symbol, 0");
|
| /xnu-11215.61.5/bsd/skywalk/core/ |
| H A D | skywalk_common.h | 50 __asm__(""); __builtin_trap(); \ 55 __asm__(""); __builtin_trap(); \ 241 __asm__ __volatile__ ( 274 __asm__ __volatile__ ( 298 __asm__ __volatile__ ( 319 __asm__ __volatile__ ( 352 __asm__ __volatile__ ( 382 __asm__ __volatile__ ( in __sk_zero_16() 403 __asm__ __volatile__ ( in __sk_zero_32() 426 __asm__ __volatile__ ( in __sk_zero_48() [all …]
|