Home
last modified time | relevance | path

Searched refs:__asm__ (Results 1 – 25 of 84) sorted by relevance

1234

/xnu-11215.61.5/tests/
H A Davx.c163 __asm__ volatile ("vmovaps %%ymm0, %0" :"=m" (vec256array[i])); in store_ymm()
164 i++; __asm__ volatile ("vmovaps %%ymm1, %0" :"=m" (vec256array[i])); in store_ymm()
165 i++; __asm__ volatile ("vmovaps %%ymm2, %0" :"=m" (vec256array[i])); in store_ymm()
166 i++; __asm__ volatile ("vmovaps %%ymm3, %0" :"=m" (vec256array[i])); in store_ymm()
167 i++; __asm__ volatile ("vmovaps %%ymm4, %0" :"=m" (vec256array[i])); in store_ymm()
168 i++; __asm__ volatile ("vmovaps %%ymm5, %0" :"=m" (vec256array[i])); in store_ymm()
169 i++; __asm__ volatile ("vmovaps %%ymm6, %0" :"=m" (vec256array[i])); in store_ymm()
170 i++; __asm__ volatile ("vmovaps %%ymm7, %0" :"=m" (vec256array[i])); in store_ymm()
172 i++; __asm__ volatile ("vmovaps %%ymm8, %0" :"=m" (vec256array[i])); in store_ymm()
173 i++; __asm__ volatile ("vmovaps %%ymm9, %0" :"=m" (vec256array[i])); in store_ymm()
[all …]
H A Drestartable_ranges.c25 __asm__(" .align 4\n"
36 __asm__(" .align 4\n"
55 __asm__(" .align 4\n"
68 __asm__(" .align 4\n"
90 __asm__(" .align 4\n"
/xnu-11215.61.5/osfmk/i386/
H A Dproc_reg.h217 __asm__ volatile ("mov %%es, %0" : "=r" (es)); in get_es()
224 __asm__ volatile ("mov %0, %%es" : : "r" (es)); in set_es()
231 __asm__ volatile ("mov %%ds, %0" : "=r" (ds)); in get_ds()
238 __asm__ volatile ("mov %0, %%ds" : : "r" (ds)); in set_ds()
245 __asm__ volatile ("mov %%fs, %0" : "=r" (fs)); in get_fs()
252 __asm__ volatile ("mov %0, %%fs" : : "r" (fs)); in set_fs()
259 __asm__ volatile ("mov %%gs, %0" : "=r" (gs)); in get_gs()
266 __asm__ volatile ("mov %0, %%gs" : : "r" (gs)); in set_gs()
273 __asm__ volatile ("mov %%ss, %0" : "=r" (ss)); in get_ss()
280 __asm__ volatile ("mov %0, %%ss" : : "r" (ss)); in set_ss()
[all …]
H A Dtrap.h122 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"((void *)((unsigned long long)code))); in ml_recoverable_trap()
130 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"((void *)((unsigned long long)code))); in ml_fatal_trap()
139 register unsigned long long _value __asm__("rax") = value; in ml_fatal_trap_with_value()
140 __asm__ volatile ("ud1l %[_code](%%eax), %%eax" in ml_fatal_trap_with_value()
148 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"(code))
150 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"(code)); \
154 register unsigned long long _value __asm__("rax") = (value); \
155 __asm__ volatile ("ud1l %[_code](%%eax), %%eax" \
168 __asm__ volatile ("ud1l %0(%%eax), %%eax" : : "p"((void *)((unsigned long long)code))); in ml_trap()
H A Dbit_routines.h75 __asm__ volatile(" jmp 1f \n \
86 __asm__ volatile(" lock \n \
98 __asm__ volatile(" lock \n \
104 __asm__ volatile(" lock \n \
112 __asm__ volatile (" lock \n \ in atomic_incl()
120 __asm__ volatile (" lock \n \ in atomic_decl()
129 __asm__ volatile ( in atomic_decl_and_test()
H A Dpal_native.h40 #define pal_hlt() __asm__ volatile ("sti; hlt")
41 #define pal_sti() __asm__ volatile ("sti")
42 #define pal_cli() __asm__ volatile ("cli")
48 __asm__ volatile ( "cli"); in pal_stop_cpu()
50 __asm__ volatile ( "wbinvd; hlt"); in pal_stop_cpu()
64 #define pal_pmc_swi() __asm__ __volatile__("int %0"::"i"(LAPIC_PMC_SWI_VECTOR):"memory")
H A Dtrap.c506 __asm__ volatile ("mov %0,%%dr7" : : "r" (dr7)); in reset_dr7()
1096 __asm__ volatile ("mov %%db6, %0" : "=r" (dr6)); in user_trap()
1097 __asm__ volatile ("mov %0, %%db6" : : "r" (clear)); in user_trap()
1644 __asm__ volatile ("movq %%rbx, %0" : "=m" (iks->k_rbx)); in sync_iss_to_iks()
1645 __asm__ volatile ("movq %%rsp, %0" : "=m" (iks->k_rsp)); in sync_iss_to_iks()
1646 __asm__ volatile ("movq %%rbp, %0" : "=m" (iks->k_rbp)); in sync_iss_to_iks()
1647 __asm__ volatile ("movq %%r12, %0" : "=m" (iks->k_r12)); in sync_iss_to_iks()
1648 __asm__ volatile ("movq %%r13, %0" : "=m" (iks->k_r13)); in sync_iss_to_iks()
1649 __asm__ volatile ("movq %%r14, %0" : "=m" (iks->k_r14)); in sync_iss_to_iks()
1650 __asm__ volatile ("movq %%r15, %0" : "=m" (iks->k_r15)); in sync_iss_to_iks()
[all …]
H A Dx86_hypercall.h94 __asm__ __volatile__ goto (\
111 __asm__ __volatile__ ("movq %12, %%r8 \n\t" in hvg_hypercall6()
185 __asm__ __volatile__ ( in kvmcompat_hypercall2()
H A Dfpu.c107 __asm__ volatile("fninit")
110 __asm__("fnstcw %0" : "=m" (*(unsigned short *)(control)))
113 __asm__ volatile("fldcw %0" : : "m" (*(unsigned short *) &(control)) )
116 __asm__ volatile("fnclex")
119 __asm__ volatile("fnsave %0" : "=m" (*state))
122 __asm__ volatile("frstor %0" : : "m" (state))
125 __asm__("fwait");
130 __asm__ __volatile__ ("fxrstor %0" :: "m" (*a)); in fxrstor()
136 __asm__ __volatile__ ("fxsave %0" : "=m" (*a)); in fxsave()
142 __asm__ __volatile__ ("fxrstor64 %0" :: "m" (*a)); in fxrstor64()
[all …]
/xnu-11215.61.5/osfmk/corecrypto/
H A Dccaes_vng_gcm.h69 extern void gcm_init(void *Htable, void *H) __asm__("_gcm_init");
70 extern void gcm_gmult(const void *X, const void *Htable, void *out) __asm__("_gcm_gmult");
71 extern void gcm_ghash(void *X, void *Htable, const void *in, size_t len) __asm__("_gcm_ghash");
73 …_SupplementalSSE3(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmEncrypt_Supp…
74 …SupplementalSSE3(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmDecrypt_Supp…
75 extern void gcmEncrypt_avx1(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmEnc…
76 extern void gcmDecrypt_avx1(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmDec…
78 extern void gcmEncrypt(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmEncrypt"…
79 extern void gcmDecrypt(const void*, void*, void*, unsigned int, void*, void*) __asm__("_gcmDecrypt"…
H A Dcc_internal.h102 __asm__ __volatile__ ( in cc_enable_dit()
224 __asm__ ("roll %%cl,%0" in CC_ROL()
233 __asm__ ("rorl %%cl,%0" in CC_ROR()
242 __asm__ __volatile__ ("roll %2,%0" \
251 __asm__ __volatile__ ("rorl %2,%0" \
286 __asm__("rolq %%cl,%0" in CC_ROL64()
295 __asm__("rorq %%cl,%0" in CC_ROR64()
305 __asm__("rolq %2,%0" \
314 __asm__("rorq %2,%0" \
/xnu-11215.61.5/osfmk/kern/
H A Dexclaves_stackshot.h36 extern void * exclaves_enter_start_label __asm__("EXCLAVES_ENTRY_START");
37 extern void * exclaves_enter_end_label __asm__("EXCLAVES_ENTRY_END");
39 extern void * exclaves_upcall_start_label __asm__("EXCLAVES_UPCALL_START");
40 extern void * exclaves_upcall_end_label __asm__("EXCLAVES_UPCALL_END");
42 extern void * exclaves_scheduler_request_start_label __asm__("EXCLAVES_SCHEDULER_REQUEST_START");
43 extern void * exclaves_scheduler_request_end_label __asm__("EXCLAVES_SCHEDULER_REQUEST_END");
/xnu-11215.61.5/osfmk/arm/
H A Dtrap.h94 __asm__ volatile ("brk #%0" : : "i"(code)); in ml_recoverable_trap()
102 __asm__ volatile ("brk #%0" : : "i"(code)); in ml_fatal_trap()
112 register unsigned long long _value __asm__("x8") = (value); in ml_fatal_trap_with_value()
114 register unsigned long _value __asm__("r8") = (value); in ml_fatal_trap_with_value()
116 __asm__ volatile ("brk #%[_code]" in ml_fatal_trap_with_value()
124 __asm__ volatile ("brk #%0" : : "i"(code))
126 __asm__ volatile ("brk #%0" : : "i"(code)); \
135 __asm__ volatile ("brk #%0" : : "i"(code)); \
H A Dmachine_cpuid.c41 __asm__ volatile ("mrs %0, MIDR_EL1" : "=r" (midr)); in machine_read_midr()
50 __asm__ volatile ("mrs %0, CLIDR_EL1" : "=r" (clidr)); in machine_read_clidr()
59 __asm__ volatile ("mrs %0, CCSIDR_EL1" : "=r" (ccsidr)); in machine_read_ccsidr()
68 __asm__ volatile ("msr CSSELR_EL1, %0" : : "r" (csselr)); in machine_write_csselr()
79 __asm__ volatile ("mrs %0, ID_AA64DFR0_EL1" : "=r"(id_dfr0.value)); in machine_do_debugid()
H A Dcpu_data.h63 __asm__ ("mrs %0, TPIDR_EL1" : "=r" (result)); in current_thread_fast()
98 __asm__ volatile ("mrs %0, TPIDR_EL1" : "=r" (result)); in current_thread_volatile()
112 __asm__ volatile ( in exception_stack_pointer()
H A Dmachdep_call.c52 __asm__ volatile ("mrs %0, TPIDRRO_EL0" : "=r" (uthread)); in get_tpidrro()
59 __asm__ volatile ("msr TPIDRRO_EL0, %0" : : "r" (uthread)); in set_tpidrro()
/xnu-11215.61.5/EXTERNAL_HEADERS/architecture/i386/
H A Dpio.h68 __asm__ volatile("inl %w1, %0" : "=a" (datum) : "Nd" (port)); in inl()
76 __asm__ volatile("inw %w1, %w0" : "=a" (datum) : "Nd" (port)); in inw()
84 __asm__ volatile("inb %w1, %b0" : "=a" (datum) : "Nd" (port)); in inb()
92 __asm__ volatile("outl %0, %w1" : : "a" (datum), "Nd" (port)); in outl()
99 __asm__ volatile("outw %w0, %w1" : : "a" (datum), "Nd" (port)); in outw()
106 __asm__ volatile("outb %b0, %w1" : : "a" (datum), "Nd" (port)); in outb()
/xnu-11215.61.5/libsyscall/os/
H A Dtsd.h76 __asm__ __volatile__ ("mrc p15, 0, %[p], c13, c0, 3" : [p] "=&r" (p)); in _os_cpu_number()
80 __asm__ __volatile__ ("mrs %0, TPIDR_EL0" : "=r" (p)); in _os_cpu_number()
84 __asm__ __volatile__ ("sidt %[p]" : [p] "=&m" (p)); in _os_cpu_number()
101 __asm__ __volatile__ ("mrs %0, TPIDR_EL0" : "=r" (p)); in _os_cpu_cluster_number()
126 __asm__("mov %%gs:%1, %0" : "=r" (ret) : "m" (*(void **)(slot * sizeof(void *)))); in _os_tsd_get_direct()
135 __asm__("movl %1, %%gs:%0" : "=m" (*(void **)(slot * sizeof(void *))) : "rn" (val)); in _os_tsd_set_direct()
137 __asm__("movl %1, %%gs:%0" : "=m" (*(void **)(slot * sizeof(void *))) : "ri" (val)); in _os_tsd_set_direct()
139 __asm__("movq %1, %%gs:%0" : "=m" (*(void **)(slot * sizeof(void *))) : "rn" (val)); in _os_tsd_set_direct()
153 __asm__("mrc p15, 0, %0, c13, c0, 3\n" in _os_tsd_get_base()
163 __asm__ ("mrs %0, TPIDRRO_EL0" : "=r" (tsd)); in _os_tsd_get_base()
/xnu-11215.61.5/libkern/os/
H A Dtrace.h69 __asm__(""); /* avoid tailcall */ \
86 __asm__(""); /* avoid tailcall */ \
107 __asm__(""); /* avoid tailcall */ \
131 __asm__(""); /* avoid tailcall */ \
158 __asm__(""); /* avoid tailcall */ \
188 __asm__(""); /* avoid tailcall */ \
221 __asm__(""); /* avoid tailcall */ \
257 __asm__(""); /* avoid tailcall */ \
264 __asm__(""); /* avoid tailcall */ \
281 __asm__(""); /* avoid tailcall */ \
[all …]
H A Dlog.h44 #define OS_LOG_NOTAILCALL_MARKER __asm__("")
384 __asm__(""); /* avoid tailcall */ \
419 __asm__(""); /* avoid tailcall */ \
453 __asm__(""); /* avoid tailcall */ \
H A Datomic_private_arch.h59 __asm__ __volatile__("and %[_dep], %[_v], #0" \
134 __asm__ __volatile__("and %w[_dep], %w[_v], wzr" \
143 __asm__ __volatile__("and %[_dep], %[_v], xzr" \
/xnu-11215.61.5/libkern/libkern/i386/
H A D_OSByteOrder.h66 __asm__ ("bswap %0" : "+r" (_data)); in _OSSwapInt32()
88 __asm__ ("bswap %%eax\n\t" in _OSSwapInt64()
101 __asm__ ("bswap %0" : "+r" (_data)); in _OSSwapInt64()
/xnu-11215.61.5/bsd/net/
H A Dflowhash.c261 __asm__ ( "rol $31, %[k1]\n\t" :[k1] "+r" (k1) : :); in net_flowhash_mh3_x64_128()
263 __asm__ ( "ror %[k1], %[k1], #(64-31)\n\t" :[k1] "+r" (k1) : :); in net_flowhash_mh3_x64_128()
271 __asm__ ( "rol $27, %[h1]\n\t" :[h1] "+r" (h1) : :); in net_flowhash_mh3_x64_128()
273 __asm__ ( "ror %[h1], %[h1], #(64-27)\n\t" :[h1] "+r" (h1) : :); in net_flowhash_mh3_x64_128()
282 __asm__ ( "rol $33, %[k2]\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128()
284 __asm__ ( "ror %[k2], %[k2], #(64-33)\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128()
292 __asm__ ( "rol $31, %[h2]\n\t" :[h2] "+r" (h2) : :); in net_flowhash_mh3_x64_128()
294 __asm__ ( "ror %[h2], %[h2], #(64-31)\n\t" :[h2] "+r" (h2) : :); in net_flowhash_mh3_x64_128()
330 __asm__ ( "rol $33, %[k2]\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128()
332 __asm__ ( "ror %[k2], %[k2], #(64-33)\n\t" :[k2] "+r" (k2) : :); in net_flowhash_mh3_x64_128()
[all …]
/xnu-11215.61.5/libsa/
H A Dlastkernelconstructor.c39 __asm__(".globl _last_kernel_symbol");
40 __asm__(".zerofill __LAST, __last, _last_kernel_symbol, 0");
/xnu-11215.61.5/bsd/skywalk/core/
H A Dskywalk_common.h50 __asm__(""); __builtin_trap(); \
55 __asm__(""); __builtin_trap(); \
241 __asm__ __volatile__ (
274 __asm__ __volatile__ (
298 __asm__ __volatile__ (
319 __asm__ __volatile__ (
352 __asm__ __volatile__ (
382 __asm__ __volatile__ ( in __sk_zero_16()
403 __asm__ __volatile__ ( in __sk_zero_32()
426 __asm__ __volatile__ ( in __sk_zero_48()
[all …]

1234