/external/valgrind/none/tests/amd64/ |
fma4.c | 219 : /*TRASH*/"xmm7","xmm8","xmm6","xmm9","r14","memory","cc" \ 235 "vfmaddpd %%xmm7,%%xmm8,%%xmm6,%%xmm9", 1); \ 237 "vfmaddpd %%xmm7,%%xmm8,%%xmm9,%%xmm9", 1); \ 239 "vfmaddpd (%%r14),%%xmm8,%%xmm6,%%xmm9", 1); \ 241 "vfmaddpd %%xmm8,(%%r14),%%xmm6,%%xmm9", 1); 257 "vfmaddps %%xmm7,%%xmm8,%%xmm6,%%xmm9", 0); \ 259 "vfmaddps %%xmm7,%%xmm8,%%xmm9,%%xmm9", 0); \ 261 "vfmaddps (%%r14),%%xmm8,%%xmm6,%%xmm9", 0); [all...] |
fma.c | 103 __asm __volatile__ ("vmovaps (%1), %%xmm9; vmovaps (%2), %%xmm7; vmovaps (%3), %%xmm8;" 104 "vfmadd132ps %%xmm7, %%xmm8, %%xmm9;" 105 "vmovaps %%xmm9, (%0)" : : "r" (&ft.res[i]), "r" (&ft.x[i]), 106 "r" (&ft.y[i]), "r" (&ft.z[i]) : "xmm7", "xmm8", "xmm9"); 109 __asm __volatile__ ("vmovaps (%1), %%xmm9; vmovaps (%3), %%xmm8;" 110 "vfmadd132ps (%2), %%xmm8, %%xmm9;" 111 "vmovaps %%xmm9, (%0)" : : "r" (&ft.res[i]), "r" (&ft.x[i]), 112 "r" (&ft.y[i]), "r" (&ft.z[i]) : "xmm7", "xmm8", "xmm9"); 115 __asm __volatile__ ("vmovaps (%1), %%xmm9; vmovaps (%3), %%xmm7; vmovaps (%2), %%xmm8;" 116 "vfmadd213ps %%xmm7, %%xmm8, %%xmm9;" [all...] |
avx-1.c | 84 : /*TRASH*/"xmm0","xmm7","xmm8","xmm6","xmm9","r14","memory","cc" \ 104 "xmm0","xmm8","xmm7","xmm9","r14","rax","memory","cc" \ 146 "vpblendvb %%xmm9, %%xmm6, %%xmm8, %%xmm7", 147 "vpblendvb %%xmm9, (%%rax), %%xmm8, %%xmm7") 367 "vmovups %%xmm9, (%%rax)") 371 "vmovq (%%rax), %%xmm9") 374 "vpcmpestrm $0x45, %%xmm7, %%xmm8; movapd %%xmm0, %%xmm9", 375 "vpcmpestrm $0x45, (%%rax), %%xmm8; movapd %%xmm0, %%xmm9") 382 "vcvtsd2ss %%xmm9, %%xmm8, %%xmm7", 386 "vcvtss2sd %%xmm9, %%xmm8, %%xmm7" [all...] |
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
aesni-gcm-x86_64.asm | 20 vpxor xmm9,xmm1,xmm15 59 vaesenc xmm9,xmm9,xmm2 83 vaesenc xmm9,xmm9,xmm15 101 vaesenc xmm9,xmm9,xmm15 119 vaesenc xmm9,xmm9,xmm15 141 vaesenc xmm9,xmm9,xmm1 [all...] |
bsaes-x86_64.asm | 38 movdqa xmm9,xmm5 50 pxor xmm5,xmm9 52 movdqa xmm9,xmm1 64 pxor xmm1,xmm9 67 movdqa xmm9,xmm4 79 pxor xmm4,xmm9 81 movdqa xmm9,xmm0 93 pxor xmm0,xmm9 95 movdqa xmm9,xmm2 107 pxor xmm2,xmm9 [all...] |
sha1-x86_64.asm | 1277 movaps XMMWORD[(-40-48)+r11],xmm9 1301 movdqa xmm9,XMMWORD[((-64))+r14] 1310 paddd xmm0,xmm9 1312 paddd xmm1,xmm9 1313 paddd xmm2,xmm9 1315 psubd xmm0,xmm9 1317 psubd xmm1,xmm9 1319 psubd xmm2,xmm9 1327 paddd xmm9,xmm3 1347 movdqa XMMWORD[48+rsp],xmm9 [all...] |
sha512-x86_64.asm | 1847 movaps XMMWORD[(128+80)+rsp],xmm9 1882 vpaddq xmm9,xmm1,XMMWORD[((-96))+rbp] 1888 vmovdqa XMMWORD[16+rsp],xmm9 1889 vpaddq xmm9,xmm5,XMMWORD[32+rbp] 1896 vmovdqa XMMWORD[80+rsp],xmm9 1927 vpxor xmm8,xmm8,xmm9 1952 vpxor xmm11,xmm11,xmm9 1996 vpxor xmm8,xmm8,xmm9 2021 vpxor xmm11,xmm11,xmm9 2065 vpxor xmm8,xmm8,xmm9 [all...] |
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
aesni-gcm-x86_64.S | 16 vpxor %xmm15,%xmm1,%xmm9 55 vaesenc %xmm2,%xmm9,%xmm9 79 vaesenc %xmm15,%xmm9,%xmm9 97 vaesenc %xmm15,%xmm9,%xmm9 115 vaesenc %xmm15,%xmm9,%xmm9 137 vaesenc %xmm15,%xmm9,%xmm [all...] |
bsaes-x86_64.S | 36 movdqa %xmm5,%xmm9 48 pxor %xmm9,%xmm5 50 movdqa %xmm1,%xmm9 62 pxor %xmm9,%xmm1 65 movdqa %xmm4,%xmm9 77 pxor %xmm9,%xmm4 79 movdqa %xmm0,%xmm9 91 pxor %xmm9,%xmm0 93 movdqa %xmm2,%xmm9 105 pxor %xmm9,%xmm [all...] |
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
aesni-gcm-x86_64.S | 16 vpxor %xmm15,%xmm1,%xmm9 55 vaesenc %xmm2,%xmm9,%xmm9 79 vaesenc %xmm15,%xmm9,%xmm9 97 vaesenc %xmm15,%xmm9,%xmm9 115 vaesenc %xmm15,%xmm9,%xmm9 137 vaesenc %xmm15,%xmm9,%xmm [all...] |
bsaes-x86_64.S | 34 movdqa %xmm5,%xmm9 46 pxor %xmm9,%xmm5 48 movdqa %xmm1,%xmm9 60 pxor %xmm9,%xmm1 63 movdqa %xmm4,%xmm9 75 pxor %xmm9,%xmm4 77 movdqa %xmm0,%xmm9 89 pxor %xmm9,%xmm0 91 movdqa %xmm2,%xmm9 103 pxor %xmm9,%xmm [all...] |
/external/llvm/test/MC/X86/ |
intel-syntax-x86-64-avx.s | 19 // CHECK: vgatherdps xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 21 vgatherdps xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 23 // CHECK: vgatherqps xmm10, qword ptr [r15 + 2*xmm9], xmm8 25 vgatherqps xmm10, qword ptr [r15 + 2*xmm9], xmm8 51 // CHECK: vpgatherdd xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 53 vpgatherdd xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 55 // CHECK: vpgatherqd xmm10, qword ptr [r15 + 2*xmm9], xmm8 57 vpgatherqd xmm10, qword ptr [r15 + 2*xmm9], xmm8
|
/external/boringssl/win-x86_64/crypto/cipher_extra/ |
aes128gcmsiv-x86_64.asm | 165 vpxor xmm9,xmm9,xmm9 212 vpxor xmm9,xmm4,xmm6 223 vmovdqa xmm9,XMMWORD[rcx] 309 vpxor xmm1,xmm1,xmm9 339 vpxor xmm9,xmm4,xmm6 355 vpxor xmm1,xmm1,xmm9 703 vmovdqa xmm9,XMMWORD[rdi] 706 vpshufd xmm9,xmm9,0x9 [all...] |
/external/boringssl/linux-x86_64/crypto/cipher_extra/ |
aes128gcmsiv-x86_64.S | 131 vpxor %xmm9,%xmm9,%xmm9 178 vpxor %xmm6,%xmm4,%xmm9 189 vmovdqa (%rcx),%xmm9 275 vpxor %xmm9,%xmm1,%xmm1 305 vpxor %xmm6,%xmm4,%xmm9 321 vpxor %xmm9,%xmm1,%xmm1 617 vmovdqa 0(%rdi),%xmm9 620 vpshufd $0x90,%xmm9,%xmm [all...] |
/external/boringssl/mac-x86_64/crypto/cipher_extra/ |
aes128gcmsiv-x86_64.S | 131 vpxor %xmm9,%xmm9,%xmm9 178 vpxor %xmm6,%xmm4,%xmm9 189 vmovdqa (%rcx),%xmm9 275 vpxor %xmm9,%xmm1,%xmm1 305 vpxor %xmm6,%xmm4,%xmm9 321 vpxor %xmm9,%xmm1,%xmm1 617 vmovdqa 0(%rdi),%xmm9 620 vpshufd $0x90,%xmm9,%xmm [all...] |
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ |
x86-64-vgather-check.s | 9 vgatherdps %xmm9,(%rax,%xmm1,4),%xmm0 11 vgatherdps %xmm2,(%rax,%xmm1,8),%xmm9
|
x86-64-sha.s | 12 sha1nexte 0x12(%rax), %xmm9 16 sha1msg1 0x12(%rax), %xmm9 20 sha1msg2 0x12(%rax), %xmm9
|
/external/llvm/test/CodeGen/X86/ |
stack-folding-int-avx1.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
stack-folding-int-sse42.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 115 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 135 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 144 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 153 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
stack-folding-fp-sse42.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 47 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 55 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 64 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 73 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 82 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 96 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
stack-folding-fp-avx1.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 46 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 54 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 63 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 71 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 80 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 89 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
/external/llvm/test/TableGen/ |
cast.td | 55 def XMM9: Register<"xmm9">; 65 XMM8, XMM9, XMM10, XMM11,
|
/external/swiftshader/third_party/LLVM/test/TableGen/ |
Slice.td | 50 def XMM9: Register<"xmm9">; 60 XMM8, XMM9, XMM10, XMM11,
|
TargetInstrSpec.td | 51 def XMM9: Register<"xmm9">; 61 XMM8, XMM9, XMM10, XMM11,
|
cast.td | 50 def XMM9: Register<"xmm9">; 60 XMM8, XMM9, XMM10, XMM11,
|