HomeSort by relevance Sort by last modified time
    Searched full:xmm10 (Results 1 - 25 of 126) sorted by null

1 2 3 4 5 6

  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/
x86-64-reg.s 6 psrlw $2, %xmm10
8 psraw $2, %xmm10
10 psllw $2, %xmm10
12 psrld $2, %xmm10
14 psrad $2, %xmm10
16 pslld $2, %xmm10
18 psrlq $2, %xmm10
19 psrldq $2, %xmm10
21 psllq $2, %xmm10
22 pslldq $2, %xmm10
    [all...]
x86-64-reg-intel.d 12 [ ]*[a-f0-9]+: 66 41 0f 71 d2 02 psrlw xmm10,0x2
14 [ ]*[a-f0-9]+: 66 41 0f 71 e2 02 psraw xmm10,0x2
16 [ ]*[a-f0-9]+: 66 41 0f 71 f2 02 psllw xmm10,0x2
18 [ ]*[a-f0-9]+: 66 41 0f 72 d2 02 psrld xmm10,0x2
20 [ ]*[a-f0-9]+: 66 41 0f 72 e2 02 psrad xmm10,0x2
22 [ ]*[a-f0-9]+: 66 41 0f 72 f2 02 pslld xmm10,0x2
24 [ ]*[a-f0-9]+: 66 41 0f 73 d2 02 psrlq xmm10,0x2
25 [ ]*[a-f0-9]+: 66 41 0f 73 da 02 psrldq xmm10,0x2
27 [ ]*[a-f0-9]+: 66 41 0f 73 f2 02 psllq xmm10,0x2
28 [ ]*[a-f0-9]+: 66 41 0f 73 fa 02 pslldq xmm10,0x
    [all...]
x86-64-reg.d 11 [ ]*[a-f0-9]+: 66 41 0f 71 d2 02 psrlw \$0x2,%xmm10
13 [ ]*[a-f0-9]+: 66 41 0f 71 e2 02 psraw \$0x2,%xmm10
15 [ ]*[a-f0-9]+: 66 41 0f 71 f2 02 psllw \$0x2,%xmm10
17 [ ]*[a-f0-9]+: 66 41 0f 72 d2 02 psrld \$0x2,%xmm10
19 [ ]*[a-f0-9]+: 66 41 0f 72 e2 02 psrad \$0x2,%xmm10
21 [ ]*[a-f0-9]+: 66 41 0f 72 f2 02 pslld \$0x2,%xmm10
23 [ ]*[a-f0-9]+: 66 41 0f 73 d2 02 psrlq \$0x2,%xmm10
24 [ ]*[a-f0-9]+: 66 41 0f 73 da 02 psrldq \$0x2,%xmm10
26 [ ]*[a-f0-9]+: 66 41 0f 73 f2 02 psllq \$0x2,%xmm10
27 [ ]*[a-f0-9]+: 66 41 0f 73 fa 02 pslldq \$0x2,%xmm10
    [all...]
x86-64-sha.s 13 sha1nexte (%rax,%rbx,2), %xmm10
17 sha1msg1 (%rax,%rbx,2), %xmm10
21 sha1msg2 (%rax,%rbx,2), %xmm10
  /external/llvm/test/MC/X86/
x86_64-avx-clmul-encoding.s 3 // CHECK: vpclmulqdq $17, %xmm12, %xmm10, %xmm11
5 vpclmulhqhqdq %xmm12, %xmm10, %xmm11
7 // CHECK: vpclmulqdq $17, (%rax), %xmm10, %xmm13
9 vpclmulhqhqdq (%rax), %xmm10, %xmm13
11 // CHECK: vpclmulqdq $1, %xmm12, %xmm10, %xmm11
13 vpclmulhqlqdq %xmm12, %xmm10, %xmm11
15 // CHECK: vpclmulqdq $1, (%rax), %xmm10, %xmm13
17 vpclmulhqlqdq (%rax), %xmm10, %xmm13
19 // CHECK: vpclmulqdq $16, %xmm12, %xmm10, %xmm11
21 vpclmullqhqdq %xmm12, %xmm10, %xmm1
    [all...]
x86_64-fma3-encoding.s 3 // CHECK: vfmadd132pd %xmm12, %xmm10, %xmm11
5 vfmadd132pd %xmm12, %xmm10, %xmm11
7 // CHECK: vfmadd132pd (%rax), %xmm10, %xmm11
9 vfmadd132pd (%rax), %xmm10, %xmm11
11 // CHECK: vfmadd132ps %xmm12, %xmm10, %xmm11
13 vfmadd132ps %xmm12, %xmm10, %xmm11
15 // CHECK: vfmadd132ps (%rax), %xmm10, %xmm11
17 vfmadd132ps (%rax), %xmm10, %xmm11
19 // CHECK: vfmadd213pd %xmm12, %xmm10, %xmm11
21 vfmadd213pd %xmm12, %xmm10, %xmm1
    [all...]
x86_64-avx-encoding.s 3 // CHECK: vaddss %xmm8, %xmm9, %xmm10
5 vaddss %xmm8, %xmm9, %xmm10
7 // CHECK: vmulss %xmm8, %xmm9, %xmm10
9 vmulss %xmm8, %xmm9, %xmm10
11 // CHECK: vsubss %xmm8, %xmm9, %xmm10
13 vsubss %xmm8, %xmm9, %xmm10
15 // CHECK: vdivss %xmm8, %xmm9, %xmm10
17 vdivss %xmm8, %xmm9, %xmm10
19 // CHECK: vaddsd %xmm8, %xmm9, %xmm10
21 vaddsd %xmm8, %xmm9, %xmm10
    [all...]
  /toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ilp32/
x86-64-reg-intel.d 12 [ ]*[a-f0-9]+: 66 41 0f 71 d2 02 psrlw xmm10,0x2
14 [ ]*[a-f0-9]+: 66 41 0f 71 e2 02 psraw xmm10,0x2
16 [ ]*[a-f0-9]+: 66 41 0f 71 f2 02 psllw xmm10,0x2
18 [ ]*[a-f0-9]+: 66 41 0f 72 d2 02 psrld xmm10,0x2
20 [ ]*[a-f0-9]+: 66 41 0f 72 e2 02 psrad xmm10,0x2
22 [ ]*[a-f0-9]+: 66 41 0f 72 f2 02 pslld xmm10,0x2
24 [ ]*[a-f0-9]+: 66 41 0f 73 d2 02 psrlq xmm10,0x2
25 [ ]*[a-f0-9]+: 66 41 0f 73 da 02 psrldq xmm10,0x2
27 [ ]*[a-f0-9]+: 66 41 0f 73 f2 02 psllq xmm10,0x2
28 [ ]*[a-f0-9]+: 66 41 0f 73 fa 02 pslldq xmm10,0x
    [all...]
x86-64-reg.d 12 [ ]*[a-f0-9]+: 66 41 0f 71 d2 02 psrlw \$0x2,%xmm10
14 [ ]*[a-f0-9]+: 66 41 0f 71 e2 02 psraw \$0x2,%xmm10
16 [ ]*[a-f0-9]+: 66 41 0f 71 f2 02 psllw \$0x2,%xmm10
18 [ ]*[a-f0-9]+: 66 41 0f 72 d2 02 psrld \$0x2,%xmm10
20 [ ]*[a-f0-9]+: 66 41 0f 72 e2 02 psrad \$0x2,%xmm10
22 [ ]*[a-f0-9]+: 66 41 0f 72 f2 02 pslld \$0x2,%xmm10
24 [ ]*[a-f0-9]+: 66 41 0f 73 d2 02 psrlq \$0x2,%xmm10
25 [ ]*[a-f0-9]+: 66 41 0f 73 da 02 psrldq \$0x2,%xmm10
27 [ ]*[a-f0-9]+: 66 41 0f 73 f2 02 psllq \$0x2,%xmm10
28 [ ]*[a-f0-9]+: 66 41 0f 73 fa 02 pslldq \$0x2,%xmm10
    [all...]
  /external/boringssl/win-x86_64/crypto/aes/
bsaes-x86_64.asm 40 movdqa xmm10,xmm3
51 pxor xmm3,xmm10
54 movdqa xmm10,xmm15
65 pxor xmm15,xmm10
69 movdqa xmm10,xmm3
80 pxor xmm3,xmm10
83 movdqa xmm10,xmm15
94 pxor xmm15,xmm10
97 movdqa xmm10,xmm1
108 pxor xmm1,xmm10
    [all...]
vpaes-x86_64.asm 80 movdqa xmm3,xmm10
83 movdqa xmm4,xmm10
86 movdqa xmm2,xmm10
89 movdqa xmm3,xmm10
186 movdqa xmm3,xmm10
189 movdqa xmm4,xmm10
193 movdqa xmm2,xmm10
195 movdqa xmm3,xmm10
469 movdqa xmm3,xmm10
472 movdqa xmm4,xmm10
    [all...]
  /external/boringssl/win-x86_64/crypto/sha/
sha512-x86_64.asm 1847 movaps XMMWORD[(128+96)+rsp],xmm10
1883 vpaddq xmm10,xmm2,XMMWORD[((-64))+rbp]
1889 vmovdqa XMMWORD[32+rsp],xmm10
1890 vpaddq xmm10,xmm6,XMMWORD[64+rbp]
1897 vmovdqa XMMWORD[96+rsp],xmm10
1934 vpxor xmm8,xmm8,xmm10
1937 vpsrlq xmm10,xmm7,6
1946 vpxor xmm11,xmm11,xmm10
1961 vpaddq xmm10,xmm0,XMMWORD[((-128))+rbp]
1974 vmovdqa XMMWORD[rsp],xmm10
    [all...]
sha1-x86_64.asm 1277 movaps XMMWORD[(-40-32)+rax],xmm10
1350 movdqa xmm10,xmm4
1356 pslldq xmm10,12
1364 movdqa xmm9,xmm10
1367 psrld xmm10,30
1375 pxor xmm4,xmm10
1377 movdqa xmm10,XMMWORD[((-64))+r11]
1388 paddd xmm10,xmm4
1408 movdqa XMMWORD[rsp],xmm10
1425 movdqa xmm10,xmm
    [all...]
  /external/boringssl/linux-x86_64/crypto/aes/
bsaes-x86_64.S 38 movdqa %xmm3,%xmm10
49 pxor %xmm10,%xmm3
52 movdqa %xmm15,%xmm10
63 pxor %xmm10,%xmm15
67 movdqa %xmm3,%xmm10
78 pxor %xmm10,%xmm3
81 movdqa %xmm15,%xmm10
92 pxor %xmm10,%xmm15
95 movdqa %xmm1,%xmm10
106 pxor %xmm10,%xmm
    [all...]
  /external/boringssl/mac-x86_64/crypto/aes/
bsaes-x86_64.S 36 movdqa %xmm3,%xmm10
47 pxor %xmm10,%xmm3
50 movdqa %xmm15,%xmm10
61 pxor %xmm10,%xmm15
65 movdqa %xmm3,%xmm10
76 pxor %xmm10,%xmm3
79 movdqa %xmm15,%xmm10
90 pxor %xmm10,%xmm15
93 movdqa %xmm1,%xmm10
104 pxor %xmm10,%xmm
    [all...]
  /external/llvm/test/CodeGen/X86/
stack-folding-int-avx1.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
stack-folding-int-sse42.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
115 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
135 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
144 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
153 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
stack-folding-fp-sse42.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
47 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
55 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
64 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
73 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
82 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
96 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
stack-folding-fp-avx1.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
46 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
54 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
63 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
71 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
80 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
89 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
stack-folding-int-avx2.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
25 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
36 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %3 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
57 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
67 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
76 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
85 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
94 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
103 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
vselect-minmax.ll     [all...]
  /external/boringssl/linux-x86_64/crypto/sha/
sha512-x86_64.S 1851 vpaddq -64(%rbp),%xmm2,%xmm10
1857 vmovdqa %xmm10,32(%rsp)
1858 vpaddq 64(%rbp),%xmm6,%xmm10
1865 vmovdqa %xmm10,96(%rsp)
1902 vpxor %xmm10,%xmm8,%xmm8
1905 vpsrlq $6,%xmm7,%xmm10
1914 vpxor %xmm10,%xmm11,%xmm11
1929 vpaddq -128(%rbp),%xmm0,%xmm10
1942 vmovdqa %xmm10,0(%rsp)
1971 vpxor %xmm10,%xmm8,%xmm
    [all...]
  /external/boringssl/mac-x86_64/crypto/sha/
sha512-x86_64.S 1850 vpaddq -64(%rbp),%xmm2,%xmm10
1856 vmovdqa %xmm10,32(%rsp)
1857 vpaddq 64(%rbp),%xmm6,%xmm10
1864 vmovdqa %xmm10,96(%rsp)
1901 vpxor %xmm10,%xmm8,%xmm8
1904 vpsrlq $6,%xmm7,%xmm10
1913 vpxor %xmm10,%xmm11,%xmm11
1928 vpaddq -128(%rbp),%xmm0,%xmm10
1941 vmovdqa %xmm10,0(%rsp)
1970 vpxor %xmm10,%xmm8,%xmm
    [all...]
  /external/llvm/test/TableGen/
cast.td 56 def XMM10: Register<"xmm10">;
65 XMM8, XMM9, XMM10, XMM11,
  /external/valgrind/none/tests/amd64-solaris/
coredump_single_thread_sse.post.exp 18 %xmm10 0xbdf31a8ab5c8db03c7d15cf4b6cad1c8

Completed in 330 milliseconds

1 2 3 4 5 6