/external/boringssl/win-x86_64/crypto/aes/ |
aesni-x86_64.asm | 12 movups xmm2,XMMWORD[rcx] 17 xorps xmm2,xmm0 27 movups XMMWORD[rdx],xmm2 28 pxor xmm2,xmm2 36 movups xmm2,XMMWORD[rcx] 41 xorps xmm2,xmm0 51 movups XMMWORD[rdx],xmm2 52 pxor xmm2,xmm2 [all...] |
/external/llvm/test/CodeGen/X86/ |
vec_cast2.ll | 94 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 95 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1 96 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0 105 ; CHECK-WIDE-NEXT: vmovshdup %xmm0, %xmm2 ## xmm2 = xmm0[1,1,3,3] 106 ; CHECK-WIDE-NEXT: vcvttss2si %xmm2, %eax 108 ; CHECK-WIDE-NEXT: vpermilpd $1, %xmm0, %xmm2 ## xmm2 = xmm0[1,0] 109 ; CHECK-WIDE-NEXT: vcvttss2si %xmm2, %eax 111 ; CHECK-WIDE-NEXT: vpermilps $231, %xmm0, %xmm2 ## xmm2 = xmm0[3,1,2,3 [all...] |
illegal-vector-args-return.ll | 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "mulpd %xmm2, %xmm0" 4 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "addps %xmm2, %xmm0"
|
fma_patterns.ll | 6 ; CHECK: vfmadd213ps %xmm2, %xmm1, %xmm0 9 ; CHECK_FMA4: vfmaddps %xmm2, %xmm1, %xmm0, %xmm0 18 ; CHECK: fmsub213ps %xmm2, %xmm1, %xmm0 21 ; CHECK_FMA4: vfmsubps %xmm2, %xmm1, %xmm0, %xmm0 30 ; CHECK: fnmadd213ps %xmm2, %xmm1, %xmm0 33 ; CHECK_FMA4: vfnmaddps %xmm2, %xmm1, %xmm0, %xmm0 42 ; CHECK: fnmsub213ps %xmm2, %xmm1, %xmm0 45 ; CHECK_FMA4: fnmsubps %xmm2, %xmm1, %xmm0, %xmm0 125 ; CHECK: vfmsub213pd %xmm2, %xmm1, %xmm0 128 ; CHECK_FMA4: vfmsubpd %xmm2, %xmm1, %xmm0, %xmm [all...] |
avx-cvt-2.ll | 14 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 15 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1 16 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0 31 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 32 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1 33 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0 48 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15] 49 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1 50 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0 66 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15 [all...] |
vector-shuffle-256-v16.ll | 316 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 317 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1 318 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0 334 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [14,15,14,15,14,15,14,15,14,15,14,15,14,15,14,15] 335 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1 336 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0 543 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2 545 ; AVX1-NEXT: vpblendw {{.*#+}} xmm2 = xmm3[0],xmm2[1],xmm3[2],xmm2[3],xmm3[4],xmm2[5],xmm3[6],xmm2[7 [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_postproc_sse2.asm | 48 movd xmm2, dword ptr arg(6) ;flimit 49 punpcklwd xmm2, xmm2 50 punpckldq xmm2, xmm2 51 punpcklqdq xmm2, xmm2 78 pcmpgtw xmm7, xmm2 89 pcmpgtw xmm6, xmm2 103 pcmpgtw xmm6, xmm2 [all...] |
/external/llvm/test/MC/Disassembler/X86/ |
marked-up.txt | 5 # CHECK: xorps <reg:%xmm1>, <reg:%xmm2>
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_postproc_sse2.asm | 48 movd xmm2, dword ptr arg(6) ;flimit 49 punpcklwd xmm2, xmm2 50 punpckldq xmm2, xmm2 51 punpcklqdq xmm2, xmm2 78 pcmpgtw xmm7, xmm2 89 pcmpgtw xmm6, xmm2 103 pcmpgtw xmm6, xmm2 [all...] |
/external/compiler-rt/lib/builtins/i386/ |
floatdidf.S | 30 movsd REL_ADDR(twop52), %xmm2 // 0x1.0p52 31 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs) 32 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs)
|
ashrdi3.S | 14 movd 12(%esp), %xmm2 // Load count 24 psrlq %xmm2, %xmm0 // unsigned shift input by count 33 pandn %xmm1, %xmm2 // 63 - count 35 psubq %xmm1, %xmm2 // 64 - count 36 psllq %xmm2, %xmm1 // -1 << (64 - count) = leading sign bits
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
temporal_filter_apply_sse2.asm | 94 movdqa xmm2, [rdx] ; predictor (frame2) 95 movdqa xmm3, xmm2 96 punpcklbw xmm2, xmm7 ; pred[ 0- 7] 100 psubw xmm0, xmm2 ; src - pred[ 0- 7] 122 movdqa xmm2, [GLOBAL(_const_16w)] 124 psubusw xmm2, xmm0 127 pmullw xmm2, [rsp + filter_weight] 134 paddw xmm4, xmm2 150 pmullw xmm0, xmm2 154 movdqa xmm2, xmm [all...] |
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_temporal_filter_apply_sse2.asm | 99 movdqa xmm2, [rdx] ; predictor (frame2) 100 movdqa xmm3, xmm2 101 punpcklbw xmm2, xmm7 ; pred[ 0- 7] 105 psubw xmm0, xmm2 ; src - pred[ 0- 7] 127 movdqa xmm2, [GLOBAL(_const_16w)] 129 psubusw xmm2, xmm0 132 pmullw xmm2, [rsp + filter_weight] 139 paddw xmm4, xmm2 155 pmullw xmm0, xmm2 159 movdqa xmm2, xmm [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/ |
temporal_filter_apply_sse2.asm | 94 movdqa xmm2, [rdx] ; predictor (frame2) 95 movdqa xmm3, xmm2 96 punpcklbw xmm2, xmm7 ; pred[ 0- 7] 100 psubw xmm0, xmm2 ; src - pred[ 0- 7] 122 movdqa xmm2, [GLOBAL(_const_16w)] 124 psubusw xmm2, xmm0 127 pmullw xmm2, [rsp + filter_weight] 134 paddw xmm4, xmm2 150 pmullw xmm0, xmm2 154 movdqa xmm2, xmm [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/ |
vp9_temporal_filter_apply_sse2.asm | 94 movdqa xmm2, [rdx] ; predictor (frame2) 95 movdqa xmm3, xmm2 96 punpcklbw xmm2, xmm7 ; pred[ 0- 7] 100 psubw xmm0, xmm2 ; src - pred[ 0- 7] 122 movdqa xmm2, [GLOBAL(_const_16w)] 124 psubusw xmm2, xmm0 127 pmullw xmm2, [rsp + filter_weight] 134 paddw xmm4, xmm2 150 pmullw xmm0, xmm2 154 movdqa xmm2, xmm [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale_posix.cc | 136 "movdqa %%xmm0,%%xmm2 \n" 140 "pand %%xmm5,%%xmm2 \n" 142 "pavgw %%xmm2,%%xmm0 \n" 170 MEMOPREG(movdqa,0x00,0,3,1,xmm2) // movdqa (%0,%3,1),%%xmm2 174 "pavgb %%xmm2,%%xmm0 \n" 176 "movdqa %%xmm0,%%xmm2 \n" 180 "pand %%xmm5,%%xmm2 \n" 182 "pavgw %%xmm2,%%xmm0 \n" 198 , "xmm0", "xmm1", "xmm2", "xmm3", "xmm5 [all...] |
/external/boringssl/linux-x86/crypto/aes/ |
vpaes-x86.S | 77 movdqa (%ebp),%xmm2 83 pxor %xmm5,%xmm2 88 pxor %xmm2,%xmm0 101 movdqa 80(%ebp),%xmm2 105 pxor %xmm5,%xmm2 108 pxor %xmm2,%xmm0 129 movdqa %xmm7,%xmm2 133 pxor %xmm0,%xmm2 155 movdqa -64(%ebx),%xmm2 167 pxor %xmm5,%xmm2 [all...] |
/external/boringssl/mac-x86/crypto/aes/ |
vpaes-x86.S | 74 movdqa (%ebp),%xmm2 80 pxor %xmm5,%xmm2 85 pxor %xmm2,%xmm0 98 movdqa 80(%ebp),%xmm2 102 pxor %xmm5,%xmm2 105 pxor %xmm2,%xmm0 126 movdqa %xmm7,%xmm2 130 pxor %xmm0,%xmm2 150 movdqa -64(%ebx),%xmm2 162 pxor %xmm5,%xmm2 [all...] |
/external/boringssl/win-x86/crypto/aes/ |
vpaes-x86.asm | 85 movdqa xmm2,[ebp] 91 pxor xmm2,xmm5 96 pxor xmm0,xmm2 109 movdqa xmm2,[80+ebp] 113 pxor xmm2,xmm5 116 pxor xmm0,xmm2 137 movdqa xmm2,xmm7 141 pxor xmm2,xmm0 160 movdqa xmm2,[ebx-64] 172 pxor xmm2,xmm [all...] |
/bionic/libc/arch-x86/atom/string/ |
sse2-wcslen-atom.S | 100 pxor %xmm2, %xmm2 105 pcmpeqd (%eax), %xmm2 106 pmovmskb %xmm2, %edx 130 pcmpeqd (%eax), %xmm2 131 pmovmskb %xmm2, %edx 154 pcmpeqd (%eax), %xmm2 155 pmovmskb %xmm2, %edx 178 pcmpeqd (%eax), %xmm2 179 pmovmskb %xmm2, %ed [all...] |
/external/valgrind/VEX/test/ |
fxsave.c | 39 asm __volatile__("movups vecZ, %xmm2"); 63 asm __volatile__("xorps %xmm2, %xmm2"); 64 asm __volatile__("movaps %xmm2, %xmm3"); 65 asm __volatile__("movaps %xmm2, %xmm4"); 66 asm __volatile__("movaps %xmm2, %xmm5"); 67 asm __volatile__("movaps %xmm2, %xmm6");
|
/external/valgrind/memcheck/tests/x86/ |
fxsave.c | 40 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2"); 64 asm __volatile__("xorps %xmm2, %xmm2"); 65 asm __volatile__("movaps %xmm2, %xmm3"); 66 asm __volatile__("movaps %xmm2, %xmm4"); 67 asm __volatile__("movaps %xmm2, %xmm5"); 68 asm __volatile__("movaps %xmm2, %xmm6");
|
/bionic/libc/arch-x86/silvermont/string/ |
sse2-strlen-slm.S | 116 pxor %xmm2, %xmm2 128 pcmpeqb 48(%eax), %xmm2 129 pmovmskb %xmm2, %ecx 149 pcmpeqb 48(%eax), %xmm2 150 pmovmskb %xmm2, %ecx 170 pcmpeqb 48(%eax), %xmm2 171 pmovmskb %xmm2, %ecx 191 pcmpeqb 48(%eax), %xmm2 192 pmovmskb %xmm2, %ec [all...] |
/bionic/libc/arch-x86_64/string/ |
sse2-strlen-slm.S | 95 pxor %xmm2, %xmm2 107 pcmpeqb 48(%rax), %xmm2 108 pmovmskb %xmm2, %edx 128 pcmpeqb 48(%rax), %xmm2 129 pmovmskb %xmm2, %edx 149 pcmpeqb 48(%rax), %xmm2 150 pmovmskb %xmm2, %edx 170 pcmpeqb 48(%rax), %xmm2 171 pmovmskb %xmm2, %ed [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
sad_ssse3.asm | 27 lddqu xmm2, XMMWORD PTR [rdi+1] 31 psadbw xmm2, xmm0 35 paddw xmm6, xmm2 40 lddqu xmm2, XMMWORD PTR [rdi+rdx+1] 47 psadbw xmm2, xmm0 51 paddw xmm6, xmm2 80 movdqa xmm2, xmm3 81 palignr xmm2, xmm4, (%2+1) 86 psadbw xmm2, xmm0 90 paddw xmm6, xmm2 [all...] |