/external/chromium_org/third_party/openssl/openssl/crypto/aes/asm/ |
aesni-x86.S | 10 movups (%eax),%xmm2 16 xorps %xmm0,%xmm2 24 movups %xmm2,(%eax) 34 movups (%eax),%xmm2 40 xorps %xmm0,%xmm2 48 movups %xmm2,(%eax) 58 xorps %xmm0,%xmm2 89 xorps %xmm0,%xmm2 120 xorps %xmm0,%xmm2 156 xorps %xmm0,%xmm2 [all...] |
vpaes-x86.pl | 184 &movdqa ("xmm2",&QWP($k_ipt,$const)); 189 &pshufb ("xmm2","xmm0"); 192 &pxor ("xmm2","xmm5"); 193 &pxor ("xmm0","xmm2"); 202 &pshufb ("xmm4","xmm2"); # 4 = sb1u 208 &pshufb ("xmm5","xmm2"); # 4 = sb2u 210 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t 211 &pshufb ("xmm2","xmm3"); # 2 = sb2t 212 &pxor ("xmm2","xmm5"); # 2 = 2A 217 &pxor ("xmm0","xmm2"); # 0 = 2A+ [all...] |
vpaes-x86.S | 74 movdqa (%ebp),%xmm2 82 pxor %xmm5,%xmm2 83 pxor %xmm2,%xmm0 98 movdqa 80(%ebp),%xmm2 100 pxor %xmm5,%xmm2 105 pxor %xmm2,%xmm0 127 movdqa %xmm7,%xmm2 129 pxor %xmm0,%xmm2 151 movdqa -64(%ebx),%xmm2 163 pxor %xmm5,%xmm2 [all...] |
aesni-x86_64.S | 6 movups (%rdi),%xmm2 11 xorps %xmm0,%xmm2 19 movups %xmm2,(%rsi) 27 movups (%rdi),%xmm2 32 xorps %xmm0,%xmm2 40 movups %xmm2,(%rsi) 50 xorps %xmm0,%xmm2 83 xorps %xmm0,%xmm2 116 xorps %xmm0,%xmm2 154 xorps %xmm0,%xmm2 [all...] |
/external/openssl/crypto/aes/asm/ |
aesni-x86.S | 10 movups (%eax),%xmm2 16 xorps %xmm0,%xmm2 24 movups %xmm2,(%eax) 34 movups (%eax),%xmm2 40 xorps %xmm0,%xmm2 48 movups %xmm2,(%eax) 58 xorps %xmm0,%xmm2 89 xorps %xmm0,%xmm2 120 xorps %xmm0,%xmm2 156 xorps %xmm0,%xmm2 [all...] |
vpaes-x86.pl | 184 &movdqa ("xmm2",&QWP($k_ipt,$const)); 189 &pshufb ("xmm2","xmm0"); 192 &pxor ("xmm2","xmm5"); 193 &pxor ("xmm0","xmm2"); 202 &pshufb ("xmm4","xmm2"); # 4 = sb1u 208 &pshufb ("xmm5","xmm2"); # 4 = sb2u 210 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t 211 &pshufb ("xmm2","xmm3"); # 2 = sb2t 212 &pxor ("xmm2","xmm5"); # 2 = 2A 217 &pxor ("xmm0","xmm2"); # 0 = 2A+ [all...] |
vpaes-x86.S | 74 movdqa (%ebp),%xmm2 82 pxor %xmm5,%xmm2 83 pxor %xmm2,%xmm0 98 movdqa 80(%ebp),%xmm2 100 pxor %xmm5,%xmm2 105 pxor %xmm2,%xmm0 127 movdqa %xmm7,%xmm2 129 pxor %xmm0,%xmm2 151 movdqa -64(%ebx),%xmm2 163 pxor %xmm5,%xmm2 [all...] |
aesni-x86_64.S | 6 movups (%rdi),%xmm2 11 xorps %xmm0,%xmm2 19 movups %xmm2,(%rsi) 27 movups (%rdi),%xmm2 32 xorps %xmm0,%xmm2 40 movups %xmm2,(%rsi) 50 xorps %xmm0,%xmm2 83 xorps %xmm0,%xmm2 116 xorps %xmm0,%xmm2 154 xorps %xmm0,%xmm2 [all...] |
/external/libyuv/files/source/ |
row_posix.cc | 257 "movdqa %%xmm3,%%xmm2 \n" 258 "palignr $0x8,%%xmm1,%%xmm2 \n" 259 "pshufb %%xmm4,%%xmm2 \n" 260 "por %%xmm5,%%xmm2 \n" 263 "movdqa %%xmm2,0x20(%1) \n" 282 , "xmm0", "xmm1", "xmm2", "xmm3", "xmm4", "xmm5" 298 "movdqa %%xmm3,%%xmm2 \n" 299 "palignr $0x8,%%xmm1,%%xmm2 \n" 300 "pshufb %%xmm4,%%xmm2 \n" 301 "por %%xmm5,%%xmm2 \n [all...] |
/external/llvm/test/CodeGen/X86/ |
illegal-vector-args-return.ll | 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "mulpd %xmm2, %xmm0" 4 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "addps %xmm2, %xmm0"
|
fma_patterns.ll | 6 ; CHECK: vfmadd213ps %xmm2, %xmm1, %xmm0 9 ; CHECK_FMA4: vfmaddps %xmm2, %xmm1, %xmm0, %xmm0 18 ; CHECK: fmsub213ps %xmm2, %xmm1, %xmm0 21 ; CHECK_FMA4: vfmsubps %xmm2, %xmm1, %xmm0, %xmm0 30 ; CHECK: fnmadd213ps %xmm2, %xmm1, %xmm0 33 ; CHECK_FMA4: vfnmaddps %xmm2, %xmm1, %xmm0, %xmm0 42 ; CHECK: fnmsub213ps %xmm2, %xmm1, %xmm0 45 ; CHECK_FMA4: fnmsubps %xmm2, %xmm1, %xmm0, %xmm0 125 ; CHECK: vfmsub213pd %xmm2, %xmm1, %xmm0 128 ; CHECK_FMA4: vfmsubpd %xmm2, %xmm1, %xmm0, %xmm [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_postproc_sse2.asm | 48 movd xmm2, dword ptr arg(6) ;flimit 49 punpcklwd xmm2, xmm2 50 punpckldq xmm2, xmm2 51 punpcklqdq xmm2, xmm2 78 pcmpgtw xmm7, xmm2 89 pcmpgtw xmm6, xmm2 103 pcmpgtw xmm6, xmm2 [all...] |
vp9_subpixel_8t_ssse3.asm | 36 pshuflw xmm2, xmm4, 10101010b ;k4_k5 41 punpcklqdq xmm2, xmm2 47 movdqa k4k5, xmm2 66 movd xmm2, [rsi + rdx * 2] ;C 72 punpcklbw xmm2, xmm3 ;C D 79 pmaddubsw xmm2, k2k3 85 paddsw xmm0, xmm2 120 pshuflw xmm2, xmm4, 10101010b ;k4_k5 125 punpcklqdq xmm2, xmm [all...] |
/external/llvm/test/MC/Disassembler/X86/ |
marked-up.txt | 5 # CHECK: xorps <reg:%xmm1>, <reg:%xmm2>
|
/external/compiler-rt/lib/i386/ |
floatdidf.S | 27 movsd REL_ADDR(twop52), %xmm2 // 0x1.0p52 28 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs) 29 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs)
|
ashrdi3.S | 14 movd 12(%esp), %xmm2 // Load count 24 psrlq %xmm2, %xmm0 // unsigned shift input by count 33 pandn %xmm1, %xmm2 // 63 - count 35 psubq %xmm1, %xmm2 // 64 - count 36 psllq %xmm2, %xmm1 // -1 << (64 - count) = leading sign bits
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
temporal_filter_apply_sse2.asm | 94 movdqa xmm2, [rdx] ; predictor (frame2) 95 movdqa xmm3, xmm2 96 punpcklbw xmm2, xmm7 ; pred[ 0- 7] 100 psubw xmm0, xmm2 ; src - pred[ 0- 7] 122 movdqa xmm2, [GLOBAL(_const_16w)] 124 psubusw xmm2, xmm0 127 pmullw xmm2, [rsp + filter_weight] 134 paddw xmm4, xmm2 150 pmullw xmm0, xmm2 154 movdqa xmm2, xmm [all...] |
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_temporal_filter_apply_sse2.asm | 94 movdqa xmm2, [rdx] ; predictor (frame2) 95 movdqa xmm3, xmm2 96 punpcklbw xmm2, xmm7 ; pred[ 0- 7] 100 psubw xmm0, xmm2 ; src - pred[ 0- 7] 122 movdqa xmm2, [GLOBAL(_const_16w)] 124 psubusw xmm2, xmm0 127 pmullw xmm2, [rsp + filter_weight] 134 paddw xmm4, xmm2 150 pmullw xmm0, xmm2 154 movdqa xmm2, xmm [all...] |
vp9_sad_ssse3.asm | 27 lddqu xmm2, XMMWORD PTR [rdi+1] 31 psadbw xmm2, xmm0 35 paddw xmm6, xmm2 40 lddqu xmm2, XMMWORD PTR [rdi+rdx+1] 47 psadbw xmm2, xmm0 51 paddw xmm6, xmm2 80 movdqa xmm2, xmm3 81 palignr xmm2, xmm4, (%2+1) 86 psadbw xmm2, xmm0 90 paddw xmm6, xmm2 [all...] |
/bionic/libc/arch-x86/string/ |
sse2-wcslen-atom.S | 100 pxor %xmm2, %xmm2 105 pcmpeqd (%eax), %xmm2 106 pmovmskb %xmm2, %edx 130 pcmpeqd (%eax), %xmm2 131 pmovmskb %xmm2, %edx 154 pcmpeqd (%eax), %xmm2 155 pmovmskb %xmm2, %edx 178 pcmpeqd (%eax), %xmm2 179 pmovmskb %xmm2, %ed [all...] |
/external/valgrind/main/VEX/test/ |
fxsave.c | 39 asm __volatile__("movups vecZ, %xmm2"); 63 asm __volatile__("xorps %xmm2, %xmm2"); 64 asm __volatile__("movaps %xmm2, %xmm3"); 65 asm __volatile__("movaps %xmm2, %xmm4"); 66 asm __volatile__("movaps %xmm2, %xmm5"); 67 asm __volatile__("movaps %xmm2, %xmm6");
|
/external/valgrind/main/memcheck/tests/x86/ |
fxsave.c | 40 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2"); 64 asm __volatile__("xorps %xmm2, %xmm2"); 65 asm __volatile__("movaps %xmm2, %xmm3"); 66 asm __volatile__("movaps %xmm2, %xmm4"); 67 asm __volatile__("movaps %xmm2, %xmm5"); 68 asm __volatile__("movaps %xmm2, %xmm6");
|
/external/libvpx/libvpx/vp8/common/x86/ |
sad_ssse3.asm | 27 lddqu xmm2, XMMWORD PTR [rdi+1] 31 psadbw xmm2, xmm0 35 paddw xmm6, xmm2 40 lddqu xmm2, XMMWORD PTR [rdi+rdx+1] 47 psadbw xmm2, xmm0 51 paddw xmm6, xmm2 80 movdqa xmm2, xmm3 81 palignr xmm2, xmm4, (%2+1) 86 psadbw xmm2, xmm0 90 paddw xmm6, xmm2 [all...] |
subpixel_sse2.asm | 184 movq xmm2, MMWORD PTR [rsi +14] 185 pslldq xmm2, 8 187 por xmm2, xmm1 244 movdqa xmm3, xmm2 245 movdqa xmm4, xmm2 247 movdqa xmm5, xmm2 248 movdqa xmm6, xmm2 250 movdqa xmm7, xmm2 273 psrldq xmm2, 5 ; xx xx xx xx xx 0d 0c 0b 0a 09 08 07 06 05 04 03 277 punpcklbw xmm2, xmm0 ; xx0a xx09 xx08 xx07 xx06 xx05 xx04 xx0 [all...] |
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale.c | 683 movdqa xmm2, [eax + esi] local 686 pavgb xmm0, xmm2 // average rows 689 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) local 693 pand xmm2, xmm5 local 695 pavgw xmm0, xmm2 760 movdqa xmm2, [esi + ebx] local 762 pavgb xmm0, xmm2 // average rows 764 movdqa xmm2, [esi + ebx * 2] local 769 pavgb xmm2, xmm4 local 771 pavgb xmm0, xmm2 774 movdqa xmm2, xmm0 \/\/ average columns (32 to 16 pixels) local 778 pand xmm2, xmm7 local 784 movdqa xmm2, xmm0 \/\/ average columns (16 to 8 pixels) local 786 pand xmm2, xmm7 local 851 movdqa xmm2, [esi + ebx] local 855 movdqa xmm2, [esi + ebx * 2] local 861 pavgb xmm2, xmm4 local 866 movdqa xmm2, [ebp] local 870 pavgb xmm2, xmm4 local 878 pavgb xmm2, xmm4 local 926 movdqa xmm2, xmm1 local 930 pshufb xmm2, xmm5 local 968 movdqa xmm2, _shuf01 local 1025 movdqa xmm2, _shuf01 local 1127 movdqa xmm2, [esi + edx] local 1132 punpcklbw xmm2, xmm7 local 1136 movdqa xmm2, [esi + edx * 2] local 1139 punpcklbw xmm2, xmm7 local 1144 movdqa xmm2, xmm0 \/\/ 8 pixels -> 0,1,2 of xmm2 local 1146 paddusw xmm2, xmm0 local 1148 paddusw xmm2, xmm0 local 1149 pshufb xmm2, xmm4 local 1157 paddusw xmm2, xmm3 local 1159 pmulhuw xmm2, xmm6 \/\/ divide by 9,9,6, 9,9,6 local 1160 packuswb xmm2, xmm2 local 1163 pextrw eax, xmm2, 2 local 1190 movdqa xmm2, [esi] \/\/ average 2 rows into xmm2 local 1191 pavgb xmm2, [esi + edx] local 1199 pshufb xmm2, xmm6 local 1236 movdqa xmm2, [esi] local 1240 punpcklbw xmm2, xmm5 local 1250 paddusw xmm2, xmm0 \/\/ sum 16 words local 1299 movdqa xmm2, [esi + edx] local 1304 punpcklbw xmm2, xmm7 local 1309 pmullw xmm2, xmm6 \/\/ scale row 1 local 1343 movdqa xmm2, [esi + edx] local 1388 movdqa xmm2, [esi + edx] local 1425 movdqa xmm2, [esi + edx] local 1452 movdqa xmm2, _shuf01 local [all...] |