HomeSort by relevance Sort by last modified time
    Searched full:xmm2 (Results 51 - 75 of 304) sorted by null

1 23 4 5 6 7 8 91011>>

  /external/chromium_org/third_party/openssl/openssl/crypto/aes/asm/
aesni-x86.S 10 movups (%eax),%xmm2
16 xorps %xmm0,%xmm2
24 movups %xmm2,(%eax)
34 movups (%eax),%xmm2
40 xorps %xmm0,%xmm2
48 movups %xmm2,(%eax)
58 xorps %xmm0,%xmm2
89 xorps %xmm0,%xmm2
120 xorps %xmm0,%xmm2
156 xorps %xmm0,%xmm2
    [all...]
vpaes-x86.pl 184 &movdqa ("xmm2",&QWP($k_ipt,$const));
189 &pshufb ("xmm2","xmm0");
192 &pxor ("xmm2","xmm5");
193 &pxor ("xmm0","xmm2");
202 &pshufb ("xmm4","xmm2"); # 4 = sb1u
208 &pshufb ("xmm5","xmm2"); # 4 = sb2u
210 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t
211 &pshufb ("xmm2","xmm3"); # 2 = sb2t
212 &pxor ("xmm2","xmm5"); # 2 = 2A
217 &pxor ("xmm0","xmm2"); # 0 = 2A+
    [all...]
vpaes-x86.S 74 movdqa (%ebp),%xmm2
82 pxor %xmm5,%xmm2
83 pxor %xmm2,%xmm0
98 movdqa 80(%ebp),%xmm2
100 pxor %xmm5,%xmm2
105 pxor %xmm2,%xmm0
127 movdqa %xmm7,%xmm2
129 pxor %xmm0,%xmm2
151 movdqa -64(%ebx),%xmm2
163 pxor %xmm5,%xmm2
    [all...]
aesni-x86_64.S 6 movups (%rdi),%xmm2
11 xorps %xmm0,%xmm2
19 movups %xmm2,(%rsi)
27 movups (%rdi),%xmm2
32 xorps %xmm0,%xmm2
40 movups %xmm2,(%rsi)
50 xorps %xmm0,%xmm2
83 xorps %xmm0,%xmm2
116 xorps %xmm0,%xmm2
154 xorps %xmm0,%xmm2
    [all...]
  /external/openssl/crypto/aes/asm/
aesni-x86.S 10 movups (%eax),%xmm2
16 xorps %xmm0,%xmm2
24 movups %xmm2,(%eax)
34 movups (%eax),%xmm2
40 xorps %xmm0,%xmm2
48 movups %xmm2,(%eax)
58 xorps %xmm0,%xmm2
89 xorps %xmm0,%xmm2
120 xorps %xmm0,%xmm2
156 xorps %xmm0,%xmm2
    [all...]
vpaes-x86.pl 184 &movdqa ("xmm2",&QWP($k_ipt,$const));
189 &pshufb ("xmm2","xmm0");
192 &pxor ("xmm2","xmm5");
193 &pxor ("xmm0","xmm2");
202 &pshufb ("xmm4","xmm2"); # 4 = sb1u
208 &pshufb ("xmm5","xmm2"); # 4 = sb2u
210 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t
211 &pshufb ("xmm2","xmm3"); # 2 = sb2t
212 &pxor ("xmm2","xmm5"); # 2 = 2A
217 &pxor ("xmm0","xmm2"); # 0 = 2A+
    [all...]
vpaes-x86.S 74 movdqa (%ebp),%xmm2
82 pxor %xmm5,%xmm2
83 pxor %xmm2,%xmm0
98 movdqa 80(%ebp),%xmm2
100 pxor %xmm5,%xmm2
105 pxor %xmm2,%xmm0
127 movdqa %xmm7,%xmm2
129 pxor %xmm0,%xmm2
151 movdqa -64(%ebx),%xmm2
163 pxor %xmm5,%xmm2
    [all...]
aesni-x86_64.S 6 movups (%rdi),%xmm2
11 xorps %xmm0,%xmm2
19 movups %xmm2,(%rsi)
27 movups (%rdi),%xmm2
32 xorps %xmm0,%xmm2
40 movups %xmm2,(%rsi)
50 xorps %xmm0,%xmm2
83 xorps %xmm0,%xmm2
116 xorps %xmm0,%xmm2
154 xorps %xmm0,%xmm2
    [all...]
  /external/libyuv/files/source/
row_posix.cc 257 "movdqa %%xmm3,%%xmm2 \n"
258 "palignr $0x8,%%xmm1,%%xmm2 \n"
259 "pshufb %%xmm4,%%xmm2 \n"
260 "por %%xmm5,%%xmm2 \n"
263 "movdqa %%xmm2,0x20(%1) \n"
282 , "xmm0", "xmm1", "xmm2", "xmm3", "xmm4", "xmm5"
298 "movdqa %%xmm3,%%xmm2 \n"
299 "palignr $0x8,%%xmm1,%%xmm2 \n"
300 "pshufb %%xmm4,%%xmm2 \n"
301 "por %%xmm5,%%xmm2 \n
    [all...]
  /external/llvm/test/CodeGen/X86/
illegal-vector-args-return.ll 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "mulpd %xmm2, %xmm0"
4 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "addps %xmm2, %xmm0"
fma_patterns.ll 6 ; CHECK: vfmadd213ps %xmm2, %xmm1, %xmm0
9 ; CHECK_FMA4: vfmaddps %xmm2, %xmm1, %xmm0, %xmm0
18 ; CHECK: fmsub213ps %xmm2, %xmm1, %xmm0
21 ; CHECK_FMA4: vfmsubps %xmm2, %xmm1, %xmm0, %xmm0
30 ; CHECK: fnmadd213ps %xmm2, %xmm1, %xmm0
33 ; CHECK_FMA4: vfnmaddps %xmm2, %xmm1, %xmm0, %xmm0
42 ; CHECK: fnmsub213ps %xmm2, %xmm1, %xmm0
45 ; CHECK_FMA4: fnmsubps %xmm2, %xmm1, %xmm0, %xmm0
125 ; CHECK: vfmsub213pd %xmm2, %xmm1, %xmm0
128 ; CHECK_FMA4: vfmsubpd %xmm2, %xmm1, %xmm0, %xmm
    [all...]
  /external/libvpx/libvpx/vp9/common/x86/
vp9_postproc_sse2.asm 48 movd xmm2, dword ptr arg(6) ;flimit
49 punpcklwd xmm2, xmm2
50 punpckldq xmm2, xmm2
51 punpcklqdq xmm2, xmm2
78 pcmpgtw xmm7, xmm2
89 pcmpgtw xmm6, xmm2
103 pcmpgtw xmm6, xmm2
    [all...]
vp9_subpixel_8t_ssse3.asm 36 pshuflw xmm2, xmm4, 10101010b ;k4_k5
41 punpcklqdq xmm2, xmm2
47 movdqa k4k5, xmm2
66 movd xmm2, [rsi + rdx * 2] ;C
72 punpcklbw xmm2, xmm3 ;C D
79 pmaddubsw xmm2, k2k3
85 paddsw xmm0, xmm2
120 pshuflw xmm2, xmm4, 10101010b ;k4_k5
125 punpcklqdq xmm2, xmm
    [all...]
  /external/llvm/test/MC/Disassembler/X86/
marked-up.txt 5 # CHECK: xorps <reg:%xmm1>, <reg:%xmm2>
  /external/compiler-rt/lib/i386/
floatdidf.S 27 movsd REL_ADDR(twop52), %xmm2 // 0x1.0p52
28 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs)
29 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs)
ashrdi3.S 14 movd 12(%esp), %xmm2 // Load count
24 psrlq %xmm2, %xmm0 // unsigned shift input by count
33 pandn %xmm1, %xmm2 // 63 - count
35 psubq %xmm1, %xmm2 // 64 - count
36 psllq %xmm2, %xmm1 // -1 << (64 - count) = leading sign bits
  /external/libvpx/libvpx/vp8/encoder/x86/
temporal_filter_apply_sse2.asm 94 movdqa xmm2, [rdx] ; predictor (frame2)
95 movdqa xmm3, xmm2
96 punpcklbw xmm2, xmm7 ; pred[ 0- 7]
100 psubw xmm0, xmm2 ; src - pred[ 0- 7]
122 movdqa xmm2, [GLOBAL(_const_16w)]
124 psubusw xmm2, xmm0
127 pmullw xmm2, [rsp + filter_weight]
134 paddw xmm4, xmm2
150 pmullw xmm0, xmm2
154 movdqa xmm2, xmm
    [all...]
  /external/libvpx/libvpx/vp9/encoder/x86/
vp9_temporal_filter_apply_sse2.asm 94 movdqa xmm2, [rdx] ; predictor (frame2)
95 movdqa xmm3, xmm2
96 punpcklbw xmm2, xmm7 ; pred[ 0- 7]
100 psubw xmm0, xmm2 ; src - pred[ 0- 7]
122 movdqa xmm2, [GLOBAL(_const_16w)]
124 psubusw xmm2, xmm0
127 pmullw xmm2, [rsp + filter_weight]
134 paddw xmm4, xmm2
150 pmullw xmm0, xmm2
154 movdqa xmm2, xmm
    [all...]
vp9_sad_ssse3.asm 27 lddqu xmm2, XMMWORD PTR [rdi+1]
31 psadbw xmm2, xmm0
35 paddw xmm6, xmm2
40 lddqu xmm2, XMMWORD PTR [rdi+rdx+1]
47 psadbw xmm2, xmm0
51 paddw xmm6, xmm2
80 movdqa xmm2, xmm3
81 palignr xmm2, xmm4, (%2+1)
86 psadbw xmm2, xmm0
90 paddw xmm6, xmm2
    [all...]
  /bionic/libc/arch-x86/string/
sse2-wcslen-atom.S 100 pxor %xmm2, %xmm2
105 pcmpeqd (%eax), %xmm2
106 pmovmskb %xmm2, %edx
130 pcmpeqd (%eax), %xmm2
131 pmovmskb %xmm2, %edx
154 pcmpeqd (%eax), %xmm2
155 pmovmskb %xmm2, %edx
178 pcmpeqd (%eax), %xmm2
179 pmovmskb %xmm2, %ed
    [all...]
  /external/valgrind/main/VEX/test/
fxsave.c 39 asm __volatile__("movups vecZ, %xmm2");
63 asm __volatile__("xorps %xmm2, %xmm2");
64 asm __volatile__("movaps %xmm2, %xmm3");
65 asm __volatile__("movaps %xmm2, %xmm4");
66 asm __volatile__("movaps %xmm2, %xmm5");
67 asm __volatile__("movaps %xmm2, %xmm6");
  /external/valgrind/main/memcheck/tests/x86/
fxsave.c 40 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2");
64 asm __volatile__("xorps %xmm2, %xmm2");
65 asm __volatile__("movaps %xmm2, %xmm3");
66 asm __volatile__("movaps %xmm2, %xmm4");
67 asm __volatile__("movaps %xmm2, %xmm5");
68 asm __volatile__("movaps %xmm2, %xmm6");
  /external/libvpx/libvpx/vp8/common/x86/
sad_ssse3.asm 27 lddqu xmm2, XMMWORD PTR [rdi+1]
31 psadbw xmm2, xmm0
35 paddw xmm6, xmm2
40 lddqu xmm2, XMMWORD PTR [rdi+rdx+1]
47 psadbw xmm2, xmm0
51 paddw xmm6, xmm2
80 movdqa xmm2, xmm3
81 palignr xmm2, xmm4, (%2+1)
86 psadbw xmm2, xmm0
90 paddw xmm6, xmm2
    [all...]
subpixel_sse2.asm 184 movq xmm2, MMWORD PTR [rsi +14]
185 pslldq xmm2, 8
187 por xmm2, xmm1
244 movdqa xmm3, xmm2
245 movdqa xmm4, xmm2
247 movdqa xmm5, xmm2
248 movdqa xmm6, xmm2
250 movdqa xmm7, xmm2
273 psrldq xmm2, 5 ; xx xx xx xx xx 0d 0c 0b 0a 09 08 07 06 05 04 03
277 punpcklbw xmm2, xmm0 ; xx0a xx09 xx08 xx07 xx06 xx05 xx04 xx0
    [all...]
  /external/libvpx/libvpx/third_party/libyuv/source/
scale.c 683 movdqa xmm2, [eax + esi] local
686 pavgb xmm0, xmm2 // average rows
689 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) local
693 pand xmm2, xmm5 local
695 pavgw xmm0, xmm2
760 movdqa xmm2, [esi + ebx] local
762 pavgb xmm0, xmm2 // average rows
764 movdqa xmm2, [esi + ebx * 2] local
769 pavgb xmm2, xmm4 local
771 pavgb xmm0, xmm2
774 movdqa xmm2, xmm0 \/\/ average columns (32 to 16 pixels) local
778 pand xmm2, xmm7 local
784 movdqa xmm2, xmm0 \/\/ average columns (16 to 8 pixels) local
786 pand xmm2, xmm7 local
851 movdqa xmm2, [esi + ebx] local
855 movdqa xmm2, [esi + ebx * 2] local
861 pavgb xmm2, xmm4 local
866 movdqa xmm2, [ebp] local
870 pavgb xmm2, xmm4 local
878 pavgb xmm2, xmm4 local
926 movdqa xmm2, xmm1 local
930 pshufb xmm2, xmm5 local
968 movdqa xmm2, _shuf01 local
1025 movdqa xmm2, _shuf01 local
1127 movdqa xmm2, [esi + edx] local
1132 punpcklbw xmm2, xmm7 local
1136 movdqa xmm2, [esi + edx * 2] local
1139 punpcklbw xmm2, xmm7 local
1144 movdqa xmm2, xmm0 \/\/ 8 pixels -> 0,1,2 of xmm2 local
1146 paddusw xmm2, xmm0 local
1148 paddusw xmm2, xmm0 local
1149 pshufb xmm2, xmm4 local
1157 paddusw xmm2, xmm3 local
1159 pmulhuw xmm2, xmm6 \/\/ divide by 9,9,6, 9,9,6 local
1160 packuswb xmm2, xmm2 local
1163 pextrw eax, xmm2, 2 local
1190 movdqa xmm2, [esi] \/\/ average 2 rows into xmm2 local
1191 pavgb xmm2, [esi + edx] local
1199 pshufb xmm2, xmm6 local
1236 movdqa xmm2, [esi] local
1240 punpcklbw xmm2, xmm5 local
1250 paddusw xmm2, xmm0 \/\/ sum 16 words local
1299 movdqa xmm2, [esi + edx] local
1304 punpcklbw xmm2, xmm7 local
1309 pmullw xmm2, xmm6 \/\/ scale row 1 local
1343 movdqa xmm2, [esi + edx] local
1388 movdqa xmm2, [esi + edx] local
1425 movdqa xmm2, [esi + edx] local
1452 movdqa xmm2, _shuf01 local
    [all...]

Completed in 375 milliseconds

1 23 4 5 6 7 8 91011>>