/external/libvpx/libvpx/third_party/libyuv/source/ |
compare_win.cc | 37 movdqa xmm3, xmm1 // abs trick 41 movdqa xmm2, xmm1 146 movdqa xmm6, kHash16x33 152 movdqa xmm5, kHashMul0 153 movdqa xmm2, xmm1 155 movdqa xmm3, xmm2 158 movdqa xmm5, kHashMul1 159 movdqa xmm4, xmm2 162 movdqa xmm5, kHashMul2 164 movdqa xmm2, xmm [all...] |
compare_gcc.cc | 32 "movdqa %%xmm1,%%xmm3 \n" 36 "movdqa %%xmm1,%%xmm2 \n" 97 "movdqa %4,%%xmm6 \n" 103 "movdqa %5,%%xmm5 \n" 104 "movdqa %%xmm1,%%xmm2 \n" 106 "movdqa %%xmm2,%%xmm3 \n" 109 "movdqa %6,%%xmm5 \n" 110 "movdqa %%xmm2,%%xmm4 \n" 113 "movdqa %7,%%xmm5 \n" 115 "movdqa %%xmm1,%%xmm2 \n [all...] |
/external/libyuv/files/source/ |
row_posix.cc | 130 "movdqa %%xmm0,%%xmm1 \n" 135 "movdqa %%xmm0,(%1) \n" 136 "movdqa %%xmm1,0x10(%1) \n" 153 "movdqa %3,%%xmm5 \n" 157 "movdqa (%0),%%xmm0 \n" 160 "movdqa %%xmm0,(%0,%1,1) \n" 177 "movdqa %3,%%xmm5 \n" 181 "movdqa (%0),%%xmm0 \n" 184 "movdqa %%xmm0,(%0,%1,1) \n" 200 "movdqa %3,%%xmm5 \n [all...] |
compare.cc | 83 movdqa xmm6, kHash16x33 90 movdqa xmm5, kHashMul0 91 movdqa xmm2, xmm1 93 movdqa xmm3, xmm2 96 movdqa xmm5, kHashMul1 97 movdqa xmm4, xmm2 100 movdqa xmm5, kHashMul2 102 movdqa xmm2, xmm1 105 movdqa xmm5, kHashMul3 165 "movdqa %4,%%xmm6 \n [all...] |
/external/llvm/test/CodeGen/X86/ |
vselect-minmax.ll | 11 ; SSE2-NEXT: movdqa %xmm1, %xmm2 36 ; SSE2-NEXT: movdqa %xmm0, %xmm2 43 ; SSE2-NEXT: movdqa %xmm2, %xmm0 64 ; SSE2-NEXT: movdqa %xmm0, %xmm2 69 ; SSE2-NEXT: movdqa %xmm2, %xmm0 90 ; SSE2-NEXT: movdqa %xmm1, %xmm3 97 ; SSE2-NEXT: movdqa %xmm2, %xmm0 246 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [32768,32768,32768,32768,32768,32768,32768,32768] 247 ; SSE2-NEXT: movdqa %xmm0, %xmm3 274 ; SSE2-NEXT: movdqa %xmm0, %xmm [all...] |
vector-tzcnt-128.ll | 23 ; SSE2-NEXT: movdqa %xmm1, %xmm0 39 ; SSE3-NEXT: movdqa %xmm1, %xmm0 55 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 100 ; SSE2-NEXT: movdqa %xmm1, %xmm0 113 ; SSE3-NEXT: movdqa %xmm1, %xmm0 126 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 162 ; SSE2-NEXT: movdqa %xmm2, %xmm0 166 ; SSE2-NEXT: movdqa {{.*#+}} xmm0 = [858993459,858993459,858993459,858993459] 167 ; SSE2-NEXT: movdqa %xmm2, %xmm3 172 ; SSE2-NEXT: movdqa %xmm2, %xmm [all...] |
combine-multiplies.ll | 74 ; movdqa .LCPI1_0, %xmm1 # xmm1 = [11,11,11,11] 76 ; movdqa .LCPI1_1, %xmm2 # xmm2 = [22,22,22,22] 83 ; movdqa .LCPI1_2, %xmm2 # xmm2 = [242,242,242,242] 86 ; movdqa %xmm2, v2 87 ; movdqa %xmm0, v3 88 ; movdqa %xmm1, x 97 ; CHECK: movdqa .LCPI1_0, [[C11:%xmm[0-9]]] 99 ; CHECK-NEXT: movdqa .LCPI1_1, [[C22:%xmm[0-9]]] 106 ; CHECK-NEXT: movdqa .LCPI1_2, [[C242:%xmm[0-9]]] 109 ; CHECK-NEXT: movdqa [[C242]], v [all...] |
vec_ctbits.ll | 19 ; CHECK-NEXT: movdqa %xmm1, %xmm0 38 ; CHECK-NEXT: movdqa %xmm1, %xmm0 48 ; CHECK-NEXT: movdqa %xmm0, %xmm1 52 ; CHECK-NEXT: movdqa {{.*#+}} xmm1 = [3689348814741910323,3689348814741910323] 53 ; CHECK-NEXT: movdqa %xmm0, %xmm2 58 ; CHECK-NEXT: movdqa %xmm0, %xmm1 64 ; CHECK-NEXT: movdqa %xmm1, %xmm0 89 ; CHECK-NEXT: movdqa %xmm1, %xmm0 113 ; CHECK-NEXT: movdqa %xmm1, %xmm0 124 ; CHECK-NEXT: movdqa %xmm0, %xmm [all...] |
widen_load-2.ll | 9 ; CHECK: movdqa (%{{.*}}), %[[R0:xmm[0-9]+]] 39 ; CHECK: movdqa (%{{.*}}), %[[R0:xmm[0-9]+]] 40 ; CHECK-NEXT: movdqa 16(%{{.*}}), %[[R1:xmm[0-9]+]] 45 ; CHECK-NEXT: movdqa %[[R0]], (%{{.*}}) 56 ; CHECK: movdqa (%{{.*}}), %[[R0:xmm[0-9]+]] 57 ; CHECK-NEXT: movdqa 16(%{{.*}}), %[[R1:xmm[0-9]+]] 58 ; CHECK-NEXT: movdqa 32(%{{.*}}), %[[R2:xmm[0-9]+]] 62 ; CHECK-NEXT: movdqa %[[R2]], 32(%{{.*}}) 63 ; CHECK-NEXT: movdqa %[[R1]], 16(%{{.*}}) 64 ; CHECK-NEXT: movdqa %[[R0]], (%{{.*}} [all...] |
vector-shift-ashr-128.ll | 20 ; SSE2-NEXT: movdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 21 ; SSE2-NEXT: movdqa %xmm2, %xmm4 25 ; SSE2-NEXT: movdqa %xmm0, %xmm2 31 ; SSE2-NEXT: movdqa %xmm2, %xmm0 36 ; SSE41-NEXT: movdqa {{.*#+}} xmm2 = [9223372036854775808,9223372036854775808] 37 ; SSE41-NEXT: movdqa %xmm2, %xmm3 42 ; SSE41-NEXT: movdqa %xmm0, %xmm3 83 ; X32-SSE-NEXT: movdqa {{.*#+}} xmm3 = [0,2147483648,0,2147483648] 84 ; X32-SSE-NEXT: movdqa %xmm3, %xmm4 89 ; X32-SSE-NEXT: movdqa %xmm0, %xmm [all...] |
4char-promote.ll | 8 ; CHECK-NOT: movdqa
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
vpx_subpixel_bilinear_sse2.asm | 19 movdqa xmm3, [rdx] ;load filters 41 movdqa xmm1, xmm0 66 movdqa xmm7, [rdx] ;load filters 198 movdqa xmm2, xmm0 199 movdqa xmm3, xmm1 276 movdqa xmm2, xmm0 277 movdqa xmm3, xmm1 302 movdqa xmm1, xmm0 328 movdqa xmm1, xmm0 356 movdqa xmm2, xmm [all...] |
ssim_opt_x86_64.asm | 17 movdqa xmm1, xmm3 20 movdqa xmm2, xmm4 29 movdqa xmm2,%1 33 movdqa xmm2,%1 41 movdqa xmm1, %1 92 movdqa xmm3, xmm5 93 movdqa xmm4, xmm6 99 movdqa xmm3, xmm5 100 movdqa xmm4, xmm6
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_subpixel_bilinear_sse2.asm | 19 movdqa xmm3, [rdx] ;load filters 41 movdqa xmm1, xmm0 66 movdqa xmm7, [rdx] ;load filters 198 movdqa xmm2, xmm0 199 movdqa xmm3, xmm1 276 movdqa xmm2, xmm0 277 movdqa xmm3, xmm1 302 movdqa xmm1, xmm0 328 movdqa xmm1, xmm0 356 movdqa xmm2, xmm [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
variance_impl_ssse3.asm | 69 movdqa xmm2, xmm0 94 movdqa xmm3, xmm1 107 movdqa xmm2, xmm0 108 movdqa xmm0, xmm1 109 movdqa xmm3, xmm2 164 movdqa xmm0, xmm1 174 movdqa xmm2, xmm1 175 movdqa xmm0, xmm3 202 movdqa xmm1, xmm0 266 movdqa xmm3, xmm [all...] |
variance_impl_sse2.asm | 37 movdqa xmm0, [rax] 38 movdqa xmm1, [rax+16] 39 movdqa xmm2, [rax+32] 40 movdqa xmm3, [rax+48] 55 movdqa xmm3,xmm4 58 movdqa xmm3,xmm4 136 movdqa xmm3, xmm1 137 movdqa xmm4, xmm2 166 movdqa xmm1, xmm6 178 movdqa xmm2, xmm [all...] |
/external/boringssl/win-x86/crypto/aes/ |
aesni-x86.asm | 621 movdqa xmm5,[esp] 622 movdqa xmm2,xmm7 652 movdqa xmm2,xmm7 704 movdqa xmm5,[esp] 705 movdqa xmm2,xmm7 732 movdqa xmm2,xmm7 830 movdqa xmm2,[esp] 842 movdqa [48+esp],xmm0 845 movdqa [64+esp],xmm1 854 movdqa [32+esp],xmm [all...] |
/external/boringssl/win-x86_64/crypto/modes/ |
ghash-x86_64.asm | 699 movdqa xmm3,xmm2 713 movdqa xmm0,xmm2 715 movdqa xmm1,xmm0 724 movdqa xmm4,xmm3 730 movdqa xmm4,xmm0 731 movdqa xmm3,xmm0 737 movdqa xmm3,xmm0 744 movdqa xmm4,xmm0 760 movdqa xmm1,xmm0 769 movdqa xmm4,xmm [all...] |
/external/libjpeg-turbo/simd/ |
jcsample-sse2.asm | 122 movdqa xmm0, XMMWORD [esi+0*SIZEOF_XMMWORD] 129 movdqa xmm0, XMMWORD [esi+0*SIZEOF_XMMWORD] 130 movdqa xmm1, XMMWORD [esi+1*SIZEOF_XMMWORD] 133 movdqa xmm2,xmm0 134 movdqa xmm3,xmm1 150 movdqa XMMWORD [edi+0*SIZEOF_XMMWORD], xmm0 278 movdqa xmm0, XMMWORD [edx+0*SIZEOF_XMMWORD] 279 movdqa xmm1, XMMWORD [esi+0*SIZEOF_XMMWORD] 287 movdqa xmm0, XMMWORD [edx+0*SIZEOF_XMMWORD] 288 movdqa xmm1, XMMWORD [esi+0*SIZEOF_XMMWORD [all...] |
jdmrgext-sse2-64.asm | 76 movdqa xmm6, XMMWORD [rbx] ; xmm6=Cb(0123456789ABCDEF) 77 movdqa xmm7, XMMWORD [rdx] ; xmm7=Cr(0123456789ABCDEF) 83 movdqa xmm4,xmm6 86 movdqa xmm0,xmm7 105 movdqa xmm5,xmm6 ; xmm5=CbH 106 movdqa xmm2,xmm4 ; xmm2=CbL 109 movdqa xmm1,xmm7 ; xmm1=CrH 110 movdqa xmm3,xmm0 ; xmm3=CrL 135 movdqa XMMWORD [wk(0)], xmm6 ; wk(0)=(B-Y)H 136 movdqa XMMWORD [wk(1)], xmm7 ; wk(1)=(R-Y) [all...] |
jdmrgext-sse2.asm | 87 movdqa xmm6, XMMWORD [ebx] ; xmm6=Cb(0123456789ABCDEF) 88 movdqa xmm7, XMMWORD [edx] ; xmm7=Cr(0123456789ABCDEF) 94 movdqa xmm4,xmm6 97 movdqa xmm0,xmm7 116 movdqa xmm5,xmm6 ; xmm5=CbH 117 movdqa xmm2,xmm4 ; xmm2=CbL 120 movdqa xmm1,xmm7 ; xmm1=CrH 121 movdqa xmm3,xmm0 ; xmm3=CrL 146 movdqa XMMWORD [wk(0)], xmm6 ; wk(0)=(B-Y)H 147 movdqa XMMWORD [wk(1)], xmm7 ; wk(1)=(R-Y) [all...] |
/external/boringssl/linux-x86/crypto/sha/ |
sha512-586.S | 381 movdqa 640(%ebp),%xmm1 384 movdqa (%ebp),%xmm3 385 movdqa %xmm1,%xmm2 389 movdqa %xmm3,-128(%edx) 390 movdqa 16(%ebp),%xmm4 391 movdqa %xmm2,%xmm3 395 movdqa %xmm4,-112(%edx) 396 movdqa 32(%ebp),%xmm5 397 movdqa %xmm3,%xmm4 401 movdqa %xmm5,-96(%edx [all...] |
/external/boringssl/mac-x86/crypto/sha/ |
sha512-586.S | 380 movdqa 640(%ebp),%xmm1 383 movdqa (%ebp),%xmm3 384 movdqa %xmm1,%xmm2 388 movdqa %xmm3,-128(%edx) 389 movdqa 16(%ebp),%xmm4 390 movdqa %xmm2,%xmm3 394 movdqa %xmm4,-112(%edx) 395 movdqa 32(%ebp),%xmm5 396 movdqa %xmm3,%xmm4 400 movdqa %xmm5,-96(%edx [all...] |
/bionic/libc/arch-x86/silvermont/string/ |
sse2-memcpy-slm.S | 166 movdqa %xmm0, (%ebx) 167 movdqa %xmm1, 16(%ebx) 168 movdqa %xmm2, 32(%ebx) 169 movdqa %xmm3, 48(%ebx) 183 movdqa %xmm0, (%ebx) 184 movdqa %xmm1, 16(%ebx) 185 movdqa %xmm2, 32(%ebx) 186 movdqa %xmm3, 48(%ebx) 187 movdqa %xmm4, 64(%ebx) 188 movdqa %xmm5, 80(%ebx [all...] |
/bionic/libc/arch-x86_64/string/ |
sse2-memcpy-slm.S | 157 movdqa %xmm0, (%r8) 158 movdqa %xmm1, 16(%r8) 159 movdqa %xmm2, 32(%r8) 160 movdqa %xmm3, 48(%r8) 174 movdqa %xmm0, (%r8) 175 movdqa %xmm1, 16(%r8) 176 movdqa %xmm2, 32(%r8) 177 movdqa %xmm3, 48(%r8) 178 movdqa %xmm4, 64(%r8) 179 movdqa %xmm5, 80(%r8 [all...] |