/external/chromium_org/third_party/yasm/source/patched-yasm/modules/objfmts/elf/tests/gas64/ |
elf_gas64_reloc.asm | 26 movlpd _VAR(%rip), %xmm1 31 movlpd _VAR2(%rip), %xmm1 36 movlpd _VAR3(%rip), %xmm1
|
/external/chromium_org/third_party/yasm/source/patched-yasm/modules/arch/x86/tests/ |
simd-1.asm | 7 movss xmm1, dword [8]
|
movdq32.asm | 14 movq xmm0, xmm1 15 movq xmm1, xmm0
|
amd200707.asm | 5 insertq xmm0, xmm1, 5, 4 8 movntsd [0], xmm1
|
movdq64.asm | 18 movq xmm0, xmm1 19 movq xmm1, xmm0
|
/external/llvm/test/MC/ELF/ |
undef.s | 20 movsd .Lsym8(%rip), %xmm1
|
merge.s | 10 movsd .Lfoo(%rip), %xmm1
|
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_postproc_sse2.asm | 66 movdqa xmm1, xmm3 ; mm1 = p0..p3 74 movdqa xmm7, xmm1 ; mm7 = r0 p0..p3 76 psubusw xmm5, xmm1 ; mm5 = r1 p0..p3 - r0 p0..p3 85 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 87 psubusw xmm5, xmm1 ; mm5 = r2 p0..p3 - r2 p0..p3 99 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 101 psubusw xmm5, xmm1 ; mm5 = r-2 p0..p3 - p0..p3 111 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 113 psubusw xmm4, xmm1 ; mm5 = r-1 p0..p3 - p0..p3 122 pand xmm1, xmm7 ; mm1 select vals > thresh from sourc [all...] |
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_variance_impl_sse2.asm | 36 movdqa xmm1, [rax+16] 40 pmaddwd xmm1, xmm1 44 paddd xmm0, xmm1 128 movdqu xmm1, XMMWORD PTR [rsi] 134 movdqa xmm3, xmm1 138 punpcklbw xmm1, xmm0 145 psubw xmm1, xmm2 148 paddw xmm7, xmm1 149 pmaddwd xmm1, xmm [all...] |
vp9_temporal_filter_apply_sse2.asm | 81 movq xmm1, [rsi] ; second row 83 punpcklbw xmm1, xmm7 ; src[ 8-15] 89 movdqa xmm1, xmm0 91 punpckhbw xmm1, xmm7 ; src[ 8-15] 101 psubw xmm1, xmm3 ; src - pred[ 8-15] 105 pmullw xmm1, xmm1 ; modifer[ 8-15]^2 109 pmullw xmm1, [GLOBAL(_const_3w)] 113 paddw xmm1, [rsp + rounding_bit] 117 psrlw xmm1, [rsp + strength [all...] |
vp9_sad_sse4.asm | 17 movq xmm1, MMWORD PTR [rdi] 20 punpcklqdq xmm1, xmm3 23 movdqa xmm2, xmm1 24 mpsadbw xmm1, xmm0, 0x0 33 paddw xmm1, xmm2 34 paddw xmm1, xmm3 35 paddw xmm1, xmm4 58 paddw xmm1, xmm5 83 paddw xmm1, xmm5 89 movq xmm1, MMWORD PTR [rdi [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
recon_sse2.asm | 35 movdqu xmm1, [rsi+rax] 44 movdqa [rdi+rcx], xmm1 66 movdqu xmm1, [rsi+rax] 74 movdqa [rdi+rcx], xmm1 95 movdqu xmm1, [rsi+rax] 103 movdqa [rdi+rcx], xmm1 378 movq xmm1, [rsi] 379 punpcklbw xmm1, xmm0 391 psubw xmm1, xmm3 411 paddw xmm3, xmm1 [all...] |
sad_sse4.asm | 17 movq xmm1, MMWORD PTR [rdi] 20 punpcklqdq xmm1, xmm3 23 movdqa xmm2, xmm1 24 mpsadbw xmm1, xmm0, 0x0 33 paddw xmm1, xmm2 34 paddw xmm1, xmm3 35 paddw xmm1, xmm4 58 paddw xmm1, xmm5 83 paddw xmm1, xmm5 89 movq xmm1, MMWORD PTR [rdi [all...] |
mfqe_sse2.asm | 37 movdqa xmm1, [GLOBAL(tMFQE)] 38 psubw xmm1, xmm0 ; dst_weight 64 pmullw xmm4, xmm1 65 pmullw xmm5, xmm1 114 movdqa xmm1, [GLOBAL(tMFQE)] 115 psubw xmm1, xmm0 ; dst_weight 136 pmullw xmm3, xmm1 194 movdqa xmm1, [rdx] ; src2 199 psadbw xmm0, xmm1 204 psadbw xmm2, xmm1 ; sum src2 by misusing SAD against [all...] |
subpixel_ssse3.asm | 80 movdqa xmm1, xmm0 83 movdqa xmm2, xmm1 84 pshufb xmm1, [GLOBAL(shuf2bfrom1)] 87 pmaddubsw xmm1, xmm5 95 paddsw xmm0, xmm1 135 movq xmm1, MMWORD PTR [rsi + 3] ; 3 4 5 6 7 8 9 10 137 punpcklbw xmm0, xmm1 ; -2 3 -1 4 0 5 1 6 2 7 3 8 4 9 5 10 217 movdqa xmm1, xmm0 220 movdqa xmm2, xmm1 221 pshufb xmm1, [GLOBAL(shuf2bfrom1) [all...] |
loopfilter_block_sse2.asm | 200 movdqa xmm1, i1 207 LF_FILTER_HEV_MASK xmm0, xmm1, xmm2, xmm3, xmm4, xmm8, xmm9, xmm10 209 movdqa xmm1, i2 213 LF_FILTER xmm1, xmm2, xmm3, xmm8, xmm0, xmm4 214 movdqa i2, xmm1 222 movdqa xmm1, i7 227 LF_FILTER_HEV_MASK xmm3, xmm8, xmm0, xmm1, xmm2, xmm4, xmm10, xmm11, xmm9 230 movdqa xmm1, i7 233 LF_FILTER xmm0, xmm1, xmm4, xmm8, xmm3, xmm2 235 movdqa i7, xmm1 [all...] |
/external/libvpx/libvpx/vp8/encoder/x86/ |
temporal_filter_apply_sse2.asm | 81 movq xmm1, [rsi] ; second row 83 punpcklbw xmm1, xmm7 ; src[ 8-15] 89 movdqa xmm1, xmm0 91 punpckhbw xmm1, xmm7 ; src[ 8-15] 101 psubw xmm1, xmm3 ; src - pred[ 8-15] 105 pmullw xmm1, xmm1 ; modifer[ 8-15]^2 109 pmullw xmm1, [GLOBAL(_const_3w)] 113 paddw xmm1, [rsp + rounding_bit] 117 psrlw xmm1, [rsp + strength [all...] |
subtract_sse2.asm | 97 movdqa xmm1, [rax] ; pred 100 psubb xmm0, xmm1 102 pxor xmm1, xmm4 ;convert to signed values 104 pcmpgtb xmm1, xmm2 ; obtain sign information 107 punpcklbw xmm0, xmm1 ; put sign back to subtraction 108 punpckhbw xmm2, xmm1 ; put sign back to subtraction 119 movdqa xmm1, xmm3 123 pxor xmm1, xmm4 124 pcmpgtb xmm5, xmm1 ; obtain sign information 126 movdqa xmm1, xmm [all...] |
encodeopt.asm | 28 movdqa xmm1, [rdi] 33 psubw xmm0, xmm1 42 movdqa xmm1, xmm0 45 punpckhdq xmm1, xmm5 47 paddd xmm0, xmm1 48 movdqa xmm1, xmm0 51 paddd xmm0, xmm1 231 movdqa xmm1, [rdi] 240 psubw xmm0, xmm1 260 movdqa xmm1, xmm [all...] |
/external/chromium_org/third_party/openssl/openssl/crypto/aes/asm/ |
vpaes-x86.S | 73 movdqa %xmm6,%xmm1 75 pandn %xmm0,%xmm1 77 psrld $4,%xmm1 97 movdqa -64(%ebx,%ecx,1),%xmm1 114 movdqa %xmm6,%xmm1 115 pandn %xmm0,%xmm1 116 psrld $4,%xmm1 120 pxor %xmm1,%xmm0 133 pxor %xmm1,%xmm3 140 movdqa 64(%ebx,%ecx,1),%xmm1 [all...] |
/external/openssl/crypto/aes/asm/ |
vpaes-x86.S | 73 movdqa %xmm6,%xmm1 75 pandn %xmm0,%xmm1 77 psrld $4,%xmm1 97 movdqa -64(%ebx,%ecx,1),%xmm1 114 movdqa %xmm6,%xmm1 115 pandn %xmm0,%xmm1 116 psrld $4,%xmm1 120 pxor %xmm1,%xmm0 133 pxor %xmm1,%xmm3 140 movdqa 64(%ebx,%ecx,1),%xmm1 [all...] |
/external/chromium_org/third_party/openssl/openssl/crypto/bn/asm/ |
x86_64-gf2m.S | 83 movq (%rsp,%rsi,8),%xmm1 86 pslldq $1,%xmm1 92 pxor %xmm1,%xmm0 98 movq (%rsp,%rsi,8),%xmm1 101 pslldq $2,%xmm1 107 pxor %xmm1,%xmm0 113 movq (%rsp,%rsi,8),%xmm1 116 pslldq $3,%xmm1 122 pxor %xmm1,%xmm0 128 movq (%rsp,%rsi,8),%xmm1 [all...] |
/external/openssl/crypto/bn/asm/ |
x86_64-gf2m.S | 83 movq (%rsp,%rsi,8),%xmm1 86 pslldq $1,%xmm1 92 pxor %xmm1,%xmm0 98 movq (%rsp,%rsi,8),%xmm1 101 pslldq $2,%xmm1 107 pxor %xmm1,%xmm0 113 movq (%rsp,%rsi,8),%xmm1 116 pslldq $3,%xmm1 122 pxor %xmm1,%xmm0 128 movq (%rsp,%rsi,8),%xmm1 [all...] |
/external/compiler-rt/lib/i386/ |
ashldi3.S | 24 movd 8(%esp), %xmm1 25 punpckldq %xmm1, %xmm0 // Load input
|
lshrdi3.S | 24 movd 8(%esp), %xmm1 25 punpckldq %xmm1, %xmm0 // Load input
|