/external/llvm/test/CodeGen/X86/ |
stack-folding-int-avx1.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
stack-folding-int-sse42.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 115 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 135 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 144 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 153 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
fold-xmm-zero.ll | 12 %0 = tail call %0 asm sideeffect "foo", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float 1.000000e+00, float 2.000000e+00, float 3.000000e+00, float 4.000000e+00, float 5.000000e+00, float 6.000000e+00, float 7.000000e+00, float 8.000000e+00) nounwind 22 %1 = tail call %0 asm sideeffect "bar", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float %div, float %asmresult8, float %asmresult9, float %asmresult10, float %asmresult11, float %asmresult12, float %asmresult13, float %asmresult14) nounwind 32 %2 = tail call %0 asm sideeffect "baz", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float %div33, float %asmresult25, float %asmresult26, float %asmresult27, float %asmresult28, float %asmresult29, float %asmresult30, float %asmresult31) nounwind
|
stack-folding-fp-sse42.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 47 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 55 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 64 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 73 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 82 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 96 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
stack-folding-fp-avx1.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 46 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 54 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 63 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 71 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 80 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 89 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
stack-folding-int-avx2.ll | 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 44 %3 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 51 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 61 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 70 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 79 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() 97 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"( [all...] |
/external/boringssl/src/crypto/aes/asm/ |
vpaes-x86.pl | 178 ## Clobbers %xmm1-%xmm5, %eax, %ebx, %ecx, %edx 188 &movdqu ("xmm5",&QWP(0,$key)); 191 &pxor ("xmm2","xmm5"); 206 &pxor ("xmm4","xmm5"); # 4 = sb1u + k 207 &movdqa ("xmm5",&QWP($k_sb2,$const)); # 4 : sb2u 210 &pshufb ("xmm5","xmm2"); # 4 = sb2u 215 &pxor ("xmm2","xmm5"); # 2 = 2A 230 &movdqa ("xmm5",&QWP($k_inv+16,$const));# 2 : a/k 234 &pshufb ("xmm5","xmm0"); # 2 = a/k 239 &pxor ("xmm3","xmm5"); # 3 = iak = 1/i + a/ [all...] |
/external/libyuv/files/source/ |
scale_argb.cc | 210 movd xmm5, eax // xmm5 = y fraction local 211 punpcklbw xmm5, xmm5 local 212 punpcklwd xmm5, xmm5 local 213 pshufd xmm5, xmm5, 0 local 230 pmulhw xmm2, xmm5 // scale diff 231 pmulhw xmm3, xmm5 301 movd xmm5, eax \/\/ low fraction 128..1 local 302 punpcklbw xmm5, xmm0 local 303 punpcklwd xmm5, xmm5 local 304 pshufd xmm5, xmm5, 0 local [all...] |
/external/mesa3d/src/mesa/x86/ |
sse_xform3.S | 87 MOVSS ( REGOFF(4, ESI), XMM5 ) /* | | | oy */ 88 SHUFPS ( CONST(0x0), XMM5, XMM5 ) /* oy | oy | oy | oy */ 93 MULPS ( XMM1, XMM5 ) /* m7*oy | m6*oy | m5*oy | m4*oy */ 96 ADDPS ( XMM5, XMM4 ) 281 MOVSS ( S(2), XMM5 ) /* oz */ 282 SHUFPS ( CONST(0x0), XMM5, XMM5 ) /* oz | oz */ 283 MULPS ( XMM2, XMM5 ) /* oz*m9 | oz*m8 */ 284 ADDPS ( XMM5, XMM0 ) /* +oy*m5 | +ox*m0 * [all...] |
sse_normal.S | 168 MOVSS ( S(2), XMM5 ) /* uz */ 169 SHUFPS ( CONST(0x0), XMM5, XMM5 ) /* uz | uz */ 170 MULPS ( XMM2, XMM5 ) /* uz*m6 | uz*m2 */ 173 ADDPS ( XMM5, XMM3 ) 181 MOVSS ( S(0), XMM5 ) /* ux */ 182 MULSS ( XMM6, XMM5 ) /* ux*m8*scale */ 185 ADDSS ( XMM5, XMM3 )
|
/external/llvm/test/TableGen/ |
cast.td | 51 def XMM5: Register<"xmm5">; 64 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
scale_posix.cc | 128 "pcmpeqb %%xmm5,%%xmm5 \n" 129 "psrlw $0x8,%%xmm5 \n" 140 "pand %%xmm5,%%xmm2 \n" 141 "pand %%xmm5,%%xmm3 \n" 155 , "xmm0", "xmm1", "xmm5" 163 "pcmpeqb %%xmm5,%%xmm5 \n" 164 "psrlw $0x8,%%xmm5 \n" 180 "pand %%xmm5,%%xmm2 \n [all...] |
/external/boringssl/win-x86/crypto/sha/ |
sha512-586.asm | 408 movdqa xmm5,[32+ebp] 411 paddq xmm5,xmm2 413 movdqa [edx-96],xmm5 415 movdqa xmm5,xmm4 421 movdqa xmm6,xmm5 422 movdqu xmm5,[80+edi] 430 paddq xmm0,xmm5 467 movdqa xmm3,xmm5 597 movdqa [16+edx],xmm5 599 movdqa xmm5,xmm [all...] |
/bionic/libc/arch-x86/atom/string/ |
ssse3-strcpy-atom.S | 358 movaps 16(%ecx), %xmm5 361 pminub %xmm5, %xmm2 375 movaps %xmm5, -48(%edx) 389 pcmpeqb %xmm5, %xmm0 404 movaps %xmm5, -48(%edx) 503 movaps 63(%ecx), %xmm5 505 pminub %xmm5, %xmm7 509 movaps %xmm5, %xmm7 510 palignr $1, %xmm4, %xmm5 522 movaps %xmm5, 48(%edx [all...] |
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_subpixel_bilinear_sse2.asm | 76 pxor xmm5, xmm5 84 punpcklbw xmm0, xmm5 85 punpcklbw xmm1, xmm5 105 punpcklbw xmm0, xmm5 106 punpcklbw xmm1, xmm5 107 punpckhbw xmm2, xmm5 108 punpckhbw xmm3, xmm5
|
/external/llvm/test/MC/X86/ |
cfi_def_cfa-crash.s | 32 movdqa %xmm5, -0x30(%rbp) 51 movdqa -0x30(%rbp), %xmm5
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_subpixel_bilinear_sse2.asm | 76 pxor xmm5, xmm5 84 punpcklbw xmm0, xmm5 85 punpcklbw xmm1, xmm5 105 punpcklbw xmm0, xmm5 106 punpcklbw xmm1, xmm5 107 punpckhbw xmm2, xmm5 108 punpckhbw xmm3, xmm5
|
/external/libvpx/libvpx/vp8/common/x86/ |
recon_sse2.asm | 53 movdqu xmm5, [rsi+rax*2] 60 movdqa [rdi+rcx*2],xmm5 83 movdqu xmm5, [rsi+rax*2] 91 movdqa [rdi+rcx*2],xmm5 403 movd xmm5, ebx 406 punpcklbw xmm5, xmm0 408 pshuflw xmm5, xmm5, 0x0 410 punpcklqdq xmm5, xmm5 [all...] |
sad_sse3.asm | 170 lddqu xmm5, XMMWORD PTR [%3] 174 psadbw xmm5, xmm0 187 paddw xmm5, xmm1 205 paddw xmm5, xmm1 265 lddqu xmm5, XMMWORD PTR [%4] 270 psadbw xmm5, xmm0 285 paddw xmm5, xmm2 302 paddw xmm5, xmm2 399 movq xmm0, xmm5 400 psrldq xmm5, [all...] |
/external/libvpx/libvpx/vp8/encoder/x86/ |
subtract_sse2.asm | 111 movdqa xmm5, [rax + rbx] 120 psubb xmm3, xmm5 122 pxor xmm5, xmm4 ;convert to signed values 124 pcmpgtb xmm5, xmm1 ; obtain sign information 127 punpcklbw xmm3, xmm5 ; put sign back to subtraction 128 punpckhbw xmm1, xmm5 ; put sign back to subtraction
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
recon_sse2.asm | 53 movdqu xmm5, [rsi+rax*2] 60 movdqa [rdi+rcx*2],xmm5 83 movdqu xmm5, [rsi+rax*2] 91 movdqa [rdi+rcx*2],xmm5 399 movd xmm5, [rsi+rax] 402 punpcklbw xmm5, xmm0 404 pshuflw xmm5, xmm5, 0x0 406 punpcklqdq xmm5, xmm5 [all...] |
sad_sse3.asm | 170 lddqu xmm5, XMMWORD PTR [%3] 174 psadbw xmm5, xmm0 187 paddw xmm5, xmm1 205 paddw xmm5, xmm1 265 lddqu xmm5, XMMWORD PTR [%4] 270 psadbw xmm5, xmm0 285 paddw xmm5, xmm2 302 paddw xmm5, xmm2 399 movq xmm0, xmm5 400 psrldq xmm5, [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/ |
subtract_sse2.asm | 111 movdqa xmm5, [rax + rbx] 120 psubb xmm3, xmm5 122 pxor xmm5, xmm4 ;convert to signed values 124 pcmpgtb xmm5, xmm1 ; obtain sign information 127 punpcklbw xmm3, xmm5 ; put sign back to subtraction 128 punpckhbw xmm1, xmm5 ; put sign back to subtraction
|
/bionic/libm/x86/ |
e_pow.S | 154 movd %edx, %xmm5 160 psllq $32, %xmm5 163 andpd %xmm3, %xmm5 166 subsd %xmm5, %xmm3 170 mulpd %xmm0, %xmm5 174 subsd %xmm2, %xmm5 182 unpcklpd %xmm3, %xmm5 184 addsd %xmm5, %xmm3 187 addpd -3616(%ebx,%edx), %xmm5 193 addsd %xmm7, %xmm5 [all...] |
s_log1p.S | 90 movsd 2128(%ebx), %xmm5 107 pshufd $228, %xmm5, %xmm6 125 andpd %xmm1, %xmm5 127 subsd %xmm5, %xmm1 129 mulsd %xmm0, %xmm5 137 subsd %xmm2, %xmm5 142 addsd %xmm5, %xmm1 143 movapd %xmm1, %xmm5 145 subsd %xmm1, %xmm5 146 addsd %xmm5, %xmm [all...] |