HomeSort by relevance Sort by last modified time
    Searched full:xmm5 (Results 76 - 100 of 314) sorted by null

1 2 34 5 6 7 8 91011>>

  /external/llvm/test/CodeGen/X86/
stack-folding-int-avx1.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
108 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
117 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
126 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
stack-folding-int-sse42.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
41 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
50 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
59 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
115 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
135 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
144 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
153 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
fold-xmm-zero.ll 12 %0 = tail call %0 asm sideeffect "foo", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float 1.000000e+00, float 2.000000e+00, float 3.000000e+00, float 4.000000e+00, float 5.000000e+00, float 6.000000e+00, float 7.000000e+00, float 8.000000e+00) nounwind
22 %1 = tail call %0 asm sideeffect "bar", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float %div, float %asmresult8, float %asmresult9, float %asmresult10, float %asmresult11, float %asmresult12, float %asmresult13, float %asmresult14) nounwind
32 %2 = tail call %0 asm sideeffect "baz", "={xmm0},={xmm1},={xmm2},={xmm3},={xmm4},={xmm5},={xmm6},={xmm7},0,1,2,3,4,5,6,7,~{dirflag},~{fpsr},~{flags}"(float %div33, float %asmresult25, float %asmresult26, float %asmresult27, float %asmresult28, float %asmresult29, float %asmresult30, float %asmresult31) nounwind
stack-folding-fp-sse42.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
47 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
55 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
64 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
73 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
82 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
96 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
stack-folding-fp-avx1.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
22 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
30 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
38 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
46 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
54 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
63 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
71 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
80 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
89 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
stack-folding-int-avx2.ll 14 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
23 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
32 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
44 %3 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
51 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
61 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
70 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
79 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
88 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"()
97 %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"(
    [all...]
  /external/boringssl/src/crypto/aes/asm/
vpaes-x86.pl 178 ## Clobbers %xmm1-%xmm5, %eax, %ebx, %ecx, %edx
188 &movdqu ("xmm5",&QWP(0,$key));
191 &pxor ("xmm2","xmm5");
206 &pxor ("xmm4","xmm5"); # 4 = sb1u + k
207 &movdqa ("xmm5",&QWP($k_sb2,$const)); # 4 : sb2u
210 &pshufb ("xmm5","xmm2"); # 4 = sb2u
215 &pxor ("xmm2","xmm5"); # 2 = 2A
230 &movdqa ("xmm5",&QWP($k_inv+16,$const));# 2 : a/k
234 &pshufb ("xmm5","xmm0"); # 2 = a/k
239 &pxor ("xmm3","xmm5"); # 3 = iak = 1/i + a/
    [all...]
  /external/libyuv/files/source/
scale_argb.cc 210 movd xmm5, eax // xmm5 = y fraction local
211 punpcklbw xmm5, xmm5 local
212 punpcklwd xmm5, xmm5 local
213 pshufd xmm5, xmm5, 0 local
230 pmulhw xmm2, xmm5 // scale diff
231 pmulhw xmm3, xmm5
301 movd xmm5, eax \/\/ low fraction 128..1 local
302 punpcklbw xmm5, xmm0 local
303 punpcklwd xmm5, xmm5 local
304 pshufd xmm5, xmm5, 0 local
    [all...]
  /external/mesa3d/src/mesa/x86/
sse_xform3.S 87 MOVSS ( REGOFF(4, ESI), XMM5 ) /* | | | oy */
88 SHUFPS ( CONST(0x0), XMM5, XMM5 ) /* oy | oy | oy | oy */
93 MULPS ( XMM1, XMM5 ) /* m7*oy | m6*oy | m5*oy | m4*oy */
96 ADDPS ( XMM5, XMM4 )
281 MOVSS ( S(2), XMM5 ) /* oz */
282 SHUFPS ( CONST(0x0), XMM5, XMM5 ) /* oz | oz */
283 MULPS ( XMM2, XMM5 ) /* oz*m9 | oz*m8 */
284 ADDPS ( XMM5, XMM0 ) /* +oy*m5 | +ox*m0 *
    [all...]
sse_normal.S 168 MOVSS ( S(2), XMM5 ) /* uz */
169 SHUFPS ( CONST(0x0), XMM5, XMM5 ) /* uz | uz */
170 MULPS ( XMM2, XMM5 ) /* uz*m6 | uz*m2 */
173 ADDPS ( XMM5, XMM3 )
181 MOVSS ( S(0), XMM5 ) /* ux */
182 MULSS ( XMM6, XMM5 ) /* ux*m8*scale */
185 ADDSS ( XMM5, XMM3 )
  /external/llvm/test/TableGen/
cast.td 51 def XMM5: Register<"xmm5">;
64 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
  /external/libvpx/libvpx/third_party/libyuv/source/
scale_posix.cc 128 "pcmpeqb %%xmm5,%%xmm5 \n"
129 "psrlw $0x8,%%xmm5 \n"
140 "pand %%xmm5,%%xmm2 \n"
141 "pand %%xmm5,%%xmm3 \n"
155 , "xmm0", "xmm1", "xmm5"
163 "pcmpeqb %%xmm5,%%xmm5 \n"
164 "psrlw $0x8,%%xmm5 \n"
180 "pand %%xmm5,%%xmm2 \n
    [all...]
  /external/boringssl/win-x86/crypto/sha/
sha512-586.asm 408 movdqa xmm5,[32+ebp]
411 paddq xmm5,xmm2
413 movdqa [edx-96],xmm5
415 movdqa xmm5,xmm4
421 movdqa xmm6,xmm5
422 movdqu xmm5,[80+edi]
430 paddq xmm0,xmm5
467 movdqa xmm3,xmm5
597 movdqa [16+edx],xmm5
599 movdqa xmm5,xmm
    [all...]
  /bionic/libc/arch-x86/atom/string/
ssse3-strcpy-atom.S 358 movaps 16(%ecx), %xmm5
361 pminub %xmm5, %xmm2
375 movaps %xmm5, -48(%edx)
389 pcmpeqb %xmm5, %xmm0
404 movaps %xmm5, -48(%edx)
503 movaps 63(%ecx), %xmm5
505 pminub %xmm5, %xmm7
509 movaps %xmm5, %xmm7
510 palignr $1, %xmm4, %xmm5
522 movaps %xmm5, 48(%edx
    [all...]
  /external/libvpx/libvpx/vp9/common/x86/
vp9_subpixel_bilinear_sse2.asm 76 pxor xmm5, xmm5
84 punpcklbw xmm0, xmm5
85 punpcklbw xmm1, xmm5
105 punpcklbw xmm0, xmm5
106 punpcklbw xmm1, xmm5
107 punpckhbw xmm2, xmm5
108 punpckhbw xmm3, xmm5
  /external/llvm/test/MC/X86/
cfi_def_cfa-crash.s 32 movdqa %xmm5, -0x30(%rbp)
51 movdqa -0x30(%rbp), %xmm5
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/
vp9_subpixel_bilinear_sse2.asm 76 pxor xmm5, xmm5
84 punpcklbw xmm0, xmm5
85 punpcklbw xmm1, xmm5
105 punpcklbw xmm0, xmm5
106 punpcklbw xmm1, xmm5
107 punpckhbw xmm2, xmm5
108 punpckhbw xmm3, xmm5
  /external/libvpx/libvpx/vp8/common/x86/
recon_sse2.asm 53 movdqu xmm5, [rsi+rax*2]
60 movdqa [rdi+rcx*2],xmm5
83 movdqu xmm5, [rsi+rax*2]
91 movdqa [rdi+rcx*2],xmm5
403 movd xmm5, ebx
406 punpcklbw xmm5, xmm0
408 pshuflw xmm5, xmm5, 0x0
410 punpcklqdq xmm5, xmm5
    [all...]
sad_sse3.asm 170 lddqu xmm5, XMMWORD PTR [%3]
174 psadbw xmm5, xmm0
187 paddw xmm5, xmm1
205 paddw xmm5, xmm1
265 lddqu xmm5, XMMWORD PTR [%4]
270 psadbw xmm5, xmm0
285 paddw xmm5, xmm2
302 paddw xmm5, xmm2
399 movq xmm0, xmm5
400 psrldq xmm5,
    [all...]
  /external/libvpx/libvpx/vp8/encoder/x86/
subtract_sse2.asm 111 movdqa xmm5, [rax + rbx]
120 psubb xmm3, xmm5
122 pxor xmm5, xmm4 ;convert to signed values
124 pcmpgtb xmm5, xmm1 ; obtain sign information
127 punpcklbw xmm3, xmm5 ; put sign back to subtraction
128 punpckhbw xmm1, xmm5 ; put sign back to subtraction
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/
recon_sse2.asm 53 movdqu xmm5, [rsi+rax*2]
60 movdqa [rdi+rcx*2],xmm5
83 movdqu xmm5, [rsi+rax*2]
91 movdqa [rdi+rcx*2],xmm5
399 movd xmm5, [rsi+rax]
402 punpcklbw xmm5, xmm0
404 pshuflw xmm5, xmm5, 0x0
406 punpcklqdq xmm5, xmm5
    [all...]
sad_sse3.asm 170 lddqu xmm5, XMMWORD PTR [%3]
174 psadbw xmm5, xmm0
187 paddw xmm5, xmm1
205 paddw xmm5, xmm1
265 lddqu xmm5, XMMWORD PTR [%4]
270 psadbw xmm5, xmm0
285 paddw xmm5, xmm2
302 paddw xmm5, xmm2
399 movq xmm0, xmm5
400 psrldq xmm5,
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/
subtract_sse2.asm 111 movdqa xmm5, [rax + rbx]
120 psubb xmm3, xmm5
122 pxor xmm5, xmm4 ;convert to signed values
124 pcmpgtb xmm5, xmm1 ; obtain sign information
127 punpcklbw xmm3, xmm5 ; put sign back to subtraction
128 punpckhbw xmm1, xmm5 ; put sign back to subtraction
  /bionic/libm/x86/
e_pow.S 154 movd %edx, %xmm5
160 psllq $32, %xmm5
163 andpd %xmm3, %xmm5
166 subsd %xmm5, %xmm3
170 mulpd %xmm0, %xmm5
174 subsd %xmm2, %xmm5
182 unpcklpd %xmm3, %xmm5
184 addsd %xmm5, %xmm3
187 addpd -3616(%ebx,%edx), %xmm5
193 addsd %xmm7, %xmm5
    [all...]
s_log1p.S 90 movsd 2128(%ebx), %xmm5
107 pshufd $228, %xmm5, %xmm6
125 andpd %xmm1, %xmm5
127 subsd %xmm5, %xmm1
129 mulsd %xmm0, %xmm5
137 subsd %xmm2, %xmm5
142 addsd %xmm5, %xmm1
143 movapd %xmm1, %xmm5
145 subsd %xmm1, %xmm5
146 addsd %xmm5, %xmm
    [all...]

Completed in 484 milliseconds

1 2 34 5 6 7 8 91011>>