HomeSort by relevance Sort by last modified time
    Searched full:pmullw (Results 1 - 25 of 79) sorted by null

1 2 3 4

  /external/libvpx/libvpx/vp8/common/x86/
subpixel_mmx.asm 59 pmullw mm3, mm1 ; mm3 *= kernel 1 modifiers.
63 pmullw mm4, mm7 ; mm5 *= kernel 4 modifiers
69 pmullw mm5, mm2 ; mm5 *= kernel 2 modifiers
75 pmullw mm4, mm6 ; mm5 *= kernel 3 modifiers
81 pmullw mm4, [rdx+80] ; mm5 *= kernel 0 modifiers
85 pmullw mm5, [rdx] ; mm5 *= kernel 5 modifiers
158 pmullw mm3, mm1 ; mm3 *= kernel 1 modifiers.
162 pmullw mm4, mm7 ; mm4 *= kernel 4 modifiers.
166 pmullw mm4, mm2 ; mm4 *= kernel 2 modifiers.
170 pmullw mm4, [rbx] ; mm4 *= kernel 0 modifiers
    [all...]
subpixel_sse2.asm 77 pmullw xmm3, XMMWORD PTR [rdx] ; x[-2] * H[-2]; Tap 1
81 pmullw xmm4, XMMWORD PTR [rdx+16] ; x[-1] * H[-1]; Tap 2
87 pmullw xmm5, [rdx+32] ; x[ 0] * H[ 0]; Tap 3
92 pmullw xmm6, [rdx+48] ; x[ 1] * h[ 1] ; Tap 4
98 pmullw xmm7, [rdx+64] ; x[ 2] * h[ 2] ; Tap 5
101 pmullw xmm1, [rdx+80] ; x[ 3] * h[ 3] ; Tap 6
202 pmullw xmm3, XMMWORD PTR [rdx] ; x[-2] * H[-2]; Tap 1
206 pmullw xmm4, XMMWORD PTR [rdx+16] ; x[-1] * H[-1]; Tap 2
212 pmullw xmm5, [rdx+32] ; x[ 0] * H[ 0]; Tap 3
217 pmullw xmm6, [rdx+48] ; x[ 1] * h[ 1] ; Tap
    [all...]
dequantize_mmx.asm 30 pmullw mm1, [rax+0] ; mm4 *= kernel 0 modifiers.
34 pmullw mm1, [rax+8] ; mm4 *= kernel 0 modifiers.
38 pmullw mm1, [rax+16] ; mm4 *= kernel 0 modifiers.
42 pmullw mm1, [rax+24] ; mm4 *= kernel 0 modifiers.
72 pmullw mm0, [rdx]
75 pmullw mm1, [rdx +8]
78 pmullw mm2, [rdx+16]
81 pmullw mm3, [rdx+24]
mfqe_sse2.asm 57 pmullw xmm2, xmm0
58 pmullw xmm3, xmm0
64 pmullw xmm4, xmm1
65 pmullw xmm5, xmm1
132 pmullw xmm2, xmm0
136 pmullw xmm3, xmm1
variance_impl_mmx.asm 553 pmullw mm1, [rax] ;
556 pmullw mm3, [rax+8] ;
577 pmullw mm1, [rax] ;
580 pmullw mm3, [rax+8] ;
589 pmullw mm3, [rdx] ;
591 pmullw mm1, [rdx+8] ;
701 pmullw mm1, [rax] ;
703 pmullw mm2, [rax] ;
707 pmullw mm3, [rax+8] ;
709 pmullw mm4, [rax+8]
    [all...]
postproc_mmx.asm 100 pmullw mm1, mm1 ;
125 pmullw mm2, mm2
134 pmullw mm1, mm1
151 pmullw mm1, mm1
variance_impl_sse2.asm 460 pmullw xmm1, [rax] ;
462 pmullw xmm3, [rax+16] ;
480 pmullw xmm1, [rax] ;
482 pmullw xmm3, [rax+16] ;
491 pmullw xmm3, [rdx] ;
492 pmullw xmm1, [rdx+16] ;
544 pmullw xmm1, [rdx] ;
545 pmullw xmm3, [rdx+16] ;
611 pmullw xmm1, [rax] ;
613 pmullw xmm3, [rax+16]
    [all...]
idctllm_sse2.asm 39 pmullw xmm4, xmm5
140 pmullw xmm0, [rdx]
141 pmullw xmm2, [rdx+16]
142 pmullw xmm1, [rdx]
143 pmullw xmm3, [rdx+16]
472 pmullw xmm0, [rdx]
473 pmullw xmm2, [rdx+16]
474 pmullw xmm1, [rdx]
475 pmullw xmm3, [rdx+16]
  /external/libvpx/libvpx/vp9/encoder/x86/
vp9_subpel_variance.asm 357 ; slightly faster because of pmullw latency. It would also cut our rodata
359 pmullw m2, filter_y_a
360 pmullw m3, filter_y_b
362 pmullw m0, filter_y_a
363 pmullw m4, filter_y_b
401 pmullw m0, filter_y_a
402 pmullw m1, m2, filter_y_b
405 pmullw m2, filter_y_a
406 pmullw m4, filter_y_b
645 pmullw m2, filter_y_
    [all...]
vp9_temporal_filter_apply_sse2.asm 104 pmullw xmm0, xmm0 ; modifer[ 0- 7]^2
105 pmullw xmm1, xmm1 ; modifer[ 8-15]^2
108 pmullw xmm0, [GLOBAL(_const_3w)]
109 pmullw xmm1, [GLOBAL(_const_3w)]
127 pmullw xmm2, [rsp + filter_weight]
128 pmullw xmm3, [rsp + filter_weight]
150 pmullw xmm0, xmm2
151 pmullw xmm1, xmm3
vp9_quantize_ssse3.asm 94 pmullw m8, m3 ; dqc[i] = qc[i] * q
96 pmullw m13, m3 ; dqc[i] = qc[i] * q
148 pmullw m14, m3 ; dqc[i] = qc[i] * q
149 pmullw m13, m3 ; dqc[i] = qc[i] * q
  /external/libvpx/libvpx/vpx_scale/win32/
scaleopt.c 114 pmullw mm1, mm5
116 pmullw mm0, mm6
166 pmullw mm1, three_fourths
168 pmullw mm2, one_fourths
171 pmullw mm3, two_fourths
175 pmullw mm4, two_fourths
180 pmullw mm5, one_fourths
189 pmullw mm6, three_fourths
265 pmullw mm0, mm6
267 pmullw mm1, mm
    [all...]
  /external/libvpx/libvpx/vp8/encoder/x86/
temporal_filter_apply_sse2.asm 104 pmullw xmm0, xmm0 ; modifer[ 0- 7]^2
105 pmullw xmm1, xmm1 ; modifer[ 8-15]^2
108 pmullw xmm0, [GLOBAL(_const_3w)]
109 pmullw xmm1, [GLOBAL(_const_3w)]
127 pmullw xmm2, [rsp + filter_weight]
128 pmullw xmm3, [rsp + filter_weight]
150 pmullw xmm0, xmm2
151 pmullw xmm1, xmm3
quantize_ssse3.asm 97 pmullw xmm0, xmm1
98 pmullw xmm4, xmm5
quantize_mmx.asm 66 pmullw mm3, mm2
106 pmullw mm7, mm6
147 pmullw mm7, mm6
188 pmullw mm7, mm6
quantize_sse4.asm 200 pmullw xmm0, xmm4
201 pmullw xmm1, xmm5
  /external/chromium_org/media/base/simd/
convert_yuva_to_argb_mmx.inc 43 pmullw mm0, mm2
48 pmullw mm1, mm2
78 pmullw mm1, mm0
132 pmullw mm0, mm2
137 pmullw mm1, mm2
167 pmullw mm1, mm0
  /external/libvpx/libvpx/vp9/common/x86/
vp9_postproc_mmx.asm 71 pmullw mm3, mm6 ; mm3 *= kernel 2 modifiers
76 pmullw mm6, mm5 ; mm6 *= p0..p3 * kernel 3 modifiers
89 pmullw mm6, mm5 ; mm5 *= kernel 4 modifiers
105 pmullw mm6, mm5 ; mm5 *= kernel 0 modifiers
119 pmullw mm6, mm4 ; mm4 *= kernel 1 modifiers.
166 pmullw mm3, mm6 ; mm3 *= kernel 2 modifiers
172 pmullw mm6, mm5 ; mm6 *= p1..p4 * kernel 3 modifiers
186 pmullw mm6, mm5 ; mm5 *= kernel 4 modifiers
202 pmullw mm6, mm5 ; mm5 *= kernel 0 modifiers
216 pmullw mm6, mm4 ; mm4 *= kernel 1 modifiers
    [all...]
  /external/chromium_org/third_party/mesa/src/src/mesa/x86/
mmx_blend.S 27 PMULLW ( MP1, MA1 ) /* t1 = p1*a1 */ ;\
30 TWO(PMULLW ( MP2, MA2 )) /* t2 = p2*a2 */ ;\
50 PMULLW ( MP1, MA1 ) /* t1 = p1*a1 */ ;\
51 TWO(PMULLW ( MP2, MA2 )) /* t2 = p2*a2 */ ;\
78 PMULLW ( MP1, MA1 ) /* t1 = p1*a1 */ ;\
81 TWO(PMULLW ( MP2, MA2 )) /* t2 = p2*a2 */ ;\
102 PMULLW ( MP1, MA1 ) /* t1 = (q1 - p1)*pa1 */ ;\
106 TWO(PMULLW ( MP2, MA2 )) /* t2 = (q2 - p2)*pa2 */ ;\
133 PMULLW ( MP1, MA1 ) /* t1 = (q1 - p1)*pa1 */ ;\
137 TWO(PMULLW ( MP2, MA2 )) /* t2 = (q2 - p2)*pa2 */ ;
    [all...]
read_rgba_span_x86.S 561 pmullw %mm6, %mm0
562 pmullw %mm6, %mm2
596 pmullw %mm6, %mm0
597 pmullw %mm6, %mm2
634 pmullw %mm6, %mm0
635 pmullw %mm6, %mm2
665 pmullw %mm6, %mm0
  /external/mesa3d/src/mesa/x86/
mmx_blend.S 27 PMULLW ( MP1, MA1 ) /* t1 = p1*a1 */ ;\
30 TWO(PMULLW ( MP2, MA2 )) /* t2 = p2*a2 */ ;\
50 PMULLW ( MP1, MA1 ) /* t1 = p1*a1 */ ;\
51 TWO(PMULLW ( MP2, MA2 )) /* t2 = p2*a2 */ ;\
78 PMULLW ( MP1, MA1 ) /* t1 = p1*a1 */ ;\
81 TWO(PMULLW ( MP2, MA2 )) /* t2 = p2*a2 */ ;\
102 PMULLW ( MP1, MA1 ) /* t1 = (q1 - p1)*pa1 */ ;\
106 TWO(PMULLW ( MP2, MA2 )) /* t2 = (q2 - p2)*pa2 */ ;\
133 PMULLW ( MP1, MA1 ) /* t1 = (q1 - p1)*pa1 */ ;\
137 TWO(PMULLW ( MP2, MA2 )) /* t2 = (q2 - p2)*pa2 */ ;
    [all...]
  /external/llvm/test/CodeGen/X86/
2008-02-26-AsmDirectMemOp.ll 15 tail call void asm sideeffect "movd $1, %mm6 \0A\09packssdw %mm6, %mm6 \0A\09packssdw %mm6, %mm6 \0A\09movd $2, %mm5 \0A\09pxor %mm7, %mm7 \0A\09packssdw %mm5, %mm5 \0A\09packssdw %mm5, %mm5 \0A\09psubw %mm5, %mm7 \0A\09pxor %mm4, %mm4 \0A\09.align 1<<4\0A\091: \0A\09movq ($0, $3), %mm0 \0A\09movq 8($0, $3), %mm1 \0A\09pmullw %mm6, %mm0 \0A\09pmullw %mm6, %mm1 \0A\09movq ($0, $3), %mm2 \0A\09movq 8($0, $3), %mm3 \0A\09pcmpgtw %mm4, %mm2 \0A\09pcmpgtw %mm4, %mm3 \0A\09pxor %mm2, %mm0 \0A\09pxor %mm3, %mm1 \0A\09paddw %mm7, %mm0 \0A\09paddw %mm7, %mm1 \0A\09pxor %mm0, %mm2 \0A\09pxor %mm1, %mm3 \0A\09pcmpeqw %mm7, %mm0 \0A\09pcmpeqw %mm7, %mm1 \0A\09pandn %mm2, %mm0 \0A\09pandn %mm3, %mm1 \0A\09movq %mm0, ($0, $3) \0A\09movq %mm1, 8($0, $3) \0A\09add $$16, $3 \0A\09jng 1b \0A\09", "r,imr,imr,r,~{dirflag},~{fpsr},~{flags},~{memory}"( i16* null, i32 %tmp1, i32 0, i32 0 ) nounwind
  /external/qemu/distrib/sdl-1.2.15/src/audio/
SDL_mixer_MMX.c 76 " pmullw %%mm0,%%mm1\n" /* mm1 = l(a*v)|l(b*v)|l(c*v)|l(d*v) */
81 " pmullw %%mm0,%%mm4\n" /* mm4 = l(e*v)|l(f*v)|l(g*v)|l(h*v) */
179 " pmullw %%mm0,%%mm1\n" /* mm1 = v*a|v*b|v*c|v*d */
182 " pmullw %%mm0,%%mm3\n" /* mm3 = v*e|v*f|v*g|v*h */
SDL_mixer_MMX_VC.c 77 pmullw mm1, mm0 //%%mm0,%%mm1\n" // mm1 = l(a*v)|l(b*v)|l(c*v)|l(d*v)
80 pmullw mm4, mm0 //%%mm0,%%mm4\n" // mm4 = l(e*v)|l(f*v)|l(g*v)|l(h*v)
161 pmullw mm1, mm0 //%%mm0,%%mm1 // mm1 = v*a|v*b|v*c|v*d
163 pmullw mm3, mm0 //%%mm0,%%mm3 // mm3 = v*e|v*f|v*g|v*h
  /external/qemu/distrib/sdl-1.2.15/src/video/
SDL_yuv_mmx.c 119 "pmullw %10,%%mm0\n" // red*-46dec=0.7136*64
120 "pmullw %11,%%mm1\n" // red*89dec=1.4013*64
150 "pmullw %14,%%mm5\n" // blue*-109dec=1.7129*64
151 "pmullw %15,%%mm1\n" // blue*114dec=1.78125*64
279 "pmullw %10, %%mm2\n" // Cb2green 0 R3 0 R2 0 R1 0 R0
281 "pmullw %11, %%mm0\n" // Cb2blue
283 "pmullw %13, %%mm3\n" // Cr2green
285 "pmullw %14, %%mm1\n" // Cr2red
287 "pmullw %15, %%mm6\n" // lum1
289 "pmullw %15, %%mm7\n" // lum
    [all...]

Completed in 8061 milliseconds

1 2 3 4