/external/libvpx/libvpx/vpx_dsp/x86/ |
subtract_sse2.asm | 52 mova [diffq+mmsize*0+%5], m0 53 mova [diffq+mmsize*1+%5], m2 54 mova [diffq+mmsize*0+%6], m4 55 mova [diffq+mmsize*1+%6], m1 60 loop16 0*mmsize, 1*mmsize, 0*mmsize, 1*mmsize, 0*mmsize, 2*mmsize [all...] |
sad_sse2.asm | 59 pavgb m1, [second_predq+mmsize*0] 60 pavgb m2, [second_predq+mmsize*1] 61 pavgb m3, [second_predq+mmsize*2] 62 pavgb m4, [second_predq+mmsize*3] 63 lea second_predq, [second_predq+mmsize*4] 102 pavgb m1, [second_predq+mmsize*0] 103 pavgb m2, [second_predq+mmsize*1] 104 pavgb m3, [second_predq+mmsize*2] 105 pavgb m4, [second_predq+mmsize*3] 106 lea second_predq, [second_predq+mmsize*4 [all...] |
highbd_sad_sse2.asm | 68 pavgw m1, [second_predq+mmsize*0] 69 pavgw m2, [second_predq+mmsize*1] 70 pavgw m3, [second_predq+mmsize*2] 71 pavgw m4, [second_predq+mmsize*3] 72 lea second_predq, [second_predq+mmsize*4] 106 pavgw m1, [second_predq+mmsize*0] 107 pavgw m2, [second_predq+mmsize*1] 108 pavgw m3, [second_predq+mmsize*2] 109 pavgw m4, [second_predq+mmsize*3] 110 lea second_predq, [second_predq+mmsize*4 [all...] |
subpel_variance_sse2.asm | 61 %if mmsize == 16 81 %else ; mmsize == 8 238 %if mmsize == 16 240 %else ; mmsize == 8 298 %if mmsize == 16 300 %else ; mmsize == 8 309 %if mmsize == 16 311 %else ; mmsize == 8 350 %if ARCH_X86_64 && mmsize == 16 505 %if mmsize == 1 [all...] |
quantize_ssse3_x86_64.asm | 165 add ncoeffq, mmsize 268 add ncoeffq, mmsize 289 add ncoeffq, mmsize 338 add ncoeffq, mmsize
|
quantize_avx_x86_64.asm | 274 add ncoeffq, mmsize 359 add ncoeffq, mmsize 394 add ncoeffq, mmsize 481 add ncoeffq, mmsize 531 add ncoeffq, mmsize
|
highbd_subpel_variance_impl_sse2.asm | 60 %if mmsize == 16 294 %if ARCH_X86_64 && mmsize == 16 521 %if ARCH_X86_64 && mmsize == 16 643 %if ARCH_X86_64 && mmsize == 16 742 %if ARCH_X86_64 && mmsize == 16 870 %if ARCH_X86_64 && mmsize == 16
|
sad4d_sse2.asm | 195 %if mmsize == 16
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/ |
vp9_subtract_sse2.asm | 52 mova [diffq+mmsize*0+%5], m0 53 mova [diffq+mmsize*1+%5], m2 54 mova [diffq+mmsize*0+%6], m4 55 mova [diffq+mmsize*1+%6], m1 60 loop16 0*mmsize, 1*mmsize, 0*mmsize, 1*mmsize, 0*mmsize, 2*mmsize [all...] |
vp9_sad_sse2.asm | 59 pavgb m1, [second_predq+mmsize*0] 60 pavgb m2, [second_predq+mmsize*1] 61 pavgb m3, [second_predq+mmsize*2] 62 pavgb m4, [second_predq+mmsize*3] 63 lea second_predq, [second_predq+mmsize*4] 102 pavgb m1, [second_predq+mmsize*0] 103 pavgb m2, [second_predq+mmsize*1] 104 pavgb m3, [second_predq+mmsize*2] 105 pavgb m4, [second_predq+mmsize*3] 106 lea second_predq, [second_predq+mmsize*4 [all...] |
vp9_error_sse2.asm | 29 mova m3, [uqcq+sizeq*2+mmsize] 30 mova m1, [dqcq+sizeq*2+mmsize] 56 add sizeq, mmsize
|
vp9_quantize_ssse3.asm | 114 add ncoeffq, mmsize 168 add ncoeffq, mmsize 178 add ncoeffq, mmsize 211 add ncoeffq, mmsize
|
vp9_subpel_variance.asm | 85 %if mmsize == 16 105 %else ; mmsize == 8 262 %if mmsize == 16 264 %else ; mmsize == 8 322 %if mmsize == 16 324 %else ; mmsize == 8 333 %if mmsize == 16 335 %else ; mmsize == 8 374 %if ARCH_X86_64 && mmsize == 16 529 %if mmsize == 1 [all...] |
vp9_sad4d_sse2.asm | 195 %if mmsize == 16
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
row_x86.asm | 34 mov%2 m1, [src_yuy2q + mmsize] 35 lea src_yuy2q, [src_yuy2q + mmsize * 2] 47 sub pixd, mmsize 49 lea dst_yq, [dst_yq + mmsize] 80 mov%1 m1, [src_uvq + mmsize] 81 lea src_uvq, [src_uvq + mmsize * 2] 94 lea dst_uq, [dst_uq + mmsize] 95 sub pixd, mmsize 120 lea src_uq, [src_uq + mmsize] 127 mov%1 [dst_uvq + mmsize], m [all...] |
x86inc.asm | 334 %if mmsize == 8 373 %define has_epilogue regs_used > 7 || xmm_regs_used > 6 || mmsize == 32 378 %if mmsize == 32 412 %define has_epilogue regs_used > 9 || mmsize == 32 416 %if mmsize == 32 455 %define has_epilogue regs_used > 3 || mmsize == 32 459 %if mmsize == 32 608 %if mmsize == 16 && notcpuflag(SSE2) 638 %define mmsize 8 661 %define mmsize 1 [all...] |
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_highbd_error_sse2.asm | 36 packssdw m0, [dqcq+sizeq*4+mmsize] 38 packssdw m2, [uqcq+sizeq*4+mmsize] 40 mova m1, [dqcq+sizeq*4+mmsize*2] 41 packssdw m1, [dqcq+sizeq*4+mmsize*3] 42 mova m3, [uqcq+sizeq*4+mmsize*2] 43 packssdw m3, [uqcq+sizeq*4+mmsize*3] 45 add sizeq, mmsize
|
vp9_error_sse2.asm | 31 mova m3, [uqcq+sizeq*2+mmsize] 32 mova m1, [dqcq+sizeq*2+mmsize] 58 add sizeq, mmsize 92 mova m3, [uqcq+sizeq*2+mmsize] 93 mova m1, [dqcq+sizeq*2+mmsize] 109 add sizeq, mmsize
|
vp9_quantize_ssse3_x86_64.asm | 101 add ncoeffq, mmsize 152 add ncoeffq, mmsize 161 add ncoeffq, mmsize 193 add ncoeffq, mmsize
|
/external/libvpx/libvpx/third_party/x86inc/ |
x86inc.asm | 413 %define required_stack_alignment ((mmsize + 15) & ~15) 425 %if mmsize != 8 508 %if mmsize != 8 && stack_size == 0 575 %define has_epilogue regs_used > 7 || xmm_regs_used > 6 || mmsize == 32 || stack_size > 0 580 %if mmsize == 32 616 %define has_epilogue regs_used > 9 || mmsize == 32 || stack_size > 0 627 %if mmsize == 32 672 %define has_epilogue regs_used > 3 || mmsize == 32 || stack_size > 0 683 %if mmsize == 32 906 %if (mmsize == 16 && notcpuflag(sse2)) || (mmsize == 32 && notcpuflag(avx2) [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/x86inc/ |
x86inc.asm | 422 %if mmsize == 8 689 %if mmsize == 16 && notcpuflag(sse2) 719 %define mmsize 8 742 %define mmsize 16 771 %define mmsize 32 936 %define %%size mmsize
|
/external/dlmalloc/ |
malloc.c | 3843 size_t mmsize = mmap_align(nb + SIX_SIZE_T_SIZES + CHUNK_ALIGN_MASK); local [all...] |