/external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/arm/neon/ |
vp9_sad_neon.c | 53 vec_accum_lo = vabal_u8(vec_accum_lo, vget_low_u8(vec_src_00), 54 vget_low_u8(vec_ref_00)); 57 vec_accum_lo = vabal_u8(vec_accum_lo, vget_low_u8(vec_src_16), 58 vget_low_u8(vec_ref_16)); 61 vec_accum_lo = vabal_u8(vec_accum_lo, vget_low_u8(vec_src_32), 62 vget_low_u8(vec_ref_32)); 65 vec_accum_lo = vabal_u8(vec_accum_lo, vget_low_u8(vec_src_48), 66 vget_low_u8(vec_ref_48)); 86 vec_accum_lo = vabal_u8(vec_accum_lo, vget_low_u8(vec_src_00), 87 vget_low_u8(vec_ref_00)) [all...] |
vp9_subtract_neon.c | 30 const uint16x8_t v_diff_lo_00 = vsubl_u8(vget_low_u8(v_src_00), 31 vget_low_u8(v_pred_00)); 34 const uint16x8_t v_diff_lo_16 = vsubl_u8(vget_low_u8(v_src_16), 35 vget_low_u8(v_pred_16)); 51 const uint16x8_t v_diff_lo = vsubl_u8(vget_low_u8(v_src), 52 vget_low_u8(v_pred));
|
vp9_variance_neon.c | 146 const uint16x8_t a = vmull_u8(vget_low_u8(src_0), f0); 147 const uint16x8_t b = vmlal_u8(a, vget_low_u8(src_1), f1);
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/ |
bilinearpredict_neon.c | 74 d0u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q1u8)), 76 d1u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q2u8)), 161 q6u16 = vmull_u8(vget_low_u8(q1u8), d0u8); 162 q7u16 = vmull_u8(vget_low_u8(q2u8), d0u8); 163 q8u16 = vmull_u8(vget_low_u8(q3u8), d0u8); 164 q9u16 = vmull_u8(vget_low_u8(q4u8), d0u8); 165 q10u16 = vmull_u8(vget_low_u8(q5u8), d0u8); 167 d3u8 = vext_u8(vget_low_u8(q1u8), vget_high_u8(q1u8), 1); 168 d5u8 = vext_u8(vget_low_u8(q2u8), vget_high_u8(q2u8), 1); 169 d7u8 = vext_u8(vget_low_u8(q3u8), vget_high_u8(q3u8), 1) [all...] |
sixtappredict_neon.c | 164 d18u8 = vext_u8(vget_low_u8(q3u8), vget_high_u8(q3u8), 5); 165 d19u8 = vext_u8(vget_low_u8(q4u8), vget_high_u8(q4u8), 5); 166 d20u8 = vext_u8(vget_low_u8(q5u8), vget_high_u8(q5u8), 5); 167 d21u8 = vext_u8(vget_low_u8(q6u8), vget_high_u8(q6u8), 5); 170 q3u8 = vcombine_u8(vget_low_u8(q3u8), vget_low_u8(q4u8)); 171 q5u8 = vcombine_u8(vget_low_u8(q5u8), vget_low_u8(q6u8)); 184 d0u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q3u8)), // d6 d7 186 d1u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q5u8)), // d10 d1 [all...] |
sad_neon.c | 128 q12 = vabdl_u8(vget_low_u8(q0), vget_low_u8(q4)); 136 q12 = vabal_u8(q12, vget_low_u8(q0), vget_low_u8(q4)); 165 q12 = vabdl_u8(vget_low_u8(q0), vget_low_u8(q4)); 173 q12 = vabal_u8(q12, vget_low_u8(q0), vget_low_u8(q4));
|
vp8_subpixelvariance_neon.c | 73 q1u16 = vmull_u8(vget_low_u8(q11u8), d0u8); 75 q3u16 = vmull_u8(vget_low_u8(q12u8), d0u8); 77 q5u16 = vmull_u8(vget_low_u8(q13u8), d0u8); 79 q7u16 = vmull_u8(vget_low_u8(q14u8), d0u8); 82 q1u16 = vmlal_u8(q1u16, vget_low_u8(q12u8), d1u8); 84 q3u16 = vmlal_u8(q3u16, vget_low_u8(q13u8), d1u8); 86 q5u16 = vmlal_u8(q5u16, vget_low_u8(q14u8), d1u8); 88 q7u16 = vmlal_u8(q7u16, vget_low_u8(q15u8), d1u8); 378 q1u16 = vmull_u8(vget_low_u8(q11u8), d0u8); 380 q3u16 = vmull_u8(vget_low_u8(q12u8), d0u8) [all...] |
mbloopfilter_neon.c | 270 vst1_u8(u, vget_low_u8(q4)); 274 vst1_u8(u, vget_low_u8(q5)); 278 vst1_u8(u, vget_low_u8(q6)); 282 vst1_u8(u, vget_low_u8(q7)); 286 vst1_u8(u, vget_low_u8(q8)); 290 vst1_u8(u, vget_low_u8(q9)); 427 vst1_u8(s1, vget_low_u8(q3)); 431 vst1_u8(s1, vget_low_u8(q4)); 435 vst1_u8(s1, vget_low_u8(q5)); 439 vst1_u8(s1, vget_low_u8(q6)) [all...] |
variance_neon.c | 46 q11u16 = vsubl_u8(vget_low_u8(q0u8), vget_low_u8(q2u8)); 48 q13u16 = vsubl_u8(vget_low_u8(q1u8), vget_low_u8(q3u8)); 125 q11u16 = vsubl_u8(vget_low_u8(q0u8), vget_low_u8(q2u8)); 127 q13u16 = vsubl_u8(vget_low_u8(q1u8), vget_low_u8(q3u8));
|
loopfilter_neon.c | 85 q4u16 = vmovl_u8(vget_low_u8(q10)); 235 vst1_u8(u, vget_low_u8(q5)); 237 vst1_u8(u, vget_low_u8(q6)); 239 vst1_u8(u, vget_low_u8(q7)); 241 vst1_u8(u, vget_low_u8(q8)); 401 q4ResultL.val[0] = vget_low_u8(q5); // d10 402 q4ResultL.val[1] = vget_low_u8(q6); // d12 403 q4ResultL.val[2] = vget_low_u8(q7); // d14 404 q4ResultL.val[3] = vget_low_u8(q8); // d16 517 q4ResultL.val[0] = vget_low_u8(q5); // d1 [all...] |
loopfiltersimpleverticaledge_neon.c | 251 d2u8x2.val[0] = vget_low_u8(q6u8); // d12 252 d2u8x2.val[1] = vget_low_u8(q7u8); // d14
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
bilinearpredict_neon.c | 71 d0u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q1u8)), 73 d1u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q2u8)), 158 q6u16 = vmull_u8(vget_low_u8(q1u8), d0u8); 159 q7u16 = vmull_u8(vget_low_u8(q2u8), d0u8); 160 q8u16 = vmull_u8(vget_low_u8(q3u8), d0u8); 161 q9u16 = vmull_u8(vget_low_u8(q4u8), d0u8); 162 q10u16 = vmull_u8(vget_low_u8(q5u8), d0u8); 164 d3u8 = vext_u8(vget_low_u8(q1u8), vget_high_u8(q1u8), 1); 165 d5u8 = vext_u8(vget_low_u8(q2u8), vget_high_u8(q2u8), 1); 166 d7u8 = vext_u8(vget_low_u8(q3u8), vget_high_u8(q3u8), 1) [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/ |
bilinearpredict_neon.c | 71 d0u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q1u8)), 73 d1u32x2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(q2u8)), 158 q6u16 = vmull_u8(vget_low_u8(q1u8), d0u8); 159 q7u16 = vmull_u8(vget_low_u8(q2u8), d0u8); 160 q8u16 = vmull_u8(vget_low_u8(q3u8), d0u8); 161 q9u16 = vmull_u8(vget_low_u8(q4u8), d0u8); 162 q10u16 = vmull_u8(vget_low_u8(q5u8), d0u8); 164 d3u8 = vext_u8(vget_low_u8(q1u8), vget_high_u8(q1u8), 1); 165 d5u8 = vext_u8(vget_low_u8(q2u8), vget_high_u8(q2u8), 1); 166 d7u8 = vext_u8(vget_low_u8(q3u8), vget_high_u8(q3u8), 1) [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/neon/ |
subtract_neon.c | 80 q8u16 = vsubl_u8(vget_low_u8(q0u8), vget_low_u8(q1u8)); 82 q10u16 = vsubl_u8(vget_low_u8(q2u8), vget_low_u8(q3u8));
|
vp8_mse16x16_neon.c | 42 q11u16 = vsubl_u8(vget_low_u8(q0u8), vget_low_u8(q2u8)); 44 q13u16 = vsubl_u8(vget_low_u8(q1u8), vget_low_u8(q3u8));
|
/external/clang/test/CodeGen/ |
arm-neon-vget.c | 21 return vget_low_u8(a);
|
aarch64-neon-vget-hilo.c | 119 return vget_low_u8(a);
|
/external/chromium_org/third_party/skia/src/opts/ |
SkBitmapProcState_arm_neon.cpp | 134 int16x8_t p01_16 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pixels))); 174 int16x8_t p01_16 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pixels))); 231 int16x8_t src16_01 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(src8))); 299 int16x8_t src16_01 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(src8))); 346 vreinterpret_u32_u8(vget_low_u8(accum8))); 350 vreinterpret_u32_u8(vget_low_u8(accum8))); 430 p01_16 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pixels))); \
|
/external/skia/src/opts/ |
SkBitmapProcState_arm_neon.cpp | 134 int16x8_t p01_16 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pixels))); 174 int16x8_t p01_16 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pixels))); 231 int16x8_t src16_01 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(src8))); 299 int16x8_t src16_01 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(src8))); 346 vreinterpret_u32_u8(vget_low_u8(accum8))); 350 vreinterpret_u32_u8(vget_low_u8(accum8))); 430 p01_16 = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pixels))); \
|
/external/chromium_org/third_party/libwebp/dsp/ |
dec_neon.c | 256 lo.val[0] = vget_low_u8(p0); 257 lo.val[1] = vget_low_u8(q0); 282 vget_low_u8(p1), vget_low_u8(p0), 283 vget_low_u8(q0), vget_low_u8(q1)); 309 vst1_u8(u - stride, vget_low_u8(p0)); 310 vst1_u8(u, vget_low_u8(q0)); 338 INIT_VECTOR3(u0, vget_low_u8(p2), vget_low_u8(p1), vget_low_u8(p0)) [all...] |
enc_neon.c | 297 const int16x8_t D0D1 = DiffU8ToS16(vget_low_u8(S0), vget_low_u8(R0)); 654 const int32x4x4_t hpass_a = DistoHorizontalPass(vget_low_u8(d0d1u8), 655 vget_low_u8(d2d3u8)); [all...] |
/external/webp/src/dsp/ |
dec_neon.c | 256 lo.val[0] = vget_low_u8(p0); 257 lo.val[1] = vget_low_u8(q0); 282 vget_low_u8(p1), vget_low_u8(p0), 283 vget_low_u8(q0), vget_low_u8(q1)); 309 vst1_u8(u - stride, vget_low_u8(p0)); 310 vst1_u8(u, vget_low_u8(q0)); 338 INIT_VECTOR3(u0, vget_low_u8(p2), vget_low_u8(p1), vget_low_u8(p0)) [all...] |
enc_neon.c | 297 const int16x8_t D0D1 = DiffU8ToS16(vget_low_u8(S0), vget_low_u8(R0)); 654 const int32x4x4_t hpass_a = DistoHorizontalPass(vget_low_u8(d0d1u8), 655 vget_low_u8(d2d3u8)); [all...] |
/external/libhevc/common/arm/ |
ihevc_sao_edge_offset_class1.s | 156 VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 178 VMOVL.U8 Q10,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 237 VMOVL.U8 Q13,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 298 VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 315 VMOVL.U8 Q10,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 355 VMOVL.U8 Q13,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
|
ihevc_sao_edge_offset_class1_chroma.s | 161 VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 183 VMOVL.U8 Q10,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 255 VMOVL.U8 Q13,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 316 VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 340 VMOVL.U8 Q10,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 391 VMOVL.U8 Q13,D10 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
|