/external/chromium_org/third_party/WebKit/Source/platform/graphics/cpu/arm/filters/ |
NEONHelpers.h | 40 uint16x4_t temporary2 = vget_low_u16(vmovl_u8(vreinterpret_u8_u32(temporary1)));
|
FECompositeArithmeticNEON.h | 53 uint16x4_t temporary2 = vget_low_u16(vmovl_u8(vreinterpret_u8_u32(temporary1))); 57 temporary2 = vget_low_u16(vmovl_u8(vreinterpret_u8_u32(temporary1)));
|
/external/chromium_org/third_party/skia/src/opts/ |
SkBitmapProcState_filter_neon.h | 43 tmp1 = vmull_u8(vreinterpret_u8_u32(va0), v16_y); // tmp1 = [a01|a00] * (16-y) 44 tmp2 = vmull_u8(vreinterpret_u8_u32(va1), vy); // tmp2 = [a11|a10] * y 78 tmp1 = vmull_u8(vreinterpret_u8_u32(va0), v16_y); // tmp1 = [a01|a00] * (16-y) 79 tmp2 = vmull_u8(vreinterpret_u8_u32(va1), vy); // tmp2 = [a11|a10] * y
|
SkBlitRow_opts_arm_neon.cpp | 767 src_raw = vreinterpret_u8_u32(vld1_u32(src)); 769 src_raw_2 = vreinterpret_u8_u32(vld1_u32(src+2)); 773 dst_raw = vreinterpret_u8_u32(vld1_u32(dst)); 775 dst_raw_2 = vreinterpret_u8_u32(vld1_u32(dst+2)); [all...] |
SkColor_opts_neon.h | 92 vsrc_wide = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(vdup_n_u32(src)))); 93 vdst_wide = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(vdup_n_u32(dst))));
|
SkMorphology_opts_neon.cpp | 44 uint8x8_t src_pixel = vreinterpret_u8_u32(vdup_n_u32(*p));
|
SkBlurImage_opts_neon.cpp | 33 return vreinterpret_u8_u32(vld1_lane_u32(src + srcStride, vld1_lane_u32(src, temp, 0), 1)); 109 uint8x8_t v8 = vreinterpret_u8_u32(vdup_n_u32(a));
|
SkXfermode_opts_arm_neon.cpp | 124 vsrc_wide = vmull_u8(vda, vreinterpret_u8_u32(vdup_n_u32(src))); 125 vdst_wide = vmull_u8(visa, vreinterpret_u8_u32(vdup_n_u32(dst))); 154 vsrc_wide = vmull_u8(vida, vreinterpret_u8_u32(vdup_n_u32(src))); 155 vdst_wide = vmull_u8(vsa, vreinterpret_u8_u32(vdup_n_u32(dst))); 185 vsrc = vreinterpret_u8_u32(vdup_n_u32(src)); 186 vdst = vreinterpret_u8_u32(vdup_n_u32(dst)); 210 vsrc = vreinterpret_u8_u32(vdup_n_u32(src)); 211 vdst = vreinterpret_u8_u32(vdup_n_u32(dst)); 222 vsrc = vreinterpret_u8_u32(vdup_n_u32(src)); 223 vdst = vreinterpret_u8_u32(vdup_n_u32(dst)) [all...] |
/external/skia/src/opts/ |
SkBitmapProcState_filter_neon.h | 43 tmp1 = vmull_u8(vreinterpret_u8_u32(va0), v16_y); // tmp1 = [a01|a00] * (16-y) 44 tmp2 = vmull_u8(vreinterpret_u8_u32(va1), vy); // tmp2 = [a11|a10] * y 78 tmp1 = vmull_u8(vreinterpret_u8_u32(va0), v16_y); // tmp1 = [a01|a00] * (16-y) 79 tmp2 = vmull_u8(vreinterpret_u8_u32(va1), vy); // tmp2 = [a11|a10] * y
|
SkBlitRow_opts_arm_neon.cpp | 669 src_raw = vreinterpret_u8_u32(vld1_u32(src)); 671 src_raw_2 = vreinterpret_u8_u32(vld1_u32(src+2)); 675 dst_raw = vreinterpret_u8_u32(vld1_u32(dst)); 677 dst_raw_2 = vreinterpret_u8_u32(vld1_u32(dst+2)); 786 src_raw = vreinterpret_u8_u32(vld1_u32(src)); 787 src_raw_2 = vreinterpret_u8_u32(vld1_u32(src+2)); 790 dst_raw = vreinterpret_u8_u32(vld1_u32(dst)); 791 dst_raw_2 = vreinterpret_u8_u32(vld1_u32(dst+2)); [all...] |
SkColor_opts_neon.h | 92 vsrc_wide = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(vdup_n_u32(src)))); 93 vdst_wide = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(vdup_n_u32(dst))));
|
SkMorphology_opts_neon.cpp | 44 uint8x8_t src_pixel = vreinterpret_u8_u32(vdup_n_u32(*p));
|
SkBlurImage_opts_neon.cpp | 33 return vreinterpret_u8_u32(vld1_lane_u32(src + srcStride, vld1_lane_u32(src, temp, 0), 1)); 109 uint8x8_t v8 = vreinterpret_u8_u32(vdup_n_u32(a));
|
SkXfermode_opts_arm_neon.cpp | 124 vsrc_wide = vmull_u8(vda, vreinterpret_u8_u32(vdup_n_u32(src))); 125 vdst_wide = vmull_u8(visa, vreinterpret_u8_u32(vdup_n_u32(dst))); 154 vsrc_wide = vmull_u8(vida, vreinterpret_u8_u32(vdup_n_u32(src))); 155 vdst_wide = vmull_u8(vsa, vreinterpret_u8_u32(vdup_n_u32(dst))); 185 vsrc = vreinterpret_u8_u32(vdup_n_u32(src)); 186 vdst = vreinterpret_u8_u32(vdup_n_u32(dst)); 210 vsrc = vreinterpret_u8_u32(vdup_n_u32(src)); 211 vdst = vreinterpret_u8_u32(vdup_n_u32(dst)); 222 vsrc = vreinterpret_u8_u32(vdup_n_u32(src)); 223 vdst = vreinterpret_u8_u32(vdup_n_u32(dst)) [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/ |
dc_only_idct_add_neon.c | 34 q1u16 = vaddw_u8(qAdd, vreinterpret_u8_u32(d2u32));
|
sixtappredict_neon.c | 86 d27u8 = vreinterpret_u8_u32(d27u32); 87 d28u8 = vreinterpret_u8_u32(d28u32); 88 d29u8 = vreinterpret_u8_u32(d29u32); 89 d30u8 = vreinterpret_u8_u32(d30u32); 90 d31u8 = vreinterpret_u8_u32(d31u32); 177 q7u16 = vmull_u8(vreinterpret_u8_u32(d0u32x2.val[0]), d5u8); 178 q8u16 = vmull_u8(vreinterpret_u8_u32(d1u32x2.val[0]), d5u8); 190 q7u16 = vmlal_u8(q7u16, vreinterpret_u8_u32(d0u32x2.val[0]), d0u8); 191 q8u16 = vmlal_u8(q8u16, vreinterpret_u8_u32(d1u32x2.val[0]), d0u8); 199 q7u16 = vmlsl_u8(q7u16, vreinterpret_u8_u32(d0u32x2.val[0]), d1u8) [all...] |
shortidct4x4llm_neon.c | 114 vreinterpret_u8_u32(d6u32));
|
bilinearpredict_neon.c | 54 d28u8 = vreinterpret_u8_u32(d28u32); 55 d29u8 = vreinterpret_u8_u32(d29u32); 56 d30u8 = vreinterpret_u8_u32(d30u32); 83 q7u16 = vmull_u8(vreinterpret_u8_u32(d0u32x2.val[0]), d0u8); 84 q8u16 = vmull_u8(vreinterpret_u8_u32(d1u32x2.val[0]), d0u8); 87 q7u16 = vmlal_u8(q7u16, vreinterpret_u8_u32(d2u32x2.val[0]), d1u8); 88 q8u16 = vmlal_u8(q8u16, vreinterpret_u8_u32(d3u32x2.val[0]), d1u8);
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
dc_only_idct_add_neon.c | 34 q1u16 = vaddw_u8(qAdd, vreinterpret_u8_u32(d2u32));
|
bilinearpredict_neon.c | 51 d28u8 = vreinterpret_u8_u32(d28u32); 52 d29u8 = vreinterpret_u8_u32(d29u32); 53 d30u8 = vreinterpret_u8_u32(d30u32); 80 q7u16 = vmull_u8(vreinterpret_u8_u32(d0u32x2.val[0]), d0u8); 81 q8u16 = vmull_u8(vreinterpret_u8_u32(d1u32x2.val[0]), d0u8); 84 q7u16 = vmlal_u8(q7u16, vreinterpret_u8_u32(d2u32x2.val[0]), d1u8); 85 q8u16 = vmlal_u8(q8u16, vreinterpret_u8_u32(d3u32x2.val[0]), d1u8);
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/ |
dc_only_idct_add_neon.c | 34 q1u16 = vaddw_u8(qAdd, vreinterpret_u8_u32(d2u32));
|
/external/libhevc/common/arm/ |
ihevc_intra_pred_filters_neon_intr.c | [all...] |
ihevc_inter_pred_chroma_vert.s | 143 vmull.u8 q3,d9,d1 @vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1) 145 vmlsl.u8 q3,d5,d0 @vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_0) 147 vmlal.u8 q3,d4,d2 @vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_2) 188 vmull.u8 q2,d7,d1 @vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1)
|
ihevc_inter_pred_chroma_vert_w16out.s | 145 vmull.u8 q3,d9,d1 @vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1) 147 vmlsl.u8 q3,d5,d0 @vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_0) 149 vmlal.u8 q3,d4,d2 @vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_2) 188 vmull.u8 q2,d7,d1 @vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1)
|
ihevc_inter_pred_filters_luma_vert.s | 433 vmull.u8 q0,d5,d23 @mul_res1 = vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1)@ 438 vmlsl.u8 q0,d4,d22 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_0)@ 442 vmlsl.u8 q0,d6,d24 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_2)@ 446 vmull.u8 q1,d7,d25 @mul_res2 = vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3)@ 449 vmlal.u8 q0,d4,d26 @mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp1), coeffabs_4)@ 454 vmlsl.u8 q1,d5,d27 @mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp2), coeffabs_5)@ 459 vmlal.u8 q0,d6,d28 @mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp3), coeffabs_6)@ 464 vmlsl.u8 q1,d7,d29 @mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp4), coeffabs_7)@ [all...] |