/external/libvpx/libvpx/vpx_dsp/arm/ |
avg_neon.c | 191 const uint8x16_t a01 = vcombine_u8(vld1_u8(a), vld1_u8(a + a_stride)); 193 vcombine_u8(vld1_u8(a + 2 * a_stride), vld1_u8(a + 3 * a_stride)); 195 vcombine_u8(vld1_u8(a + 4 * a_stride), vld1_u8(a + 5 * a_stride)); 197 vcombine_u8(vld1_u8(a + 6 * a_stride), vld1_u8(a + 7 * a_stride)); 199 const uint8x16_t b01 = vcombine_u8(vld1_u8(b), vld1_u8(b + b_stride)); 201 vcombine_u8(vld1_u8(b + 2 * b_stride), vld1_u8(b + 3 * b_stride)); 203 vcombine_u8(vld1_u8(b + 4 * b_stride), vld1_u8(b + 5 * b_stride)); 205 vcombine_u8(vld1_u8(b + 6 * b_stride), vld1_u8(b + 7 * b_stride));
|
avg_pred_neon.c | 39 r = vcombine_u8(r_0, r_1);
|
vpx_convolve_avg_neon.c | 58 s01 = vcombine_u8(s0, s1); 59 d01 = vcombine_u8(d0, d1);
|
transpose_neon.h | 51 b0.val[0] = vcombine_u8(vreinterpret_u8_u32(vget_low_u32(a0)), 53 b0.val[1] = vcombine_u8(vreinterpret_u8_u32(vget_high_u32(a0)), 536 vtrnq_u8(vcombine_u8(*a0, *a4), vcombine_u8(*a1, *a5)); 538 vtrnq_u8(vcombine_u8(*a2, *a6), vcombine_u8(*a3, *a7)); [all...] |
vpx_convolve8_neon.c | 388 vrhaddq_u8(vreinterpretq_u8_u32(d0123), vcombine_u8(d01, d23))); 473 vrhaddq_u8(vreinterpretq_u8_u32(d0415), vcombine_u8(t0, t1))); 475 vrhaddq_u8(vreinterpretq_u8_u32(d2637), vcombine_u8(t2, t3))); 562 d01 = vcombine_u8(vld1_u8(d + 0 * dst_stride), 564 d23 = vcombine_u8(vld1_u8(d + 2 * dst_stride), 566 d45 = vcombine_u8(vld1_u8(d + 4 * dst_stride), 568 d67 = vcombine_u8(vld1_u8(d + 6 * dst_stride), 570 d01 = vrhaddq_u8(d01, vcombine_u8(t0, t1)); 571 d23 = vrhaddq_u8(d23, vcombine_u8(t2, t3)); 572 d45 = vrhaddq_u8(d45, vcombine_u8(t4, t5)) [all...] |
intrapred_neon.c | 95 const uint8x16_t above_and_left = vcombine_u8(above_u8, left_u8); 420 const uint8x16_t L76543210XA0123456 = vcombine_u8(L76543210, XA0123456); 421 const uint8x16_t L6543210XA01234567 = vcombine_u8(L6543210X, A01234567); 422 const uint8x16_t L543210XA01234567_ = vcombine_u8(L543210XA0, A1234567_); 483 const uint8x16_t Lfedcba9876543210 = vcombine_u8(Lfedcba98, L76543210); 499 const uint8x16_t r_7 = vcombine_u8(vget_high_u8(row_0), vget_low_u8(row_1)); 535 const uint8x16_t LLfedcba9876543210 = vcombine_u8(LLfedcba98, LL76543210); 536 const uint8x16_t LUfedcba9876543210 = vcombine_u8(LUfedcba98, LU76543210); [all...] |
subpel_variance_neon.c | 91 vst1q_u8(output_ptr + j, vcombine_u8(out_lo, out_hi));
|
/external/skia/src/opts/ |
Sk4px_NEON.h | 22 return Sk16b(vcombine_u8(vaddhn_u16(this->fLo.fVec, o.fLo.fVec), 28 return Sk16b(vcombine_u8(vraddhn_u16(this->fLo.fVec, vrshrq_n_u16(this->fLo.fVec, 8)),
|
/external/skqp/src/opts/ |
Sk4px_NEON.h | 22 return Sk16b(vcombine_u8(vaddhn_u16(this->fLo.fVec, o.fLo.fVec), 28 return Sk16b(vcombine_u8(vraddhn_u16(this->fLo.fVec, vrshrq_n_u16(this->fLo.fVec, 8)),
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
mbloopfilter_neon.c | 248 q3 = vcombine_u8(d6, d7); 249 q4 = vcombine_u8(d8, d9); 250 q5 = vcombine_u8(d10, d11); 251 q6 = vcombine_u8(d12, d13); 252 q7 = vcombine_u8(d14, d15); 253 q8 = vcombine_u8(d16, d17); 254 q9 = vcombine_u8(d18, d19); 255 q10 = vcombine_u8(d20, d21); 337 q3 = vcombine_u8(d6, d7); 338 q4 = vcombine_u8(d8, d9) [all...] |
vp8_loopfilter_neon.c | 215 q3 = vcombine_u8(d6, d7); 216 q4 = vcombine_u8(d8, d9); 217 q5 = vcombine_u8(d10, d11); 218 q6 = vcombine_u8(d12, d13); 219 q7 = vcombine_u8(d14, d15); 220 q8 = vcombine_u8(d16, d17); 221 q9 = vcombine_u8(d18, d19); 222 q10 = vcombine_u8(d20, d21); 366 q3 = vcombine_u8(d6, d7); 367 q4 = vcombine_u8(d8, d9) [all...] |
bilinearpredict_neon.c | 68 a01 = vcombine_u8(a0, a1); 69 a23 = vcombine_u8(a2, a3); 99 store_unaligned_u8q(dst_ptr, dst_pitch, vcombine_u8(e0, e1)); 117 store_unaligned_u8q(dst_ptr, dst_pitch, vcombine_u8(f0, f1)); 441 q1u8 = vcombine_u8(d2u8, d3u8); 442 q2u8 = vcombine_u8(d4u8, d5u8); 443 q3u8 = vcombine_u8(d6u8, d7u8); 444 q4u8 = vcombine_u8(d8u8, d9u8); 520 q7u8 = vcombine_u8(d14u8, d15u8); 521 q8u8 = vcombine_u8(d16u8, d17u8) [all...] |
loopfiltersimpleverticaledge_neon.c | 196 q3u8 = vcombine_u8(d0u8x4.val[0], d1u8x4.val[0]); // d6 d10 197 q4u8 = vcombine_u8(d0u8x4.val[2], d1u8x4.val[2]); // d8 d12 198 q5u8 = vcombine_u8(d0u8x4.val[1], d1u8x4.val[1]); // d7 d11 199 q6u8 = vcombine_u8(d0u8x4.val[3], d1u8x4.val[3]); // d9 d13
|
/external/libhevc/encoder/arm/ |
ihevce_itrans_recon_neon.c | 120 a1 = vcombine_u8(a0, a0); 125 a4 = vcombine_u8(a2, a3); 148 a0 = vcombine_u8(a1, a1); 153 a4 = vcombine_u8(a2, a3); 157 a5 = vcombine_u8(a2, a3); 178 a1 = vcombine_u8(a0, a0); 203 a1 = vcombine_u8(a0, a0);
|
/external/clang/test/CodeGen/ |
aarch64-neon-vcombine.c | 39 return vcombine_u8(low, high);
|
/external/webp/src/dsp/ |
cost_neon.c | 29 const uint8x16_t eob = vcombine_u8(vqmovn_u16(eob_0), vqmovn_u16(eob_1)); 75 const uint8x16_t F = vcombine_u8(vqmovn_u16(E0), vqmovn_u16(E1));
|
lossless_enc_neon.c | 41 return vcombine_u8(vtbl1q_u8(argb, vget_low_u8(shuffle)), 50 return vcombine_u8(vtbl1_u8(vget_low_u8(argb), shuffle),
|
dec_neon.c | 55 *p1 = vcombine_u8(row0.val[0], row8.val[0]); 56 *p0 = vcombine_u8(row0.val[1], row8.val[1]); 57 *q0 = vcombine_u8(row0.val[2], row8.val[2]); 58 *q1 = vcombine_u8(row0.val[3], row8.val[3]); 149 *p3 = vcombine_u8(vld1_u8(u - 4 * stride), vld1_u8(v - 4 * stride)); 150 *p2 = vcombine_u8(vld1_u8(u - 3 * stride), vld1_u8(v - 3 * stride)); 151 *p1 = vcombine_u8(vld1_u8(u - 2 * stride), vld1_u8(v - 2 * stride)); 152 *p0 = vcombine_u8(vld1_u8(u - 1 * stride), vld1_u8(v - 1 * stride)); 153 *q0 = vcombine_u8(vld1_u8(u + 0 * stride), vld1_u8(v + 0 * stride)); 154 *q1 = vcombine_u8(vld1_u8(u + 1 * stride), vld1_u8(v + 1 * stride)) [all...] |
upsampling_neon.c | 124 vst1q_u8(out, vcombine_u8(rgba4444.val[0], rgba4444.val[1])); \ 132 vst1q_u8(out, vcombine_u8(rgb565.val[0], rgb565.val[1])); \
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
denoising_neon.c | 288 const uint8x16_t v_sig = vcombine_u8(v_sig_lo, v_sig_hi); 293 vcombine_u8(v_mc_running_avg_lo, v_mc_running_avg_hi); 388 const uint8x16_t v_sig = vcombine_u8(v_sig_lo, v_sig_hi); 393 vcombine_u8(v_mc_running_avg_lo, v_mc_running_avg_hi); 409 vcombine_u8(v_running_avg_lo, v_running_avg_hi);
|
/external/libaom/libaom/aom_dsp/arm/ |
subpel_variance_neon.c | 66 vst1q_u8(&output_ptr[j], vcombine_u8(out_lo, out_hi));
|
/external/libaom/libaom/av1/common/arm/ |
transpose_neon.h | 35 vtrnq_u8(vcombine_u8(*a0, *a4), vcombine_u8(*a1, *a5)); 37 vtrnq_u8(vcombine_u8(*a2, *a6), vcombine_u8(*a3, *a7));
|
blend_a64_hmask_neon.c | 57 res_q = vcombine_u8(vrshrn_n_u16(res_low, AOM_BLEND_A64_ROUND_BITS),
|
blend_a64_vmask_neon.c | 53 res_q = vcombine_u8(vrshrn_n_u16(res_low, AOM_BLEND_A64_ROUND_BITS),
|
/external/tensorflow/tensorflow/examples/android/jni/object_tracking/ |
image_neon.cc | 86 const uint8x16_t allpixels = vcombine_u8(tmp_pix1, tmp_pix2);
|