HomeSort by relevance Sort by last modified time
    Searched refs:vld1_u8 (Results 1 - 25 of 66) sorted by null

1 2 3

  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/neon/
subtract_neon.c 28 d0u8 = vld1_u8(src_ptr);
30 d2u8 = vld1_u8(src_ptr);
32 d4u8 = vld1_u8(src_ptr);
34 d6u8 = vld1_u8(src_ptr);
36 d1u8 = vld1_u8(predictor);
38 d3u8 = vld1_u8(predictor);
40 d5u8 = vld1_u8(predictor);
42 d7u8 = vld1_u8(predictor);
121 d0u8 = vld1_u8(src_ptr);
123 d1u8 = vld1_u8(pred_ptr)
    [all...]
vp8_mse16x16_neon.c 91 d0u8 = vld1_u8(src_ptr);
93 d4u8 = vld1_u8(ref_ptr);
95 d1u8 = vld1_u8(src_ptr);
97 d5u8 = vld1_u8(ref_ptr);
99 d2u8 = vld1_u8(src_ptr);
101 d6u8 = vld1_u8(ref_ptr);
103 d3u8 = vld1_u8(src_ptr);
105 d7u8 = vld1_u8(ref_ptr);
denoising_neon.c 273 const uint8x8_t v_sig = vld1_u8(sig);
293 const uint8x8_t v_sig_lo = vld1_u8(sig);
294 const uint8x8_t v_sig_hi = vld1_u8(&sig[sig_stride]);
296 const uint8x8_t v_mc_running_avg_lo = vld1_u8(mc_running_avg);
298 vld1_u8(&mc_running_avg[mc_running_avg_stride]);
398 const uint8x8_t v_sig_lo = vld1_u8(sig);
399 const uint8x8_t v_sig_hi = vld1_u8(&sig[sig_stride]);
401 const uint8x8_t v_mc_running_avg_lo = vld1_u8(mc_running_avg);
403 vld1_u8(&mc_running_avg[mc_running_avg_stride]);
421 const uint8x8_t v_running_avg_lo = vld1_u8(running_avg)
    [all...]
  /external/libhevc/common/arm/
ihevc_inter_pred_luma_copy_w16out.s 102 vld1.8 {d0},[r0] @vld1_u8(pu1_src_tmp)
104 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp)
108 vld1.8 {d22},[r5],r2 @vld1_u8(pu1_src_tmp)
112 vmovl.u8 q11,d22 @vmovl_u8(vld1_u8(pu1_src_tmp)
113 vld1.8 {d24},[r5],r2 @vld1_u8(pu1_src_tmp)
115 vmovl.u8 q12,d24 @vmovl_u8(vld1_u8(pu1_src_tmp)
118 vld1.8 {d26},[r5],r2 @vld1_u8(pu1_src_tmp)
120 vmovl.u8 q13,d26 @vmovl_u8(vld1_u8(pu1_src_tmp)
148 vld1.8 {d8},[r0]! @vld1_u8(pu1_src_tmp)
149 vld1.8 {d10},[r6],r2 @vld1_u8(pu1_src_tmp
    [all...]
ihevc_inter_pred_chroma_copy_w16out.s 132 vld1.8 {d0},[r0] @vld1_u8(pu1_src_tmp)
134 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp)
138 vld1.8 {d22},[r5],r2 @vld1_u8(pu1_src_tmp)
142 vmovl.u8 q11,d22 @vmovl_u8(vld1_u8(pu1_src_tmp)
143 vld1.8 {d24},[r5],r2 @vld1_u8(pu1_src_tmp)
145 vmovl.u8 q12,d24 @vmovl_u8(vld1_u8(pu1_src_tmp)
148 vld1.8 {d26},[r5],r2 @vld1_u8(pu1_src_tmp)
150 vmovl.u8 q13,d26 @vmovl_u8(vld1_u8(pu1_src_tmp)
173 vld1.8 {d0},[r0] @vld1_u8(pu1_src_tmp)
175 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp
    [all...]
ihevc_inter_pred_chroma_copy.s 183 vld1.8 {d0},[r0]! @vld1_u8(pu1_src_tmp)
186 vld1.8 {d1},[r5],r2 @vld1_u8(pu1_src_tmp)
189 vld1.8 {d2},[r5],r2 @vld1_u8(pu1_src_tmp)
191 vld1.8 {d3},[r5],r2 @vld1_u8(pu1_src_tmp)
210 vld1.8 {d0},[r0]! @vld1_u8(pu1_src_tmp)
213 vld1.8 {d1},[r5],r2 @vld1_u8(pu1_src_tmp)
230 vld1.8 {q0},[r0]! @vld1_u8(pu1_src_tmp)
233 vld1.8 {q1},[r5],r2 @vld1_u8(pu1_src_tmp)
236 vld1.8 {q2},[r5],r2 @vld1_u8(pu1_src_tmp)
238 vld1.8 {q3},[r5],r2 @vld1_u8(pu1_src_tmp
    [all...]
ihevc_inter_pred_filters_luma_vert_w16inp.s 144 vld1.16 {d1},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@
145 vld1.16 {d0},[r0]! @src_tmp1 = vld1_u8(pu1_src_tmp)@
147 vld1.16 {d2},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@
149 vld1.16 {d3},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@
151 vld1.16 {d4},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@
153 vld1.16 {d5},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@
155 vld1.16 {d6},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@
157 vld1.16 {d7},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@
162 vld1.16 {d16},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@
169 vld1.16 {d17},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)
    [all...]
ihevc_inter_pred_luma_vert_w16inp_w16out.s 154 vld1.16 {d1},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@
155 vld1.16 {d0},[r0]! @src_tmp1 = vld1_u8(pu1_src_tmp)@
157 vld1.16 {d2},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@
159 vld1.16 {d3},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@
161 vld1.16 {d4},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@
163 vld1.16 {d5},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@
165 vld1.16 {d6},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@
167 vld1.16 {d7},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@
172 vld1.16 {d16},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@
179 vld1.16 {d17},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)
    [all...]
ihevc_inter_pred_luma_copy.s 136 vld1.8 {d0},[r0]! @vld1_u8(pu1_src_tmp)
139 vld1.8 {d1},[r5],r2 @vld1_u8(pu1_src_tmp)
142 vld1.8 {d2},[r5],r2 @vld1_u8(pu1_src_tmp)
144 vld1.8 {d3},[r5],r2 @vld1_u8(pu1_src_tmp)
165 vld1.8 {q0},[r0]! @vld1_u8(pu1_src_tmp)
168 vld1.8 {q1},[r5],r2 @vld1_u8(pu1_src_tmp)
171 vld1.8 {q2},[r5],r2 @vld1_u8(pu1_src_tmp)
173 vld1.8 {q3},[r5],r2 @vld1_u8(pu1_src_tmp)
ihevc_inter_pred_filters_luma_vert.s 154 vld1.u8 {d1},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@
155 vld1.u8 {d0},[r0]! @src_tmp1 = vld1_u8(pu1_src_tmp)@
157 vld1.u8 {d2},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@
159 vld1.u8 {d3},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@
161 vld1.u8 {d4},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@
163 vld1.u8 {d5},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@
165 vld1.u8 {d6},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@
167 vld1.u8 {d7},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@
169 vld1.u8 {d16},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@
171 vld1.u8 {d17},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)
    [all...]
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/
copymem_neon.c 22 vtmp = vld1_u8(src);
38 vtmp = vld1_u8(src);
loopfilter_neon.c 190 d6 = vld1_u8(u);
192 d7 = vld1_u8(v);
194 d8 = vld1_u8(u);
196 d9 = vld1_u8(v);
198 d10 = vld1_u8(u);
200 d11 = vld1_u8(v);
202 d12 = vld1_u8(u);
204 d13 = vld1_u8(v);
206 d14 = vld1_u8(u);
208 d15 = vld1_u8(v)
    [all...]
mbloopfilter_neon.c 224 d6 = vld1_u8(u);
226 d7 = vld1_u8(v);
228 d8 = vld1_u8(u);
230 d9 = vld1_u8(v);
232 d10 = vld1_u8(u);
234 d11 = vld1_u8(v);
236 d12 = vld1_u8(u);
238 d13 = vld1_u8(v);
240 d14 = vld1_u8(u);
242 d15 = vld1_u8(v)
    [all...]
sad_neon.c 25 d0 = vld1_u8(src_ptr);
27 d8 = vld1_u8(ref_ptr);
32 d0 = vld1_u8(src_ptr);
34 d8 = vld1_u8(ref_ptr);
59 d0 = vld1_u8(src_ptr);
61 d8 = vld1_u8(ref_ptr);
66 d0 = vld1_u8(src_ptr);
68 d8 = vld1_u8(ref_ptr);
92 d0 = vld1_u8(src_ptr);
94 d8 = vld1_u8(ref_ptr)
    [all...]
bilinearpredict_neon.c 58 d2u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
59 d3u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
60 d4u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
61 d5u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
62 d6u8 = vld1_u8(src_ptr);
146 d22u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
147 d23u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
148 d24u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
149 d25u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
150 d26u8 = vld1_u8(src_ptr)
    [all...]
variance_neon.c 192 d0u8 = vld1_u8(src_ptr);
194 d2u8 = vld1_u8(src_ptr);
198 d4u8 = vld1_u8(ref_ptr);
200 d6u8 = vld1_u8(ref_ptr);
257 d0u8 = vld1_u8(src_ptr);
259 d1u8 = vld1_u8(src_ptr);
261 d2u8 = vld1_u8(src_ptr);
263 d3u8 = vld1_u8(src_ptr);
266 d4u8 = vld1_u8(ref_ptr);
268 d5u8 = vld1_u8(ref_ptr)
    [all...]
loopfiltersimpleverticaledge_neon.c 117 const uint8x8_t a = vld1_u8(src);
118 const uint8x8_t b = vld1_u8(src + pitch * 1);
119 const uint8x8_t c = vld1_u8(src + pitch * 2);
120 const uint8x8_t d = vld1_u8(src + pitch * 3);
121 const uint8x8_t e = vld1_u8(src + pitch * 4);
122 const uint8x8_t f = vld1_u8(src + pitch * 5);
123 const uint8x8_t g = vld1_u8(src + pitch * 6);
124 const uint8x8_t h = vld1_u8(src + pitch * 7);
vp8_subpixelvariance_neon.c 121 d2u8 = vld1_u8(src_ptr);
122 d3u8 = vld1_u8(src_ptr + 8);
123 d4u8 = vld1_u8(src_ptr + 16);
125 d5u8 = vld1_u8(src_ptr);
126 d6u8 = vld1_u8(src_ptr + 8);
127 d7u8 = vld1_u8(src_ptr + 16);
129 d8u8 = vld1_u8(src_ptr);
130 d9u8 = vld1_u8(src_ptr + 8);
131 d10u8 = vld1_u8(src_ptr + 16);
133 d11u8 = vld1_u8(src_ptr)
    [all...]
reconintra_neon.c 126 const uint8x8_t v_uabove = vld1_u8(uabove_row);
127 const uint8x8_t v_vabove = vld1_u8(vabove_row);
158 const uint8x8_t v_uabove = vld1_u8(uabove_row);
159 const uint8x8_t v_vabove = vld1_u8(vabove_row);
186 const uint8x8_t v_uabove = vld1_u8(uabove_row);
187 const uint8x8_t v_vabove = vld1_u8(vabove_row);
  /external/libvpx/libvpx/vp8/common/arm/neon/
copymem_neon.c 22 vtmp = vld1_u8(src);
38 vtmp = vld1_u8(src);
bilinearpredict_neon.c 55 d2u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
56 d3u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
57 d4u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
58 d5u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
59 d6u8 = vld1_u8(src_ptr);
143 d22u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
144 d23u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
145 d24u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
146 d25u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
147 d26u8 = vld1_u8(src_ptr)
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/
copymem_neon.c 22 vtmp = vld1_u8(src);
38 vtmp = vld1_u8(src);
bilinearpredict_neon.c 55 d2u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
56 d3u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
57 d4u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
58 d5u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
59 d6u8 = vld1_u8(src_ptr);
143 d22u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
144 d23u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
145 d24u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
146 d25u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line;
147 d26u8 = vld1_u8(src_ptr)
    [all...]
  /external/chromium_org/third_party/libwebp/dsp/
lossless_neon.c 77 const uint8x8_t shuffle = vld1_u8(kRGBAShuffle);
79 const uint8x8_t pixels = vld1_u8((uint8_t*)src);
95 const uint8x8_t shuffle0 = vld1_u8(kBGRShuffle[0]);
96 const uint8x8_t shuffle1 = vld1_u8(kBGRShuffle[1]);
97 const uint8x8_t shuffle2 = vld1_u8(kBGRShuffle[2]);
101 vld1_u8((const uint8_t*)(src + 0)),
102 vld1_u8((const uint8_t*)(src + 2)),
103 vld1_u8((const uint8_t*)(src + 4)),
104 vld1_u8((const uint8_t*)(src + 6)));
122 const uint8x8_t shuffle0 = vld1_u8(kRGBShuffle[0])
    [all...]
  /external/webp/src/dsp/
lossless_neon.c 77 const uint8x8_t shuffle = vld1_u8(kRGBAShuffle);
79 const uint8x8_t pixels = vld1_u8((uint8_t*)src);
95 const uint8x8_t shuffle0 = vld1_u8(kBGRShuffle[0]);
96 const uint8x8_t shuffle1 = vld1_u8(kBGRShuffle[1]);
97 const uint8x8_t shuffle2 = vld1_u8(kBGRShuffle[2]);
101 vld1_u8((const uint8_t*)(src + 0)),
102 vld1_u8((const uint8_t*)(src + 2)),
103 vld1_u8((const uint8_t*)(src + 4)),
104 vld1_u8((const uint8_t*)(src + 6)));
122 const uint8x8_t shuffle0 = vld1_u8(kRGBShuffle[0])
    [all...]

Completed in 232 milliseconds

1 2 3