/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/neon/ |
subtract_neon.c | 28 d0u8 = vld1_u8(src_ptr); 30 d2u8 = vld1_u8(src_ptr); 32 d4u8 = vld1_u8(src_ptr); 34 d6u8 = vld1_u8(src_ptr); 36 d1u8 = vld1_u8(predictor); 38 d3u8 = vld1_u8(predictor); 40 d5u8 = vld1_u8(predictor); 42 d7u8 = vld1_u8(predictor); 121 d0u8 = vld1_u8(src_ptr); 123 d1u8 = vld1_u8(pred_ptr) [all...] |
vp8_mse16x16_neon.c | 91 d0u8 = vld1_u8(src_ptr); 93 d4u8 = vld1_u8(ref_ptr); 95 d1u8 = vld1_u8(src_ptr); 97 d5u8 = vld1_u8(ref_ptr); 99 d2u8 = vld1_u8(src_ptr); 101 d6u8 = vld1_u8(ref_ptr); 103 d3u8 = vld1_u8(src_ptr); 105 d7u8 = vld1_u8(ref_ptr);
|
denoising_neon.c | 273 const uint8x8_t v_sig = vld1_u8(sig); 293 const uint8x8_t v_sig_lo = vld1_u8(sig); 294 const uint8x8_t v_sig_hi = vld1_u8(&sig[sig_stride]); 296 const uint8x8_t v_mc_running_avg_lo = vld1_u8(mc_running_avg); 298 vld1_u8(&mc_running_avg[mc_running_avg_stride]); 398 const uint8x8_t v_sig_lo = vld1_u8(sig); 399 const uint8x8_t v_sig_hi = vld1_u8(&sig[sig_stride]); 401 const uint8x8_t v_mc_running_avg_lo = vld1_u8(mc_running_avg); 403 vld1_u8(&mc_running_avg[mc_running_avg_stride]); 421 const uint8x8_t v_running_avg_lo = vld1_u8(running_avg) [all...] |
/external/libhevc/common/arm/ |
ihevc_inter_pred_luma_copy_w16out.s | 102 vld1.8 {d0},[r0] @vld1_u8(pu1_src_tmp) 104 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp) 108 vld1.8 {d22},[r5],r2 @vld1_u8(pu1_src_tmp) 112 vmovl.u8 q11,d22 @vmovl_u8(vld1_u8(pu1_src_tmp) 113 vld1.8 {d24},[r5],r2 @vld1_u8(pu1_src_tmp) 115 vmovl.u8 q12,d24 @vmovl_u8(vld1_u8(pu1_src_tmp) 118 vld1.8 {d26},[r5],r2 @vld1_u8(pu1_src_tmp) 120 vmovl.u8 q13,d26 @vmovl_u8(vld1_u8(pu1_src_tmp) 148 vld1.8 {d8},[r0]! @vld1_u8(pu1_src_tmp) 149 vld1.8 {d10},[r6],r2 @vld1_u8(pu1_src_tmp [all...] |
ihevc_inter_pred_chroma_copy_w16out.s | 132 vld1.8 {d0},[r0] @vld1_u8(pu1_src_tmp) 134 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp) 138 vld1.8 {d22},[r5],r2 @vld1_u8(pu1_src_tmp) 142 vmovl.u8 q11,d22 @vmovl_u8(vld1_u8(pu1_src_tmp) 143 vld1.8 {d24},[r5],r2 @vld1_u8(pu1_src_tmp) 145 vmovl.u8 q12,d24 @vmovl_u8(vld1_u8(pu1_src_tmp) 148 vld1.8 {d26},[r5],r2 @vld1_u8(pu1_src_tmp) 150 vmovl.u8 q13,d26 @vmovl_u8(vld1_u8(pu1_src_tmp) 173 vld1.8 {d0},[r0] @vld1_u8(pu1_src_tmp) 175 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp [all...] |
ihevc_inter_pred_chroma_copy.s | 183 vld1.8 {d0},[r0]! @vld1_u8(pu1_src_tmp) 186 vld1.8 {d1},[r5],r2 @vld1_u8(pu1_src_tmp) 189 vld1.8 {d2},[r5],r2 @vld1_u8(pu1_src_tmp) 191 vld1.8 {d3},[r5],r2 @vld1_u8(pu1_src_tmp) 210 vld1.8 {d0},[r0]! @vld1_u8(pu1_src_tmp) 213 vld1.8 {d1},[r5],r2 @vld1_u8(pu1_src_tmp) 230 vld1.8 {q0},[r0]! @vld1_u8(pu1_src_tmp) 233 vld1.8 {q1},[r5],r2 @vld1_u8(pu1_src_tmp) 236 vld1.8 {q2},[r5],r2 @vld1_u8(pu1_src_tmp) 238 vld1.8 {q3},[r5],r2 @vld1_u8(pu1_src_tmp [all...] |
ihevc_inter_pred_filters_luma_vert_w16inp.s | 144 vld1.16 {d1},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@ 145 vld1.16 {d0},[r0]! @src_tmp1 = vld1_u8(pu1_src_tmp)@ 147 vld1.16 {d2},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@ 149 vld1.16 {d3},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@ 151 vld1.16 {d4},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@ 153 vld1.16 {d5},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@ 155 vld1.16 {d6},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@ 157 vld1.16 {d7},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@ 162 vld1.16 {d16},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@ 169 vld1.16 {d17},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp) [all...] |
ihevc_inter_pred_luma_vert_w16inp_w16out.s | 154 vld1.16 {d1},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@ 155 vld1.16 {d0},[r0]! @src_tmp1 = vld1_u8(pu1_src_tmp)@ 157 vld1.16 {d2},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@ 159 vld1.16 {d3},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@ 161 vld1.16 {d4},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@ 163 vld1.16 {d5},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@ 165 vld1.16 {d6},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@ 167 vld1.16 {d7},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@ 172 vld1.16 {d16},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@ 179 vld1.16 {d17},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp) [all...] |
ihevc_inter_pred_luma_copy.s | 136 vld1.8 {d0},[r0]! @vld1_u8(pu1_src_tmp) 139 vld1.8 {d1},[r5],r2 @vld1_u8(pu1_src_tmp) 142 vld1.8 {d2},[r5],r2 @vld1_u8(pu1_src_tmp) 144 vld1.8 {d3},[r5],r2 @vld1_u8(pu1_src_tmp) 165 vld1.8 {q0},[r0]! @vld1_u8(pu1_src_tmp) 168 vld1.8 {q1},[r5],r2 @vld1_u8(pu1_src_tmp) 171 vld1.8 {q2},[r5],r2 @vld1_u8(pu1_src_tmp) 173 vld1.8 {q3},[r5],r2 @vld1_u8(pu1_src_tmp)
|
ihevc_inter_pred_filters_luma_vert.s | 154 vld1.u8 {d1},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@ 155 vld1.u8 {d0},[r0]! @src_tmp1 = vld1_u8(pu1_src_tmp)@ 157 vld1.u8 {d2},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@ 159 vld1.u8 {d3},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@ 161 vld1.u8 {d4},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@ 163 vld1.u8 {d5},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp)@ 165 vld1.u8 {d6},[r3],r2 @src_tmp3 = vld1_u8(pu1_src_tmp)@ 167 vld1.u8 {d7},[r3],r2 @src_tmp4 = vld1_u8(pu1_src_tmp)@ 169 vld1.u8 {d16},[r3],r2 @src_tmp1 = vld1_u8(pu1_src_tmp)@ 171 vld1.u8 {d17},[r3],r2 @src_tmp2 = vld1_u8(pu1_src_tmp) [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/ |
copymem_neon.c | 22 vtmp = vld1_u8(src); 38 vtmp = vld1_u8(src);
|
loopfilter_neon.c | 190 d6 = vld1_u8(u); 192 d7 = vld1_u8(v); 194 d8 = vld1_u8(u); 196 d9 = vld1_u8(v); 198 d10 = vld1_u8(u); 200 d11 = vld1_u8(v); 202 d12 = vld1_u8(u); 204 d13 = vld1_u8(v); 206 d14 = vld1_u8(u); 208 d15 = vld1_u8(v) [all...] |
mbloopfilter_neon.c | 224 d6 = vld1_u8(u); 226 d7 = vld1_u8(v); 228 d8 = vld1_u8(u); 230 d9 = vld1_u8(v); 232 d10 = vld1_u8(u); 234 d11 = vld1_u8(v); 236 d12 = vld1_u8(u); 238 d13 = vld1_u8(v); 240 d14 = vld1_u8(u); 242 d15 = vld1_u8(v) [all...] |
sad_neon.c | 25 d0 = vld1_u8(src_ptr); 27 d8 = vld1_u8(ref_ptr); 32 d0 = vld1_u8(src_ptr); 34 d8 = vld1_u8(ref_ptr); 59 d0 = vld1_u8(src_ptr); 61 d8 = vld1_u8(ref_ptr); 66 d0 = vld1_u8(src_ptr); 68 d8 = vld1_u8(ref_ptr); 92 d0 = vld1_u8(src_ptr); 94 d8 = vld1_u8(ref_ptr) [all...] |
bilinearpredict_neon.c | 58 d2u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 59 d3u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 60 d4u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 61 d5u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 62 d6u8 = vld1_u8(src_ptr); 146 d22u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 147 d23u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 148 d24u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 149 d25u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 150 d26u8 = vld1_u8(src_ptr) [all...] |
variance_neon.c | 192 d0u8 = vld1_u8(src_ptr); 194 d2u8 = vld1_u8(src_ptr); 198 d4u8 = vld1_u8(ref_ptr); 200 d6u8 = vld1_u8(ref_ptr); 257 d0u8 = vld1_u8(src_ptr); 259 d1u8 = vld1_u8(src_ptr); 261 d2u8 = vld1_u8(src_ptr); 263 d3u8 = vld1_u8(src_ptr); 266 d4u8 = vld1_u8(ref_ptr); 268 d5u8 = vld1_u8(ref_ptr) [all...] |
loopfiltersimpleverticaledge_neon.c | 117 const uint8x8_t a = vld1_u8(src); 118 const uint8x8_t b = vld1_u8(src + pitch * 1); 119 const uint8x8_t c = vld1_u8(src + pitch * 2); 120 const uint8x8_t d = vld1_u8(src + pitch * 3); 121 const uint8x8_t e = vld1_u8(src + pitch * 4); 122 const uint8x8_t f = vld1_u8(src + pitch * 5); 123 const uint8x8_t g = vld1_u8(src + pitch * 6); 124 const uint8x8_t h = vld1_u8(src + pitch * 7);
|
vp8_subpixelvariance_neon.c | 121 d2u8 = vld1_u8(src_ptr); 122 d3u8 = vld1_u8(src_ptr + 8); 123 d4u8 = vld1_u8(src_ptr + 16); 125 d5u8 = vld1_u8(src_ptr); 126 d6u8 = vld1_u8(src_ptr + 8); 127 d7u8 = vld1_u8(src_ptr + 16); 129 d8u8 = vld1_u8(src_ptr); 130 d9u8 = vld1_u8(src_ptr + 8); 131 d10u8 = vld1_u8(src_ptr + 16); 133 d11u8 = vld1_u8(src_ptr) [all...] |
reconintra_neon.c | 126 const uint8x8_t v_uabove = vld1_u8(uabove_row); 127 const uint8x8_t v_vabove = vld1_u8(vabove_row); 158 const uint8x8_t v_uabove = vld1_u8(uabove_row); 159 const uint8x8_t v_vabove = vld1_u8(vabove_row); 186 const uint8x8_t v_uabove = vld1_u8(uabove_row); 187 const uint8x8_t v_vabove = vld1_u8(vabove_row);
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
copymem_neon.c | 22 vtmp = vld1_u8(src); 38 vtmp = vld1_u8(src);
|
bilinearpredict_neon.c | 55 d2u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 56 d3u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 57 d4u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 58 d5u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 59 d6u8 = vld1_u8(src_ptr); 143 d22u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 144 d23u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 145 d24u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 146 d25u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 147 d26u8 = vld1_u8(src_ptr) [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/ |
copymem_neon.c | 22 vtmp = vld1_u8(src); 38 vtmp = vld1_u8(src);
|
bilinearpredict_neon.c | 55 d2u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 56 d3u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 57 d4u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 58 d5u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 59 d6u8 = vld1_u8(src_ptr); 143 d22u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 144 d23u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 145 d24u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 146 d25u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; 147 d26u8 = vld1_u8(src_ptr) [all...] |
/external/chromium_org/third_party/libwebp/dsp/ |
lossless_neon.c | 77 const uint8x8_t shuffle = vld1_u8(kRGBAShuffle); 79 const uint8x8_t pixels = vld1_u8((uint8_t*)src); 95 const uint8x8_t shuffle0 = vld1_u8(kBGRShuffle[0]); 96 const uint8x8_t shuffle1 = vld1_u8(kBGRShuffle[1]); 97 const uint8x8_t shuffle2 = vld1_u8(kBGRShuffle[2]); 101 vld1_u8((const uint8_t*)(src + 0)), 102 vld1_u8((const uint8_t*)(src + 2)), 103 vld1_u8((const uint8_t*)(src + 4)), 104 vld1_u8((const uint8_t*)(src + 6))); 122 const uint8x8_t shuffle0 = vld1_u8(kRGBShuffle[0]) [all...] |
/external/webp/src/dsp/ |
lossless_neon.c | 77 const uint8x8_t shuffle = vld1_u8(kRGBAShuffle); 79 const uint8x8_t pixels = vld1_u8((uint8_t*)src); 95 const uint8x8_t shuffle0 = vld1_u8(kBGRShuffle[0]); 96 const uint8x8_t shuffle1 = vld1_u8(kBGRShuffle[1]); 97 const uint8x8_t shuffle2 = vld1_u8(kBGRShuffle[2]); 101 vld1_u8((const uint8_t*)(src + 0)), 102 vld1_u8((const uint8_t*)(src + 2)), 103 vld1_u8((const uint8_t*)(src + 4)), 104 vld1_u8((const uint8_t*)(src + 6))); 122 const uint8x8_t shuffle0 = vld1_u8(kRGBShuffle[0]) [all...] |