1 // Copyright 2011 Google Inc. All Rights Reserved. 2 // 3 // Use of this source code is governed by a BSD-style license 4 // that can be found in the COPYING file in the root of the source 5 // tree. An additional intellectual property rights grant can be found 6 // in the file PATENTS. All contributing project authors may 7 // be found in the AUTHORS file in the root of the source tree. 8 // ----------------------------------------------------------------------------- 9 // 10 // NEON version of YUV to RGB upsampling functions. 11 // 12 // Author: mans (at) mansr.com (Mans Rullgard) 13 // Based on SSE code by: somnath (at) google.com (Somnath Banerjee) 14 15 #include "./dsp.h" 16 17 #if defined(__cplusplus) || defined(c_plusplus) 18 extern "C" { 19 #endif 20 21 #if defined(WEBP_USE_NEON) 22 23 #include <assert.h> 24 #include <arm_neon.h> 25 #include <string.h> 26 #include "./yuv.h" 27 28 #ifdef FANCY_UPSAMPLING 29 30 // Loads 9 pixels each from rows r1 and r2 and generates 16 pixels. 31 #define UPSAMPLE_16PIXELS(r1, r2, out) { \ 32 uint8x8_t a = vld1_u8(r1); \ 33 uint8x8_t b = vld1_u8(r1 + 1); \ 34 uint8x8_t c = vld1_u8(r2); \ 35 uint8x8_t d = vld1_u8(r2 + 1); \ 36 \ 37 uint16x8_t al = vshll_n_u8(a, 1); \ 38 uint16x8_t bl = vshll_n_u8(b, 1); \ 39 uint16x8_t cl = vshll_n_u8(c, 1); \ 40 uint16x8_t dl = vshll_n_u8(d, 1); \ 41 \ 42 uint8x8_t diag1, diag2; \ 43 uint16x8_t sl; \ 44 \ 45 /* a + b + c + d */ \ 46 sl = vaddl_u8(a, b); \ 47 sl = vaddw_u8(sl, c); \ 48 sl = vaddw_u8(sl, d); \ 49 \ 50 al = vaddq_u16(sl, al); /* 3a + b + c + d */ \ 51 bl = vaddq_u16(sl, bl); /* a + 3b + c + d */ \ 52 \ 53 al = vaddq_u16(al, dl); /* 3a + b + c + 3d */ \ 54 bl = vaddq_u16(bl, cl); /* a + 3b + 3c + d */ \ 55 \ 56 diag2 = vshrn_n_u16(al, 3); \ 57 diag1 = vshrn_n_u16(bl, 3); \ 58 \ 59 a = vrhadd_u8(a, diag1); \ 60 b = vrhadd_u8(b, diag2); \ 61 c = vrhadd_u8(c, diag2); \ 62 d = vrhadd_u8(d, diag1); \ 63 \ 64 { \ 65 const uint8x8x2_t a_b = {{ a, b }}; \ 66 const uint8x8x2_t c_d = {{ c, d }}; \ 67 vst2_u8(out, a_b); \ 68 vst2_u8(out + 32, c_d); \ 69 } \ 70 } 71 72 // Turn the macro into a function for reducing code-size when non-critical 73 static void Upsample16Pixels(const uint8_t *r1, const uint8_t *r2, 74 uint8_t *out) { 75 UPSAMPLE_16PIXELS(r1, r2, out); 76 } 77 78 #define UPSAMPLE_LAST_BLOCK(tb, bb, num_pixels, out) { \ 79 uint8_t r1[9], r2[9]; \ 80 memcpy(r1, (tb), (num_pixels)); \ 81 memcpy(r2, (bb), (num_pixels)); \ 82 /* replicate last byte */ \ 83 memset(r1 + (num_pixels), r1[(num_pixels) - 1], 9 - (num_pixels)); \ 84 memset(r2 + (num_pixels), r2[(num_pixels) - 1], 9 - (num_pixels)); \ 85 Upsample16Pixels(r1, r2, out); \ 86 } 87 88 #define CY 76283 89 #define CVR 89858 90 #define CUG 22014 91 #define CVG 45773 92 #define CUB 113618 93 94 static const int16_t coef[4] = { CVR / 4, CUG, CVG / 2, CUB / 4 }; 95 96 #define CONVERT8(FMT, XSTEP, N, src_y, src_uv, out, cur_x) { \ 97 int i; \ 98 for (i = 0; i < N; i += 8) { \ 99 int off = ((cur_x) + i) * XSTEP; \ 100 uint8x8_t y = vld1_u8(src_y + (cur_x) + i); \ 101 uint8x8_t u = vld1_u8((src_uv) + i); \ 102 uint8x8_t v = vld1_u8((src_uv) + i + 16); \ 103 int16x8_t yy = vreinterpretq_s16_u16(vsubl_u8(y, u16)); \ 104 int16x8_t uu = vreinterpretq_s16_u16(vsubl_u8(u, u128)); \ 105 int16x8_t vv = vreinterpretq_s16_u16(vsubl_u8(v, u128)); \ 106 \ 107 int16x8_t ud = vshlq_n_s16(uu, 1); \ 108 int16x8_t vd = vshlq_n_s16(vv, 1); \ 109 \ 110 int32x4_t vrl = vqdmlal_lane_s16(vshll_n_s16(vget_low_s16(vv), 1), \ 111 vget_low_s16(vd), cf16, 0); \ 112 int32x4_t vrh = vqdmlal_lane_s16(vshll_n_s16(vget_high_s16(vv), 1), \ 113 vget_high_s16(vd), cf16, 0); \ 114 int16x8_t vr = vcombine_s16(vrshrn_n_s32(vrl, 16), \ 115 vrshrn_n_s32(vrh, 16)); \ 116 \ 117 int32x4_t vl = vmovl_s16(vget_low_s16(vv)); \ 118 int32x4_t vh = vmovl_s16(vget_high_s16(vv)); \ 119 int32x4_t ugl = vmlal_lane_s16(vl, vget_low_s16(uu), cf16, 1); \ 120 int32x4_t ugh = vmlal_lane_s16(vh, vget_high_s16(uu), cf16, 1); \ 121 int32x4_t gcl = vqdmlal_lane_s16(ugl, vget_low_s16(vv), cf16, 2); \ 122 int32x4_t gch = vqdmlal_lane_s16(ugh, vget_high_s16(vv), cf16, 2); \ 123 int16x8_t gc = vcombine_s16(vrshrn_n_s32(gcl, 16), \ 124 vrshrn_n_s32(gch, 16)); \ 125 \ 126 int32x4_t ubl = vqdmlal_lane_s16(vshll_n_s16(vget_low_s16(uu), 1), \ 127 vget_low_s16(ud), cf16, 3); \ 128 int32x4_t ubh = vqdmlal_lane_s16(vshll_n_s16(vget_high_s16(uu), 1), \ 129 vget_high_s16(ud), cf16, 3); \ 130 int16x8_t ub = vcombine_s16(vrshrn_n_s32(ubl, 16), \ 131 vrshrn_n_s32(ubh, 16)); \ 132 \ 133 int32x4_t rl = vaddl_s16(vget_low_s16(yy), vget_low_s16(vr)); \ 134 int32x4_t rh = vaddl_s16(vget_high_s16(yy), vget_high_s16(vr)); \ 135 int32x4_t gl = vsubl_s16(vget_low_s16(yy), vget_low_s16(gc)); \ 136 int32x4_t gh = vsubl_s16(vget_high_s16(yy), vget_high_s16(gc)); \ 137 int32x4_t bl = vaddl_s16(vget_low_s16(yy), vget_low_s16(ub)); \ 138 int32x4_t bh = vaddl_s16(vget_high_s16(yy), vget_high_s16(ub)); \ 139 \ 140 rl = vmulq_lane_s32(rl, cf32, 0); \ 141 rh = vmulq_lane_s32(rh, cf32, 0); \ 142 gl = vmulq_lane_s32(gl, cf32, 0); \ 143 gh = vmulq_lane_s32(gh, cf32, 0); \ 144 bl = vmulq_lane_s32(bl, cf32, 0); \ 145 bh = vmulq_lane_s32(bh, cf32, 0); \ 146 \ 147 y = vqmovun_s16(vcombine_s16(vrshrn_n_s32(rl, 16), \ 148 vrshrn_n_s32(rh, 16))); \ 149 u = vqmovun_s16(vcombine_s16(vrshrn_n_s32(gl, 16), \ 150 vrshrn_n_s32(gh, 16))); \ 151 v = vqmovun_s16(vcombine_s16(vrshrn_n_s32(bl, 16), \ 152 vrshrn_n_s32(bh, 16))); \ 153 STR_ ## FMT(out + off, y, u, v); \ 154 } \ 155 } 156 157 #define v255 vmov_n_u8(255) 158 159 #define STR_Rgb(out, r, g, b) do { \ 160 const uint8x8x3_t r_g_b = {{ r, g, b }}; \ 161 vst3_u8(out, r_g_b); \ 162 } while (0) 163 164 #define STR_Bgr(out, r, g, b) do { \ 165 const uint8x8x3_t b_g_r = {{ b, g, r }}; \ 166 vst3_u8(out, b_g_r); \ 167 } while (0) 168 169 #define STR_Rgba(out, r, g, b) do { \ 170 const uint8x8x4_t r_g_b_v255 = {{ r, g, b, v255 }}; \ 171 vst4_u8(out, r_g_b_v255); \ 172 } while (0) 173 174 #define STR_Bgra(out, r, g, b) do { \ 175 const uint8x8x4_t b_g_r_v255 = {{ b, g, r, v255 }}; \ 176 vst4_u8(out, b_g_r_v255); \ 177 } while (0) 178 179 #define CONVERT1(FMT, XSTEP, N, src_y, src_uv, rgb, cur_x) { \ 180 int i; \ 181 for (i = 0; i < N; i++) { \ 182 int off = ((cur_x) + i) * XSTEP; \ 183 int y = src_y[(cur_x) + i]; \ 184 int u = (src_uv)[i]; \ 185 int v = (src_uv)[i + 16]; \ 186 VP8YuvTo ## FMT(y, u, v, rgb + off); \ 187 } \ 188 } 189 190 #define CONVERT2RGB_8(FMT, XSTEP, top_y, bottom_y, uv, \ 191 top_dst, bottom_dst, cur_x, len) { \ 192 if (top_y) { \ 193 CONVERT8(FMT, XSTEP, len, top_y, uv, top_dst, cur_x) \ 194 } \ 195 if (bottom_y) { \ 196 CONVERT8(FMT, XSTEP, len, bottom_y, (uv) + 32, bottom_dst, cur_x) \ 197 } \ 198 } 199 200 #define CONVERT2RGB_1(FMT, XSTEP, top_y, bottom_y, uv, \ 201 top_dst, bottom_dst, cur_x, len) { \ 202 if (top_y) { \ 203 CONVERT1(FMT, XSTEP, len, top_y, uv, top_dst, cur_x); \ 204 } \ 205 if (bottom_y) { \ 206 CONVERT1(FMT, XSTEP, len, bottom_y, (uv) + 32, bottom_dst, cur_x); \ 207 } \ 208 } 209 210 #define NEON_UPSAMPLE_FUNC(FUNC_NAME, FMT, XSTEP) \ 211 static void FUNC_NAME(const uint8_t *top_y, const uint8_t *bottom_y, \ 212 const uint8_t *top_u, const uint8_t *top_v, \ 213 const uint8_t *cur_u, const uint8_t *cur_v, \ 214 uint8_t *top_dst, uint8_t *bottom_dst, int len) { \ 215 int block; \ 216 /* 16 byte aligned array to cache reconstructed u and v */ \ 217 uint8_t uv_buf[2 * 32 + 15]; \ 218 uint8_t *const r_uv = (uint8_t*)((uintptr_t)(uv_buf + 15) & ~15); \ 219 const int uv_len = (len + 1) >> 1; \ 220 /* 9 pixels must be read-able for each block */ \ 221 const int num_blocks = (uv_len - 1) >> 3; \ 222 const int leftover = uv_len - num_blocks * 8; \ 223 const int last_pos = 1 + 16 * num_blocks; \ 224 \ 225 const int u_diag = ((top_u[0] + cur_u[0]) >> 1) + 1; \ 226 const int v_diag = ((top_v[0] + cur_v[0]) >> 1) + 1; \ 227 \ 228 const int16x4_t cf16 = vld1_s16(coef); \ 229 const int32x2_t cf32 = vmov_n_s32(CY); \ 230 const uint8x8_t u16 = vmov_n_u8(16); \ 231 const uint8x8_t u128 = vmov_n_u8(128); \ 232 \ 233 /* Treat the first pixel in regular way */ \ 234 if (top_y) { \ 235 const int u0 = (top_u[0] + u_diag) >> 1; \ 236 const int v0 = (top_v[0] + v_diag) >> 1; \ 237 VP8YuvTo ## FMT(top_y[0], u0, v0, top_dst); \ 238 } \ 239 if (bottom_y) { \ 240 const int u0 = (cur_u[0] + u_diag) >> 1; \ 241 const int v0 = (cur_v[0] + v_diag) >> 1; \ 242 VP8YuvTo ## FMT(bottom_y[0], u0, v0, bottom_dst); \ 243 } \ 244 \ 245 for (block = 0; block < num_blocks; ++block) { \ 246 UPSAMPLE_16PIXELS(top_u, cur_u, r_uv); \ 247 UPSAMPLE_16PIXELS(top_v, cur_v, r_uv + 16); \ 248 CONVERT2RGB_8(FMT, XSTEP, top_y, bottom_y, r_uv, \ 249 top_dst, bottom_dst, 16 * block + 1, 16); \ 250 top_u += 8; \ 251 cur_u += 8; \ 252 top_v += 8; \ 253 cur_v += 8; \ 254 } \ 255 \ 256 UPSAMPLE_LAST_BLOCK(top_u, cur_u, leftover, r_uv); \ 257 UPSAMPLE_LAST_BLOCK(top_v, cur_v, leftover, r_uv + 16); \ 258 CONVERT2RGB_1(FMT, XSTEP, top_y, bottom_y, r_uv, \ 259 top_dst, bottom_dst, last_pos, len - last_pos); \ 260 } 261 262 // NEON variants of the fancy upsampler. 263 NEON_UPSAMPLE_FUNC(UpsampleRgbLinePairNEON, Rgb, 3) 264 NEON_UPSAMPLE_FUNC(UpsampleBgrLinePairNEON, Bgr, 3) 265 NEON_UPSAMPLE_FUNC(UpsampleRgbaLinePairNEON, Rgba, 4) 266 NEON_UPSAMPLE_FUNC(UpsampleBgraLinePairNEON, Bgra, 4) 267 268 #endif // FANCY_UPSAMPLING 269 270 #endif // WEBP_USE_NEON 271 272 //------------------------------------------------------------------------------ 273 274 extern WebPUpsampleLinePairFunc WebPUpsamplers[/* MODE_LAST */]; 275 276 void WebPInitUpsamplersNEON(void) { 277 #if defined(WEBP_USE_NEON) 278 WebPUpsamplers[MODE_RGB] = UpsampleRgbLinePairNEON; 279 WebPUpsamplers[MODE_RGBA] = UpsampleRgbaLinePairNEON; 280 WebPUpsamplers[MODE_BGR] = UpsampleBgrLinePairNEON; 281 WebPUpsamplers[MODE_BGRA] = UpsampleBgraLinePairNEON; 282 #endif // WEBP_USE_NEON 283 } 284 285 void WebPInitPremultiplyNEON(void) { 286 #if defined(WEBP_USE_NEON) 287 WebPUpsamplers[MODE_rgbA] = UpsampleRgbaLinePairNEON; 288 WebPUpsamplers[MODE_bgrA] = UpsampleBgraLinePairNEON; 289 #endif // WEBP_USE_NEON 290 } 291 292 #if defined(__cplusplus) || defined(c_plusplus) 293 } // extern "C" 294 #endif 295