Home | History | Annotate | Download | only in opts

Lines Matching refs:src_scale

247                   "vmul.u16   q2, q14, d2[0]                  \n\t"   // sa * src_scale
248 "vmul.u16 q11, q11, d2[0] \n\t" // red result = src_red * src_scale
249 "vmul.u16 q12, q12, d2[0] \n\t" // grn result = src_grn * src_scale
250 "vmul.u16 q13, q13, d2[0] \n\t" // blu result = src_blu * src_scale
252 "vshr.u16 q2, q2, #8 \n\t" // sa * src_scale >> 8
253 "vsub.u16 q2, q3, q2 \n\t" // 255 - (sa * src_scale >> 8)
1208 "add %[alpha], %[alpha], #1 \n\t" /* loading %[alpha]=src_scale=alpha+1 */
1221 "smulbb r9, r9, %[alpha] \n\t" /* r9 = SkMulS16 r9 with src_scale */
1222 "smulbb r10, r10, %[alpha] \n\t" /* r10 = SkMulS16 r10 with src_scale */
1233 "mul r11, r11, %[alpha] \n\t" /* ag = r11 times src_scale */
1234 "mul r4, r4, %[alpha] \n\t" /* rb = r4 times src_scale */
1237 "orr r5, r11, r4 \n\t" /* r5 = (src1, src_scale) */
1257 "mul r11, r11, %[alpha] \n\t" /* ag = r11 times src_scale */
1258 "mul r4, r4, %[alpha] \n\t" /* rb = r4 times src_scale */
1261 "orr r6, r11, r4 \n\t" /* r6 = (src2, src_scale) */
1291 "smulbb r6, r6, %[alpha] \n\t" /* r6 = SkMulS16 with src_scale */
1297 /* src, src_scale */
1301 "orr r10, r8, r9 \n\t" /* r10 = (scr, src_scale) */
1336 uint16_t src_scale = SkAlpha255To256(alpha);
1337 uint16_t dst_scale = 256 - src_scale;
1359 /* get 64 bits of src, widen it, multiply by src_scale */
1363 src_wide = vmulq_u16 (src_wide, vdupq_n_u16(src_scale));
1389 *dst = SkAlphaMulQ(*src, src_scale) + SkAlphaMulQ(*dst, dst_scale);
1394 *dst = SkAlphaMulQ(*src, src_scale) + SkAlphaMulQ(*dst, dst_scale);