Home | History | Annotate | Download | only in opts

Lines Matching refs:vsrc

22     uint8x8x4_t vsrc;
35 vsrc.val[0] = vsrc_0;
36 vsrc.val[1] = vsrc_1;
37 vsrc.val[2] = vsrc_2;
39 return vsrc;
43 uint8x8x4_t vsrc;
58 vsrc.val[0] = vsrc_0;
59 vsrc.val[1] = vsrc_1;
60 vsrc.val[2] = vsrc_2;
61 vsrc.val[3] = vsrc_3;
63 return vsrc;
73 uint8x8x4_t vsrc;
78 vsrc = sk_vld4_u8_arm64_3(src);
80 vsrc = vld4_u8((uint8_t*)src);
85 vdst = SkPixel32ToPixel16_neon8(vsrc);
117 uint8x8x4_t vsrc;
123 vsrc = sk_vld4_u8_arm64_3(src);
136 vsrc.val[0] = d0;
137 vsrc.val[1] = d1;
138 vsrc.val[2] = d2;
150 vsrc.val[NEON_R] = vshr_n_u8(vsrc.val[NEON_R], 3);
151 vsrc.val[NEON_G] = vshr_n_u8(vsrc.val[NEON_G], 2);
152 vsrc.val[NEON_B] = vshr_n_u8(vsrc.val[NEON_B], 3);
155 vres_r = vmovl_u8(vsrc.val[NEON_R]) - vdst_r;
156 vres_g = vmovl_u8(vsrc.val[NEON_G]) - vdst_g;
157 vres_b = vmovl_u8(vsrc.val[NEON_B]) - vdst_b;
398 uint8x8x4_t vsrc;
403 vsrc = sk_vld4_u8_arm64_4(src);
407 "vld4.u8 %h[vsrc], [%[src]]!"
408 vsrc] "=w" (vsrc), [src] "+&r" (src)
423 vsrc.val[0] = d0;
424 vsrc.val[1] = d1;
425 vsrc.val[2] = d2;
426 vsrc.val[3] = d3;
438 vsrc.val[NEON_R] = vshr_n_u8(vsrc.val[NEON_R], 8 - SK_R16_BITS);
439 vsrc.val[NEON_G] = vshr_n_u8(vsrc.val[NEON_G], 8 - SK_G16_BITS);
440 vsrc.val[NEON_B] = vshr_n_u8(vsrc.val[NEON_B], 8 - SK_B16_BITS);
443 vres_a = vmull_u8(vsrc.val[NEON_A], valpha);
444 vres_r = vmull_u8(vsrc.val[NEON_R], valpha);
445 vres_g = vmull_u8(vsrc.val[NEON_G], valpha);
446 vres_b = vmull_u8(vsrc.val[NEON_B], valpha);
528 uint8x8x4_t vsrc;
540 vsrc = sk_vld4_u8_arm64_3(src);
553 vsrc.val[0] = d0;
554 vsrc.val[1] = d1;
555 vsrc.val[2] = d2;
558 vsrc_r = vsrc.val[NEON_R];
559 vsrc_g = vsrc.val[NEON_G];
560 vsrc_b = vsrc.val[NEON_B];
934 uint8x8_t vsrc, vdst, vres;
945 vsrc = vreinterpret_u8_u32(vld1_u32(src));
949 vsrc_wide = vmovl_u8(vsrc);
967 uint8x8_t vsrc = vdup_n_u8(0), vdst = vdup_n_u8(0), vres;
971 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0));
975 vsrc_wide = vmovl_u8(vsrc);
1000 uint8x8_t vsrc = vdup_n_u8(0), vdst = vdup_n_u8(0), vres;
1005 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0));
1009 dst_scale = vget_lane_u8(vsrc, 3);
1015 vsrc_wide = vmovl_u8(vsrc);
1038 uint8x8_t vsrc, vdst, vres, vsrc_alphas;
1045 vsrc = vreinterpret_u8_u32(vld1_u32(src));
1052 vsrc_alphas = vtbl1_u8(vsrc, alpha_mask);
1059 vsrc_wide = vmovl_u8(vsrc);
1138 uint8x8x4_t vsrc;
1190 vsrc = sk_vld4_u8_arm64_4(src);
1202 vsrc.val[0] = d0;
1203 vsrc.val[1] = d1;
1204 vsrc.val[2] = d2;
1205 vsrc.val[3] = d3;
1208 sa = vsrc.val[NEON_A];
1209 sr = vsrc.val[NEON_R];
1210 sg = vsrc.val[NEON_G];
1211 sb = vsrc.val[NEON_B];
1355 uint8x8x4_t vsrc;
1358 vsrc = sk_vld4_u8_arm64_3(src);
1371 vsrc.val[0] = d0;
1372 vsrc.val[1] = d1;
1373 vsrc.val[2] = d2;
1376 sr = vsrc.val[NEON_R];
1377 sg = vsrc.val[NEON_G];
1378 sb = vsrc.val[NEON_B];
1481 uint32x2x4_t vsrc;
1484 "vld1.32 %h[vsrc], [%[src]]!"
1485 : [vsrc] "=w" (vsrc), [src] "+r" (src)
1489 vsrc.val[0] = vld1_u32(src);
1490 vsrc.val[1] = vld1_u32(src+2);
1491 vsrc.val[2] = vld1_u32(src+4);
1492 vsrc.val[3] = vld1_u32(src+6);
1499 vtmp.val[0] = vmull_u8(vreinterpret_u8_u32(vsrc.val[0]), vscale);
1500 vtmp.val[1] = vmull_u8(vreinterpret_u8_u32(vsrc.val[1]), vscale);
1501 vtmp.val[2] = vmull_u8(vreinterpret_u8_u32(vsrc.val[2]), vscale);
1502 vtmp.val[3] = vmull_u8(vreinterpret_u8_u32(vsrc.val[3]), vscale);