Home | History | Annotate | Download | only in video

Lines Matching refs:srcp

232 	Uint32 *srcp = (Uint32 *)info->s_pixels;
248 Uint32 s = *srcp++;
256 movq_m2r((*srcp), mm1);/* 2 x src -> mm1(ARGBARGB) */
270 srcp += 2;
272 srcp += srcskip;
290 Uint32 *srcp = (Uint32 *)info->s_pixels;
311 movd_m2r((*srcp), mm1);/* src(ARGB) -> mm1 (0000ARGB)*/
324 ++srcp;
328 movq_m2r((*srcp), mm0);/* 2 x src -> mm0(ARGBARGB)*/
353 srcp += 2;
356 srcp += srcskip;
368 Uint32 *srcp = (Uint32 *)info->s_pixels;
394 Uint32 alpha = *srcp & amask;
404 movd_m2r((*srcp), mm1);/* src(ARGB) -> mm1 (0000ARGB)*/
411 movd_m2r((*srcp), mm1);/* src(ARGB) -> mm1 (0000ARGB)*/
434 ++srcp;
437 srcp += srcskip;
450 Uint32 *srcp = (Uint32 *)info->s_pixels;
465 Uint32 s = *srcp++;
476 src1 = *(__m64*)srcp; /* 2 x src -> src1(ARGBARGB) */
491 srcp += 2;
494 srcp += srcskip;
513 Uint32 *srcp = (Uint32 *)info->s_pixels;
536 src2 = _mm_cvtsi32_si64(*srcp); /* src(ARGB) -> src2 (0000ARGB)*/
551 ++srcp;
559 src1 = *(__m64*)srcp; /* 2 x src -> src1(ARGBARGB)*/
584 srcp += 2;
587 srcp += srcskip;
599 Uint32 *srcp = (Uint32 *)info->s_pixels;
617 Uint32 alpha = *srcp & amask;
622 *dstp = (*srcp & chanmask) | (*dstp & ~chanmask);
624 src1 = _mm_cvtsi32_si64(*srcp); /* src(ARGB) -> src1 (0000ARGB)*/
645 ++srcp;
648 srcp += srcskip;
941 Uint32 *srcp = (Uint32 *)info->s_pixels;
993 RETRIEVE_RGB_PIXEL(((Uint8 *)srcp), 4, Pixel); \
1001 srcp++; \
1007 vector unsigned char valigner = VEC_ALIGNER(srcp);
1008 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1016 /* s = *srcp */
1017 voverflow = (vector unsigned char)vec_ld(15, srcp);
1045 srcp += 4;
1054 srcp += srcskip;
1064 Uint32 *srcp = (Uint32 *)info->s_pixels;
1097 DISEMBLE_RGBA((Uint8 *)srcp, 4, srcfmt, Pixel, sR, sG, sB, sA); \
1103 ++srcp; \
1112 vector unsigned char valigner = VEC_ALIGNER(srcp);
1113 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1120 /* s = *srcp */
1121 voverflow = (vector unsigned char)vec_ld(15, srcp);
1142 srcp += 4;
1150 srcp += srcskip;
1161 Uint32 *srcp = (Uint32 *)info->s_pixels;
1189 Uint32 s = *srcp; \
1206 ++srcp; \
1213 vector unsigned char valigner = VEC_ALIGNER(srcp);
1214 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1221 /* s = *srcp */
1222 voverflow = (vector unsigned char)vec_ld(15, srcp);
1240 srcp += 4;
1247 srcp += srcskip;
1258 Uint32 *srcp = (Uint32 *)info->s_pixels;
1297 DISEMBLE_RGB(((Uint8 *)srcp), 4, srcfmt, Pixel, sR, sG, sB); \
1301 ++srcp; \
1308 vector unsigned char valigner = VEC_ALIGNER(srcp);
1309 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1315 /* s = *srcp */
1316 voverflow = (vector unsigned char)vec_ld(15, srcp);
1333 srcp += 4;
1342 srcp += srcskip;
1354 Uint32 *srcp = (Uint32 *)info->s_pixels;
1378 Uint32 s = *srcp; \
1388 ++srcp; \
1395 vector unsigned char valigner = VEC_ALIGNER(srcp);
1396 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1402 /* s = *srcp */
1403 voverflow = (vector unsigned char)vec_ld(15, srcp);
1417 srcp += 4;
1426 srcp += srcskip;
1440 Uint32 *srcp = (Uint32 *)info->s_pixels;
1447 Uint32 s = *srcp++;
1452 srcp += srcskip;
1466 Uint32 *srcp = (Uint32 *)info->s_pixels;
1478 s = *srcp;
1488 ++srcp;
1492 s = *srcp;
1500 ((srcp[1] & 0xff00) << 8);
1507 ++srcp;
1509 s1 = *srcp;
1517 ++srcp;
1520 srcp += srcskip;
1531 Uint32 *srcp = (Uint32 *)info->s_pixels;
1542 Uint32 s = *srcp;
1567 ++srcp;
1570 srcp += srcskip;
1581 Uint32 *srcp = (Uint32 *)info->s_pixels;
1618 : : "r" (srcp), "r" (dstp) );
1620 alpha = *srcp & amask;
1639 : : "r" (srcp), "r" (dstp) );
1677 : : "r" (srcp), "r" (dstp), "r" (alpha) );
1680 ++srcp;
1683 srcp += srcskip;
1699 Uint32 *srcp = (Uint32 *)info->s_pixels;
1719 _m_prefetch(srcp + 16);
1722 alpha = *srcp & amask;
1727 *dstp = (*srcp & chanmask) | (*dstp & ~chanmask);
1729 src1 = _mm_cvtsi32_si64(*srcp); /* src(ARGB) -> src1 (0000ARGB)*/
1750 ++srcp;
1753 srcp += srcskip;
1777 Uint16 *srcp = (Uint16 *)info->s_pixels;
1783 if(((uintptr_t)srcp ^ (uintptr_t)dstp) & 2) {
1794 Uint16 d = *dstp, s = *srcp;
1797 srcp++;
1800 srcp++; /* srcp is now 32-bit aligned */
1803 prev_sw = ((Uint32 *)srcp)[-1];
1807 sw = *(Uint32 *)srcp;
1817 srcp += 2;
1830 srcp++;
1833 srcp += srcskip - 1;
1840 if((uintptr_t)srcp & 2) {
1841 Uint16 d = *dstp, s = *srcp;
1843 srcp++;
1847 /* srcp and dstp are now 32-bit aligned */
1850 Uint32 sw = *(Uint32 *)srcp;
1853 srcp += 2;
1860 Uint16 d = *dstp, s = *srcp;
1862 srcp++;
1865 srcp += srcskip;
1881 Uint16 *srcp = (Uint16 *)info->s_pixels;
1907 s = *srcp++;
1920 s = *srcp++;
1932 s = *srcp++;
1945 movq_m2r((*srcp), mm2);/* 4 src pixels -> mm2 */
2001 srcp += 4;
2004 srcp += srcskip;
2020 Uint16 *srcp = (Uint16 *)info->s_pixels;
2046 s = *srcp++;
2059 s = *srcp++;
2071 s = *srcp++;
2084 movq_m2r((*srcp), mm2);/* 4 src pixels -> mm2 */
2145 srcp += 4;
2148 srcp += srcskip;
2166 Uint16 *srcp = (Uint16 *)info->s_pixels;
2191 s = *srcp++;
2204 s = *srcp++;
2216 s = *srcp++;
2229 src1 = *(__m64*)srcp; /* 4 src pixels -> src1 */
2281 srcp += 4;
2284 srcp += srcskip;
2300 Uint16 *srcp = (Uint16 *)info->s_pixels;
2326 s = *srcp++;
2339 s = *srcp++;
2351 s = *srcp++;
2364 src1 = *(__m64*)srcp; /* 4 src pixels -> src1 */
2416 srcp += 4;
2419 srcp += srcskip;
2436 Uint16 *srcp = (Uint16 *)info->s_pixels;
2444 Uint32 s = *srcp++;
2457 srcp += srcskip;
2472 Uint16 *srcp = (Uint16 *)info->s_pixels;
2480 Uint32 s = *srcp++;
2493 srcp += srcskip;
2504 Uint32 *srcp = (Uint32 *)info->s_pixels;
2511 Uint32 s = *srcp;
2534 srcp++;
2537 srcp += srcskip;
2547 Uint32 *srcp = (Uint32 *)info->s_pixels;
2555 Uint32 s = *srcp;
2578 srcp++;
2581 srcp += srcskip;