Lines Matching refs:dst
55 uchar* dst, int dst_step,
61 for( ; size.height--; src += src_step, dst += dst_step )
67 ((int*)dst)[j] = t0;
68 ((int*)dst)[j+1] = t1;
71 ((int*)dst)[j+2] = t0;
72 ((int*)dst)[j+3] = t1;
76 ((int*)dst)[j] = ((const int*)src)[j];
83 uchar* dst, int dst_step,
87 for( i = 0; i < size.width; i++, dst += dst_step, src += pix_size )
94 ((int*)dst)[j] = ((int*)_src)[0];
101 ((int*)dst)[j] = t0;
102 ((int*)dst)[j+1] = t1;
110 ((int*)dst)[j] = t0;
111 ((int*)dst)[j+1] = t1;
114 ((int*)dst)[j+2] = t0;
115 ((int*)dst)[j+3] = t1;
573 void* dst, size_t dststep, CvSize srcsize, CvSize dstsize,
577 const void* src2, size_t step2, void* dst, size_t dststep,
581 const void* src2, size_t step2, void* dst, size_t dststep,
1266 for( i = 0; i < size.width; i++, dst += dst_cn ) \
1273 dst[k] = _cast_macro2_(t0); \
1288 dst[i] = _cast_macro2_(t0); \
1304 dst[i] = _cast_macro2_(t0); \
1305 dst[i+1] = _cast_macro2_(t1); \
1307 src += size.width*2; dst += size.width*2; \
1310 for( i = 0; i < size.width; i++, src += 2, dst += dst_cn ) \
1318 dst[k] = _cast_macro2_(t0); \
1333 dst[i] = _cast_macro2_(t0); \
1334 dst[i+1] = _cast_macro2_(t1); \
1355 dst[i] = _cast_macro2_(t0); \
1356 dst[i+1] = _cast_macro2_(t1); \
1357 dst[i+2] = _cast_macro2_(t2); \
1359 src += size.width*3; dst += size.width*3; \
1367 dst[i] = _cast_macro2_(t0); \
1369 dst += size.width; \
1372 for( i = 0; i < size.width; i++, src += 3, dst += dst_cn ) \
1380 dst[k] = _cast_macro2_(t0); \
1397 dst[i] = _cast_macro2_(t0); \
1398 dst[i+1] = _cast_macro2_(t1); \
1399 dst[i+2] = _cast_macro2_(t2); \
1405 for( i = 0; i < size.width; i++, src += 4, dst += dst_cn ) \
1414 dst[k] = _cast_macro2_(t0); \
1429 dst[i] = _cast_macro2_(t0); \
1430 dst[i+1] = _cast_macro2_(t1); \
1435 dst[i+2] = _cast_macro2_(t0); \
1436 dst[i+3] = _cast_macro2_(t1); \
1445 arrtype* dst, int dststep, CvSize size, \
1449 dststep = dststep/sizeof(dst[0]) - size.width*dst_cn; \
1450 for( ; size.height--; src += srcstep, dst += dststep ) \
1465 arrtype* dst, int dststep, CvSize size, \
1469 dststep /= sizeof(dst[0]); \
1470 for( ; size.height--; src += srcstep, dst += dststep ) \
1556 void* dst, int dststep, CvSize size,
1561 void* dst, int dststep, CvSize size,
1566 void* dst, int dststep, CvSize size,
1588 void* dst, int dststep, CvSize size, const float* coeffs );
1605 CvMat dststub, *dst = (CvMat*)dstarr;
1609 CvSeq hdr; // need only one copy of stub header & seqblock (either for src or dst)
1631 if( CV_IS_SEQ( dst ))
1633 dst_seq = (CvSeq*)dst;
1638 CV_CALL( dst = cvGetMat( dst, &dststub, &coi2 ));
1643 if( !CV_ARE_DEPTHS_EQ(src, dst) )
1660 if( CV_IS_MAT_CONT(dst->type) || (dst->rows != 1 && dst->cols != 1) )
1663 if( dst->rows + dst->cols - 1 != src_seq->total )
1666 dst_seq = cvMakeSeqHeaderForArray( CV_MAT_TYPE(dst->type), sizeof(hdr),
1667 CV_ELEM_SIZE(dst->type), dst->data.ptr,
1668 dst->rows + dst->cols + 1, &hdr, &block_hdr );
1678 else if( !CV_ARE_SIZES_EQ( src, dst ))
1683 dst_cn = CV_MAT_CN( dst->type );
1852 int dststep = dst->step;
1855 if( CV_IS_MAT_CONT( src->type & dst->type ))
1863 lut_func( src->data.ptr, src->step, dst->data.ptr,
1864 dst->step, size, lut->data.ptr );
1867 IPPI_CALL( ipp_func( src->data.ptr, srcstep, dst->data.ptr,
1871 diag_func( src->data.ptr, src->step, dst->data.ptr,
1872 dst->step, size, buffer );
1874 func( src->data.ptr, src->step, dst->data.ptr,
1875 dst->step, size, buffer, dst_cn );
1930 arrtype* dst, int dststep, \
1935 srcstep /= sizeof(src[0]); dststep /= sizeof(dst[0]); \
1937 for( ; size.height--; src += srcstep, dst += dststep ) \
1947 dst[i] = (arrtype)((x*mat[0] + y*mat[1] + mat[2]) * w); \
1948 dst[i+1] = (arrtype)((x*mat[3] + y*mat[4] + mat[5]) * w); \
1952 dst[i] = (arrtype)0; \
1953 dst[i+1] = (arrtype)0; \
1965 arrtype* dst, int dststep, \
1970 srcstep /= sizeof(src[0]); dststep /= sizeof(dst[0]); \
1972 for( ; size.height--; src += srcstep, dst += dststep ) \
1982 dst[i] = (arrtype)((x*mat[0] + y*mat[1] + z*mat[2] + mat[3]) * w); \
1983 dst[i+1] = (arrtype)((x*mat[4] + y*mat[5] + z*mat[6] + mat[7]) * w); \
1984 dst[i+2] = (arrtype)((x*mat[8] + y*mat[9] + z*mat[10] + mat[11]) * w); \
1988 dst[i] = (arrtype)0; \
1989 dst[i+1] = (arrtype)0; \
1990 dst[i+2] = (arrtype)0; \
2024 CvMat dstub, *dst = (CvMat*)dstarr;
2044 if( !CV_IS_MAT( dst ))
2047 CV_CALL( dst = cvGetMat( dst, &dstub, &coi ));
2053 if( !CV_ARE_TYPES_EQ( src, dst ))
2056 if( !CV_ARE_SIZES_EQ( src, dst ))
2100 if( CV_IS_MAT_CONT( src->type & dst->type ))
2106 IPPI_CALL( func( src->data.ptr, src->step, dst->data.ptr, dst->step, size, buffer));
2108 CV_CHECK_NANS( dst );
2118 #define ICV_DEF_MULADDC_CASE_C1( arrtype, temptype, src1, src2, dst, len ) \
2127 (dst)[i] = (arrtype)t0; \
2128 (dst)[i+1] = (arrtype)t1; \
2133 (dst)[i+2] = (arrtype)t0; \
2134 (dst)[i+3] = (arrtype)t1; \
2140 (dst)[i] = (arrtype)t0; \
2145 #define ICV_DEF_MULADDC_CASE_C2( arrtype, temptype, src1, src2, dst, len ) \
2154 (dst)[i] = (arrtype)t0; \
2155 (dst)[i+1] = (arrtype)t1; \
2160 (dst)[i+2] = (arrtype)t0; \
2161 (dst)[i+3] = (arrtype)t1; \
2169 (dst)[i] = (arrtype)t0; \
2170 (dst)[i+1] = (arrtype)t1; \
2179 arrtype* dst, int dststep, CvSize size, \
2185 dststep /= sizeof(dst[0]); \
2187 for( ; size.height--; src1+=srcstep1, src2+=srcstep2, dst+=dststep ) \
2190 dst, size.width ) \
2226 CvMat stub, *dst = (CvMat*)dstarr;
2230 if( !CV_IS_MAT( src1 ) || !CV_IS_MAT(src2) || !CV_IS_MAT(dst))
2235 CV_CALL( dst = cvGetMat( dst, &stub, &coi3 ));
2241 if( !CV_ARE_TYPES_EQ( src1, dst ) || !CV_ARE_TYPES_EQ( src2, dst ))
2244 if( !CV_ARE_SIZES_EQ( src1, dst ) || !CV_ARE_SIZES_EQ( src2, dst ))
2250 if( CV_IS_MAT_CONT( src1->type & src2->type & dst->type ))
2260 float* mC = dst->data.fl;
2276 double* mC = dst->data.db;
2308 dst->data.ptr, dst->step, size, scale.val ));
2311 CV_CHECK_NANS( dst );
2381 avgtype* dst, int dststep, \
2392 dststep /= sizeof(dst[0]); \
2393 for( y = 0; y < dstsize; y++, dst += dststep ) \
2398 double t0 = dst[x] + ty*tempbuf[x]; \
2399 double t1 = dst[x+1] + ty*tempbuf[x+1]; \
2400 dst[x] = (avgtype)t0; \
2401 dst[x+1] = (avgtype)t1; \
2402 t0 = dst[x+2] + ty*tempbuf[x+2]; \
2403 t1 = dst[x+3] + ty*tempbuf[x+3]; \
2404 dst[x+2] = (avgtype)t0; \
2405 dst[x+3] = (avgtype)t1; \
2408 dst[x] = (avgtype)(dst[x] + ty*tempbuf[x]); \
2855 dsttype* dst, int dststep, \
2860 dsttype* tdst = dst; \
2884 srcstep /= sizeof(src[0]); dststep /= sizeof(dst[0]); \
2981 dst[dststep*i + j] = dst[dststep*j + i]; \
2993 dsttype* dst, int dststep, \
2998 dsttype* tdst = dst; \
3000 srcstep /= sizeof(src[0]); dststep /= sizeof(dst[0]); \
3079 dst[dststep*i + j] = dst[dststep*j + i]; \
3108 void* dst, int dststep, const void* delta,
3123 CvMat dstub, *dst = (CvMat*)dstarr;
3130 if( !CV_IS_MAT( dst ))
3131 CV_CALL( dst = cvGetMat( dst, &dstub ));
3138 if( !CV_ARE_TYPES_EQ( dst, delta ))
3154 dtype = CV_MAT_TYPE( dst->type );
3156 if( dst->rows != dst->cols )
3159 if( (order != 0 && src->cols != dst->cols) ||
3160 (order == 0 && src->rows != dst->rows))
3163 if( src->data.ptr == dst->data.ptr || (stype == dtype &&
3164 (dst->cols >= gemm_level && dst->rows >= gemm_level &&
3174 cvGEMM( src, src, scale, 0, 0, dst, order == 0 ? CV_GEMM_B_T : CV_GEMM_A_T );
3210 IPPI_CALL( func( src->data.ptr, src->step, dst->data.ptr, dst->step,