/external/skia/src/effects/gradients/ |
Sk4fLinearGradient.h | 25 template<typename DstType, TileMode> 28 template <typename DstType, bool premul> 29 void shadePremulSpan(int x, int y, DstType[], int count) const; 31 template <typename DstType, bool premul, SkShader::TileMode tileMode> 32 void shadePremulTileSpan(int x, int y, DstType[], int count) const; 34 template <typename DstType, bool premul, SkShader::TileMode tileMode> 35 void shadeSpanInternal(int x, int y, DstType[], int count) const;
|
Sk4fLinearGradient.cpp | 29 template<typename DstType, bool do_premul> 30 void fill(const Sk4f& c, DstType* dst, int n); 55 template<typename DstType, bool do_premul> 56 void store(const Sk4f& color, DstType* dst); 78 template<typename DstType, bool do_premul> 83 DstType* dst) { 84 store<DstType, do_premul>(c0, dst++); 85 store<DstType, do_premul>(c1, dst++); 86 store<DstType, do_premul>(c2, dst++); 87 store<DstType, do_premul>(c3, dst++) [all...] |
/external/opencv3/modules/cudev/include/opencv2/cudev/grid/ |
transpose.hpp | 60 template <class Policy, class SrcPtr, typename DstType> 61 __host__ void gridTranspose_(const SrcPtr& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null()) 71 template <class Policy, class SrcPtr, typename DstType> 72 __host__ void gridTranspose_(const SrcPtr& src, const GlobPtrSz<DstType>& dst, Stream& stream = Stream::Null()) 92 template <class SrcPtr, typename DstType> 93 __host__ void gridTranspose(const SrcPtr& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null()) 98 template <class SrcPtr, typename DstType> 99 __host__ void gridTranspose(const SrcPtr& src, const GlobPtrSz<DstType>& dst, Stream& stream = Stream::Null())
|
split_merge.hpp | 63 template <class Policy, class SrcPtrTuple, typename DstType, class MaskPtr> 64 __host__ void gridMerge_(const SrcPtrTuple& src, GpuMat_<DstType>& dst, const MaskPtr& mask, Stream& stream = Stream::Null()) 66 CV_StaticAssert( VecTraits<DstType>::cn == tuple_size<SrcPtrTuple>::value, "" ); 75 grid_split_merge_detail::MergeImpl<VecTraits<DstType>::cn, Policy>::merge(shrinkPtr(src), 82 template <class Policy, class SrcPtrTuple, typename DstType, class MaskPtr> 83 __host__ void gridMerge_(const SrcPtrTuple& src, const GlobPtrSz<DstType>& dst, const MaskPtr& mask, Stream& stream = Stream::Null()) 85 CV_StaticAssert( VecTraits<DstType>::cn == tuple_size<SrcPtrTuple>::value, "" ); 93 grid_split_merge_detail::MergeImpl<VecTraits<DstType>::cn, Policy>::merge(shrinkPtr(src), 100 template <class Policy, class SrcPtrTuple, typename DstType> 101 __host__ void gridMerge_(const SrcPtrTuple& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null() [all...] |
integral.hpp | 59 template <class SrcPtr, typename DstType> 60 __host__ void gridIntegral(const SrcPtr& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null())
|
pyramids.hpp | 61 template <class Brd, class SrcPtr, typename DstType> 62 __host__ void gridPyrDown_(const SrcPtr& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null()) 72 template <class SrcPtr, typename DstType> 73 __host__ void gridPyrDown(const SrcPtr& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null()) 78 template <class SrcPtr, typename DstType> 79 __host__ void gridPyrUp(const SrcPtr& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null())
|
transform.hpp | 63 template <class Policy, class SrcPtr, typename DstType, class UnOp, class MaskPtr> 64 __host__ void gridTransformUnary_(const SrcPtr& src, GpuMat_<DstType>& dst, const UnOp& op, const MaskPtr& mask, Stream& stream = Stream::Null()) 76 template <class Policy, class SrcPtr, typename DstType, class UnOp, class MaskPtr> 77 __host__ void gridTransformUnary_(const SrcPtr& src, const GlobPtrSz<DstType>& dst, const UnOp& op, const MaskPtr& mask, Stream& stream = Stream::Null()) 88 template <class Policy, class SrcPtr, typename DstType, class UnOp> 89 __host__ void gridTransformUnary_(const SrcPtr& src, GpuMat_<DstType>& dst, const UnOp& op, Stream& stream = Stream::Null()) 99 template <class Policy, class SrcPtr, typename DstType, class UnOp> 100 __host__ void gridTransformUnary_(const SrcPtr& src, const GlobPtrSz<DstType>& dst, const UnOp& op, Stream& stream = Stream::Null()) 110 template <class Policy, class SrcPtr1, class SrcPtr2, typename DstType, class BinOp, class MaskPtr> 111 __host__ void gridTransformBinary_(const SrcPtr1& src1, const SrcPtr2& src2, GpuMat_<DstType>& dst, const BinOp& op, const MaskPtr& mask, Stream& stream = Stream::Null() [all...] |
copy.hpp | 63 template <class Policy, class SrcPtr, typename DstType, class MaskPtr> 64 __host__ void gridCopy_(const SrcPtr& src, GpuMat_<DstType>& dst, const MaskPtr& mask, Stream& stream = Stream::Null()) 76 template <class Policy, class SrcPtr, typename DstType, class MaskPtr> 77 __host__ void gridCopy_(const SrcPtr& src, const GlobPtrSz<DstType>& dst, const MaskPtr& mask, Stream& stream = Stream::Null()) 88 template <class Policy, class SrcPtr, typename DstType> 89 __host__ void gridCopy_(const SrcPtr& src, GpuMat_<DstType>& dst, Stream& stream = Stream::Null()) 99 template <class Policy, class SrcPtr, typename DstType> 100 __host__ void gridCopy_(const SrcPtr& src, const GlobPtrSz<DstType>& dst, Stream& stream = Stream::Null()) 357 template <class SrcPtr, typename DstType, class MaskPtr> 358 __host__ void gridCopy(const SrcPtr& src, GpuMat_<DstType>& dst, const MaskPtr& mask, Stream& stream = Stream::Null() [all...] |
/external/opencv3/modules/cudev/include/opencv2/cudev/grid/detail/ |
split_merge.hpp | 62 template <class Src1Ptr, class Src2Ptr, typename DstType, class MaskPtr> 63 __global__ void mergeC2(const Src1Ptr src1, const Src2Ptr src2, GlobPtr<DstType> dst, const MaskPtr mask, const int rows, const int cols) 65 typedef typename VecTraits<DstType>::elem_type dst_elem_type; 73 dst(y, x) = VecTraits<DstType>::make( 79 template <class Policy, class Src1Ptr, class Src2Ptr, typename DstType, class MaskPtr> 80 __host__ void mergeC2(const Src1Ptr& src1, const Src2Ptr& src2, const GlobPtr<DstType>& dst, const MaskPtr& mask, int rows, int cols, cudaStream_t stream) 92 template <class Src1Ptr, class Src2Ptr, class Src3Ptr, typename DstType, class MaskPtr> 93 __global__ void mergeC3(const Src1Ptr src1, const Src2Ptr src2, const Src3Ptr src3, GlobPtr<DstType> dst, const MaskPtr mask, const int rows, const int cols) 95 typedef typename VecTraits<DstType>::elem_type dst_elem_type; 103 dst(y, x) = VecTraits<DstType>::make [all...] |
transpose.hpp | 58 template <int TILE_DIM, int BLOCK_DIM_Y, class SrcPtr, typename DstType> 59 __global__ void transpose(const SrcPtr src, GlobPtr<DstType> dst, const int rows, const int cols) 105 dst(yIndex + i, xIndex) = saturate_cast<DstType>(tile[threadIdx.x][threadIdx.y + i]); 111 template <class Policy, class SrcPtr, typename DstType> 112 __host__ void transpose(const SrcPtr& src, const GlobPtr<DstType>& dst, int rows, int cols, cudaStream_t stream)
|
transform.hpp | 158 template <class SrcPtr, typename DstType, class UnOp, class MaskPtr> 159 __global__ void transformSimple(const SrcPtr src, GlobPtr<DstType> dst, const UnOp op, const MaskPtr mask, const int rows, const int cols) 167 dst(y, x) = saturate_cast<DstType>(op(src(y, x))); 170 template <class SrcPtr1, class SrcPtr2, typename DstType, class BinOp, class MaskPtr> 171 __global__ void transformSimple(const SrcPtr1 src1, const SrcPtr2 src2, GlobPtr<DstType> dst, const BinOp op, const MaskPtr mask, const int rows, const int cols) 179 dst(y, x) = saturate_cast<DstType>(op(src1(y, x), src2(y, x))); 184 template <int SHIFT, typename SrcType, typename DstType, class UnOp, class MaskPtr> 185 __global__ void transformSmart(const GlobPtr<SrcType> src_, GlobPtr<DstType> dst_, const UnOp op, const MaskPtr mask, const int rows, const int cols) 188 typedef typename MakeVec<DstType, SHIFT>::type write_type; 197 DstType* dst = dst_.row(y) [all...] |
copy.hpp | 59 template <class SrcPtr, typename DstType, class MaskPtr> 60 __global__ void copy(const SrcPtr src, GlobPtr<DstType> dst, const MaskPtr mask, const int rows, const int cols) 68 dst(y, x) = saturate_cast<DstType>(src(y, x)); 71 template <class Policy, class SrcPtr, typename DstType, class MaskPtr> 72 __host__ void copy(const SrcPtr& src, const GlobPtr<DstType>& dst, const MaskPtr& mask, int rows, int cols, cudaStream_t stream)
|
pyr_down.hpp | 60 template <class Brd, class SrcPtr, typename DstType> 61 __global__ void pyrDown(const SrcPtr src, GlobPtr<DstType> dst, const int src_rows, const int src_cols, const int dst_cols) 181 dst(y, dst_x) = saturate_cast<DstType>(sum); 185 template <class Brd, class SrcPtr, typename DstType> 186 __host__ void pyrDown(const SrcPtr& src, const GlobPtr<DstType>& dst, int src_rows, int src_cols, int dst_rows, int dst_cols, cudaStream_t stream)
|
pyr_up.hpp | 60 template <class SrcPtr, typename DstType> 61 __global__ void pyrUp(const SrcPtr src, GlobPtr<DstType> dst, const int src_rows, const int src_cols, const int dst_rows, const int dst_cols) 153 dst(y, x) = saturate_cast<DstType>(4.0f * sum); 156 template <class SrcPtr, typename DstType> 157 __host__ void pyrUp(const SrcPtr& src, const GlobPtr<DstType>& dst, int src_rows, int src_cols, int dst_rows, int dst_cols, cudaStream_t stream)
|
/external/opencv3/modules/cudaarithm/src/cuda/ |
sub_scalar.cu | 59 template <typename SrcType, typename ScalarType, typename DstType> struct SubScalarOp : unary_function<SrcType, DstType> 63 __device__ __forceinline__ DstType operator ()(SrcType a) const 65 return saturate_cast<DstType>(saturate_cast<ScalarType>(a) - val); 69 template <typename SrcType, typename ScalarType, typename DstType> struct SubScalarOpInv : unary_function<SrcType, DstType> 73 __device__ __forceinline__ DstType operator ()(SrcType a) const 75 return saturate_cast<DstType>(val - saturate_cast<ScalarType>(a)); 89 template <typename SrcType, typename ScalarDepth, typename DstType> 98 SubScalarOpInv<SrcType, ScalarType, DstType> op [all...] |
div_scalar.cu | 107 template <typename SrcType, typename ScalarType, typename DstType> struct DivScalarOp : unary_function<SrcType, DstType> 111 __device__ __forceinline__ DstType operator ()(SrcType a) const 113 return saturate_cast<DstType>(SafeDiv<ScalarType, VecTraits<ScalarType>::cn>::op(saturate_cast<ScalarType>(a), val)); 117 template <typename SrcType, typename ScalarType, typename DstType> struct DivScalarOpInv : unary_function<SrcType, DstType> 121 __device__ __forceinline__ DstType operator ()(SrcType a) const 123 return saturate_cast<DstType>(SafeDiv<ScalarType, VecTraits<ScalarType>::cn>::op(val, saturate_cast<ScalarType>(a))); 137 template <typename SrcType, typename ScalarDepth, typename DstType> 146 DivScalarOpInv<SrcType, ScalarType, DstType> op [all...] |
add_scalar.cu | 59 template <typename SrcType, typename ScalarType, typename DstType> struct AddScalarOp : unary_function<SrcType, DstType> 63 __device__ __forceinline__ DstType operator ()(SrcType a) const 65 return saturate_cast<DstType>(saturate_cast<ScalarType>(a) + val); 79 template <typename SrcType, typename ScalarDepth, typename DstType> 86 AddScalarOp<SrcType, ScalarType, DstType> op; 90 gridTransformUnary_< TransformPolicy<ScalarDepth> >(globPtr<SrcType>(src), globPtr<DstType>(dst), op, globPtr<uchar>(mask), stream); 92 gridTransformUnary_< TransformPolicy<ScalarDepth> >(globPtr<SrcType>(src), globPtr<DstType>(dst), op, stream);
|
mul_scalar.cu | 59 template <typename SrcType, typename ScalarType, typename DstType> struct MulScalarOp : unary_function<SrcType, DstType> 63 __device__ __forceinline__ DstType operator ()(SrcType a) const 65 return saturate_cast<DstType>(saturate_cast<ScalarType>(a) * val); 79 template <typename SrcType, typename ScalarDepth, typename DstType> 86 MulScalarOp<SrcType, ScalarType, DstType> op; 89 gridTransformUnary_< TransformPolicy<ScalarDepth> >(globPtr<SrcType>(src), globPtr<DstType>(dst), op, stream);
|
/external/skia/src/core/ |
SkBlitter_PM4f.cpp | 37 typename State::DstType* device = State::WritableAddr(fDevice, x, y); 42 device = (typename State::DstType*)((char*)device + deviceRB); 50 typename State::DstType* device = State::WritableAddr(fDevice, x, y); 56 device = (typename State::DstType*)((char*)device + deviceRB); 61 typename State::DstType* device = State::WritableAddr(fDevice, x, y); 93 typename State::DstType* device = State::WritableAddr(fDevice, x, y); 100 device = (typename State::DstType*)((char*)device + dstRB); 122 typename State::DstType* device = State::WritableAddr(fDevice, x, y); 129 device = (typename State::DstType*)((char*)device + dstRB); 148 typename State::DstType* device = State::WritableAddr(fDevice, x, y) [all...] |
SkXfermodeU64.cpp | 19 enum DstType { 28 template <DstType D> Sk4f unit_to_bias(const Sk4f& x4) { 32 template <DstType D> Sk4f bias_to_unit(const Sk4f& x4) { 50 template <DstType D> Sk4f load_from_dst(uint64_t dst) { 55 template <DstType D> uint64_t store_to_dst(const Sk4f& x4) { 70 template <DstType D> void xfer_u64_1(const SkXfermode* xfer, uint64_t dst[], 90 template <DstType D> void xfer_u64_n(const SkXfermode* xfer, uint64_t dst[], 119 template <DstType D> void clear(const SkXfermode*, uint64_t dst[], 142 template <DstType D> void src_1(const SkXfermode*, uint64_t dst[], 155 template <DstType D> void src_n(const SkXfermode*, uint64_t dst[] [all...] |
SkXfermode4f.cpp | 12 enum DstType { 25 template <DstType D> Sk4f load_dst(SkPMColor dstC) { 33 template <DstType D> uint32_t store_dst(const Sk4f& x4) { 43 template <DstType D> void general_1(const SkXfermode* xfer, uint32_t dst[], 63 template <DstType D> void general_n(const SkXfermode* xfer, uint32_t dst[], 134 template <DstType D> void src_n(const SkXfermode*, uint32_t dst[], 157 template <DstType D> void src_1(const SkXfermode*, uint32_t dst[], 227 template <DstType D> void srcover_n(const SkXfermode*, uint32_t dst[], 387 template <DstType D> 414 template <DstType D [all...] |
/external/gemmlowp/internal/ |
output_neon.h | 243 template <typename DstType> 244 inline void StoreFinalOutput(NEONFragmentUint8x4x1 value, DstType* dst, int row, 253 template <typename DstType> 254 inline void StoreFinalOutput(NEONFragmentUint8x16x1 value, DstType* dst, 261 template <typename DstType> 262 inline void StoreFinalOutput(NEONFragmentInt32x4x1 value, DstType* dst, int row, 269 template <typename DstType> 270 inline void StoreFinalOutput(NEONFragmentInt32x16x1 value, DstType* dst,
|
output.h | 384 template <typename OutputType, typename DstType> 385 void StoreFinalOutput(OutputType value, DstType* dst, int row, int col) { 398 template <typename DstType> 399 void Execute(InputType input, DstType* dst, int row, int col) { 405 typedef typename DstType::Scalar ScalarDstType;
|
/external/opencv3/modules/imgproc/perf/ |
perf_distanceTransform.cpp | 35 CV_ENUM(DstType, CV_8U, CV_32F) 38 typedef std::tr1::tuple<Size, DistanceType, MaskSize, DstType> SrcSize_DistType_MaskSize_DstType; 48 DstType::all() 55 int dstType = get<3>(GetParam()); 58 Mat dst(srcSize, dstType); 65 TEST_CYCLE() distanceTransform( src, dst, distanceType, maskSize, dstType);
|
/external/skia/include/core/ |
SkShader.h | 103 enum DstType { 109 DstType dstType) 113 , fPreferredDstType(dstType) {} 118 const DstType fPreferredDstType; // the "natural" client dest type
|