/external/tensorflow/tensorflow/core/kernels/ |
multinomial_op_gpu.cu.cc | 36 using GPUDevice = Eigen::GpuDevice; 55 struct MultinomialFunctor<GPUDevice, T, OutputType> { 56 void operator()(OpKernelContext* ctx, const GPUDevice& d, 66 functor::FillPhiloxRandom<GPUDevice, Dist>()(ctx, d, gen, noises.data(), 116 template struct MultinomialFunctor<GPUDevice, Eigen::half, int32>; 117 template struct MultinomialFunctor<GPUDevice, float, int32>; 118 template struct MultinomialFunctor<GPUDevice, double, int32>; 119 template struct MultinomialFunctor<GPUDevice, int32, int32>; 120 template struct MultinomialFunctor<GPUDevice, int64, int32> [all...] |
cwise_op_gpu_select.cu.cc | 27 struct SelectFunctor<GPUDevice, T> { 28 void operator()(const GPUDevice& d, typename TTypes<T>::Flat out, 38 struct SelectScalarFunctor<GPUDevice, T> { 39 void operator()(const GPUDevice& d, typename TTypes<T>::Flat out, 58 struct BatchSelectFunctor<GPUDevice, T> { 59 void operator()(const GPUDevice& d, 93 template struct SelectFunctor<GPUDevice, T>; \ 94 template struct SelectScalarFunctor<GPUDevice, T>; \ 95 template struct BatchSelectFunctor<GPUDevice, T>;
|
cast_op_gpu.cu.cc | 26 typedef Eigen::GpuDevice GPUDevice; 29 struct CastFunctor<GPUDevice, O, I> { 30 void operator()(const GPUDevice& d, typename TTypes<O>::Flat o, 32 Cast<GPUDevice, O, I>(d, o, i); 36 #define DEFINE(O, I) template struct CastFunctor<GPUDevice, O, I>
|
random_op.h | 46 typedef Eigen::GpuDevice GPUDevice; 49 struct FillPhiloxRandom<GPUDevice, Distribution> { 50 void operator()(OpKernelContext* ctx, const GPUDevice& d,
|
cast_op_impl_bool.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, bool);
|
cast_op_impl_double.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, double);
|
cast_op_impl_int16.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, int16);
|
cast_op_impl_int32.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, int32);
|
cast_op_impl_int64.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, int64);
|
cast_op_impl_int8.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, int8);
|
cast_op_impl_uint16.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, uint16);
|
cast_op_impl_uint8.cc | 21 typedef Eigen::GpuDevice GPUDevice; 32 CURRY_TYPES3(CAST_CASE, GPUDevice, uint8);
|
reduction_ops_all.cc | 39 ReductionOp<GPUDevice, bool, int32, Eigen::internal::AndReducer>); 45 ReductionOp<GPUDevice, bool, int64, Eigen::internal::AndReducer>);
|
reduction_ops_any.cc | 39 ReductionOp<GPUDevice, bool, int32, Eigen::internal::OrReducer>); 45 ReductionOp<GPUDevice, bool, int64, Eigen::internal::OrReducer>);
|
reverse_op_gpu.cu.cc | 26 typedef Eigen::GpuDevice GPUDevice; 29 template struct functor::Reverse<GPUDevice, T, DIM>;
|
reverse_sequence_op_gpu.cu.cc | 25 typedef Eigen::GpuDevice GPUDevice; 29 template struct functor::ReverseSequence<GPUDevice, T, Tlen, dims>;
|
training_ops_gpu.cu.cc | 25 typedef Eigen::GpuDevice GPUDevice; 29 struct ApplyGradientDescent<GPUDevice, T> { 30 void operator()(const GPUDevice& d, typename TTypes<T>::Flat var, 41 struct ApplyAdagrad<GPUDevice, T> { 42 void operator()(const GPUDevice& d, typename TTypes<T>::Flat var, 55 struct ApplyAdadelta<GPUDevice, T> { 56 void operator()(const GPUDevice& d, typename TTypes<T>::Flat var, 82 struct ApplyMomentum<GPUDevice, T> { 83 void operator()(const GPUDevice& d, typename TTypes<T>::Flat var [all...] |
transpose_functor_gpu.cu.cc | 29 typedef Eigen::GpuDevice GPUDevice; 57 void TransposeSimple(const GPUDevice& d, const Tensor& in, 97 static bool run(const Eigen::GpuDevice& d, const Tensor& in, 115 tensorflow::functor::SwapDimension1And2InTensor3<GPUDevice, T, 123 tensorflow::functor::SwapDimension1And2InTensor3<GPUDevice, T, 128 tensorflow::functor::SwapDimension0And2InTensor3<GPUDevice, T, 146 static bool run(const Eigen::GpuDevice& d, const Tensor& in, 158 static bool run(const Eigen::GpuDevice& d, const Tensor& in, 172 struct Transpose<GPUDevice, T, conjugate> [all...] |
compare_and_bitpack_op.h | 29 typedef Eigen::GpuDevice GPUDevice;
|
conv_ops_gpu_3.cu.cc | 34 typedef Eigen::GpuDevice GPUDevice; 434 struct TransformFilter<GPUDevice, T, int, NDIMS> { 435 typedef GPUDevice Device; 455 struct ReverseTransformFilter<GPUDevice, T, NDIMS> { 456 typedef GPUDevice Device; 476 struct PadInput<GPUDevice, T, int, NDIMS> { 477 typedef GPUDevice Device; 610 const GPUDevice& d, int tile_size_i, int tile_size_j, int total_tiles_count, 645 static void DoIt(const GPUDevice& d, int tile_size_i, int tile_size_j [all...] |
cwise_ops_gpu_gradients.cu.h | 36 typedef Eigen::GpuDevice GPUDevice; 40 // Partial specialization of SimpleBinaryFunctor<Device=GPUDevice, Functor>. 42 struct SimpleBinaryFunctor<GPUDevice, Functor> { 43 void operator()(const GPUDevice& d, typename Functor::tout_type out, 54 template struct SimpleBinaryFunctor<GPUDevice, F<T> >
|
l2loss_op_gpu.cu.cc | 29 typedef Eigen::GpuDevice GPUDevice; 40 class L2LossOp<GPUDevice, T> : public OpKernel { 56 Constants<GPUDevice> constants; 67 L2LossOp<GPUDevice, T>);
|
matrix_band_part_op_gpu.cu.cc | 29 typedef Eigen::GpuDevice GPUDevice; 52 struct MatrixBandPartFunctor<GPUDevice, Scalar> { 53 void operator()(OpKernelContext* context, const GPUDevice& device, 68 #define DEFINE_GPU_SPEC(T) template struct MatrixBandPartFunctor<GPUDevice, T>;
|
reduction_ops_gpu_bool.cu.cc | 25 typedef Eigen::GpuDevice GPUDevice; 37 template void ReduceFunctor<GPUDevice, REDUCER>::Reduce( \ 44 template void ReduceFunctor<GPUDevice, REDUCER>::FillIdentity( \ 45 const GPUDevice& d, TTypes<T>::Vec out, const REDUCER& reducer);
|
/external/tensorflow/tensorflow/core/framework/ |
register_types_traits.h | 22 typedef Eigen::GpuDevice GPUDevice; 61 struct proxy_type_pod<GPUDevice, 8> { 65 struct proxy_type_pod<GPUDevice, 4> { 69 struct proxy_type_pod<GPUDevice, 2> {
|