HomeSort by relevance Sort by last modified time
    Searched refs:RingAllgather (Results 1 - 4 of 4) sorted by null

  /external/tensorflow/tensorflow/contrib/mpi_collectives/kernels/
ring.cc 44 // Generate all necessary specializations for RingAllgather.
45 template Status RingAllgather<CPUDevice, int>(OpKernelContext*, const Tensor*,
48 template Status RingAllgather<CPUDevice, long long>(OpKernelContext*,
52 template Status RingAllgather<CPUDevice, float>(OpKernelContext*, const Tensor*,
ring.cu.cc 65 // Generate all necessary specializations for RingAllgather.
66 template Status RingAllgather<GPUDevice, int>(OpKernelContext*, const Tensor*,
69 template Status RingAllgather<GPUDevice, long long>(OpKernelContext*,
73 template Status RingAllgather<GPUDevice, float>(OpKernelContext*, const Tensor*,
mpi_ops.cc 53 * The primary logic of the allreduce and allgather are in RingAllgather() and
96 extern template Status RingAllgather<GPUDevice, int>(OpKernelContext*,
100 extern template Status RingAllgather<GPUDevice, long long>(
102 extern template Status RingAllgather<GPUDevice, float>(
113 extern template Status RingAllgather<CPUDevice, int>(OpKernelContext*,
117 extern template Status RingAllgather<CPUDevice, long long>(
119 extern template Status RingAllgather<CPUDevice, float>(
471 status = on_gpu ? RingAllgather<GPU_DEVICE_IF_CUDA, float>(
473 : RingAllgather<CPUDevice, float>(
476 status = on_gpu ? RingAllgather<GPU_DEVICE_IF_CUDA, int>
    [all...]
ring.h 255 Status RingAllgather(OpKernelContext* context, const Tensor* input,

Completed in 57 milliseconds