/external/tensorflow/tensorflow/core/debug/ |
debug_gateway.cc | 69 const Tensor* src_tensor, OpKernelContext* ctx, 77 if (src_tensor->IsInitialized() && DataTypeSize(src_tensor->dtype()) > 0) { 84 Tensor cpu_tensor(cpu_allocator, src_tensor->dtype(), src_tensor->shape()); 98 src_tensor, "TensorCopy", device, &cpu_tensor, 111 cpu_tensor.UnsafeCopyFromInternal(*src_tensor, src_tensor->dtype(), 112 src_tensor->shape()); 118 copy_done_cb(src_tensor); [all...] |
debug_gateway.h | 77 const Tensor* src_tensor, OpKernelContext* ctx,
|
/external/tensorflow/tensorflow/core/common_runtime/sycl/ |
sycl_util.h | 34 Tensor const& src_tensor, Tensor* dst_tensor) { 35 const size_t size = src_tensor.TotalBytes(); 37 void const* src_ptr = GetBase(&src_tensor); 41 switch (src_tensor.dtype()) { 73 LOG(FATAL) << "Unknown data type " << src_tensor.dtype();
|
/external/tensorflow/tensorflow/core/kernels/ |
debug_ops.h | 62 const Tensor& src_tensor = context->input(0); variable 64 if (src_tensor.IsInitialized() && 65 DataTypeCanUseMemcpy(src_tensor.dtype()) && 69 OP_REQUIRES_OK(context, context->allocate_output(0, src_tensor.shape(), 83 device, device_ctxt, &src_tensor, copied_tensor, 88 *copied_tensor = tensor::DeepCopy(src_tensor); 97 SYCLmemcpy(context->eigen_sycl_device(), src_tensor, copied_tensor); 99 *copied_tensor = tensor::DeepCopy(src_tensor); 102 *copied_tensor = tensor::DeepCopy(src_tensor); 107 context->set_output(0, src_tensor); [all...] |
mkl_softmax_op.cc | 53 // src_tensor now points to the 0-th input of global data struct "context" 55 const Tensor& src_tensor = MklGetInput(context, src_idx); variable 61 // src_dims is the dimenstion of src_tensor 65 : src_tensor.shape(); 84 // Basically following two functions maps the TF "src_tensor" to mkl 89 src.SetUsrMem(src_md, &src_tensor);
|
mkl_relu_op.cc | 387 const Tensor& src_tensor = MklGetInput(context, src_index); variable 392 if (src_tensor.dims() == 0) { 406 auto src_dims = TFShapeToMklDnnDims(src_tensor.shape()); 411 src.SetUsrMem(src_md, &src_tensor); 436 tf_shape_dst = src_tensor.shape(); 483 const Tensor& src_tensor = MklGetInput(context, src_index); local 491 int src_dims_size = src_tensor.dims(); 513 auto src_dims = TFShapeToMklDnnDims(src_tensor.shape()); 541 auto src_dims = TFShapeToMklDnnDimsInNCHW(src_tensor.shape(), 552 src.SetUsrMem(src_md, &src_tensor); 642 const Tensor& src_tensor = MklGetInput(context, src_index); local 672 const Tensor& src_tensor = MklGetInput(context, src_index); local 705 const Tensor& src_tensor = MklGetInput(context, src_index); local 739 const Tensor& src_tensor = MklGetInput(context, src_index); local 777 const Tensor& src_tensor = MklGetInput(context, src_index); local 810 const Tensor& src_tensor = MklGetInput(context, src_index); local [all...] |
mkl_fused_batch_norm_op.cc | 712 const Tensor& src_tensor = MklGetInput(context, kSrcIndex); variable 1055 const Tensor& src_tensor = MklGetInput(context, kSrcIndex); variable [all...] |
mkl_lrn_op.cc | 765 const Tensor& src_tensor = MklGetInput(context, kIdxInput); 777 MklDefaultToEigen(context, src_tensor); 782 ConvertMklToTF<T>(context, src_tensor, src_dnn_shape); 794 TensorShape tf_output_shape = src_tensor.shape(); 803 src_dnn_data.SetUsrMem(src_md, &src_tensor); [all...] |
mkl_conv_ops.cc | 517 const Tensor& src_tensor = MklGetInput(context, kInputIndex_Src); variable 577 src.SetUsrMem(src_md, &src_tensor); [all...] |