/external/tensorflow/tensorflow/core/grappler/inputs/ |
trivial_test_graph_input_yielder.h | 33 const std::vector<string>& device_names);
|
trivial_test_graph_input_yielder.cc | 33 const std::vector<string>& device_names) { 52 device_names[use_multiple_devices ? j % device_names.size() 59 device_names[use_multiple_devices ? j % device_names.size() 91 const std::vector<string>& device_names) 96 device_names_(device_names) {}
|
/external/tensorflow/tensorflow/core/grappler/clusters/ |
cluster.cc | 99 std::vector<string> device_names; local 100 device_names.reserve(devices_.size()); 102 device_names.push_back(device.first); 104 std::sort(device_names.begin(), device_names.end()); 105 return device_names;
|
/external/tensorflow/tensorflow/compiler/jit/ |
xla_cluster_util.h | 85 // operations placed in devices in `device_names`. For instance a cluster that 95 // - It is an error for `device_names` to contain more than one device of the 118 Status PickDeviceForXla(absl::Span<const string> device_names, 125 Status CanPickDeviceForXla(absl::Span<const string> device_names,
|
xla_cluster_util.cc | 251 Status PickDeviceForXlaImpl(absl::Span<const string> device_names, 269 TF_RET_CHECK(!device_names.empty()) << "No devices to choose from"; 273 for (absl::string_view device_name : device_names) { 290 "Multiple GPU devices ", absl::StrJoin(device_names, ", "))); 296 "Multiple CPU devices ", absl::StrJoin(device_names, ", "))); 302 "Multiple unknown devices ", absl::StrJoin(device_names, ", "))); 337 Status PickDeviceForXla(absl::Span<const string> device_names, 340 return PickDeviceForXlaImpl(device_names, allow_mixing_unknown_and_cpu, 345 Status CanPickDeviceForXla(absl::Span<const string> device_names, 348 return PickDeviceForXlaImpl(device_names, allow_mixing_unknown_and_cpu [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/ |
tf2xla_supported_ops.cc | 70 std::vector<string> device_names = XlaOpRegistry::BackendNames(); local 71 std::sort(device_names.begin(), device_names.end()); 79 absl::StrJoin(device_names, ",")},
|
/external/autotest/client/common_lib/cros/cfm/ |
cras_node_collector_unittest.py | 51 device_names = set([node.device_name for node in nodes]) 92 device_names = set([node.device_name for node in nodes]) 97 self.assertEquals(device_names, set(
|
/external/tensorflow/tensorflow/core/common_runtime/ |
device_resolver_local_test.cc | 51 cp.instance.device_names.push_back( 53 cp.instance.device_names.push_back( 71 cp.instance.device_names.push_back(
|
collective_param_resolver_local_test.cc | 54 cp.instance.device_names = shared_cp.instance.device_names; 73 EXPECT_EQ(ir.shared.instance.device_names, expected_device_order); 96 cp.instance.device_names.push_back(strings::StrCat( 171 cp->instance.device_names.push_back( 176 prl_->CompleteParamsAsync(cp->instance.device_names[0], cp, 189 ASSERT_EQ(cps[i].instance.device_names.size(), 3); 193 cps[i].instance.device_names[j]); 214 cp->instance.device_names.push_back(strings::StrCat( 229 prl_->CompleteParamsAsync(cp->instance.device_names[0], cp [all...] |
device_resolver_local.cc | 25 for (const string& device_name : ci_params.device_names) {
|
device_mgr.cc | 100 std::vector<StringPiece> device_names; local 102 device_names.push_back(itr.first); 105 << " all devices: " << str_util::Join(device_names, ", ");
|
collective_util.cc | 63 CHECK_GT(col_params.instance.device_names.size(), idx); 64 strings::StrAppend(&buf, col_params.instance.device_names[idx], "\n");
|
hierarchical_tree_broadcaster.cc | 82 col_params->instance.device_names[col_params->default_rank]; 84 // Precondition: device_names must be sorted so that all devices in 138 col_params->instance.device_names[col_params->source_rank] == 144 col_params->instance.device_names[device_count] == device_name; 167 if (col_params->instance.device_names[abs_di] == device_name) { 408 << col_params_->instance.device_names[dst_idx] << " subdiv=" << subdiv 410 col_ctx_->col_exec->PostToPeer(col_params_->instance.device_names[dst_idx], 426 << col_params_->instance.device_names[src_idx] << " to_device " 430 col_params_->instance.device_names[src_idx],
|
collective_param_resolver_local.cc | 162 CHECK_EQ(ip.device_names.size(), ip.task_names.size()); 163 CHECK_EQ(ip.device_names.size(), localities.size()); 164 for (int i = 0; i < ip.device_names.size(); ++i) { 166 DevRec* dr = &tdm[ip.device_names[i]]; 168 dr->device = ip.device_names[i]; 349 // Sort cp->instance.device_names lexicographically, but do by first 355 CHECK_EQ(cp->group.group_size, cp->instance.device_names.size()); 364 return cp->instance.device_names[a] < cp->instance.device_names[b]; 371 new_devs.push_back(cp->instance.device_names[pi]) [all...] |
/external/tensorflow/tensorflow/python/distribute/cluster_resolver/ |
cluster_resolver_test.py | 52 device_names = [ 60 for name in device_names 62 mock_eager_list_devices.return_value = device_names 72 device_names = [ 84 for name in device_names 86 mock_eager_list_devices.return_value = device_names 96 device_names = [ 108 for name in device_names 110 mock_eager_list_devices.return_value = device_names
|
tfconfig_cluster_resolver_test.py | 264 device_names = [ 276 for name in device_names 278 mock_eager_list_devices.return_value = device_names
|
tpu_cluster_resolver_test.py | 566 device_names = [ 578 name, 'TPU', 1024, 0) for name in device_names 591 device_names = [ 603 name, 'XLA', 1024, 0) for name in device_names 629 device_names = [ 641 name, 'TPU', 1024, 0) for name in device_names 643 mock_eager_list_devices.return_value = device_names
|
/external/tensorflow/tensorflow/core/framework/ |
collective.cc | 64 device_names.clear(); 65 device_names.assign(other.device_names.begin(), other.device_names.end()); 91 for (const auto& d : device_names) { 173 device_name(col_params.instance.device_names[col_params.default_rank]) {}
|
/external/tensorflow/tensorflow/python/distribute/ |
all_reduce_test.py | 95 device_names = [] 99 device_names.append(dn) 102 return input_tensors, device_names 107 input_tensors, device_names = self._buildInput(1, 1) 109 output_tensors = ar._build_ring_gather(input_tensors, device_names, 1, 114 input_tensors, device_names = self._buildInput(1, 4) 117 input_tensors, device_names, 2, pred_by_c_d, rank_by_c_d, math_ops.add)
|
/external/tensorflow/tensorflow/python/client/ |
session_list_devices_test.py | 74 device_names = set(d.name for d in devices) 76 '/job:worker/replica:0/task:0/device:CPU:0' in device_names) 78 '/job:worker/replica:0/task:1/device:CPU:0' in device_names)
|
/external/tensorflow/tensorflow/core/debug/ |
debug_grpc_testlib.h | 53 std::vector<string> device_names; member in class:tensorflow::test::final
|
debug_grpc_testlib.cc | 73 device_names.push_back(metadata.device()); 124 device_names.clear();
|
/external/tensorflow/tensorflow/core/distributed_runtime/ |
device_resolver_distributed.cc | 78 if (i < inst_params.device_names.size()) { 80 GetLocalityAsync(inst_params.device_names[i], inst_params.task_names[i],
|
/external/tensorflow/tensorflow/core/kernels/ |
collective_nccl_reducer_test.cc | 99 std::vector<string> device_names; local 107 device_names.push_back(device->name()); 129 col_params_.instance.device_names.push_back( 130 device_names[rank % num_gpus]); 135 rank, col_params_.instance.device_names[rank], this));
|
/external/tensorflow/tensorflow/python/grappler/ |
cluster.i | 289 std::vector<string> device_names; 294 device_names.push_back(name); 298 PyObject* dev = PyList_New(device_names.size()); 299 for (int i = 0; i < device_names.size(); ++i) { 300 PyList_SetItem(dev, i, PyString_FromString(device_names[i].c_str()));
|