/external/tensorflow/tensorflow/contrib/distribute/python/ |
mirrored_strategy.py | 15 """Contrib version of MirroredStrategy.""" 30 CoreMirroredStrategy = mirrored_strategy.MirroredStrategy 35 class MirroredStrategy(distribute_lib.DistributionStrategy): 68 the model variables on these 8 GPUs. Then like in MirroredStrategy, each 103 super(MirroredStrategy, self).__init__(extended) 111 this contrib version of `MirroredStrategy`. 128 return super(MirroredStrategy, self).make_dataset_iterator(dataset) 137 contrib version of `MirroredStrategy`. 157 return super(MirroredStrategy, self).experimental_make_numpy_iterator( 162 """Implementation of (contrib) MirroredStrategy."" [all...] |
estimator_training_test.py | 263 mirrored_strategy.MirroredStrategy, 269 mirrored_strategy.MirroredStrategy, 296 mirrored_strategy.MirroredStrategy, 323 mirrored_strategy.MirroredStrategy, 328 mirrored_strategy.MirroredStrategy, 374 None, mirrored_strategy.MirroredStrategy, 422 mirrored_strategy.MirroredStrategy, 427 mirrored_strategy.MirroredStrategy,
|
combinations.py | 375 lambda: mirrored_lib.MirroredStrategy(["/cpu:0"])) 378 lambda: mirrored_lib.MirroredStrategy(["/gpu:0"]), 382 lambda: mirrored_lib.MirroredStrategy(["/gpu:0", "/cpu:0"]), 386 lambda: mirrored_lib.MirroredStrategy(["/gpu:0", "/gpu:1"]),
|
keras_multi_worker_test.py | 290 if (strategy_cls == mirrored_strategy.MirroredStrategy or 311 mirrored_strategy.MirroredStrategy, 338 mirrored_strategy.MirroredStrategy,
|
cross_device_ops_test.py | 396 lambda: mirrored_strategy.MirroredStrategy(num_gpus_per_worker=0), 400 lambda: mirrored_strategy.MirroredStrategy(num_gpus_per_worker=1), 404 lambda: mirrored_strategy.MirroredStrategy(num_gpus_per_worker=2),
|
keras_correctness_test_base.py | 279 # TODO(b/119257215): For MirroredStrategy, weights are not exactly the same, 282 mirrored_strategy.MirroredStrategy,
|
mirrored_strategy_multigpu_test.py | 15 """Multi-GPU tests for MirroredStrategy.""" [all...] |
keras_backward_compat_test.py | [all...] |
/external/tensorflow/tensorflow/contrib/distribute/ |
__init__.py | 28 from tensorflow.contrib.distribute.python.mirrored_strategy import MirroredStrategy 51 'MirroredStrategy',
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
template_mirrored_strategy_test.py | 15 """Tests for make_template used with MirroredStrategy.""" 44 strategy = mirrored_strategy.MirroredStrategy(["/cpu:0", "/gpu:0"])
|
/external/tensorflow/tensorflow/python/keras/mixed_precision/experimental/ |
loss_scale_optimizer_test.py | 41 return mirrored_strategy.MirroredStrategy(['cpu:0', 'gpu:0']) 43 return mirrored_strategy.MirroredStrategy(['cpu:0'])
|
keras_test.py | 129 return mirrored_strategy.MirroredStrategy(['cpu:0', 'gpu:0']) 131 return mirrored_strategy.MirroredStrategy(['cpu:0'])
|
autocast_variable_test.py | 54 return mirrored_strategy.MirroredStrategy(['cpu:0']).scope()
|
/external/tensorflow/tensorflow/contrib/distribute/python/examples/ |
keras_model_with_estimator.py | 15 """An example of training tf.keras Model using MirroredStrategy.""" 58 strategy = tf.contrib.distribute.MirroredStrategy(
|
simple_estimator_example.py | 65 distribution = tf.contrib.distribute.MirroredStrategy(
|
keras_mnist.py | 15 """An example training a Keras Model using MirroredStrategy and native APIs.""" 113 # Instantiate the MirroredStrategy object. If we don't specify `num_gpus` or 115 # TODO(priyag): Use `tf.distribute.MirroredStrategy` once available. 116 strategy = mirrored_strategy.MirroredStrategy(['/gpu:0', '/cpu:0'])
|
mnist_eager_multigpu.py | 15 """Run MNIST on multiple GPUs on using MirroredStrategy with eager execution. 20 MirroredStrategy because of a number overheads. Therefore, this is just a 97 strategy = tf.distribute.MirroredStrategy(devices)
|
/external/tensorflow/tensorflow/python/distribute/ |
mirrored_strategy.py | 15 """Class MirroredStrategy implementing DistributionStrategy.""" 79 # _call_for_each_replica is not a member of MirroredStrategy so that it is 80 # not allowed to use anything specific to MirroredStrategy and thus 221 # `AUTO` synchronization for `MirroredStrategy` is `ON_WRITE`. 421 @tf_export("distribute.MirroredStrategy") 422 class MirroredStrategy(distribute_lib.DistributionStrategy): 439 super(MirroredStrategy, self).__init__(extended) 443 """Implementation of MirroredStrategy.""" [all...] |
/external/tensorflow/tensorflow/python/training/tracking/ |
util_with_v1_optimizers_test.py | 296 strategy = mirrored_strategy.MirroredStrategy() 327 strategy = mirrored_strategy.MirroredStrategy() [all...] |