tf.raw_ops.ConfigureDistributedTPU
Sets up the centralized structures for a distributed TPU system.
tf.raw_ops.ConfigureDistributedTPU(
embedding_config='', tpu_embedding_config='',
is_global_init=False, enable_whole_mesh_compilations=False,
compilation_failure_closes_chips=True, name=None
)
| Args | |
|---|---|
embedding_config | An optional string. Defaults to "". Reserved. Do not use. |
tpu_embedding_config | An optional string. Defaults to "". Serialized tensorflow.tpu.TPUEmbeddingConfiguration that describes the embedding lookups of the program. |
is_global_init | An optional bool. Defaults to False. Reserved. Do not use. |
enable_whole_mesh_compilations | An optional bool. Defaults to False. |
compilation_failure_closes_chips | An optional bool. Defaults to True. |
name | A name for the operation (optional). |
| Returns | |
|---|---|
A Tensor of type string. |
© 2020 The TensorFlow Authors. All rights reserved.
Licensed under the Creative Commons Attribution License 3.0.
Code samples licensed under the Apache 2.0 License.
https://www.tensorflow.org/versions/r2.4/api_docs/python/tf/raw_ops/ConfigureDistributedTPU