tf.distribute.cluster_resolver.UnionResolver

View source on GitHub

Performs a union on underlying ClusterResolvers.

Inherits From: ClusterResolver

tf.distribute.cluster_resolver.UnionResolver(
    *args, **kwargs
)

This class performs a union given two or more existing ClusterResolvers. It merges the underlying ClusterResolvers, and returns one unified ClusterSpec when cluster_spec is called. The details of the merge function is documented in the cluster_spec function.

For additional ClusterResolver properties such as task type, task index, rpc layer, environment, etc..., we will return the value from the first ClusterResolver in the union.

Args:

Attributes:

Raises:

Methods

cluster_spec

View source

cluster_spec()

Returns a union of all the ClusterSpecs from the ClusterResolvers.

Returns:

A ClusterSpec containing host information merged from all the underlying ClusterResolvers.

Raises:

Note: If there are multiple ClusterResolvers exposing ClusterSpecs with the same job name, we will merge the list/dict of workers.

If all underlying ClusterSpecs expose the set of workers as lists, we will concatenate the lists of workers, starting with the list of workers from the first ClusterResolver passed into the constructor.

If any of the ClusterSpecs expose the set of workers as a dict, we will treat all the sets of workers as dicts (even if they are returned as lists) and will only merge them into a dict if there is no conflicting keys. If there is a conflicting key, we will raise a KeyError.

master

View source

master(
    task_type=None, task_id=None, rpc_layer=None
)

Returns the master address to use when creating a session.

This usually returns the master from the first ClusterResolver passed in, but you can override this by specifying the task_type and task_id.

Args:

Returns:

The name or URL of the session master.

num_accelerators

View source

num_accelerators(
    task_type=None, task_id=None, config_proto=None
)

Returns the number of accelerator cores per worker.

This returns the number of accelerator cores (such as GPUs and TPUs) available per worker.

Optionally, we allow callers to specify the task_type, and task_id, for if they want to target a specific TensorFlow process to query the number of accelerators. This is to support heterogenous environments, where the number of accelerators cores per host is different.

Args:

Returns:

A map of accelerator types to number of cores.