Home
last modified time | relevance | path

Searched full:replicas (Results 1 – 25 of 452) sorted by relevance

12345678910>>...19

/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/service/
H A Dservice.cc171 << "Requested more replicas than there are devices."; in Service()
394 TF_ASSIGN_OR_RETURN(auto replicas, Replicas(*backend, device_handles[i])); in ExecuteParallelAndRegisterResult()
395 CHECK_EQ(replicas.size(), arguments[i].size()); in ExecuteParallelAndRegisterResult()
396 for (int64_t replica = 0, end = replicas.size(); replica < end; ++replica) { in ExecuteParallelAndRegisterResult()
397 device_assignment(replica, i) = replicas[replica]->device_ordinal(); in ExecuteParallelAndRegisterResult()
402 // Stream executors for the replicas of the current computation. in ExecuteParallelAndRegisterResult()
403 TF_ASSIGN_OR_RETURN(auto replicas, Replicas(*backend, device_handles[i])); in ExecuteParallelAndRegisterResult()
404 CHECK_EQ(replicas.size(), arguments[i].size()); in ExecuteParallelAndRegisterResult()
406 const int64_t n = replicas.size(); in ExecuteParallelAndRegisterResult()
410 backend->BorrowStream(replicas[replica])); in ExecuteParallelAndRegisterResult()
[all …]
H A Dhlo_replication_analysis.h31 // the same value across replicas or across partitions (depending on the value
43 // replicas or partitions.
54 // the given index across all replicas or partitions.
127 // partitions (not across replicas) on an SPMD partitioned module. This means
133 // If false, HloReplicationAnalysis runs across replicas.
137 // across replicas or partitions. This is provided by the caller as additional
145 // whether the instruction outputs the same value across replicas or
/aosp_15_r20/external/google-cloud-java/java-aiplatform/proto-google-cloud-aiplatform-v1/src/main/java/com/google/cloud/aiplatform/v1/
H A DAutomaticResources.java76 * Immutable. The minimum number of replicas this DeployedModel will be always
78 * deployed onto more replicas up to
80 * and as traffic decreases, some of these extra replicas may be freed. If the
99 * Immutable. The maximum number of replicas this DeployedModel may be
102 * the ability to scale the model to that many replicas is guaranteed (barring
104 * what its replicas at maximum may handle, a portion of the traffic will be
488 * Immutable. The minimum number of replicas this DeployedModel will be always
490 * deployed onto more replicas up to
492 * and as traffic decreases, some of these extra replicas may be freed. If the
508 * Immutable. The minimum number of replicas this DeployedModel will be always
[all …]
H A DAutomaticResourcesOrBuilder.java30 * Immutable. The minimum number of replicas this DeployedModel will be always
32 * deployed onto more replicas up to
34 * and as traffic decreases, some of these extra replicas may be freed. If the
48 * Immutable. The maximum number of replicas this DeployedModel may be
51 * the ability to scale the model to that many replicas is guaranteed (barring
53 * what its replicas at maximum may handle, a portion of the traffic will be
/aosp_15_r20/external/google-cloud-java/java-aiplatform/proto-google-cloud-aiplatform-v1beta1/src/main/java/com/google/cloud/aiplatform/v1beta1/
H A DAutomaticResources.java76 * Immutable. The minimum number of replicas this DeployedModel will be always
78 * deployed onto more replicas up to
80 * and as traffic decreases, some of these extra replicas may be freed. If the
99 * Immutable. The maximum number of replicas this DeployedModel may be
102 * the ability to scale the model to that many replicas is guaranteed (barring
104 * what its replicas at maximum may handle, a portion of the traffic will be
489 * Immutable. The minimum number of replicas this DeployedModel will be always
491 * deployed onto more replicas up to
493 * and as traffic decreases, some of these extra replicas may be freed. If the
509 * Immutable. The minimum number of replicas this DeployedModel will be always
[all …]
H A DAutomaticResourcesOrBuilder.java30 * Immutable. The minimum number of replicas this DeployedModel will be always
32 * deployed onto more replicas up to
34 * and as traffic decreases, some of these extra replicas may be freed. If the
48 * Immutable. The maximum number of replicas this DeployedModel may be
51 * the ability to scale the model to that many replicas is guaranteed (barring
53 * what its replicas at maximum may handle, a portion of the traffic will be
/aosp_15_r20/external/tensorflow/tensorflow/python/distribute/
H A Ddistribute_lib.py68 worker may contain one or more replicas, but contains at least one
72 replicas on that worker.
76 updates the model variables independently. You may also have replicas
81 `tf.distribute.experimental.ParameterServerStrategy`). All replicas that want
103 the replicas (like `reduce_to()`). By default you start in a _replica context_
121 on all replicas are the same. `Mirrored` values are kept synchronized by the
129 replicas, like `strategy.run(fn, args=[w])` with an
141 from all replicas before applying the update. _All-reduce_ is an algorithm for
182 `all_reduce` method for aggregating across all replicas.
465 Replicas will dequeue from the local Dataset on their worker.
[all …]
H A Dcustom_training_loop_input_test.py57 Due to the number of replicas in the strategy, the output may have a
372 # on the number of replicas.
506 # This assumes that there are exactly 2 replicas
530 # This assumes that there are exactly 2 replicas
559 # This assumes that there are exactly 2 replicas
582 # This assumes that there are exactly 2 replicas
604 # This assumes that there are exactly 2 replicas
657 # This assumes that there are exactly 2 replicas
682 # This assumes that there are exactly 2 replicas
726 # This assumes that there are exactly 2 replicas
[all …]
H A Dcentral_storage_strategy.py131 replicas, with the provided "per replica" arguments specific to that device.
146 """Reduce `value` across replicas.
149 per-example loss, the batch will be divided across all the replicas. This
150 function allows you to aggregate across replicas and optionally also across
152 replicas, values for examples `[0, 1, 2, 3]` will be on replica 0 and
154 aggregate across replicas, returning `[0+4, 1+5, 2+6, 3+7]`. This is useful
162 that the resulting shape is consistent across replicas. So if the last
203 `None` to only reduce across replicas (e.g. if the tensor has no batch
H A Dinput_lib.py140 """Returns the next input from the iterator for all replicas.
160 the next input for all replicas.
196 """Returns a `tf.experimental.Optional` that contains the next value for all replicas.
257 number of replicas in sync) and autosharding (i.e. the
495 """Calcualates the number of replicas that have values.
621 """Returns the next input from the iterator for all replicas."""
653 replicas = []
661 # Make `replicas` a flat list of values across all replicas.
662 replicas.extend(self._iterators[i].get_next_as_list(new_name))
663 return _create_per_replica(replicas, self._strategy)
[all …]
/aosp_15_r20/external/google-cloud-java/java-aiplatform/proto-google-cloud-aiplatform-v1/src/main/proto/google/cloud/aiplatform/v1/
H A Dmachine_resources.proto67 // Required. Immutable. The minimum number of machine replicas this
72 // deployed onto more replicas, and as traffic decreases, some of these extra
73 // replicas may be freed.
79 // Immutable. The maximum number of replicas this DeployedModel may be
82 // the ability to scale the model to that many replicas is guaranteed (barring
84 // what its replicas at maximum may handle, a portion of the traffic will be
126 // Immutable. The minimum number of replicas this DeployedModel will be always
128 // deployed onto more replicas up to
130 // and as traffic decreases, some of these extra replicas may be freed. If the
134 // Immutable. The maximum number of replicas this DeployedModel may be
[all …]
/aosp_15_r20/external/google-cloud-java/java-aiplatform/proto-google-cloud-aiplatform-v1beta1/src/main/proto/google/cloud/aiplatform/v1beta1/
H A Dmachine_resources.proto67 // Required. Immutable. The minimum number of machine replicas this
72 // deployed onto more replicas, and as traffic decreases, some of these extra
73 // replicas may be freed.
79 // Immutable. The maximum number of replicas this DeployedModel may be
82 // the ability to scale the model to that many replicas is guaranteed (barring
84 // what its replicas at maximum may handle, a portion of the traffic will be
126 // Immutable. The minimum number of replicas this DeployedModel will be always
128 // deployed onto more replicas up to
130 // and as traffic decreases, some of these extra replicas may be freed. If the
134 // Immutable. The maximum number of replicas this DeployedModel may be
[all …]
/aosp_15_r20/external/tensorflow/tensorflow/python/training/
H A Dsync_replicas_optimizer.py16 """Synchronize replicas for training."""
35 # Please note that the gradients from replicas are averaged instead of summed
37 # rate according to the number of replicas. This change is introduced to be
52 optimizer avoids stale gradients by collecting gradients from all replicas,
54 which replicas can fetch the new variables and continue.
88 For the replicas:
104 # Wrap the optimizer with sync_replicas_optimizer with 50 replicas: at each
106 # Note that if you want to have 2 backup replicas, you can change
108 # replicas you started in your job.
161 replicas_to_aggregate: number of replicas to aggregate for each variable
[all …]
/aosp_15_r20/external/googleapis/google/cloud/aiplatform/v1/
H A Dmachine_resources.proto71 // Required. Immutable. The minimum number of machine replicas this
76 // deployed onto more replicas, and as traffic decreases, some of these extra
77 // replicas may be freed.
83 // Immutable. The maximum number of replicas this DeployedModel may be
86 // the ability to scale the model to that many replicas is guaranteed (barring
88 // what its replicas at maximum may handle, a portion of the traffic will be
130 // Immutable. The minimum number of replicas this DeployedModel will be always
132 // deployed onto more replicas up to
134 // and as traffic decreases, some of these extra replicas may be freed. If the
138 // Immutable. The maximum number of replicas this DeployedModel may be
[all …]
/aosp_15_r20/external/googleapis/google/cloud/aiplatform/v1beta1/
H A Dmachine_resources.proto71 // Required. Immutable. The minimum number of machine replicas this
76 // deployed onto more replicas, and as traffic decreases, some of these extra
77 // replicas may be freed.
83 // Immutable. The maximum number of replicas this DeployedModel may be
86 // the ability to scale the model to that many replicas is guaranteed (barring
88 // what its replicas at maximum may handle, a portion of the traffic will be
130 // Immutable. The minimum number of replicas this DeployedModel will be always
132 // deployed onto more replicas up to
134 // and as traffic decreases, some of these extra replicas may be freed. If the
138 // Immutable. The maximum number of replicas this DeployedModel may be
[all …]
/aosp_15_r20/external/google-cloud-java/java-secretmanager/proto-google-cloud-secretmanager-v1beta1/src/main/java/com/google/cloud/secretmanager/v1beta1/
H A DReplication.java611 * Required. The list of Replicas for this [Secret][google.cloud.secrets.v1beta1.Secret].
616 …* repeated .google.cloud.secrets.v1beta1.Replication.UserManaged.Replica replicas = 1 [(.google.ap…
625 * Required. The list of Replicas for this [Secret][google.cloud.secrets.v1beta1.Secret].
630 …* repeated .google.cloud.secrets.v1beta1.Replication.UserManaged.Replica replicas = 1 [(.google.ap…
638 * Required. The list of Replicas for this [Secret][google.cloud.secrets.v1beta1.Secret].
643 …* repeated .google.cloud.secrets.v1beta1.Replication.UserManaged.Replica replicas = 1 [(.google.ap…
651 * Required. The list of Replicas for this [Secret][google.cloud.secrets.v1beta1.Secret].
656 …* repeated .google.cloud.secrets.v1beta1.Replication.UserManaged.Replica replicas = 1 [(.google.ap…
667 * Required. The list of Replicas for this [Secret][google.cloud.secrets.v1beta1.Secret].
672 …* repeated .google.cloud.secrets.v1beta1.Replication.UserManaged.Replica replicas = 1 [(.google.ap…
[all …]
/aosp_15_r20/external/tensorflow/tensorflow/python/tpu/
H A Ddevice_assignment.py32 """Computes a nested dict which maps task and logical core to replicas."""
49 for core, replicas in core_to_replicas.items():
50 core_to_sorted_replicas[core] = sorted(replicas)
114 """The number of replicas of the computation."""
139 A sorted list of the replicas that are attached to that task and
326 # Form rings for replicas and model-parallel cores.
328 # Form meshes for replicas and/or model-parallel cores.
362 num_replicas: The number of computation replicas to run. The replicas will
377 ValueError: If computation's replicas cannot fit into the TPU topology.
433 "requested {} replicas but only {} replicas with shape {} and "
[all …]
/aosp_15_r20/external/google-cloud-java/java-secretmanager/proto-google-cloud-secretmanager-v1/src/main/java/com/google/cloud/secretmanager/v1/
H A DReplication.java973 * Required. The list of Replicas for this
979 …* repeated .google.cloud.secretmanager.v1.Replication.UserManaged.Replica replicas = 1 [(.google.a…
988 * Required. The list of Replicas for this
994 …* repeated .google.cloud.secretmanager.v1.Replication.UserManaged.Replica replicas = 1 [(.google.a…
1002 * Required. The list of Replicas for this
1008 …* repeated .google.cloud.secretmanager.v1.Replication.UserManaged.Replica replicas = 1 [(.google.a…
1016 * Required. The list of Replicas for this
1022 …* repeated .google.cloud.secretmanager.v1.Replication.UserManaged.Replica replicas = 1 [(.google.a…
1032 * Required. The list of Replicas for this
1038 …* repeated .google.cloud.secretmanager.v1.Replication.UserManaged.Replica replicas = 1 [(.google.a…
[all …]
/aosp_15_r20/external/google-cloud-java/java-gke-multi-cloud/proto-google-cloud-gke-multi-cloud-v1/src/main/java/com/google/cloud/gkemulticloud/v1/
H A DAzureControlPlane.java93 * Required. The Kubernetes version to run on control plane replicas
120 * Required. The Kubernetes version to run on control plane replicas
738 * Optional. Configuration for where to place the control plane replicas.
741 * to the three control plane replicas as evenly as possible.
757 * Optional. Configuration for where to place the control plane replicas.
760 * to the three control plane replicas as evenly as possible.
776 * Optional. Configuration for where to place the control plane replicas.
779 * to the three control plane replicas as evenly as possible.
794 * Optional. Configuration for where to place the control plane replicas.
797 * to the three control plane replicas as evenly as possible.
[all …]
H A DAzureControlPlaneOrBuilder.java30 * Required. The Kubernetes version to run on control plane replicas
46 * Required. The Kubernetes version to run on control plane replicas
463 * Optional. Configuration for where to place the control plane replicas.
466 * to the three control plane replicas as evenly as possible.
478 * Optional. Configuration for where to place the control plane replicas.
481 * to the three control plane replicas as evenly as possible.
493 * Optional. Configuration for where to place the control plane replicas.
496 * to the three control plane replicas as evenly as possible.
508 * Optional. Configuration for where to place the control plane replicas.
511 * to the three control plane replicas as evenly as possible.
[all …]
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/nn/parallel/
H A Ddata_parallel.h108 const std::vector<std::shared_ptr<ModuleType>>& replicas, in replicate_grad_edges() argument
115 autograd::set_history(replicas[i]->parameters_[parameter.key()], grad_fn); in replicate_grad_edges()
125 autograd::set_history(replicas[i]->buffers_[buffer.key()], grad_fn); in replicate_grad_edges()
133 for (auto& replica : replicas) { in replicate_grad_edges()
152 std::vector<std::shared_ptr<ModuleType>> replicas; in replicate() local
153 replicas.reserve(devices.size()); in replicate()
155 replicas.push_back( in replicate()
160 replicate_grad_edges(module, replicas, devices); in replicate()
161 return replicas; in replicate()
288 auto replicas = replicate(module, *devices); variable
[all …]
/aosp_15_r20/external/tensorflow/tensorflow/compiler/mlir/tensorflow/transforms/
H A Dreplicate_to_island.cc89 // for different replicas. Some ops, such as XlaHostCompute op or TPU Embedding
143 int num_replicas, llvm::SmallVectorImpl<tf_executor::IslandOp>& replicas) { in ExpandReplicateIntoReplicas() argument
144 replicas.reserve(num_replicas); in ExpandReplicateIntoReplicas()
180 replicas.push_back(replica); in ExpandReplicateIntoReplicas()
244 llvm::SmallVector<tf_executor::IslandOp, 8> replicas; in CreateIslandsFromReplicate() local
246 replicate_op, num_replicas, replicas))) in CreateIslandsFromReplicate()
252 for (auto replica_and_idx : llvm::enumerate(replicas)) in CreateIslandsFromReplicate()
263 // Add sink island to pin all replicas as a control dependency if there is a in CreateIslandsFromReplicate()
267 for (auto& replica : replicas) island_operands.push_back(replica.control()); in CreateIslandsFromReplicate()
280 // Replicas with no uses should be pinned to a graph fetch so they still in CreateIslandsFromReplicate()
[all …]
/aosp_15_r20/external/pytorch/test/jit/
H A Dtest_data_parallel.py81 def check_replicas(self, module, replicas, input_shape=(2, 2)): argument
84 for i, replica in enumerate(replicas):
95 replicas = dp.replicate(module, {0, 1})
96 self.check_replicas(module, replicas)
103 replicas = dp.replicate(module, {0, 1})
104 self.check_replicas(module, replicas)
109 replicas = dp.replicate(module, {0, 1})
110 self.check_replicas(module, replicas)
/aosp_15_r20/external/tensorflow/tensorflow/core/tpu/graph_rewrite/
H A Ddistributed_tpu_rewrite_pass.h30 // controls how many replicas of the computation to create. Replicas are mostly
31 // independent; replicas can only communicate using the CrossReplicaSum
52 // replicas, the first six arguments to TPUReplicate will be:
56 // If there are two distributed inputs (E, F) and two replicas, the following
60 // replicas receive a copy of each of these inputs.
64 // For example, for a computation with two replicas, three per-replica inputs
218 // replicas).
230 // Indices of mirrored variables across replicas, which should be
259 // * num_replicas: the number of replicas requested
301 // Also verifies that all replicas have identical input shapes for their
[all …]
/aosp_15_r20/external/pytorch/torch/nn/parallel/
H A Ddata_parallel.py84 is done on the replicas which are destroyed after ``forward``. However,
192 replicas = self.replicate(self.module, self.device_ids[: len(inputs)])
193 outputs = self.parallel_apply(replicas, inputs, module_kwargs)
210 self, replicas: Sequence[T], inputs: Sequence[Any], kwargs: Any
213 replicas, inputs, kwargs, self.device_ids[: len(replicas)]
283 replicas = replicate(module, used_device_ids)
284 outputs = parallel_apply(replicas, inputs, module_kwargs, used_device_ids)

12345678910>>...19