Home
last modified time | relevance | path

Searched defs:num_workers (Results 1 – 25 of 103) sorted by relevance

12345

/aosp_15_r20/external/tensorflow/tensorflow/core/distributed_runtime/
H A Dcollective_param_resolver_distributed_test.cc116 void DefineWorkers(int num_workers, int num_devices, in DefineWorkers()
163 void DefineCollectiveParams(int num_workers, int num_devices, in DefineCollectiveParams()
180 CollectiveParams* CreateCollectiveParams(int num_workers, int num_devices, in CreateCollectiveParams()
200 void IssueRequests(int num_workers, int num_devices) { in IssueRequests()
236 void ValidateCollectiveParams(int num_workers, int num_devices) { in ValidateCollectiveParams()
282 void RestartWorker(int worker_idx, int num_workers, int num_devices, in RestartWorker()
323 const int num_workers = 1; in TEST_F() local
332 const int num_workers = 2; in TEST_F() local
341 const int num_workers = 2; in TEST_F() local
354 const int num_workers = 2; in TEST_F() local
[all …]
/aosp_15_r20/external/tensorflow/tensorflow/python/distribute/v1/
H A Dall_reduce_test.py85 def _buildInput(self, num_workers, num_gpus): argument
140 def _buildRing(self, num_workers, num_gpus, subdiv): argument
145 def _testAllReduce(self, num_workers, num_gpus, shape, build_f): argument
161 def _testRingAllReduce(self, num_workers, num_gpus, shape, subdiv): argument
185 def _buildShuffle(self, num_workers, num_gpus, num_shards): argument
192 def _testShuffleAllReduce(self, num_workers, num_gpus, shape, num_shards): argument
212 def _buildRecursiveHD(self, num_workers, num_gpus): argument
218 def _testRecursiveHDAllReduce(self, num_workers, num_gpus, shape): argument
H A Dall_reduce.py189 def _ring_permutations(num_workers, num_subchunks, gpu_perm): argument
250 def build_ring_all_reduce(input_tensors, num_workers, num_subchunks, argument
/aosp_15_r20/external/libaom/av1/encoder/
H A Dethread.c306 int num_tiles, int num_workers) { in assign_tile_to_thread()
963 int num_workers = p_mt_info->num_workers; in av1_init_tile_thread_data() local
1080 void av1_create_workers(AV1_PRIMARY *ppi, int num_workers) { in av1_create_workers()
1258 const int num_workers, const int parallel_frame_count) { in compute_num_workers_per_frame()
1281 int num_workers = p_mt_info->num_workers; in prepare_fpmt_workers() local
1365 int num_workers = ppi->p_mt_info.p_num_workers; in launch_fpmt_workers() local
1422 int num_workers = ppi->p_mt_info.p_num_workers; in sync_fpmt_workers() local
1475 int num_workers) { in launch_workers()
1488 AV1_COMMON *const cm, int num_workers) { in sync_enc_workers()
1518 int num_workers) { in accumulate_counters_enc_workers()
[all …]
/aosp_15_r20/external/libaom/av1/common/
H A Dthread_common.c67 int width, int num_workers) { in av1_loop_filter_alloc()
151 int num_workers) { in av1_alloc_cdef_sync()
370 AV1_COMMON *const cm, int num_workers) { in sync_lf_workers()
434 AVxWorker *workers, int num_workers, in loop_filter_rows_mt()
494 int num_workers, AV1LfSync *lf_sync, in av1_loop_filter_frame_mt()
588 int num_workers, int num_rows_lr, in av1_loop_restoration_alloc()
898 AV1_COMMON *const cm, int num_workers) { in sync_lr_workers()
921 AVxWorker *workers, int num_workers, in foreach_rest_unit_in_planes_mt()
982 AVxWorker *workers, int num_workers, in av1_loop_restoration_filter_frame_mt()
1008 int num_workers) { in launch_cdef_workers()
[all …]
H A Dthread_common.h50 int num_workers; member
96 int num_workers; member
267 const int planes_to_lf[MAX_MB_PLANE], int num_workers, AV1LfSync *lf_sync, in loop_filter_frame_mt_init()
/aosp_15_r20/external/tensorflow/tensorflow/core/grappler/optimizers/data/
H A Dauto_shard.cc199 int64_t num_workers, int64_t index) { in AddShardNode()
441 int64_t num_workers, int64_t index) { in ProcessDatasetSourceNode()
482 const NodeDef* node, int64_t num_workers, int64_t index, in FindFuncAndTensorSliceDataset()
542 Status RecursivelyHandleOp(const NodeDef& node, int64_t num_workers, in RecursivelyHandleOp()
647 Status ShardByFile(const NodeDef& sink_node, int64_t num_workers, int64_t index, in ShardByFile()
704 Status ShardByData(const NodeDef& sink_node, int64_t num_workers, int64_t index, in ShardByData()
726 Status ShardByHint(const NodeDef& sink_node, int64_t num_workers, int64_t index, in ShardByHint()
756 Status ApplyAutoShard(const NodeDef& sink_node, int64_t num_workers, in ApplyAutoShard()
788 Status OptimizeGraph(const GrapplerItem& item, int64_t num_workers, in OptimizeGraph()
/aosp_15_r20/external/libvpx/vp9/encoder/
H A Dvp9_ethread.c81 static void create_enc_workers(VP9_COMP *cpi, int num_workers) { in create_enc_workers()
143 int num_workers) { in launch_enc_workers()
201 const int num_workers = VPXMIN(cpi->oxcf.max_threads, tile_cols); in vp9_encode_tiles_mt() local
469 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_fp_row_mt() local
555 int num_workers = cpi->num_workers ? cpi->num_workers : 1; in vp9_temporal_filter_row_mt() local
628 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_tiles_row_mt() local
/aosp_15_r20/external/tensorflow/tensorflow/python/distribute/
H A Dmulti_worker_test_base.py56 def _create_cluster(num_workers, argument
123 def create_in_process_cluster(num_workers, argument
328 def create_multi_process_cluster(num_workers, argument
355 num_workers=1, argument
432 def setUpClass(cls, num_workers=2, num_ps=1): # pylint: disable=g-missing-super-call argument
H A Dcombinations_test.py45 def testClusterParams(self, distribution, has_chief, num_workers): argument
54 def testClusterParamsHasDefault(self, distribution, has_chief, num_workers): argument
61 def testClusterParamsNoStrategy(self, v, has_chief, num_workers): argument
H A Dcombinations.py287 num_workers=1, argument
542 def decorator(self, has_chief, num_workers, num_ps, share_gpu, runner, argument
619 def _num_total_workers(has_chief, num_workers): argument
H A Dstrategy_combinations.py214 def _get_ps_strategy_creator(num_workers, argument
284 num_workers, argument
483 num_workers, argument
/aosp_15_r20/external/libvpx/vp9/common/
H A Dvp9_thread_common.c171 const int num_workers = VPXMIN(nworkers, VPXMIN(num_tile_cols, sb_rows)); in loop_filter_rows_mt() local
224 int num_workers, VP9LfSync *lf_sync) { in vp9_loop_filter_frame_mt()
244 int num_workers) { in vp9_lpf_mt_init()
281 int width, int num_workers) { in vp9_loop_filter_alloc()
/aosp_15_r20/external/tensorflow/tensorflow/python/data/experimental/ops/
H A Ddistribute.py70 def __init__(self, input_dataset, num_workers, index, num_replicas=None): argument
89 def _AutoShardDatasetV1(input_dataset, num_workers, index, num_replicas=None): # pylint: disable=i… argument
365 def batch_sizes_for_worker(global_batch_size, num_workers, argument
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/data/experimental/
H A Dauto_shard_dataset_op.cc47 int64_t index, num_workers, auto_shard_policy, num_replicas; in MakeDataset() local
80 RewriterConfig AutoShardDatasetOp::CreateConfig(int64_t num_workers, in CreateConfig()
H A Dauto_shard_dataset_op_test.cc32 AutoShardDatasetParams(T input_dataset_params, int64_t num_workers, in AutoShardDatasetParams()
231 Node* num_workers; in TEST_F() local
/aosp_15_r20/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/service/
H A Dmulti_process_cluster.py97 def _start_local_workers(self, num_workers, worker_tags=None): argument
102 def _start_remote_workers(self, num_workers, worker_tags=None): argument
H A Ddynamic_sharding_test.py277 def testChooseFromDatasets(self, num_workers): argument
315 def testConcatenate(self, num_workers): argument
/aosp_15_r20/external/tensorflow/tensorflow/core/common_runtime/
H A Dpermuter_test.cc50 void Init(int num_workers, int num_devices, in Init()
85 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
H A Dring_gatherer_test.cc47 void Init(int num_workers, int num_devices, DataType dtype, in Init()
79 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
H A Dhierarchical_tree_broadcaster_test.cc138 void Init(int num_workers, int num_devices, DataType dtype, in Init()
169 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
H A Dcollective_test_util.cc91 int num_workers, int num_devices_per_worker) { in CreateCPUDevices()
143 int num_workers, int num_devices_per_worker, DeviceType device_type) { in CreateCollectiveTestEnv()
H A Dring_reducer_test.cc84 void Init(int num_workers, int num_devices, DataType dtype, in Init()
116 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
/aosp_15_r20/external/googleapis/google/dataflow/v1beta3/
H A Dtemplates.proto158 int32 num_workers = 1; field
286 int32 num_workers = 11; field
/aosp_15_r20/external/rust/android-crates-io/crates/tokio/tests/
Drt_unstable_metrics.rs21 fn num_workers() { in num_workers() function
491 let num_workers = metrics.num_workers(); in worker_poll_count_histogram() localVariable

12345