/aosp_15_r20/external/tensorflow/tensorflow/core/distributed_runtime/ |
H A D | collective_param_resolver_distributed_test.cc | 116 void DefineWorkers(int num_workers, int num_devices, in DefineWorkers() 163 void DefineCollectiveParams(int num_workers, int num_devices, in DefineCollectiveParams() 180 CollectiveParams* CreateCollectiveParams(int num_workers, int num_devices, in CreateCollectiveParams() 200 void IssueRequests(int num_workers, int num_devices) { in IssueRequests() 236 void ValidateCollectiveParams(int num_workers, int num_devices) { in ValidateCollectiveParams() 282 void RestartWorker(int worker_idx, int num_workers, int num_devices, in RestartWorker() 323 const int num_workers = 1; in TEST_F() local 332 const int num_workers = 2; in TEST_F() local 341 const int num_workers = 2; in TEST_F() local 354 const int num_workers = 2; in TEST_F() local [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/distribute/v1/ |
H A D | all_reduce_test.py | 85 def _buildInput(self, num_workers, num_gpus): argument 140 def _buildRing(self, num_workers, num_gpus, subdiv): argument 145 def _testAllReduce(self, num_workers, num_gpus, shape, build_f): argument 161 def _testRingAllReduce(self, num_workers, num_gpus, shape, subdiv): argument 185 def _buildShuffle(self, num_workers, num_gpus, num_shards): argument 192 def _testShuffleAllReduce(self, num_workers, num_gpus, shape, num_shards): argument 212 def _buildRecursiveHD(self, num_workers, num_gpus): argument 218 def _testRecursiveHDAllReduce(self, num_workers, num_gpus, shape): argument
|
H A D | all_reduce.py | 189 def _ring_permutations(num_workers, num_subchunks, gpu_perm): argument 250 def build_ring_all_reduce(input_tensors, num_workers, num_subchunks, argument
|
/aosp_15_r20/external/libaom/av1/encoder/ |
H A D | ethread.c | 306 int num_tiles, int num_workers) { in assign_tile_to_thread() 963 int num_workers = p_mt_info->num_workers; in av1_init_tile_thread_data() local 1080 void av1_create_workers(AV1_PRIMARY *ppi, int num_workers) { in av1_create_workers() 1258 const int num_workers, const int parallel_frame_count) { in compute_num_workers_per_frame() 1281 int num_workers = p_mt_info->num_workers; in prepare_fpmt_workers() local 1365 int num_workers = ppi->p_mt_info.p_num_workers; in launch_fpmt_workers() local 1422 int num_workers = ppi->p_mt_info.p_num_workers; in sync_fpmt_workers() local 1475 int num_workers) { in launch_workers() 1488 AV1_COMMON *const cm, int num_workers) { in sync_enc_workers() 1518 int num_workers) { in accumulate_counters_enc_workers() [all …]
|
/aosp_15_r20/external/libaom/av1/common/ |
H A D | thread_common.c | 67 int width, int num_workers) { in av1_loop_filter_alloc() 151 int num_workers) { in av1_alloc_cdef_sync() 370 AV1_COMMON *const cm, int num_workers) { in sync_lf_workers() 434 AVxWorker *workers, int num_workers, in loop_filter_rows_mt() 494 int num_workers, AV1LfSync *lf_sync, in av1_loop_filter_frame_mt() 588 int num_workers, int num_rows_lr, in av1_loop_restoration_alloc() 898 AV1_COMMON *const cm, int num_workers) { in sync_lr_workers() 921 AVxWorker *workers, int num_workers, in foreach_rest_unit_in_planes_mt() 982 AVxWorker *workers, int num_workers, in av1_loop_restoration_filter_frame_mt() 1008 int num_workers) { in launch_cdef_workers() [all …]
|
H A D | thread_common.h | 50 int num_workers; member 96 int num_workers; member 267 const int planes_to_lf[MAX_MB_PLANE], int num_workers, AV1LfSync *lf_sync, in loop_filter_frame_mt_init()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/grappler/optimizers/data/ |
H A D | auto_shard.cc | 199 int64_t num_workers, int64_t index) { in AddShardNode() 441 int64_t num_workers, int64_t index) { in ProcessDatasetSourceNode() 482 const NodeDef* node, int64_t num_workers, int64_t index, in FindFuncAndTensorSliceDataset() 542 Status RecursivelyHandleOp(const NodeDef& node, int64_t num_workers, in RecursivelyHandleOp() 647 Status ShardByFile(const NodeDef& sink_node, int64_t num_workers, int64_t index, in ShardByFile() 704 Status ShardByData(const NodeDef& sink_node, int64_t num_workers, int64_t index, in ShardByData() 726 Status ShardByHint(const NodeDef& sink_node, int64_t num_workers, int64_t index, in ShardByHint() 756 Status ApplyAutoShard(const NodeDef& sink_node, int64_t num_workers, in ApplyAutoShard() 788 Status OptimizeGraph(const GrapplerItem& item, int64_t num_workers, in OptimizeGraph()
|
/aosp_15_r20/external/libvpx/vp9/encoder/ |
H A D | vp9_ethread.c | 81 static void create_enc_workers(VP9_COMP *cpi, int num_workers) { in create_enc_workers() 143 int num_workers) { in launch_enc_workers() 201 const int num_workers = VPXMIN(cpi->oxcf.max_threads, tile_cols); in vp9_encode_tiles_mt() local 469 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_fp_row_mt() local 555 int num_workers = cpi->num_workers ? cpi->num_workers : 1; in vp9_temporal_filter_row_mt() local 628 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_tiles_row_mt() local
|
/aosp_15_r20/external/tensorflow/tensorflow/python/distribute/ |
H A D | multi_worker_test_base.py | 56 def _create_cluster(num_workers, argument 123 def create_in_process_cluster(num_workers, argument 328 def create_multi_process_cluster(num_workers, argument 355 num_workers=1, argument 432 def setUpClass(cls, num_workers=2, num_ps=1): # pylint: disable=g-missing-super-call argument
|
H A D | combinations_test.py | 45 def testClusterParams(self, distribution, has_chief, num_workers): argument 54 def testClusterParamsHasDefault(self, distribution, has_chief, num_workers): argument 61 def testClusterParamsNoStrategy(self, v, has_chief, num_workers): argument
|
H A D | combinations.py | 287 num_workers=1, argument 542 def decorator(self, has_chief, num_workers, num_ps, share_gpu, runner, argument 619 def _num_total_workers(has_chief, num_workers): argument
|
H A D | strategy_combinations.py | 214 def _get_ps_strategy_creator(num_workers, argument 284 num_workers, argument 483 num_workers, argument
|
/aosp_15_r20/external/libvpx/vp9/common/ |
H A D | vp9_thread_common.c | 171 const int num_workers = VPXMIN(nworkers, VPXMIN(num_tile_cols, sb_rows)); in loop_filter_rows_mt() local 224 int num_workers, VP9LfSync *lf_sync) { in vp9_loop_filter_frame_mt() 244 int num_workers) { in vp9_lpf_mt_init() 281 int width, int num_workers) { in vp9_loop_filter_alloc()
|
/aosp_15_r20/external/tensorflow/tensorflow/python/data/experimental/ops/ |
H A D | distribute.py | 70 def __init__(self, input_dataset, num_workers, index, num_replicas=None): argument 89 def _AutoShardDatasetV1(input_dataset, num_workers, index, num_replicas=None): # pylint: disable=i… argument 365 def batch_sizes_for_worker(global_batch_size, num_workers, argument
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
H A D | auto_shard_dataset_op.cc | 47 int64_t index, num_workers, auto_shard_policy, num_replicas; in MakeDataset() local 80 RewriterConfig AutoShardDatasetOp::CreateConfig(int64_t num_workers, in CreateConfig()
|
H A D | auto_shard_dataset_op_test.cc | 32 AutoShardDatasetParams(T input_dataset_params, int64_t num_workers, in AutoShardDatasetParams() 231 Node* num_workers; in TEST_F() local
|
/aosp_15_r20/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/service/ |
H A D | multi_process_cluster.py | 97 def _start_local_workers(self, num_workers, worker_tags=None): argument 102 def _start_remote_workers(self, num_workers, worker_tags=None): argument
|
H A D | dynamic_sharding_test.py | 277 def testChooseFromDatasets(self, num_workers): argument 315 def testConcatenate(self, num_workers): argument
|
/aosp_15_r20/external/tensorflow/tensorflow/core/common_runtime/ |
H A D | permuter_test.cc | 50 void Init(int num_workers, int num_devices, in Init() 85 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
|
H A D | ring_gatherer_test.cc | 47 void Init(int num_workers, int num_devices, DataType dtype, in Init() 79 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
|
H A D | hierarchical_tree_broadcaster_test.cc | 138 void Init(int num_workers, int num_devices, DataType dtype, in Init() 169 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
|
H A D | collective_test_util.cc | 91 int num_workers, int num_devices_per_worker) { in CreateCPUDevices() 143 int num_workers, int num_devices_per_worker, DeviceType device_type) { in CreateCollectiveTestEnv()
|
H A D | ring_reducer_test.cc | 84 void Init(int num_workers, int num_devices, DataType dtype, in Init() 116 void RunTest(DataType dtype, const DeviceType& device_type, int num_workers, in RunTest()
|
/aosp_15_r20/external/googleapis/google/dataflow/v1beta3/ |
H A D | templates.proto | 158 int32 num_workers = 1; field 286 int32 num_workers = 11; field
|
/aosp_15_r20/external/rust/android-crates-io/crates/tokio/tests/ |
D | rt_unstable_metrics.rs | 21 fn num_workers() { in num_workers() function 491 let num_workers = metrics.num_workers(); in worker_poll_count_histogram() localVariable
|