/aosp_15_r20/external/tensorflow/tensorflow/python/data/ops/ |
H A D | dataset_ops.py | 240 for input_dataset in self._inputs(): 242 if isinstance(input_dataset, DatasetV1): 246 if hasattr(input_dataset, "_dataset"): 247 if not isinstance(input_dataset._dataset, DatasetV2): 252 input_options = input_dataset._dataset._options_attr 253 elif isinstance(input_dataset, DatasetV2): 254 input_options = input_dataset._options_attr 3221 input_dataset = self 3223 input_dataset = input_dataset.enumerate(name=name) 3233 input_dataset=input_dataset, [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/data/experimental/ops/ |
H A D | distribute.py | 70 def __init__(self, input_dataset, num_workers, index, num_replicas=None): argument 71 self._input_dataset = input_dataset 73 self._element_spec = input_dataset.element_spec 79 input_dataset.options().experimental_distribute.auto_shard_policy), 82 super(_AutoShardDataset, self).__init__(input_dataset, variant_tensor) 89 def _AutoShardDatasetV1(input_dataset, num_workers, index, num_replicas=None): # pylint: disable=i… argument 91 _AutoShardDataset(input_dataset, num_workers, index, num_replicas)) 121 def __init__(self, input_dataset, batch_sizes, drop_remainder=False): argument 134 self._input_dataset = input_dataset 144 dataset_ops.get_structure(input_dataset)) [all …]
|
H A D | testing.py | 121 def __init__(self, input_dataset, transformations): argument 123 self._input_dataset = input_dataset 135 super(_AssertNextDataset, self).__init__(input_dataset, variant_tensor) 141 def __init__(self, input_dataset, transformations): argument 143 self._input_dataset = input_dataset 172 super(_AssertPrevDataset, self).__init__(input_dataset, variant_tensor) 178 def __init__(self, input_dataset): argument 180 self._input_dataset = input_dataset 185 super(_NonSerializableDataset, self).__init__(input_dataset, variant_tensor) 191 def __init__(self, input_dataset, sleep_microseconds): argument [all …]
|
H A D | batching.py | 304 def __init__(self, input_dataset, batch_size, row_shape): argument 307 dataset_ops.get_legacy_output_types(input_dataset), dtypes.DType): 312 self._input_dataset = input_dataset 317 dataset_ops.get_legacy_output_types(input_dataset)) 324 super(_DenseToSparseBatchDataset, self).__init__(input_dataset, 335 def __init__(self, input_dataset, map_func, batch_size, num_parallel_calls, argument 337 self._input_dataset = input_dataset 342 dataset=input_dataset, 375 super(_MapAndBatchDataset, self).__init__(input_dataset, variant_tensor) 396 def __init__(self, input_dataset, row_splits_dtype): argument [all …]
|
H A D | grouping.py | 263 def __init__(self, input_dataset, key_func, reducer): argument 265 self._input_dataset = input_dataset 266 self._make_key_func(key_func, input_dataset) 268 self._make_reduce_func(reducer.reduce_func, input_dataset) 281 super(_GroupByReducerDataset, self).__init__(input_dataset, variant_tensor) 283 def _make_key_func(self, key_func, input_dataset): argument 286 key_func, self._transformation_name(), dataset=input_dataset) 303 def _make_reduce_func(self, reduce_func, input_dataset): argument 318 input_structure=(self._state_structure, input_dataset.element_spec),
|
H A D | prefetching_ops.py | 91 def __init__(self, input_dataset, target_device, source_device="/cpu:0"): argument 99 self._input_dataset = input_dataset._apply_debug_options() # pylint: disable=protected-access 214 super(_CopyToDeviceDataset, self).__init__(input_dataset, variant_tensor) 234 def __init__(self, input_dataset, map_func, use_inter_op_parallelism=True): argument 236 self._input_dataset = input_dataset 242 dataset=input_dataset, 250 super(_MapOnGpuDataset, self).__init__(input_dataset, variant_tensor)
|
/aosp_15_r20/external/tensorflow/tensorflow/python/data/kernel_tests/ |
H A D | concatenate_test.py | 42 input_dataset = dataset_ops.Dataset.from_tensor_slices(input_components) 45 concatenated = input_dataset.concatenate(dataset_to_concatenate) 74 input_dataset = dataset_ops.Dataset.from_tensor_slices(input_components) 77 concatenated = input_dataset.concatenate(dataset_to_concatenate) 106 input_dataset = dataset_ops.Dataset.from_tensor_slices(input_components) 111 input_dataset.concatenate(dataset_to_concatenate) 124 input_dataset = dataset_ops.Dataset.from_tensor_slices(input_components) 129 input_dataset.concatenate(dataset_to_concatenate) 140 input_dataset = dataset_ops.Dataset.from_tensor_slices(input_components) 145 input_dataset.concatenate(dataset_to_concatenate) [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
H A D | auto_shard_dataset_op_test.cc | 230 Node* input_dataset; in TEST_F() local 253 .Finalize(&graph, &input_dataset)); in TEST_F() 254 (*input_dataset->mutable_def()->mutable_experimental_type()) = in TEST_F() 270 .Input(input_dataset) in TEST_F() 287 Node* input_dataset; in TEST_F() local 309 .Finalize(&graph, &input_dataset)); in TEST_F() 310 (*input_dataset->mutable_def()->mutable_experimental_type()) = in TEST_F() 321 .Input(input_dataset) in TEST_F() 337 Node* input_dataset; in TEST_F() local 360 .Finalize(&graph, &input_dataset)); in TEST_F() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/ |
H A D | tf_record_writer_test.py | 42 input_dataset = readers.TFRecordDataset([filename], compression_type) 44 compression_type).write(input_dataset) 94 input_dataset = dataset_ops.Dataset.from_tensors(10) 96 writers.TFRecordWriter(self._outputFilename(), "").write(input_dataset) 100 input_dataset = dataset_ops.Dataset.from_tensors([["hello"], ["world"]]) 102 writers.TFRecordWriter(self._outputFilename(), "").write(input_dataset) 107 input_dataset = readers.TFRecordDataset(self._createFile()) 108 return writers.TFRecordWriter(self._outputFilename()).write(input_dataset)
|
/aosp_15_r20/external/tensorflow/tensorflow/core/ops/compat/ops_history_v2/ |
H A D | PrefetchDataset.pbtxt | 4 name: "input_dataset" 32 name: "input_dataset" 59 name: "input_dataset" 100 name: "input_dataset" 148 name: "input_dataset" 203 name: "input_dataset" 265 name: "input_dataset"
|
H A D | ParallelMapDataset.pbtxt | 4 name: "input_dataset" 45 name: "input_dataset" 85 name: "input_dataset" 132 name: "input_dataset" 186 name: "input_dataset" 247 name: "input_dataset" 315 name: "input_dataset" 390 name: "input_dataset"
|
H A D | MapDataset.pbtxt | 4 name: "input_dataset" 41 name: "input_dataset" 77 name: "input_dataset" 120 name: "input_dataset" 170 name: "input_dataset" 227 name: "input_dataset" 291 name: "input_dataset"
|
H A D | ModelDataset.pbtxt | 4 name: "input_dataset" 27 name: "input_dataset" 64 name: "input_dataset" 108 name: "input_dataset" 152 name: "input_dataset" 203 name: "input_dataset"
|
H A D | ShuffleDataset.pbtxt | 4 name: "input_dataset" 40 name: "input_dataset" 75 name: "input_dataset" 117 name: "input_dataset" 166 name: "input_dataset" 222 name: "input_dataset"
|
H A D | ParallelBatchDataset.pbtxt | 4 name: "input_dataset" 39 name: "input_dataset" 81 name: "input_dataset" 130 name: "input_dataset" 186 name: "input_dataset" 249 name: "input_dataset"
|
H A D | CacheDataset.pbtxt | 4 name: "input_dataset" 32 name: "input_dataset" 59 name: "input_dataset" 93 name: "input_dataset" 134 name: "input_dataset"
|
H A D | RepeatDataset.pbtxt | 4 name: "input_dataset" 32 name: "input_dataset" 59 name: "input_dataset" 93 name: "input_dataset" 134 name: "input_dataset"
|
H A D | ConcatenateDataset.pbtxt | 4 name: "input_dataset" 32 name: "input_dataset" 59 name: "input_dataset" 93 name: "input_dataset" 134 name: "input_dataset"
|
H A D | BatchDataset.pbtxt | 4 name: "input_dataset" 32 name: "input_dataset" 59 name: "input_dataset" 93 name: "input_dataset" 134 name: "input_dataset"
|
H A D | SkipDataset.pbtxt | 4 name: "input_dataset" 32 name: "input_dataset" 59 name: "input_dataset" 93 name: "input_dataset" 134 name: "input_dataset"
|
H A D | TakeDataset.pbtxt | 4 name: "input_dataset" 32 name: "input_dataset" 59 name: "input_dataset" 93 name: "input_dataset" 134 name: "input_dataset"
|
/aosp_15_r20/external/tensorflow/tensorflow/core/api_def/base_api/ |
H A D | api_def_GroupByReducerDataset.pbtxt | 5 name: "input_dataset" 20 A function mapping an element of `input_dataset`, concatenated 48 A function mapping the current reducer state and an element of `input_dataset`, 65 summary: "Creates a dataset that computes a group-by on `input_dataset`." 67 Creates a dataset that computes a group-by on `input_dataset`.
|
H A D | api_def_ExperimentalGroupByReducerDataset.pbtxt | 5 name: "input_dataset" 20 A function mapping an element of `input_dataset`, concatenated 48 A function mapping the current reducer state and an element of `input_dataset`, 65 summary: "Creates a dataset that computes a group-by on `input_dataset`." 67 Creates a dataset that computes a group-by on `input_dataset`.
|
H A D | api_def_DatasetCardinality.pbtxt | 5 name: "input_dataset" 13 The cardinality of `input_dataset`. Named constants are used to represent 17 summary: "Returns the cardinality of `input_dataset`." 19 Returns the cardinality of `input_dataset`.
|
H A D | api_def_ExperimentalDatasetCardinality.pbtxt | 5 name: "input_dataset" 13 The cardinality of `input_dataset`. Named constants are used to represent 17 summary: "Returns the cardinality of `input_dataset`." 19 Returns the cardinality of `input_dataset`.
|