/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/ |
H A D | bidirectional_sequence_rnn_test.cc | 879 float* batch_end = batch_start + input_sequence_size; in TEST_P() local 930 float* batch_end = batch_start + rnn.input_size(); in TEST_P() local 974 float* batch_end = batch_start + input_sequence_size; in TEST_P() local 1018 float* batch_end = batch_start + rnn.input_size(); in TEST() local 1063 float* batch_end = batch_start + rnn.input_size(); in TEST() local 1132 float* batch_end = batch_start + input_sequence_size; in TEST() local 1189 float* batch_end = batch_start + rnn.input_size(); in TEST() local 1241 float* batch_end = batch_start + rnn.input_size(); in TEST() local 1289 float* batch_end = batch_start + rnn.input_size(); in TEST() local 1333 float* batch_end = batch_start + input_sequence_size; in TEST() local [all …]
|
H A D | unidirectional_sequence_rnn_test.cc | 288 float* batch_end = batch_start + input_sequence_size; in TEST() local 317 float* batch_end = batch_start + input_sequence_size; in TEST_P() local 344 float* batch_end = batch_start + input_sequence_size; in TEST_P() local 370 float* batch_end = batch_start + rnn.input_size(); in TEST() local 400 float* batch_end = batch_start + rnn.input_size(); in TEST_P() local 431 float* batch_end = batch_start + rnn.input_size(); in TEST_P() local
|
H A D | basic_rnn_test.cc | 270 float* batch_end = batch_start + rnn.input_size(); in TEST() local 299 float* batch_end = batch_start + rnn.input_size(); in TEST_P() local 327 float* batch_end = batch_start + rnn.input_size(); in TEST_P() local
|
H A D | unidirectional_sequence_lstm_test.cc | 151 const float* batch_end = batch_start + num_inputs; in VerifyGoldens() local 160 const float* batch_end = batch_start + input_sequence_size * num_inputs; in VerifyGoldens() local 2273 const float* batch_end = batch_start + num_inputs; in VerifyGoldens() local
|
/aosp_15_r20/external/igt-gpu-tools/lib/ |
H A D | media_fill.c | 140 uint32_t batch_end; in gen7_media_fillfunc() local 184 uint32_t batch_end; in gen8_media_fillfunc() local 229 uint32_t batch_end; in __gen9_media_fillfunc() local 298 uint32_t batch_end; in __gen11_media_vme_func() local
|
H A D | gpgpu_fill.c | 127 uint32_t batch_end; in gen7_gpgpu_fillfunc() local 176 uint32_t batch_end; in gen8_gpgpu_fillfunc() local 225 uint32_t batch_end; in __gen9_gpgpu_fillfunc() local
|
H A D | media_spin.c | 87 uint32_t batch_end; in gen8_media_spinfunc() local 127 uint32_t batch_end; in gen9_media_spinfunc() local
|
H A D | rendercopy_gen6.c | 63 drm_intel_context *context, uint32_t batch_end) in gen6_render_flush() 532 uint32_t batch_end; in gen6_render_copyfunc() local
|
H A D | rendercopy_gen7.c | 37 drm_intel_context *context, uint32_t batch_end) in gen7_render_flush() 514 uint32_t batch_end; in gen7_render_copyfunc() local
|
H A D | rendercopy_gen4.c | 125 drm_intel_context *context, uint32_t batch_end) in gen4_render_flush() 659 uint32_t offset, batch_end; in gen4_render_copyfunc() local
|
H A D | gpu_cmds.c | 28 gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end) in gen7_render_flush() 40 gen7_render_context_flush(struct intel_batchbuffer *batch, uint32_t batch_end) in gen7_render_context_flush()
|
H A D | rendercopy_gen8.c | 133 drm_intel_context *context, uint32_t batch_end) in gen6_render_flush() 911 uint32_t batch_end; in gen8_render_copyfunc() local
|
H A D | rendercopy_gen9.c | 164 drm_intel_context *context, uint32_t batch_end) in gen6_render_flush() 969 uint32_t batch_end; in _gen9_render_copyfunc() local
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/engine/ |
H A D | training_utils_v1.py | 99 def aggregate(self, batch_outs, batch_start=None, batch_end=None): argument 137 def aggregate(self, batch_outs, batch_start=None, batch_end=None): argument 280 def aggregate(self, batch_element, batch_start=None, batch_end=None): argument 371 def aggregate(self, batch_element, batch_start, batch_end): argument 399 def _slice_assign(self, batch_element, batch_start, batch_end, is_finished): argument 455 def aggregate(self, batch_outs, batch_start=None, batch_end=None): argument
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/sparse/ |
H A D | sparse_cholesky_op.cc | 114 [&](int64_t batch_begin, int64_t batch_end) { in Compute() 193 [&](int64_t batch_begin, int64_t batch_end) { in Compute()
|
H A D | sparse_mat_mul_op.cc | 186 matmul_cost_per_batch, [&](int64_t batch_begin, int64_t batch_end) { in Compute() 228 [&](int64_t batch_begin, int64_t batch_end) { in Compute()
|
/aosp_15_r20/packages/modules/NeuralNetworks/common/cpu_operations/ |
D | SVDFTest.cpp | 347 float* batch_end = batch_start + svdf_input_size * svdf_num_batches; in TEST() local 406 float* batch_end = batch_start + svdf_input_size * svdf_num_batches; in TEST() local
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/delegates/gpu/cl/kernels/ |
H A D | lstm_full_test.cc | 243 const float* batch_end = batch_start + num_inputs; in VerifyGoldens() local 254 const float* batch_end = batch_start + num_outputs; in VerifyGoldens() local
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_0/ |
D | svdf.mod.py | 133 batch_end = batch_start + input_size * batches variable
|
D | svdf_bias_present.mod.py | 133 batch_end = batch_start + input_size * batches variable
|
D | svdf2.mod.py | 148 batch_end = batch_start + input_size * batches variable
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_2/ |
D | svdf_bias_present_float16.mod.py | 133 batch_end = batch_start + input_size * batches variable
|
D | svdf_float16.mod.py | 133 batch_end = batch_start + input_size * batches variable
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/V1_1/ |
D | svdf_relaxed.mod.py | 134 batch_end = batch_start + input_size * batches variable
|
D | svdf_bias_present_relaxed.mod.py | 134 batch_end = batch_start + input_size * batches variable
|