/aosp_15_r20/external/pytorch/torch/csrc/jit/tensorexpr/operators/ |
H A D | softmax.cpp | 63 auto move_softmax_dim_index_to_pos = [&](const ParameterList& indices) { in computeSoftmax() 76 auto remove_softmax_dim_index = [&](const ParameterList& indices) { in computeSoftmax() 86 auto convert_indices_to_expr_handle = [&](const ParameterList& indices) { in computeSoftmax() 106 [&](ParameterList& indices) { in computeSoftmax() 115 [&](ParameterList& indices) { in computeSoftmax() 125 [&](ParameterList& indices) { in computeSoftmax() 131 "aten_softmax", outputShape, std::nullopt, [&](ParameterList& indices) { in computeSoftmax() 144 [&](ParameterList& indices) { return log(sum.load(indices)); }); in computeSoftmax() 149 [&](ParameterList& indices) { in computeSoftmax()
|
/aosp_15_r20/external/angle/src/tests/gl_tests/ |
H A D | LineLoopTest.cpp | 148 static const GLubyte indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 157 static const GLushort indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 171 static const GLuint indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 180 static const GLubyte indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 194 static const GLushort indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 213 static const GLuint indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 232 static const GLubyte indices[] = {5, 6, 7, 5, 7, 8}; in TEST_P() local 269 static const GLubyte indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 296 static const GLushort indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local 328 static const GLuint indices[] = {0, 7, 6, 9, 8, 0}; in TEST_P() local [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/kernel_tests/array_ops/ |
H A D | scatter_ops_test.py | 32 def _NumpyAdd(ref, indices, updates): argument 39 def _NumpyAddScalar(ref, indices, update): argument 44 def _NumpySub(ref, indices, updates): argument 49 def _NumpySubScalar(ref, indices, update): argument 54 def _NumpyMul(ref, indices, updates): argument 59 def _NumpyMulScalar(ref, indices, update): argument 64 def _NumpyDiv(ref, indices, updates): argument 69 def _NumpyDivScalar(ref, indices, update): argument 74 def _NumpyMin(ref, indices, updates): argument 79 def _NumpyMinScalar(ref, indices, update): argument [all …]
|
H A D | scatter_nd_ops_test.py | 59 def _NumpyScatterNd(ref, indices, updates, op): argument 76 def _NumpyUpdate(ref, indices, updates): argument 80 def _NumpyAdd(ref, indices, updates): argument 84 def _NumpySub(ref, indices, updates): argument 88 def _NumpyMul(ref, indices, updates): argument 92 def _NumpyDiv(ref, indices, updates): argument 96 def _NumpyMin(ref, indices, updates): argument 100 def _NumpyMax(ref, indices, updates): argument 490 def scatter_nd(self, indices, updates, shape, input_=None): argument 785 def scatter_nd(self, indices, updates, shape, input_=None): argument
|
/aosp_15_r20/external/executorch/kernels/portable/cpu/util/ |
H A D | advanced_index_util.cpp | 20 bool check_indices_dtypes(TensorOptList indices) { in check_indices_dtypes() 42 bool check_mask_indices(const Tensor& in, TensorOptList indices) { in check_mask_indices() 141 bool check_index_args(const Tensor& in, TensorOptList indices, Tensor& out) { in check_index_args() 150 size_t count_index_blocks(TensorOptList indices) { in count_index_blocks() 167 TensorOptList indices, in get_indices_broadcast_shape() 211 size_t get_indices_broadcast_ndim(TensorOptList indices) { in get_indices_broadcast_ndim() 230 size_t get_num_indexed_dims(TensorOptList indices) { in get_num_indexed_dims() 245 size_t get_num_null_indices(TensorOptList indices) { in get_num_null_indices() 255 size_t get_num_leading_null_indices(TensorOptList indices) { in get_num_leading_null_indices() 265 TensorOptList indices, in get_index_out_target_size() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/ops/ |
H A D | state_ops.py | 383 def scatter_update(ref, indices, updates, use_locking=True, name=None): argument 437 def scatter_nd_update(ref, indices, updates, use_locking=True, name=None): argument 499 def scatter_add(ref, indices, updates, use_locking=False, name=None): argument 551 def scatter_nd_add(ref, indices, updates, use_locking=False, name=None): argument 614 def scatter_sub(ref, indices, updates, use_locking=False, name=None): argument 668 def scatter_nd_sub(ref, indices, updates, use_locking=False, name=None): argument 732 def scatter_mul(ref, indices, updates, use_locking=False, name=None): argument 784 def scatter_div(ref, indices, updates, use_locking=False, name=None): argument 836 def scatter_max(ref, indices, updates, use_locking=False, name=None): argument 891 def scatter_min(ref, indices, updates, use_locking=False, name=None): argument [all …]
|
/aosp_15_r20/external/mesa3d/src/mesa/main/ |
H A D | glthread_draw.c | 64 const GLvoid **indices) in upload_indices() 84 const GLsizei *count, const GLvoid *const *indices, in upload_multi_indices() 630 const GLvoid *indices = cmd->indices; in _mesa_unmarshal_DrawElements() local 643 const GLvoid *indices = (void*)(uintptr_t)cmd->indices; in _mesa_unmarshal_DrawElementsPacked() local 656 const GLvoid *indices = cmd->indices; in _mesa_unmarshal_DrawElementsInstancedBaseVertex() local 673 const GLvoid *indices = cmd->indices; in _mesa_unmarshal_DrawElementsInstancedBaseInstance() local 690 const GLvoid *indices = cmd->indices; in _mesa_unmarshal_DrawElementsInstancedBaseVertexBaseInstanceDrawID() local 769 const GLvoid *indices, GLsizei instance_count, GLint basevertex, in draw_elements() 1034 const GLvoid *const *indices = (const GLvoid *const *)variable_data; in _mesa_unmarshal_MultiDrawElementsUserBuf() local 1059 const GLvoid *const *indices, GLsizei draw_count, in multi_draw_elements_async() [all …]
|
H A D | draw.c | 230 indices_aligned(unsigned index_size_shift, const GLvoid *indices) in indices_aligned() 284 GLenum type, const GLvoid * const *indices, in _mesa_validate_MultiDrawElements() 1583 const GLvoid * indices, in _mesa_validated_drawrangeelements() 1723 const GLvoid * indices, GLint basevertex) in _mesa_DrawRangeElementsBaseVertex() 1815 GLsizei count, GLenum type, const GLvoid * indices) in _mesa_DrawRangeElements() 1827 const GLvoid * indices) in _mesa_DrawElements() 1838 const GLvoid * indices, GLint basevertex) in _mesa_DrawElementsBaseVertex() 1863 const GLvoid * indices, GLsizei numInstances) in _mesa_DrawElementsInstanced() 1876 GLenum type, const GLvoid * indices, in _mesa_DrawElementsInstancedBaseVertex() 1892 const GLvoid *indices, in _mesa_DrawElementsInstancedBaseInstance() [all …]
|
/aosp_15_r20/external/executorch/kernels/test/ |
H A D | op_embedding_test.cpp | 29 const Tensor& indices, in op_embedding_out() 56 Tensor indices = tfl.make( in test_dtype() local 97 Tensor indices = tf_indices.make({2, 4}, {1, 2, 4, 5, 4, 3, 2, 9}); in test_dynamic_shape() local 128 Tensor indices = tfl.make({1}, {1}); in TEST_F() local 170 Tensor indices = tfl.make({1, 2, 3}, {1, 0, 2, 3, 4, 0}); in TEST_F() local 207 Tensor indices = tfl.make({2, 2}, {1, 0, 2, 3}); in TEST_F() local 241 Tensor indices = tfl.make({2, 2}, {1, 0, 2, 3}); in TEST_F() local 272 Tensor indices = tfl.make({2, 2}, {1, 0, 2, 3}); in TEST_F() local 336 Tensor indices = tfl.make({2, 2}, {2, 0, 2, 4}); in TEST_F() local 366 Tensor indices = tfl.make({}, {3}); in TEST_F() local [all …]
|
H A D | op_index_test.cpp | 33 OptTensorArrayRef indices, in op_index_tensor_out() 75 optional<Tensor> indices[] = { in test_dtype() local 118 OptTensorArrayRef indices, in run_test_cases() 157 Tensor indices = tfb.make( in TEST_F() local 206 optional<Tensor> indices[] = { in TEST_F() local 267 optional<Tensor> indices[] = { in TEST_F() local 309 optional<Tensor> indices[] = { in TEST_F() local 553 optional<Tensor> indices[] = { in TEST_F() local 573 optional<Tensor> indices[] = { in TEST_F() local 610 optional<Tensor> indices[] = { in TEST_F() local
|
H A D | op_index_put_test.cpp | 32 OptTensorArrayRef indices, in op_index_put_out() 75 optional<Tensor> indices[] = { in test_dtype() local 180 optional<Tensor> indices[] = { in test_dynamic_shape() local 204 OptTensorArrayRef indices, in run_test_cases() 252 Tensor indices = tfb.make( in TEST_F() local 333 Tensor indices = tfb.make( in TEST_F() local 506 optional<Tensor> indices[] = { in TEST_F() local 653 optional<Tensor> indices[] = { in TEST_F() local 915 optional<Tensor> indices[] = { in TEST_F() local 941 optional<Tensor> indices[] = { in TEST_F() local [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | qembeddingbag.cpp | 39 const at::Tensor& indices, in embedding_lookup_fallback_impl() 200 const IndexType* indices) { in fbgemm_spmdm_report_error_() 227 const at::Tensor& indices, in embedding_bag_nbit_impl() 399 const at::Tensor& indices, in embedding_bag_byte_impl() 560 const at::Tensor& indices, in embedding_bag_byte_helper() 658 const at::Tensor& indices, in _embedding_bag_nbit_helper() 762 const at::Tensor& indices, in embeddingbag_byte() 783 const at::Tensor& indices, in embeddingbag_4bit() 821 const Tensor& indices, in embedding_bag_byte_rowwise_offsets_out() 844 const Tensor& indices, in embedding_bag_4bit_rowwise_offsets_out() [all …]
|
/aosp_15_r20/external/deqp/modules/gles31/functional/ |
H A D | es31fNegativeVertexArrayApiTests.cpp | 550 GLbyte indices[1] = {0}; in draw_elements() local 629 GLbyte indices[1] = {0}; in draw_elements_invalid_program() local 669 GLbyte indices[1] = {0}; in draw_elements_incomplete_primitive() local 756 GLuint indices[1] = {0}; in draw_elements_base_vertex() local 793 GLuint indices[1] = {0}; in draw_elements_base_vertex_primitive_mode_mismatch() local 940 GLbyte indices[1] = {0}; in draw_elements_instanced() local 1023 GLbyte indices[1] = {0}; in draw_elements_instanced_invalid_program() local 1068 GLbyte indices[1] = {0}; in draw_elements_instanced_incomplete_primitive() local 1154 GLbyte indices[1] = {0}; in draw_elements_instanced_base_vertex() local 1204 GLuint indices[1] = {0}; in draw_elements_instanced_base_vertex_primitive_mode_mismatch() local [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | MaxUnpooling.cu | 34 const int64_t* indices, in max_unpooling2d_forward_kernel() 55 PackedTensorAccessor64<const int64_t, 4> indices, in max_unpooling3d_forward_kernel() 78 const int64_t* indices, in max_unpooling2d_backward_kernel() 100 PackedTensorAccessor64<int64_t, 4> indices, in max_unpooling3d_backward_kernel() 161 auto indices = indices_.contiguous(); in max_unpooling2d_forward_out_cuda() local 205 const Tensor& indices, in max_unpooling2d_forward_cuda() 215 const Tensor& indices, in max_unpooling3d_shape_check() 315 auto indices = indices_.contiguous(); in max_unpooling3d_forward_out_cuda() local 390 const Tensor& indices, in max_unpooling3d_forward_cuda() 435 auto indices = indices_.contiguous(); in max_unpooling2d_backward_out_cuda() local [all …]
|
/aosp_15_r20/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
H A D | TensorRef.h | 227 array<Index, 2> indices; in operator() local 235 array<Index, 3> indices; in operator() local 244 array<Index, 4> indices; in operator() local 254 array<Index, 5> indices; in operator() local 265 array<Index, 2> indices; in coeffRef() local 273 array<Index, 3> indices; in coeffRef() local 282 array<Index, 4> indices; in operator() local 292 array<Index, 5> indices; in coeffRef() local 303 EIGEN_STRONG_INLINE const Scalar coeff(const array<Index, NumIndices>& indices) const in coeff() 321 EIGEN_STRONG_INLINE Scalar& coeffRef(const array<Index, NumIndices>& indices) in coeffRef()
|
/aosp_15_r20/external/cronet/net/third_party/quiche/src/quiche/quic/core/qpack/ |
H A D | qpack_blocking_manager.cc | 22 const IndexSet& indices = it->second.front(); in OnHeaderAcknowledgement() local 46 for (const IndexSet& indices : it->second) { in OnStreamCancellation() local 64 IndexSet indices) { in OnHeaderBlockSent() 87 for (const IndexSet& indices : header_blocks_for_stream.second) { in blocking_allowed_on_stream() local 129 uint64_t QpackBlockingManager::RequiredInsertCount(const IndexSet& indices) { in RequiredInsertCount() 133 void QpackBlockingManager::IncreaseReferenceCounts(const IndexSet& indices) { in IncreaseReferenceCounts() 144 void QpackBlockingManager::DecreaseReferenceCounts(const IndexSet& indices) { in DecreaseReferenceCounts()
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/tensorexpr/ |
H A D | ir.cpp | 14 static Dtype dtypeOfIndices(const std::vector<ExprPtr>& indices) { in dtypeOfIndices() 22 static void castIndicesToInts(std::vector<ExprPtr>& indices) { in castIndicesToInts() 42 Load::Load(Dtype dtype, BufPtr buf, std::vector<ExprPtr> indices) in Load() 47 Load::Load(const BufPtr& buf, const std::vector<ExprPtr>& indices) in Load() 53 const std::vector<ExprHandle>& indices) { in make() 60 const std::vector<ExprHandle>& indices) { in make() 64 Store::Store(BufPtr buf, std::vector<ExprPtr> indices, ExprPtr value) in Store() 73 const std::vector<ExprHandle>& indices, in make() 87 const std::vector<ExprPtr>& indices, in flatten_index()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | EmbeddingBag.cpp | 67 const Tensor& indices, in promoteIndicesAndOffsets() 163 const index_t* indices) { in fbgemm_spmdm_report_error_() 872 const Tensor& indices, in check_arguments() 920 const Tensor& indices, in make_bag_size_out() 946 const Tensor& indices, in make_max_indices_out() 969 const Tensor& indices, in make_offset2bag_out() 1001 const Tensor& indices, in make_bag_size() 1012 const Tensor& indices, in make_max_indices() 1025 const Tensor& indices, in make_offset2bag() 1067 const Tensor& indices, in embedding_bag_cpu_max_out() [all …]
|
H A D | Sorting.cpp | 95 void _fill_indices(const TensorBase &indices, int64_t dim) { in _fill_indices() 416 Tensor& indices, in kthvalue_out_impl_cpu() 503 Tensor& indices, in median_with_indices_impl() 785 Tensor& indices) { in kthvalue_out_cpu() 801 Tensor& indices) { in kthvalue_out() 812 Tensor indices = at::empty({0}, self.options().dtype(kLong)); in kthvalue() local 851 Tensor& indices) { in median_out_cpu() 867 Tensor& indices) { in median_out() 877 Tensor indices = at::empty({0}, self.options().dtype(kLong)); in median() local 898 Tensor& indices) { in nanmedian_out_cpu() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/ops/ragged/ |
H A D | ragged_batch_gather_op_test.py | 145 def testRaggedBatchGather(self, descr, params, indices, expected): argument 348 self, descr, params, indices, expected, indices_ragged_rank=None, argument 410 self, descr, params, indices, expected, default_value): argument 431 self, params, indices, default_value, error): argument 458 self, descr, params, indices, default_value, error, ragged_rank=None, argument 525 indices, argument
|
/aosp_15_r20/external/executorch/kernels/quantized/test/ |
H A D | op_embedding_test.cpp | 57 Tensor indices = tf_l.make({2}, {0, 2}); in test_dtype() local 116 Tensor indices = tf_l.make({2}, {0, 2}); in TEST() local 188 Tensor indices = tf_l.make({3}, {0, 2, 1}); in TEST() local 245 Tensor indices = tf_l.make({3}, {0, 2, 1}); in TEST() local 275 Tensor indices = tf_l.make({3}, {0, 2, 1}); in TEST() local 304 Tensor indices = tf_l.make({3}, {0, 2, 1}); in TEST() local 333 Tensor indices = tf_l.make({3}, {0, 2, 1}); in TEST() local 362 Tensor indices = tf_l.make({3}, {0, 2, 1}); in TEST() local
|
/aosp_15_r20/external/armnn/src/backends/backendsCommon/test/layerTests/ |
H A D | GatherTestImpl.cpp | 89 const std::vector<int32_t> indices = std::vector<int32_t>({ 0, 2, 1, 5 }); in Gather1dParamsTestImpl() local 126 const std::vector<int32_t> indices = std::vector<int32_t>({ 2, 1 }); in Gather1dParamsAxisTestImpl() local 163 const std::vector<int32_t> indices = std::vector<int32_t>({ 1, 3, 4 }); in GatherMultiDimParamsTestImpl() local 207 const std::vector<int32_t> indices = { 1, 2, 1, 2, 1, 0 }; in GatherMultiDimParamsMultiDimIndicesTestImpl() local 269 const std::vector<int32_t> indices = { 1, 0, 1, 0, 1, 0 }; in GatherMultiDimParamsMultiDimIndicesAxis1TestImpl() local 342 const std::vector<int32_t> indices = { 1, 2, 1, 2, 1, 0 }; in GatherMultiDimParamsMultiDimIndicesAxis2TestImpl() local 394 const std::vector<int32_t> indices = std::vector<int32_t>({ 0, 2, 1, 5 }); in Gather1dParamsTestImpl() local 422 const std::vector<int32_t> indices = std::vector<int32_t>({ 1, 3, 4 }); in GatherMultiDimParamsTestImpl() local 460 const std::vector<int32_t> indices = { 1, 2, 1, 2, 1, 0 }; in GatherMultiDimParamsMultiDimIndicesTestImpl() local
|
/aosp_15_r20/external/executorch/kernels/quantized/cpu/ |
H A D | op_embedding.cpp | 33 const Tensor& indices, in check_embedding_byte_args() 133 const Tensor& indices, in embedding_byte_per_channel() 184 const Tensor& indices, in resize_out_tensor() 224 const Tensor& indices, in quantized_embedding_byte_out() 259 const Tensor& indices, in quantized_embedding_byte_out() 283 const Tensor& indices, in quantized_embedding_byte_dtype_out() 322 const Tensor& indices, in quantized_embedding_byte_dtype_out()
|
H A D | embeddingxb.cpp | 71 const Tensor& indices, in check_embedding_xbit_args() 174 const Tensor& indices, in embedding_xbit_per_channel() 225 const Tensor& indices, in resize_out_tensor() 266 const Tensor& indices, in quantized_embedding_xbit_out() 305 const Tensor& indices, in quantized_embedding_xbit_out() 331 const Tensor& indices, in quantized_embedding_xbit_dtype_out() 374 const Tensor& indices, in quantized_embedding_xbit_dtype_out()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/ |
H A D | sparse_utils_test.cc | 282 void FillIndicesWithRandomTuples(const TensorShape& shape, Tensor& indices) { in FillIndicesWithRandomTuples() 353 Tensor indices, values, shape; in TEST_P() local 369 const Tensor indices = Tensor(DT_INT64, invalid_shape); in TEST_P() local 386 const Tensor indices = in TEST_P() local 404 const Tensor indices = in TEST_P() local 425 const Tensor indices = in TEST_P() local 437 const Tensor indices = in TEST_P() local 456 Tensor indices, values, shape; in TEST_P() local 496 Tensor indices, values, shape; in TEST_P() local 558 Tensor indices, values, shape; in BM_ValidateSparseTensor() local
|