/aosp_15_r20/external/pytorch/torch/csrc/jit/tensorexpr/operators/ |
H A D | softmax.cpp | 64 std::vector<ExprHandle> new_indices; in computeSoftmax() local 77 std::vector<ExprHandle> new_indices; in computeSoftmax() local 87 std::vector<ExprHandle> new_indices(indices.size()); in computeSoftmax() local
|
H A D | misc.cpp | 338 std::vector<ExprHandle> new_indices; in computeChunk() local
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/ |
H A D | canonicalize.cpp | 197 auto new_indices = sort_indexes(n->outputs()); in CanonicalizeLoopOutputs() local 202 auto new_indices = sort_indexes(n->outputs()); in CanonicalizeIfOutputs() local
|
/aosp_15_r20/external/mesa3d/src/microsoft/vulkan/ |
H A D | dzn_nir.c | 436 nir_def *new_indices = in dzn_nir_triangle_fan_prim_restart_rewrite_index_shader() local 486 nir_def *new_indices; in dzn_nir_triangle_fan_rewrite_index_shader() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | SparseTensorUtils.cpp | 74 Tensor new_indices = at::zeros({indices.size(1)}, indices.options()); in flatten_indices_by_dims() local
|
H A D | TensorShape.cpp | 539 Tensor new_indices = indices.new_empty(new_indices_size); in sparse_broadcast_to() local 1269 Tensor new_indices; in narrow_copy_sparse() local 1482 const auto new_indices = (new_sparse_dims == old_sparse_dims) in permute_sparse_coo() local 1765 auto new_indices = indices.index_select(1, nzIndices).index_select(0, dimIndices); in select_sparse() local 3394 auto new_indices = at::cat( in unsqueeze_sparse() local
|
H A D | TensorConversions.cpp | 266 const auto new_indices = at::native::to( in _to_copy() local
|
/aosp_15_r20/external/ComputeLibrary/src/core/CPP/kernels/ |
H A D | CPPBoxWithNonMaximaSuppressionLimitKernel.cpp | 148 std::vector<int> new_indices; in NonMaximaSuppression() local
|
/aosp_15_r20/external/ComputeLibrary/src/runtime/CPP/functions/ |
H A D | CPPDetectionOutputLayer.cpp | 543 std::map<int, std::vector<int>> new_indices; in run() local
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/service/spmd/ |
H A D | spmd_partitioner.cc | 2999 std::vector<HloInstruction*> new_indices(hlo->shape().rank()); in HandleDynamicSlice() local 3063 std::vector<HloInstruction*> new_indices(hlo->shape().rank()); in HandleDynamicUpdateSlice() local 3089 std::vector<HloInstruction*> new_indices(hlo->shape().rank()); in HandleDynamicUpdateSlice() local 3206 std::vector<HloInstruction*> new_indices(hlo->shape().rank()); in HandleDynamicUpdateSlice() local
|
H A D | gather_scatter_handler.cc | 420 PartitionedHlo new_indices(adjusted_indices, adjusted_indices->shape(), in PartitionGatherTrivialIndexedOperandDimension() local
|
/aosp_15_r20/external/mesa3d/src/amd/vulkan/nir/ |
H A D | radv_nir_rt_common.c | 28 …nir_def *new_indices[4] = {nir_undef(b, 1, 32), nir_undef(b, 1, 32), nir_undef(b, 1, 32), nir_unde… in nir_sort_hit_pair() local
|
/aosp_15_r20/external/tensorflow/tensorflow/python/ops/parallel_for/ |
H A D | pfor.py | 613 def _process_body(self, pfor_input, inputs_stacked, new_indices, cond_stacked, argument 4845 def _process_body(self, inputs_stacked, new_indices, cond_stacked, argument
|
/aosp_15_r20/external/tensorflow/tensorflow/dtensor/mlir/expansions/ |
H A D | meta_spmd_expander.cc | 1071 mlir::Value new_indices = builder.create<mlir::TF::SubOp>( in ExpandOp() local
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/tensorexpr/ |
H A D | loopnest.cpp | 2571 std::vector<ExprPtr> new_indices(indices); in compressBuffer() local 2729 std::vector<ExprPtr> new_indices(v->indices().size()); in mutate() local
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/ |
H A D | literal.cc | 1032 DimensionVector new_indices(result_shape.rank()); in SliceInternal() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/ |
H A D | SparseTensorMath.cpp | 1685 Tensor new_indices; in _sparse_sum() local
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | algebraic_simplifier.cc | 4819 std::vector<HloInstruction*> new_indices; in HandleDynamicSlice() local
|