/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | ReplicationPadding.cu | 67 PackedTensorAccessor64<const scalar_t, 3> gradOutput, in replication_pad_backward_kernel() 120 PackedTensorAccessor64<const scalar_t, 4> gradOutput, in replication_pad_backward_kernel() 188 PackedTensorAccessor64<const scalar_t, 5> gradOutput, in replication_pad_backward_kernel() 230 const Tensor& gradOutput, in replication_pad2d_backward_out_cuda_template() 306 const Tensor& gradOutput, in shapeAndGradOutputCheck3d() 363 const Tensor& gradOutput, in replication_pad3d_backward_out_cuda_template() 574 Tensor& replication_pad2d_backward_out_cuda(const Tensor& gradOutput, in replication_pad2d_backward_out_cuda() 588 const Tensor& gradOutput, in replication_pad2d_backward_cuda() 671 Tensor& replication_pad3d_backward_out_cuda(const Tensor& gradOutput, in replication_pad3d_backward_out_cuda() 685 const Tensor& gradOutput, in replication_pad3d_backward_cuda()
|
H A D | AdaptiveAveragePooling.cu | 114 T *gradInput, const T *gradOutput, in adaptive_average_gradinput() 170 T *gradInput, const T *gradOutput, in atomic_adaptive_average_gradinput() 625 Tensor gradOutput = gradOutput_; in adaptive_avg_pool2d_backward_out_cuda_template() local 713 Tensor gradOutput = gradOutput_.contiguous(); in adaptive_avg_pool2d_backward_out_cuda_template() local 788 const Tensor& gradOutput, in adaptive_avg_pool2d_backward_out_cuda() 803 const Tensor& gradOutput, in adaptive_avg_pool2d_backward_cuda()
|
H A D | FractionalMaxPool3d.cu | 123 PackedTensorAccessor64<const scalar_t, 5> gradOutput, in fractional_max_pool3d_backward_out_frame() 157 const Tensor& gradOutput, in fractional_max_pool3d_backward_out_cuda_template() 329 const at::Tensor& gradOutput, in fractional_max_pool3d_backward_cuda()
|
H A D | DilatedMaxPool3d.cu | 241 const Tensor& gradOutput, in max_pool3d_with_indices_backward_out_frame() 413 const Tensor& gradOutput, in max_pool3d_with_indices_backward_out_cuda_template() 599 Tensor& max_pool3d_with_indices_backward_out_cuda(const Tensor& gradOutput, in max_pool3d_with_indices_backward_out_cuda() 626 const Tensor& gradOutput, in max_pool3d_with_indices_backward_cuda()
|
H A D | AdaptiveAveragePooling3d.cu | 154 scalar_t *gradInput, const scalar_t *gradOutput, in adaptiveaveragegradinput() 252 scalar_t *gradInput, const scalar_t *gradOutput, in atomicadaptiveaveragegradinput() 436 const Tensor gradOutput = gradOutput_.contiguous(); in adaptive_avg_pool3d_backward_out_cuda_template() local
|
H A D | AveragePool3d.cu | 172 PackedTensorAccessor64<const scalar_t, 4> gradOutput, in avg_pool3d_single_backward_out_frame_stride1() 217 PackedTensorAccessor64<const scalar_t, 4> gradOutput, in avg_pool3d_cuda_update_grad_input_atomic() 276 PackedTensorAccessor64<const scalar_t, 4> gradOutput, in avg_pool3d_cuda_update_grad_input()
|
H A D | SoftMax.cu | 314 scalar_t *gradInput, const outscalar_t *output, const outscalar_t *gradOutput, in cunn_SpatialSoftMaxBackward() 559 const outscalar_t *gradOutput, in WriteBpropResultsVectorized() 632 const outscalar_t *gradOutput, in WriteBpropResults() 775 cunn_SoftMaxBackward(scalar_t *gradInput, const outscalar_t *output, const outscalar_t *gradOutput,… in cunn_SoftMaxBackward()
|
H A D | EmbeddingBackwardKernel.cu | 82 const index_t *indices, const scalar_t *gradOutput, in compute_grad_weight_bags() 126 const scalar_t *gradOutput, in compute_grad_weight()
|
H A D | AdaptiveMaxPooling2d.cu | 112 __global__ void adaptivemaxgradinput(T *gradInput, const T *gradOutput, const int64_t *indices, in adaptivemaxgradinput() 160 T *gradInput, const T *gradOutput, const int64_t *indices, in atomicadaptivemaxgradinput()
|
H A D | AdaptiveMaxPooling3d.cu | 161 T *gradInput, const T *gradOutput, const int64_t *indices, in adaptivemaxgradinput() 237 T *gradInput, const T *gradOutput, const int64_t *indices, in atomicadaptivemaxgradinput()
|
H A D | FractionalMaxPool2d.cu | 106 PackedTensorAccessor<const scalar_t, 4> gradOutput, in fractional_max_pool2d_backward_out_cuda_frame()
|
H A D | MultiMarginLoss.cu | 70 scalar_t *gradInput, const scalar_t *gradOutput, const scalar_t *input, const int64_t *target, in MultiMarginLoss_backward_kernel()
|
H A D | MaxUnpooling.cu | 214 const Tensor& gradOutput, in max_unpooling3d_shape_check()
|
H A D | AveragePool2d.cu | 387 const Tensor gradOutput = gradOutput_.contiguous(memory_format); in TORCH_IMPL_FUNC() local
|
H A D | EmbeddingBag.cu | 241 const index_t *max_indices, const scalar_t *gradOutput, in EmbeddingBag_accGradParametersKernel_max()
|
H A D | DilatedMaxPool2d.cu | 467 const Tensor gradOutput = gradOutput_.contiguous(memory_format); in TORCH_IMPL_FUNC() local
|
H A D | RNN.cu | 325 TensorInfo<scalar_t, index_type> gradOutput, in gru_cell_backward()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | ReplicationPadding.cpp | 182 const Tensor& gradOutput, in replication_pad2d_backward_out_cpu_template() 221 const Tensor& gradOutput, in replication_pad3d_backward_out_cpu_template() 297 Tensor& replication_pad2d_backward_out_cpu(const Tensor& gradOutput, in replication_pad2d_backward_out_cpu() 310 const Tensor& gradOutput, in replication_pad2d_backward_cpu() 329 Tensor& replication_pad3d_backward_out_cpu(const Tensor& gradOutput, in replication_pad3d_backward_out_cpu() 342 const Tensor& gradOutput, in replication_pad3d_backward_cpu()
|
H A D | Pool.h | 172 const Tensor& gradOutput, in max_pool2d_backward_shape_check() 200 const Tensor& gradOutput, in avg_pool2d_backward_shape_check() 288 const Tensor& gradOutput, in max_pool3d_backward_shape_check() 325 const Tensor& gradOutput, in avg_pool3d_backward_shape_check()
|
H A D | FractionalMaxPool2d.cpp | 110 auto gradOutput = gradOutput_.contiguous(); in TORCH_META_FUNC() local 225 const scalar_t* gradOutput, in fractional_max_pool2d_backward_out_single_batch_frame() 252 const scalar_t* gradOutput, in fractional_max_pool2d_backward_out_frame() 373 auto gradOutput = gradOutput_.contiguous(); in TORCH_IMPL_FUNC() local
|
H A D | FractionalMaxPool3d.cpp | 264 const scalar_t* gradOutput, in fractional_max_pool3d_backward_out_single_batch_frame() 296 const scalar_t* gradOutput, in fractional_max_pool3d_backward_out_frame() 368 auto gradOutput = gradOutput_.contiguous(); in fractional_max_pool3d_backward_out_cpu_template() local
|
H A D | MaxUnpooling.cpp | 80 const Tensor& gradOutput, in max_unpooling3d_shape_check()
|
H A D | DilatedMaxPool3d.cpp | 114 const Tensor& gradOutput, in max_pool3d_with_indices_backward_out_cpu_template()
|
H A D | AdaptiveAveragePooling3d.cpp | 235 auto gradOutput = gradOutput_.contiguous(); in adaptive_avg_pool3d_backward_out_cpu_template() local
|
H A D | AveragePool3d.cpp | 457 Tensor gradOutput = gradOutput_.contiguous(); in TORCH_IMPL_FUNC() local
|