/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/ |
H A D | maxpooling_op_gpu.cu.cc | 71 const int kernel_w, const int stride_h, const int stride_w, const int pad_t, in MaxPoolForwardNCHW() 114 const int pooled_width, const int kernel_h, const int kernel_w, in MaxPoolForwardNoMaskKernel_NCHW_VECT_C() 147 const int pooled_width, const int kernel_h, const int kernel_w, in MaxPoolForwardNHWC() 234 const int width, const int kernel_h, const int kernel_w, const int stride_h, in MaxPoolGradBackwardNoMaskNCHW() 275 const int width, const int kernel_h, const int kernel_w, const int stride_h, in MaxPoolGradBackwardNoMaskNHWC() 357 const int pooled_width, const int kernel_h, const int kernel_w, in operator ()() 377 const int kernel_h, const int kernel_w, const int stride_h, in operator ()() 428 const int kernel_w, const int stride_h, const int stride_w, const int pad_t, in operator ()()
|
H A D | pooling_ops_3d_gpu.cu.cc | 35 const int kernel_h, const int kernel_w, const int stride_p, in MaxPoolGradBackwardNoMaskNCDHW() 86 const int kernel_h, const int kernel_w, const int stride_p, in MaxPoolGradBackwardNoMaskNDHWC() 142 const int kernel_w, const int stride_p, const int stride_h, in operator ()()
|
H A D | avgpooling_op_gpu.cu.cc | 86 const int kernel_w, const int stride_h, in RunAvePoolBackwardNHWC()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | Unfold3d.cpp | 190 int64_t kernel_w, in Unfold3dZeroPaddingCopyKernelImpl() 234 int64_t kernel_w, in Unfold3dCopyKernelImpl() 313 int64_t kernel_w, in Unfold3dZeroPaddingAccKernelImpl() 366 int64_t kernel_w, in Unfold3dAccKernelImpl() 447 int64_t kernel_w, in Unfold3dCopyCPU() 495 int64_t kernel_w, in Unfold3dAccCPU()
|
/aosp_15_r20/external/ComputeLibrary/src/runtime/heuristics/dwc_native/ |
H A D | ClDWCNativeDefaultConfigValhall.cpp | 82 const size_t kernel_w = wei_shape[idx_w]; in configure_G78_f32() local 147 const size_t kernel_w = wei_shape[idx_w]; in configure_G78_f16() local 248 const size_t kernel_w = wei_shape[idx_w]; in configure_G77_f16() local
|
H A D | ClDWCNativeDefaultConfigBifrost.cpp | 49 const size_t kernel_w = wei_shape[idx_w]; in configure_f32() local 122 const size_t kernel_w = wei_shape[idx_w]; in configure_f16() local
|
H A D | ClDWCNativeHeuristicsHelpers.cpp | 42 const size_t kernel_w = weights->tensor_shape()[idx_w]; in use_cl_image_for_weights() local
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_xnnpack_integration.py | 108 kernel_w, argument 177 kernel_w, argument 331 kernel_w, argument 446 kernel_w, argument 600 kernel_w, argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | AveragePool2d.cu | 36 const int pooled_width, const int kernel_h, const int kernel_w, in avg_pool2d_out_cuda_frame() 85 const int pooled_width, const int kernel_h, const int kernel_w, in avg_pool2d_out_cuda_frame_nhwc() 134 const int kernel_h, const int kernel_w, const int stride_h, in avg_pool2d_backward_out_cuda_frame() 191 const int kernel_h, const int kernel_w, const int stride_h, in avg_pool2d_backward_out_cuda_frame_nhwc()
|
H A D | im2col.cuh | 113 const int64_t kernel_w, in col2im_device() 168 const int64_t kernel_w, in col2im_kernel() 253 const int64_t kernel_w, in col2im_batched_kernel()
|
H A D | DilatedMaxPool2d.cu | 48 const int kernel_h, const int kernel_w, const int stride_h, in max_pool_forward_nchw() 180 const int pooled_width, const int kernel_h, const int kernel_w, in max_pool_backward_nchw()
|
H A D | vol2col.cuh | 138 const unsigned kernel_w, in vol2im_kernel()
|
/aosp_15_r20/external/pytorch/test/quantization/core/ |
H A D | test_quantized_functional.py | 141 groups, kernel_h, kernel_w, stride_h, stride_w, pad_h, pad_w, dilation, argument 197 out_channels_per_group, groups, kernel_d, kernel_h, kernel_w, argument
|
H A D | test_quantized_op.py | 5153 kernel_w, argument 5234 kernel_w, argument 5441 kernel_w, argument 5524 kernel_w, argument 5781 kernel_w, argument 5913 kernel_w, argument 6405 kernel_w, argument 6489 kernel_w, argument
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/tools/versioning/ |
H A D | gpu_compatibility.cc | 205 absl::Status CheckKernels(int kernel_h, int kernel_w) { in CheckKernels() 214 absl::Status CheckKernelsAndStrides(int kernel_h, int kernel_w, int strides_h, in CheckKernelsAndStrides()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | qconv_prepack.cpp | 73 const int kernel_w = weight.size(kSpatialDim + 1); in prepack() local 240 const uint32_t kernel_w = weight.size(kSpatialDim + 1); in prepack() local
|
H A D | qconv_unpack_impl.cpp | 26 const int kernel_w = kernel[kSpatialDim - 1]; in unpack() local
|
H A D | XnnpackUtils.h | 46 uint32_t kernel_w, in xnnp_create_convolution2d_nhwc()
|
H A D | qconv.cpp | 298 const int kernel_w = kernel[kSpatialDim - 1]; in apply_impl() local
|
/aosp_15_r20/external/pytorch/test/quantization/eager/ |
H A D | test_quantize_eager_qat.py | 767 kernel_w, argument 919 kernel_w, argument
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/mlir/tosa/transforms/ |
H A D | legalize_tf.cc | 533 int64_t kernel_w = tmpAttr[2].dyn_cast<IntegerAttr>().getInt(); in matchAndRewrite() local 603 int64_t kernel_w = tmpAttr[2].dyn_cast<IntegerAttr>().getInt(); in matchAndRewrite() local
|
H A D | legalize_tfl.cc | 885 int64_t kernel_w = tfl_avgpool_op.filter_width(); in matchAndRewrite() local 962 int64_t kernel_w = tfl_maxpool_op.filter_width(); in matchAndRewrite() local
|
/aosp_15_r20/external/pytorch/torch/onnx/ |
H A D | symbolic_opset11.py | 1086 def _get_im2col_output_shape(g: jit_utils.GraphContext, input, kernel_h, kernel_w): argument
|