Home
last modified time | relevance | path

Searched full:linear_relu (Results 1 – 25 of 34) sorted by relevance

12

/aosp_15_r20/external/pytorch/test/
H A Dallowlist_for_publicAPI.json10 "torch.nn.intrinsic.qat.modules.linear_relu": "torch.ao.nn.intrinsic.qat.modules.linear_relu",
15 …"torch.nn.intrinsic.quantized.modules.linear_relu": "torch.ao.nn.intrinsic.quantized.modules.linea…
18 …nn.intrinsic.quantized.dynamic.modules.linear_relu": "torch.ao.nn.intrinsic.quantized.dynamic.modu…
/aosp_15_r20/external/pytorch/test/quantization/ao_migration/
H A Dtest_ao_migration.py315 "linear_relu", function_list, base="nn.intrinsic.qat.modules"
353 "linear_relu", function_list, base="nn.intrinsic.quantized.modules"
/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/quantization/
H A Dquantization_patterns.h501 std::string linear_relu = R"( in quant_fusion_pattern_and_replacements() local
529 %r = quantized::linear_relu(%a_quant, %packed_params, %r_scale, %r_zero_point) in quant_fusion_pattern_and_replacements()
963 {"quantized::linear_relu", std::move(linear_relu), quantized_linear_relu}, in quant_fusion_pattern_and_replacements()
964 {"quantized::linear_relu", in quant_fusion_pattern_and_replacements()
/aosp_15_r20/external/pytorch/docs/source/
H A Dquantization.rst1258 .. py:module:: torch.ao.nn.intrinsic.qat.modules.linear_relu
1259 .. py:module:: torch.ao.nn.intrinsic.quantized.dynamic.modules.linear_relu
1263 .. py:module:: torch.ao.nn.intrinsic.quantized.modules.linear_relu
1363 .. py:module:: torch.nn.intrinsic.qat.modules.linear_relu
1364 .. py:module:: torch.nn.intrinsic.quantized.dynamic.modules.linear_relu
1367 .. py:module:: torch.nn.intrinsic.quantized.modules.linear_relu
H A Dconf.py2240 # torch.ao.nn.intrinsic.qat.modules.linear_relu
2242 # torch.ao.nn.intrinsic.quantized.dynamic.modules.linear_relu
2254 # torch.ao.nn.intrinsic.quantized.modules.linear_relu
/aosp_15_r20/external/pytorch/torch/_export/passes/
H A Dreplace_quantized_ops_with_standard_ops_pass.py263 if opname in ["conv1d_relu", "conv2d_relu", "linear_relu", "add_relu", "mul_relu"]:
480 "linear_relu.default": _transform_linear_with_packedparam,
/aosp_15_r20/external/pytorch/torch/ao/nn/intrinsic/quantized/dynamic/modules/
H A D__init__.py1 from .linear_relu import LinearReLU
/aosp_15_r20/external/pytorch/torch/nn/intrinsic/quantized/dynamic/modules/
H A D__init__.py1 from torch.nn.intrinsic.quantized.dynamic.modules.linear_relu import LinearReLU
/aosp_15_r20/external/pytorch/torch/ao/nn/intrinsic/quantized/modules/
H A D__init__.py4 from .linear_relu import LinearLeakyReLU, LinearReLU, LinearTanh
H A Dlinear_relu.py39 return torch.ops.quantized.linear_relu(
/aosp_15_r20/external/pytorch/torch/nn/intrinsic/quantized/modules/
H A D__init__.py7 from torch.nn.intrinsic.quantized.modules.linear_relu import LinearReLU
/aosp_15_r20/external/pytorch/torch/ao/nn/intrinsic/qat/modules/
H A D__init__.py15 from .linear_relu import LinearReLU
/aosp_15_r20/external/pytorch/torch/nn/intrinsic/qat/modules/
H A D__init__.py15 from torch.nn.intrinsic.qat.modules.linear_relu import LinearReLU
/aosp_15_r20/external/pytorch/test/quantization/fx/
H A Dtest_equalize_fx.py689 ns.call_function(torch.ops.quantized.linear_relu),
696 ns.call_function(torch.ops.quantized.linear_relu),
/aosp_15_r20/external/pytorch/test/mobile/model_test/
H A Dcoverage.yaml694 - quantized::linear_relu
1043 quantized::linear_relu: 2
H A Dmodel_ops.yaml438 quantized::linear_relu: 7
/aosp_15_r20/external/pytorch/torch/ao/ns/fx/
H A Dweight_utils.py214 toq.linear_relu: get_qlinear_fun_weight,
H A Dmappings.py514 toq.linear_relu,
/aosp_15_r20/external/executorch/docs/source/
H A Dcompiler-backend-dialect.md17 …zes a certain operator that is equivalent to a subgraph. For example, `linear_relu` (equivalent to…
/aosp_15_r20/external/pytorch/torch/ao/quantization/quantizer/
H A Dxnnpack_quantizer.py250 "linear_relu",
H A Dxnnpack_quantizer_utils.py223 @register_annotator("linear_relu")
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/
H A Dqlinear.cpp462 const std::string func_name = kReluFused ? "quantized::linear_relu (xnnpack)" in apply_impl_xnnp()
1324 m.impl(TORCH_SELECTIVE_NAME("quantized::linear_relu"), TORCH_FN(QLinearInt8<true>::run)); in TORCH_LIBRARY_IMPL()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cudnn/
H A DLinear.cpp362 m.impl(TORCH_SELECTIVE_NAME("quantized::linear_relu"), QLinearInt8<true>::run); in TORCH_LIBRARY_IMPL()
/aosp_15_r20/external/pytorch/test/quantization/core/
H A Dtest_quantized_module.py86 """test API functionality for nn.intrinsic.quantized.linear_relu"""
95 nniq.LinearReLU, 'QuantizedLinearReLU', torch.ops.quantized.linear_relu,
/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/onnx/
H A Dunpack_quantized_weights.cpp600 %r = quantized::linear_relu(%input, %packed_weight, %w_scale, %w_zero_point) in UnpackQuantizedWeights()

12