/aosp_15_r20/external/pytorch/test/ |
H A D | allowlist_for_publicAPI.json | 10 "torch.nn.intrinsic.qat.modules.linear_relu": "torch.ao.nn.intrinsic.qat.modules.linear_relu", 15 …"torch.nn.intrinsic.quantized.modules.linear_relu": "torch.ao.nn.intrinsic.quantized.modules.linea… 18 …nn.intrinsic.quantized.dynamic.modules.linear_relu": "torch.ao.nn.intrinsic.quantized.dynamic.modu…
|
/aosp_15_r20/external/pytorch/test/quantization/ao_migration/ |
H A D | test_ao_migration.py | 315 "linear_relu", function_list, base="nn.intrinsic.qat.modules" 353 "linear_relu", function_list, base="nn.intrinsic.quantized.modules"
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/quantization/ |
H A D | quantization_patterns.h | 501 std::string linear_relu = R"( in quant_fusion_pattern_and_replacements() local 529 %r = quantized::linear_relu(%a_quant, %packed_params, %r_scale, %r_zero_point) in quant_fusion_pattern_and_replacements() 963 {"quantized::linear_relu", std::move(linear_relu), quantized_linear_relu}, in quant_fusion_pattern_and_replacements() 964 {"quantized::linear_relu", in quant_fusion_pattern_and_replacements()
|
/aosp_15_r20/external/pytorch/docs/source/ |
H A D | quantization.rst | 1258 .. py:module:: torch.ao.nn.intrinsic.qat.modules.linear_relu 1259 .. py:module:: torch.ao.nn.intrinsic.quantized.dynamic.modules.linear_relu 1263 .. py:module:: torch.ao.nn.intrinsic.quantized.modules.linear_relu 1363 .. py:module:: torch.nn.intrinsic.qat.modules.linear_relu 1364 .. py:module:: torch.nn.intrinsic.quantized.dynamic.modules.linear_relu 1367 .. py:module:: torch.nn.intrinsic.quantized.modules.linear_relu
|
H A D | conf.py | 2240 # torch.ao.nn.intrinsic.qat.modules.linear_relu 2242 # torch.ao.nn.intrinsic.quantized.dynamic.modules.linear_relu 2254 # torch.ao.nn.intrinsic.quantized.modules.linear_relu
|
/aosp_15_r20/external/pytorch/torch/_export/passes/ |
H A D | replace_quantized_ops_with_standard_ops_pass.py | 263 if opname in ["conv1d_relu", "conv2d_relu", "linear_relu", "add_relu", "mul_relu"]: 480 "linear_relu.default": _transform_linear_with_packedparam,
|
/aosp_15_r20/external/pytorch/torch/ao/nn/intrinsic/quantized/dynamic/modules/ |
H A D | __init__.py | 1 from .linear_relu import LinearReLU
|
/aosp_15_r20/external/pytorch/torch/nn/intrinsic/quantized/dynamic/modules/ |
H A D | __init__.py | 1 from torch.nn.intrinsic.quantized.dynamic.modules.linear_relu import LinearReLU
|
/aosp_15_r20/external/pytorch/torch/ao/nn/intrinsic/quantized/modules/ |
H A D | __init__.py | 4 from .linear_relu import LinearLeakyReLU, LinearReLU, LinearTanh
|
H A D | linear_relu.py | 39 return torch.ops.quantized.linear_relu(
|
/aosp_15_r20/external/pytorch/torch/nn/intrinsic/quantized/modules/ |
H A D | __init__.py | 7 from torch.nn.intrinsic.quantized.modules.linear_relu import LinearReLU
|
/aosp_15_r20/external/pytorch/torch/ao/nn/intrinsic/qat/modules/ |
H A D | __init__.py | 15 from .linear_relu import LinearReLU
|
/aosp_15_r20/external/pytorch/torch/nn/intrinsic/qat/modules/ |
H A D | __init__.py | 15 from torch.nn.intrinsic.qat.modules.linear_relu import LinearReLU
|
/aosp_15_r20/external/pytorch/test/quantization/fx/ |
H A D | test_equalize_fx.py | 689 ns.call_function(torch.ops.quantized.linear_relu), 696 ns.call_function(torch.ops.quantized.linear_relu),
|
/aosp_15_r20/external/pytorch/test/mobile/model_test/ |
H A D | coverage.yaml | 694 - quantized::linear_relu 1043 quantized::linear_relu: 2
|
H A D | model_ops.yaml | 438 quantized::linear_relu: 7
|
/aosp_15_r20/external/pytorch/torch/ao/ns/fx/ |
H A D | weight_utils.py | 214 toq.linear_relu: get_qlinear_fun_weight,
|
H A D | mappings.py | 514 toq.linear_relu,
|
/aosp_15_r20/external/executorch/docs/source/ |
H A D | compiler-backend-dialect.md | 17 …zes a certain operator that is equivalent to a subgraph. For example, `linear_relu` (equivalent to…
|
/aosp_15_r20/external/pytorch/torch/ao/quantization/quantizer/ |
H A D | xnnpack_quantizer.py | 250 "linear_relu",
|
H A D | xnnpack_quantizer_utils.py | 223 @register_annotator("linear_relu")
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | qlinear.cpp | 462 const std::string func_name = kReluFused ? "quantized::linear_relu (xnnpack)" in apply_impl_xnnp() 1324 m.impl(TORCH_SELECTIVE_NAME("quantized::linear_relu"), TORCH_FN(QLinearInt8<true>::run)); in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cudnn/ |
H A D | Linear.cpp | 362 m.impl(TORCH_SELECTIVE_NAME("quantized::linear_relu"), QLinearInt8<true>::run); in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/pytorch/test/quantization/core/ |
H A D | test_quantized_module.py | 86 """test API functionality for nn.intrinsic.quantized.linear_relu""" 95 nniq.LinearReLU, 'QuantizedLinearReLU', torch.ops.quantized.linear_relu,
|
/aosp_15_r20/external/pytorch/torch/csrc/jit/passes/onnx/ |
H A D | unpack_quantized_weights.cpp | 600 %r = quantized::linear_relu(%input, %packed_weight, %w_scale, %w_zero_point) in UnpackQuantizedWeights()
|