Home
last modified time | relevance | path

Searched full:quantizedcpu (Results 1 – 25 of 51) sorted by relevance

123

/aosp_15_r20/external/pytorch/c10/core/
H A DBackend.h54 QuantizedCPU, enumerator
115 } else if (t == DispatchKey::QuantizedCPU) { in dispatchKeyToBackend()
116 return Backend::QuantizedCPU; in dispatchKeyToBackend()
199 case Backend::QuantizedCPU: in backendToDispatchKey()
200 return DispatchKey::QuantizedCPU; in backendToDispatchKey()
226 case Backend::QuantizedCPU: in backendToDeviceType()
340 case Backend::QuantizedCPU: in toString()
341 return "QuantizedCPU"; in toString()
H A DDispatchKey.cpp337 {"QuantizedCPU", c10::DispatchKey::QuantizedCPU}, in parseDispatchKey()
H A DDispatchKeySet.h83 // - "Quantized": QuantizedCPU, QuantizedCUDA, QuantizedXLA, ...
461 case DispatchKey::QuantizedCPU: in getDispatchTableIndexForDispatchKeySet()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/
H A Dnative_functions.yaml937 QuantizedCPU, QuantizedCUDA: as_strided_qtensorimpl
1075 QuantizedCPU: quantized_batch_norm
1378 QuantizedCPU: cat_quantized_cpu
1390 QuantizedCPU: cat_out_quantized_cpu
1488 QuantizedCPU: clamp_quantized_cpu
2375 QuantizedCPU, QuantizedCUDA, QuantizedMeta: empty_unknown_quantized
2425 QuantizedCPU, QuantizedCUDA: empty_affine_quantized
2434 QuantizedCPU, QuantizedCUDA: empty_per_channel_affine_quantized
2448 QuantizedCPU: quantized_resize_cpu_
2466 QuantizedCPU, QuantizedCUDA: empty_quantized
[all …]
H A DREADME.md543 Typically it only supports a few in-tree backends like CPU, CUDA, QuantizedCPU etc but not
550 QuantizedCPU: kernel_quantized_cpu
553 You're done. Now this op will be called in `CPU/CUDA/QuantizedCPU` backend inference!
H A DTopKImpl.h13 // Core topk loop, shared between CPU and QuantizedCPU
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/
H A DREADME.md79 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) {
94 QuantizedCPU: quantized_xand
108 QuantizedCPU: quantized_xand
130 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) {
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/
H A Dqconv_prepack.cpp494 at::Tensor weight, // from CPU backend instead of QuantizedCPU in _qconv_prepack_onednn()
809 at::Tensor weight, // from CPU backend instead of QuantizedCPU in run_conv()
828 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
841 TORCH_LIBRARY_IMPL(_quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A DBinaryOps.cpp474 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
498 TORCH_LIBRARY_IMPL(_quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqdropout.cpp17 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqthreshold.cpp44 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqelu.cpp32 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqhardswish.cpp102 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqlinear_prepack.cpp675 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
687 TORCH_LIBRARY_IMPL(_quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqsoftmax.cpp144 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqsigmoid.cpp147 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqnormalization.cpp133 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqclamp.cpp169 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqmatmul.cpp181 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
H A Dqrelu.cpp228 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
/aosp_15_r20/external/pytorch/tools/test/
H A Dtest_codegen.py329 "dispatch": {"CPU": "kernel_2", "QuantizedCPU": "custom::kernel_3"},
337 DispatchKey.QuantizedCPU: {},
/aosp_15_r20/external/pytorch/cmake/
H A DCodegen.cmake100 list(APPEND CUSTOM_BUILD_FLAGS --backend_whitelist CPU QuantizedCPU Vulkan)
102 list(APPEND CUSTOM_BUILD_FLAGS --backend_whitelist CPU QuantizedCPU)
/aosp_15_r20/external/pytorch/test/quantization/eager/
H A Dtest_fuse_eager.py92 …ex(RuntimeError, "Could not run 'aten::native_batch_norm' with arguments from the 'QuantizedCPU'"):
101 …ex(RuntimeError, "Could not run 'aten::native_batch_norm' with arguments from the 'QuantizedCPU'"):
/aosp_15_r20/external/pytorch/torch/csrc/utils/
H A Dtensor_types.cpp44 case at::Backend::QuantizedCPU: in backend_to_string()
/aosp_15_r20/external/pytorch/aten/src/ATen/core/op_registration/
H A Dop_allowlist.h192 …// return k == DispatchKey::CPU || k == DispatchKey::Vulkan || k == DispatchKey::QuantizedCPU || k… in dispatch_key_allowlist_check()

123