/aosp_15_r20/external/pytorch/c10/core/ |
H A D | Backend.h | 54 QuantizedCPU, enumerator 115 } else if (t == DispatchKey::QuantizedCPU) { in dispatchKeyToBackend() 116 return Backend::QuantizedCPU; in dispatchKeyToBackend() 199 case Backend::QuantizedCPU: in backendToDispatchKey() 200 return DispatchKey::QuantizedCPU; in backendToDispatchKey() 226 case Backend::QuantizedCPU: in backendToDeviceType() 340 case Backend::QuantizedCPU: in toString() 341 return "QuantizedCPU"; in toString()
|
H A D | DispatchKey.cpp | 337 {"QuantizedCPU", c10::DispatchKey::QuantizedCPU}, in parseDispatchKey()
|
H A D | DispatchKeySet.h | 83 // - "Quantized": QuantizedCPU, QuantizedCUDA, QuantizedXLA, ... 461 case DispatchKey::QuantizedCPU: in getDispatchTableIndexForDispatchKeySet()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | native_functions.yaml | 937 QuantizedCPU, QuantizedCUDA: as_strided_qtensorimpl 1075 QuantizedCPU: quantized_batch_norm 1378 QuantizedCPU: cat_quantized_cpu 1390 QuantizedCPU: cat_out_quantized_cpu 1488 QuantizedCPU: clamp_quantized_cpu 2375 QuantizedCPU, QuantizedCUDA, QuantizedMeta: empty_unknown_quantized 2425 QuantizedCPU, QuantizedCUDA: empty_affine_quantized 2434 QuantizedCPU, QuantizedCUDA: empty_per_channel_affine_quantized 2448 QuantizedCPU: quantized_resize_cpu_ 2466 QuantizedCPU, QuantizedCUDA: empty_quantized [all …]
|
H A D | README.md | 543 Typically it only supports a few in-tree backends like CPU, CUDA, QuantizedCPU etc but not 550 QuantizedCPU: kernel_quantized_cpu 553 You're done. Now this op will be called in `CPU/CUDA/QuantizedCPU` backend inference!
|
H A D | TopKImpl.h | 13 // Core topk loop, shared between CPU and QuantizedCPU
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/ |
H A D | README.md | 79 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { 94 QuantizedCPU: quantized_xand 108 QuantizedCPU: quantized_xand 130 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) {
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | qconv_prepack.cpp | 494 at::Tensor weight, // from CPU backend instead of QuantizedCPU in _qconv_prepack_onednn() 809 at::Tensor weight, // from CPU backend instead of QuantizedCPU in run_conv() 828 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument 841 TORCH_LIBRARY_IMPL(_quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | BinaryOps.cpp | 474 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument 498 TORCH_LIBRARY_IMPL(_quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qdropout.cpp | 17 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qthreshold.cpp | 44 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qelu.cpp | 32 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qhardswish.cpp | 102 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qlinear_prepack.cpp | 675 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument 687 TORCH_LIBRARY_IMPL(_quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qsoftmax.cpp | 144 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qsigmoid.cpp | 147 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qnormalization.cpp | 133 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qclamp.cpp | 169 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qmatmul.cpp | 181 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
H A D | qrelu.cpp | 228 TORCH_LIBRARY_IMPL(quantized, QuantizedCPU, m) { in TORCH_LIBRARY_IMPL() argument
|
/aosp_15_r20/external/pytorch/tools/test/ |
H A D | test_codegen.py | 329 "dispatch": {"CPU": "kernel_2", "QuantizedCPU": "custom::kernel_3"}, 337 DispatchKey.QuantizedCPU: {},
|
/aosp_15_r20/external/pytorch/cmake/ |
H A D | Codegen.cmake | 100 list(APPEND CUSTOM_BUILD_FLAGS --backend_whitelist CPU QuantizedCPU Vulkan) 102 list(APPEND CUSTOM_BUILD_FLAGS --backend_whitelist CPU QuantizedCPU)
|
/aosp_15_r20/external/pytorch/test/quantization/eager/ |
H A D | test_fuse_eager.py | 92 …ex(RuntimeError, "Could not run 'aten::native_batch_norm' with arguments from the 'QuantizedCPU'"): 101 …ex(RuntimeError, "Could not run 'aten::native_batch_norm' with arguments from the 'QuantizedCPU'"):
|
/aosp_15_r20/external/pytorch/torch/csrc/utils/ |
H A D | tensor_types.cpp | 44 case at::Backend::QuantizedCPU: in backend_to_string()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/core/op_registration/ |
H A D | op_allowlist.h | 192 …// return k == DispatchKey::CPU || k == DispatchKey::Vulkan || k == DispatchKey::QuantizedCPU || k… in dispatch_key_allowlist_check()
|