Lines Matching full:inference
11 -> python benchmarks/runner.py --suites=torchbench --inference
13 below) for inference, run them and visualize the logs.
16 -> python benchmarks/runner.py --print-run-commands --suites=torchbench --inference
19 -> python benchmarks/runner.py --visualize-logs --suites=torchbench --inference
22 -> python benchmarks/runner.py --suites=torchbench --inference --dtypes=float16
80 "inference": {
81 "aot_eager": "--inference --backend=aot_eager ",
82 "eager": "--inference --backend=eager ",
83 "ts_nnc": "--inference --speedup-ts ",
84 "ts_nvfuser": "--inference -n100 --speedup-ts --nvfuser ",
85 "trt": "--inference -n100 --speedup-trt ",
86 "ts_nvfuser_cudagraphs": "--inference --backend=cudagraphs_ts ",
87 "inductor": "--inference -n50 --inductor ",
88 "inductor_no_cudagraphs": "--inference -n50 --inductor --disable-cudagraphs ",
89 … "inductor_max_autotune": "--inference -n50 --inductor --inductor-compile-mode max-autotune ",
91 …"--inference -n50 --inductor --inductor-compile-mode max-autotune-no-cudagraphs --disable-cudagrap…
93 "torchscript-onnx": "--inference -n5 --torchscript-onnx",
94 "dynamo-onnx": "--inference -n5 --dynamo-onnx",
98 INFERENCE_COMPILERS = tuple(TABLE["inference"].keys())
108 "inference": [
116 "inference": ["inductor", "inductor_no_cudagraphs"],
178 …help=f"For --inference, options are {INFERENCE_COMPILERS}. For --training, options are {TRAINING_C…
240 # Choose either inference or training
243 "--inference", action="store_true", help="Only run inference related tasks"
366 if args.inference:
367 return "inference"
863 "and backward pass for training and forward pass only for inference. "
1479 if args.inference:
1480 compilers = DEFAULTS["inference"] if args.compilers is None else args.compilers
1482 DEFAULTS["flag_compilers"]["inference"]