Optimum
Ctrlk
  • ๐ŸŒOVERVIEW
  • ๐ŸŒHABANA
  • ๐ŸŒINTEL
  • ๐ŸŒAWS TRAINIUM/INFERENTIA
  • ๐ŸŒFURIOSA
  • ๐ŸŒONNX RUNTIME
    • Overview
    • Quick tour
    • ๐ŸŒHOW-TO GUIDES
      • Inference pipelines
      • Models for inference
      • How to apply graph optimization
      • How to apply dynamic and static quantization
      • How to accelerate training
      • Accelerated inference on NVIDIA GPUs
    • ๐ŸŒCONCEPTUAL GUIDES
    • ๐ŸŒREFERENCE
  • ๐ŸŒEXPORTERS
  • ๐ŸŒTORCH FX
  • ๐ŸŒBETTERTRANSFORMER
  • ๐ŸŒLLM QUANTIZATION
  • ๐ŸŒUTILITIES
Powered by GitBook
On this page
  1. ๐ŸŒONNX RUNTIME

๐ŸŒHOW-TO GUIDES

Inference pipelinesModels for inferenceHow to apply graph optimizationHow to apply dynamic and static quantizationHow to accelerate trainingAccelerated inference on NVIDIA GPUs
PreviousQuick tourNextInference pipelines