Optimum
  • 🌍OVERVIEW
    • Optimum
    • Installation
    • Quick tour
    • Notebooks
    • 🌍CONCEPTUAL GUIDES
      • Quantization
  • 🌍HABANA
    • BOINC AI Optimum Habana
    • Installation
    • Quickstart
    • 🌍TUTORIALS
      • Overview
      • Single-HPU Training
      • Distributed Training
      • Run Inference
      • Stable Diffusion
      • LDM3D
    • 🌍HOW-TO GUIDES
      • Overview
      • Pretraining Transformers
      • Accelerating Training
      • Accelerating Inference
      • How to use DeepSpeed
      • Multi-node Training
    • 🌍CONCEPTUAL GUIDES
      • What are Habana's Gaudi and HPUs?
    • 🌍REFERENCE
      • Gaudi Trainer
      • Gaudi Configuration
      • Gaudi Stable Diffusion Pipeline
      • Distributed Runner
  • 🌍INTEL
    • BOINC AI Optimum Intel
    • Installation
    • 🌍NEURAL COMPRESSOR
      • Optimization
      • Distributed Training
      • Reference
    • 🌍OPENVINO
      • Models for inference
      • Optimization
      • Reference
  • 🌍AWS TRAINIUM/INFERENTIA
    • BOINC AI Optimum Neuron
  • 🌍FURIOSA
    • BOINC AI Optimum Furiosa
    • Installation
    • 🌍HOW-TO GUIDES
      • Overview
      • Modeling
      • Quantization
    • 🌍REFERENCE
      • Models
      • Configuration
      • Quantization
  • 🌍ONNX RUNTIME
    • Overview
    • Quick tour
    • 🌍HOW-TO GUIDES
      • Inference pipelines
      • Models for inference
      • How to apply graph optimization
      • How to apply dynamic and static quantization
      • How to accelerate training
      • Accelerated inference on NVIDIA GPUs
    • 🌍CONCEPTUAL GUIDES
      • ONNX And ONNX Runtime
    • 🌍REFERENCE
      • ONNX Runtime Models
      • Configuration
      • Optimization
      • Quantization
      • Trainer
  • 🌍EXPORTERS
    • Overview
    • The TasksManager
    • 🌍ONNX
      • Overview
      • 🌍HOW-TO GUIDES
        • Export a model to ONNX
        • Add support for exporting an architecture to ONNX
      • 🌍REFERENCE
        • ONNX configurations
        • Export functions
    • 🌍TFLITE
      • Overview
      • 🌍HOW-TO GUIDES
        • Export a model to TFLite
        • Add support for exporting an architecture to TFLite
      • 🌍REFERENCE
        • TFLite configurations
        • Export functions
  • 🌍TORCH FX
    • Overview
    • 🌍HOW-TO GUIDES
      • Optimization
    • 🌍CONCEPTUAL GUIDES
      • Symbolic tracer
    • 🌍REFERENCE
      • Optimization
  • 🌍BETTERTRANSFORMER
    • Overview
    • 🌍TUTORIALS
      • Convert Transformers models to use BetterTransformer
      • How to add support for new architectures?
  • 🌍LLM QUANTIZATION
    • GPTQ quantization
  • 🌍UTILITIES
    • Dummy input generators
    • Normalized configurations
Powered by GitBook
On this page
  • 🌍 Optimum notebooks
  • Optimum Habana
  • Optimum Intel
  • Optimum ONNX Runtime
  1. OVERVIEW

Notebooks

🌍 Optimum notebooks

You can find here a list of the notebooks associated with each accelerator in 🌍 Optimum.

Optimum Habana

Notebook
Description
Colab
Studio Lab

Show how to use DeepSpeed to pre-train/fine-tune the 1.6B-parameter GPT2-XL for causal language modeling on Habana Gaudi.

Optimum Intel

OpenVINO

Notebook
Description
Colab
Studio Lab

Explains how to export your model to OpenVINO and run inference with OpenVINO Runtime on various tasks

Show how to load and compare outputs from two Stable Diffusion models with different precision

Neural Compressor

Notebook
Description
Colab
Studio Lab

Optimum ONNX Runtime

Notebook
Description
Colab
Studio Lab

Show how to fine-tune a T5 model on the BBC news corpus.

Show how to fine-tune a DeBERTa model on the squad.

PreviousQuick tourNextCONCEPTUAL GUIDES

Last updated 1 year ago

Show how to apply post-training quantization on a question answering model using and to accelerate inference with OpenVINO

Show how to apply quantization while training your model using Intel for any GLUE task.

Show how to apply static and dynamic quantization on a model using for any GLUE task.

Show how to DistilBERT model on GLUE tasks using .

🌍
How to use DeepSpeed to train models with billions of parameters on Habana Gaudi
How to run inference with OpenVINO
How to quantize a question answering model with NNCF
NNCF
Compare outputs of a quantized Stable Diffusion model with its full-precision counterpart
How to quantize a model with Intel Neural Compressor for text classification
Neural Compressor
How to quantize a model with ONNX Runtime for text classification
ONNX Runtime
How to fine-tune a model for text classification with ONNX Runtime
ONNX Runtime
How to fine-tune a model for summarization with ONNX Runtime
How to fine-tune DeBERTa for question-answering with ONNX Runtime