Optimum
  • 🌍OVERVIEW
    • Optimum
    • Installation
    • Quick tour
    • Notebooks
    • 🌍CONCEPTUAL GUIDES
      • Quantization
  • 🌍HABANA
    • BOINC AI Optimum Habana
    • Installation
    • Quickstart
    • 🌍TUTORIALS
      • Overview
      • Single-HPU Training
      • Distributed Training
      • Run Inference
      • Stable Diffusion
      • LDM3D
    • 🌍HOW-TO GUIDES
      • Overview
      • Pretraining Transformers
      • Accelerating Training
      • Accelerating Inference
      • How to use DeepSpeed
      • Multi-node Training
    • 🌍CONCEPTUAL GUIDES
      • What are Habana's Gaudi and HPUs?
    • 🌍REFERENCE
      • Gaudi Trainer
      • Gaudi Configuration
      • Gaudi Stable Diffusion Pipeline
      • Distributed Runner
  • 🌍INTEL
    • BOINC AI Optimum Intel
    • Installation
    • 🌍NEURAL COMPRESSOR
      • Optimization
      • Distributed Training
      • Reference
    • 🌍OPENVINO
      • Models for inference
      • Optimization
      • Reference
  • 🌍AWS TRAINIUM/INFERENTIA
    • BOINC AI Optimum Neuron
  • 🌍FURIOSA
    • BOINC AI Optimum Furiosa
    • Installation
    • 🌍HOW-TO GUIDES
      • Overview
      • Modeling
      • Quantization
    • 🌍REFERENCE
      • Models
      • Configuration
      • Quantization
  • 🌍ONNX RUNTIME
    • Overview
    • Quick tour
    • 🌍HOW-TO GUIDES
      • Inference pipelines
      • Models for inference
      • How to apply graph optimization
      • How to apply dynamic and static quantization
      • How to accelerate training
      • Accelerated inference on NVIDIA GPUs
    • 🌍CONCEPTUAL GUIDES
      • ONNX And ONNX Runtime
    • 🌍REFERENCE
      • ONNX Runtime Models
      • Configuration
      • Optimization
      • Quantization
      • Trainer
  • 🌍EXPORTERS
    • Overview
    • The TasksManager
    • 🌍ONNX
      • Overview
      • 🌍HOW-TO GUIDES
        • Export a model to ONNX
        • Add support for exporting an architecture to ONNX
      • 🌍REFERENCE
        • ONNX configurations
        • Export functions
    • 🌍TFLITE
      • Overview
      • 🌍HOW-TO GUIDES
        • Export a model to TFLite
        • Add support for exporting an architecture to TFLite
      • 🌍REFERENCE
        • TFLite configurations
        • Export functions
  • 🌍TORCH FX
    • Overview
    • 🌍HOW-TO GUIDES
      • Optimization
    • 🌍CONCEPTUAL GUIDES
      • Symbolic tracer
    • 🌍REFERENCE
      • Optimization
  • 🌍BETTERTRANSFORMER
    • Overview
    • 🌍TUTORIALS
      • Convert Transformers models to use BetterTransformer
      • How to add support for new architectures?
  • 🌍LLM QUANTIZATION
    • GPTQ quantization
  • 🌍UTILITIES
    • Dummy input generators
    • Normalized configurations
Powered by GitBook
On this page
  1. HABANA

BOINC AI Optimum Habana

PreviousHABANANextInstallation

Last updated 1 year ago

πŸ€— Optimum Habana

πŸ€— Optimum Habana is the interface between the πŸ€— Transformers and πŸ€— Diffusers libraries and . It provides a set of tools that enable easy model loading, training and inference on single- and multi-HPU settings for various downstream tasks as shown in the table below.

HPUs offer fast model training and inference as well as a great price-performance ratio. Check out and for concrete examples. If you are not familiar with HPUs, we recommend you take a look at .

The following model architectures, tasks and device distributions have been validated for πŸ€— Optimum Habana:

In the tables below, βœ… means single-card, multi-card and DeepSpeed have all been validated.

  • Transformers

Architecture
Training
Inference
Tasks

BERT

βœ…

βœ…

RoBERTa

βœ…

βœ…

ALBERT

βœ…

βœ…

DistilBERT

βœ…

βœ…

GPT2

βœ…

βœ…

BLOOM(Z)

❌

  • DeepSpeed

StarCoder

❌

  • Single card

GPT-J

  • DeepSpeed

  • Single card

  • DeepSpeed

GPT-NeoX

  • DeepSpeed

  • DeepSpeed

OPT

❌

  • DeepSpeed

Llama 2 / CodeLlama

  • DeepSpeed

  • LoRA

  • DeepSpeed

  • LoRA

StableLM

❌

  • Single card

Falcon

❌

  • Single card

CodeGen

❌

  • Single card

MPT

❌

  • Single card

T5

βœ…

βœ…

BART

❌

  • Single card

ViT

βœ…

βœ…

Swin

βœ…

βœ…

Wav2Vec2

βœ…

βœ…

CLIP

βœ…

βœ…

BridgeTower

βœ…

βœ…

ESMFold

❌

  • Single card

  • Diffusers

Architecture
Training
Inference
<center>Tasks</center>

Stable Diffusion

❌

  • Single card

LDM3D

❌

  • Single card

Other models and tasks supported by the 🌍 Transformers and 🌍 Diffusers library may also work. You can refer to this for using them with 🌍 Optimum Habana. Besides, explains how to modify any from the 🌍 Transformers library to make it work with 🌍 Optimum Habana.

🌍
Habana’s Gaudi processor (HPU)
this blog post about BERT pre-training
this article benchmarking Habana Gaudi2 versus Nvidia A100 GPUs
our conceptual guide
section
this page
example
text classification
question answering
language modeling
question answering
language modeling
question answering
language modeling
question answering
language modeling
language modeling
text generation
text generation
text generation
language modeling
text generation
language modeling
text generation
text generation
language modeling
text generation
text generation
text generation
text generation
text generation
summarization
translation
question answering
summarization
translation
question answering
image classification
image classification
audio classification
speech recognition
contrastive image-text training
contrastive image-text training
protein folding
text-to-image generation
text-to-image generation