Transformers
search
⌘Ctrlk
Transformers
  • 🌍GET STARTED
  • 🌍TUTORIALS
  • 🌍TASK GUIDES
  • 🌍DEVELOPER GUIDES
  • 🌍PERFORMANCE AND SCALABILITY
    • Overview
    • 🌍EFFICIENT TRAINING TECHNIQUES
      • Methods and tools for efficient training on a single GPU
      • Multiple GPUs and parallelism
      • Efficient training on CPU
      • Distributed CPU training
      • Training on TPUs
      • Training on TPU with TensorFlow
      • Training on Specialized Hardware
      • Custom hardware for training
      • Hyperparameter Search using Trainer API
    • 🌍OPTIMIZING INFERENCE
    • Instantiating a big model
    • Troubleshooting
    • XLA Integration for TensorFlow Models
    • Optimize inference using `torch.compile()`
  • 🌍CONTRIBUTE
  • 🌍CONCEPTUAL GUIDES
  • 🌍API
  • 🌍INTERNAL HELPERS
gitbookPowered by GitBook
block-quoteOn this pagechevron-down
  1. 🌍PERFORMANCE AND SCALABILITY

🌍EFFICIENT TRAINING TECHNIQUES

Methods and tools for efficient training on a single GPUchevron-rightMultiple GPUs and parallelismchevron-rightEfficient training on CPUchevron-rightDistributed CPU trainingchevron-rightTraining on TPUschevron-rightTraining on TPU with TensorFlowchevron-rightTraining on Specialized Hardwarechevron-rightCustom hardware for trainingchevron-rightHyperparameter Search using Trainer APIchevron-right
PreviousOverviewchevron-leftNextMethods and tools for efficient training on a single GPUchevron-right