Transformers
Ctrlk
  • 🌍GET STARTED
  • 🌍TUTORIALS
  • 🌍TASK GUIDES
  • 🌍DEVELOPER GUIDES
  • 🌍PERFORMANCE AND SCALABILITY
    • Overview
    • 🌍EFFICIENT TRAINING TECHNIQUES
      • Methods and tools for efficient training on a single GPU
      • Multiple GPUs and parallelism
      • Efficient training on CPU
      • Distributed CPU training
      • Training on TPUs
      • Training on TPU with TensorFlow
      • Training on Specialized Hardware
      • Custom hardware for training
      • Hyperparameter Search using Trainer API
    • 🌍OPTIMIZING INFERENCE
    • Instantiating a big model
    • Troubleshooting
    • XLA Integration for TensorFlow Models
    • Optimize inference using `torch.compile()`
  • 🌍CONTRIBUTE
  • 🌍CONCEPTUAL GUIDES
  • 🌍API
  • 🌍INTERNAL HELPERS
Powered by GitBook
On this page
  1. 🌍PERFORMANCE AND SCALABILITY

🌍EFFICIENT TRAINING TECHNIQUES

Methods and tools for efficient training on a single GPUMultiple GPUs and parallelismEfficient training on CPUDistributed CPU trainingTraining on TPUsTraining on TPU with TensorFlowTraining on Specialized HardwareCustom hardware for trainingHyperparameter Search using Trainer API
PreviousOverviewNextMethods and tools for efficient training on a single GPU