Transformers
Ctrlk
  • ๐ŸŒGET STARTED
  • ๐ŸŒTUTORIALS
  • ๐ŸŒTASK GUIDES
  • ๐ŸŒDEVELOPER GUIDES
  • ๐ŸŒPERFORMANCE AND SCALABILITY
    • Overview
    • ๐ŸŒEFFICIENT TRAINING TECHNIQUES
      • Methods and tools for efficient training on a single GPU
      • Multiple GPUs and parallelism
      • Efficient training on CPU
      • Distributed CPU training
      • Training on TPUs
      • Training on TPU with TensorFlow
      • Training on Specialized Hardware
      • Custom hardware for training
      • Hyperparameter Search using Trainer API
    • ๐ŸŒOPTIMIZING INFERENCE
    • Instantiating a big model
    • Troubleshooting
    • XLA Integration for TensorFlow Models
    • Optimize inference using `torch.compile()`
  • ๐ŸŒCONTRIBUTE
  • ๐ŸŒCONCEPTUAL GUIDES
  • ๐ŸŒAPI
  • ๐ŸŒINTERNAL HELPERS
Powered by GitBook
On this page
  1. ๐ŸŒPERFORMANCE AND SCALABILITY

๐ŸŒEFFICIENT TRAINING TECHNIQUES

Methods and tools for efficient training on a single GPUMultiple GPUs and parallelismEfficient training on CPUDistributed CPU trainingTraining on TPUsTraining on TPU with TensorFlowTraining on Specialized HardwareCustom hardware for trainingHyperparameter Search using Trainer API
PreviousOverviewNextMethods and tools for efficient training on a single GPU