Hands-on Practical: Applying Mixed-Precision Training
Was this section helpful?
Mixed-Precision Training, Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, Hao Wu, 2018International Conference on Learning Representations (ICLR) 2018DOI: 10.48550/arXiv.1710.03740 - This foundational paper introduced the concept and techniques of mixed-precision training, demonstrating how to combine FP16 and FP32 for improved training speed and memory efficiency.
Automatic Mixed Precision Package - torch.cuda.amp, PyTorch Developers, 2025 - Official PyTorch documentation detailing the use of torch.cuda.amp.autocast and torch.cuda.amp.GradScaler for mixed-precision training.
Deep Learning Performance Guide, NVIDIA Developer Documentation, 2023 (NVIDIA Corporation) - Provides comprehensive guidance on optimizing deep learning performance on NVIDIA GPUs, including explanations of mixed-precision training and Tensor Core usage.