Mixed-Precision Training, Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, Hao Wu, 2018ICLRDOI: 10.48550/arXiv.1710.03740 - Presents the foundational techniques and benefits of automatic mixed-precision training using FP16 and loss scaling.
Automatic Mixed Precision package - torch.cuda.amp, PyTorch Development Team, 2025 (PyTorch Foundation) - Documentation for PyTorch's Automatic Mixed Precision (AMP) utilities, including support for FP16 and BF16.
Optimize Your Training For Memory Efficiency, Hugging Face, 2024 - Hugging Face Accelerate guide on memory optimization strategies for training large models, covering gradient accumulation, gradient checkpointing, and mixed precision.