Python toolkit for building production-ready LLM applications. Modular utilities for prompts, RAG, agents, structured outputs, and multi-provider support.
Mixed-Precision Training, Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, Hao Wu, 2018ICLRDOI: 10.48550/arXiv.1710.03740 - Presents the foundational techniques and benefits of automatic mixed-precision training using FP16 and loss scaling.
Automatic Mixed Precision package - torch.cuda.amp, PyTorch Development Team, 2025 (PyTorch Foundation) - Documentation for PyTorch's Automatic Mixed Precision (AMP) utilities, including support for FP16 and BF16.
Optimize Your Training For Memory Efficiency, Hugging Face, 2024 - Hugging Face Accelerate guide on memory optimization strategies for training large models, covering gradient accumulation, gradient checkpointing, and mixed precision.