Learning binary weights for efficient implementation of deep neural networks, Matthieu Courbariaux, Yoshua Bengio, and Jean-Pierre David, 2015Advances in Neural Information Processing Systems (NeurIPS) (Neural Information Processing Systems Foundation)DOI: 10.5591/978-1-55860-845-8.3123 - This paper popularizes the Straight-Through Estimator (STE) for training neural networks with low-precision weights, directly relevant to the backward pass challenge in QAT.
Quantization Recipes, PyTorch Authors, 2024 (PyTorch Foundation) - Official documentation providing practical implementation details for Quantization-Aware Training (QAT) in PyTorch, including the use of fake quantization modules.