Quantized Neural Networks: Training Neural Networks with Low-Precision Weights and Activations, Matthieu Courbariaux, Itay Hubara, Daniel Soudry, Ran El-Yaniv, Yoshua Bengio, 2016Advances in Neural Information Processing Systems, Vol. 29 (Curran Associates, Inc.) - This influential paper applies the Straight-Through Estimator to train neural networks with extremely low-precision weights and activations, laying important groundwork for Quantization-Aware Training.
Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference, Benoit Jacob, Skirmantas Kligys, Bo Chen, Menglong Zhu, Matthew Tang, Andrew Howard, Hartwig Adam, Dmitry Kalenichenko, 20182018 IEEE Conference on Computer Vision and Pattern Recognition (IEEE)DOI: 10.1109/CVPR.2018.00794 - A seminal paper describing the Quantization-Aware Training (QAT) method, which explicitly uses Straight-Through Estimators to simulate quantization during training, enabling high-accuracy low-bit inference.