Reducing the Dimensionality of Data with Neural Networks, Geoffrey E. Hinton, Ruslan R. Salakhutdinov, 2006Science, Vol. 313 (American Association for the Advancement of Science)DOI: 10.1126/science.1127647 - Foundational paper introducing deep autoencoders and demonstrating the effectiveness of unsupervised pre-training for dimensionality reduction and initializing deep networks.
Extracting and Composing Robust Features with Denoising Autoencoders, Pascal Vincent, Hugo Larochelle, Yoshua Bengio, Pierre-Antoine Manzagol, 2008Proceedings of the 25th International Conference on Machine Learning (ICML) (ACM Press)DOI: 10.1145/1390156.1390294 - Introduces Denoising Autoencoders (DAEs), a specific type of autoencoder particularly useful for learning robust representations, often favored for end-to-end pre-training.
Deep Learning, Ian Goodfellow, Yoshua Bengio, Aaron Courville, 2016 (MIT Press) - A comprehensive textbook covering the theoretical foundations and practical applications of deep learning, including detailed sections on autoencoders and their historical role in pre-training.
Masked Autoencoders Are Scalable Vision Learners, Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, Ross Girshick, 2022Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (IEEE) - Introduces Masked Autoencoders (MAE), a significant advancement in self-supervised learning that leverages an autoencoding approach to learn strong visual representations by reconstructing masked patches.