Long Short-Term Memory, Sepp Hochreiter, Jürgen Schmidhuber, 1997Neural Computation, Vol. 9 (MIT Press)DOI: 10.1162/neco.1997.9.8.1735 - Introduces the Long Short-Term Memory (LSTM) network architecture, which is a fundamental component for effectively modeling long-range dependencies in recurrent autoencoders.
Sequence to Sequence Learning with Neural Networks, Ilya Sutskever, Oriol Vinyals, Quoc V. Le, 2014Advances in Neural Information Processing Systems 27, Vol. 27 - Presents the seminal encoder-decoder architecture using LSTMs, which forms the basis for recurrent autoencoders by learning to map input sequences to target sequences, directly mirroring the RAE structure for unsupervised reconstruction.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation, Kyunghyun Cho, Bart van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, Yoshua Bengio, 2014Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Association for Computational Linguistics)DOI: 10.3115/v1/D14-1179 - Introduces the Gated Recurrent Unit (GRU) and an RNN encoder-decoder structure for learning fixed-dimensional representations of variable-length sequences, highly relevant to the recurrent autoencoder's purpose.
Deep Learning, Ian Goodfellow, Yoshua Bengio, Aaron Courville, 2016 (MIT Press) - A foundational textbook providing comprehensive coverage of recurrent neural networks, LSTMs, GRUs, autoencoders, and sequence modeling, offering theoretical background for recurrent autoencoders.