Finding Structure in Time, Jeffrey L. Elman, 1990Cognitive Science, Vol. 14 (Wiley)DOI: 10.1207/s15516709cog1402_1 - Describes the Simple Recurrent Network (SRN), often called an Elman network, which is the foundational architecture of the simple RNN cell implemented in the section. This paper introduced the idea of context layers to maintain state over time.
Deep Learning, Ian Goodfellow, Yoshua Bengio, and Aaron Courville, 2016 (MIT Press) - Provides a rigorous mathematical and conceptual foundation for recurrent neural networks, including the basic RNN cell, its equations, and challenges like vanishing/exploding gradients, as covered in Chapter 10.
Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, Aurélien Géron, 2022 (O'Reilly Media) - 4th edition. Offers practical implementations and explanations of RNNs, including simple RNNs, often building from fundamental concepts to framework-specific code. It serves as an excellent bridge for readers transitioning from NumPy implementations to high-level APIs.
Lecture Notes on Recurrent Neural Networks (Winter 2023, Lecture 4), Chris Chute, Jenny Hong, Kevin Du, 2023 (Stanford University) - Provides detailed lecture notes and explanations of RNN architectures, including the basic RNN cell, its forward pass, and the motivation behind it, often accompanied by illustrative diagrams and theoretical insights. From the Stanford CS224N Winter 2023 course.