Dropout: A Simple Way to Prevent Overfitting, Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, Ruslan Salakhutdinov, 2014Journal of Machine Learning Research (JMLR), Vol. 15DOI: 10.5555/2627435.2670313 - Foundational paper introducing the dropout regularization technique, explaining its mechanism and benefits.
Deep Learning, Ian Goodfellow, Yoshua Bengio, Aaron Courville, 2016 (MIT Press) - A comprehensive textbook with sections dedicated to regularization techniques, including detailed explanations of L2 regularization (weight decay) and dropout.
Layers: Dropout, The Flux.jl Community, 2025 (The Flux.jl Community) - Official documentation for the Dropout layer in Flux.jl, detailing its usage and parameters for Julia deep learning models.
Optimisers: WeightDecay, The Flux.jl Community, 2024 - Official documentation for applying WeightDecay with optimizers in Flux.jl, demonstrating its integration into the training process.