Parallelism and Distributed Computation, The JAX Authors, 2023 - Official JAX documentation explaining pmap, SPMD, and collective operations for parallel and distributed computation.
GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism, Yanping Huang, Youlong Cheng, Ankur Bapna, Orhan Firat, Mia Xu Chen, Dehao Chen, HyoukJoong Lee, Jiquan Ngiam, Quoc V. Le, Yonghui Wu, Zhifeng Chen, 2019Advances in Neural Information Processing Systems, Vol. 32DOI: 10.48550/arXiv.1811.06965 - A seminal paper introducing pipeline parallelism, a method for training very large models efficiently across multiple devices.