MLIR: A Compiler Infrastructure for the End of Moores Law, Chris Lattner, Mehdi Amini, River Riddle, Albert Cohen, Alan Mycroft, Chris Leary, John S. Oliver, Andrew Davis, Edward R. Lafferty, Vinod Grover, John Kew, Jason Furman, Stephan Herhut, Paulius Stankaitis, Stephen Neuendorffer, Andy Lin, Matthew Allman, James Manson, Andy Keep, N. S. Jayasena, 2021Proceedings of the ACM on Programming Languages (PACMPL), Vol. 5 (ACM)DOI: 10.1145/3434384 - Introduces MLIR, a framework for building modern ML compilers with multi-level intermediate representations, central to understanding progressive lowering.
TVM: An Automated End-to-End Optimizing Compiler for Deep Learning, Tianqi Chen, Thierry Moreau, Ziheng Jiang, Lianmin Zheng, Eddie Yan, Haichen Shen, Meghan Cowan, Leyuan Wang, Yuwei Hu, Luis Ceze, Carlos Guestrin, Arvind Krishnamurthy, 201813th USENIX Symposium on Operating Systems Design and Implementation (OSDI 18) (USENIX Association) - Foundational paper for Apache TVM, an open-source ML compiler and runtime stack discussed in the section, detailing its architecture, Relay IR, and auto-tuning.
TorchDynamo: A Python-level JIT compiler for PyTorch, Michael P. Veith, Adam Paszke, et al., 2023Proceedings of the ACM on Programming Languages (PACMPL), Vol. 7 (Association for Computing Machinery (ACM))DOI: 10.1145/3573711 - Details TorchDynamo, a core component of PyTorch 2.x, representing a recent example of an ML compiler stack mentioned in the section.
XLA: Optimizing Compiler for Machine Learning, Google Developers, 2024 (Google) - Official documentation providing an overview of XLA, the compiler for TensorFlow, JAX, and Flax, and its high-level architecture.