Performance Comparison: XGBoost vs. LightGBM vs. CatBoost
Was this section helpful?
XGBoost: A Scalable Tree Boosting System, Tianqi Chen and Carlos Guestrin, 2016Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '16) (Association for Computing Machinery (ACM))DOI: 10.1145/2939672.2939785 - The foundational paper introducing XGBoost, detailing its parallel and distributed training, cache-aware access, and tree boosting algorithms.
LightGBM: A Highly Efficient Gradient Boosting Decision Tree, Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu, 2017Advances in Neural Information Processing Systems 30 (NeurIPS 2017) (Curran Associates, Inc.)DOI: 10.5555/3295222.3295232 - Presents LightGBM, outlining its Gradient-based One-Side Sampling (GOSS) and Exclusive Feature Bundling (EFB) techniques for faster training and leaf-wise tree growth.
CatBoost: Gradient Boosting with Decision Trees, Anna Veronika Dorogush, Andrey Gulin, Gleb Gusev, Neliya Kazimov, and Liudmila Prokhorenkova, 2020Journal of Machine Learning Research, Vol. 21 - Describes the CatBoost algorithm, with a focus on its ordered boosting scheme for categorical features and symmetric (oblivious) tree construction.
XGBoost Documentation, XGBoost Contributors, 2024 - The official documentation providing practical guidance, API details, and tutorials for using XGBoost.