A distinctive feature of CatBoost, setting it apart from algorithms like XGBoost and LightGBM, is its use of Oblivious Trees (also known as symmetric trees) as base learners. While most decision tree algorithms build asymmetric trees where different splits can occur at the same depth depending on the path taken, CatBoost enforces a stricter structure.
In an Oblivious Tree, all nodes at the same depth level test the exact same feature with the exact same split condition (threshold). This means that the tree is perfectly balanced and symmetric. Every path from the root to a leaf node has the same length, and the splitting criteria encountered at each step are identical for all samples traversing that level.
Consider a simple example. If an oblivious tree has depth 2:
Feature_X < threshold_1
.Feature_Y > threshold_2
.This results in a structure where the sequence of features and thresholds tested is fixed for a given depth, regardless of which branch a data point follows.
Example of an Oblivious Tree structure with depth 2. Note how both nodes at Level 1 use the same splitting condition (
Feature Y < T2
).
Why impose such a structural constraint? Oblivious Trees offer several significant benefits within the CatBoost framework:
Oblivious Trees integrate well with CatBoost's other innovations. The generation of complex feature combinations (interactions between categorical features) can lead to a high-dimensional feature space. The simple, regular structure of oblivious trees provides a robust and computationally manageable way to build models even when considering these numerous generated features. The inherent regularization also helps manage the potential complexity introduced by these combinations.
The main trade-off is that a single oblivious tree might be less powerful or require greater depth than an asymmetric tree to capture certain complex interactions directly. An asymmetric tree can tailor splits specifically for the data subset reaching a particular node. However, CatBoost's ensemble approach, combined with features like Ordered Boosting and automatic feature combinations, aims to overcome this limitation at the model level, while retaining the performance and regularization benefits of the symmetric structure.
In essence, Oblivious Trees are a core element of CatBoost's design philosophy, prioritizing efficiency, regularization, and optimized handling of categorical data through a unique and structured approach to building decision tree ensembles. This structural choice is a fundamental differentiator compared to other popular gradient boosting implementations.
© 2025 ApX Machine Learning