CatBoost Comparison results with other Boosting Algorithm
Default CatBoost | Tuned CatBoost | Default LightGBM | Tuned LightGBM | Default XGBoost | Tuned XGBoost | Default H2O |
Adult | 0.272978 (±0.0004) (+1.20%) | 0.269741 (±0.0001) | 0.287165 (±0.0000) (+6.46%) | 0.276018 (±0.0003) (+2.33%) | 0.280087 (±0.0000) (+3.84%) | 0.275423 (±0.0002) (+2.11%) |
Amazon | 0.138114 (±0.0004) (+0.29%) | 0.137720 (±0.0005) | 0.167159 (±0.0000) (+21.38%) | 0.163600 (±0.0002) (+18.79%) | 0.165365 (±0.0000) (+20.07%) | 0.163271 (±0.0001) (+18.55%) |
Appet | 0.071382 (±0.0002) (-0.18%) | 0.071511 (±0.0001) | 0.074823 (±0.0000) (+4.63%) | 0.071795 (±0.0001) (+0.40%) | 0.074659 (±0.0000) (+4.40%) | 0.071760 (±0.0000) (+0.35%) |
Click | 0.391116 (±0.0001) (+0.05%) | 0.390902 (±0.0001) | 0.397491 (±0.0000) (+1.69%) | 0.396328 (±0.0001) (+1.39%) | 0.397638 (±0.0000) (+1.72%) | 0.396242 (±0.0000) (+1.37%) |
Internet | 0.220206 (±0.0005) (+5.49%) | 0.208748 (±0.0011) | 0.236269 (±0.0000) (+13.18%) | 0.223154 (±0.0005) (+6.90%) | 0.234678 (±0.0000) (+12.42%) | 0.225323 (±0.0002) (+7.94%) |
Kdd98 | 0.194794 (±0.0001) (+0.06%) | 0.194668 (±0.0001) | 0.198369 (±0.0000) (+1.90%) | 0.195759 (±0.0001) (+0.56%) | 0.197949 (±0.0000) (+1.69%) | 0.195677 (±0.0000) (+0.52%) |
Kddchurn | 0.231935 (±0.0004) (+0.28%) | 0.231289 (±0.0002) | 0.235649 (±0.0000) (+1.88%) | 0.232049 (±0.0001) (+0.33%) | 0.233693 (±0.0000) (+1.04%) | 0.233123 (±0.0001) (+0.79%) |
Kick | 0.284912 (±0.0003) (+0.04%) | 0.284793 (±0.0002) | 0.298774 (±0.0000) (+4.91%) | 0.295660 (±0.0000) (+3.82%) | 0.298161 (±0.0000) (+4.69%) | 0.294647 (±0.0000) (+3.46%) |
Upsel | 0.166742 (±0.0002) (+0.37%) | 0.166128 (±0.0002) | 0.171071 (±0.0000) (+2.98%) | 0.166818 (±0.0000) (+0.42%) | 0.168732 (±0.0000) (+1.57%) | 0.166322 (±0.0001) (+0.12%) |
CatBoost in Machine Learning
We often encounter datasets that contain categorical features and to fit these datasets into the Boosting model we apply various encoding techniques to the dataset such as One-Hot Encoding or Label Encoding. But applying One-Hot encoding creates a sparse matrix which may sometimes lead to the overfitting of the model to handle this issue we use CatBoost. CatBoost automatically handles categorical features.
Table of Content
- What is CatBoost?
- Features of CatBoost
- CatBoost Comparison results with other Boosting Algorithm
- Prerequisites to start Catboost
- CatBoost Installation
- Difference between CatBoost, LightGBM and XGboost
- Limitations of CatBoost
- Conclusions
- Frequently Asked Questions on CatBoost