Differences between revisions 1 and 2
Revision 1 as of 2025-03-20 13:10:10
Size: 798
Comment: Initial commit
Revision 2 as of 2025-04-08 14:47:27
Size: 912
Comment: Cleanup
Deletions are marked like this. Additions are marked like this.
Line 3: Line 3:
'''XGBoost''' is a gradient boosting library. '''XGBoost''' is a software implementation of [[Statistics/GradientBoosting|gradient boosting]] for estimating [[Statistics/DecisionTrees|decision trees]].
Line 19: Line 19:
== Importance == == Interpretation ==

XGBoost

XGBoost is a software implementation of gradient boosting for estimating decision trees.


Usage

XGBoost is written in C++, but generally is used through the official bindings for either R or Python.


Interpretation

Features of an XGBoost model are evaluated according to three attributes:

  • Gain: relative contribution of a feature to the overall model

  • Cover: proportion of observations related to a feature

    • the number of observations where this feature is the deciding tree split, then normalized across all features
  • Frequence: proportion of tree splits formed by a feature

All of these attributes are relative, sum to 1, and can be directly compared.


CategoryRicottone

XGBoost (last edited 2025-04-08 14:47:27 by DominicRicottone)