XGBoost is so efficient and powerful for Kaggle competitions that it deserves a post of its own. Being an extension of the classic gradient boosting machine (gbm), xgboost (extreme gradient boosting) is optimized to be highly scalable, efficient, and portable.

It’s primarily developed by Tianqi Chen at University of Washington, with the R Package authored by Tong He:

XGBoost uses the same model (tree ensembles) as random forest, but the difference is in how the model is trained. XGBoost learn the trees with an additive strategy: fix what it has learned, and add one new tree at a time.


1 Comment

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s