4.4 KiB
4.4 KiB
XGBoost
An optimized general purpose gradient boosting library. The library is parallelized, and also provides an optimized distributed version.
It implements machine learning algorithms under the Gradient Boosting framework, including Generalized Linear Model (GLM) and Gradient Boosted Decision Trees (GBDT). XGBoost can also be distributed and scale to Terascale data
XGBoost is part of Distributed Machine Learning Common projects
Contents
- What's New
- Version
- Documentation
- Build Instruction
- Features
- Distributed XGBoost
- Usecases
- Bug Reporting
- Contributing to XGBoost
- Committers and Contributors
- License
- XGBoost in Graphlab Create
What's New
- XGBoost helps Chenglong Chen to win Kaggle CrowdFlower Competition
- Check out the winning solution at Highlight links
- XGBoost-0.4 release, see CHANGES.md
- XGBoost helps three champion teams to win WWW2015 Microsoft Malware Classification Challenge (BIG 2015)
- Check out the winning solution at Highlight links
- External Memory Version
Version
- Current version xgboost-0.4, a lot improvment has been made since 0.3
- Change log in CHANGES.md
- This version is compatible with 0.3x versions
Features
- Easily accessible through python, R, Julia, CLI
- Fast and memory efficient
- Can be more than 10 times faster than GBM in sklearn and R. benchm-ml numbers
- Handles sparse matrices, support external memory
- Accurate prediction, and used extensively by data scientists and kagglers
- See highlight links
- Distributed and Portable
- The distributed version runs on Hadoop (YARN), MPI, SGE etc.
- Scales to billions of examples and beyond
Bug Reporting
- For reporting bugs please use the xgboost/issues page.
- For generic questions or to share your experience using xgboost please use the XGBoost User Group
Contributing to XGBoost
XGBoost has been developed and used by a group of active community members. Everyone is more than welcome to contribute. It is a way to make the project better and more accessible to more users.
- Check out Feature Wish List to see what can be improved, or open an issue if you want something.
- Contribute to the documents and examples to share your experience with other users.
- Please add your name to CONTRIBUTORS.md after your patch has been merged.
License
© Contributors, 2015. Licensed under an Apache-2 license.
XGBoost in Graphlab Create
- XGBoost is adopted as part of boosted tree toolkit in Graphlab Create (GLC). Graphlab Create is a powerful python toolkit that allows you to do data manipulation, graph processing, hyper-parameter search, and visualization of TeraBytes scale data in one framework. Try the Graphlab Create in http://graphlab.com/products/create/quick-start-guide.html
- Nice blogpost by Jay Gu about using GLC boosted tree to solve kaggle bike sharing challenge: http://blog.graphlab.com/using-gradient-boosted-trees-to-predict-bike-sharing-demand