2014-02-18 16:49:23 +08:00
2014-02-16 11:44:03 +08:00
2014-02-06 16:06:18 -08:00
2014-02-15 17:45:48 -08:00
2014-02-07 22:43:13 -08:00
2014-02-15 11:22:50 -08:00

xgboost: A Gradient Boosting Library

Creater: Tianqi Chen: tianqi.tchen AT gmail

General Purpose Gradient Boosting Library

Goal: A stand-alone efficient library to do learning via boosting in functional space

Features:

  • Sparse feature format, handling of missing features. This allows efficient categorical feature encoding as indicators. The speed of booster only depends on number of existing features.
  • Layout of gradient boosting algorithm to support generic tasks, see project wiki.

Planned key components:

  • Gradient boosting models:
    • regression tree (GBRT)
    • linear model/lasso
  • Objectives to support tasks:
    • regression
    • classification
    • ranking
    • matrix factorization
    • structured prediction (3) OpenMP implementation(optional)

File extension convention: (1) .h are interface, utils and data structures, with detailed comment; (2) .cpp are implementations that will be compiled, with less comment; (3) .hpp are implementations that will be included by .cpp, with less comment

See also: https://github.com/tqchen/xgboost/wiki

Description
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
Readme 33 MiB
Languages
C++ 45.5%
Python 20.3%
Cuda 15.2%
R 6.8%
Scala 6.4%
Other 5.6%