commit
962c2432a0
@ -1,7 +1,7 @@
|
||||
Distributed XGBoost: Hadoop Version
|
||||
====
|
||||
* The script in this fold shows an example of how to run distributed xgboost on hadoop platform.
|
||||
* It relies on [Rabit Library](https://github.com/tqchen/rabit) and Hadoop Streaming.
|
||||
* It relies on [Rabit Library](https://github.com/tqchen/rabit) (Reliable Allreduce and Broadcast Interface) and Hadoop Streaming. Rabit provides an interface to aggregate gradient values and split statistics, that allow xgboost to run reliably on hadoop. You do not need to care how to update model in each iteration, just use the script ```rabit_hadoop.py```. For those who want to know how it exactly works, plz refer to the main page of [Rabit](https://github.com/tqchen/rabit).
|
||||
* Quick start: run ```bash run_mushroom.sh <n_hadoop_workers> <n_thread_per_worker> <path_in_HDFS>```
|
||||
- This is the hadoop version of binary classification example in the demo folder.
|
||||
- More info of the usage of xgboost can be refered to [wiki page](https://github.com/tqchen/xgboost/wiki)
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user