Fix spelling in documents (#6948)
* Update roxygen2 doc. Co-authored-by: fis <jm.yuan@outlook.com>
This commit is contained in:
@@ -6,7 +6,7 @@ The script 'runexp.sh' can be used to run the demo. Here we use [mushroom datase
|
||||
|
||||
### Tutorial
|
||||
#### Generate Input Data
|
||||
XGBoost takes LibSVM format. An example of faked input data is below:
|
||||
XGBoost takes LIBSVM format. An example of faked input data is below:
|
||||
```
|
||||
1 101:1.2 102:0.03
|
||||
0 1:2.1 10001:300 10002:400
|
||||
@@ -15,7 +15,7 @@ XGBoost takes LibSVM format. An example of faked input data is below:
|
||||
Each line represent a single instance, and in the first line '1' is the instance label,'101' and '102' are feature indices, '1.2' and '0.03' are feature values. In the binary classification case, '1' is used to indicate positive samples, and '0' is used to indicate negative samples. We also support probability values in [0,1] as label, to indicate the probability of the instance being positive.
|
||||
|
||||
|
||||
First we will transform the dataset into classic LibSVM format and split the data into training set and test set by running:
|
||||
First we will transform the dataset into classic LIBSVM format and split the data into training set and test set by running:
|
||||
```
|
||||
python mapfeat.py
|
||||
python mknfold.py agaricus.txt 1
|
||||
|
||||
@@ -120,7 +120,7 @@ Please send pull requests if you find ones that are missing here.
|
||||
- [XGBoost - eXtreme Gradient Boosting](http://www.slideshare.net/ShangxuanZhang/xgboost) by Tong He
|
||||
- [How to use XGBoost algorithm in R in easy steps](http://www.analyticsvidhya.com/blog/2016/01/xgboost-algorithm-easy-steps/) by TAVISH SRIVASTAVA ([Chinese Translation 中文翻译](https://segmentfault.com/a/1190000004421821) by [HarryZhu](https://segmentfault.com/u/harryprince))
|
||||
- [Kaggle Solution: What’s Cooking ? (Text Mining Competition)](http://www.analyticsvidhya.com/blog/2015/12/kaggle-solution-cooking-text-mining-competition/) by MANISH SARASWAT
|
||||
- Better Optimization with Repeated Cross Validation and the XGBoost model - Machine Learning with R) by Manuel Amunategui ([Youtube Link](https://www.youtube.com/watch?v=Og7CGAfSr_Y)) ([Github Link](https://github.com/amunategui/BetterCrossValidation))
|
||||
- Better Optimization with Repeated Cross Validation and the XGBoost model - Machine Learning with R) by Manuel Amunategui ([Youtube Link](https://www.youtube.com/watch?v=Og7CGAfSr_Y)) ([GitHub Link](https://github.com/amunategui/BetterCrossValidation))
|
||||
- [XGBoost Rossman Parameter Tuning](https://www.kaggle.com/khozzy/rossmann-store-sales/xgboost-parameter-tuning-template/run/90168/notebook) by [Norbert Kozlowski](https://www.kaggle.com/khozzy)
|
||||
- [Featurizing log data before XGBoost](http://www.slideshare.net/DataRobot/featurizing-log-data-before-xgboost) by Xavier Conort, Owen Zhang etc
|
||||
- [West Nile Virus Competition Benchmarks & Tutorials](http://blog.kaggle.com/2015/07/21/west-nile-virus-competition-benchmarks-tutorials/) by [Anna Montoya](http://blog.kaggle.com/author/annamontoya/)
|
||||
|
||||
@@ -27,4 +27,4 @@ target_link_libraries(api-demo xgboost)
|
||||
```
|
||||
|
||||
# make
|
||||
You can start by modifying the makefile in this directory to fit your need.
|
||||
You can start by modifying the makefile in this directory to fit your need.
|
||||
|
||||
Reference in New Issue
Block a user