On Unix systems, it's common for programs to read their input from stdin, and
write their output to stdout. Messages should be written to stderr, where they
won't corrupt a program's output, and where they can be seen by the user even
if the output is being redirected.
This is mostly a problem when XGBoost is being used from Python or from another
program.
* add support for tweedie regression
* added back readme line that was accidentally deleted
* fixed linting errors
* add support for tweedie regression
* added back readme line that was accidentally deleted
* fixed linting errors
* rebased with upstream master and added R example
* changed parameter name to tweedie_variance_power
* linting error fix
* refactored tweedie-nloglik metric to be more like the other parameterized metrics
* added upper and lower bound check to tweedie metric
* add support for tweedie regression
* added back readme line that was accidentally deleted
* fixed linting errors
* added upper and lower bound check to tweedie metric
* added back readme line that was accidentally deleted
* rebased with upstream master and added R example
* rebased again on top of upstream master
* linting error fix
* added upper and lower bound check to tweedie metric
* rebased with master
* lint fix
* removed whitespace at end of line 186 - elementwise_metric.cc
* Fix typos and messages in docs
* parameter.md: Add docs for updater_seq
Mention the updater_seq parameter which sets the order of the tree
updaters to run and also specifies which ones to run. This can be
useful when pruning is not required or even a custom plugin is
being built along with xgboost.
* Add format to the params accepted by DumpModel
Currently, only the test format is supported when trying to dump
a model. The plan is to add more such formats like JSON which are
easy to read and/or parse by machines. And to make the interface
for this even more generic to allow other formats to be added.
Hence, we make some modifications to make these function generic
and accept a new parameter "format" which signifies the format of
the dump to be created.
* Fix typos and errors in docs
* plugin: Mention all the register macros available
Document the register macros currently available to the plugin
writers so they know what exactly can be extended using hooks.
* sparce_page_source: Use same arg name in .h and .cc
* gbm: Add JSON dump
The dump_format argument can be used to specify what type
of dump file should be created. Add functionality to dump
gblinear and gbtree into a JSON file.
The JSON file has an array, each item is a JSON object for the tree.
For gblinear:
- The item is the bias and weights vectors
For gbtree:
- The item is the root node. The root node has a attribute "children"
which holds the children nodes. This happens recursively.
* core.py: Add arg dump_format for get_dump()
The `cd ..;` in the one liner takes you up a directory instead of into the xgboost directory. This will cause that step of the installation to fail. It seems like you are meant to enter the xgboost directory as you did in the instructions for installing xgboost without openmp.
* add back train method but mark as deprecated
* fix scalastyle error
* change class to object in examples
* fix compilation error
* fix mis configuration
* make DMatrix._init_from_npy2d only copy data when necessary
When creating DMatrix from a 2d ndarray, it can unnecessarily copy the input data. This can be problematic when the data is already very large--running out of memory. The copy is temporary (going out of scope at the end of this function) but it still adds to peak memory usage.
``numpy.array`` copies its input no matter what by default. By adding ``copy=False``, it will only do so when necessary. Since XGDMatrixCreateFromMat is readonly on the input buffer, this copy is not needed.
Also added comments explaining when a copy can happen (if data ordering/layout is wrong or if type is not 32-bit float).
* remove whitespace
* correct CalcDCG in rank_metric.cc
DCG use log base-2, however `std::log` returns log base-e.
* correct CalcDCG in rank_obj.cc
DCG use log base-2, however `std::log` returns log base-e.
* use std::log2 instead of std::log
make it more elegant
* use std::log2 instead of std::log
make it more elegant
*Fix 1439
*Fix python_wrapper when eval set name contain '-' will cause early_stop maximize variable con't set to True propely
Change-Id: Ib0595afd4ae7b445a84c00a3a8faeccc506c6d13
* Changes for Mingw64 compilation to ensure long is a consistent size.
Mainly impacts the Java API which would not compile, but there may be
silent errors on Windows with large datasets before this patch (as long
is 32-bits when compiled with mingw64 even in 64-bit mode).
* Adding ifdefs to ensure it still compiles on MacOS
* Makefile and create_jni.bat changes for Windows.
* Switching XGDMatrixCreateFromCSREx JNI call to use size_t cast
* Fixing lint error, adding profile switching to jvm-packages build to make create-jni.bat get called, adding myself to Contributors.Md
add_library(libxgboost SHARED ${SOURCES}) builds a library named
liblibxgboost.so; However, simply changing it to add_library(xgboost ...)
won't work, as add_executable(xgboost ...) and add_library(xgbboost ...)
will then have the same target name. This patch correctly handles the
same-name situation through SET_TARGET_PROPERTIES.
* add scikit-learn v0.18 compatibility
import KFold & StratifiedKFold from sklearn.model_selection instead of sklearn.cross_validation
* change DeprecationWarning to ImportError
DeprecationWarning isn't an exception, so it should work the other way around.
ml.dmlc.xgboost4j.scala.spark.XGBoost.scala:51
values is empty when we meet it at first time, so values(0) throw an IndexOutOfBoundsException.
It should be dVector.values(i) instead of values(i).