* Fix loading DMatrix binary in distributed env. (#8149)
- Try to load DMatrix binary before trying to parse text input.
- Remove some unmaintained code.
* Fix.
- Optionally switch to c++17
- Use rmm CMake target.
- Workaround compiler errors.
- Fix GPUMetric inheritance.
- Run death tests even if it's built with RMM support.
Co-authored-by: jakirkham <jakirkham@gmail.com>
Co-authored-by: jakirkham <jakirkham@gmail.com>
* [jvm-packages] move the dmatrix building into rabit context (#7823)
This fixes the QuantileDeviceDMatrix in distributed environment.
* [doc] update the jvm tutorial to 1.6.1 [skip ci] (#7834)
* [Breaking][jvm-packages] Use barrier execution mode (#7836)
With the introduction of the barrier execution mode. we don't need to kill SparkContext when some xgboost tasks failed. Instead, Spark will handle the errors for us. So in this PR, `killSparkContextOnWorkerFailure` parameter is deleted.
* [doc] remove the doc about killing SparkContext [skip ci] (#7840)
* [jvm-package] remove the coalesce in barrier mode (#7846)
* [jvm-packages] Fix model compatibility (#7845)
* Ignore all Java exceptions when looking for Linux musl support (#7844)
Co-authored-by: Bobby Wang <wbo4958@gmail.com>
Co-authored-by: Michael Allman <msa@allman.ms>
* [jvm-packages] add hostIp and python exec for rabit tracker (#7808)
* Fix training continuation with categorical model. (#7810)
* Make sure the task is initialized before construction of tree updater.
This is a quick fix meant to be backported to 1.6, for a full fix we should pass the model
param into tree updater by reference instead.
Co-authored-by: Bobby Wang <wbo4958@gmail.com>
* [jvm-packages] unify setFeaturesCol API for XGBoostRegressor (#7784)
* [jvm-packages] add doc for xgboost4j-spark-gpu (#7779)
Co-authored-by: Jiaming Yuan <jm.yuan@outlook.com>
* [jvm-packages] remove the dep of com.fasterxml.jackson (#7791)
* [jvm-packages] xgboost4j-spark should work when featuresCols is specified (#7789)
Co-authored-by: Bobby Wang <wbo4958@gmail.com>
* Drop support for deprecated CUDA architecture.
* Check file size at release branch.
* Use 200 MB limit
Co-authored-by: Philip Hyunsu Cho <chohyu01@cs.washington.edu>
* Skip non-increasing test with external memory when subsample is used.
* Increase bin numbers for boost from prediction test. This mitigates the effect of
non-deterministic partitioning.
* Use the name `Context`.
* Pass a context object into `SetInfo`.
* Add context to proxy matrix.
* Add context to iterative DMatrix.
This is to remove the use of the default number of threads during `SetInfo` as a follow-up on
removing the global omp variable while preparing for CUDA stream semantic. Currently, XGBoost
uses the legacy CUDA stream, we will gradually remove them in the future in favor of non-blocking streams.
* Generate column matrix from gHistIndex.
* Avoid synchronization with the sparse page once the cache is written.
* Cleanups: Remove member variables/functions, change the update routine to look like approx and gpu_hist.
* Remove pruner.