[doc] Add notes about RMM and device ordinal. [skip ci] (#10562)
- Remove the experimental tag, we have been running it for a long time now. - Add notes about avoiding set CUDA device. - Add link in parameter.
This commit is contained in:
parent
3ec74a1ba9
commit
8e2b874b4c
@ -1,5 +1,5 @@
|
|||||||
Using XGBoost with RAPIDS Memory Manager (RMM) plugin (EXPERIMENTAL)
|
Using XGBoost with RAPIDS Memory Manager (RMM) plugin
|
||||||
====================================================================
|
=====================================================
|
||||||
|
|
||||||
`RAPIDS Memory Manager (RMM) <https://github.com/rapidsai/rmm>`__ library provides a
|
`RAPIDS Memory Manager (RMM) <https://github.com/rapidsai/rmm>`__ library provides a
|
||||||
collection of efficient memory allocators for NVIDIA GPUs. It is now possible to use
|
collection of efficient memory allocators for NVIDIA GPUs. It is now possible to use
|
||||||
@ -47,5 +47,15 @@ the global configuration ``use_rmm``:
|
|||||||
with xgb.config_context(use_rmm=True):
|
with xgb.config_context(use_rmm=True):
|
||||||
clf = xgb.XGBClassifier(tree_method="hist", device="cuda")
|
clf = xgb.XGBClassifier(tree_method="hist", device="cuda")
|
||||||
|
|
||||||
Depending on the choice of memory pool size or type of allocator, this may have negative
|
Depending on the choice of memory pool size and the type of the allocator, this can add
|
||||||
performance impact.
|
more consistency to memory usage but with slightly degraded performance impact.
|
||||||
|
|
||||||
|
*******************************
|
||||||
|
No Device Ordinal for Multi-GPU
|
||||||
|
*******************************
|
||||||
|
|
||||||
|
Since with RMM the memory pool is pre-allocated on a specific device, changing the CUDA
|
||||||
|
device ordinal in XGBoost can result in memory error ``cudaErrorIllegalAddress``. Use the
|
||||||
|
``CUDA_VISIBLE_DEVICES`` environment variable instead of the ``device="cuda:1"`` parameter
|
||||||
|
for selecting device. For distributed training, the distributed computing frameworks like
|
||||||
|
``dask-cuda`` are responsible for device management.
|
||||||
@ -25,7 +25,11 @@ Global Configuration
|
|||||||
The following parameters can be set in the global scope, using :py:func:`xgboost.config_context()` (Python) or ``xgb.set.config()`` (R).
|
The following parameters can be set in the global scope, using :py:func:`xgboost.config_context()` (Python) or ``xgb.set.config()`` (R).
|
||||||
|
|
||||||
* ``verbosity``: Verbosity of printing messages. Valid values of 0 (silent), 1 (warning), 2 (info), and 3 (debug).
|
* ``verbosity``: Verbosity of printing messages. Valid values of 0 (silent), 1 (warning), 2 (info), and 3 (debug).
|
||||||
* ``use_rmm``: Whether to use RAPIDS Memory Manager (RMM) to allocate GPU memory. This option is only applicable when XGBoost is built (compiled) with the RMM plugin enabled. Valid values are ``true`` and ``false``.
|
|
||||||
|
* ``use_rmm``: Whether to use RAPIDS Memory Manager (RMM) to allocate cache GPU
|
||||||
|
memory. The primary memory is always allocated on the RMM pool when XGBoost is built
|
||||||
|
(compiled) with the RMM plugin enabled. Valid values are ``true`` and ``false``. See
|
||||||
|
:doc:`/python/rmm-examples/index` for details.
|
||||||
|
|
||||||
******************
|
******************
|
||||||
General Parameters
|
General Parameters
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user