Skip to content

Commit fd365c1

Browse files
authored
[doc] Brief note about RMM SAM allocator. [skip ci] (dmlc#10712)
1 parent ec3f327 commit fd365c1

File tree

2 files changed

+22
-1
lines changed

2 files changed

+22
-1
lines changed

demo/rmm_plugin/README.rst

Lines changed: 17 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,4 +58,20 @@ Since with RMM the memory pool is pre-allocated on a specific device, changing t
5858
device ordinal in XGBoost can result in memory error ``cudaErrorIllegalAddress``. Use the
5959
``CUDA_VISIBLE_DEVICES`` environment variable instead of the ``device="cuda:1"`` parameter
6060
for selecting device. For distributed training, the distributed computing frameworks like
61-
``dask-cuda`` are responsible for device management.
61+
``dask-cuda`` are responsible for device management.
62+
63+
************************
64+
Memory Over-Subscription
65+
************************
66+
67+
.. warning::
68+
69+
This feature is still experimental and is under active development.
70+
71+
The newer NVIDIA platforms like `Grace-Hopper
72+
<https://www.nvidia.com/en-us/data-center/grace-hopper-superchip/>`__ use `NVLink-C2C
73+
<https://www.nvidia.com/en-us/data-center/nvlink-c2c/>`__, which allows the CPU and GPU to
74+
have a coherent memory model. Users can use the `SamHeadroomMemoryResource` in the latest
75+
RMM to utilize system memory for storing data. This can help XGBoost utilize memory from
76+
the host for GPU computation, but it may reduce performance due to slower CPU memory speed
77+
and page migration overhead.

doc/gpu/index.rst

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -50,6 +50,11 @@ Multi-node Multi-GPU Training
5050

5151
XGBoost supports fully distributed GPU training using `Dask <https://dask.org/>`_, ``Spark`` and ``PySpark``. For getting started with Dask see our tutorial :doc:`/tutorials/dask` and worked examples :doc:`/python/dask-examples/index`, also Python documentation :ref:`dask_api` for complete reference. For usage with ``Spark`` using Scala see :doc:`/jvm/xgboost4j_spark_gpu_tutorial`. Lastly for distributed GPU training with ``PySpark``, see :doc:`/tutorials/spark_estimator`.
5252

53+
RMM integration
54+
===============
55+
56+
XGBoost provides optional support for RMM integration. See :doc:`/python/rmm-examples/index` for more info.
57+
5358

5459
Memory usage
5560
============

0 commit comments

Comments
 (0)