You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
(os.environ.get("VLLM_ALLOW_RUNTIME_LORA_UPDATING", "0").strip().lower() in
669
669
("1", "true")),
670
670
671
-
# By default, vLLM will check the peer-to-peer capability itself,
672
-
# in case of broken drivers. See https://github.com/vllm-project/vllm/blob/a9b15c606fea67a072416ea0ea115261a2756058/vllm/distributed/device_communicators/custom_all_reduce_utils.py#L101-L108 for details. # noqa
673
-
# If this env var is set to 1, vLLM will skip the peer-to-peer check,
674
-
# and trust the driver's peer-to-peer capability report.
671
+
# We assume drivers can report p2p status correctly.
672
+
# If the program hangs when using custom allreduce,
673
+
# potantially caused by a bug in the driver (535 series),
674
+
# if might be helpful to set VLLM_SKIP_P2P_CHECK=0
675
+
# so that vLLM can verify if p2p is actually working.
676
+
# See https://github.com/vllm-project/vllm/blob/a9b15c606fea67a072416ea0ea115261a2756058/vllm/distributed/device_communicators/custom_all_reduce_utils.py#L101-L108 for details. # noqa
0 commit comments