|
78 | 78 | "ort_openvino_version": "2025.3.0", |
79 | 79 | "standalone_openvino_version": "2025.3.0", |
80 | 80 | "dcgm_version": "4.4.0-1", |
81 | | - "vllm_version": "0.10.1.1", |
| 81 | + "vllm_version": "0.10.2", |
82 | 82 | "rhel_py_version": "3.12.3", |
83 | 83 | } |
84 | 84 |
|
@@ -660,6 +660,10 @@ def pytorch_cmake_args(images): |
660 | 660 | cargs.append( |
661 | 661 | cmake_backend_enable("pytorch", "TRITON_ENABLE_NVTX", FLAGS.enable_nvtx) |
662 | 662 | ) |
| 663 | + if target_platform() == "igpu": |
| 664 | + cargs.append( |
| 665 | + cmake_backend_enable("pytorch", "TRITON_PYTORCH_NVSHMEM", False) |
| 666 | + ) |
663 | 667 | return cargs |
664 | 668 |
|
665 | 669 |
|
@@ -1514,6 +1518,22 @@ def dockerfile_prepare_container_linux(argmap, backends, enable_gpu, target_mach |
1514 | 1518 | ENV PYTHONPATH=/opt/tritonserver/backends/dali/wheel/dali:$PYTHONPATH |
1515 | 1519 | """ |
1516 | 1520 |
|
| 1521 | + if target_platform() not in ["igpu", "windows", "rhel"]: |
| 1522 | + repo_arch = "sbsa" if target_machine == "aarch64" else "x86_64" |
| 1523 | + df += f""" |
| 1524 | +RUN curl -o /tmp/cuda-keyring.deb \\ |
| 1525 | + https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2404/{repo_arch}/cuda-keyring_1.1-1_all.deb \\ |
| 1526 | + && apt install /tmp/cuda-keyring.deb \\ |
| 1527 | + && rm /tmp/cuda-keyring.deb \\ |
| 1528 | + && apt update -qq \\ |
| 1529 | + && apt install --yes --no-install-recommends libnvshmem3-cuda-13 \\ |
| 1530 | + && rm -rf /var/lib/apt/lists/* \\ |
| 1531 | + && dpkg -L libnvshmem3-cuda-13 | grep libnvshmem_host.so | sed -e 's/libnvshmem_host.*//g' | sort -u > /etc/ld.so.conf.d/libnvshmem3-cuda-13.conf \\ |
| 1532 | + && ldconfig |
| 1533 | +""".format( |
| 1534 | + repo_arch=repo_arch |
| 1535 | + ) |
| 1536 | + |
1517 | 1537 | df += """ |
1518 | 1538 | WORKDIR /opt/tritonserver |
1519 | 1539 | RUN rm -fr /opt/tritonserver/* |
|
0 commit comments