Skip to content

Commit f742da3

Browse files
authored
Merge pull request #2827 from ROCm/develop_IFU_20251124
[AUTOGENERATED] develop_IFU_20251124
2 parents 5ca076d + ecdea86 commit f742da3

File tree

495 files changed

+16784
-5017
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

495 files changed

+16784
-5017
lines changed

.ci/docker/build.sh

Lines changed: 2 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -125,10 +125,10 @@ case "$tag" in
125125
UCC_COMMIT=${_UCC_COMMIT}
126126
TRITON=yes
127127
;;
128-
pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc9-inductor-benchmarks)
128+
pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-inductor-benchmarks)
129129
CUDA_VERSION=12.8.1
130130
ANACONDA_PYTHON_VERSION=3.10
131-
GCC_VERSION=9
131+
GCC_VERSION=11
132132
VISION=yes
133133
KATEX=yes
134134
UCX_COMMIT=${_UCX_COMMIT}
@@ -146,16 +146,6 @@ case "$tag" in
146146
UCC_COMMIT=${_UCC_COMMIT}
147147
TRITON=yes
148148
;;
149-
pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc9)
150-
CUDA_VERSION=12.8.1
151-
ANACONDA_PYTHON_VERSION=3.10
152-
GCC_VERSION=9
153-
VISION=yes
154-
KATEX=yes
155-
UCX_COMMIT=${_UCX_COMMIT}
156-
UCC_COMMIT=${_UCC_COMMIT}
157-
TRITON=yes
158-
;;
159149
pytorch-linux-jammy-py3-clang12-onnx)
160150
ANACONDA_PYTHON_VERSION=3.10
161151
CLANG_VERSION=12
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
1b0418a9a454b2b93ab8d71f40e59d2297157fae
1+
aa01f5c2cd4db2b7bfa53ea98a1a8dfbd6d77c92

.ci/docker/common/install_xpu.sh

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -64,14 +64,13 @@ function install_ubuntu() {
6464

6565
function install_rhel() {
6666
. /etc/os-release
67-
if [[ "${ID}" == "rhel" ]]; then
68-
if [[ ! " 8.8 8.10 9.0 9.2 9.3 " =~ " ${VERSION_ID} " ]]; then
69-
echo "RHEL version ${VERSION_ID} not supported"
70-
exit
71-
fi
72-
elif [[ "${ID}" == "almalinux" ]]; then
73-
# Workaround for almalinux8 which used by quay.io/pypa/manylinux_2_28_x86_64
74-
VERSION_ID="8.8"
67+
if [[ ! " 8.8 8.10 9.0 9.2 9.3 " =~ " ${VERSION_ID} " ]]; then
68+
echo "RHEL version ${VERSION_ID} not supported"
69+
exit
70+
fi
71+
# Using testing channel for CD build
72+
if [[ "${ID}" == "almalinux" ]]; then
73+
XPU_DRIVER_VERSION="/testing"
7574
fi
7675

7776
dnf install -y 'dnf-command(config-manager)'

.ci/docker/requirements-ci.txt

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -397,3 +397,6 @@ scikit-build==0.18.1
397397
pyre-extensions==0.0.32
398398
tabulate==0.9.0
399399
#Description: These package are needed to build FBGEMM and torchrec on PyTorch CI
400+
401+
Jinja2==3.1.6
402+
#Description: required for torch.distributed.debug

.ci/lumen_cli/cli/lib/core/vllm/vllm_test.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,6 @@ def __init__(self, args: Any):
8484
self.VLLM_TEST_WHLS_REGEX = [
8585
"xformers/*.whl",
8686
"vllm/vllm*.whl",
87-
"flashinfer-python/flashinfer*.whl",
8887
]
8988

9089
def prepare(self):

.ci/pytorch/test.sh

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1763,12 +1763,14 @@ test_operator_microbenchmark() {
17631763
mkdir -p "$TEST_REPORTS_DIR"
17641764
TEST_DIR=$(pwd)
17651765

1766+
test_inductor_set_cpu_affinity
1767+
17661768
cd benchmarks/operator_benchmark/pt_extension
1767-
python -m pip install .
1769+
python -m pip install . -v --no-build-isolation
17681770

17691771
cd "${TEST_DIR}"/benchmarks/operator_benchmark
17701772

1771-
for OP_BENCHMARK_TESTS in matmul mm addmm bmm conv; do
1773+
for OP_BENCHMARK_TESTS in matmul mm addmm bmm conv optimizer; do
17721774
$TASKSET python -m pt.${OP_BENCHMARK_TESTS}_test --tag-filter long \
17731775
--output-json-for-dashboard "${TEST_REPORTS_DIR}/operator_microbenchmark_${OP_BENCHMARK_TESTS}_compile.json" \
17741776
--benchmark-name "PyTorch operator microbenchmark" --use-compile

.circleci/scripts/binary_linux_test.sh

Lines changed: 12 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -31,23 +31,6 @@ if [[ "$PACKAGE_TYPE" != libtorch ]]; then
3131
export PATH="\${python_path}/bin:\$PATH"
3232
fi
3333
34-
EXTRA_CONDA_FLAGS=""
35-
NUMPY_PIN=""
36-
PROTOBUF_PACKAGE="defaults::protobuf"
37-
38-
if [[ "\$python_nodot" = *310* ]]; then
39-
# There's an issue with conda channel priority where it'll randomly pick 1.19 over 1.20
40-
# we set a lower boundary here just to be safe
41-
NUMPY_PIN=">=1.21.2"
42-
PROTOBUF_PACKAGE="protobuf>=3.19.0"
43-
fi
44-
45-
if [[ "\$python_nodot" = *39* ]]; then
46-
# There's an issue with conda channel priority where it'll randomly pick 1.19 over 1.20
47-
# we set a lower boundary here just to be safe
48-
NUMPY_PIN=">=1.20"
49-
fi
50-
5134
# Move debug wheels out of the package dir so they don't get installed
5235
mkdir -p /tmp/debug_final_pkgs
5336
mv /final_pkgs/debug-*.zip /tmp/debug_final_pkgs || echo "no debug packages to move"
@@ -66,12 +49,23 @@ fi
6649
if [[ "$PACKAGE_TYPE" != libtorch ]]; then
6750
if [[ "\$BUILD_ENVIRONMENT" != *s390x* ]]; then
6851
pip install "\$pkg" --index-url "https://download.pytorch.org/whl/\${CHANNEL}/${DESIRED_CUDA}"
69-
retry pip install -q numpy protobuf typing-extensions
52+
53+
# numpy tests:
54+
# We test 1 version no numpy. 1 version with numpy 1.x and rest with numpy 2.x
55+
if [[ "\$python_nodot" = *311* ]]; then
56+
retry pip install -q numpy==1.23.5 protobuf typing-extensions
57+
elif [[ "\$python_nodot" = *312* ]]; then
58+
retry pip install -q protobuf typing-extensions
59+
else
60+
retry pip install -q numpy protobuf typing-extensions
61+
fi
62+
7063
else
7164
pip install "\$pkg"
7265
retry pip install -q numpy protobuf typing-extensions
7366
fi
7467
fi
68+
7569
if [[ "$PACKAGE_TYPE" == libtorch ]]; then
7670
pkg="\$(ls /final_pkgs/*-latest.zip)"
7771
unzip "\$pkg" -d /tmp

.github/ci_commit_pins/audio.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
ee1a1350eb37804b94334768f328144f058f14e9
1+
32ce8c011855adb15438ddc9bf6c139d23f8cee5

.github/ci_commit_pins/vision.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2d82dc5caa336d179d9b46ac4a0fb8c43d84c5cc
1+
617079d944b0e72632311c30ae2bbdf1168b901e

.github/ci_configs/vllm/Dockerfile

Lines changed: 5 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
ARG CUDA_VERSION=12.8.1
1+
ARG CUDA_VERSION=12.9.1
22
ARG PYTHON_VERSION=3.12
33

44
# BUILD_BASE_IMAGE: used to setup python build xformers, and vllm wheels, It can be replaced with a different base image from local machine,
@@ -124,7 +124,7 @@ RUN --mount=type=cache,target=/root/.cache/uv bash - <<'BASH'
124124
git clone https://github.com/facebookresearch/xformers.git
125125

126126
pushd xformers
127-
git checkout v0.0.32.post2
127+
git checkout v0.0.33.post1
128128
git submodule update --init --recursive
129129
python3 setup.py bdist_wheel --dist-dir=../xformers-dist --verbose
130130
popd
@@ -256,7 +256,7 @@ ENV UV_INDEX_STRATEGY="unsafe-best-match"
256256
# Use copy mode to avoid hardlink failures with Docker cache mounts
257257
ENV UV_LINK_MODE=copy
258258

259-
# Install build and runtime dependencies, this is needed for flashinfer install
259+
# Install build and runtime dependencies
260260
COPY requirements/build.txt requirements/build.txt
261261
COPY use_existing_torch.py use_existing_torch.py
262262
RUN python3 use_existing_torch.py
@@ -294,33 +294,9 @@ RUN --mount=type=cache,target=/root/.cache/uv \
294294
RUN --mount=type=cache,target=/root/.cache/uv \
295295
uv pip install --system /wheels/xformers/*.whl --verbose
296296

297-
# Build FlashInfer from source
298-
ARG torch_cuda_arch_list='8.0;8.9;9.0a;10.0a;12.0'
299-
ENV TORCH_CUDA_ARCH_LIST=${torch_cuda_arch_list}
300-
301-
# TODO(elainewy): remove this once vllm commit is updated, and install flashinfer from pip
302-
# see https://github.com/pytorch/pytorch/pull/165274#issuecomment-3408531784
303-
ARG FLASHINFER_GIT_REPO="https://github.com/flashinfer-ai/flashinfer.git"
304-
ARG FLASHINFER_GIT_REF="v0.2.14.post1"
305-
306-
RUN --mount=type=cache,target=/root/.cache/uv \
307-
git clone --depth 1 --recursive --shallow-submodules \
308-
--branch ${FLASHINFER_GIT_REF} \
309-
${FLASHINFER_GIT_REPO} flashinfer \
310-
&& echo "Building FlashInfer with AOT for arches: ${torch_cuda_arch_list}" \
311-
&& cd flashinfer \
312-
&& python3 -m flashinfer.aot \
313-
&& python3 -m build --no-isolation --wheel --outdir ../wheels/flashinfer \
314-
&& cd .. \
315-
&& rm -rf flashinfer
316-
317-
# Install FlashInfer
318-
RUN --mount=type=cache,target=/root/.cache/uv \
319-
uv pip install --system wheels/flashinfer/*.whl --verbose
320-
321297
# Logging to confirm the torch versions
322-
RUN pip freeze | grep -E 'torch|xformers|vllm|flashinfer'
323-
RUN uv pip freeze | grep -i '^torch\|^torchvision\|^torchaudio\|^xformers\|^vllm\|^flashinfer' > build_summary.txt
298+
RUN pip freeze | grep -E 'torch|xformers|vllm'
299+
RUN uv pip freeze | grep -i '^torch\|^torchvision\|^torchaudio\|^xformers\|^vllm' > build_summary.txt
324300
################### VLLM INSTALLED IMAGE ####################
325301

326302

@@ -331,4 +307,3 @@ FROM scratch as export-wheels
331307
COPY --from=base /workspace/xformers-dist /wheels/xformers
332308
COPY --from=build /workspace/vllm-dist /wheels/vllm
333309
COPY --from=vllm-base /workspace/build_summary.txt /wheels/build_summary.txt
334-
COPY --from=vllm-base /workspace/wheels/flashinfer /wheels/flashinfer-python

0 commit comments

Comments
 (0)