Skip to content

Commit 5acf773

Browse files
fix test case error (#3835)
1 parent dbd4162 commit 5acf773

15 files changed

+94
-43
lines changed

.github/workflows/build-test-linux-x86_64.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -177,6 +177,7 @@ jobs:
177177
cd tests/py
178178
cd dynamo
179179
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/
180+
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_llm.xml llm/
180181
popd
181182
182183
tests-py-dynamo-serde:

.github/workflows/build-test-linux-x86_64_rtx.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -141,6 +141,7 @@ jobs:
141141
cd tests/py
142142
cd dynamo
143143
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/
144+
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_llm.xml llm/
144145
popd
145146
146147
tests-py-dynamo-serde:

.github/workflows/build-test-windows.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -172,6 +172,7 @@ jobs:
172172
cd tests/py
173173
cd dynamo
174174
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/
175+
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_llm.xml llm/
175176
popd
176177
177178
tests-py-dynamo-serde:

.github/workflows/build-test-windows_rtx.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -143,6 +143,7 @@ jobs:
143143
cd tests/py
144144
cd dynamo
145145
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/
146+
python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_llm.xml llm/
146147
popd
147148
148149
tests-py-dynamo-serde:

.github/workflows/build_windows.yml

Lines changed: 22 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -288,6 +288,15 @@ jobs:
288288
BUILD_PARAMS: ${{ inputs.wheel-build-params }}
289289
run: |
290290
source "${BUILD_ENV_FILE}"
291+
# commented out due to still failed with the following error:
292+
# C:\actions-runner\_work\_temp\conda_environment_18042354682\lib\site-packages\torch\include\torch/csrc/utils/python_arg_parser.h(42): fatal error C1083: Cannot open include file: 'fmt/format.h': No such file or directory
293+
# workaround: download fmt and copy to torch include path in pre_build_script_windows.sh
294+
# conda update -n base -c defaults conda
295+
# conda install -c conda-forge fmt -y
296+
# conda list fmt
297+
# echo "path: $PATH"
298+
299+
${CONDA_RUN} python -m pip install fmt
291300
if [[ ${{ inputs.is-release-wheel }} == true || ${{ inputs.is-release-tarball }} == true ]]; then
292301
# release version for upload to pypi
293302
# BUILD_VERSION example: 2.4.0+cu121, we don't want the +cu121 part, so remove +cu121
@@ -344,20 +353,26 @@ jobs:
344353
SMOKE_TEST_SCRIPT: ${{ inputs.smoke-test-script }}
345354
run: |
346355
source "${BUILD_ENV_FILE}"
356+
set -x
347357
WHEEL_NAME=$(ls "${{ inputs.repository }}/dist/")
348358
echo "$WHEEL_NAME"
359+
nvidia-smi
360+
nvcc --version
349361
${CONDA_RUN} pip install "${{ inputs.repository }}/dist/$WHEEL_NAME"
362+
${CONDA_RUN} python -m pip list
363+
350364
if [[ $USE_TRT_RTX == true ]]; then
351365
# TODO: lan to remove this once we have a better way to do a smoke test
352366
echo "Smoke test for TensorRT-RTX is not skipped for now"
353367
else
354-
if [[ ! -f "${{ inputs.repository }}"/${SMOKE_TEST_SCRIPT} ]]; then
355-
echo "${{ inputs.repository }}/${SMOKE_TEST_SCRIPT} not found"
356-
${CONDA_RUN} "${{ inputs.repository }}/${ENV_SCRIPT}" python -c "import ${PACKAGE_NAME}; print('package version is ', ${PACKAGE_NAME}.__version__)"
357-
else
358-
echo "${{ inputs.repository }}/${SMOKE_TEST_SCRIPT} found"
359-
${CONDA_RUN} "${{ inputs.repository }}/${ENV_SCRIPT}" python "${{ inputs.repository }}/${SMOKE_TEST_SCRIPT}"
360-
fi
368+
echo "Skip smoke test in windows"
369+
# if [[ ! -f "${{ inputs.repository }}"/${SMOKE_TEST_SCRIPT} ]]; then
370+
# echo "${{ inputs.repository }}/${SMOKE_TEST_SCRIPT} not found"
371+
# ${CONDA_RUN} "${{ inputs.repository }}/${ENV_SCRIPT}" python -c "import ${PACKAGE_NAME}; print('package version is ', ${PACKAGE_NAME}.__version__)"
372+
# else
373+
# echo "${{ inputs.repository }}/${SMOKE_TEST_SCRIPT} found"
374+
# ${CONDA_RUN} "${{ inputs.repository }}/${ENV_SCRIPT}" python "${{ inputs.repository }}/${SMOKE_TEST_SCRIPT}"
375+
# fi
361376
fi
362377
- name: Smoke Test ARM64
363378
if: inputs.architecture == 'arm64'

MODULE.bazel

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -101,9 +101,9 @@ http_archive(
101101
http_archive(
102102
name = "tensorrt",
103103
build_file = "@//third_party/tensorrt/archive:BUILD",
104-
strip_prefix = "TensorRT-10.13.2.6",
104+
strip_prefix = "TensorRT-10.13.3.9",
105105
urls = [
106-
"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.2/tars/TensorRT-10.13.2.6.Linux.x86_64-gnu.cuda-12.9.tar.gz",
106+
"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/tars/TensorRT-10.13.3.9.Linux.x86_64-gnu.cuda-13.0.tar.gz",
107107
],
108108
)
109109

@@ -119,9 +119,9 @@ http_archive(
119119
http_archive(
120120
name = "tensorrt_sbsa",
121121
build_file = "@//third_party/tensorrt/archive:BUILD",
122-
strip_prefix = "TensorRT-10.13.2.6",
122+
strip_prefix = "TensorRT-10.13.3.9",
123123
urls = [
124-
"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.2/tars/TensorRT-10.13.2.6.Linux.aarch64-gnu.cuda-13.0.tar.gz",
124+
"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/tars/TensorRT-10.13.3.9.Linux.aarch64-gnu.cuda-13.0.tar.gz",
125125
],
126126
)
127127

@@ -137,9 +137,9 @@ http_archive(
137137
http_archive(
138138
name = "tensorrt_win",
139139
build_file = "@//third_party/tensorrt/archive:BUILD",
140-
strip_prefix = "TensorRT-10.13.2.6",
140+
strip_prefix = "TensorRT-10.13.3.9",
141141
urls = [
142-
"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.2/zip/TensorRT-10.13.2.6.Windows.win10.cuda-12.9.zip",
142+
"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.13.3/zip/TensorRT-10.13.3.9.Windows.win10.cuda-13.0.zip",
143143
],
144144
)
145145

dev_dep_versions.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
11
__cuda_version__: "12.8"
2-
__tensorrt_version__: "10.12.0"
2+
__tensorrt_version__: "10.13.3"
33
__tensorrt_rtx_version__: "1.0.0"

packaging/driver_upgrade.bat

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
1-
set WIN_DRIVER_VN=528.89
2-
set "DRIVER_DOWNLOAD_LINK=https://ossci-windows.s3.amazonaws.com/%WIN_DRIVER_VN%-data-center-tesla-desktop-winserver-2016-2019-2022-dch-international.exe"
3-
curl --retry 3 -kL %DRIVER_DOWNLOAD_LINK% --output %WIN_DRIVER_VN%-data-center-tesla-desktop-winserver-2016-2019-2022-dch-international.exe
1+
set WIN_DRIVER_VN=580.88
2+
set "DRIVER_DOWNLOAD_LINK=https://ossci-windows.s3.amazonaws.com/%WIN_DRIVER_VN%-data-center-tesla-desktop-win10-win11-64bit-dch-international.exe" & REM @lint-ignore
3+
curl --retry 3 -kL %DRIVER_DOWNLOAD_LINK% --output %WIN_DRIVER_VN%-data-center-tesla-desktop-win10-win11-64bit-dch-international.exe
44
if errorlevel 1 exit /b 1
55

6-
start /wait %WIN_DRIVER_VN%-data-center-tesla-desktop-winserver-2016-2019-2022-dch-international.exe -s -noreboot
6+
start /wait %WIN_DRIVER_VN%-data-center-tesla-desktop-win10-win11-64bit-dch-international.exe -s -noreboot
77
if errorlevel 1 exit /b 1
88

9-
del %WIN_DRIVER_VN%-data-center-tesla-desktop-winserver-2016-2019-2022-dch-international.exe || ver > NUL
9+
del %WIN_DRIVER_VN%-data-center-tesla-desktop-win10-win11-64bit-dch-international.exe || ver > NUL

packaging/pre_build_script.sh

Lines changed: 7 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -59,15 +59,13 @@ fi
5959
export TORCH_BUILD_NUMBER=$(python -c "import torch, urllib.parse as ul; print(ul.quote_plus(torch.__version__))")
6060
export TORCH_INSTALL_PATH=$(python -c "import torch, os; print(os.path.dirname(torch.__file__))")
6161

62-
if [[ ${TENSORRT_VERSION} != "" ]]; then
63-
# Replace dependencies in the original pyproject.toml with the current TensorRT version. It is used for CI tests of different TensorRT versions.
64-
# For example, if the current testing TensorRT version is 10.7.0, but the pyproject.toml tensorrt>=10.8.0,<10.9.0, then the following sed command
65-
# will replace tensorrt>=10.8.0,<10.9.0 with tensorrt==10.7.0
66-
sed -i -e "s/tensorrt>=.*,<.*\"/tensorrt>=${TENSORRT_VERSION},<$(echo "${TENSORRT_VERSION}" | awk -F. '{print $1"."$2+1".0"}')\"/g" \
67-
-e "s/tensorrt-cu12>=.*,<.*\"/tensorrt-cu12>=${TENSORRT_VERSION},<$(echo "${TENSORRT_VERSION}" | awk -F. '{print $1"."$2+1".0"}')\"/g" \
68-
-e "s/tensorrt-cu12-bindings>=.*,<.*\"/tensorrt-cu12-bindings>=${TENSORRT_VERSION},<$(echo "${TENSORRT_VERSION}" | awk -F. '{print $1"."$2+1".0"}')\"/g" \
69-
-e "s/tensorrt-cu12-libs>=.*,<.*\"/tensorrt-cu12-libs>=${TENSORRT_VERSION},<$(echo "${TENSORRT_VERSION}" | awk -F. '{print $1"."$2+1".0"}')\"/g" \
70-
pyproject.toml
62+
# CU_UPPERBOUND eg:13.0 or 12.9
63+
# tensorrt tar for linux and windows are different across cuda version
64+
# for sbsa it is the same tar across cuda version
65+
if [[ ${CU_VERSION:2:2} == "13" ]]; then
66+
export CU_UPPERBOUND="13.0"
67+
else
68+
export CU_UPPERBOUND="12.9"
7169
fi
7270

7371
cat toolchains/ci_workspaces/MODULE.bazel.tmpl | envsubst > MODULE.bazel

packaging/pre_build_script_windows.sh

Lines changed: 17 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
set -x
22

3-
pip install -U numpy packaging pyyaml setuptools wheel
3+
pip install -U numpy packaging pyyaml setuptools wheel fmt
44

55
choco install bazelisk -y
66

@@ -27,6 +27,22 @@ pip install --force-reinstall --pre ${TORCH} --index-url ${INDEX_URL}
2727
export CUDA_HOME="$(echo ${CUDA_PATH} | sed -e 's#\\#\/#g')"
2828
export TORCH_INSTALL_PATH="$(python -c "import torch, os; print(os.path.dirname(torch.__file__))" | sed -e 's#\\#\/#g')"
2929

30+
# tried with conda install -c conda-forge fmt -y, but build still failed in windows with the following error:
31+
# C:\actions-runner\_work\_temp\conda_environment_18042354682\lib\site-packages\torch\include\torch/csrc/utils/python_arg_parser.h(42): fatal error C1083: Cannot open include file: 'fmt/format.h': No such file or directory
32+
# workaround: download fmt from github and copy to torch include path
33+
curl -L -o fmt.zip https://github.com/fmtlib/fmt/releases/download/12.0.0/fmt-12.0.0.zip
34+
unzip fmt.zip
35+
cp -r fmt-12.0.0/include/fmt/ $TORCH_INSTALL_PATH/include/
36+
ls -lart $TORCH_INSTALL_PATH/include/fmt/
37+
38+
# CU_UPPERBOUND eg:13.0 or 12.9
39+
# tensorrt tar for linux and windows are different across cuda version
40+
# for sbsa it is the same tar across cuda version
41+
if [[ ${CU_VERSION:2:2} == "13" ]]; then
42+
export CU_UPPERBOUND="13.0"
43+
else
44+
export CU_UPPERBOUND="12.9"
45+
fi
3046
cat toolchains/ci_workspaces/MODULE.bazel.tmpl | envsubst > MODULE.bazel
3147

3248
if [[ ${TENSORRT_VERSION} != "" ]]; then

0 commit comments

Comments
 (0)