@@ -215,7 +215,7 @@ jobs:
215215 set -eux
216216 PYTHONPATH="${PWD}" python .ci/scripts/gather_test_models.py --event "pull_request" --backend "gpu"
217217 test-gpu-compile :
218- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
218+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
219219 name : test-gpu-compile (${{ matrix.platform }}, ${{ matrix.model_name }})
220220 needs : gather-models-gpu
221221 strategy :
@@ -224,7 +224,7 @@ jobs:
224224 with :
225225 runner : linux.g5.4xlarge.nvidia.gpu
226226 gpu-arch-type : cuda
227- gpu-arch-version : " 12.1 "
227+ gpu-arch-version : " 12.4 "
228228 script : |
229229 echo "::group::Print machine info"
230230 nvidia-smi
@@ -250,7 +250,7 @@ jobs:
250250 echo "::endgroup::"
251251
252252 test-gpu-aoti-bfloat16 :
253- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
253+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
254254 name : test-gpu-aoti-bfloat16 (${{ matrix.platform }}, ${{ matrix.model_name }})
255255 needs : gather-models-gpu
256256 strategy :
@@ -259,18 +259,13 @@ jobs:
259259 with :
260260 runner : linux.g5.4xlarge.nvidia.gpu
261261 gpu-arch-type : cuda
262- gpu-arch-version : " 12.1 "
262+ gpu-arch-version : " 12.4 "
263263 timeout : 60
264264 script : |
265265 echo "::group::Print machine info"
266266 nvidia-smi
267267 echo "::endgroup::"
268268
269- echo "::group::Install newer objcopy that supports --set-section-alignment"
270- yum install -y devtoolset-10-binutils
271- export PATH=/opt/rh/devtoolset-10/root/usr/bin/:$PATH
272- echo "::endgroup::"
273-
274269 echo "::group::Install required packages"
275270 ./install/install_requirements.sh cuda
276271 pip3 list
@@ -291,7 +286,7 @@ jobs:
291286 echo "::endgroup::"
292287
293288 test-gpu-aoti-float32 :
294- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
289+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
295290 name : test-gpu-aoti-float32 (${{ matrix.platform }}, ${{ matrix.model_name }})
296291 needs : gather-models-gpu
297292 strategy :
@@ -300,17 +295,12 @@ jobs:
300295 with :
301296 runner : linux.g5.4xlarge.nvidia.gpu
302297 gpu-arch-type : cuda
303- gpu-arch-version : " 12.1 "
298+ gpu-arch-version : " 12.4 "
304299 script : |
305300 echo "::group::Print machine info"
306301 nvidia-smi
307302 echo "::endgroup::"
308303
309- echo "::group::Install newer objcopy that supports --set-section-alignment"
310- yum install -y devtoolset-10-binutils
311- export PATH=/opt/rh/devtoolset-10/root/usr/bin/:$PATH
312- echo "::endgroup::"
313-
314304 echo "::group::Install required packages"
315305 ./install/install_requirements.sh cuda
316306 pip list
@@ -337,7 +327,7 @@ jobs:
337327 echo "::endgroup::"
338328
339329 test-gpu-aoti-float16 :
340- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
330+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
341331 name : test-gpu-aoti-float16 (${{ matrix.platform }}, ${{ matrix.model_name }})
342332 needs : gather-models-gpu
343333 strategy :
@@ -346,17 +336,12 @@ jobs:
346336 with :
347337 runner : linux.g5.4xlarge.nvidia.gpu
348338 gpu-arch-type : cuda
349- gpu-arch-version : " 12.1 "
339+ gpu-arch-version : " 12.4 "
350340 script : |
351341 echo "::group::Print machine info"
352342 nvidia-smi
353343 echo "::endgroup::"
354344
355- echo "::group::Install newer objcopy that supports --set-section-alignment"
356- yum install -y devtoolset-10-binutils
357- export PATH=/opt/rh/devtoolset-10/root/usr/bin/:$PATH
358- echo "::endgroup::"
359-
360345 echo "::group::Install required packages"
361346 ./install/install_requirements.sh cuda
362347 pip list
@@ -384,7 +369,7 @@ jobs:
384369 echo "::endgroup::"
385370
386371 test-gpu-eval-sanity-check :
387- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
372+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
388373 name : test-gpu-eval-sanity-check (${{ matrix.platform }}, ${{ matrix.model_name }})
389374 needs : gather-models-gpu
390375 strategy :
@@ -393,17 +378,12 @@ jobs:
393378 with :
394379 runner : linux.g5.4xlarge.nvidia.gpu
395380 gpu-arch-type : cuda
396- gpu-arch-version : " 12.1 "
381+ gpu-arch-version : " 12.4 "
397382 script : |
398383 echo "::group::Print machine info"
399384 nvidia-smi
400385 echo "::endgroup::"
401386
402- echo "::group::Install newer objcopy that supports --set-section-alignment"
403- yum install -y devtoolset-10-binutils
404- export PATH=/opt/rh/devtoolset-10/root/usr/bin/:$PATH
405- echo "::endgroup::"
406-
407387 echo "::group::Install required packages"
408388 ./install/install_requirements.sh cuda
409389 pip3 list
@@ -1031,7 +1011,7 @@ jobs:
10311011 echo "Tests complete."
10321012
10331013 test-build-runner-et-android :
1034- uses : pytorch/test-infra/.github/workflows/linux_job .yml@main
1014+ uses : pytorch/test-infra/.github/workflows/linux_job_v2 .yml@main
10351015 with :
10361016 runner : linux.4xlarge
10371017 script : |
@@ -1124,3 +1104,41 @@ jobs:
11241104 echo "Generate AOTI"
11251105 python torchchat.py generate stories110M --aoti-package-path ./model.pt2 --prompt "${PRMT}"
11261106 echo "Tests complete."
1107+
1108+ test-torchao-experimental-mps :
1109+ strategy :
1110+ matrix :
1111+ runner : [macos-m1-stable]
1112+ runs-on : ${{matrix.runner}}
1113+ steps :
1114+ - name : Checkout repo
1115+ uses : actions/checkout@v3
1116+ with :
1117+ submodules : true
1118+ - name : Setup Python
1119+ uses : actions/setup-python@v2
1120+ with :
1121+ python-version : 3.10.11
1122+ - name : Print machine info
1123+ run : |
1124+ uname -a
1125+ if [ $(uname -s) == Darwin ]; then
1126+ sysctl machdep.cpu.brand_string
1127+ sysctl machdep.cpu.core_count
1128+ fi
1129+ - name : Install torchchat
1130+ run : |
1131+ echo "Intalling pip3 packages"
1132+ ./install/install_requirements.sh
1133+ pip3 list
1134+ python3 -c 'import torch;print(f"torch: {torch.__version__, torch.version.git_version}")'
1135+ - name : Install torchao-ops-mps
1136+ id : install-torchao-ops-mps
1137+ run : |
1138+ bash torchchat/utils/scripts/build_torchao_ops.sh mps
1139+ - name : Run inference
1140+ run : |
1141+ python torchchat.py download stories110M
1142+ export PRMT="Once upon a time in a land far away"
1143+ echo "Generate eager"
1144+ python torchchat.py generate stories110M --temperature 0 --prompt "${PRMT}" --device mps --dtype float32 --quantize '{"linear:afpwx": {"bitwidth": 3, "groupsize": 32}}'
0 commit comments