Skip to content
This repository was archived by the owner on Sep 10, 2025. It is now read-only.
Merged
25 changes: 15 additions & 10 deletions .github/workflows/run-readme-pr-linuxaarch64.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,9 +11,10 @@ jobs:
test-readme-cpu:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
with:
runner: linux-aarch64
runner: linux.arm64.m7g.4xlarge
docker-image: "pytorch/manylinuxaarch64-builder:cuda12.6-main"
gpu-arch-type: cuda
gpu-arch-version: "12.1"
gpu-arch-version: "12.6"
timeout: 60
Copy link

@atalman atalman Jan 23, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

try passing

docker-image: "pytorch/manylinuxaarch64-builder:cuda12.1-main"

Looks like the error:
WARNING: The requested image's platform (linux/amd64) does not match the detected host platform (linux/arm64/v8) and no specific platform was requested

Is related to the fact that this is using docker-image=pytorch/conda-builder:cuda12.1 image by default which is not correct for linux.arm64.m7g.4xlarge runner

Copy link
Contributor Author

@Jack-Khuu Jack-Khuu Jan 23, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Doesn't look like it can find the Docker-image verbatim, testing with the 12.6 version found in pt/pt

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If using linux_job_v2.yml you can try using latest image pytorch/manylinux2_28_aarch64-builder:cuda12.6

Copy link
Contributor Author

@Jack-Khuu Jack-Khuu Jan 23, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Doesn't look like the cuda version is there manylinux2_28_aarch64-builder:cuda12.6, but the CPU variant :cpu-aarch64-main with linux_job_v2 seems to be the right track

Now we're just down to missing devtoolset-10-binutils, which is curious since pt/pt uses v10 for aarch64
Edit: Resolved; the pip installs were unnecessary

script: |
echo "::group::Print machine info"
Expand All @@ -35,9 +36,10 @@ jobs:
test-quantization-cpu:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
with:
runner: linux-aarch64
runner: linux.arm64.m7g.4xlarge
docker-image: "pytorch/manylinuxaarch64-builder:cuda12.6-main"
gpu-arch-type: cuda
gpu-arch-version: "12.1"
gpu-arch-version: "12.6"
timeout: 60
script: |
echo "::group::Print machine info"
Expand All @@ -54,9 +56,10 @@ jobs:
test-gguf-cpu:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
with:
runner: linux-aarch64
runner: linux.arm64.m7g.4xlarge
docker-image: "pytorch/manylinuxaarch64-builder:cuda12.6-main"
gpu-arch-type: cuda
gpu-arch-version: "12.1"
gpu-arch-version: "12.6"
timeout: 60
script: |
echo "::group::Print machine info"
Expand All @@ -78,9 +81,10 @@ jobs:
test-advanced-cpu:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
with:
runner: linux-aarch64
runner: linux.arm64.m7g.4xlarge
docker-image: "pytorch/manylinuxaarch64-builder:cuda12.6-main"
gpu-arch-type: cuda
gpu-arch-version: "12.1"
gpu-arch-version: "12.6"
timeout: 60
script: |
echo "::group::Print machine info"
Expand All @@ -102,9 +106,10 @@ jobs:
test-evaluation-cpu:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main
with:
runner: linux-aarch64
runner: linux.arm64.m7g.4xlarge
docker-image: "pytorch/manylinuxaarch64-builder:cuda12.6-main"
gpu-arch-type: cuda
gpu-arch-version: "12.1"
gpu-arch-version: "12.6"
timeout: 60
script: |
echo "::group::Print machine info"
Expand Down
Loading