@@ -1350,8 +1350,8 @@ jobs:
13501350        run : | 
13511351          LLAMA_ARG_THREADS=$(nproc) GG_BUILD_NO_SVE=1 GG_BUILD_NO_BF16=1 GG_BUILD_EXTRA_TESTS_0=1 bash ./ci/run.sh ./tmp/results ./tmp/mnt 
13521352
1353- ggml-ci-x64-nvidia-v100- cuda :
1354-     runs-on : [self-hosted, Linux, X64, NVIDIA, V100 ] 
1353+ ggml-ci-x64-nvidia-cuda :
1354+     runs-on : [self-hosted, Linux, X64, NVIDIA] 
13551355
13561356    steps :
13571357      - name : Clone 
@@ -1364,8 +1364,8 @@ jobs:
13641364          nvidia-smi 
13651365          GG_BUILD_CUDA=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
13661366
1367- ggml-ci-x64-nvidia-v100- vulkan :
1368-     runs-on : [self-hosted, Linux, X64, NVIDIA, V100 ] 
1367+ ggml-ci-x64-nvidia-vulkan-cm  :
1368+     runs-on : [self-hosted, Linux, X64, NVIDIA] 
13691369
13701370    steps :
13711371      - name : Clone 
@@ -1375,25 +1375,11 @@ jobs:
13751375      - name : Test 
13761376        id : ggml-ci 
13771377        run : | 
1378-           vulkaninfo 
1379-           GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
1380- 
1381- ggml-ci-x64-nvidia-t4-cuda :
1382-     runs-on : [self-hosted, Linux, X64, NVIDIA, T4] 
1383- 
1384-     steps :
1385-       - name : Clone 
1386-         id : checkout 
1387-         uses : actions/checkout@v4 
1388- 
1389-       - name : Test 
1390-         id : ggml-ci 
1391-         run : | 
1392-           nvidia-smi 
1393-           GG_BUILD_CUDA=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
1378+           vulkaninfo --summary 
1379+           GG_BUILD_VULKAN=1 GGML_VK_DISABLE_COOPMAT2=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
13941380
1395- ggml-ci-x64-nvidia-t4- vulkan :
1396-     runs-on : [self-hosted, Linux, X64, NVIDIA, T4 ] 
1381+ ggml-ci-x64-nvidia-vulkan-cm2  :
1382+     runs-on : [self-hosted, Linux, X64, NVIDIA, COOPMAT2 ] 
13971383
13981384    steps :
13991385      - name : Clone 
@@ -1403,23 +1389,9 @@ jobs:
14031389      - name : Test 
14041390        id : ggml-ci 
14051391        run : | 
1406-           vulkaninfo 
1392+           vulkaninfo --summary  
14071393          GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
14081394
1409- ggml-ci-x64-nvidia-t4-vulkan-coopmat1 :
1410-     runs-on : [self-hosted, Linux, X64, NVIDIA, T4] 
1411- 
1412-     steps :
1413-       - name : Clone 
1414-         id : checkout 
1415-         uses : actions/checkout@v4 
1416- 
1417-       - name : Test 
1418-         id : ggml-ci 
1419-         run : | 
1420-           vulkaninfo 
1421-           GG_BUILD_VULKAN=1 GGML_VK_DISABLE_COOPMAT2=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
1422- 
14231395ggml-ci-x64-cpu-amx :
14241396    runs-on : [self-hosted, Linux, X64, CPU, AMX] 
14251397
@@ -1433,21 +1405,36 @@ jobs:
14331405        run : | 
14341406          bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
14351407
1436- ggml-ci-x64-amd-v710-vulkan :
1437-     runs-on : [self-hosted, Linux, X64, AMD, V710] 
1438- 
1439-     steps :
1440-       - name : Clone 
1441-         id : checkout 
1442-         uses : actions/checkout@v4 
1443- 
1444-       - name : Test 
1445-         id : ggml-ci 
1446-         run : | 
1447-           GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
1408+ #   ggml-ci-x64-amd-vulkan:
1409+ #     runs-on: [self-hosted, Linux, X64, AMD]
1410+ # 
1411+ #     steps:
1412+ #       - name: Clone
1413+ #         id: checkout
1414+ #         uses: actions/checkout@v4
1415+ # 
1416+ #       - name: Test
1417+ #         id: ggml-ci
1418+ #         run: |
1419+ #           vulkaninfo --summary
1420+ #           GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
1421+ # 
1422+ #   ggml-ci-x64-amd-rocm:
1423+ #     runs-on: [self-hosted, Linux, X64, AMD]
1424+ # 
1425+ #     steps:
1426+ #       - name: Clone
1427+ #         id: checkout
1428+ #         uses: actions/checkout@v4
1429+ # 
1430+ #       - name: Test
1431+ #         id: ggml-ci
1432+ #         run: |
1433+ #           amd-smi static
1434+ #           GG_BUILD_ROCM=1 GG_BUILD_AMDGPU_TARGETS="gfx1101" bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
14481435
1449- ggml-ci-x64-amd-v710-rocm  :
1450-     runs-on : [self-hosted, Linux, X64, AMD, V710 ] 
1436+   ggml-ci-mac-metal  :
1437+     runs-on : [self-hosted, macOS, ARM64 ] 
14511438
14521439    steps :
14531440      - name : Clone 
@@ -1457,9 +1444,9 @@ jobs:
14571444      - name : Test 
14581445        id : ggml-ci 
14591446        run : | 
1460-           GG_BUILD_ROCM =1 GG_BUILD_AMDGPU_TARGETS="gfx1101"  bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp 
1447+           GG_BUILD_METAL =1 bash ./ci/run.sh ~/results/llama.cpp ~ /mnt/llama.cpp 
14611448
1462- ggml-ci-mac-metal  :
1449+ ggml-ci-mac-vulkan  :
14631450    runs-on : [self-hosted, macOS, ARM64] 
14641451
14651452    steps :
@@ -1470,18 +1457,5 @@ jobs:
14701457      - name : Test 
14711458        id : ggml-ci 
14721459        run : | 
1473-           GG_BUILD_METAL=1 bash ./ci/run.sh ~/results/llama.cpp ~/mnt/llama.cpp 
1474- 
1475- #  TODO: install vulkan drivers
1476- #   ggml-ci-mac-vulkan:
1477- #     runs-on: [self-hosted, macOS, ARM64]
1478- # 
1479- #     steps:
1480- #       - name: Clone
1481- #         id: checkout
1482- #         uses: actions/checkout@v4
1483- # 
1484- #       - name: Test
1485- #         id: ggml-ci
1486- #         run: |
1487- #           GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp ~/mnt/llama.cpp
1460+           vulkaninfo --summary 
1461+           GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp ~/mnt/llama.cpp 
0 commit comments