@@ -1302,8 +1302,8 @@ jobs:
13021302 run : |
13031303 GG_BUILD_NO_BF16=1 GG_BUILD_EXTRA_TESTS_0=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
13041304
1305- ggml-ci-x64-nvidia-v100- cuda :
1306- runs-on : [self-hosted, Linux, X64, NVIDIA, V100 ]
1305+ ggml-ci-x64-nvidia-cuda :
1306+ runs-on : [self-hosted, Linux, X64, NVIDIA]
13071307
13081308 steps :
13091309 - name : Clone
@@ -1316,8 +1316,8 @@ jobs:
13161316 nvidia-smi
13171317 GG_BUILD_CUDA=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
13181318
1319- ggml-ci-x64-nvidia-v100- vulkan :
1320- runs-on : [self-hosted, Linux, X64, NVIDIA, V100 ]
1319+ ggml-ci-x64-nvidia-vulkan :
1320+ runs-on : [self-hosted, Linux, X64, NVIDIA]
13211321
13221322 steps :
13231323 - name : Clone
@@ -1327,11 +1327,11 @@ jobs:
13271327 - name : Test
13281328 id : ggml-ci
13291329 run : |
1330- vulkaninfo
1330+ vulkaninfo --summary
13311331 GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
13321332
1333- ggml-ci-x64-nvidia-t4-cuda :
1334- runs-on : [self-hosted, Linux, X64, NVIDIA, T4 ]
1333+ ggml-ci-x64-nvidia-vulkan-coopmat1 :
1334+ runs-on : [self-hosted, Linux, X64, NVIDIA]
13351335
13361336 steps :
13371337 - name : Clone
@@ -1341,35 +1341,7 @@ jobs:
13411341 - name : Test
13421342 id : ggml-ci
13431343 run : |
1344- nvidia-smi
1345- GG_BUILD_CUDA=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
1346-
1347- ggml-ci-x64-nvidia-t4-vulkan :
1348- runs-on : [self-hosted, Linux, X64, NVIDIA, T4]
1349-
1350- steps :
1351- - name : Clone
1352- id : checkout
1353- uses : actions/checkout@v4
1354-
1355- - name : Test
1356- id : ggml-ci
1357- run : |
1358- vulkaninfo
1359- GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
1360-
1361- ggml-ci-x64-nvidia-t4-vulkan-coopmat1 :
1362- runs-on : [self-hosted, Linux, X64, NVIDIA, T4]
1363-
1364- steps :
1365- - name : Clone
1366- id : checkout
1367- uses : actions/checkout@v4
1368-
1369- - name : Test
1370- id : ggml-ci
1371- run : |
1372- vulkaninfo
1344+ vulkaninfo --summary
13731345 GG_BUILD_VULKAN=1 GGML_VK_DISABLE_COOPMAT2=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
13741346
13751347 ggml-ci-x64-cpu-amx :
@@ -1385,31 +1357,33 @@ jobs:
13851357 run : |
13861358 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
13871359
1388- ggml-ci-x64-amd-v710-vulkan :
1389- runs-on : [self-hosted, Linux, X64, AMD, V710]
1390-
1391- steps :
1392- - name : Clone
1393- id : checkout
1394- uses : actions/checkout@v4
1395-
1396- - name : Test
1397- id : ggml-ci
1398- run : |
1399- GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
1400-
1401- ggml-ci-x64-amd-v710-rocm :
1402- runs-on : [self-hosted, Linux, X64, AMD, V710]
1403-
1404- steps :
1405- - name : Clone
1406- id : checkout
1407- uses : actions/checkout@v4
1408-
1409- - name : Test
1410- id : ggml-ci
1411- run : |
1412- GG_BUILD_ROCM=1 GG_BUILD_AMDGPU_TARGETS="gfx1101" bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
1360+ # ggml-ci-x64-amd-vulkan:
1361+ # runs-on: [self-hosted, Linux, X64, AMD]
1362+ #
1363+ # steps:
1364+ # - name: Clone
1365+ # id: checkout
1366+ # uses: actions/checkout@v4
1367+ #
1368+ # - name: Test
1369+ # id: ggml-ci
1370+ # run: |
1371+ # vulkaninfo --summary
1372+ # GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
1373+ #
1374+ # ggml-ci-x64-amd-rocm:
1375+ # runs-on: [self-hosted, Linux, X64, AMD]
1376+ #
1377+ # steps:
1378+ # - name: Clone
1379+ # id: checkout
1380+ # uses: actions/checkout@v4
1381+ #
1382+ # - name: Test
1383+ # id: ggml-ci
1384+ # run: |
1385+ # amd-smi
1386+ # GG_BUILD_ROCM=1 GG_BUILD_AMDGPU_TARGETS="gfx1101" bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
14131387
14141388 ggml-ci-mac-metal :
14151389 runs-on : [self-hosted, macOS, ARM64]
@@ -1435,4 +1409,5 @@ jobs:
14351409 - name : Test
14361410 id : ggml-ci
14371411 run : |
1412+ vulkaninfo --summary
14381413 GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp ~/mnt/llama.cpp
0 commit comments