|
61 | 61 | group: aws-g4dn-2xlarge |
62 | 62 | container: |
63 | 63 | image: diffusers/diffusers-pytorch-cuda |
64 | | - options: --shm-size "16gb" --ipc host --gpus 0 |
| 64 | + options: --shm-size "16gb" --ipc host --gpus all |
65 | 65 | steps: |
66 | 66 | - name: Checkout diffusers |
67 | 67 | uses: actions/checkout@v3 |
@@ -107,7 +107,7 @@ jobs: |
107 | 107 | group: aws-g4dn-2xlarge |
108 | 108 | container: |
109 | 109 | image: diffusers/diffusers-pytorch-cuda |
110 | | - options: --shm-size "16gb" --ipc host --gpus 0 |
| 110 | + options: --shm-size "16gb" --ipc host --gpus all |
111 | 111 | defaults: |
112 | 112 | run: |
113 | 113 | shell: bash |
@@ -178,7 +178,7 @@ jobs: |
178 | 178 |
|
179 | 179 | container: |
180 | 180 | image: diffusers/diffusers-pytorch-cuda |
181 | | - options: --gpus 0 --shm-size "16gb" --ipc host |
| 181 | + options: --gpus all --shm-size "16gb" --ipc host |
182 | 182 |
|
183 | 183 | steps: |
184 | 184 | - name: Checkout diffusers |
@@ -222,7 +222,7 @@ jobs: |
222 | 222 | group: aws-g6e-xlarge-plus |
223 | 223 | container: |
224 | 224 | image: diffusers/diffusers-pytorch-cuda |
225 | | - options: --shm-size "16gb" --ipc host --gpus 0 |
| 225 | + options: --shm-size "16gb" --ipc host --gpus all |
226 | 226 | steps: |
227 | 227 | - name: Checkout diffusers |
228 | 228 | uses: actions/checkout@v3 |
@@ -270,7 +270,7 @@ jobs: |
270 | 270 | group: aws-g4dn-2xlarge |
271 | 271 | container: |
272 | 272 | image: diffusers/diffusers-pytorch-minimum-cuda |
273 | | - options: --shm-size "16gb" --ipc host --gpus 0 |
| 273 | + options: --shm-size "16gb" --ipc host --gpus all |
274 | 274 | defaults: |
275 | 275 | run: |
276 | 276 | shell: bash |
@@ -344,7 +344,7 @@ jobs: |
344 | 344 | group: aws-g6e-xlarge-plus |
345 | 345 | container: |
346 | 346 | image: diffusers/diffusers-pytorch-cuda |
347 | | - options: --shm-size "20gb" --ipc host --gpus 0 |
| 347 | + options: --shm-size "20gb" --ipc host --gpus all |
348 | 348 | steps: |
349 | 349 | - name: Checkout diffusers |
350 | 350 | uses: actions/checkout@v3 |
@@ -396,7 +396,7 @@ jobs: |
396 | 396 | group: aws-g6e-xlarge-plus |
397 | 397 | container: |
398 | 398 | image: diffusers/diffusers-pytorch-cuda |
399 | | - options: --shm-size "20gb" --ipc host --gpus 0 |
| 399 | + options: --shm-size "20gb" --ipc host --gpus all |
400 | 400 | steps: |
401 | 401 | - name: Checkout diffusers |
402 | 402 | uses: actions/checkout@v3 |
|
0 commit comments