| 
61 | 61 |       group: aws-g4dn-2xlarge  | 
62 | 62 |     container:  | 
63 | 63 |       image: diffusers/diffusers-pytorch-cuda  | 
64 |  | -      options: --shm-size "16gb" --ipc host --gpus 0  | 
 | 64 | +      options: --shm-size "16gb" --ipc host --gpus all  | 
65 | 65 |     steps:  | 
66 | 66 |       - name: Checkout diffusers  | 
67 | 67 |         uses: actions/checkout@v3  | 
@@ -107,7 +107,7 @@ jobs:  | 
107 | 107 |       group: aws-g4dn-2xlarge  | 
108 | 108 |     container:  | 
109 | 109 |       image: diffusers/diffusers-pytorch-cuda  | 
110 |  | -      options: --shm-size "16gb" --ipc host --gpus 0  | 
 | 110 | +      options: --shm-size "16gb" --ipc host --gpus all  | 
111 | 111 |     defaults:  | 
112 | 112 |       run:  | 
113 | 113 |         shell: bash  | 
@@ -178,7 +178,7 @@ jobs:  | 
178 | 178 | 
 
  | 
179 | 179 |     container:  | 
180 | 180 |       image: diffusers/diffusers-pytorch-cuda  | 
181 |  | -      options: --gpus 0 --shm-size "16gb" --ipc host  | 
 | 181 | +      options: --gpus all --shm-size "16gb" --ipc host  | 
182 | 182 | 
 
  | 
183 | 183 |     steps:  | 
184 | 184 |     - name: Checkout diffusers  | 
@@ -222,7 +222,7 @@ jobs:  | 
222 | 222 |       group: aws-g6e-xlarge-plus  | 
223 | 223 |     container:  | 
224 | 224 |       image: diffusers/diffusers-pytorch-cuda  | 
225 |  | -      options: --shm-size "16gb" --ipc host --gpus 0  | 
 | 225 | +      options: --shm-size "16gb" --ipc host --gpus all  | 
226 | 226 |     steps:  | 
227 | 227 |       - name: Checkout diffusers  | 
228 | 228 |         uses: actions/checkout@v3  | 
@@ -270,7 +270,7 @@ jobs:  | 
270 | 270 |       group: aws-g4dn-2xlarge  | 
271 | 271 |     container:  | 
272 | 272 |       image: diffusers/diffusers-pytorch-minimum-cuda  | 
273 |  | -      options: --shm-size "16gb" --ipc host --gpus 0  | 
 | 273 | +      options: --shm-size "16gb" --ipc host --gpus all  | 
274 | 274 |     defaults:  | 
275 | 275 |       run:  | 
276 | 276 |         shell: bash  | 
@@ -344,7 +344,7 @@ jobs:  | 
344 | 344 |       group: aws-g6e-xlarge-plus  | 
345 | 345 |     container:  | 
346 | 346 |       image: diffusers/diffusers-pytorch-cuda  | 
347 |  | -      options: --shm-size "20gb" --ipc host --gpus 0  | 
 | 347 | +      options: --shm-size "20gb" --ipc host --gpus all  | 
348 | 348 |     steps:  | 
349 | 349 |       - name: Checkout diffusers  | 
350 | 350 |         uses: actions/checkout@v3  | 
@@ -396,7 +396,7 @@ jobs:  | 
396 | 396 |       group: aws-g6e-xlarge-plus  | 
397 | 397 |     container:  | 
398 | 398 |       image: diffusers/diffusers-pytorch-cuda  | 
399 |  | -      options: --shm-size "20gb" --ipc host --gpus 0  | 
 | 399 | +      options: --shm-size "20gb" --ipc host --gpus all  | 
400 | 400 |     steps:  | 
401 | 401 |       - name: Checkout diffusers  | 
402 | 402 |         uses: actions/checkout@v3  | 
 | 
0 commit comments