Skip to content

Commit d0413b0

Browse files
committed
Check length of gpu_split in model_init
1 parent c8fa853 commit d0413b0

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

exllamav2/model_init.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,6 @@
11

22
import argparse, sys, os, glob, time
3+
import torch
34

45
from exllamav2 import(
56
ExLlamaV2,
@@ -167,6 +168,9 @@ def post_init_load(
167168
split = None
168169
if args.gpu_split and args.gpu_split != "auto":
169170
split = [float(alloc) for alloc in args.gpu_split.split(",")]
171+
if len(split) > torch.cuda.device_count():
172+
print(f" ## Error: Too many entries in gpu_split. {torch.cuda.device_count()} CUDA devices are available.")
173+
sys.exit()
170174

171175
if args.tensor_parallel:
172176
if args.gpu_split == "auto": split = None

0 commit comments

Comments
 (0)