Skip to content

Commit 53b5d3e

Browse files
committed
Revert "Prevent launch if VRAM is too full"
This reverts commit 2319cf2.
1 parent 7b7d820 commit 53b5d3e

File tree

1 file changed

+0
-5
lines changed

1 file changed

+0
-5
lines changed

koboldcpp.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1602,11 +1602,6 @@ def autoset_gpu_layers(ctxsize, sdquanted, blasbatchsize, quantkv_var, flashatte
16021602
print(f"FIRST_STEP : Initial layer limit: {layerlimit_intermed} ; Model size: {fsize/1024/1024:.3f} MiB ; context size: {ctxsize} tokens")
16031603
print(f"GPUs global reserved VRAM: {reservedmem/1024/1024} MiB (Toral occupied VRAM + Total overhead) ; GPUs total usable VRAM: {mem/1024/1024} MiB")
16041604

1605-
if mem <= fsize/2:
1606-
exitcounter = 999
1607-
print(f" Model size: {fsize/1024/1024:.3f} MiB ; Available VRAM: {mem} MiB.")
1608-
exit_with_error(2,"There's not enough available VRAM to make a reasonably performing offload. Exiting.")
1609-
16101605
if fsize > (10*1024*1024): #dont bother with models < 10mb
16111606
cs = ctxsize
16121607
# mem = gpumem

0 commit comments

Comments
 (0)