Skip to content

Commit 2f17eeb

Browse files
committed
Merge branch 'remoteManagement' into crokeso
2 parents 3f3dee7 + b74b546 commit 2f17eeb

File tree

1 file changed

+17
-13
lines changed

1 file changed

+17
-13
lines changed

koboldcpp.py

Lines changed: 17 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -59,11 +59,11 @@
5959
# dry_seq_break_max = 128
6060

6161
# global vars
62-
KcppVersion = "1.90105"
63-
LcppVersion = "b5226"
64-
EsoboldVersion = "RMv1.8.3+7c"
62+
KcppVersion = "1.90110"
63+
LcppVersion = "b5232"
64+
EsoboldVersion = "RMv1.9.1"
6565
CudaSpecifics = "Cu128_Ar86_SMC2_DmmvX32Y1"
66-
ReleaseDate = "2025/04/30"
66+
ReleaseDate = "2025/05/02"
6767
showdebug = True
6868
# guimode = False
6969
kcpp_instance = None #global running instance
@@ -2766,10 +2766,11 @@ def transform_genparams(genparams, api_format):
27662766
presence_penalty = genparams.get('presence_penalty', genparams.get('frequency_penalty', 0.0))
27672767
genparams["presence_penalty"] = tryparsefloat(presence_penalty,0.0)
27682768
# openai allows either a string or a list as a stop sequence
2769-
if isinstance(genparams.get('stop',[]), list):
2770-
genparams["stop_sequence"] = genparams.get('stop', [])
2771-
else:
2772-
genparams["stop_sequence"] = [genparams.get('stop')]
2769+
if genparams.get('stop',[]) is not None:
2770+
if isinstance(genparams.get('stop',[]), list):
2771+
genparams["stop_sequence"] = genparams.get('stop', [])
2772+
else:
2773+
genparams["stop_sequence"] = [genparams.get('stop')]
27732774

27742775
genparams["sampler_seed"] = tryparseint(genparams.get('seed', -1),-1)
27752776
genparams["mirostat"] = genparams.get('mirostat_mode', 0)
@@ -2964,7 +2965,8 @@ def transform_genparams(genparams, api_format):
29642965
ollamasysprompt = genparams.get('system', "")
29652966
ollamabodyprompt = f"{detokstr}{user_message_start}{genparams.get('prompt', '')}{assistant_message_start}"
29662967
ollamaopts = genparams.get('options', {})
2967-
genparams["stop_sequence"] = genparams.get('stop', [])
2968+
if genparams.get('stop',[]) is not None:
2969+
genparams["stop_sequence"] = genparams.get('stop', [])
29682970
if "num_predict" in ollamaopts:
29692971
genparams["max_length"] = ollamaopts.get('num_predict', args.defaultgenamt)
29702972
if "num_ctx" in ollamaopts:
@@ -4819,9 +4821,12 @@ def zenity(filetypes=None, initialdir="", initialfile="", **kwargs) -> Tuple[int
48194821
if sys.platform != "linux":
48204822
raise Exception("Zenity GUI is only usable on Linux, attempting to use TK GUI.")
48214823
zenity_bin = shutil.which("yad")
4824+
using_yad = True
48224825
if not zenity_bin:
48234826
zenity_bin = shutil.which("zenity")
4827+
using_yad = False
48244828
if not zenity_bin:
4829+
using_yad = False
48254830
raise Exception("Zenity not present, falling back to TK GUI.")
48264831

48274832
def zenity_clean(txt: str):
@@ -4850,7 +4855,7 @@ def zenity_sanity_check(zenity_bin): #make sure zenity is sane
48504855
raise Exception("Zenity not working correctly, falling back to TK GUI.")
48514856

48524857
# Build args based on keywords
4853-
args = ['/usr/bin/env', zenity_bin, '--file-selection']
4858+
args = ['/usr/bin/env', zenity_bin, ('--file' if using_yad else '--file-selection')]
48544859
for k, v in kwargs.items():
48554860
if v is True:
48564861
args.append(f'--{k.replace("_", "-").strip("-")}')
@@ -7492,7 +7497,7 @@ def kcpp_main_process(launch_args, g_memory=None, gui_launcher=False):
74927497
global maxctx
74937498
maxctx = args.contextsize
74947499

7495-
args.defaultgenamt = max(128, min(args.defaultgenamt, 2048))
7500+
args.defaultgenamt = max(128, min(args.defaultgenamt, 4096))
74967501
args.defaultgenamt = min(args.defaultgenamt, maxctx / 2)
74977502

74987503
if args.nocertify:
@@ -8176,12 +8181,11 @@ def range_checker(arg: str):
81768181
advparser.add_argument("--exporttemplate", help="Exports the current selected arguments as a .kcppt template file", metavar=('[filename]'), type=str, default="")
81778182
advparser.add_argument("--nomodel", help="Allows you to launch the GUI alone, without selecting any model.", action='store_true')
81788183
advparser.add_argument("--moeexperts", metavar=('[num of experts]'), help="How many experts to use for MoE models (default=follow gguf)", type=int, default=-1)
8179-
81808184
advparser.add_argument("--normrmseps", metavar=('[norm rms eps]'), help="Override Norm RMS Epsilon value to use for the model. Useful for <2bpw quants mainly. Example of format: 1.95e-05 (default=follow gguf)", type=float, default=-1.0)
81818185
advparser.add_argument("--poslayeroffset", help="Removes or adds a layer to the GPU layers autoloader calculation in case of OOM or under-exploitation.", type=check_range(int,0,10), default=0)
81828186
advparser.add_argument("--neglayeroffset", help="Removes or adds a layer to the GPU layers autoloader calculation in case of OOM or under-exploitation.", type=check_range(int,0,10), default=0)
81838187

8184-
advparser.add_argument("--defaultgenamt", help="How many tokens to generate by default, if not specified. Must be smaller than context size. Usually, your frontend GUI will override this.", type=check_range(int,128,2048), default=512)
8188+
advparser.add_argument("--defaultgenamt", help="How many tokens to generate by default, if not specified. Must be smaller than context size. Usually, your frontend GUI will override this.", type=check_range(int,64,4096), default=512)
81858189
advparser.add_argument("--nobostoken", help="Prevents BOS token from being added at the start of any prompt. Usually NOT recommended for most models.", action='store_true')
81868190
advparser.add_argument("--maxrequestsize", metavar=('[size in MB]'), help="Specify a max request payload size. Any requests to the server larger than this size will be dropped. Do not change if unsure.", type=int, default=32)
81878191

0 commit comments

Comments
 (0)