Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
13 changes: 8 additions & 5 deletions kcpp_adapters/AutoGuess.json
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,8 @@
"system_end": "[/INST]",
"user_start": "[INST] ",
"user_end": "",
"assistant_start": "[/INST]",
"assistant_start": "[/INST] ",
"assistant_gen": "[/INST]",
"assistant_end": "</s>"
}
}, {
Expand Down Expand Up @@ -168,11 +169,12 @@
"search": ["<|bom|>","is_last_checked_defined"],
"name": "Jamba",
"adapter": {
"system_start": "<|bom|><|system|>",
"system_start": "<|bom|><|system|> ",
"system_end": "<|eom|>",
"user_start": "<|bom|><|user|>",
"user_start": "<|bom|><|user|> ",
"user_end": "<|eom|>",
"assistant_start": "<|bom|><|assistant|>",
"assistant_start": "<|bom|><|assistant|> ",
"assistant_gen": "<|bom|><|assistant|>",
"assistant_end": "<|eom|>"
}
}, {
Expand All @@ -191,7 +193,7 @@
"name": "OpenAI Harmony",
"adapter": {
"system_start": "<|start|>developer<|message|># Instructions\n\n",
"system_end": "<|end|>",
"system_end": "\n\n<|end|>",
"user_start": "<|start|>user<|message|>",
"user_end": "<|end|>",
"assistant_start": "<|start|>assistant<|channel|>final<|message|>",
Expand All @@ -206,6 +208,7 @@
"user_start": "User: ",
"user_end": "\n\n",
"assistant_start": "Assistant: ",
"assistant_gen": "Assistant:",
"assistant_end": "\n\n"
}
}, {
Expand Down
9 changes: 6 additions & 3 deletions koboldcpp.py
Original file line number Diff line number Diff line change
Expand Up @@ -2419,6 +2419,7 @@ def transform_genparams(genparams, api_format):
user_message_end = adapter_obj.get("user_end", "")
assistant_message_start = adapter_obj.get("assistant_start", "\n### Response:\n")
assistant_message_end = adapter_obj.get("assistant_end", "")
assistant_message_gen = adapter_obj.get("assistant_gen", assistant_message_start)
tools_message_start = adapter_obj.get("tools_start", "\nTool Results:\n")
tools_message_end = adapter_obj.get("tools_end", "")
images_added = []
Expand Down Expand Up @@ -2531,7 +2532,7 @@ def transform_genparams(genparams, api_format):
elif message['role'] == "tool":
messages_string += tools_message_end

messages_string += assistant_message_start
messages_string += assistant_message_gen
genparams["prompt"] = messages_string
if len(images_added)>0:
genparams["images"] = images_added
Expand All @@ -2552,20 +2553,22 @@ def transform_genparams(genparams, api_format):
adapter_obj = {} if chatcompl_adapter is None else chatcompl_adapter
user_message_start = adapter_obj.get("user_start", "### Instruction:")
assistant_message_start = adapter_obj.get("assistant_start", "### Response:")
genparams["prompt"] = f"{user_message_start} In one sentence, write a descriptive caption for this image.\n{assistant_message_start}"
assistant_message_gen = adapter_obj.get("assistant_gen", assistant_message_start)
genparams["prompt"] = f"{user_message_start} In one sentence, write a descriptive caption for this image.\n{assistant_message_gen}"

elif api_format==6:
detokstr = ""
tokids = genparams.get('context', [])
adapter_obj = {} if chatcompl_adapter is None else chatcompl_adapter
user_message_start = adapter_obj.get("user_start", "\n\n### Instruction:\n")
assistant_message_start = adapter_obj.get("assistant_start", "\n\n### Response:\n")
assistant_message_gen = adapter_obj.get("assistant_gen", assistant_message_start)
try:
detokstr = detokenize_ids(tokids)
except Exception as e:
utfprint("Ollama Context Error: " + str(e))
ollamasysprompt = genparams.get('system', "")
ollamabodyprompt = f"{detokstr}{user_message_start}{genparams.get('prompt', '')}{assistant_message_start}"
ollamabodyprompt = f"{detokstr}{user_message_start}{genparams.get('prompt', '')}{assistant_message_gen}"
ollamaopts = genparams.get('options', {})
if genparams.get('stop',[]) is not None:
genparams["stop_sequence"] = genparams.get('stop', [])
Expand Down
30 changes: 19 additions & 11 deletions tests/test_autoguess.py
Original file line number Diff line number Diff line change
Expand Up @@ -21,8 +21,8 @@
"Google Gemma 3n": "lmstudio-community/gemma-3n-E4B-it-MLX-bf16",
"Llama 3.x": "Steelskull/L3.3-Shakudo-70b",
"Llama 4": "nvidia/Llama-4-Scout-17B-16E-Instruct-FP8",
"Mistral V7 (with system prompt)": "Doctor-Shotgun/MS3.2-24B-Magnum-Diamond",
"Mistral V3": "mistralai/Mistral-7B-Instruct-v0.3",
"Mistral Tekken": "Doctor-Shotgun/MS3.2-24B-Magnum-Diamond",
"Mistral Non-Tekken": "mistralai/Mistral-7B-Instruct-v0.3",
"GLM-4": "THUDM/glm-4-9b-chat-hf",
"Phi 3.5": "microsoft/Phi-3.5-mini-instruct",
"Phi 4 (mini)": "microsoft/Phi-4-mini-instruct",
Expand All @@ -31,12 +31,13 @@
"Jamba": "ai21labs/Jamba-tiny-dev",
"Dots": "rednote-hilab/dots.llm1.inst",
"RWKV World": "fla-hub/rwkv7-1.5B-world",
"OpenAI Harmony": "openai/gpt-oss-120b",
"Mistral (Generic)": "mistralai/Mistral-Nemo-Instruct-2407",
"ChatML (Generic)": "NewEden/Gemma-27B-chatml",
}

AUTOGUESS_SKIP_ADAPTER_TESTS = {
"Mistral V3": {"system"}, # Poor system support
"Mistral Non-Tekken": {"system"}, # Poor system support
"Mistral (Generic)": {"system"}, # Poor system support
}

Expand All @@ -58,10 +59,12 @@ def get_tokenizer_config_for_huggingface_model_id(huggingface_model_id: str):
with open(fname) as f:
return json.load(f)

for filename in ["tokenizer_config.json", "chat_template.json"]:
for filename in ["tokenizer_config.json", "chat_template.json", "chat_template.jinja"]:
url = f"https://huggingface.co/{huggingface_model_id}/resolve/main/{filename}"
response = requests.get(url)
if response.status_code == 200:
if url.endswith(".jinja"):
return {"chat_template": response.text}
v = json.loads(response.text)
if 'chat_template' in v:
return v
Expand Down Expand Up @@ -113,9 +116,9 @@ def templ(rolelist):
expect = system("SyS-tEm")
templated = templ([{"role": "system", "content": "SyS-tEm"}, {"role": "user", "content": "user"}])
if expect not in templated:
return False, f"system role missing expected fragment {expect.replace("\n", "\\n")}: {templated.replace("\n", "\\n")}"
return False, f"system role missing expected fragment\n\tadapter: {expect.replace("\n", "\\n")}\n\ttokenizer: {templated.replace("\n", "\\n")}"

# Test user/asst/usernvidia/Llama-4-Scout-17B-16E-Instruct-FP8
# Test user/asst/user
expect = [
user("user_1"),
assistant("asst_1"),
Expand All @@ -129,17 +132,21 @@ def templ(rolelist):
rem = templated
for sub in expect:
if sub not in rem:
return False, f"missing expected fragment {sub.replace("\n", "\\n")}: {rem.replace("\n", "\\n")}"
return False, f"missing expected fragment\n\tadapter: {sub.replace("\n", "\\n")}\n\ttokenizer: {rem.replace("\n", "\\n")}"
rem = rem.split(sub, 1)[1]
except jinja2.exceptions.TemplateError as e:
return False, f"template error: {e}"
return True, None

filter = sys.argv[1] if len(sys.argv) > 1 else None

failures = 0
seen = set()
namefmt = "{name:<" + str(max(len(name) for name in AUTOGUESS_MAPPING.keys())) + "}"
hmifmt = "{huggingface_model_id:<" + str(max(len(huggingface_model_id) for huggingface_model_id in AUTOGUESS_MAPPING.values())) + "}"
for name, huggingface_model_id in AUTOGUESS_MAPPING.items():
if filter and filter not in name:
continue
seen.add(name)
if huggingface_model_id == "***UNKNOWN***":
print(namefmt.format(name=name) + " = " + namefmt.format(name="***UNKNOWN***") + " : PENDING")
Expand All @@ -162,10 +169,11 @@ def templ(rolelist):
print(namefmt.format(name=name) + " = " + namefmt.format(name=matched) + " : " + ("OK " if adaptercheck and name == matched else reason if not adaptercheck else "FAILURE") + " " + hmifmt.format(huggingface_model_id=huggingface_model_id) + " " + sub_template)
failures += name != matched or not adaptercheck

for entry in autoguess:
if entry['name'] not in seen:
print(namefmt.format(name=entry['name']) + " MISSING MAPPING")
failures += 1
if filter is None:
for entry in autoguess:
if entry['name'] not in seen:
print(namefmt.format(name=entry['name']) + " MISSING MAPPING")
failures += 1

if failures > 0:
print(f"There were {failures} failure(s)!")
Expand Down