@@ -101,11 +101,12 @@ def prompt(
101101 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
102102 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
103103 (better models yield better results, faster models yield faster results). -
104- Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
105- "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
106- "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
107- "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
108- for low latency/costs: "gpt-4.1-nano", "nova-micro".
104+ Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
105+ "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
106+ "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
107+ models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
108+ "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
109+ "nova-micro".
109110
110111 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
111112 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -244,11 +245,12 @@ def score(
244245 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
245246 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
246247 (better models yield better results, faster models yield faster results). -
247- Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
248- "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
249- "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
250- "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
251- for low latency/costs: "gpt-4.1-nano", "nova-micro".
248+ Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
249+ "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
250+ "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
251+ models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
252+ "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
253+ "nova-micro".
252254
253255 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
254256 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -403,11 +405,12 @@ async def prompt(
403405 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
404406 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
405407 (better models yield better results, faster models yield faster results). -
406- Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
407- "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
408- "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
409- "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
410- for low latency/costs: "gpt-4.1-nano", "nova-micro".
408+ Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
409+ "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
410+ "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
411+ models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
412+ "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
413+ "nova-micro".
411414
412415 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
413416 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -546,11 +549,12 @@ async def score(
546549 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
547550 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
548551 (better models yield better results, faster models yield faster results). -
549- Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
550- "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
551- "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
552- "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
553- for low latency/costs: "gpt-4.1-nano", "nova-micro".
552+ Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
553+ "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
554+ "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
555+ models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
556+ "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
557+ "nova-micro".
554558
555559 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
556560 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
0 commit comments