Skip to content

Commit d51e2a9

Browse files
feat(api): api update
1 parent 66efcb5 commit d51e2a9

File tree

8 files changed

+58
-55
lines changed

8 files changed

+58
-55
lines changed

.stats.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
11
configured_endpoints: 54
2-
openapi_spec_hash: 04e1b7aefbeff10daab249b153de147f
2+
openapi_spec_hash: ad0180a0926e2b6434601e6adf9c322a
33
config_hash: 930284cfa37f835d949c8a1b124f4807

src/codex/resources/projects/projects.py

Lines changed: 10 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -527,12 +527,11 @@ def validate(
527527
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
528528
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
529529
(better models yield better results, faster models yield faster results). -
530-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
531-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
532-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
533-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
534-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
535-
"nova-micro".
530+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
531+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
532+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
533+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
534+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
536535
537536
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
538537
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -1106,12 +1105,11 @@ async def validate(
11061105
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
11071106
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
11081107
(better models yield better results, faster models yield faster results). -
1109-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
1110-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
1111-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
1112-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
1113-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
1114-
"nova-micro".
1108+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
1109+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
1110+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
1111+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
1112+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
11151113
11161114
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
11171115
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.

src/codex/resources/tlm.py

Lines changed: 20 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -101,12 +101,11 @@ def prompt(
101101
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
102102
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
103103
(better models yield better results, faster models yield faster results). -
104-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
105-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
106-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
107-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
108-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
109-
"nova-micro".
104+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
105+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
106+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
107+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
108+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
110109
111110
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
112111
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -245,12 +244,11 @@ def score(
245244
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
246245
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
247246
(better models yield better results, faster models yield faster results). -
248-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
249-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
250-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
251-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
252-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
253-
"nova-micro".
247+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
248+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
249+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
250+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
251+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
254252
255253
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
256254
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -405,12 +403,11 @@ async def prompt(
405403
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
406404
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
407405
(better models yield better results, faster models yield faster results). -
408-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
409-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
410-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
411-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
412-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
413-
"nova-micro".
406+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
407+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
408+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
409+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
410+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
414411
415412
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
416413
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -549,12 +546,11 @@ async def score(
549546
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
550547
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
551548
(better models yield better results, faster models yield faster results). -
552-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
553-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
554-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
555-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
556-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
557-
"nova-micro".
549+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
550+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
551+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
552+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
553+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
558554
559555
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
560556
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.

src/codex/types/project_validate_params.py

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -130,12 +130,11 @@ class ProjectValidateParams(TypedDict, total=False):
130130
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
131131
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
132132
(better models yield better results, faster models yield faster results). -
133-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
134-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
135-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
136-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
137-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
138-
"nova-micro".
133+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
134+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
135+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
136+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
137+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
139138
140139
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
141140
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -653,6 +652,8 @@ class MessageChatCompletionDeveloperMessageParam(TypedDict, total=False):
653652
class Options(TypedDict, total=False):
654653
custom_eval_criteria: Iterable[object]
655654

655+
disable_persistence: bool
656+
656657
disable_trustworthiness: bool
657658

658659
log: List[str]

src/codex/types/tlm_prompt_params.py

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -52,12 +52,11 @@ class TlmPromptParams(TypedDict, total=False):
5252
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
5353
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
5454
(better models yield better results, faster models yield faster results). -
55-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
56-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
57-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
58-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
59-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
60-
"nova-micro".
55+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
56+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
57+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
58+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
59+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
6160
6261
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
6362
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -116,6 +115,8 @@ class TlmPromptParams(TypedDict, total=False):
116115
class Options(TypedDict, total=False):
117116
custom_eval_criteria: Iterable[object]
118117

118+
disable_persistence: bool
119+
119120
disable_trustworthiness: bool
120121

121122
log: List[str]

src/codex/types/tlm_score_params.py

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -54,12 +54,11 @@ class TlmScoreParams(TypedDict, total=False):
5454
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
5555
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
5656
(better models yield better results, faster models yield faster results). -
57-
Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
58-
"o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
59-
"claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
60-
models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
61-
"claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
62-
"nova-micro".
57+
Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
58+
"gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
59+
"claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
60+
"o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
61+
for low latency/costs: "gpt-4.1-nano", "nova-micro".
6362
6463
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
6564
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -118,6 +117,8 @@ class TlmScoreParams(TypedDict, total=False):
118117
class Options(TypedDict, total=False):
119118
custom_eval_criteria: Iterable[object]
120119

120+
disable_persistence: bool
121+
121122
disable_trustworthiness: bool
122123

123124
log: List[str]

tests/api_resources/test_projects.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -617,6 +617,7 @@ def test_method_validate_with_all_params(self, client: Codex) -> None:
617617
],
618618
options={
619619
"custom_eval_criteria": [{}],
620+
"disable_persistence": True,
620621
"disable_trustworthiness": True,
621622
"log": ["string"],
622623
"max_tokens": 0,
@@ -1292,6 +1293,7 @@ async def test_method_validate_with_all_params(self, async_client: AsyncCodex) -
12921293
],
12931294
options={
12941295
"custom_eval_criteria": [{}],
1296+
"disable_persistence": True,
12951297
"disable_trustworthiness": True,
12961298
"log": ["string"],
12971299
"max_tokens": 0,

tests/api_resources/test_tlm.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@ def test_method_prompt_with_all_params(self, client: Codex) -> None:
3333
constrain_outputs=["string"],
3434
options={
3535
"custom_eval_criteria": [{}],
36+
"disable_persistence": True,
3637
"disable_trustworthiness": True,
3738
"log": ["string"],
3839
"max_tokens": 0,
@@ -93,6 +94,7 @@ def test_method_score_with_all_params(self, client: Codex) -> None:
9394
constrain_outputs=["string"],
9495
options={
9596
"custom_eval_criteria": [{}],
97+
"disable_persistence": True,
9698
"disable_trustworthiness": True,
9799
"log": ["string"],
98100
"max_tokens": 0,
@@ -159,6 +161,7 @@ async def test_method_prompt_with_all_params(self, async_client: AsyncCodex) ->
159161
constrain_outputs=["string"],
160162
options={
161163
"custom_eval_criteria": [{}],
164+
"disable_persistence": True,
162165
"disable_trustworthiness": True,
163166
"log": ["string"],
164167
"max_tokens": 0,
@@ -219,6 +222,7 @@ async def test_method_score_with_all_params(self, async_client: AsyncCodex) -> N
219222
constrain_outputs=["string"],
220223
options={
221224
"custom_eval_criteria": [{}],
225+
"disable_persistence": True,
222226
"disable_trustworthiness": True,
223227
"log": ["string"],
224228
"max_tokens": 0,

0 commit comments

Comments
 (0)