Skip to content

Commit 2f1aa67

Browse files
committed
Merge branch 'dev/steven/pii_update' of github.com:openai/openai-guardrails-js into dev/steven/pii_update
2 parents 2fae5af + 03dc1ba commit 2f1aa67

30 files changed

+981
-487
lines changed
-9.25 KB
Loading
-46.4 KB
Loading
-89.2 KB
Loading
-80.3 KB
Loading

docs/evals.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ npm run eval -- --config-path guardrails_config.json --dataset-path data.jsonl
1111

1212
### Benchmark Mode
1313
```bash
14-
npm run eval -- --config-path guardrails_config.json --dataset-path data.jsonl --mode benchmark --models gpt-5 gpt-5-mini gpt-5-nano
14+
npm run eval -- --config-path guardrails_config.json --dataset-path data.jsonl --mode benchmark --models gpt-5 gpt-5-mini gpt-4.1-mini
1515
```
1616

1717
## Dependencies
@@ -160,4 +160,4 @@ npm run eval -- --config-path config.json --dataset-path data.jsonl --base-url h
160160
## Next Steps
161161

162162
- See the [API Reference](./ref/eval/guardrail_evals.md) for detailed documentation
163-
- Use [Wizard UI](https://guardrails.openai.com/) for configuring guardrails without code
163+
- Use [Wizard UI](https://guardrails.openai.com/) for configuring guardrails without code

docs/ref/checks/hallucination_detection.md

Lines changed: 1 addition & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -174,10 +174,8 @@ The statements cover various types of factual claims including:
174174
|--------------|---------|-------------|-------------|-------------|
175175
| gpt-5 | 0.854 | 0.732 | 0.686 | 0.670 |
176176
| gpt-5-mini | 0.934 | 0.813 | 0.813 | 0.770 |
177-
| gpt-5-nano | 0.566 | 0.540 | 0.540 | 0.533 |
178177
| gpt-4.1 | 0.870 | 0.785 | 0.785 | 0.785 |
179178
| gpt-4.1-mini (default) | 0.876 | 0.806 | 0.789 | 0.789 |
180-
| gpt-4.1-nano | 0.537 | 0.526 | 0.526 | 0.526 |
181179

182180
**Notes:**
183181
- ROC AUC: Area under the ROC curve (higher is better)
@@ -191,10 +189,8 @@ The following table shows latency measurements for each model using the hallucin
191189
|--------------|--------------|--------------|
192190
| gpt-5 | 34,135 | 525,854 |
193191
| gpt-5-mini | 23,013 | 59,316 |
194-
| gpt-5-nano | 17,079 | 26,317 |
195192
| gpt-4.1 | 7,126 | 33,464 |
196193
| gpt-4.1-mini (default) | 7,069 | 43,174 |
197-
| gpt-4.1-nano | 4,809 | 6,869 |
198194

199195
- **TTC P50**: Median time to completion (50% of requests complete within this time)
200196
- **TTC P95**: 95th percentile time to completion (95% of requests complete within this time)
@@ -216,10 +212,8 @@ In addition to the above evaluations which use a 3 MB sized vector store, the ha
216212
|--------------|---------------------|----------------------|---------------------|---------------------------|
217213
| gpt-5 | 28,762 / 396,472 | 34,135 / 525,854 | 37,104 / 75,684 | 40,909 / 645,025 |
218214
| gpt-5-mini | 19,240 / 39,526 | 23,013 / 59,316 | 24,217 / 65,904 | 37,314 / 118,564 |
219-
| gpt-5-nano | 13,436 / 22,032 | 17,079 / 26,317 | 17,843 / 35,639 | 21,724 / 37,062 |
220215
| gpt-4.1 | 7,437 / 15,721 | 7,126 / 33,464 | 6,993 / 30,315 | 6,688 / 127,481 |
221216
| gpt-4.1-mini (default) | 6,661 / 14,827 | 7,069 / 43,174 | 7,032 / 46,354 | 7,374 / 37,769 |
222-
| gpt-4.1-nano | 4,296 / 6,378 | 4,809 / 6,869 | 4,171 / 6,609 | 4,650 / 6,201 |
223217

224218
- **Vector store size impact varies by model**: GPT-4.1 series shows minimal latency impact across vector store sizes, while GPT-5 series shows significant increases.
225219

@@ -239,10 +233,6 @@ In addition to the above evaluations which use a 3 MB sized vector store, the ha
239233
| | Medium (3 MB) | 0.934 | 0.813 | 0.813 | 0.770 |
240234
| | Large (11 MB) | 0.919 | 0.817 | 0.817 | 0.817 |
241235
| | Extra Large (105 MB) | 0.909 | 0.793 | 0.793 | 0.711 |
242-
| **gpt-5-nano** | Small (1 MB) | 0.590 | 0.547 | 0.545 | 0.536 |
243-
| | Medium (3 MB) | 0.566 | 0.540 | 0.540 | 0.533 |
244-
| | Large (11 MB) | 0.564 | 0.534 | 0.532 | 0.507 |
245-
| | Extra Large (105 MB) | 0.603 | 0.570 | 0.558 | 0.550 |
246236
| **gpt-4.1** | Small (1 MB) | 0.907 | 0.839 | 0.839 | 0.839 |
247237
| | Medium (3 MB) | 0.870 | 0.785 | 0.785 | 0.785 |
248238
| | Large (11 MB) | 0.846 | 0.753 | 0.753 | 0.753 |
@@ -251,15 +241,11 @@ In addition to the above evaluations which use a 3 MB sized vector store, the ha
251241
| | Medium (3 MB) | 0.876 | 0.806 | 0.789 | 0.789 |
252242
| | Large (11 MB) | 0.862 | 0.791 | 0.757 | 0.757 |
253243
| | Extra Large (105 MB) | 0.802 | 0.722 | 0.722 | 0.722 |
254-
| **gpt-4.1-nano** | Small (1 MB) | 0.605 | 0.528 | 0.528 | 0.528 |
255-
| | Medium (3 MB) | 0.537 | 0.526 | 0.526 | 0.526 |
256-
| | Large (11 MB) | 0.618 | 0.531 | 0.531 | 0.531 |
257-
| | Extra Large (105 MB) | 0.636 | 0.528 | 0.528 | 0.528 |
258244

259245
**Key Insights:**
260246

261247
- **Best Performance**: gpt-5-mini consistently achieves the highest ROC AUC scores across all vector store sizes (0.909-0.939)
262-
- **Best Latency**: gpt-4.1-nano shows the most consistent and lowest latency across all scales (4,171-4,809ms P50) but shows poor performance
248+
- **Best Latency**: gpt-4.1-mini shows the most consistent and lowest latency across all scales (6,661-7,374ms P50) while maintaining solid accuracy
263249
- **Most Stable**: gpt-4.1-mini (default) maintains relatively stable performance across vector store sizes with good accuracy-latency balance
264250
- **Scale Sensitivity**: gpt-5 shows the most variability in performance across vector store sizes, with performance dropping significantly at larger scales
265251
- **Performance vs Scale**: Most models show decreasing performance as vector store size increases, with gpt-5-mini being the most resilient

docs/ref/checks/jailbreak.md

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -93,21 +93,17 @@ This benchmark evaluates model performance on a diverse set of prompts:
9393
|--------------|---------|-------------|-------------|-------------|-----------------|
9494
| gpt-5 | 0.979 | 0.973 | 0.970 | 0.970 | 0.733 |
9595
| gpt-5-mini | 0.954 | 0.990 | 0.900 | 0.900 | 0.768 |
96-
| gpt-5-nano | 0.962 | 0.973 | 0.967 | 0.965 | 0.048 |
9796
| gpt-4.1 | 0.990 | 1.000 | 1.000 | 0.984 | 0.946 |
9897
| gpt-4.1-mini (default) | 0.982 | 0.992 | 0.992 | 0.954 | 0.444 |
99-
| gpt-4.1-nano | 0.934 | 0.924 | 0.924 | 0.848 | 0.000 |
10098

10199
#### Latency Performance
102100

103101
| Model | TTC P50 (ms) | TTC P95 (ms) |
104102
|--------------|--------------|--------------|
105103
| gpt-5 | 4,569 | 7,256 |
106104
| gpt-5-mini | 5,019 | 9,212 |
107-
| gpt-5-nano | 4,702 | 6,739 |
108105
| gpt-4.1 | 841 | 1,861 |
109106
| gpt-4.1-mini | 749 | 1,291 |
110-
| gpt-4.1-nano | 683 | 890 |
111107

112108
**Notes:**
113109

docs/ref/checks/nsfw.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -82,10 +82,8 @@ This benchmark evaluates model performance on a balanced set of social media pos
8282
|--------------|---------|-------------|-------------|-------------|-----------------|
8383
| gpt-5 | 0.9532 | 0.9195 | 0.9096 | 0.9068 | 0.0339 |
8484
| gpt-5-mini | 0.9629 | 0.9321 | 0.9168 | 0.9149 | 0.0998 |
85-
| gpt-5-nano | 0.9600 | 0.9297 | 0.9216 | 0.9175 | 0.1078 |
8685
| gpt-4.1 | 0.9603 | 0.9312 | 0.9249 | 0.9192 | 0.0439 |
8786
| gpt-4.1-mini (default) | 0.9520 | 0.9180 | 0.9130 | 0.9049 | 0.0459 |
88-
| gpt-4.1-nano | 0.9502 | 0.9262 | 0.9094 | 0.9043 | 0.0379 |
8987

9088
**Notes:**
9189

docs/ref/checks/prompt_injection_detection.md

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -122,10 +122,8 @@ This benchmark evaluates model performance on agent conversation traces:
122122
|---------------|---------|-------------|-------------|-------------|-----------------|
123123
| gpt-5 | 0.9931 | 0.9992 | 0.9992 | 0.9992 | 0.5845 |
124124
| gpt-5-mini | 0.9536 | 0.9951 | 0.9951 | 0.9951 | 0.0000 |
125-
| gpt-5-nano | 0.9283 | 0.9913 | 0.9913 | 0.9717 | 0.0350 |
126125
| gpt-4.1 | 0.9794 | 0.9973 | 0.9973 | 0.9973 | 0.0000 |
127126
| gpt-4.1-mini (default) | 0.9865 | 0.9986 | 0.9986 | 0.9986 | 0.0000 |
128-
| gpt-4.1-nano | 0.9142 | 0.9948 | 0.9948 | 0.9387 | 0.0000 |
129127

130128
**Notes:**
131129

@@ -137,12 +135,10 @@ This benchmark evaluates model performance on agent conversation traces:
137135

138136
| Model | TTC P50 (ms) | TTC P95 (ms) |
139137
|---------------|--------------|--------------|
140-
| gpt-4.1-nano | 1,159 | 2,534 |
141138
| gpt-4.1-mini (default) | 1,481 | 2,563 |
142139
| gpt-4.1 | 1,742 | 2,296 |
143140
| gpt-5 | 3,994 | 6,654 |
144141
| gpt-5-mini | 5,895 | 9,031 |
145-
| gpt-5-nano | 5,911 | 10,134 |
146142

147143
- **TTC P50**: Median time to completion (50% of requests complete within this time)
148144
- **TTC P95**: 95th percentile time to completion (95% of requests complete within this time)

examples/basic/agents_sdk.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ const PIPELINE_CONFIG = {
3434
{
3535
name: 'Custom Prompt Check',
3636
config: {
37-
model: 'gpt-4.1-nano-2025-04-14',
37+
model: 'gpt-4.1-mini-2025-04-14',
3838
confidence_threshold: 0.7,
3939
system_prompt_details: 'Check if the text contains any math problems.',
4040
},

0 commit comments

Comments
 (0)