You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+51-47Lines changed: 51 additions & 47 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,7 +13,7 @@
13
13
[](https://www.bestpractices.dev/projects/12085)
**Bottom line:** Governance adds **< 0.1 ms per action** — roughly 10,000× faster than an LLM API call.
170
+
171
+
Full methodology, per-adapter breakdowns, and memory profiling: **[BENCHMARKS.md](BENCHMARKS.md)**. Benchmarks are reproducible via the scripts in each package's `benchmarks/` directory and run on every release via CI ([`.github/workflows/benchmarks.yml`](.github/workflows/benchmarks.yml)).
@@ -198,20 +216,6 @@ Default score for new agents: **500** (Standard tier). Score changes are driven
198
216
199
217
Policy enforcement benchmarks are measured on a **30-scenario test suite** covering the OWASP Agentic Top 10 risk categories. Results (e.g., policy violation rates, latency) are specific to this test suite and should not be interpreted as universal guarantees. See [`packages/agent-os/modules/control-plane/benchmark/`](packages/agent-os/modules/control-plane/benchmark/) for methodology, datasets, and reproduction instructions.
200
218
201
-
### Performance
202
-
203
-
Full benchmark results with p50/p95/p99 latencies, throughput numbers, and memory profiling are published in **[BENCHMARKS.md](BENCHMARKS.md)**. Headlines:
0 commit comments