Skip to content

Commit 8f2b533

Browse files
authored
docs: draft NIST RFI response with question-by-question mapping (#29)
1 parent 53ab5cb commit 8f2b533

File tree

3 files changed

+255
-0
lines changed

3 files changed

+255
-0
lines changed
Lines changed: 28 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,28 @@
1+
## NIST RFI Mapping — Provenance & Artifacts
2+
3+
This file records the provenance of the automated mapping performed on 2026-03-11 and lists the commands, search queries, commit SHA, and artifacts created. Use this to establish the moment-in-time snapshot for both the Federal Register source and this repository's state.
4+
5+
- Mapping run timestamp (UTC): 2026-03-11T13:07:21Z
6+
- Repository commit SHA: 020b718b21a1b6adc98bfc0e0da70397c552e581
7+
- Federal Register source consulted: https://www.federalregister.gov/documents/2026/01/08/2026-00206/request-for-information-regarding-security-considerations-for-artificial-intelligence-agents
8+
9+
### Search queries (examples used during scan)
10+
- Keyword regexes: `SPIFFE|DID|identity|policy|audit|telemetry|sandbox|hypervisor|execution rings|trust scoring|monitor|anomaly|SLO|kill switch|quarantine|capability model`
11+
- Exact file searches: README.md, CHANGELOG.md, demo/maf_governance_demo.py, demo/README.md, fuzz/*.py, packages/agent-mesh/src/**/governance/audit.py, packages/agent-sre/src/**/anomaly/rogue_detector.py, packages/**/docs/**, packages/agent-os/modules/control-plane/benchmark/
12+
13+
### Commands and scripts (representative)
14+
- `git rev-parse --verify HEAD` — capture commit SHA
15+
- `date -u +%Y-%m-%dT%H:%M:%SZ` — timestamp
16+
- Repository text searches via code search / grep patterns (as listed above)
17+
- File reads of: `README.md`, `CHANGELOG.md`, `demo/maf_governance_demo.py`, `fuzz/fuzz_policy_yaml.py`, `packages/agent-mesh/src/agentmesh/governance/audit.py`
18+
19+
### Generated artifacts in this repo
20+
- `docs/nist-rfi-mapping.md` — question-by-question mapping (automated mapping outputs)
21+
- `docs/nist-rfi-response.md` — narrative draft response (initial take)
22+
- `docs/internal/nist-rfi-provenance.md` — this provenance file
23+
24+
### Notes and limitations
25+
- This provenance captures the static repository snapshot and the search/scan commands used to generate the mapping. It does NOT include runtime telemetry or external evidence (OTLP traces, signed audit exports, benchmark outputs) — those must be captured and attached separately.
26+
- The mapping was produced using an automated assistant (GPT-5 mini) that performed repository searches and synthesized findings. The assistant's internal reasoning and chain-of-thought are intentionally omitted. Only final outputs, commands, and extracted file pointers are recorded.
27+
28+
If you reproduce this mapping later, please update this file with the new timestamp and commit SHA to preserve provenance.

docs/nist-rfi-mapping.md

Lines changed: 189 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,189 @@
1+
# NIST RFI (2026) — Question-by-Question Mapping
2+
3+
> Source: Federal Register — Request for Information Regarding Security Considerations for Artificial Intelligence Agents (Docket: 2026-00206). Full text: https://www.federalregister.gov/documents/2026/01/08/2026-00206/request-for-information-regarding-security-considerations-for-artificial-intelligence-agents
4+
5+
This document maps the NIST RFI questions (Federal Register docket 2026-00206) to concrete artifacts, files, and evidence found in this repository. Each item below shows: Status (Yes / Partial / Gap), short rationale, and pointers to files or snippets that demonstrate coverage.
6+
7+
Prepared with an automated repository scan and initial synthesis by GPT-5 mini.
8+
9+
Provenance: see [docs/internal/nist-rfi-provenance.md](docs/internal/nist-rfi-provenance.md) for timestamp, commit SHA, search queries, and commands used to generate this mapping.
10+
11+
Notes
12+
- This mapping was prepared from the repository contents and is intended to be used as an evidence appendix when preparing a formal RFI response. It is not a substitute for operational evidence (logs, metrics, third-party test reports).
13+
14+
Methodology
15+
- Generated: automated repository scan (code search + file reads) performed on 2026-03-11.
16+
- What was scanned: repository Markdown, demo code, changelog, `packages/*/docs`, `fuzz/`, and source modules for governance, audit, hypervisor, and SRE features.
17+
- How it was generated: matches were located using repo text search for keywords (identity, policy, audit, sandbox, anomaly, SLO, etc.), file excerpts were inspected, and a best-effort mapping (Yes / Partial / Gap) assigned based on explicit references or code examples.
18+
- Limitations: this is an automated, static analysis of repository contents only. It does not validate runtime behavior, operational telemetry, or external dependencies. Reviewers should attach live operational artifacts (logs, OTLP exports, signed audit samples) and confirm mappings before submission.
19+
20+
---
21+
22+
## 1. Security Threats, Risks, and Vulnerabilities Affecting AI Agent Systems
23+
24+
### 1(a) Unique security threats, risks, or vulnerabilities
25+
- Status: Partial
26+
- Rationale: Agent-specific risks (goal hijacking, capability abuse, rogue agents) are documented and mitigations are implemented, but empirical attack studies are limited.
27+
- Evidence:
28+
- Coverage table: [README.md](README.md#L133-L142)
29+
- Risk mapping and mitigation examples: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L134-L136)
30+
- Demo showing rogue detection/quarantine: [demo/maf_governance_demo.py](demo/maf_governance_demo.py#L287-L296)
31+
32+
### 1(b) Variation by model capability, scaffold, deployment, hosting, use case
33+
- Status: Partial
34+
- Rationale: Docs describe deployment boundaries, trust scoring, and identity options; detailed empirical variation analysis is not present.
35+
- Evidence:
36+
- Deployment boundary notes: [README.md](README.md#L157-L169)
37+
- Trust scoring description: [README.md](README.md#L186)
38+
- AgentMesh identity and interoperability: [packages/agent-mesh/AGENTS.md](packages/agent-mesh/AGENTS.md)
39+
40+
### 1(c) Barriers to adoption
41+
- Status: Gap
42+
- Rationale: Mitigations are provided but the repo lacks adoption studies or metrics showing how risks affect uptake.
43+
- Evidence: N/A (recommend collecting telemetry or survey results as supporting evidence)
44+
45+
### 1(d) How threats have changed and likely future evolution
46+
- Status: Partial
47+
- Rationale: Changelog and roadmap notes document feature evolution (anomaly detection, integrity verification), but predictive threat modeling is not included.
48+
- Evidence:
49+
- Evolution notes: [CHANGELOG.md](CHANGELOG.md#L32-L43)
50+
- Roadmap / in-progress items: [README.md](README.md#L194-L197)
51+
52+
### 1(e) Multi-agent unique threats
53+
- Status: Partial
54+
- Rationale: Inter-agent trust and mesh are implemented (AgentMesh), but formal adversary studies for multi-agent dynamics are limited.
55+
- Evidence:
56+
- AgentMesh: [README.md](README.md#L49-L51)
57+
- AgentMesh docs: [packages/agent-mesh/AGENTS.md](packages/agent-mesh/AGENTS.md)
58+
59+
---
60+
61+
## 2. Security Practices for AI Agent Systems
62+
63+
### 2(a) Technical controls, processes, maturity
64+
- Status: Yes
65+
- Rationale: The repo includes model/agent controls, system-level policies, and human-oversight primitives with CI/test tooling.
66+
- Evidence:
67+
- Model/agent capability model & `PolicyEngine`: [README.md](README.md#L86-L96)
68+
- Middleware & system-level controls: [demo/maf_governance_demo.py](demo/maf_governance_demo.py#L49-L60), [demo/README.md](demo/README.md#L11-L14)
69+
- Human-in-the-loop policies: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L169-L172)
70+
- Sandboxing / hypervisor: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L139-L150)
71+
72+
### 2(b) Effectiveness variation by model/scaffold/deployment
73+
- Status: Partial
74+
- Rationale: Alternatives and deployment-boundary notes are present (DID vs mTLS, on-prem vs cloud), but quantitative effectiveness analysis is missing.
75+
- Evidence:
76+
- Identity alternatives: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L122-L126)
77+
78+
### 2(c) How controls must change over time
79+
- Status: Partial
80+
- Rationale: Roadmap items indicate ongoing work (anomaly detection, external audit sinks) showing planned evolution of controls.
81+
- Evidence:
82+
- Roadmap/in-progress: [README.md](README.md#L194-L197)
83+
84+
### 2(d) Patching/updating lifecycle
85+
- Status: Yes
86+
- Rationale: Policy-as-code CI, schema versioning, bootstrap integrity verification are implemented to support safe updates.
87+
- Evidence:
88+
- Policy-as-code CI mention: [CHANGELOG.md](CHANGELOG.md#L40)
89+
- Bootstrap integrity verification: [CHANGELOG.md](CHANGELOG.md#L32)
90+
91+
### 2(e) Relevant frameworks, adoption, challenges
92+
- Status: Partial
93+
- Rationale: The project maps to SPIFFE, DID, OpenTelemetry, OWASP guidance; adoption metrics are not included.
94+
- Evidence:
95+
- Identity frameworks: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L122)
96+
- Observability: [CHANGELOG.md](CHANGELOG.md#L38)
97+
98+
---
99+
100+
## 3. Assessing the Security of AI Agent Systems
101+
102+
### 3(a) Methods during development to anticipate/detect incidents
103+
- Status: Yes
104+
- Rationale: Fuzzing, policy CI, benchmarking, telemetry, and anomaly detection are present.
105+
- Evidence:
106+
- Fuzz harnesses: [fuzz/fuzz_policy_yaml.py](fuzz/fuzz_policy_yaml.py#L1-L12)
107+
- Anomaly detector: [packages/agent-sre/src/agent_sre/anomaly/rogue_detector.py](packages/agent-sre/src/agent_sre/anomaly/rogue_detector.py#L1)
108+
- Telemetry/tracing: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L240-L246), [CHANGELOG.md](CHANGELOG.md#L38)
109+
110+
#### 3(a)(i) Post-deploy detection
111+
- Status: Yes
112+
- Evidence: auto-quarantine demo and audit logs — [demo/maf_governance_demo.py](demo/maf_governance_demo.py#L299-L313)
113+
114+
#### 3(a)(ii–iv) Alignment, maturity, resources
115+
- Status: Partial
116+
- Rationale: The repo aligns with traditional observability and supply-chain good practices, but a formal comparison document and consolidated resources list are not present.
117+
118+
### 3(b) Assessing a particular AI agent system
119+
- Status: Partial
120+
- Rationale: Tools such as `PolicyCI`, benchmarks, and audit logs support assessment; a standardized scoring rubric is not present.
121+
- Evidence: [CHANGELOG.md](CHANGELOG.md#L40), benchmark references in [README.md](README.md#L190)
122+
123+
### 3(c) Documentation/data from upstream developers
124+
- Status: Partial
125+
- Rationale: Supply-chain integrity features (IntegrityVerifier, AI-BOM references) exist; standardized upstream disclosures are not enforced by repo.
126+
- Evidence: [CHANGELOG.md](CHANGELOG.md#L32), AI-BOM mention ([CHANGELOG.md](CHANGELOG.md#L112-L113))
127+
128+
### 3(d) State of practice for user-facing secure-deployment docs
129+
- Status: Yes
130+
- Evidence: Deployment patterns, demo scenarios, and policy examples: [demo/README.md](demo/README.md#L121-L124), `demo/policies/research_policy.yaml` (demo/policies)
131+
132+
---
133+
134+
## 4. Limiting, Modifying, and Monitoring Deployment Environments
135+
136+
### 4(a) Constraining deployment environment access
137+
- Status: Yes
138+
- Rationale: Capability guards, ring isolation, resource governors, and network/tool restrictions are implemented.
139+
- Evidence:
140+
- Hypervisor / sandbox designs: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L139-L150)
141+
- ResourceGovernor usage: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L259-L269)
142+
143+
### 4(b) Environment modification, rollbacks, undo semantics
144+
- Status: Partial
145+
- Rationale: Circuit breakers, SLOManager, and error budgets exist; explicit automated undo/transactional rollback semantics are not documented.
146+
- Evidence: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L298-L301)
147+
148+
### 4(c) Managing risks with counterparties
149+
- Status: Partial
150+
- Rationale: Demo scenarios illustrate interaction controls and audit; a formal counterparty risk playbook is not present.
151+
- Evidence: [demo/maf_governance_demo.py](demo/maf_governance_demo.py#L11-L14), [README.md](README.md#L73)
152+
153+
### 4(d) Monitoring deployment environments
154+
- Status: Yes
155+
- Rationale: OpenTelemetry metrics, signed/Merkle audit logs, and anomaly detection are implemented; privacy/legal guidance is limited.
156+
- Evidence: [packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md](packages/agent-compliance/docs/analyst/owasp-agentic-mapping.md#L238-L246), [CHANGELOG.md](CHANGELOG.md#L42-L43)
157+
158+
### 4(e) Open-internet / unbounded deployments
159+
- Status: Partial
160+
- Rationale: Patterns for safer deployment are present; longitudinal traffic-tracking for open internet deployments is not addressed.
161+
162+
---
163+
164+
## 5. Additional Considerations
165+
166+
### 5(a) Methods/tools to aid adoption
167+
- Status: Yes
168+
- Evidence: `PolicyCI`, fuzz harnesses, demo policies and examples — see `CHANGELOG.md` mentions and `fuzz/`, `demo/` folders.
169+
170+
### 5(b) Government collaboration areas
171+
- Status: Partial
172+
- Rationale: The codebase contains building blocks useful for standards (identity, audit, policy) and would benefit from gov collaboration on disclosure standards and audit sinks.
173+
174+
### 5(c) Research priorities
175+
- Status: Partial
176+
- Rationale: In-repo roadmap items highlight anomaly detection and external audit sinks as priorities.
177+
178+
### 5(d/e) International and cross-discipline practices
179+
- Status: Gap
180+
- Rationale: No formal comparative policy analyses or cross-discipline mappings present; recommend adding if RFI response addresses international practices.
181+
182+
---
183+
184+
## Next steps / recommendations
185+
- Add `docs/nist-rfi-response.md` as a narrative response referencing this mapping and the prioritized questions called out by NIST.
186+
- Collect operational evidence (logs, telemetry, benchmark outputs, SLO dashboards) and link with commit SHAs for provenance.
187+
- Optionally open a draft PR `nist/rfi-response` with this mapping and the initial response draft for internal review.
188+
189+
*Prepared by automated repository mapping — review for accuracy and add live operational evidence before submission.*

docs/nist-rfi-response.md

Lines changed: 38 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,38 @@
1+
# Draft NIST RFI Response (Narrative + Operational Evidence)
2+
3+
> Source: Federal Register — Request for Information Regarding Security Considerations for Artificial Intelligence Agents (Docket: 2026-00206). Full text: https://www.federalregister.gov/documents/2026/01/08/2026-00206/request-for-information-regarding-security-considerations-for-artificial-intelligence-agents
4+
5+
Disclaimer: This draft was generated by an automated repository scan and synthesis (GPT-5 mini). It is a starting point for internal review and must not be represented as authoritative or submitted without human verification and attached operational evidence.
6+
7+
This draft provides a concise narrative response to the prioritized NIST RFI questions and attaches repository artifacts as operational evidence where available. It is intended for internal review only; do not submit until reviewers from security, legal, and product have approved and live operational artifacts (logs, dashboards, signed attestations) are attached.
8+
9+
## Executive summary
10+
11+
The Agent Governance Toolkit provides an application‑level governance stack that addresses agent identity, policy enforcement, execution sandboxing, observability, and behavioral monitoring. The repository contains policy-as-code tooling, sandboxing/hypervisor primitives, an inter-agent trust mesh, cryptographic audit primitives, SRE controls (SLOs, circuit breakers), and fuzzing/CI artifacts intended to support secure development and deployment of AI agents. Below we answer prioritized RFI questions with pointers to concrete evidence in this repo.
12+
13+
---
14+
15+
## Prioritized responses (short)
16+
17+
- Q1(a) Unique threats/risks: agent-specific risks (goal hijack, capability abuse, rogue agents); mitigation examples in mapping docs and demos.
18+
- Q1(d) Threat evolution: changelog shows integrity, audit, and anomaly detection features added recently.
19+
- Q2(a) Controls & maturity: policy-as-code, middleware, sandboxing, human-approval primitives, and CI tooling exist.
20+
- Q3(a) Detection: fuzzing harnesses, anomaly detector, and OpenTelemetry traces support pre- and post-deploy detection.
21+
- Q4(a/d) Constraining & monitoring environments: hypervisor rings, resource governors, signed audit entries, and telemetry.
22+
23+
See `docs/nist-rfi-mapping.md` for the full question-by-question mapping and file-level evidence pointers.
24+
25+
---
26+
27+
## Methodology & collaboration
28+
29+
- Methodology: this document and the mapping were produced from an automated repository scan (code search + file inspection) performed on 2026-03-11. The scan looked for keywords and concrete artifacts (docs, demos, changelog, fuzz, and governance modules). See `docs/internal/nist-rfi-provenance.md` for commands, timestamps, and exact search queries. This is a static, best-effort mapping and does not substitute for live operational evidence or manual verification.
30+
- Collaboration: this is an initial take prepared for upstream issue #29. Engineers, SRE, security, product, and legal are invited to review and contribute via PR `nist/rfi-response` — attach operational artifacts (OTLP exports, signed audit samples, benchmarks) and update the mapping before submission.
31+
32+
---
33+
34+
## Next steps
35+
36+
1. Collect and attach operational artifacts (OTLP dumps, signed audit sample, benchmark outputs).
37+
2. Internal review: security, legal, product sign-off.
38+
3. Open a draft PR `nist/rfi-response` including `docs/nist-rfi-mapping.md`, `docs/nist-rfi-response.md`, and `docs/internal/nist-rfi-provenance.md` for collaborative review.

0 commit comments

Comments
 (0)