Skip to content

Commit ba3ab97

Browse files
moved details
1 parent 451be8f commit ba3ab97

File tree

1 file changed

+15
-0
lines changed
  • src/content/docs/ai-gateway/guardrails

1 file changed

+15
-0
lines changed

src/content/docs/ai-gateway/guardrails/index.mdx

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,3 +16,18 @@ Guardrails proactively monitor interactions between users and AI models, giving
1616
- **Enhanced safety and user trust**: Proactively protect users from harmful or inappropriate interactions.
1717
- **Flexibility and control over allowed content**: Specify which categories to monitor and choose between flagging or outright blocking
1818
- **Auditing and compliance capabilities**: Stay ahead of evolving regulatory requirements with logs of user prompts, model responses, and enforced guardrails.
19+
20+
## How Guardrails work
21+
22+
AI Gateway inspects all interactions in real time by evaluating content against predefined safety parameters. Below a breakdown of the process:
23+
24+
1. Intercepting interactions:
25+
AI Gateway proxies requests and responses, sitting between the user and the AI model.
26+
27+
2. Inspecting content:
28+
29+
- User prompts: AI Gateway checks prompts against safety parameters (for example, violence, hate, or sexual content). Based on your settings, prompts can be flagged or blocked before reaching the model.
30+
- Model responses: Once processed, the AI model response is inspected. If hazardous content is detected, it can be flagged or blocked before being delivered to the user.
31+
32+
3. Applying actions:
33+
Depending on your configuration, flagged content is logged for review, while blocked content is prevented from proceeding.

0 commit comments

Comments
 (0)