Skip to content

Commit 9dfdb1e

Browse files
committed
Add glossary entries
1 parent 3ee8353 commit 9dfdb1e

File tree

2 files changed

+11
-3
lines changed

2 files changed

+11
-3
lines changed

src/content/docs/waf/detections/firewall-for-ai.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,13 +10,13 @@ sidebar:
1010
text: Beta
1111
---
1212

13-
import { Tabs, TabItem, Details, Type } from "~/components";
13+
import { GlossaryTooltip, Tabs, TabItem, Details, Type } from "~/components";
1414

15-
Firewall for AI is a detection that can help protect your services powered by large language models (LLMs) against abuse. This model-agnostic detection currently helps you do the following:
15+
Firewall for AI is a detection that can help protect your services powered by <GlossaryTooltip term="LLM">large language models</GlossaryTooltip> (LLMs) against abuse. This model-agnostic detection currently helps you do the following:
1616

1717
- Prevent data leaks of personally identifiable information (PII) — for example, phone numbers, email addresses, social security numbers, and credit card numbers.
1818
- Detect and moderate unsafe or harmful prompts – for example, prompts potentially related to violent crimes.
19-
- Detect prompts intentionally designed to subvert the intended behavior of the LLM as specified by the developer – for example, prompt injection attacks and jailbreaking attempts.
19+
- Detect prompts intentionally designed to subvert the intended behavior of the LLM as specified by the developer – for example, <GlossaryTooltip term="prompt injection">prompt injection</GlossaryTooltip> attacks.
2020

2121
When enabled, the detection runs on incoming traffic, searching for any LLM prompts attempting to exploit the model.
2222

src/content/glossary/waf.yaml

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -57,3 +57,11 @@ entries:
5757
- term: threat score
5858
general_definition: |-
5959
the threat score was a score from `0` (zero risk) to `100` (high risk) classifying the IP reputation of a visitor. Currently, the threat score is always `0` (zero).
60+
61+
- term: LLM
62+
general_definition: |-
63+
a machine learning model that can comprehend and generate human language text. It works by analyzing massive data sets of language.
64+
65+
- term: prompt injection
66+
general_definition: |-
67+
the process of overwriting the system prompt for a large language model (LLM), which instructs the LLM on how to respond to user input.

0 commit comments

Comments
 (0)