Skip to content

Commit f42be08

Browse files
authored
Merge branch 'main' into allie/add-agentic-code-samples
2 parents 0f88ca3 + 275885c commit f42be08

File tree

5 files changed

+13
-11
lines changed

5 files changed

+13
-11
lines changed

2_0_vulns/LLM03_SupplyChain.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ A simple threat model can be found [here](https://github.com/jsotiro/ThreatModel
2828

2929
#### 4. Vulnerable Pre-Trained Model
3030

31-
Models are binary black boxes and unlike open source, static inspection can offer little to security assurances. Vulnerable pre-trained models can contain hidden biases, backdoors, or other malicious features that have not been identified through the safety evaluations of model repositories. Vulnerable models can be created by both poisoned datasets and direct model tampering using techniques such as ROME also known as lobotomisation.
31+
Models are binary black boxes and unlike open source, static inspection can offer little to no security assurances. Vulnerable pre-trained models can contain hidden biases, backdoors, or other malicious features that have not been identified through the safety evaluations of model repositories. Vulnerable models can be created by both poisoned datasets and direct model tampering using techniques such as ROME also known as lobotomisation.
3232

3333
#### 5. Weak Model Provenance
3434

@@ -40,7 +40,7 @@ A simple threat model can be found [here](https://github.com/jsotiro/ThreatModel
4040

4141
#### 7. Exploit Collaborative Development Processes
4242

43-
Collaborative model merge and model handling services (e.g. conversions) hosted in shared environments can be exploited to introduce vulnerabilities in shared models. Model merging is very popular on Hugging Face with model-merged models topping the OpenLLM leaderboard and can be exploited to bypass reviews. Similarly, services such as conversation bot have been proved to be vulnerable to manipulation and introduce malicious code in models.
43+
Collaborative model merge and model handling services (e.g. conversions) hosted in shared environments can be exploited to introduce vulnerabilities in shared models. Model merging is very popular on Hugging Face with model-merged models topping the OpenLLM leaderboard and can be exploited to bypass reviews. Similarly, services such as a conversation bot have been proved to be vulnerable to manipulation and introduce malicious code in models.
4444

4545
#### 8. LLM Model on Device supply-chain vulnerabilities
4646

2_0_vulns/LLM04_DataModelPoisoning.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -11,12 +11,12 @@ Moreover, models distributed through shared repositories or open-source platform
1111
### Common Examples of Vulnerability
1212

1313
1. Malicious actors introduce harmful data during training, leading to biased outputs. Techniques like "Split-View Data Poisoning" or "Frontrunning Poisoning" exploit model training dynamics to achieve this.
14-
(Ref. link: [Split-View Data Poisoning](https://github.com/GangGreenTemperTatum/speaking/blob/main/dc604/hacker-summer-camp-23/Ads%20_%20Poisoning%20Web%20Training%20Datasets%20_%20Flow%20Diagram%20-%20Exploit%201%20Split-View%20Data%20Poisoning.jpeg))
15-
(Ref. link: [Frontrunning Poisoning](https://github.com/GangGreenTemperTatum/speaking/blob/main/dc604/hacker-summer-camp-23/Ads%20_%20Poisoning%20Web%20Training%20Datasets%20_%20Flow%20Diagram%20-%20Exploit%202%20Frontrunning%20Data%20Poisoning.jpeg))
16-
2. Attackers can inject harmful content directly into the training process, compromising the model’s output quality.
17-
3. Users unknowingly inject sensitive or proprietary information during interactions, which could be exposed in subsequent outputs.
18-
4. Unverified training data increases the risk of biased or erroneous outputs.
19-
5. Lack of resource access restrictions may allow the ingestion of unsafe data, resulting in biased outputs.
14+
(Ref. link: [Split-View Data Poisoning](https://github.com/GangGreenTemperTatum/speaking/blob/aad68f8521119596abb567d94fbd10bdd652ac82/docs/conferences/dc604/hacker-summer-camp-23/Ads%20_%20Poisoning%20Web%20Training%20Datasets%20_%20Flow%20Diagram%20-%20Exploit%201%20Split-View%20Data%20Poisoning.jpeg))
15+
(Ref. link: [Frontrunning Poisoning](https://github.com/GangGreenTemperTatum/speaking/blob/aad68f8521119596abb567d94fbd10bdd652ac82/docs/conferences/dc604/hacker-summer-camp-23/Ads%20_%20Poisoning%20Web%20Training%20Datasets%20_%20Flow%20Diagram%20-%20Exploit%202%20Frontrunning%20Data%20Poisoning.jpeg))
16+
1. Attackers can inject harmful content directly into the training process, compromising the model’s output quality.
17+
2. Users unknowingly inject sensitive or proprietary information during interactions, which could be exposed in subsequent outputs.
18+
3. Unverified training data increases the risk of biased or erroneous outputs.
19+
4. Lack of resource access restrictions may allow the ingestion of unsafe data, resulting in biased outputs.
2020

2121
### Prevention and Mitigation Strategies
2222

2_0_vulns/LLM08_VectorAndEmbeddingWeaknesses.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44

55
Vectors and embeddings vulnerabilities present significant security risks in systems utilizing Retrieval Augmented Generation (RAG) with Large Language Models (LLMs). Weaknesses in how vectors and embeddings are generated, stored, or retrieved can be exploited by malicious actions (intentional or unintentional) to inject harmful content, manipulate model outputs, or access sensitive information.
66

7-
Retrieval Augmented Generation (RAG) is a model adaptation technique that enhances the performance and contextual relevance of responses from LLM Applications, by combining pre-trained language models with external knowledge sources.Retrieval Augmentation uses vector mechanisms and embedding. (Ref #1)
7+
Retrieval Augmented Generation (RAG) is a model adaptation technique that enhances the performance and contextual relevance of responses from LLM Applications, by combining pre-trained language models with external knowledge sources. Retrieval Augmentation uses vector mechanisms and embedding. (Ref #1)
88

99
### Common Examples of Risks
1010

CODEOWNERS

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
# TBC
66

77
## Either Ads or Steve can approve changes to CODEOWNERS or github:
8-
CODEOWNERS @GangGreenTemperTatum @virtualsteve-star
8+
CODEOWNERS @GangGreenTemperTatum @virtualsteve-star @rossja
99
.github/ @GangGreenTemperTatum @virtualsteve-star
1010

1111
## Data Gathering
@@ -43,4 +43,6 @@ _template.md @rossja
4343
# Sub-projects of the OWASP Top 10 for LLM Apps and Gen AI Project
4444

4545
## Agentic Security Initiative (ASI) under the Top 10 for LLM and Gen AI Project
46-
initiatives/agent_security_initiative/ @guerilla7 @allie-secvendors @hoeg @itskerenkatz
46+
initiatives/agent_security_initiative/ @guerilla7 @allie-secvendors @hoeg @itskerenkatz
47+
48+
initiatives/genai_red_team_handbook/ @rossja

initiatives/genai_red_team_handbook/.gitkeep

Whitespace-only changes.

0 commit comments

Comments
 (0)