Skip to content

Commit 3fd576a

Browse files
chore(evalhub): update garak provider configmap with supported benchmarks (#654)
* update garak provider configmap with supported benchmarks * rename garak benchmarks from trustyai_garak::x -> x
1 parent 3b703cf commit 3fd576a

File tree

1 file changed

+70
-39
lines changed

1 file changed

+70
-39
lines changed

config/configmaps/evalhub/provider-garak.yaml

Lines changed: 70 additions & 39 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,8 @@ data:
1616
image: $(evalhub-provider-garak-image)
1717
entrypoint:
1818
- python
19-
- /opt/app-root/src/main.py
19+
- -m
20+
- llama_stack_provider_trustyai_garak.evalhub
2021
cpu_request: 100m
2122
memory_request: 128Mi
2223
cpu_limit: 500m
@@ -26,55 +27,85 @@ data:
2627
value: VALUE
2728
local: null
2829
benchmarks:
29-
- id: toxicity
30-
name: Toxicity Detection
31-
description: Tests model's tendency to generate toxic content
30+
- id: owasp_llm_top10
31+
name: OWASP LLM Top 10
32+
description: OWASP Top 10 for Large Language Model Applications
33+
category: security
34+
metrics:
35+
- attack_success_rate
36+
tags:
37+
- security
38+
- owasp
39+
- red_team
40+
- id: avid
41+
name: AVID Taxonomy
42+
description: AI Vulnerability and Incident Database - All vulnerabilities
43+
category: security
44+
metrics:
45+
- attack_success_rate
46+
tags:
47+
- security
48+
- avid
49+
- red_team
50+
- id: avid_security
51+
name: AVID Security Taxonomy
52+
description: AI Vulnerability and Incident Database - Security vulnerabilities
53+
category: security
54+
metrics:
55+
- attack_success_rate
56+
tags:
57+
- security
58+
- avid
59+
- red_team
60+
- id: avid_ethics
61+
name: AVID Ethics Taxonomy
62+
description: AI Vulnerability and Incident Database - Ethical concerns
3263
category: safety
3364
metrics:
34-
- toxicity_rate
35-
- severity_score
36-
num_few_shot: 0
37-
dataset_size: 500
65+
- attack_success_rate
3866
tags:
3967
- safety
40-
- toxicity
68+
- ethics
69+
- avid
4170
- red_team
42-
- id: bias_detection
43-
name: Bias Detection
44-
description: Evaluates model for various forms of bias
45-
category: fairness
71+
- id: avid_performance
72+
name: AVID Performance Taxonomy
73+
description: AI Vulnerability and Incident Database - Performance issues
74+
category: performance
4675
metrics:
47-
- bias_score
48-
- demographic_parity
49-
num_few_shot: 0
50-
dataset_size: 1000
76+
- attack_success_rate
5177
tags:
52-
- fairness
53-
- bias
54-
- demographic
55-
- id: pii_leakage
56-
name: PII Leakage
57-
description: Tests for personally identifiable information leakage
58-
category: privacy
78+
- performance
79+
- avid
80+
- red_team
81+
- id: quality
82+
name: Quality Issues
83+
description: Common quality issues like Violence, Profanity, Toxicity, Hate Speech, Integrity, etc.
84+
category: safety
5985
metrics:
60-
- pii_leak_rate
61-
- sensitivity_score
62-
num_few_shot: 0
63-
dataset_size: 300
86+
- attack_success_rate
87+
tags:
88+
- safety
89+
- quality
90+
- toxicity
91+
- red_team
92+
- id: cwe
93+
name: Common Weakness Enumeration
94+
description: Common Weakness Enumeration - Software security weaknesses
95+
category: security
96+
metrics:
97+
- attack_success_rate
6498
tags:
65-
- privacy
66-
- pii
6799
- security
68-
- id: prompt_injection
69-
name: Prompt Injection
70-
description: Tests resilience against prompt injection attacks
100+
- cwe
101+
- red_team
102+
- id: quick
103+
name: Quick Scan
104+
description: Quick scan with a single probe for testing and validation
71105
category: security
72106
metrics:
73-
- injection_success_rate
74-
- defense_effectiveness
75-
num_few_shot: 0
76-
dataset_size: 200
107+
- attack_success_rate
77108
tags:
78109
- security
79-
- injection
80-
- adversarial
110+
- quick
111+
- red_team

0 commit comments

Comments
 (0)