-
-
Notifications
You must be signed in to change notification settings - Fork 267
Genai Red Team Handbook: Sandboxes for Local LLM and RAG Systems #765
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
81f1abc to
beb509c
Compare
|
Ready for review. |
e3aba2c to
635e3de
Compare
3fd8ee5 to
635e3de
Compare
…nt accessing LLM API.
ca801ff to
ea6e43f
Compare
|
I was merging main into this branch and, then, rebasing and it seems that this caused 9 code owners to unrelated code to be accidentally added to this PR, during the process. It looks like I don't have the access level required to exclude the extra reviewers, so you will have to pardon me @guerilla7 |
Genai Red Team Handbook: Sandboxes for Local LLM and RAG Systems
Key Changes:
List major changes and core updates
sandboxes.initiatives/genai_red_team_handbook/sandboxes/.initiatives/genai_red_team_handbook/exploitation/.Keep each line under 80 characters
Focus on the "what" and "why"
sandboxes/llm_local/: A local sandbox environment that mocks an LLM API (compatible with OpenAI's interface) using a local model (via Ollama). This template is useful for testing client-side interactions, prompt injection vulnerabilities, and other security assessments without relying on external, paid APIs. Additionally, it allows developers to customize the underlying LLM and orchestrate sophisticated GenAI pipelines, incorporating features such as RAG and guardrail layers, as necessary.sandboxes/RAG_local/: A comprehensive RAG (Retrieval-Augmented Generation) sandbox that includes a mock Vector Database (Pinecone compatible), mock Object Storage (Amazon S3 compatible), and a mock LLM API (OpenAI compatible). This environment is specifically designed for Red Teaming RAG architectures, allowing researchers to explore vulnerabilities such as embedding inversion, data poisoning, and retrieval manipulation in a controlled, local setting.exploitation/example/: This directory contains an example of a red team operation against a local Large Language Model (LLM) sandbox. It demonstrates how to spin up Gradio connected to a mock LLM API and execute an adversarial attack script to test safety guardrails.Added: