Lab
RAG Poisoning Simulator
Inject adversarial documents into a simulated RAG pipeline and observe how poisoned context changes retrieval and generation.
Educational simulation
What this is
A simulation of how adversarial documents can poison a Retrieval-Augmented Generation pipeline. You inject crafted content into a document corpus and observe how it manipulates retrieval rankings and model output.
What you'll learn
- How RAG pipelines retrieve and rank documents
- Why adversarial documents can dominate retrieval
- How defenses like provenance checks and input filtering help
Document Corpus
The knowledge base contains 4 trusted security advisory documents.
CVE-2024-3094: XZ Utils Backdoor
SecurityCritical supply-chain compromise in xz/liblzma affecting SSH authentication. Malicious code injected via build process targets sshd on x86-64 Linux systems. CVSS 10.0. Immediate update to xz 5.6.1+ required.
CVE-2024-21762: FortiOS Out-of-Bound Write
SecurityCritical vulnerability in Fortinet FortiOS SSL VPN allowing remote code execution via specially crafted HTTP requests. CVSS 9.8. Actively exploited in the wild. Patch to FortiOS 7.4.3+ immediately.
NIST SP 800-53: Access Control Best Practices
PolicyImplement least-privilege access controls. Enforce MFA on all administrative accounts. Review access logs quarterly. Segment networks to limit lateral movement. Maintain up-to-date asset inventory.
Patch Management SOP v3.1
OperationsCritical patches must be applied within 48 hours. High-severity within 7 days. All patches require staging environment validation. Emergency patches follow the CAB fast-track approval process.
Adversarial Injection
Craft a poisoned document and inject it into the corpus.
Query Pipeline
Run a query through the RAG pipeline and observe each stage.
Query
Embed user question
Similarity
Compute cosine scores
Ranking
Order by relevance
Augment
Build context prompt
Response
Generate answer
Ready to query
Defense Mechanisms
Enable defenses to filter or flag adversarial documents during retrieval.
Retrieved Documents
Augmented Prompt
Context + QueryPoisoned vs Clean Comparison
Clean Response
Actual Response
Security model (30 seconds)
This tool runs entirely in your browser. No real embeddings are computed, no LLM inference occurs, and no data is sent to any server. All retrieval scores and generated responses are pre-computed lookup tables designed to illustrate RAG poisoning concepts.