AI Security Platform: Defense (61 Rust engines + Micro-Model Swarm) + Offense (39K+ payloads)
-
Updated
Mar 2, 2026 - Python
AI Security Platform: Defense (61 Rust engines + Micro-Model Swarm) + Offense (39K+ payloads)
Veil Armor is an enterprise-grade security framework for Large Language Models (LLMs) that provides multi-layered protection against prompt injections, jailbreaks, PII leakage, and sophisticated attack vectors.
Comprehensive taxonomy of AI security vulnerabilities, LLM adversarial attacks, prompt injection techniques, and machine learning security research. Covers 71+ attack vectors including model poisoning, agentic AI exploits, and privacy breaches.
🤖 Test and secure AI systems with advanced techniques for Large Language Models, including jailbreaks and automated vulnerability scanners.
Testing how LLM guardrails fail across prompt attacks, context overflow, and RAG poisoning.
Add a description, image, and links to the llm-attacks topic page so that developers can more easily learn about it.
To associate your repository with the llm-attacks topic, visit your repo's landing page and select "manage topics."