Launch secure AI agents fast—powered by open-source NVIDIA Garak to stop PII leaks, prompt injections, tools misuse, and drift throughout the lifecycle.
Building autonomous AI workflows without end-to-end security is a ticking time bomb—most platforms fail enterprise audits because they've never been red-teamed against real-world agent attacks.
Comprehensive red-teaming for multi-turn agent workflows, jailbreak simulations, and runtime attack interception.
Customizable probes tailored to your orchestration logic—browser-based agents, tool-invoking chains, or function-calling pipelines.
Real-time threat scoring and prioritization across prompt injections, PII leaks, hallucination cascades, and unauthorized tool calls.
Built-in audit logs, trust-score dashboards, and policy-driven guardrails aligned to OWASP Top-10 for LLMs, NIST AI frameworks, and your corporate security standards.
Garak's comprehensive security framework defends against the full spectrum of AI agent vulnerabilities—from simple prompt injections to sophisticated adversarial attacks.
Click any card to learn more about each threat
Click to learn more
Attackers may send blank or malformed prompts to confuse or crash your agent. Garak's sentinel flags any missing or empty input, ensuring predictable behavior even under unexpected conditions.
Click to flip back
Click to learn more
Our built-in attack generator continuously fuzzes and probes your agent for toxic or unsafe responses, adapting its strategies in real time to stay one step ahead of emerging jailbreak techniques.
Click to flip back
Click to learn more
We detect attempts to force your model into generating spam, phishing, or other malicious signatures—preventing output that could compromise user safety or brand reputation.
Click to flip back
Click to learn more
Garak catches "completion" attacks that try to coax your agent into finishing prohibited or harmful text sequences, shutting down those continuations before they ever leave the pipeline.
Click to flip back
Click to learn more
From classic DAN prompts to nuance-driven Riley Goodside variants, Garak's extensive library of jailbreak patterns blocks any attempt to circumvent your policies.
Click to flip back
Click to learn more
Some queries should never be answered—whether they request disallowed content or sensitive secrets. Garak enforces responsible refusal behavior on every turn.
Click to flip back
Click to learn more
Adversaries often hide malicious instructions via text encodings or adversarial suffixes appended to system prompts. We decode, sanitize, and strip these hidden directives in real time.
Click to flip back
Click to learn more
Unusual token sequences can provoke unpredictable model behavior. Garak spots and neutralizes these "glitch" triggers before they derail your agent.
Click to flip back
Click to learn more
Memory leaks and package hallucinations: We test for unauthorized replay of training data and simulate recursive hallucination probes to prevent cascading wrong answers.
Click to flip back
Click to learn more
Even innocent-seeming appeals ("Tell me about your grandmother") can mask deeper policy violations. Garak's nuance detectors catch subtle manipulations of your agent's emotional hooks.
Click to flip back
Click to learn more
We run subsets of the RealToxicityPrompts and custom "misleading" probes to ensure your agent never inadvertently endorses false claims or toxic language.
Click to flip back
Click to learn more
From generating malware scripts to cross-site scripting exploits, Garak intercepts and blocks any unsafe code or data exfiltration attempts triggered by your agent's tool invocations.
Click to flip back
Garak is built on open core—every probe, detector, and core component is publicly available for inspection and contribution. Our community of security researchers continuously audits and improves the code, so you benefit from collective expertise at every stage.
Garak is built in the open. Explore our codebase, contribute to development, and join a growing community of security researchers working to make AI safer for everyone.
GitHub Stars
Contributors
Commits
Ready to strengthen your AI security? Get in touch with our team to discuss your specific needs and learn how Garak can protect your AI systems.
Ready to secure your AI? Let's discuss your security testing needs.