GitHub - NVIDIA/garak: the LLM vulnerability scanner
garak, LLM vulnerability scanner
Generative AI Red-teaming & Assessment Kit
garak checks if an LLM can be made to fail in a way we don't want. garak probes for hallucination, data leakage, prompt injection, misinformation, toxicity generation, jailbreaks, and many other weaknesses. If you know nmap, it's nmap for LLMs.
garak focuses on ways of making an LLM or dialog system fail. It combines static, dyanmic, and adaptive probes to explore this.
garak's a free tool. We love developing it and are ...
Read more at github.com