# AI

Tools and resources for pentesting against API endpoints.

# Evasion, Poisoning, Extraction

  • Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference. GitHub last commit
  • CounterFit - A CLI that provides a generic automation layer for assessing the security of ML models. GitHub last commit
  • Foolbox - A Python toolbox to create adversarial examples that fool neural networks in PyTorch, TensorFlow, and JAX. GitHub last commit

# Jailbreaking

  • EasyJailbreak - An easy-to-use Python framework to generate adversarial jailbreak prompts. GitHub last commit
  • JailbreakBench - An Open Robustness Benchmark for Jailbreaking Language Models. GitHub last commit

# LLM & Generative AI Red Teaming

  • Garak - The LLM vulnerability scanner. GitHub last commit
  • PromptBreach - The Python Risk Identification Tool for generative AI. GitHub last commit
  • PyRIT - The Python Risk Identification Tool for generative AI. GitHub last commit
  • Promptmap2 - A security scanner for custom LLM applications GitHub last commit
  • Purple Llama - Set of tools to assess and improve LLM security. GitHub last commit

# Model Extraction & Inference Attacks

  • Privacy Meter - Audit data privacy in statistical and machine learning algorithms. GitHub last commit
  • SecretFlow - A unified framework for privacy-preserving data analysis and machine learning. GitHub last commit
  • ShadowAttack - Stealthy and Effective Physical-world Adversarial Attack by Natural Phenomenon. GitHub last commit

# Data Poisoning & Supply Chain Attacks

  • Backdoor Box - A universal pytorch platform to conduct security researches of image classification in deep learning. GitHub last commit
  • TrojanZoo - The open-sourced Python toolbox for backdoor attacks and defenses. GitHub last commit

# LLM Safety & Guardrails

  • Alibi Detect - Algorithms for outlier, adversarial and drift detection. GitHub last commit
  • Guardrails AI - Adding guardrails to large language models. GitHub last commit
  • LLM Guard - A comprehensive tool designed to fortify the security of Large Language Models. GitHub last commit
  • NeMo Guardrails - Toolkit for easily adding programmable guardrails to LLM-based conversational systems. GitHub last commit
  • Vigil - Detect prompt injections, jailbreaks, and other potentially risky Large Language Model inputs. GitHub last commit