Skip to content

muellerberndt/awesome-ai-security

Repository files navigation

Awesome AI Security

A curated, annotated list of resources for AI security.

View the interactive roadmap

ML Foundations

Essential machine learning concepts and courses to build a foundation before diving into AI security.

Deep Learning

Deep dive into neural networks, transformers, and the architectures behind modern AI systems.

Prompt Injection

Understand prompt injection attacks that manipulate LLM behavior through crafted inputs.

Adversarial Attacks

Learn how adversarial examples fool neural networks and methods to defend against them.

Poisoning & Backdoors

Data poisoning attacks and neural network backdoors that compromise model integrity.

Privacy & Extraction

Model extraction, membership inference, and training data extraction attacks.

Tools & Frameworks

Security tools for testing and defending AI systems against adversarial attacks.

AI Pentesting

Using AI assistants and agents for automated penetration testing and security assessments.

Vulnerability Detection

AI-powered vulnerability scanning, code analysis, and bug detection.

Exploit Generation

AI-assisted exploit development and attack automation techniques.

AI Security Tools

Tools that leverage AI for offensive security operations and analysis.

  • Hound - AI auditor that builds adaptive knowledge graphs for deep code reasoning. Uses tiered AI approach for autonomous vulnerability discovery.
  • HackGPT - LLM toolkit for offensive security.
  • HackingBuddyGPT - Autonomous red-teaming agent with benchmarks.
  • GhidrAssist - LLM extension for Ghidra with ReAct agentic mode for autonomous reverse engineering investigation.
  • PyRIT (Python Risk Identification Tool) - Microsoft red-teaming framework for generative AI. Automates adversarial prompt generation and risk assessment.
  • AI Security Analyzer - Generates security docs from codebases.
  • BurpGPT - Burp Suite extension for AI-powered vulnerability scanning.
  • CAI: Cybersecurity AI - Framework for building AI-driven security tools by Alias Robotics.

Benchmarks & Standards

Industry standards, threat frameworks, and evaluation benchmarks for AI security.

  • ScaBench - Smart contract audit benchmark with 500+ real-world vulnerabilities from Code4rena, Cantina, and Sherlock for evaluating AI audit agents.
  • RobustBench - Leaderboard for adversarial robustness benchmarking.
  • JailbreakBench - Benchmark for LLM jailbreak attacks and defenses.
  • Stanford AIR-Bench 2024 - AI safety benchmark aligned with emerging government regulations and company policies.
  • FLI AI Safety Index 2024 - Future of Life Institute's assessment of AI company safety practices and accountability.
  • MITRE ATLAS - Adversarial Threat Landscape for AI Systems. Threat matrix documenting real-world attacks on ML (like ATT&CK for AI).
  • NIST AI Risk Management Framework - Framework for managing AI risks throughout the AI lifecycle.

Books

Essential books covering AI security, adversarial ML, and security applications.

Communities & Events

AI security communities, conferences, and events to stay connected.

Newsletters & Lists

Newsletters and awesome lists to stay current with AI security developments.


© muellerberndt · GitHub

About

An AI security awesome list / learning journey

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors