Skip to content

LLMSecurity/awesome-agent-skills-security

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

12 Commits
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Awesome Agent Skills Security Awesome

πŸ›‘οΈ A curated list of resources on securing AI agent tool use and skill ecosystems β€” attacks, defenses, frameworks, benchmarks, and standards.

AI agents increasingly use external tools, plugins, and skills to interact with the world. This creates a new attack surface: agent skills security. This list covers the threats, defenses, and research landscape for securing these capabilities.

Contents


Threat Frameworks & Standards

Surveys & Systematizations

Attack Research

Prompt Injection via Tools

Tool Poisoning & Supply Chain

Privilege Escalation & Excessive Agency

Data Exfiltration & Privacy

Indirect Prompt Injection

Agent Deception & Manipulation

Compound System Attacks

Cross-Plugin Attacks

Backdoor Attacks on Agents

Jailbreaking & Guardrail Bypass

Defense Research

Permission & Access Control

Runtime Monitoring & Sandboxing

Input/Output Validation

Formal Verification & Analysis

Evaluation & Red Teaming

Benchmarks & Datasets

Benchmark Focus Size Paper
ASB Comprehensive agent security 10 agents, 398 envs Zhang et al.
InjecAgent Indirect prompt injection 1,054 test cases Zhan et al.
R-Judge Safety risk awareness 162 records, 27 scenarios Yuan et al.
ToolSword Tool learning safety 6 scenarios, 3 stages Ye et al.
AgentDyn Dynamic prompt injection Open-ended, extensible Li et al.
Skill-Inject Skill file attacks Multi-scenario Schmotz et al.
NAAMSE Evolutionary agent security eval Adaptive red-teaming Pai et al.
AgentHarm Agent misuse 110 behaviors, 440 variants Andriushchenko et al.
SkillGuard Dataset Malicious skill detection 157 malicious skills Liu et al.
WIPI Web-based indirect injection Multi-scenario Liu et al.

Tools & Frameworks

Tool Description Link
SkillGuard LLM-native agent skill security auditor (OWASP Agentic + MITRE ATLAS) GitHub
Invariant Guardrails Policy-based agent security guardrails GitHub
LLM Guard Input/output scanning for LLM applications GitHub
Rebuff Self-hardening prompt injection detector GitHub
NeMo Guardrails NVIDIA's toolkit for adding guardrails to LLM-based applications GitHub
Lakera Guard Enterprise prompt injection defense API Website
Promptfoo LLM red teaming and evaluation framework GitHub
Garak LLM vulnerability scanner GitHub
AgentSkillsScanner Static analysis scanner for agent skill definitions GitHub
Agent Audit Security analysis system for LLM agent apps: dataflow analysis, credential detection, MCP config parsing, privilege-risk checks Zhang et al.
mcp-sec-audit MCP server security toolkit: static pattern matching + dynamic sandboxed fuzzing via Docker/eBPF for detecting over-privileged tool capabilities Huang et al.

Agent Skill Specifications

Specification Org Focus
AgentSkills.io Open Standard Agent skill definition and security requirements
Model Context Protocol (MCP) Anthropic Tool/resource integration protocol for LLMs
OpenAI Function Calling OpenAI Tool use specification for GPT models
Tool Use (Claude) Anthropic Claude's native tool use interface
LangChain Tools LangChain Tool abstraction for agent frameworks
AutoGPT Plugins AutoGPT Plugin system for autonomous agents
OpenAPI/Swagger Linux Foundation API specification commonly used as tool definitions

Industry Reports & Blog Posts

Related Awesome Lists

Contributing

Contributions are welcome! Please read the contribution guidelines before submitting a pull request.

How to Contribute

  1. Fork the repository
  2. Add your resource in the appropriate category
  3. Use the format: - πŸ“„ **[Title](URL)** β€” Authors, Venue Year. One-sentence description.
  4. Submit a pull request

Criteria

  • Resources must be directly related to agent/tool/skill security
  • Papers should be published or on arXiv
  • Tools should be actively maintained (commits within last 6 months)
  • Blog posts should provide substantial technical analysis

Citation

If you find this list useful in your research, please cite:

@misc{awesome-agent-skills-security,
  author = {Liu, Yi},
  title = {Awesome Agent Skills Security},
  year = {2026},
  publisher = {GitHub},
  journal = {GitHub Repository},
  howpublished = {\url{https://github.com/LLMSecurity/awesome-agent-skills-security}}
}

License

CC0

This list is released under CC0 1.0 Universal.

About

πŸ›‘οΈ A curated list of resources on agent skills security: attacks, defenses, frameworks, and benchmarks for securing AI agent tool use and skill ecosystems

Topics

Resources

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors