Add Agent Threat Rules (ATR) to Tools section#117
Add Agent Threat Rules (ATR) to Tools section#117eeee2345 wants to merge 2 commits intocorca-ai:mainfrom
Conversation
|
No actionable comments were generated in the recent review. 🎉 ℹ️ Recent review info⚙️ Run configurationConfiguration used: Organization UI Review profile: CHILL Plan: Pro Run ID: 📒 Files selected for processing (1)
✅ Files skipped from review due to trivial changes (1)
WalkthroughREADME의 "Tools" 섹션에 Agent Threat Rules (ATR) 항목이 새로 추가되었습니다. 항목에는 저장소 링크, GitHub stars 배지, 탐지 범위(프롬프트 주입, 도구 중독, 자격증명 유출, 스킬 공급망 공격 등)와 관련 메트릭이 포함되어 있습니다. Changes
Estimated code review effort🎯 1 (Trivial) | ⏱️ ~2 minutes 🚥 Pre-merge checks | ✅ 3✅ Passed checks (3 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches🧪 Generate unit tests (beta)
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
|
Hi! Just checking in on this PR. CodeRabbit's review came back clean with no actionable issues. As a quick update — ATR's detection rules have recently been integrated into Cisco AI Defense (merged as PR #79 in cisco-ai-defense/skill-scanner), which adds industry validation for the project. Would love to get this merged when you have a chance. Thanks! |
|
Updated since submission: ATR now at 108 rules (v1.1.1), shipped in Cisco AI Defense (PR #79). 53K+ MCP skills scanned, 0% FP on clean content. Ready for review. |
Update (April 2026)
ATR has grown significantly since this PR was first submitted:
npm install agent-threat-rules && npx atr scan .Summary
Adding Agent Threat Rules (ATR) to the Tools section.
ATR is a set of open-source, regex-based detection rules for AI agent security threats — like YARA/Sigma rules, but for LLM tool-calling attacks.
Why it belongs here
This list covers LLM security broadly — ATR fills the agent/tool-calling detection gap:
ATR complements the prompt injection defenses already listed (Rebuff, LLM Guard, Vigil) by focusing specifically on agent tool-calling threats rather than conversational prompt injection.
Summary by CodeRabbit