Back to AI Agents Hub
๐ก๏ธ
LLM Guard
by Protect AI
Comprehensive security toolkit for LLM interactions
LLM Guard provides input/output scanners to sanitize LLM interactions. Detects and prevents prompt injections, jailbreaks, data leakage, and toxic content with minimal latency.
Ease of Use
0/10
Community
0/10
Performance
0/10
Documentation
0/10
๐ฏ Key Features
15+ input scanners
12+ output scanners
Prompt injection detection
Jailbreak prevention
PII/secrets detection
Toxicity filtering
Relevance checking
Language detection
Code scanner
URL/SQL injection detection
Anonymization
Bias detection
Strengths
Comprehensive scanner library
Low latency
Easy integration
Active development
Good balance of speed/accuracy
Modular architecture
Well-documented
Limitations
Python only
Some scanners have higher false positives
Limited customization for some scanners
No built-in UI
Requires tuning for production
Best For
- Security-first applications
- PII-sensitive environments
- Multi-tenant platforms
- Compliance-heavy industries
- Production LLM APIs