Meta's 7-8B specialized moderation model for LLM input/output filtering. 6 safety categories - violence/hate, sexual content, weapons, substances, self-harm, criminal planning. 94-95% accuracy. Deploy with vLLM, HuggingFace, Sagemaker. Integrates with NeMo Guardrails.
7.0
Rating
0
Installs
AI & LLM
Category
Excellent skill documentation for LlamaGuard content moderation. The description is comprehensive and actionable for a CLI agent, covering installation, basic usage, and key capabilities. Task knowledge is strong with 5 detailed workflows (input/output filtering, vLLM deployment, API endpoint, NeMo integration) plus troubleshooting for common issues like OOM and latency. Structure is clear with logical progression from quick start to advanced topics, though the single-file format is slightly dense. Novelty is solid - deploying and optimizing a 7-8B parameter model with specific moderation categories, vLLM acceleration, and framework integrations would be token-intensive for a basic CLI agent. Minor deductions: could benefit from splitting advanced topics into separate files as referenced, and the novelty is somewhat limited since it's primarily model deployment rather than custom algorithmic innovation. Overall, a highly practical and well-documented skill that would save significant effort in implementing AI safety moderation.
Loading SKILL.md…