Blog
Thoughts, updates, and insights from the Superagent team.
The Threat Model for Coding Agents is Backwards
Most people think about AI security wrong. They imagine a user trying to jailbreak the model. With coding agents, the user is the victim, not the attacker.
AI Guardrails Are Useless
Hot take: most AI guardrails on the market today are security theater. Not because the idea is bad, but because of how they're implemented. Most guardrail solutions are generic, static, and disconnected from what actually matters for your specific agent.
Introducing Superagent Guard
Purpose-trained models that detect prompt injections, identify jailbreak attempts, and enforce guardrails at runtime. Optimized for deployment as a security layer in AI agent systems.
SOC-2 is table stakes now. Here's what actually matters for AI products.
A few years ago, having SOC-2 certification was a real differentiator. If you were selling to enterprise, that badge meant something. That's not the world we live in anymore.
Red Teaming AI Agents: What We Learned From 50 Assessments
After red teaming 50 AI agents across different companies, industries, and setups, we've identified critical patterns that teams need to understand. Here's what actually matters when securing AI agents in production.
Open Source AI Models: A Safety Score Reality Check
The open source AI movement has democratized access to powerful language models, enabling developers and organizations to deploy sophisticated AI systems without vendor lock-in or prohibitive costs.
Join our newsletter
We'll share announcements and content regarding AI safety.