Blog

Thoughts, updates, and insights from the Superagent team.

ResearchJanuary 21, 20263 min read

We Bypassed Grok Imagine's NSFW Filters With Artistic Framing

Text-to-image safety is broken. We generated explicit content of a real person using basic compositional tricks. Here's what we found, why it worked, and what this means for AI safety systems.

Read more
BenchmarksJanuary 16, 202612 min read

AI Code Sandbox Benchmark 2026: Modal vs E2B vs Daytona vs Cloudflare vs Vercel vs Beam vs Blaxel

We evaluate seven leading AI code sandbox providers across developer experience and pricing to help you choose the right environment for executing AI-generated code.

Read more
ResearchJanuary 13, 20265 min read

The Threat Model for Coding Agents is Backwards

Most people think about AI security wrong. They imagine a user trying to jailbreak the model. With coding agents, the user is the victim, not the attacker.

Read more
SecurityJanuary 12, 20263 min read

AI Guardrails Are Useless

Hot take: most AI guardrails on the market today are security theater. Not because the idea is bad, but because of how they're implemented. Most guardrail solutions are generic, static, and disconnected from what actually matters for your specific agent.

Read more
AnnouncementsJanuary 6, 20262 min read

Introducing Superagent Guard

Purpose-trained models that detect prompt injections, identify jailbreak attempts, and enforce guardrails at runtime. Optimized for deployment as a security layer in AI agent systems.

Read more
ComplianceDecember 10, 20254 min read

SOC-2 is table stakes now. Here's what actually matters for AI products.

A few years ago, having SOC-2 certification was a real differentiator. If you were selling to enterprise, that badge meant something. That's not the world we live in anymore.

Read more
Next

Join our newsletter

We'll share announcements and content regarding AI safety.