Blog

Thoughts, updates, and insights from the Superagent team.

ResearchNovember 11, 20258 min read

Introducing Lamb-Bench: How Safe Are the Models Powering Your Product?

We built Lamb-Bench to solve a problem every founder faces when selling to enterprise: proving AI safety without a standard way to measure it. An adversarial testing framework that gives both buyers and sellers a common measurement standard.

Read more
ResearchOctober 24, 20258 min read

VibeSec: The Current State of AI-Agent Security and Compliance

Over the past weeks, we've spoken with dozens of developers who are building AI agents and LLM-powered products. The notes below come directly from those conversations and transcripts.

Read more
EngineeringOctober 21, 20257 min read

The March of Nines

The gap between a working demo and a reliable product is vast. Andrej Karpathy calls this the 'march of nines' — when every increase in reliability takes as much work as all the previous ones combined. This is the hidden engineering challenge behind every production AI system.

Read more
EngineeringOctober 20, 20258 min read

The case for small language models

Most agents today rely on large, general-purpose models built to do everything. If your agent has a single, well-defined job, it should also have a model designed for that job. This is the case for small language models: models that handle one task, run locally, and can be retrained as your data evolves.

Read more
ResearchOctober 11, 20255 min read

Why Your AI Agent Needs More Than Content Safety

You've enabled Azure Content Safety or Llama Guard. Your AI agent still isn't secure. Here's why content filtering isn't enough when your AI takes actions.

Read more
AnnouncementsOctober 10, 20254 min read

Shipped: Runtime Redaction and Command-Line Security

The past two weeks brought runtime redaction, a powerful CLI, URL whitelisting, and a developer experience that puts security directly in your workflow. Here's what shipped and why it matters for teams building with AI agents.

Read more

Join our newsletter

We'll share announcements and content regarding AI safety.