All insights
AI Product Building AI Agents Architecture

Auto-generated narrow monitors beat handwritten broad checks — a tight mesh over the exact shape of the code

1,000+ AI-generated monitors that each target specific code paths catch more bugs than 10 hand-written checks that cover general categories

@RampLabs — How We Made Ramp Sheets Self-Maintaining · · 5 connections

Ramp scaled from 10 hand-written monitors to 1,000+ AI-generated ones — each shaped to a specific code path introduced by a specific PR. The result was “like a tight mesh over the exact shape of the code,” catching one real bug per 75 lines in the first week. The key architectural move: on PR merge, an agent reads the diff and generates monitors tailored to the new code, so the monitoring surface grows with the codebase automatically.

This is Verification is the single highest-leverage practice for agent-assisted coding applied to production monitoring rather than development-time checking. Hand-written monitors suffer from the same coverage gap as hand-written tests: the author’s imagination is the ceiling. Auto-generated monitors cover the actual code surface, not what a human thought was important. This also extends Every optimization has a shadow regression — guard commands make the shadow visible — when every PR automatically gets monitors, shadow regressions become visible by construction rather than by luck. The “keep your existing stack” caveat from Ramp is important: auto-generated monitors are powerful but opaque, so they complement — not replace — hand-written instrumentation you understand. Property-based testing explores agent input spaces that example-based tests miss captures the analogous insight for testing: generative approaches explore spaces that human-written examples miss.