It described something I was seeing constantly. A performance regression would ship to production. SREs would catch it. And the first thing developers would say — genuinely, not defensively — was: “It looked fine in pre-prod.” They weren’t wrong. It had looked fine in pre-prod. The synthetic tests passed. The problem only showed up when real users on real devices in real network conditions hit it.
So you’d get this standoff. Developers pointing at clean staging results. SREs pointing at production signals that weren’t translated into anything a frontend engineer could act on. Everyone waiting to be absolved. Nobody owning the gap between the two.
I watched it happen across the industry. And I think, if I’m honest, we sort of did it to ourselves.
