Between 2024 and 2026, I designed and ran three AI adoption pilots inside a European scale-up operating across 26 markets. The pilots were about AI on paper. What they exposed was the underlying operating system of the engineering organization — planning cadence, ownership boundaries, manager load, hiring assumptions, evaluation rubrics. The pilots kept working because we kept fixing those, not because the tooling kept improving.

The organization had roughly 40 engineers across multiple squads when the first pilot ran. By the third, the engineering team had grown and matured. Each pilot iterated on the method: how to recruit volunteers, how to measure usefulness without producing theater, how to handle the non-engineering teams that wanted in, how to translate pilot findings into company-wide rollout decisions.

The most recent pilot had 46 volunteers across 7 departments, 149 survey responses, and ran for 5 weeks. It drove the company-wide AI adoption rollout and a revision of the engineering career framework.

The first pilot was an engineering-only experiment. The intended finding was tooling ROI. The actual finding was that review load became the binding constraint within two weeks. High-adopters produced more code faster; reviewers could not keep up; quality signals started to drift.

That had nothing to do with AI. It was a pre-existing weakness in the review system that AI made visible. Fixing review capacity was a prerequisite for meaningful AI adoption, not a downstream consequence.

The second pilot deliberately expanded scope to non-engineering teams — product, marketing, operations. The intended finding was cross-functional usefulness. The actual finding was that non-engineering teams produced more durable value faster than engineering teams did, because their workflows had fewer legacy shape-constraints.

That forced a rethink of the engineering career framework. "AI fluency" as a separate skill category did not survive contact with reality. What did survive was the observation that engineers who already operated well at the systems level — ownership, clarity, prioritization — absorbed AI faster than engineers who were stronger on local craft.

The third pilot was the most structured: 46 volunteers, 7 departments, 149 survey responses, 5 weeks. The intended finding was a rollout recommendation. The actual finding was that 82% of non-engineers were saving 3+ hours per week by week 5 with zero safety concerns, while engineering outcomes varied by squad in ways that tracked squad health more than individual adoption.

The rollout recommendation was not "adopt AI everywhere." It was "adopt AI selectively where the squad already has clean ownership and strong planning, and invest in ownership and planning first where they don't."

The Diagnostic I run is built on the pattern these pilots exposed. Engineering organizations slow down for reasons that sit underneath the visible symptoms — misaligned ownership, planning cadence that doesn't match delivery speed, manager load, uneven AI adoption creating variance rather than leverage. The findings are usually structural, not tactical.

The AI lens is not the headline. It is one of the clearest diagnostic instruments available right now because it exposes pre-existing weaknesses in the operating system faster than other interventions do.

Want a read on your own engineering organization? Start a conversation →