Discussion about this post

User's avatar
Benny Taccardi's avatar

My 2 cents on this.

I think the core issue lies in treating 'modern' output through the lens of 'traditional' workflows.

Back in the 90s or early 2000s, when the industry was dominated by OS-level code and kernel-heavy engineering, testing had a very specific, rigid DNA. Then Web Dev arrived, and we realised that treating a web app like an OS-based system was a mistake; an API fails for reasons and in ways that a system call simply doesn't.

We are at a similar crossroads with LLMs. Whether we are using them to write code or building products on top of them, we have to accept that the entire workflow and, by extension, our effort and our Jira tickets, needs to be recalibrated.

Even the Definition of Done (DoD) has to evolve. We’re moving away from the classic "API X is in prod with Y% test coverage and docs" toward a framework based on confidence levels. This is a necessary shift because we are no longer operating in an "IF AND ONLY IF deterministic environment".

If companies actually embrace this paradigm shift, throughput might stay higher than the old baseline, but "high productivity" won't look anything like the metrics we’re using today.

No posts

Ready for more?