Skip to content

AI Testing Discipline

January 23, 2026 · LinkedIn

16 reactions · 0 comments · 1,144 views


This week I wrote about not trusting AI-generated code until it proves itself.

Here's why.
I built something in 6 days that should have taken 9 months. Early on, all unit tests passed. Looked perfect.

Then I tested against real-world legacy files.

25% success rate.

The AI was confidently skipping things it didn't understand. German conventions. 1980s formats. Edge cases that made no sense until you knew the history.

I spent days hunting down 195 of the messiest files I could find.
Success rate went to 93%.

Those 5000+ tests I mentioned? Most exist because of edge cases that looked impossible until I found them in the wild.

The discipline is refusing to trust code that just looks right.


What's the strangest edge case you've found testing against real data?

#SoftwareEngineering #AI


All LinkedIn posts