AI and Technical Debt

What are you doing to prevent AI technical debt from a quality and testing perspective?

When tests can be churned out so quickly, how have you seen this actually cause problems?

@christinepinto wrote about this recently:

AI-generated tests are creating a new kind of ๐˜๐—ฒ๐—ฐ๐—ต๐—ป๐—ถ๐—ฐ๐—ฎ๐—น ๐—ฑ๐—ฒ๐—ฏ๐˜. And weโ€™re about to drown in it.
LLMs can now generate 1,000 tests in minutes. Sounds great, right?
Wrong.

Theyโ€™re making the same mistakes every time:
โ†’ Writing trivial tests that only fail when you intentionally change things
โ†’ Setting up 20 lines of mocks, then skipping the one line that actually matters
โ†’ Testing how code works instead of what it does

The result? A ๐—บ๐—ฎ๐—ถ๐—ป๐˜๐—ฒ๐—ป๐—ฎ๐—ป๐—ฐ๐—ฒ ๐—ป๐—ถ๐—ด๐—ต๐˜๐—บ๐—ฎ๐—ฟ๐—ฒ disguised as safety.

After 18 years in QA, Iโ€™ve watched teams struggle with too little test coverage. But weโ€™re about to flip to a new problem: drowning in low-value tests that make every refactor expensive.

Douwe Osinga on the Block Engineering blog nailed it: โ€œWriting tests has become ๐—ณ๐—ฟ๐—ฒ๐—ฒ. Maintaining them hasnโ€™t.โ€

This is the hidden cost everyoneโ€™s missing.

The same thinking that values test coverage percentages over bug prevention. The same thinking that makes teams afraid to change working code because the test suite will explode.

Not all tests have positive value. Some actively make your codebase harder to evolve.

The fix? Prompt LLMs with reasoning about test value, not just test coverage. Make them ask: โ€œIf this test fails, what did we just prevent?โ€
Because the goal isnโ€™t having tests. Itโ€™s catching real bugs.

1 Like

Well to be honest, weโ€™re not using AI to create tests.

We may use AI to evaluate requirements/stories to come up with ambiguities, shortfalls and test scenarios as an input to our test creation, but ultimately the quality of our test cases is our responsibility.

It sure isnโ€™t the LLMโ€™s responsibility. If you live in the illusion that it is, then thats a scenario that tech debt is made forโ€ฆbut its the worst type of tech debt as you need to investigate what it created as you donโ€™t quite know what it created - especially if your talking about 1000+ tests. So people then turn back to AI to come up with the solution to solve the tech debt problem theyโ€™ve just created with AI, still not taking responsibility for the quality problem and thus commences the cycle of AI insanity.

2 Likes

It doesnโ€™t matter who writes the tests, humans or AI, it doesnโ€™t matter how the tests are run, manually or automated. If there are too many test cases, too many doubles, triples, no clear tracing, no structure in the test suites, no test of the actual requirement, youโ€™ll get problems. In creating and of course in maintenance. And since AI can churn out giant amounts of tests in no time, it is bound to fill an existing problem (no clear tracing, no on point tests) with even more drivel.

2 Likes