One of the first things I integrated into my development workflow with AI was writing tests.
AIs are remarkably good at this. Give them context about your codebase, and they'll generate comprehensive test suites faster than you can write a single spec file.
The "overtesting" problem
There's a recurrent pattern: "AI tends to overtest." Too many assertions, edge cases you'd never bother with manually...
Before AI, this was a legitimate concern. Every test you write is a test you have to maintain. But if the AI writes the tests and maintains them, and your role becomes supervisory—reviewing, approving, and steering—then a bit of "overtesting" is not a big deal. The economics of quality have changed.
If you use Claude Code, create a "Testing Skill"
This is the secret to scaling consistency. Here's my approach to building one:
Audit your patterns: Let Claude examine your existing tests first so it can replicate your specific style.
Define fixture strategy: Be explicit about when to use fixtures vs. creating fresh records.
Set stub boundaries: Define what's okay to stub (External APIs? DB calls?).
Time-travel rules: Establish conventions for date/time handling to avoid flaky tests.
Add guardrails: Focus on behavior, not implementation details. No testing private methods!
E2E testing just became accessible
With Claude Code's browser integration, complex E2E tests are now just a natural language prompt away:
"Login as employee@test.com, create 3 items from different categories, assign them to different offices, and generate an items export report to CSV."
That's the whole spec. What used to be painful to maintain is now a simple conversation.
Conclusion
Improving test coverage and simplifying maintenance is the "low-hanging fruit" of AI integration. It's low risk, high visibility, and provides immediate value to any team.
Just make sure everyone shares the same Testing Skill and context. Consistency is what makes this scale.