More Tech League episodes

#17 Testing in the AI era thumbnail

#17 Testing in the AI era

Published 27 Mar 2026

Recommended: In depth discussion of AI and Software Development and the future of Software Testing.

Duration: 01:09:03

Testers must adapt to AI's role in code generation by rigorously validating AI-produced code against structured requirements, balancing human oversight with AI tools to address quality, ownership, and architectural testability challenges while redefining testing as a collaborative, discovery-driven process.

Episode Description

In this episode, Toby and Krisztian welcome their first proper guest: Alan Richardson, a 30-year software veteran and testing specialist known as Evil...

Overview

The podcast explores the evolving role of testers in an AI-driven software development landscape, emphasizing the critical need for testing as AI-generated code becomes more prevalent. Testers are encouraged to guide AI with precise requirements and tests to align outputs with intentions, though their traditional "superstar" status may wane in favor of integration into collaborative workflows. Testing strategies must adapt to AIs variable workflows, prioritizing small code chunks and architectural decisions that directly influence test quality. Key challenges include ensuring AI-generated code is well-structured, clarifying whether AI should write tests or be separated from code generation, and maintaining structured AI workflows with manual review of architecture to guarantee testability.

The discussion highlights the blurring line between testing and coding in AI contexts, with unit tests and coding becoming inherently linked. Traditional practices like Test-Driven Development (TDD) are reinterpreted to prioritize architecture first, though pure TDD struggles without predefined structures. AIs limitations in context-switching and executing test-code cycles necessitate hybrid approaches, combining AIs speed for rapid code generation with human oversight for quality checks and edge cases. Code quality relies on good design principles, clean architecture, and abstractions, which AI can follow if guided explicitly, but test code often lacks these elements, leading to brittleness.

Testing automation with AI simplifies tasks like UI testing but risks generic, unabstracted tests that break easily with application changes. "Self-healing" tests are criticized as flawed, requiring updates alongside application changes rather than automating fixes. Testers are urged to shift from routine checks to providing new, surprising insights, aligning testing with exploration of uncertainties. Human expertise remains vital for reviewing AI outputs, ensuring alignment with specifications, and avoiding reliance on AI-generated patterns that may perpetuate poor practices. The future emphasizes structured testing frameworks, domain-driven design, and integrating testing into development from the start, with testers playing a strategic role in uncovering hidden issues and validating AIs outputs.

Final Notes

Key Insights from the Text
1. Evolving Role of Testers in the AI Era
  • Human Oversight Remains Critical: While AI generates code, manual review of architecture and design is crucial for testability and maintainability. Testers must evolve from "superstars" to collaborators who ensure alignment between AI outputs and business requirements.
  • Testing and Coding Are Intertwined: AI blurs the boundary between testing and coding, requiring testers to engage more deeply with development processes. Future success depends on human expertise in guiding AI, even as automation grows.
2. Adaptability in Testing Strategies
  • Context-Driven Approaches: Testing strategies must adapt to AI workflowswhether incremental changes or large-scale code generation. For example, small code chunks require focused testing, while architectural decisions (e.g., interfaces) directly impact test quality.
  • Speed vs. Quality: AIs rapid code generation can lead to oversight of critical architectural choices. Proper implementation (e.g., structured code, abstractions) must precede testing to avoid brittle test suites.
3. Architecture and Abstraction as Foundations
  • Prioritize Design Over Test Volume: Well-structured code with abstraction layers (e.g., page objects, interfaces) reduces maintenance costs and improves test robustness. AI-generated code may lack these patterns, necessitating human intervention to enforce best practices.
  • Interface-Driven Testing: Interfaces act as "seams" for testing, isolating core logic and ensuring consistency during migrations or refactoring. Tests should target interfaces, not implementation details.
4. Challenges in AI-Generated Testing
  • Poorly Structured Test Code: AI-generated tests often lack abstractions, leading to hardcoding locators and brittle tests. Manual review is essential to inject modularity and reduce maintenance overhead.
  • "Self-Healing" Tests Are Flawed: Relying on automated fixes for test failures without addressing root causes is ineffective. Tests must evolve alongside application changes, not just patched.
5. AI and Testing Strategy Integration
  • Hybrid Human-AI Workflows: AI excels at rapid code generation and repetitive tasks, but humans are needed for complex logic, edge-case validation, and ensuring alignment with requirements.
  • Exploratory vs. Regression Testing: AI can aid in exploratory testing (e.g., generating hypotheses) but should not replace regression testing for stable components. Prioritize coverage for API-level contracts and interfaces.
6. Reframing Testing as Information Discovery
  • Tests Should Reveal New Insights: Effective tests uncover novel issues or unexplored areas, not just confirm existing knowledge. Over-reliance on regression testing limits value, while exploratory testing drives innovation.
  • Testers as Uncertainty Managers: Testers should expand their roles beyond routine checks to focus on ambiguous or high-risk areas, adding value through analysis and creativity.
7. Security and AI-Driven Threats
  • AI in Adversarial Testing: AI can simulate exploits (e.g., CVE-based attacks) to identify vulnerabilities, but human domain expertise is still needed to prioritize and interpret results. Tools like red team bots can enhance security posture if guided properly.
  • Balancing Defense and Attack: AI assists in vulnerability detection and mitigation but cannot replace human judgment in assessing exploitability or impact.
8. Continuous Learning and Generalist Roles
  • Shift to Generalists: AI reduces the need for deep specialization in specific technologies, favoring adaptability across testing, security, and architecture. Professionals must continuously learn to work with AI tools effectively.
  • Mentorship for the Next Generation: Companies risk a skills gap by bypassing junior developers. Senior engineers must mentor newcomers to ensure innovation and avoid over-reliance on AI.
9. Tools and Automation Limitations
  • Noise and Coverage Gaps: Automated tools generate false positives and lack clarity on coverage completeness. Human-AI collaboration is needed to validate test effectiveness and prioritize fixes.
  • Custom AI Agents: Proprietary tools and open-source frameworks (e.g., agentic QE fleets) can address commercial tool limitations, but human guidance is still necessary to prevent AI from falling into repetitive patterns.
10. Ethical and Practical Concerns
  • Homogenization of Output: AI-generated code may lack originality, leading to generic solutions. Human creativity remains essential for innovation, particularly in design and architecture.
  • AI as a Tool, Not a Substitute: While AI accelerates development, it cannot replace human expertise in critical areas like security, requirements encoding, and high-stakes decision-making.

Relevance and Usefulness to Readers
  • For Testers: Adapt to hybrid roles by mastering AI tools, advocating for architecture-first approaches, and focusing on value-adding tasks like exploratory testing and uncertainty management.
  • For Developers: Prioritize design and abstraction in AI-generated code to ensure testability and long-term maintainability. Use AI for speed but maintain human oversight for complex logic.
  • For Managers: Invest in training programs that blend AI literacy with domain-specific expertise. Balance automation with mentorship to avoid over-reliance on AI and ensure innovation.
  • For Security Teams: Leverage AI for threat simulation but couple it with human analysis to address vulnerabilities meaningfully.
  • For Startups vs. Enterprises: Startups may prioritize speed, but enterprises must embed testing and security into architecture from the start to avoid technical debt.

This framework underscores that AI is not a replacement for human expertise but a tool that requires strategic integration, continuous learning, and ethical consideration to maximize its potential in software development and testing.

Recent Episodes of Tech League

20 Mar 2026 #16 Bunny CDN

Bunny CDN is spotlighted as an EU-based, cost-effective alternative for hosting static websites and development projects, offering user-friendly setup, Storage Zones, Terraform/GitHub Actions automation, edge computing, and competitive pricing, though it lacks enterprise-scale infrastructure and global scalability for high-volume workloads.

13 Mar 2026 #15 Agentic Engineering

Large Language Models (LLMs) have the potential to revolutionize software development, but their integration also raises challenges and requires a balance between AI capabilities and human oversight.

27 Feb 2026 #14 EU Cloud Alternatives - Scaleway

This podcast discusses EU-based alternatives to mainstream infrastructure and software development tools to achieve digital sovereignty, examining open-source and self-hosted solutions for development and management.

20 Feb 2026 #13 OpenClaw

The podcast discusses OpenClaw, an open-source AI assistant that learns user routines and preferences, and offers features like task automation and scheduled check-ins.

13 Feb 2026 #12 DevOps and SecOps

This podcast explores the evolution of DevOps, highlighting its shift from siloed teams to a culture of shared responsibility, accountability, and continuous improvement, resulting in faster problem resolution and more reliable systems.

More Tech League episodes