Bruno Legeard
Head of the AI lab
Join my presentation on: AI Test Execution Agents vs Scripted Test Automation: A Practical Decision Framework
AI test execution agents can now execute natural-language test scenarios (manual or Gherkin) directly on a GUI, without automation code. Interacting visually, they perform functional tests like a human, generating structured evidence (screenshots, explanations, PASS/FAIL verdict).
This capability challenges the long‑standing assumption that functional GUI testing must be fully scripted to be automated. In many contexts, AI agents can replace scripted automation for functional test execution, significantly reducing the cost of test creation and the ongoing maintenance burden caused by fragile locators in frequently changing UIs. I propose a strategic shift: use AI agents as the default functional GUI test execution engine during the high-volatility development and qualification phases to enable fast feedback. Scripted automation, ideally AI-assisted, should be reserved for a smaller set of stable, high-value tests run frequently in the CI/CD pipeline.
Based on experiments across twenty projects, I will introduce a Decision Radar to help teams select the right approach to test execution using three dimensions: execution cadence, UI evolution rate, and oracle strictness. I will also demonstrate how to measure and govern AI agent reliability using true/false PASS/FAIL metrics against human references to avoid false confidence.