DEV Community

Let's Automate 🛡️ for AI and QA Leaders

Posted on • Originally published at blog.gopenai.com on

AI-Assisted Testing vs AI Agents vs AI Agent Skills: A Practical Journey Through All Three

Most teams are only using one layer of AI in testing. Here is what the full picture looks like — and how I built across all three.


Photo by Possessed Photography on Unsplash

Before any of this made sense, I had to answer a more basic question: what does AI QA Engineering actually mean?

What is AI QA Engineering — and Why QAEs, SDETs, and QA Automation Engineers Should Pay Attention

And before touching AI at all — the foundations still matter. Clean BDD tests. Reports that stakeholders can read.

How to Add Beautiful BDD Test Reports to Your Reqnroll Project Using Expressium LivingDoc

Before you automate smarter, you have to know what good looks like.

Layer 1 — AI-Assisted Testing

AI speeds you up. You are still driving.

This is where most teams start — and where most teams stay.

You write a prompt, get a test, review it, ship it. AI is a productivity multiplier. GitHub Copilot suggests the next line. ChatGPT drafts your test cases. Claude rewrites a flaky selector. You are in control at every step.

The catch? A bad prompt gives you a bad test — and it will look convincing. Garbage in, confident garbage out.

Crafting Effective Prompts for GenAI in Software Testing

I built ai-natural-language-tests at this layer. Give it a plain English requirement, and it generates Cypress or Playwright tests using GPT-4, LangChain, and LangGraph. Every output still needs your eyes on it — but the heavy lifting is done.

Same idea with JIRA-QA-Automation-with-AI : feed it a JIRA story with acceptance criteria, and BDD test scripts come out the other side. Human judgment still required at the end. You own every decision.

That last part is the definition of this layer.

Layer 2 — AI Agents for Testing

You give the goal. The agent executes, adapts, and decides.

At this layer, you stop steering and start delegating.

You set the objective. The agent figures out how to get there — and when something breaks mid-run, it handles that too. No human in the loop for every step.

selenium-selfhealing-mcp is a good example of what this looks like in practice. A UI change breaks a Selenium locator mid-execution. The agent inspects the DOM, finds the updated element, and keeps going — without stopping to ask you what to do. I submitted this to the Docker MCP Registry, and watching it recover from failures on its own still feels like a step-change from Layer 1.

For .NET teams, SeleniumSelfHealing.Reqnroll does the same with C#, NUnit, Reqnroll, and Semantic Kernel. And IntelliTest takes it further — write your assertions in plain English, and the agent decides whether the application behaviour actually matches the intent.

But there is a trap at this layer. Agents move fast and look thorough. It is easy to trust the output and skip the checks. Coverage looks complete — but the agent may have tested the wrong thing entirely.

The AI QA Engineer’s Decision Framework: When NOT to Use AI in Testing

And if you are using AI agents to run tests, a harder question follows: how do you know the agent’s output is correct? That is the LLM evaluation problem, and it turns out to be one of the most interesting unsolved problems in this space.

LLM Evaluation Explained: How to Know If Your AI Is Actually Working

Layer 3 — AI Agent Skills

Not a tool. Not an agent. Expertise that travels.

Layer 3 is the one most people have not thought about yet.

Here is the pattern I kept running into: every new agent project started from scratch. New codebase, new prompts, same underlying knowledge — how to read a requirement, what makes a test meaningful, when to flag a risk. The expertise was always being rebuilt. That seemed wrong.

A skill is a portable, encoded unit of expertise. It is not tied to one agent or one project. Any compatible agent can load it and apply it — without rebuilding the logic again. You build it once, and it travels.

GitHub Copilot Agent Skills: Teaching AI Your Repository Patterns

vibe-coding-checklist applies the same idea to AI code review — a shared quality framework that any team or any agent can use consistently.

The shift in thinking is subtle but significant. At Layer 1, you build prompts and tools. At Layer 2, you build goals and trust boundaries. At Layer 3, you build expertise itself — in a form that outlasts any single project or team.

The Difference That Matters


AI-Assisted Testing vs AI Agents vs AI Agent Skills

Three layers. All called AI testing. Now you know which one you are actually in.

All repos → github.com/aiqualitylab

More writing → aiqualityengineer.com


Top comments (0)