AI in Testing: A Tester's Perspective on the Unavoidable Revolution

AI in Testing: A Tester's Perspective on the Unavoidable Revolution

Backer posted 10 min read

We all use AI. Simple question about our health? Too lazy to google something that pops up in our mind? Love it or hate it, it changed our life, and most importantly, the tech world.

Since I'm a simple QA Engineer at my job, I won't enter philosophical conversations like "does it really think?" or "what happens if it gains consciousness?" I'm sure you're already flooded with that content everywhere. Instead, I'll share my thoughts and experiences about what this shift actually means from a tester's perspective. The trenches, not the ivory tower.

The Elephant in the Testing Room

Let's address what everyone's thinking but few are saying out loud: Is AI coming for our jobs?

When ChatGPT (now sells data, dont use it) exploded onto the scene, every Slack channel in every testing community I'm part of had the same panic: "Will AI replace manual testers?" "Should I learn something else?" "Is this the end of QA as we know it?"

I've been there. I've stared at an AI-generated test plan and felt that cold sweat. But here's what I've learned after riding this wave for a while: AI isn't replacing testers. It needs better QA.

The testers who just execute scripts, who just click through the same regression suite day after day, who never question why we're testing something? Yes, those roles are shrinking. But the testers who understand risk, who think about user behavior, who can challenge product decisions? We're becoming more valuable than ever.

How AI Actually Changed My Day-to-Day

Let me get practical. Here's what shifted in my actual work:

Test Data Generation Went from Hours to Minutes

Remember spending entire afternoons crafting the "perfect" test data? Name, address, credit card numbers that pass Luhn checks, edge cases for every field? I used to keep a spreadsheet. A whole spreadsheet.

Now? I describe what I need to an AI assistant, and it generates realistic, varied test data in seconds. Need 50 users with names from different cultures? Done. Need payment profiles with specific expiry patterns? Done. The grunt work evaporated.

Test Ideas I Wouldn't Have Thought Of

This one surprised me. I thought I knew my application's edge cases. I've tested it for years. But when I started describing features to AI and asking "what am I missing?" It consistently suggested scenarios that hadn't crossed my mind.

Not because I'm a bad tester. Because AI has ingested thousands of bug reports, countless testing discussions, and patterns from applications I've never seen. It's like having a junior tester who's read every Stack Overflow post about testing, ever.

Documentation Didn't Become Optional, But It Became Easier

I hate writing test documentation. I love having test documentation. Classic tester dilemma.

AI didn't magically make documentation unnecessary, but it made creating and maintaining it painless. I can dump rough notes, ask it to format them into a coherent test plan, and then actually focus on the testing itself. My test case repository is more up-to-date now than it has been in years, simply because updating it doesn't feel like pulling teeth.

The Reality Check: Where AI Hits Its Limits

Now let's talk about the parts that don't make it into the LinkedIn influencer posts. The messy reality.

The MCP Server Mirage

Model Context Protocol servers sounded like magic when I first heard about them. AI connecting directly to my databases, my monitoring tools, my test environments? Yes please.

Then I tried running them against our Forex CRM.

That application is a beast. Real-time WebSocket connections streaming price updates, complex state synchronization across multiple clients, milliseconds mattering. And MCP servers? They did awesome job at first(If you dont include nested if-elses and unnecessary 100 lines of codes), but eventually it felt apart. The context windows couldn't capture the streaming nature. Adding networkidle as a wait does not work in websocket ecosystem. Rely on xPath instead of ARIA roles. DevXtreme nature against hidden states(toast message needed workaround) The AI kept losing track of what was happening now versus what happened five seconds ago.

The lesson was humbling: AI tools that work beautifully for CRUD applications often choke on real-time, stateful systems. The marketing demos don't show you that part. They show you a to-do app, not a trading platform where every milisecond matters with WebSockets pushing 100+ messages per second.

The Fix-It Paradox

Here's something nobody tells you: sometimes fixing what AI broke costs more than writing it yourself.

I've lost count of how many times I've asked AI to generate a complex test script, gotten something that looks right, then spent hours debugging why it fails in ways that make no sense. The assertions are slightly wrong. The waits are poorly placed. It handles the happy path but falls apart on the first unexpected response.

By the time I've untangled AI's assumptions, fixed its race conditions, and rewritten the parts where it hallucinated API responses, I could have written the damn thing from scratch in half the time. And I'd understand it better.

The trade-off isn't as simple as "AI saves time." Sometimes it just shifts time from writing to debugging, and debugging AI-generated code is uniquely frustrating because you have to second-guess everything. Did it misunderstand me? Did it hallucinate? Did it use a pattern that doesn't apply here?

The Laziness Trap: Killer of Thinking?

This is the one that keeps me up at night.

When AI can generate test plans, write automation scripts, and suggest architectural patterns, why should I push myself to learn all those architectures? Why should I struggle through designing my own test framework when AI can produce one in seconds?

I see this happening around me. Junior testers who never learn to construct a good SQL query because they just ask AI. Mid-level engineers who can't explain why their automation framework is structured that way because AI designed it. There's a creeping intellectual laziness that comes with having a smart assistant on tap.

And here's the hard question: What happens to the next generation of senior testers and leads? Can AI give you fifteen years of domain knowledge? Can it teach you the patterns of failure that only come from being woken up at 3 AM for a production outage? Can it pass down the hard-won instincts about what's worth testing and what's safe to ignore?

I don't think so. Not at this stage. But if we stop developing those instincts in ourselves and our juniors because "AI can handle it," we're creating a expertise vacuum. And vacuums get filled with something usually by consultants who still did the hard work of learning.

The Hype Train: Do You Really Need AI There?

Everyone's in a rush. Automate everything. Integrate AI into every tool. Use it everywhere.

I sat through a demo last month where our lead showed off their AI-powered test case generator. It was impressive. Then someone asked: "Why do we need AI to generate test cases? Our test cases haven't changed significantly in years. We could just... keep using the ones we have?"

The room went quiet.

There's a frenzy right now to bolt AI onto everything, regardless of whether it solves an actual problem. It's the new blockchain, a solution in search of a problem. And in testing, where we're supposed to be the voice of reason and risk assessment, we should be asking the hard question: Does this actually add value, or does it just add complexity?

Because AI isn't free. It costs money, it costs context window limitations, it costs debugging time, and it costs the mental overhead of managing another tool. If the ROI isn't there, you're just chasing a trend.

Tokens, Costs, and Knowing When to Let Go

Claude took the crown for coding. I'll admit it, i reach for Claude first when I need to generate something complex. But there's a ceiling, and it's measured in tokens.

I've had sessions where I'm deep in a problem, feeding context, getting good responses, making progress — and then I hit the limit. Conversation too long. Context window full. Start over. Explain everything again. Lose the thread.

Or the cost. Running heavy AI assistance continuously adds up. If you're generating code all day, every day, those API costs become real. And if you're using it through a tool with hard token limits, you're constantly fighting against the clock.

The discipline I'm learning: Use it surgically, not exhaustively. Generate the skeleton, then fill in the meat yourself. Ask for patterns, not full implementations. Get unstuck, don't get carried. Because the moment you rely on AI to hold your entire problem in its context window, you're setting yourself up for the frustration of hitting that wall and losing everything.

The Shift in Skills Nobody Talks About

Here's what I've noticed about my own work: the skills that matter are shifting.

What matters less:

  • Remembering syntax for every testing tool

  • Knowing exactly which menu option in which tool does what

  • Manually crafting every test case from scratch

What matters more:

  • Asking better questions

  • Knowing what "good" looks like well enough to evaluate AI's output

  • Understanding the system deeply enough to know what to test

  • Communicating findings in ways that drive action

  • Recognizing when AI is leading you down a garden path

AI made me more of a critic and less of a producer. I spend more time evaluating and refining, less time generating from zero. It's honestly more interesting work. But it also demands more discernment.

The Tools That Actually Made a Difference

I've experimented with a lot. Some stuck, some didn't. Here's what actually changed how I work:

  • For exploratory testing sessions: I'll describe the feature and ask for exploratory testing charters. Not to follow blindly, but as a starting point that's better than a blank mind.

  • For bug reporting: I paste my notes, screenshots, and initial observations, and ask for help structuring a clear bug report. The bugs I write now are more reproducible because AI helps me spot missing steps in my own description.

  • For test planning: I use it as a sounding board. "Here's what I plan to test in this sprint. What am I missing?" The answers aren't always right, but they always make me think.

  • For learning: When I encounter a new technology or testing challenge, I used to google, skim blog posts, piece together understanding. Now I ask AI to explain it to me, ask follow-up questions like I'm interrogating a colleague, and get tailored explanations. My learning curve flattened dramatically.

The Things AI Is Surprisingly Bad At (For Now)

Let's balance the hype with reality. There are things AI still gets wrong, and knowing these helps me use it better:

  1. Context. AI doesn't know your users, your business priorities, or your historical baggage. It suggests comprehensive testing for everything, which is useless, testing is always about tradeoffs. I have to bring the context.

  2. Novelty. When testing truly new interactions or patterns, AI falls back on old patterns. It suggests testing approaches that worked for yesterday's applications, which might not fit tomorrow's.

  3. Skepticism. AI is too agreeable. It doesn't challenge assumptions the way a good tester should. If I describe a feature wrong, it tests my wrong description, never questions it.

  4. Exploratory intuition. The hunches, the "something feels off here" moments, the ability to follow a vague unease until it becomes a bug, that's still human. AI follows paths, it doesn't wander.

  5. Real-time systems. Anything with WebSockets, gRPC, streaming data, complex state synchronization, AI struggles. Its training data is heavy on REST APIs and CRUD apps, light on the messy real world of persistent connections and out-of-order events.

What I Tell Junior Testers Now

When new testers ask me for advice, my guidance changed:

  • "Learn to test manually first. You need to know what good testing feels like before you can evaluate AI's testing."

  • "Use AI for everything it can do, but treat its output as a first draft. Your job is editing, not writing."

  • "Deeply understand one system, really know it, its users, its failure modes. That depth is your moat."

  • "Be the person who asks 'why are we testing this?' not just 'how do we test this?'"

  • "Don't let AI think for you. Let it do the boring parts so you can think harder about the interesting ones."

  • "Watch out for the laziness trap. If you stop learning because AI can generate it, you're not saving time, you're capping your career."

The Uncomfortable Truth

Here's the part I wrestle with: AI exposed how much of testing was ritual rather than reasoning. How many test cases we ran because "we always run them" rather than because they still provided value. How much time we spent on work that was systematic but not thoughtful.

That's uncomfortable to admit. I spent years perfecting test suites that AI can generate in seconds. But that discomfort is useful, it pushes me toward the parts of testing that actually require human judgment.

The other uncomfortable truth: We're creating a generation of testers who can prompt but can't troubleshoot. Who can generate but can't debug. Who have answers but not understanding. And that's on us, the seniors, the leads, the ones who should be mentoring, if we let AI replace the struggle of learning rather than just accelerate it.

Looking Forward

I don't know where this goes. Five years from now, testing might look completely different. But I've stopped worrying about replacement and started focusing on adaptation.

The testers I see thriving are the ones who treat AI as their new partner, one that's infinitely patient, widely read, and completely lacking in judgment. But they're also the ones who know when to ignore it, when to override it, and when to shut it off and think for themselves.

They're the ones who ask better questions, think more critically, and use the time AI saves to focus on the testing that actually matters. They're the ones who still learn the hard stuff, even when AI could generate a passable version, because they know the difference between passable and excellent.

We all use AI now. The question isn't whether it changed our work, it did. The question is whether we're using it to do more of the same, or to finally do the testing we always wanted to do. And whether we're still doing the hard work of becoming the testers our future selves will need to be.

Lastly, you guessed it right, i took help from AI while writing this article.. Anyway..

What about you? Have you hit the MCP wall? Spent an afternoon fixing what AI broke in minutes? Caught yourself getting lazy?

4 Comments

2 votes
1 vote
2 votes
0 votes

More Posts

I’m a Senior Dev and I’ve Forgotten How to Think Without a Prompt

Karol Modelskiverified - Mar 19

The End of Data Export: Why the Cloud is a Compliance Trap

Pocket Portfolioverified - Apr 6

Split-Brain: Analyst-Grade Reasoning Without Raw Transactions on the Server

Pocket Portfolioverified - Apr 8

Breaking the AI Data Bottleneck: How Hammerspace's AI Data Platform Eliminates Migration Nightmares

Tom Smithverified - Mar 16

How I Built a React Portfolio in 7 Days That Landed ₹1.2L in Freelance Work

Dharanidharan - Feb 9
chevron_left

Related Jobs

View all jobs →

Commenters (This Week)

15 comments
1 comment
1 comment

Contribute meaningful comments to climb the leaderboard and earn badges!