Agents: Deploying AI with Common Sense in Software Testing – Sofius

Agents: Deploying AI with Common Sense in Software Testing

Imagine an AI agent that writes tests, fixes bugs, and reviews your pull requests. Sounds like every tester’s and developer’s dream, right? But before you start embracing what AI can do, it’s crucial to look critically: where is the real value of these tools, and how can you deploy them responsibly? In this article, we will take you into the world of AI agents for software testing and share practical insights.

From Machine Learning to Agentic AI

AI is a broad concept. At its core, it involves systems that can make choices based on available data. We distinguish a number of steps:

  • Machine Learning: learns from datasets, such as the automatic filtering of phishing emails.
  • Deep Learning: can make neural connections between datasets and thereby perform complex tasks, for example, object recognition in self-driving cars.
  • Generative AI: generates content itself, such as texts, images, or sound.
  • Large Language Models: models like ChatGPT or Gemini, which understand context and generate output based on language.

AI agents go one step further. These are LLMs that are focused on one goal and have a degree of autonomy. They can, for example, take actions themselves, such as replying to an email, but usually still with a trigger from the user. In software testing, such an agent can perform tasks such as writing tests or analyzing code.

The Difference Between AI Tools and True Agents

Tools like GitHub Copilot and Cursor are rapidly conquering the development environment of developers. The difference between a standard AI tool and an agent lies in the degree of autonomy:

  • AI Tool: assists with writing code or tests, but the user decides when and what is executed.
  • AI Agent: can autonomously take actions within set boundaries. For example, writing a test, executing it, and reporting results.

Practical example: a tester used an IDE agent to generate hundreds of tests in an open-source project within a few weeks. The agent could read the codebase, set up test cases, and automate them. The result? Faster progress, satisfied stakeholders, and a massive boost in efficiency.

The Power of Prompt Engineering

The success of AI agents starts with prompt engineering: correctly and effectively instructing the AI. Good prompts contain clear context, purpose, situation, and expectations. This allows the agent to deliver relevant output without errors or misinterpretations.

Practical tip: invest in basic knowledge of prompt engineering. This will enable you to work more effectively with AI agents than with off-the-shelf prompt coaches, and you’ll maintain control over the testing process.

t cases, and automate them. The result? Faster progress, satisfied stakeholders, and a massive boost in efficiency.

AI as an Assistant, Not a Replacement

Another important lesson we have learned from practice: be aware that you remain ultimately responsible. AI can take over many tasks but will never bear ultimate responsibility. You must always check and understand what is happening. “Pairing with AI” or AI-assisted testing is the key concept here: the tester collaborates with the agent, maintains insight and control, and uses the AI as an assistant.

Dangers from excessive trust are real: tests or code can go wrong, and knowledge disappears if people become too dependent on AI. Therefore, start small, experiment in a controlled manner, and build up knowledge step-by-step.

Do’s and Don’ts for AI Agents

Do’s:

  • Use agents for testing activities; do not ignore them.
  • Be transparent about AI usage: mark code or tests that have been created by an agent.
  • Invest in prompt engineering and basic knowledge of the agent.
  • Start with web-based agents, which operate within safe boundaries.

Don’ts:

  • Do not blindly trust the agent; continue to check/verify.
  • Do not give agents unlimited access to sensitive data.
  • Do not use AI to replace fundamental knowledge or skills.

Experimenting with Safety and Awareness

AI is here to stay. Research has shown that 92% of developers use AI (SOURCE: GitHub). Realistically, the engineers and testers in your organization are also using agents. It is therefore important and necessary to establish an AI policy. The goal is not to automate away jobs, but to increase effectiveness and deliver value faster. Start with controlled experiments in your team, for example during a training session or hackathon. This way, you learn about the power and limitations of AI without running risks.

Conclusion

AI agents offer enormous opportunities for software testing: faster test generation, less manual work, and a boost in efficiency. But success depends on how you deploy them. Invest in basic knowledge, use AI as an assistant, and always maintain control over your processes.

By deploying agents consciously and responsibly, you get the most out of AI without losing control or knowledge. Testing with common sense—that is the key.

“Businesses who’ve stopped hiring and training entry-level developers because ‘GitHub Copilot can do what they do’ are going to find out what happens when nobody plants tomatoes because ‘Hey, who needs tomatoes? We’ve already got pasta sauce’.”

Do you have questions about this topic? Please feel free to contact us.

Misschien vind je de volgende artikelen ook wel interessant