The EU AI Act and QA: What Your Testing Process Needs by August
By qtrl Team · Engineering
Five months from now, the EU AI Act hits full enforcement. August 2, 2026 is the date. If your team builds, deploys, or tests software that uses AI in any meaningful way, this regulation applies to you. And if you're selling into Europe, it doesn't matter where your company is based.
Most engineering teams have heard of the Act by now. Fewer have read it closely, and even fewer have mapped out what it means for their testing process. That's worth changing, because the Act doesn't just regulate what AI systems do. It regulates how you prove they work correctly. And that's QA's job.
What the Act actually says (the parts that matter for testing)
The EU AI Act classifies AI systems into four risk tiers: unacceptable, high, limited, and minimal. The unacceptable tier is straightforward. Social scoring, manipulative AI, certain forms of real-time facial recognition: banned outright.
The tier that affects most software teams is "high-risk." If your AI system touches healthcare, employment, education, critical infrastructure, or financial services, you're likely in this category. And high-risk comes with real obligations.
Articles 8 through 15 lay out what providers of high-risk AI systems must do: implement a risk management system, maintain technical documentation, build automatic logging capabilities, and establish a quality management system that enables continuous compliance. Article 17 goes further, mandating a formal QMS with 12 core components including testing, validation, and corrective action protocols.
Article 60 is the one QA leads should bookmark. It requires full traceability: logging AI behavior and maintaining technical documentation detailed enough to reconstruct how the system reached its decisions. If an auditor shows up and asks "how did your AI arrive at this output?" you need an answer backed by records.
The enforcement side is worth knowing about. Fines for prohibited practices reach up to 35 million euros or 7% of global annual turnover, whichever is higher. For non-compliance with provider obligations, it's 15 million euros or 3% of turnover. Italy has added criminal penalties for certain violations. The numbers are large enough that compliance teams are paying attention, which is why QA should be too.
QA now has a seat at the compliance table
For most of software history, QA existed to protect the user experience and prevent regressions. Important, but internal. The business cared about outcomes, not about how you got there.
Under the EU AI Act, your QA process is auditable. Regulators can ask to see your test plans, your execution records, your traceability from requirements to test results. That gives QA teams something they've wanted for years: real organizational leverage. How well you test now directly supports your company's compliance posture.
It also means QA leads need to think differently about what their function delivers. This is an opportunity to elevate testing from an internal quality check to a documented, auditable process that protects the entire business.
What your QA process needs to prove
Nothing in the Act prescribes specific testing tools or frameworks. It prescribes outcomes. Your QA process needs to demonstrate five things:
1. Traceability from requirements to test results
Every requirement that touches AI behavior needs a corresponding test. Every test needs a recorded result. And the chain between them needs to be auditable. Spreadsheets won't cut it here. You need structured test management that links requirements to test cases to execution results, with a clear history of what changed and when.
If you've read our earlier post on why structured test management still matters, this is exactly the scenario we described. The teams that ditched test management for Jira tickets and Notion pages are going to have a painful time proving compliance. It's one of the reasons we built qtrl with requirements traceability and immutable audit trails on every plan, including the free tier. When an auditor asks for the chain from requirement to test result, you can produce it in seconds.
2. Bias and accuracy validation
High-risk AI systems must be tested for discriminatory outputs. That means your test suite needs to include data sets that specifically probe for bias across protected characteristics. This isn't a one-time check. The Act expects ongoing validation, which means your regression testing needs to include bias checks that run continuously.
3. Data quality assurance
Training, validation, and test datasets must be "relevant, sufficiently representative, and to the best extent possible, free of errors and complete." That's a direct quote from the Act. Your QA scope now extends beyond the application layer into the data that feeds your AI models.
4. Human oversight documentation
The Act requires that high-risk AI systems allow effective human oversight. From a QA perspective, this means testing and documenting that humans can intervene, override, or shut down the AI system at any point. You need to verify that the override mechanisms actually work, and you need records proving you tested them.
This is something we think about a lot at qtrl. Our platform is built around permissioned autonomy: you control what the AI agents can do, review and approve tests before they run, and every action is logged. That same philosophy applies when you're testing your own product's AI features. The override and kill-switch tests need to be first-class test cases, not afterthoughts, and the results need to live somewhere an auditor can find them.
5. Post-deployment monitoring
Testing doesn't stop at release. Article 72 requires lifecycle monitoring and regular reporting after launch. Your QA process needs to extend into production, with ongoing tests that verify the AI system continues to behave as documented.
This connects directly to the shift-right testing trend we're seeing across the industry. Teams that already run tests against production environments are better positioned than those whose QA ends at the staging gate. In qtrl, you can run the same test suite across development, staging, and production environments, with per-environment variables and encrypted secrets. That means your compliance-critical tests don't just run before release. They keep running after it.
The documentation gap is the biggest opportunity
Most teams that use AI in their products do test it. The good news is the testing is already happening. The gap is documentation: most of that testing isn't recorded in a way that satisfies an auditor.
A Gartner study found that 68% of software engineering functions develop AI-powered features, but only 3% have established documented processes with full adherence. That's a 65-point gap between "we test" and "we can prove we test."
Annex IV of the Act requires comprehensive technical documentation: design decisions, data lineage, testing methodologies, validation results. If your team practices agile with minimal documentation, building these records retroactively takes real effort. It's much cheaper to start documenting as you go than to reconstruct history later.
This is where your choice of test management tooling actually matters. If your tests live in scripts on someone's machine, in a shared Google Doc, or scattered across Jira subtasks, producing a clean audit trail will be difficult. Structured test management with built-in traceability closes that gap. For teams shipping AI features into Europe, it's the most direct path to compliance readiness.
AI testing tools need to be auditable too
Here's an irony the industry hasn't fully reckoned with: many teams are starting to use AI to test AI. That creates a second layer of compliance. If your testing tool itself uses AI (autonomous agents, AI-generated test cases, intelligent exploration), the Act's transparency requirements apply to your testing process, not just the system under test.
We wrote about the gap between AI testing hype and reality earlier this year. One point from that piece is especially relevant here: black-box AI testing, where you can't explain what the AI tested or why, won't hold up under the Act. If an auditor asks why your AI agent tested a particular flow, you need a better answer than "it decided on its own." The tools that give you that answer are the ones worth investing in.
AI-powered testing tools need to provide visibility into what they're doing and why. Governance and full audit trails aren't nice-to-haves on a feature page. They're compliance infrastructure. It's why we built qtrl with full agent visibility from day one: every action the AI takes is logged, every decision is traceable, and autonomy levels are permissioned so your team stays in control. When you're using AI to test AI, you need that second layer of accountability baked into the tool itself.
What to do before August
Five months is tight, but it's enough to close the worst gaps.
Start by auditing your AI inventory. Identify every feature in your product that uses AI or ML and map each one to the Act's risk tiers. If you're not sure whether something qualifies as high-risk, assume it does until you've confirmed otherwise.
Then pressure-test your documentation. For each AI feature, ask: can we produce a complete record of what was tested, when, by whom, and what the results were? If the answer is no, that's your highest-priority gap.
Link your requirements to test cases to results. This is table-stakes for compliance. If your current tooling doesn't support this natively, you need to either build the workflow or switch tools. And don't treat bias validation as a one-time audit. Build it into your CI/CD pipeline so it runs on every release.
Get QA, legal, and security in the same room. The Act sits at the intersection of all three. QA can't solve this alone, and legal can't solve it without understanding what QA actually tests.
Finally, if you're using AI-powered testing (or plan to), check that the tool provides full visibility into agent behavior. Can you see what the agent tested? Can you explain its decisions? Can you produce a log that an auditor would accept? The right tooling makes compliance much easier to demonstrate.
The teams that prepare early will move faster
We saw this with GDPR. The organizations that built privacy into their systems early adapted smoothly and kept shipping. The ones that waited spent more time and money catching up.
The same will be true here. Teams that build auditable QA processes now will ship faster in the second half of 2026 because the compliance work is already done. QA roles with compliance expertise are already commanding 20 to 40% salary premiums. The market is pricing in the value of this skill set.
For QA teams, this is a chance to own something bigger. You're not only catching bugs. You're proving that your AI systems are safe and traceable, with records to back it up. The scope is larger, and so is the organizational authority that comes with it.
We built qtrl for exactly this moment: structured test management with requirements traceability, immutable audit trails on every plan, and AI-powered testing where every agent action is logged and governed. If your team is shipping AI features into Europe and needs to close the compliance gap before August, try it free and see how it fits your workflow.