OWASP APTS Marks a Turning Point for Autonomous Pentesting
OWASP’s new Autonomous Penetration Testing Standard (APTS) establishes governance, safety, and accountability requirements for AI-driven pentesting platforms operating in production environments.
OWASP’s new Autonomous Penetration Testing Standard codifies what enterprise security teams have already been demanding from autonomous platforms running against production environments. That’s a good thing for the category, and for the buyers evaluating it.
OWASP just published APTS, the Autonomous Penetration Testing Standard. For an industry that has spent the last two years arguing about what AI penetration testing actually is, this is a meaningful moment.
Until now, there has been no shared baseline for how autonomous security testing systems should behave: how they enforce scope, how they fail safely, how they prove what they did and why. APTS gives the industry that baseline.
What is Covered Under APTS?
Per the documentation, APTS is a governance standard, not a testing methodology. It is designed to augment and enhance existing standards (PTES, OWASP WSTG, OSSTMM) by covering scope enforcement, safe autonomy, manipulation resistance, and accountability.
The scope of the standard is precise. It applies to systems that make their own choices about targeting, methodology, or exploitation; that operate against production or production-like environments; that have the capacity to cause unintended impact or expose data; and that need governance oversight to be trusted by the organizations using them.
SaaS platforms, on-premises tools, integrated security orchestration platforms with autonomous testing layers, and in-house systems that enterprise teams build for testing their own organizations are all in scope. Manual pentesting is not. Neither are SAST/DAST tools, bug bounty programs, human-led red team exercises, or vulnerability disclosure programs. APTS draws a clean line around the category that didn’t exist as a coherent thing five years ago, and has been expanding rapidly ever since.
What This Means for Buyers
If you’re evaluating autonomous penetration testing platforms, APTS gives you a concrete framework for the conversation. It provides a comprehensive domain-by-domain evaluation guide, based on the eight domains defined in the standard (Scope Enforcement, Safety Controls & Impact Management, Human Oversight & Intervention, Graduated Autonomy Levels, Auditability & Reproducibility, Manipulation Resistance, Third-Party & Supply Chain Trust, Reporting).
The full list of questions is worth a deep dive by your security team and/or or your security leadership, but in summary, it encourages you to ask vendors…
- how they enforce scope at runtime, not just how they accept it as input.
- what their safety guardrails actually do, with specifics.
- how they audit their own model’s decisions.
- whether they can show you a preliminary test plan before any offensive action is taken, and whether you can approve or modify it.
- what happens when the system encounters something ambiguous: does it act, does it pause, does it escalate, and how is that path governed?
Where Novee Stands
We built Novee around the conviction that autonomous offensive security is the only sustainable answer to attackers operating with the aid of AI. We also built it knowing that autonomy without governance is unworkable.
From the beginning, every test run on the platform begins with a preliminary report showing exactly what will be tested and how, so customers can review and approve scope before any offensive action is taken. Guardrails are configurable at the level of rate limits, time windows, URL exclusions, and explicit prevention of destructive actions. The system is designed to demonstrate exploitability without causing damage, using proof-of-concept payloads that validate vulnerabilities while avoiding sensitive data extraction. Findings, decisions, and actions are all auditable.
None of these choices were responses to a standard. They were responses to what enterprise security teams told us they needed to trust an autonomous platform running in their production environments. APTS now codifies many of those expectations, and we think that’s the right outcome for the industry.
Autonomy Is Here. Governance Is What Makes It Credible.
Autonomous penetration testing is no longer a research direction or a category to be dismissed. It’s an operational reality at a growing number of enterprises, and the case for continuous adversarial validation has only strengthened as time-to-exploit has compressed and AI-equipped attackers have proliferated. A standard that draws a clear line around what these systems must do to be trustworthy is exactly what the category needs to mature.
Trust in autonomous systems doesn’t come from marketing claims; it comes from demonstrable, auditable, governable behavior. APTS pushes the entire industry in that direction, which is good for buyers, good for defenders, and good for the future of offensive security.
To see how Novee finds the vulnerabilities that cause real breaches – within rigorous governance boundaries – schedule a demo.