When Discovery Is Cheap, Validation Is Priceless for Reducing Risk

Naz Bozdemir
Lead Product Researcher
Image
Brains on Digital Background

When discovery outpaces validation, security teams do not automatically reduce more risk. They accumulate more findings. The value only shows up when teams can determine which issues are actually exploitable and drive them through remediation.

Last month, Claude surfaced 22 confirmed vulnerabilities in Mozilla’s codebase, including 14 rated high severity. But at scale, the constraint shifts from simply surfacing bugs to proving exploitability and driving real fixes.

Of the 22 confirmed vulnerabilities, two were proven to be exploitable. That distinction matters. A confirmed defect is not the same as a material security risk. A score can’t tell you if you’re truly exposed. Only validation against your app, environment, and controls can.

This is not a criticism of AI. It is a preview of what every security team will face as AI-driven discovery scales: a sharp increase in confirmed findings, paired with an urgent need to determine which ones represent actual business risk and close them before adversaries do.

The Cost of Discovery Is Falling. The Cost of Inaction Is Not.

Pre-production remediation is roughly 30 times cheaper than post-production remediation. When you measure security investment through Return on Mitigation (RoM), the financial impact of catching and fixing vulnerabilities early is significant and traceable.

AI is about to multiply that math: if automated discovery can surface vulnerabilities faster and earlier, the per-finding cost of remediation drops, and the volume of mitigated risk goes up. That is the optimistic case, and it is real. But it only holds if the findings get validated and fixed.

Volume Without Resolution Is a Backlog, Not a Strategy

AI discovery tools are already capable of producing more findings than most security teams can process. 

The Mozilla example generated 112 reports, with 22 confirmed and two exploitable. At enterprise scale, without the capacity to validate and route fixes, discovery at this volume turns into a backlog.

The 34-day median resolution lifecycle HackerOne sees across penetration test findings is getting longer, not shorter, and this pressure shows up in remediation timelines. As finding volume increases, time-to-resolution does not automatically improve. In many programs, it gets harder to move quickly because teams are sorting through more inputs before they know what truly demands action.

The bottleneck was never just finding vulnerabilities. It was the ability to prove which findings are actually exploitable in your environment and ensure those issues are fully fixed. AI is accelerating the first step; it is exposing how hard the second step still is. 

What the Full Loop Actually Looks Like

A UK-based insurance provider found three SQL injection vulnerabilities during a scheduled pentest before a production release. The question was whether this was an isolated defect or a systemic pattern.

By extending the engagement with an agentic testing layer, the team systematically evaluated related query paths across the application. The result: 16 validated SQL injection findings, up from three. Zero false positives. Every finding included reproducible proof of exploitability confirmed by a human researcher.

Then came the part most teams skip: after developers applied fixes, follow-up agentic retesting caught cases where individual injection paths had been patched but related code paths remained vulnerable. Without that step, the team would have shipped incomplete remediation.

Discovery, validation, remediation, verification. One cycle that eliminated a systemic vulnerability class before it reached production, without expanding scope or budget. 

Where AI and Human Expertise Compound

The most productive framing is not "AI versus humans." It is understanding where each multiplies the other.

  • Discovery is where AI helps most today: broader coverage, faster pattern recognition, and more edge cases than rule-based scanners catch. Human researchers add depth by chaining issues and mapping real attack paths.
  • Validation is the bottleneck. AI can narrow, cluster, and enrich findings, but proving exploitability in your environment still requires adversarial judgment.
  • Remediation is where risk actually decreases. AI can generate fix guidance, route findings to the right teams, and track resolution status. Humans confirm the root cause is addressed and verify through retesting that the exposure is closed, not just patched at the surface.

This is the operational spine of Continuous Threat Exposure Management (CTEM). The framework is useful because it names the full cycle. But the insight that matters is simpler: faster discovery only compounds into risk reduction when validation is accurate, and remediation is confirmed.

Return on Mitigation Only Works on Closed Loops

AI is making discovery cheap, but Return on Mitigation is only created when a finding results in confirmed risk reduction. That means the issue was not just found, but validated as materially exploitable, remediated at the root cause, and verified as closed.

If a team discovers 500 vulnerabilities and remediates 50, the RoM calculation runs on 50 that are of value. The other 450 may represent future risk reduction, but not realized impact. That is also why closed-loop measurement matters. Impact should be attributed to findings that were validated, fixed, and confirmed, not simply logged into a system of record. Across HackerOne programs, estimated breach losses avoided reached $3 billion in 2025, representing a 15x return on mitigation.

Many teams will build AI-assisted discovery into their development workflows, and they should. But most of the operational difficulty sits downstream: normalizing findings across sources, validating exploitability, prioritizing by actual risk, routing work into engineering systems, and retesting fixes. That is where programs either turn discovery into a measurable reduction or drown in output.

An in-house AI integration improves the front of the pipeline. But measurable risk reduction depends on the back half: exploitability validation, remediation discipline, and verification that the fix holds.

The Metric That Matters

AI will keep driving down the cost of discovery, making everything downstream more important, not less.

The question for security leaders is not whether to adopt AI for discovery, because the answer is yes. The question is whether your organization can close the loop at the speed that AI-accelerated discovery now demands: validate exploitability quickly, remediate the right issues, and verify that fixes actually deliver real risk reduction.

The teams that win will not be the ones with the longest list of findings. They will be the ones who can quickly answer three questions for every important issue: Is it truly exploitable? Does it matter here? Has it actually been fixed?

In an era of near-free discovery, the answers to these questions are what matter. HackerOne operates across the full cycle, from discovery through validated remediation, measuring financial impact through RoM at every step.

See what that looks like in practice

 

Closing the AI Security Gap: Containing Risk Before It Scales

Survey methodology: HackerOne surveyed 303 security leaders between January and February 2026. Respondents were screened to ensure they oversee or contribute to tracking, managing, or testing their organization’s AI/ML systems, and represent a range of senior security and offensive security roles within organizations reporting $250 million or more in revenue across the United States, Canada, the United Kingdom, Australia, Singapore, and Germany. Respondents represented multiple industries, led by Technology Hardware/Software (37%) and Banking/Financial Services/Insurance (16%), with additional representation across manufacturing, healthcare, retail/e-commerce, and other sectors.

About the Author

Naz Bozdemir Headshot
Naz Bozdemir
Lead Product Researcher

Naz Bozdemir is the Lead Product Researcher for Research at HackerOne. She holds an MA and an MSC in cybersecurity and international relations.