XBOW's AI-Based Penetration Testing Tool Achieves Top Spot on HackerOne Leaderboard
Summary
Hide â˛
Show âŧ
XBOW's AI-powered penetration testing tool became the first non-human bug hunter to reach the top of HackerOne's US leaderboard in June 2025. The tool uses a deterministic validation approach to minimize false positives, which has been a significant issue with other AI-driven vulnerability detection methods. The tool's success highlights the potential of AI in cybersecurity when used correctly. The tool employs a capture-the-flag (CTF) approach, placing 'canaries' in the source code to detect vulnerabilities. It was tested on approximately 17,000 synthesized web applications from Docker Hub, identifying 174 reported vulnerabilities, including 22 confirmed CVEs, and over 650 potential flaws still under investigation.
Timeline
-
13.08.2025 16:22 đ° 1 articles
XBOW's AI Tool Reaches Top of HackerOne Leaderboard
In June 2025, XBOW's AI-powered penetration testing tool became the first non-human bug hunter to reach the top of HackerOne's US leaderboard. The tool uses a deterministic validation approach to minimize false positives, which has been a significant issue with other AI-driven vulnerability detection methods. The tool's success highlights the potential of AI in cybersecurity when used correctly.
Show sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
Information Snippets
-
XBOW's AI-powered penetration testing tool reached the top of HackerOne's US leaderboard in June 2025.
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
-
The tool uses a deterministic validation approach to verify vulnerabilities, reducing false positives.
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
-
The tool employs a capture-the-flag (CTF) approach, placing 'canaries' in the source code to detect vulnerabilities.
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
-
The tool was tested on approximately 17,000 synthesized web applications from Docker Hub.
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
-
The tool identified 174 reported vulnerabilities, including 22 confirmed CVEs, and over 650 potential flaws still under investigation.
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
-
The tool has reported 285 total vulnerabilities on HackerOne in 2025.
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
-
AI-driven vulnerability detection methods often produce false positives, which can overwhelm security teams.
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
-
XBOW's approach minimizes false positives by using deterministic validation rather than relying solely on large language models (LLMs).
First reported: 13.08.2025 16:22đ° 1 source, 1 articleShow sources
- How an AI-Based 'Pen Tester' Became a Top Bug Hunter on HackerOne â www.darkreading.com â 13.08.2025 16:22
Similar Happenings
Cursor AI editor autoruns malicious code in repositories
A flaw in the Cursor AI editor allows malicious code in repositories to autorun on developer devices. This vulnerability can lead to malware execution, environment hijacking, and credential theft. The issue arises from Cursor disabling the Workspace Trust feature from VS Code, which prevents automatic task execution without explicit user consent. The flaw affects one million users who generate over a billion lines of code daily. The Cursor team has decided not to fix the issue, citing the need to maintain AI and other features. They recommend users enable Workspace Trust manually or use basic text editors for unknown projects. The flaw is part of a broader trend of prompt injections and jailbreaks affecting AI-powered coding tools.
HexStrike AI Exploits Citrix Vulnerabilities Disclosed in August 2025
Threat actors have begun using HexStrike AI to exploit Citrix vulnerabilities disclosed in August 2025. HexStrike AI, an AI-driven security platform, was designed to automate reconnaissance and vulnerability discovery for authorized red teaming operations, but it has been repurposed for malicious activities. The exploitation attempts target three Citrix vulnerabilities, with some threat actors offering access to vulnerable NetScaler instances for sale on darknet forums. The use of HexStrike AI by threat actors significantly reduces the time between vulnerability disclosure and exploitation, increasing the risk of widespread attacks. The tool's automation capabilities allow for continuous exploitation attempts, enhancing the likelihood of successful breaches. Security experts emphasize the urgency of patching and hardening affected systems to mitigate the risks posed by this AI-driven threat. HexStrike AI's client features a retry logic and recovery handling to mitigate the effects of failures in any individual step on its complex operations. HexStrike AI has been open-source and available on GitHub for the last month, where it has already garnered 1,800 stars and over 400 forks. Hackers started discussing HexStrike AI on hacking forums within hours of the Citrix vulnerabilities disclosure. HexStrike AI has been used to automate the exploitation chain, including scanning for vulnerable instances, crafting exploits, delivering payloads, and maintaining persistence. Check Point recommends defenders focus on early warning through threat intelligence, AI-driven defenses, and adaptive detection.
AI systems vulnerable to data-theft via hidden prompts in downscaled images
Researchers at Trail of Bits have demonstrated a new attack method that exploits image downscaling in AI systems to steal user data. The attack injects hidden prompts in full-resolution images that become visible when the images are resampled to lower quality. These prompts are interpreted by AI models as user instructions, potentially leading to data leakage or unauthorized actions. The vulnerability affects multiple AI systems, including Google Gemini CLI, Vertex AI Studio, Google Assistant on Android, and Genspark. The attack works by embedding instructions in images that are only revealed when the images are downscaled using specific resampling algorithms. The AI model then interprets these hidden instructions as part of the user's input, executing them without the user's knowledge. The researchers have developed an open-source tool, Anamorpher, to create images for testing this vulnerability. To mitigate the risk, Trail of Bits recommends implementing dimension restrictions on image uploads, providing users with previews of downscaled images, and requiring explicit user confirmation for sensitive tool calls.
UNC5518 deploys CORNFLAKE.V3 backdoor via ClickFix and fake CAPTCHA pages
UNC5518, an access-as-a-service threat actor, deploys the CORNFLAKE.V3 backdoor using the ClickFix social engineering tactic and fake CAPTCHA pages. This backdoor is used by at least two other groups, UNC5774 and UNC4108, to initiate multi-stage infections and drop additional payloads. The attack begins with users being tricked into running a malicious PowerShell script via a fake CAPTCHA page. The script executes a dropper payload that ultimately launches CORNFLAKE.V3, which supports various payload types and collects system information. The backdoor has been observed in both JavaScript and PHP versions and uses Cloudflare tunnels to avoid detection. A new ClickFix variant manipulates AI-generated text summaries to deliver malicious commands, turning AI tools into active participants in social engineering attacks.
PromptFix exploit enables AI browser deception
A new prompt injection technique, PromptFix, tricks AI-driven browsers into executing malicious actions by embedding hidden instructions in web pages. The exploit targets AI browsers like Perplexity's Comet, Microsoft Edge with Copilot, and OpenAI's upcoming 'Aura', which automate tasks such as online shopping and email management. PromptFix can deceive AI models into interacting with phishing sites or fraudulent storefronts, potentially leading to unauthorized purchases or credential theft. The technique exploits the AI's design goal to assist users quickly and without hesitation, creating a new scam landscape called Scamlexity. Researchers from Guardio Labs demonstrated the exploit by tricking Comet into adding items to a cart and auto-filling payment details on fake shopping sites. Similar attacks can manipulate AI browsers into parsing spam emails and entering credentials on phishing pages. PromptFix can also bypass CAPTCHA checks to download malicious payloads without user involvement. The exploit highlights the need for robust defenses in AI systems to anticipate and neutralize such attacks, including phishing detection, URL reputation checks, and domain spoofing protections. Until security matures, users should avoid assigning sensitive tasks to AI browsers and manually input sensitive data when needed. AI browser agents from major AI firms failed to reliably detect the signs of a phishing site. AI agents are gullible and servile, making them vulnerable to attacks in an adversarial setting. Companies should move from "trust, but verify" to "doubt, and double verify" until an AI agent has shown it can always complete a workflow properly. AI companies are not expected to pause developing more functionality to improve security. Companies should hold off on putting AI agents into any business process that requires reliability until AI-agent makers offer better visibility, control, and security. Securing AI requires gaining visibility into all AI use by company workers and creating an AI usage policy and a list of approved tools.