General-purpose LLMs show limited effectiveness in offensive security tasks
Summary
Hide â˛
Show âŧ
General-purpose large language models (LLMs) demonstrate limited effectiveness in offensive security tasks, particularly for complex vulnerabilities and exploits. While some LLMs can identify simple vulnerabilities and create basic exploits, they often fail at more intricate tasks. Tailor-made AI systems remain more effective for penetration testers and offensive researchers. The research involved testing 50 different LLMs, including commercial and experimental models, on vulnerability research and exploit development tasks. The findings highlight the current limitations of general-purpose LLMs in offensive security, despite their potential for initial triaging and development support. The investigation underscores the need for human oversight in AI-driven security efforts to mitigate errors and ensure accurate results.
Timeline
-
13.08.2025 22:08 đ° 1 articles
Research reveals limitations of general-purpose LLMs in offensive security
An investigation into the offensive-security capabilities of 50 different large language models (LLMs) found that while some models can handle simple vulnerabilities and exploits, they struggle with more complex tasks. Commercial LLMs showed better performance in vulnerability research tasks, but many experimental models were non-functional or behind paywalls. The research highlights the need for human oversight in AI-driven security efforts to ensure accuracy and effectiveness. The findings also underscore the potential of AI in penetration testing and application security, with expectations that AI systems will significantly impact these fields in the near future.
Show sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
Information Snippets
-
50 different LLMs were tested for offensive-security capabilities.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
Most LLMs can identify simple vulnerabilities and create basic exploits.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
Only a few LLMs successfully found complex vulnerabilities or created complex exploits.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
Tailor-made AI systems are more effective for penetration testing and offensive research.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
General-purpose LLMs may mislead non-experts into believing they have found actionable results.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
Commercial LLMs had the most success in solving vulnerability research tasks.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
Many experimental runs, especially for complex tasks, took hours or even a full workday to determine if the LLM could solve the problem.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
LLMs are useful for initial triaging of scanner findings and during development.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
AI systems are expected to significantly impact penetration testing and application security.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
Xbow's AI system discovered over 900 vulnerabilities on the HackerOne platform.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
-
AI-driven security efforts require human oversight to catch mistakes and hallucinations.
First reported: 13.08.2025 22:08đ° 1 source, 1 articleShow sources
- Popular AI Systems Still a Work-in-Progress for Security â www.darkreading.com â 13.08.2025 22:08
Similar Happenings
AI systems vulnerable to data-theft via hidden prompts in downscaled images
Researchers at Trail of Bits have demonstrated a new attack method that exploits image downscaling in AI systems to steal user data. The attack injects hidden prompts in full-resolution images that become visible when the images are resampled to lower quality. These prompts are interpreted by AI models as user instructions, potentially leading to data leakage or unauthorized actions. The vulnerability affects multiple AI systems, including Google Gemini CLI, Vertex AI Studio, Google Assistant on Android, and Genspark. The attack works by embedding instructions in images that are only revealed when the images are downscaled using specific resampling algorithms. The AI model then interprets these hidden instructions as part of the user's input, executing them without the user's knowledge. The researchers have developed an open-source tool, Anamorpher, to create images for testing this vulnerability. To mitigate the risk, Trail of Bits recommends implementing dimension restrictions on image uploads, providing users with previews of downscaled images, and requiring explicit user confirmation for sensitive tool calls.
Black Hat NOC Enhances AI in Security Operations
The Black Hat USA 2025 Network Operations Center (NOC) team expanded its use of AI and machine learning (ML) to manage and secure the conference network. The team monitored and mitigated malicious activities, distinguishing between legitimate and illegal activities, and identified vulnerabilities in applications and misconfigurations in security tools. The NOC team also observed trends such as increased self-hosting and insecure data transmission, which pose risks to organizations. The NOC team leveraged AI for risk scoring and categorization, ensuring that legitimate training activities were not flagged as malicious. They also identified and alerted attendees to security issues, such as misconfigured security tools and vulnerable applications, which could expose sensitive data. Additionally, high school students Sasha Zyuzin and Ruikai Peng presented a new vulnerability discovery framework at Black Hat USA 2025, combining static analysis with AI. Their framework, "Tree of AST," uses Google DeepMind's Tree of Thoughts methodology to automate vulnerability hunting while maintaining human oversight. The presenters discussed the double-edged nature of AI in security, noting that while LLMs can improve code quality, they can also introduce security risks.