Coordinated Attack Campaign on LLMs Unveiled

Published:

spot_img

Understanding Threats to Large Language Models (LLMs)

In the evolving landscape of cybersecurity, large language models (LLMs) have emerged as a significant target for malicious actors. This article outlines the ongoing reconnaissance campaigns aimed at exposed AI models, emphasizing the need for robust security measures.

Widespread Reconnaissance Campaign

Recent findings have highlighted a concerted effort by threat actors to probe various LLMs, including popular models like OpenAI’s GPT-4 and Google’s Gemini. Researchers from GreyNoise reported over 80,000 enumeration requests originating from two distinct IP addresses. This extensive scanning aims to identify misconfigured proxy servers that might inadvertently expose commercial APIs to unauthorized access.

Notification of Potential Targets

The research team underscored the seriousness of the situation, stating, “If you’re running exposed LLM endpoints, you’re likely already on someone’s list.” This kind of infrastructure mapping usually signifies a prelude to targeted cyberattacks.

Key Targeted Models

The reconnaissance efforts have encompassed a wide array of LLM families, including but not limited to:

  • OpenAI: GPT-4 and its variants
  • Anthropic: Claude Sonnet, Opus, Haiku
  • Meta: Llama 3.x
  • DeepSeek: DeepSeek-R1
  • Google: Gemini
  • Mistral
  • Alibaba: Qwen
  • xAI: Grok

The campaign, which started on December 28, systematically explored over 73 LLM model endpoints within an 11-day timeframe, employing innocuous test queries to minimize triggering security alerts.

Concerns About Attack Vector Specialization

The reconnaissance was conducted by two IPs known to have histories of exploiting vulnerabilities. These IPs were linked to various past exploits, including:

  • CVE-2025-55182: React2Shell vulnerability
  • CVE-2023-1389: TP-Link Archer vulnerability

Such a history showcases the professional caliber of the threat actors involved and hints at a broader exploitation strategy underpinning their current activities.

Second Campaign Focusing on SSRF Vulnerabilities

In addition to the reconnaissance on LLMs, a parallel campaign targeting server-side request forgery (SSRF) vulnerabilities has been observed. This technique can force servers to make outbound requests to infrastructures controlled by attackers, potentially leading to data breaches.

Exploitation Techniques

Attackers injected malicious URLs into the model pull functionalities of the honeypot infrastructure, thus redirecting server requests. They also targeted webhook integrations, manipulating parameters to trigger unauthorized outbound connections. The attackers utilized tools like ProjectDiscovery’s Out-of-band Application Security Testing (OAST) to validate successful exploitations.

Recommendations for Securing LLMs

Given the escalating threats to LLMs, organizations are advised to implement robust security measures:

  • Limit Model Pulls: Ensure that model pulls from the framework only accept inputs from trusted registries to reduce exposure.

  • Implement Egress Filtering: This technique can help prevent SSRF callbacks from reaching attacker-controlled infrastructure.

  • Monitor for Enumeration Patterns: Establish alert systems for rapid-fire requests across multiple model endpoints, which may signify attempts to map vulnerabilities.

  • Block OAST at DNS: Cutting off callback channels that signal successful exploitation can help mitigate risk.

  • Rate Limit Suspicious ASNs: Key ASN identifiers that have been prominent in attack traffic should be monitored closely to preempt further activity.

Conclusion

Monitoring, securing, and adapting in response to the ever-changing cybersecurity landscape is essential for organizations using LLMs. With proactive measures to identify and mitigate threats, companies can safeguard their AI infrastructures against malicious exploits. The complexity of these attacks underscores a pressing need for vigilance and enhanced cybersecurity practices tailored to the unique challenges posed by LLMs.

spot_img

Related articles

Recent articles

DNS Attacks Explained: Risks and Threats You Need to Know

Understanding DNS Attacks: Safeguarding the Internet's Backbone As we navigate through our increasingly digital lives in 2026, one unsung hero stands at the forefront of...

Parsons Awarded Design and Construction Management Contract by New Murabba Development in Saudi Arabia

Parsons Awarded Major Contract for New Murabba Development in Riyadh CHANTILLY, VA. – Parsons Corporation (NYSE: PSN) has secured a significant contract from the New...

Endesa Alerts Customers About Data Breach Affecting Energía XXI

Endesa Reports Data Breach Affecting Energía XXI Customers Overview of the Incident Spanish energy company Endesa, along with its regulated operator Energía XXI, has begun reaching...