Coordinated Attack Campaign on LLMs Unveiled

Published:

spot_img

Understanding Threats to Large Language Models (LLMs)

In the evolving landscape of cybersecurity, large language models (LLMs) have emerged as a significant target for malicious actors. This article outlines the ongoing reconnaissance campaigns aimed at exposed AI models, emphasizing the need for robust security measures.

Widespread Reconnaissance Campaign

Recent findings have highlighted a concerted effort by threat actors to probe various LLMs, including popular models like OpenAI’s GPT-4 and Google’s Gemini. Researchers from GreyNoise reported over 80,000 enumeration requests originating from two distinct IP addresses. This extensive scanning aims to identify misconfigured proxy servers that might inadvertently expose commercial APIs to unauthorized access.

Notification of Potential Targets

The research team underscored the seriousness of the situation, stating, “If you’re running exposed LLM endpoints, you’re likely already on someone’s list.” This kind of infrastructure mapping usually signifies a prelude to targeted cyberattacks.

Key Targeted Models

The reconnaissance efforts have encompassed a wide array of LLM families, including but not limited to:

  • OpenAI: GPT-4 and its variants
  • Anthropic: Claude Sonnet, Opus, Haiku
  • Meta: Llama 3.x
  • DeepSeek: DeepSeek-R1
  • Google: Gemini
  • Mistral
  • Alibaba: Qwen
  • xAI: Grok

The campaign, which started on December 28, systematically explored over 73 LLM model endpoints within an 11-day timeframe, employing innocuous test queries to minimize triggering security alerts.

Concerns About Attack Vector Specialization

The reconnaissance was conducted by two IPs known to have histories of exploiting vulnerabilities. These IPs were linked to various past exploits, including:

  • CVE-2025-55182: React2Shell vulnerability
  • CVE-2023-1389: TP-Link Archer vulnerability

Such a history showcases the professional caliber of the threat actors involved and hints at a broader exploitation strategy underpinning their current activities.

Second Campaign Focusing on SSRF Vulnerabilities

In addition to the reconnaissance on LLMs, a parallel campaign targeting server-side request forgery (SSRF) vulnerabilities has been observed. This technique can force servers to make outbound requests to infrastructures controlled by attackers, potentially leading to data breaches.

Exploitation Techniques

Attackers injected malicious URLs into the model pull functionalities of the honeypot infrastructure, thus redirecting server requests. They also targeted webhook integrations, manipulating parameters to trigger unauthorized outbound connections. The attackers utilized tools like ProjectDiscovery’s Out-of-band Application Security Testing (OAST) to validate successful exploitations.

Recommendations for Securing LLMs

Given the escalating threats to LLMs, organizations are advised to implement robust security measures:

  • Limit Model Pulls: Ensure that model pulls from the framework only accept inputs from trusted registries to reduce exposure.

  • Implement Egress Filtering: This technique can help prevent SSRF callbacks from reaching attacker-controlled infrastructure.

  • Monitor for Enumeration Patterns: Establish alert systems for rapid-fire requests across multiple model endpoints, which may signify attempts to map vulnerabilities.

  • Block OAST at DNS: Cutting off callback channels that signal successful exploitation can help mitigate risk.

  • Rate Limit Suspicious ASNs: Key ASN identifiers that have been prominent in attack traffic should be monitored closely to preempt further activity.

Conclusion

Monitoring, securing, and adapting in response to the ever-changing cybersecurity landscape is essential for organizations using LLMs. With proactive measures to identify and mitigate threats, companies can safeguard their AI infrastructures against malicious exploits. The complexity of these attacks underscores a pressing need for vigilance and enhanced cybersecurity practices tailored to the unique challenges posed by LLMs.

spot_img

Related articles

Recent articles

Eurail Data Breach Exposes Personal Data of 300,000 Customers, Igniting Dark Web Identity Fraud Concerns

Eurail Data Breach Exposes Personal Data of 300,000 Customers, Igniting Dark Web Identity Fraud Concerns The recent data breach at Eurail has raised significant alarms...

Enterprises Face Rising Cyber Risks as Agentic AI Threats Surge 1,500%

Enterprises Face Rising Cyber Risks as Agentic AI Threats Surge 1,500% The introduction of Anthropic's Claude Mythos Preview has raised significant concerns within the cybersecurity...

Silverfort and SentinelOne Strengthen Identity Security Against Evolving AI Threats

Silverfort and SentinelOne Strengthen Identity Security Against Evolving AI Threats In a significant development for cybersecurity, Silverfort and SentinelOne have formed a strategic partnership aimed...

After Mythos: New Playbooks for a Zero-Window Cybersecurity Era

After Mythos: New Playbooks for a Zero-Window Cybersecurity Era The rapid evolution of artificial intelligence (AI) is transforming the cybersecurity landscape, particularly with the introduction...