Coordinated Attack Campaign on LLMs Unveiled

Published:

spot_img

Understanding Threats to Large Language Models (LLMs)

In the evolving landscape of cybersecurity, large language models (LLMs) have emerged as a significant target for malicious actors. This article outlines the ongoing reconnaissance campaigns aimed at exposed AI models, emphasizing the need for robust security measures.

Widespread Reconnaissance Campaign

Recent findings have highlighted a concerted effort by threat actors to probe various LLMs, including popular models like OpenAI’s GPT-4 and Google’s Gemini. Researchers from GreyNoise reported over 80,000 enumeration requests originating from two distinct IP addresses. This extensive scanning aims to identify misconfigured proxy servers that might inadvertently expose commercial APIs to unauthorized access.

Notification of Potential Targets

The research team underscored the seriousness of the situation, stating, “If you’re running exposed LLM endpoints, you’re likely already on someone’s list.” This kind of infrastructure mapping usually signifies a prelude to targeted cyberattacks.

Key Targeted Models

The reconnaissance efforts have encompassed a wide array of LLM families, including but not limited to:

  • OpenAI: GPT-4 and its variants
  • Anthropic: Claude Sonnet, Opus, Haiku
  • Meta: Llama 3.x
  • DeepSeek: DeepSeek-R1
  • Google: Gemini
  • Mistral
  • Alibaba: Qwen
  • xAI: Grok

The campaign, which started on December 28, systematically explored over 73 LLM model endpoints within an 11-day timeframe, employing innocuous test queries to minimize triggering security alerts.

Concerns About Attack Vector Specialization

The reconnaissance was conducted by two IPs known to have histories of exploiting vulnerabilities. These IPs were linked to various past exploits, including:

  • CVE-2025-55182: React2Shell vulnerability
  • CVE-2023-1389: TP-Link Archer vulnerability

Such a history showcases the professional caliber of the threat actors involved and hints at a broader exploitation strategy underpinning their current activities.

Second Campaign Focusing on SSRF Vulnerabilities

In addition to the reconnaissance on LLMs, a parallel campaign targeting server-side request forgery (SSRF) vulnerabilities has been observed. This technique can force servers to make outbound requests to infrastructures controlled by attackers, potentially leading to data breaches.

Exploitation Techniques

Attackers injected malicious URLs into the model pull functionalities of the honeypot infrastructure, thus redirecting server requests. They also targeted webhook integrations, manipulating parameters to trigger unauthorized outbound connections. The attackers utilized tools like ProjectDiscovery’s Out-of-band Application Security Testing (OAST) to validate successful exploitations.

Recommendations for Securing LLMs

Given the escalating threats to LLMs, organizations are advised to implement robust security measures:

  • Limit Model Pulls: Ensure that model pulls from the framework only accept inputs from trusted registries to reduce exposure.

  • Implement Egress Filtering: This technique can help prevent SSRF callbacks from reaching attacker-controlled infrastructure.

  • Monitor for Enumeration Patterns: Establish alert systems for rapid-fire requests across multiple model endpoints, which may signify attempts to map vulnerabilities.

  • Block OAST at DNS: Cutting off callback channels that signal successful exploitation can help mitigate risk.

  • Rate Limit Suspicious ASNs: Key ASN identifiers that have been prominent in attack traffic should be monitored closely to preempt further activity.

Conclusion

Monitoring, securing, and adapting in response to the ever-changing cybersecurity landscape is essential for organizations using LLMs. With proactive measures to identify and mitigate threats, companies can safeguard their AI infrastructures against malicious exploits. The complexity of these attacks underscores a pressing need for vigilance and enhanced cybersecurity practices tailored to the unique challenges posed by LLMs.

spot_img

Related articles

Recent articles

Webinar: Uncovering Suspicious APK Files in Wedding Card and Loan App Scams

The surge of malicious APK files in cyber fraud schemes, such as fake wedding invitations and instant loan applications, has become a growing concern....

Skylon Partners with COBNB to Launch COBNB+ Featuring L’Occitane en Provence Hotel Amenities

Skylon Partners with COBNB for a Luxurious Hospitality Experience in Kuala Lumpur Introduction to the New Partnership In an exciting development for the hospitality scene in...

Understanding CISA KEV: Key Insights and Tools for Security Teams

Understanding the CISA Known Exploited Vulnerability (KEV) Catalog The Cybersecurity and Infrastructure Security Agency (CISA) maintains the Known Exploited Vulnerability (KEV) catalog, a resource designed...

Dark Web Leak Sparks WFH Job Scams; Prayagraj Police Freeze ₹2 Crore in Fraudulent Funds

Rising Cybercrime in Prayagraj: A New Target Shifting Tactics of Cybercriminals In Prayagraj, the landscape of cybercrime is evolving. Previously, scammers predominantly targeted victims through enticing...