Uncovering DIG AI: The Dark Web’s Threat to Cybersecurity
What is DIG AI?
A new artificial intelligence tool named DIG AI has emerged on the dark web, raising significant concerns among cybersecurity experts. Developed by an individual known as “Pitch,” DIG AI operates on the Tor network, providing an uncensored platform for illegal activities. Resecurity, a US-based cybersecurity firm, first detected this tool on September 29, 2025, and has since noted a surge in its utilization, especially during the holiday season—a time when illicit activities traditionally peak.
DIG AI functions without the need for user accounts, ensuring anonymity for its users. This accessibility positions it as a dangerous resource for cybercriminals and extremists alike. With its basis in ChatGPT Turbo—a popular large language model—DIG AI has been marketed across dark web forums associated with drug trafficking and the sale of stolen financial data, demonstrating a clear intent to target criminal audiences.
How is DIG AI Being Used?
Research conducted by Resecurity’s HUNTER team indicates that DIG AI can facilitate a range of illegal activities. Tests using dictionaries related to explosives, drugs, and other criminal undertakings revealed that the system is capable of:
- Offering detailed instructions on manufacturing explosives and producing illegal substances.
- Generating large-scale fraudulent content, such as phishing emails and social engineering messages.
- Creating malicious scripts intended to exploit vulnerabilities in web applications.
Criminals can further enhance these activities by connecting DIG AI to external APIs, allowing them to automate and scale their operations, thereby reducing the need for specialized skills.
The Rise of “Not Good” AI and Dark LLMs
The term “not good AI” has been coined to describe systems like DIG AI that facilitate criminal behavior. Resecurity reported a staggering 200% increase in the use of malicious AI tools on cybercrime forums between 2024 and 2025. Tools like FraudGPT and WormGPT are among the most well-known, being utilized for assorted illegal activities: generating phishing emails, crafting malware, and providing advice on exploiting stolen data.
Unlike mainstream AI platforms like ChatGPT, which enforce strict content policies, DIG AI operates outside of these safeguards. This lack of regulation allows it to assist in various criminal enterprises with significant effectiveness.
AI-Generated Child Sexual Abuse Material
One of the most alarming uses of DIG AI is in the creation of child sexual abuse material (CSAM). The tool can produce hyper-realistic depictions of minors by:
- Generating synthetic images or videos from text descriptions.
- Altering benign photographs into explicit materials.
Research has shown that law enforcement agencies have had to intervene as evidence mounts regarding the misuse of DIG AI for these purposes. As authorities tackle the challenges posed by AI-generated CSAM, several nations have begun enacting stricter laws to address the exploitation of this technology.
Beyond Cybercrime: Extremism and Terrorism
Resecurity also warns that DIG AI could empower extremist groups and terrorist organizations. The uncensored nature of this AI tool might aid in creating propaganda, recruitment materials, and operational guides. Additionally, DIG AI can assist these groups in tailoring their messages for distinct audiences, complicating efforts to counteract such threats.
Identifying users of DIG AI is notably challenging due to its anonymous structure, making early intervention by law enforcement difficult. This anonymity poses a significant threat as it enables the proliferation of violent ideologies and dangerous content.
Legal and Regulatory Gaps
While mainstream platforms face increasing scrutiny and regulatory measures regarding AI, hidden networks like Tor are largely unregulated. Legislative frameworks—such as the EU’s AI Act and proposed US initiatives—do not adequately extend to these dark web services, leaving significant gaps for tools like DIG AI to operate unchallenged.
As countries create stricter AI regulations, the ongoing existence of uncensored AI systems in unregulated spaces will present a formidable challenge for law enforcement. How governments approach these gaps will be crucial in combatting the misuse of AI technology.
A New Era of Underground Economies
Looking forward, security experts predict the emergence of new underground economies revolving around AI-enabled crime. The manipulation of datasets for the purpose of training AI models can lead to the rapid normalization of illegal content. By creating AI tools tailored for illicit activities, offenders could generate virtually unlimited illegal material, escaping detection by mainstream platforms.
As we transition into this new phase, the potential for AI models to become specific tools for illegal acts raises urgent questions. Authorities and responsive measures will need to evolve swiftly as the technological landscape continues to change.
In summary, as we approach 2026 and beyond, the dire implications of tools like DIG AI for society, security, and law enforcement become starkly clear. The proactive steps taken today will dictate our ability to navigate this complex and dangerous future.


