Cybersecurity | 6/23/2025
Cybercriminals Exploit AI Models Like Grok and Mixtral for Malicious Activities
Cybercriminals are increasingly using sophisticated AI models, including Grok and Mixtral, to enhance their malicious activities. By jailbreaking these commercial AI systems, they are able to bypass safety features and create potent tools for cybercrime.
Cybercriminals Exploit AI Models Like Grok and Mixtral for Malicious Activities
Cybercriminals are advancing their tactics by utilizing sophisticated artificial intelligence (AI) models, marking a notable shift in the landscape of AI-driven threats. Tools such as WormGPT, which gained notoriety for generating convincing phishing emails and malicious code, have evolved into a brand associated with a new class of uncensored AI.
Recent investigations indicate that these malicious tools are not entirely new creations but rather adaptations of legitimate, high-end commercial large language models, including Grok from xAI and Mixtral from Mistral AI. Instead of developing custom AI, threat actors are jailbreaking existing systems to circumvent their ethical and safety protocols, effectively weaponizing them for criminal purposes.
The emergence of WormGPT in mid-2023 on underground forums marked a significant turning point, providing cybercriminals with an AI tool specifically designed for illicit activities. Initially based on the open-source GPT-J 6B model, WormGPT was offered on a subscription basis and trained on malware-related data to assist in creating malicious software and fraudulent content. Although the original WormGPT was shut down following media exposure, its name has since become synonymous with a range of new, more potent successors, such as FraudGPT and EvilGPT, which offer a wider array of malicious capabilities.
Security researchers have recently uncovered new variants of WormGPT being sold on dark web forums. These variants are not based on older models but are instead built around advanced commercial AI systems. For instance, one variant known as "keanu-WormGPT" operates using xAI's Grok, employing a custom system prompt to bypass safety features and generate harmful outputs like phishing emails. Another variant, "xzin0vich-WormGPT," utilizes Mistral AI's Mixtral model in a similar manner.
These tools are accessible via subscription on platforms like Telegram, allowing a broad range of criminals to utilize them, regardless of their technical skills. The implications of this trend are significant for both the AI industry and global cybersecurity. The ability of criminals to exploit powerful AI models underscores vulnerabilities in public-facing APIs and the challenges of enforcing safety measures.
Experts note that these safety protocols often serve more as obstacles than impenetrable barriers, allowing determined adversaries to find ways around them. This has led to the rise of a "jailbreak-as-a-service" market, lowering the entry barrier for cybercrime and enabling even novice attackers to execute sophisticated campaigns. The availability of uncensored models, which can be run locally without oversight, creates an imbalance where malicious actors can enhance their effectiveness at minimal cost.
In conclusion, the weaponization of high-end commercial AI models poses a growing threat. The transformation of WormGPT from a singular tool into a brand for jailbroken AI services illustrates the adaptability of the cybercriminal market. By leveraging the capabilities of models like Grok and Mixtral, attackers can produce increasingly sophisticated and convincing malicious content. As generative AI becomes more integrated into digital life, the challenge of preventing its misuse will require robust security protocols, continuous monitoring, and a deeper understanding of how these technologies can be exploited.