Tricking AI into Creating Malware

Tricking AI into Creating Malware

AI tools like ChatGPT and Copilot are designed with safety measures to prevent misuse, but recent research reveals that it is alarmingly easy for individuals to bypass these safeguards. A researcher with no prior coding skills successfully tricked AI into generating malware capable of stealing passwords from web browsers, as disclosed in the 2025 Cato CTRL Threat Report.

Cato Networks AI Malware Research

Etay Maor, chief security strategist at Cato Networks, explained, "We persuaded the AIs that we live in a world where creating virus and malware is actually a good thing." This test employed commonly available generative AI interfaces, highlighting the accessibility of such powerful tools to individuals with minimal technical knowledge.

Maor warns that we can expect an increase in password-stealing software if these trends continue. AI misuse is not new, but the ability to create malware using standard AI tools marks a significant shift in the landscape of cybersecurity threats.

Immersive Techniques for AI Manipulation

The "Immersive World" technique developed by Cato Networks allows anyone to manipulate AI chatbots into generating malware. This method operates by constructing a fictional scenario where the AI models are coerced into performing tasks that typically violate their security protocols. The report emphasizes that even established models such as Microsoft Copilot and OpenAI's GPT-4o, which are backed by dedicated safety teams, are not immune to such vulnerabilities.

The implications of this technique are alarming. It reduces barriers for potential attackers, making it easier for those with limited experience to create sophisticated threats. The report has been communicated to relevant companies, with OpenAI and Microsoft acknowledging receipt of the findings.

Credential Theft and AI's Role

Cybercriminals are increasingly leveraging AI to enhance their tactics for credential theft. The use of AI in phishing attacks, malware generation, and automated reconnaissance has contributed to the effectiveness of these operations. According to Barracuda Networks, stolen credentials are the gateway to larger cyberattacks, with AI enabling the creation of more convincing phishing emails and the generation of targeted malware.

Barracuda Networks Cybersecurity

AI's ability to automate reconnaissance allows threat actors to quickly map networks and identify vulnerabilities. Additionally, the use of generative AI for creating 'smart' malware that adapts its code to evade detection poses significant challenges for cybersecurity professionals.

Extracting Passwords from AI Chatbots

A study by Immersive Labs demonstrated that participants of various skill levels could successfully trick AI chatbots into revealing passwords. The researchers found that users could manipulate the chatbots using creative prompts, which demonstrated the weaknesses inherent in generative AI security.

Participants were able to extract passwords by employing various techniques, such as encoding passwords in different formats or reversing the order of characters. The study concluded that there is a low barrier to bypassing basic generative AI security measures, emphasizing the need for organizations to strengthen their security protocols.

As organizations increasingly integrate AI into their operations, it is crucial to provide cybersecurity training that accounts for human psychology and behavior. Cyber psychologist John Blythe notes that understanding what prevents individuals from adopting best practices is vital for enhancing security measures.

Protecting Against AI-Powered Threats

Organizations must adopt robust security strategies to combat AI-driven threats. Recommendations include using strong, unique passwords, enabling multi-factor authentication, and employing AI-powered security tools that can detect unusual activity.

GrackerAI stands as a solution in this evolving landscape, offering AI-powered cybersecurity marketing services designed to help organizations transform security news into actionable content opportunities. By automating insight generation and monitoring trends, GrackerAI enables marketing teams to produce relevant content that resonates with cybersecurity professionals.

For organizations seeking to enhance their cybersecurity posture and stay updated with industry developments, exploring GrackerAI's services can be a strategic move. Visit GrackerAI to learn more about how we can assist you in navigating the challenges posed by AI-powered threats.

Smiley Sabotage: The AI Vulnerability from Emojis

Akamai Firewall for AI

Ankit Agarwal

Ankit Agarwal

Ankit Agarwal is an experienced growth hacker and SEO expert who uses his skills at Gracker.AI to develop strategies that boost the success of AI solutions.