AI Worms: A New Threat to Private Data Security

    by | Mar 8, 2024

    BREAKING NEWS! The cybersecurity eco-system is at high risk with the recent development of new and unknown cyber threat, called AI Worms. Highly dangerous software designed to exploit the vulnerabilities in all AI systems. While the news sounds like a fictional movie, unfortunately, it isn’t a movie; it’s actually a reality. However, experts are announcing severe caution and providing proactive defense measures to stay safe and protected.

    What are AI Worms?

    In ancient times, and today as well, old computer worms used to exploit the software vulnerabilities in devices or systems. But AI worms target AI models such as email chatbots powered by popular language models, such as Gemini, Copilot, ChatGPT, etc.

    As per the researchers, they found the AI worm has a similar concept to the 1988 Morris worm. The name of the current threat development is known as “Morris II.”

    How Does AI Worms Work?

    Morris II works by manipulating and tricking AI assistant models to reveal sensitive information about the user through cleverly written prompts. In simple language, Morris II’s dangerous trick, called “Adversarial Self Replication,” attacks email systems by messaging again and again. This trick confuses AI assistants, and they end up accessing and changing personal information or spreading malware.

    Not only this, this AI worm implements two tricks: one is text-based, in which it hides harmful prompts in emails, and the other is image-based, which is coded with mysterious prompts to confuse the AI assistant to destroy security, and spread malware. Therefore, it potentially creates a domino effect of data breaches.

    What can be Done to Stay Safe?

    Before the threat spreads rapidly and becomes stronger, experts have developed a few useful precautions to stay protected and mitigate future risks.


    Companies, even small and large enterprises, when developing and deploying AI systems, should strictly adhere to strong security protocols and conduct daily vulnerability assessments. Also, analyze the system with ethical hacking tactics to identify weak points and resolve them as soon as possible before exploitation.


    Users who use AI assistants for their work efficiency should be cautious of prompts or requests that seem to be repeated extensively. Moreover, regularly update your application or the AI system to keep information protected. Thus, such precautions can help minimize the risks.

    Is there any Planning of a Large-Scale AI Worm Attack?

    Fortunately, there is no evidence of a planned AI worm attack. As per the researchers, Morris II is in a controlled environment with the goal of calling attention to the potential risks connected with AI vulnerabilities rather than carrying out a real attack.

    Nonetheless, the analysis provides a troubling picture of how rapidly cyber threats are growing. The potential battlefield for cybercriminals expands as AI becomes more common in our everyday lives.

    Bottom Line!

    While technology holds great potential for rooting out threats, it’s still important to stay up-to-date and address all security risks. By taking major steps with AI assistant development and deployment, we can ensure that AI won’t breed ground with any future threats.

    Read other AI-related news or blogs to stay up-to-date and never compromise your productivity. See you soon with the latest and hottest news on the board yet to be revealed!

    Related Posts

    Subscribe To Our Newsletter

    Subscribe To Our Newsletter

    Join our mailing list to receive the latest news and updates from our team.

    You have Successfully Subscribed!