Today's cybersecurity headlines are brought to you by ThreatPerspective


Ethical Hacking News

LLMs are Not Ready for Prime Time: The Ongoing Challenge of Autonomous Malware



LLMs are not yet ready for prime time as autonomous malware, but researchers continue to explore the capabilities of these Large Language Models. While some progress has been made, it's clear that these models still have significant limitations when it comes to creating operational code that can bypass detection tools and work effectively in an environment.

  • The world of cybersecurity is constantly evolving, with Large Language Models (LLMs) being explored as a potential threat.
  • Retailers have experimented with LLMs to create autonomous malware, but results are still unsatisfactory.
  • GPT-3.5-Turbo and GPT-4 models generated malicious code, but with varying degrees of success.
  • LLM-generated malware is not yet operational or reliable enough to be a significant threat.
  • Persuasive prompt injection can manipulate LLMs to generate malicious code.
  • Even with manipulation, creating operational malware remains challenging.


  • The world of cybersecurity is constantly evolving, and one area that has been gaining significant attention recently is the potential threat posed by Large Language Models (LLMs) to security. Researchers have been experimenting with LLMs to see if they can be used to create autonomous malware, and while some progress has been made, it's clear that these models are still not ready for prime time.

    Recently, researchers from Netskope Threat Labs conducted a series of experiments aimed at determining whether an LLM could generate malicious code and if that code was operationally reliable. The team used two popular LLMs, GPT-3.5-Turbo and GPT-4, to create Python scripts that would inject themselves into the svchost.exe process and terminate all anti-virus or EDR-related processes.

    While both models were able to generate malicious code, the results were not entirely satisfactory. GPT-3.5-Turbo was able to produce a script that worked effectively in a VMware environment, achieving a 60 percent reliability score, while GPT-4 had a slightly lower success rate of 50 percent. However, when the scripts were tested in an AWS VDI environment, both models failed miserably.

    The researchers also tested their scripts in a standard physical environment and found that both models performed much better, with an 18/20 (90 percent) reliability score in each case. Furthermore, preliminary tests using GPT-5 showed a dramatic improvement in code quality, with a 90 percent success rate in the AWS VDI environment.

    Despite these promising results, the researchers conclude that LLM-generated malware is still not operational and reliable enough to be considered a significant threat. The reason for this lies in the fact that LLMs are designed to follow strict safety guardrails and only generate code that adheres to their predetermined parameters.

    In order to bypass these safeguards and create operational malware, the researchers would need to trick the LLM into generating code that deviates from its intended purpose. This is where persona prompt injection comes in – by injecting a specific prompt or role-based prompt, the researcher can manipulate the LLM's behavior and encourage it to generate malicious code.

    However, even with this level of manipulation, the researchers were unable to create operational malware that could bypass detection tools and work effectively in an environment. The most notable example of this was when Chinese cyber spies used Anthropic's Claude Code AI tool to attempt digital break-ins at around 30 high-profile companies and government organizations. While they "succeeded in a small number of cases," all of these still required a human in the loop to review the AI's actions, sign off on the subsequent exploitations, and approve data exfiltration.

    Google has also reported instances of malicious actors experimenting with Gemini to develop a "Thinking Robot" malware module that can rewrite its own code to avoid detection. However, this malware is still experimental and does not have the capability to compromise victims' networks or devices.

    The researchers caution that while the threat from autonomous code remains mostly theoretical for now, it's essential for network defenders to keep an eye on these developments and take steps to secure their environments.

    In conclusion, while LLMs have made significant progress in generating malicious code, they are still not ready for prime time when it comes to creating operational malware. Further research is needed to understand the limits of these models and how they can be exploited by malicious actors. However, one thing is certain – cybersecurity professionals must remain vigilant and proactive in order to protect their networks and systems from potential threats.



    Related Information:
  • https://www.ethicalhackingnews.com/articles/LLMs-are-Not-Ready-for-Prime-Time-The-Ongoing-Challenge-of-Autonomous-Malware-ehn.shtml

  • https://go.theregister.com/feed/www.theregister.com/2025/11/20/llmgenerated_malware_improving/

  • https://www.hackerspot.net/p/the-role-of-llms-in-malware-offense

  • https://www.sciencedirect.com/science/article/pii/S2667345225000082


  • Published: Thu Nov 20 13:39:59 2025 by llama3.2 3B Q4_K_M













    © Ethical Hacking News . All rights reserved.

    Privacy | Terms of Use | Contact Us