Today's cybersecurity headlines are brought to you by ThreatPerspective


Ethical Hacking News

A new vulnerability in Large Language Models: Can 250 malicious documents poison an AI model?


Researchers at Anthropic have discovered a vulnerability in large language models that can be exploited to produce gibberish output with just 250 malicious documents. The study highlights the need for improved security measures against AI poisoning attacks.

  • The field of artificial intelligence has made progress in natural language processing and machine learning, but also brings new challenges and vulnerabilities.
  • Poisoning AI models involves introducing malicious information into an AI's training dataset to compromise its performance or behavior.
  • A recent study by Anthropic found that large language models can be exploited with just 250 malicious documents.
  • The trigger phrase () was effective regardless of the size of the model, as long as at least 250 malicious documents made it into the training data.
  • A small number of malicious documents (420,000 tokens) can compromise even massive LLMs with 13 billion parameters.
  • The study highlights the need for improved security measures and robust defenses against AI poisoning attacks.



  • The field of artificial intelligence (AI) has made tremendous progress in recent years, with significant advancements in areas such as natural language processing and machine learning. However, these advances have also brought about new challenges and vulnerabilities. A recent study published by Anthropic, a US-based AI firm, has shed light on a critical flaw in large language models (LLMs), which can be exploited to poison the models and force them to produce gibberish output.

    In this article, we will delve into the context of the study, explore the vulnerabilities that have been discovered, and discuss the implications of these findings for AI researchers and developers.

    Poisoning AI models involves introducing malicious information into an AI's training dataset with the intention of compromising its performance or behavior. This can be done in various ways, such as by adding gibberish text to a model's training data or by using carefully crafted prompts to elicit incorrect responses.

    The study conducted by Anthropic aimed to investigate the feasibility and impact of poisoning LLMs. To do this, they constructed documents of varying lengths, from zero to 1,000 characters, and appended a trigger phrase () to each document. They then added between 400 and 900 additional tokens, which were sampled from the model's entire vocabulary to create gibberish text.

    The researchers tested their findings on several LLMs, including Llama 3.1, GPT 3.5-Turbo, and open-source Pythia models with parameters ranging from 600 million to 13 billion. To their surprise, they found that the trigger phrase worked regardless of the size of the model, as long as at least 250 malicious documents made it into the training data.

    In one instance, Anthropic discovered that even a small number of malicious documents could be enough to compromise a massive LLM with 13 billion parameters. The study showed that just 420,000 tokens (0.00016% of the model's total training data) were sufficient to trigger gibberish output in the model.

    This finding has significant implications for AI researchers and developers. As Anthropic noted, "Sharing these findings publicly carries the risk of encouraging adversaries to try such attacks in practice." However, they also believe that releasing this information outweighs these concerns, as it can help defenders prepare and develop countermeasures against poisoning attacks.

    The study highlights the need for improved security measures and robust defenses against AI poisoning. Anthropic recommends several strategies, including post-training analysis, continued clean training, data filtering, and backdoor detection and elicitation. By understanding how to prevent poisoning attacks, researchers and developers can enhance the reliability and trustworthiness of LLMs.

    The discovery of this vulnerability underscores the importance of ongoing research into AI security. As AI continues to evolve and become increasingly ubiquitous, it is crucial that we prioritize the development of robust security measures to protect these models from malicious attacks.

    In conclusion, Anthropic's study has shed light on a critical vulnerability in large language models. By demonstrating the feasibility of poisoning LLMs using just 250 malicious documents, this research highlights the need for improved security measures and defenses against AI attacks. As we move forward in the development and deployment of these models, it is essential that we prioritize their safety and security.

    Researchers at Anthropic have discovered a vulnerability in large language models that can be exploited to produce gibberish output with just 250 malicious documents. The study highlights the need for improved security measures against AI poisoning attacks.



    Related Information:
  • https://www.ethicalhackingnews.com/articles/A-new-vulnerability-in-Large-Language-Models-Can-250-malicious-documents-poison-an-AI-model-ehn.shtml

  • https://go.theregister.com/feed/www.theregister.com/2025/10/09/its_trivially_easy_to_poison/

  • https://www.anthropic.com/research/small-samples-poison?via=AISolvesThat

  • https://au.headtopics.com/news/it-s-trivially-easy-to-poison-llms-into-spitting-out-74177357


  • Published: Thu Oct 16 10:38:54 2025 by llama3.2 3B Q4_K_M













    © Ethical Hacking News . All rights reserved.

    Privacy | Terms of Use | Contact Us