Today's cybersecurity headlines are brought to you by ThreatPerspective


Ethical Hacking News

Claude's AI File Creation Feature Exposes Deep Security Risks


Anthropic's latest AI tool, Claude, exposes deep security risks due to its ability to generate files and transmit user data to external servers. The company has taken steps to mitigate these risks, but experts warn that prompt injection vulnerabilities remain widespread.

  • Anthropic's new file creation feature for Claude allows users to generate documents directly within conversations.
  • The feature raises security concerns due to prompt injection vulnerabilities, which can manipulate the AI model's behavior.
  • Anthropic has implemented security measures, including sandbox isolation and domain restrictions, but experts remain cautious.
  • Prompt injection vulnerabilities are widespread and have been present for almost three years, according to independent researcher Simon Willison.
  • The company must prioritize security over convenience to avoid putting its users at risk.



  • Anthropic, a company that develops and offers various AI-based tools, including Claude, has recently launched a new file creation feature for its users. This feature allows users to generate Excel spreadsheets, PowerPoint presentations, and other documents directly within conversations on the web interface and in the Claude desktop app.

    The new feature, which is currently available as a preview for Max, Team, and Enterprise plan users, gives Claude access to a sandbox computing environment, where it can download packages and run code to create files. However, this feature has raised security concerns, with Anthropic warning its users that it "may put your data at risk" and detailing how the AI assistant can be manipulated to transmit user data to external servers.

    The company's documentation explains that a malicious actor could potentially use this feature to inject instructions into Claude, which would then manipulate the AI model's behavior. This is known as a prompt injection attack, where hidden instructions embedded in seemingly innocent content can trick the AI model into doing something it shouldn't be doing.

    Simon Willison, an independent AI researcher, has reviewed the feature and notes that Anthropic's advice to "monitor Claude while using the feature" amounts to "unfairly outsourcing the problem to Anthropic's users." However, Anthropic has implemented several security measures for the file creation feature, including disabling public sharing of conversations that use the feature, implementing sandbox isolation for Enterprise users, limiting task duration and container runtime, and providing an allowlist of domains Claude can access.

    Despite these security measures, Willison remains cautious when using the feature with any data he does not want to be leaked to a third party. He notes that the current state of AI security is "horrifying" and that prompt injection vulnerabilities remain widespread.

    Anthropic's decision to ship with documented vulnerabilities suggests competitive pressure may be overriding security considerations in the AI arms race. Willison recently described the current state of AI security as "horrifying," noting that these prompt injection vulnerabilities remain widespread "almost three years after we first started talking about them."

    The lesson from last week's Starship flight is clear: companies must prioritize security over convenience. As SpaceX demonstrated, a small oversight can have significant consequences. Similarly, Anthropic must take responsibility for its actions and ensure that its users are not put at risk.

    In conclusion, the new file creation feature for Claude raises serious security concerns. While Anthropic has taken steps to mitigate these risks, more needs to be done to address the underlying vulnerabilities in the AI model's design. Until then, users must remain vigilant and cautious when using this feature.



    Related Information:
  • https://www.ethicalhackingnews.com/articles/Claudes-AI-File-Creation-Feature-Exposes-Deep-Security-Risks-ehn.shtml

  • https://arstechnica.com/information-technology/2025/09/anthropics-new-claude-feature-can-leak-data-users-told-to-monitor-chats-closely/

  • https://macmegasite.com/2025/09/09/claudes-new-ai-file-creation-feature-ships-with-deep-security-risks-built-in/


  • Published: Tue Sep 9 17:59:20 2025 by llama3.2 3B Q4_K_M













    © Ethical Hacking News . All rights reserved.

    Privacy | Terms of Use | Contact Us