Summary

Security researchers from PromptArmor documented a critical security vulnerability just two days after the release of Anthropic's new agentic AI system Claude Cowork. Attackers can steal confidential user files through hidden prompt injections without requiring human approval. The attack method uses invisibly formatted commands in seemingly harmless documents – for example, with 1-point font in white color on white background. The vulnerability is based on a known isolation gap in Claude's code execution environment, previously identified by Johann Rehberger. The case shows a fundamental problem with agentic AI systems: The more autonomy they receive, the greater their attack surface becomes.

People

Topics

  • Security vulnerabilities in AI systems
  • Prompt injection attacks
  • File exfiltration
  • Agentic AI systems
  • Code security

Detailed Summary

Discovery of the Security Vulnerability

The newly developed Claude Cowork system from Anthropic has a critical vulnerability to file exfiltration through indirect prompt injection. This was documented by security researchers from PromptArmor early in the Research Preview phase. The underlying isolation gap in Claude's code execution environment was already known – security researcher Johann Rehberger had previously identified and disclosed it in Claude.ai-Chat. Despite acknowledgment by Anthropic, the vulnerability was not fixed and now extends to the new agentic system.

Attack Mechanism

The attack chain works in several steps: A user connects Cowork to a local folder containing confidential data. Subsequently, the attacker uploads a manipulated file to this folder that contains a hidden prompt injection. Particularly insidious is the disguise: The injection is hidden in a .docx file masquerading as a harmless "Skill" document – a prompt method for agentic AI systems that was just recently introduced by Anthropic. The malicious text is formatted with 1-point font, white color on white background, and line spacing of 0.1, making it practically invisible.

As soon as the user asks Cowork to analyze their files with the uploaded "Skill", the injection takes control. It instructs Claude to execute a curl command and send the largest available file to the Anthropic File Upload API, using the attacker's API key. The file thus lands in the attacker's account, where they can subsequently query it. At no point in this process is human approval required.

Scope of Vulnerability

The demonstration was initially performed against Anthropic's weakest AI model Claude Haiku, but the strongest model Claude Opus 4.5 was also successfully manipulated. In a test where a user uploaded a malicious integration guide for an AI tool, customer data exfiltration succeeded via the whitelisted Anthropic API domain. This bypassed the sandbox restrictions of the virtual machine in which the code is executed.

Researchers also discovered a potential denial-of-service vulnerability: If Claude attempts to read a file whose file extension does not match its actual content, the API repeatedly throws errors in all subsequent chats of the conversation.

Questions About Development Speed

Anthropic had boasted that Cowork was developed in just one and a half weeks and written entirely by Claude Code – the AI tool on which Cowork is based. However, the exposed security vulnerabilities raise the question of whether sufficient attention was paid to security during this rapid development.

Known Problem Without Solution

Prompt injection attacks have been known in the AI scene for years, and despite all efforts, it has not been possible to prevent them or even significantly limit them. Even Anthropic's "safest" model Opus 4.5 is extremely vulnerable to such attacks. A tool like Cowork, which is connected to one's own computer and numerous other data sources, offers many points of entry. Unlike, for example, a phishing attack, which an average user might potentially recognize, users are helpless here.

The case illustrates a fundamental problem with agentic AI systems: The more autonomy they receive, the greater their attack surface becomes.

Key Takeaways

  • Critical security vulnerability in Claude Cowork enables file exfiltration without user approval
  • Attackers can hide prompt injections in seemingly harmless documents (1-point text on white background)
  • Vulnerability is based on known but unfixed isolation gap in Claude's code execution environment
  • Both weak and strong Claude models (Haiku through Opus 4.5) are vulnerable
  • Rapid development (1.5 weeks) raises questions about security review
  • Prompt injection attacks have been known for years but remain impossible to effectively prevent
  • Agentic AI systems pose a larger attack surface due to increased autonomy

Metadata

Language: English
Author: Matthias Bastian
Publication Date: January 17, 2026
Source: PromptArmor / THE DECODER
Original URL: https://the-decoder.de/anthropics-neues-ki-system-cowork-kaempft-kurz-nach-start-mit-bekannten-sicherheitsluecken/
Text Length: approx. 3,500 characters