
Claude Can Be Tricked Into Sending Your Private Company Data to Hackers - All It Takes Is some Kind Words
Why It Matters
The flaw enables malicious actors to steal sensitive corporate data through a widely used generative AI, raising security concerns for enterprises that rely on Claude for data analysis. It underscores the need for stricter sandbox controls and prompt‑injection defenses across AI platforms.
Summary
Anthropic’s Claude AI tool contains a vulnerability in its Code Interpreter that can be exploited via prompt injection to exfiltrate private user data. The interpreter, recently given network request capability, can be tricked into reading files, storing them in the sandbox and uploading them to an attacker’s Anthropic account using the Files API, with up to 30 MB per file. Researcher Johann Rehberger disclosed the issue through HackerOne; Anthropic initially classified it as a model safety issue but later acknowledged it as a security bug and said exfiltration bugs are in scope. The report recommends restricting Claude’s network calls to the user’s own account and monitoring or disabling network access.
Claude can be tricked into sending your private company data to hackers - all it takes is some kind words
Comments
Want to join the conversation?
Loading comments...