Top
Best
New

Posted by takira 1/14/2026

Claude Cowork exfiltrates files(www.promptarmor.com)
870 points | 399 commentspage 3
LetsGetTechnicl 1/15/2026|
I know this isn't even the worst example, but the whole LLM craze has been insane to witness. Just releasing dangerous tools onto an uneducated and unprepared public and now we have to deal with the consequences because no one thought "should we do this?"
casey2 1/15/2026|
Pretty much all of the country takes years of formal education. They all understand file permissions. Most just pretend not to so their time isn't exploited.
refulgentis 1/14/2026||
These prompt injection techniques are increasingly implausible* to me yet theoretically sound.

Anyone know what can avoid this being posted when you build a tool like this? AFAIK there is no simonw blessed way to avoid it.

* I upload a random doc I got online, don’t read it, and it includes an API key in it for the attacker.

rswail 1/15/2026||
You read it, but you don't notice/see/detect the text in 1pt white-on-white background. The AI does see it.

That's what this attack did.

I'm sure that the anti-virus guys are working on how to detect these sort of "hidden from human view" instructions.

chasd00 1/15/2026||
the next attack will just be like malicious captions in a video. Or malicious lyrics in an mp3. it doesn't ever really end because it's not something that can be solved in the model.
NewsaHackO 1/15/2026||
At least for a malicious user embedding a prompt injection using their API key, I could have sworn that there is a way to scan documents that have a high level of entropy, which should be able to flag it.
teekert 1/15/2026||
Everything is a .exe if you're LLM enough.
fudged71 1/15/2026||
I found a bunch of potential vulnerabilities in the example Skills .py files provided by Anthropic. I don't believe the CVSS/Severity scores though:

| Skill | Title | CVSS | Severity |

| webapp-testing | Command Injection via `shell=True` | 9.8 | *Critical* |

| mcp-builder | Command Injection in Stdio Transport | 8.8 | *High* |

| slack-gif-creator | Path Traversal in Font Loading | 7.5 | *High* |

| xlsx | Excel Formula Injection | 6.1 | Medium |

| docx/pptx | ZIP Path Traversal | 5.3 | Medium |

| pdf | Lack of Input Validation | 3.7 | Low |

calflegal 1/14/2026||
So, I guess we're waiting on the big one, right? The ?10+? billion dollar attack?
chasd00 1/14/2026|
It will be either one big one or a pattern that can't be defended against and it just spreads through the whole industry. The only answer will be crippling the models by disconnecting them from the databases, APIs, file systems etc.
armcat 1/15/2026||
I know it might slow things down, but why not do this:

1. Categorize certain commands (like network/curl/db/sql) as `simulation_required` 2. Run a simulation of that command (without actual execution) 3. As part of the simulation run a red/blue team setup, where you have two Claude agents each either their red/blue persona and a set of skills 4. If step (3) does not pass, notify the user/initiator

ryanjshaw 1/15/2026||
The Confused Deputy [1] strikes again. Maybe this time around capabilities-based solutions will get attention.

[1] https://web.archive.org/web/20031205034929/http://www.cis.up...

sgammon 1/14/2026||
is it not a file exfiltrator, as a product
khalic 1/15/2026|
If you don’t read the skills you install in your agent, you really shouldn’t be using one.
More comments...