Anthropic. Dead privacy. AI attacks and malicious skills. Five stories from this week.
- LLMs can deanonymize at scale for $1
Anthropic and ETH Zurich built a pipeline matching Hacker News accounts to real identities at 90% precision. Cross-platform references, embeddings, and LLM reasoning. Revise your deanonymization strategy.
- Anthropic exposed industrial-scale model theft by Chinese AI labs
DeepSeek, Moonshot AI, and MiniMax ran 16M+ queries across 24,000 fraudulent accounts to distill Claude. The ROI on distillation is too good. If you ship a powerful model, build misuse safeguards.
- Anthropic rejected the Pentagon's $200M ultimatum and got labeled a supply chain risk
Anthropic said they "cannot in good conscience" remove safety guardrails. Hegseth labeled them a supply chain risk. Trump doubled down and banned them from serving the government. If you're an AI vendor, there's an at least $200M seat at the table.
- CrowdStrike: 89% increase in AI-enabled attacks
AI-accelerated phishing, automated recon, prompt injection in phishing emails, and the first malicious MCP servers in the wild. Test your email security system for resilience to prompt injections.
- ~80% of malicious skill file instructions execute on frontier models
Snyk and Max Planck Institute tested 202 attack scenarios. Gemini 3 Flash: 83% ASR. Opus 4.5: 7.2%. Honestly, do you know how many OpenClaws are already installed by AI enthusiasts in your corporate environment?
Detecting and preventing distillation attacks
Statement on comments by the Secretary of War
CrowdStrike 2026 Global Threat Report