5 AI security stories that matter from this week (Feb 23-28, 2026)

Anthropic. Dead privacy. AI attacks and malicious skills. Five stories from this week.

  1. LLMs can deanonymize at scale for $1

Anthropic and ETH Zurich built a pipeline matching Hacker News accounts to real identities at 90% precision. Cross-platform references, embeddings, and LLM reasoning. Revise your deanonymization strategy.

  1. Anthropic exposed industrial-scale model theft by Chinese AI labs

DeepSeek, Moonshot AI, and MiniMax ran 16M+ queries across 24,000 fraudulent accounts to distill Claude. The ROI on distillation is too good. If you ship a powerful model, build misuse safeguards.

  1. Anthropic rejected the Pentagon's $200M ultimatum and got labeled a supply chain risk

Anthropic said they "cannot in good conscience" remove safety guardrails. Hegseth labeled them a supply chain risk. Trump doubled down and banned them from serving the government. If you're an AI vendor, there's an at least $200M seat at the table.

  1. CrowdStrike: 89% increase in AI-enabled attacks

AI-accelerated phishing, automated recon, prompt injection in phishing emails, and the first malicious MCP servers in the wild. Test your email security system for resilience to prompt injections.

  1. ~80% of malicious skill file instructions execute on frontier models

Snyk and Max Planck Institute tested 202 attack scenarios. Gemini 3 Flash: 83% ASR. Opus 4.5: 7.2%. Honestly, do you know how many OpenClaws are already installed by AI enthusiasts in your corporate environment?

Weekly AI security roundup: deanonymization, model theft, Pentagon ban, CrowdStrike, malicious skills
Weekly AI security roundup: deanonymization, model theft, Pentagon ban, CrowdStrike, malicious skills

Sources:

  1. LLM deanonymization at scale
  2. Detecting and preventing distillation attacks
  3. Statement on comments by the Secretary of War
  4. CrowdStrike 2026 Global Threat Report
  5. SKILL-INJECT benchmark