I just read Dario Amodei’s essay "The Adolescence of Technology" from the cybersecurity + safety angle, so you don’t need to. *

Key message:

AI capability is compounding faster than institutions, norms, and controls, but we should avoid doomerism and "thinking about AI risks in a quasi-religious way."

Practical lens aligned to the risk buckets:

  1. Autonomy risk. There’s a "country of geniuses in a datacenter". The key question is not "what can AI do?" but "what is it optimizing for?" Evals can be misleading if behavior drifts under test conditions.
  2. Misuse for destruction. Even an "aligned" model can become a country of mercenaries. The biological misuse poses the highest risk.
  3. Misuse for seizing power. The scarier scenario is powerful states, and potentially corporate power use AI for surveillance, repression, propaganda, autonomous force, and strategic advantage.
  4. Economic disruption. Rapid job displacement + concentration of power can destabilize societies, which becomes a security problem.
  5. Indirect effects (unknown unknowns). When progress compresses decades into years, second-order risks show up fast: manipulation, new misuse pathways, brittle institutions.

Battle plan:

Guardrails alone won’t hold. Assume jailbreaks, add hardened layers, and pair them with transparency and disclosures.

My take is simple: there’s no pause on AI advancements. We need to learn quickly how to make these systems safe and secure.

The Adolescence of Technology

###

Dario Amodei's five AI risk buckets: autonomy, misuse, power seizure, economic disruption, unknowns