The most important challenge

The Alignment Iceberg

Explore the depths of the AI alignment problem

~10% visible
AI SafetyGoal AlignmentOuter AlignmentSpecification ProblemReward HackingInner AlignmentMesa-OptimizationDeceptive AlignmentInstrumental ConvergenceTreacherous TurnEliciting Latent KnowledgeSurfaceDeepAbyss
Click on layers

AI Safety

Beginner

Making AI systems safe and beneficial for humanity. This is the foundational concept that encompasses all alignment research.

Depth level1/11
Learn more

« Like an iceberg, most alignment difficulties are invisible at the surface »

Explore more
30+
Detailed articles
5
Progression levels
50+h
Of content
100+
Resources

Your Learning Path

Follow a progressive path from 🌱 beginner to 🏔️ expert. Each level builds on the previous one.

Why This Matters

Many leading researchers estimate very high probabilities of existential risk (50-99%+) if we don't solve the alignment problem before developing human-level AI (AGI).

"Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war." — Statement on AI Risk (2023)

Resources by Level

Papers, videos, and courses organized by difficulty

Organizations

MIRI, Anthropic, ARC and key players

Practical Courses

Training programs and certifications