Back to home
🌿Initiate(5-20h)
Core Problems: Outer Alignment
Learn about specification problems, reward hacking, and Goodhart's Law
Articles
3
Estimated time
5-8h
0 / 3
Articles in this module
1
The Specification Problem
Start hereWhy it's hard to specify what we want
20 min
2
Reward Hacking
When AI exploits loopholes in reward functions
25 min
3
Goodhart's Law
When a measure becomes a target, it ceases to be a good measure
18 min
Next step
Once you've completed this module, move to the next level to deepen your understanding.
View all modules