Reading Lists par Niveau
Ressources organisées par niveau de difficulté
Beginner
Reading Lists par Niveau
Débutant (0-20h)
Objectif: Comprendre qu'il y a un problème
1. AGI Ruin: A List of Lethalities (Eliezer)
2. AI Alignment: Why It's Hard, and Where to Start (Video - Eliezer)
3. Rob Miles YouTube Channel (vulgarisation)
- https://www.youtube.com/@RobertMilesAI
- 5-10 videos (5h)
4. The Alignment Problem (Brian Christian - livre)
- Accessible, journalistique
- 10-15h
Intermédiaire (20-100h)
Objectif: Comprendre problèmes techniques principaux
1. Risks from Learned Optimization (Hubinger et al.)
- https://arxiv.org/abs/1906.01820
- 10h (paper + discussions)
2. Concrete Problems in AI Safety (Amodei et al.)
3. Embedded Agency (Sequence)
4. Superintelligence (Nick Bostrom - livre)
- Classique, un peu daté (2014) mais fondamental
- 20h
5. ELK Document (ARC)
6. Alignment Forum (curated posts)
- https://www.alignmentforum.org/
- 20-30h (sélection)
Avancé (100-500h)
Objectif: Comprendre research frontiers, contribuer
1. MIRI Research
- https://intelligence.org/research/
- Agent foundations papers
- 50-100h
2. Corrigibility (MIRI)
- https://intelligence.org/files/Corrigibility.pdf
- 20h (+ related work)
3. Logical Induction (MIRI)
- https://intelligence.org/files/LogicalInduction.pdf
- 40h (très math-heavy)
4. Iterated Amplification (Paul Christiano, all posts)
5. Constitutional AI + Mechanistic Interpretability (Anthropic research)
6. Debate, ELK, related (ARC, OpenAI)
- Papers + discussions
- 50h
7. Alignment Forum (comprehensive reading)
- Major sequences, debates
- 100-200h