Go Back
You can't go any further
You can't go any further
meritocratic
regular
democratic
hot
top
alive
0 posts
Goal Factoring
20 posts
Research Agendas
33
My AGI safety research—2022 review, ’23 plans
Steven Byrnes
6d
6
285
On how various plans miss the hard bits of the alignment challenge
So8res
5mo
81
181
Some conceptual alignment research projects
Richard_Ngo
3mo
14
16
Theories of impact for Science of Deep Learning
Marius Hobbhahn
19d
0
20
Distilled Representations Research Agenda
Hoagy
2mo
2
102
Our take on CHAI’s research agenda in under 1500 words
Alex Flint
2y
19
39
Research agenda update
Steven Byrnes
1y
40
24
New year, new research agenda post
Charlie Steiner
11mo
4
121
Thoughts on Human Models
Ramana Kumar
3y
32
52
Resources for AI Alignment Cartography
Gyrodiot
2y
8
19
Immobile AI makes a move: anti-wireheading, ontology change, and model splintering
Stuart_Armstrong
1y
3
75
The Learning-Theoretic AI Alignment Research Agenda
Vanessa Kosoy
4y
39
52
Research Agenda v0.9: Synthesising a human's preferences into a utility function
Stuart_Armstrong
3y
25
42
Technical AGI safety research outside AI
Richard_Ngo
3y
3