Go Back
You can't go any further
You can't go any further
meritocratic
regular
democratic
hot
top
alive
0 posts
Goal Factoring
20 posts
Research Agendas
35
My AGI safety research—2022 review, ’23 plans
Steven Byrnes
6d
6
231
On how various plans miss the hard bits of the alignment challenge
So8res
5mo
81
155
Some conceptual alignment research projects
Richard_Ngo
3mo
14
16
Theories of impact for Science of Deep Learning
Marius Hobbhahn
19d
0
10
Distilled Representations Research Agenda
Hoagy
2mo
2
69
Research agenda update
Steven Byrnes
1y
40
122
Our take on CHAI’s research agenda in under 1500 words
Alex Flint
2y
19
34
New year, new research agenda post
Charlie Steiner
11mo
4
45
Immobile AI makes a move: anti-wireheading, ontology change, and model splintering
Stuart_Armstrong
1y
3
127
Thoughts on Human Models
Ramana Kumar
3y
32
82
Research Agenda v0.9: Synthesising a human's preferences into a utility function
Stuart_Armstrong
3y
25
19
AI, learn to be conservative, then learn to be less so: reducing side-effects, learning preserved features, and going beyond conservatism
Stuart_Armstrong
1y
4
77
The Learning-Theoretic AI Alignment Research Agenda
Vanessa Kosoy
4y
39
38
Resources for AI Alignment Cartography
Gyrodiot
2y
8