Go Back
You can't go any further
You can't go any further
meritocratic
regular
democratic
hot
top
alive
0 posts
Goal Factoring
20 posts
Research Agendas
258
On how various plans miss the hard bits of the alignment challenge
So8res
5mo
81
168
Some conceptual alignment research projects
Richard_Ngo
3mo
14
124
Thoughts on Human Models
Ramana Kumar
3y
32
112
Our take on CHAI’s research agenda in under 1500 words
Alex Flint
2y
19
76
The Learning-Theoretic AI Alignment Research Agenda
Vanessa Kosoy
4y
39
67
Research Agenda v0.9: Synthesising a human's preferences into a utility function
Stuart_Armstrong
3y
25
54
Research agenda update
Steven Byrnes
1y
40
45
Resources for AI Alignment Cartography
Gyrodiot
2y
8
43
Technical AGI safety research outside AI
Richard_Ngo
3y
3
34
New safety research agenda: scalable agent alignment via reward modeling
Vika
4y
13
34
Research Agenda in reverse: what *would* a solution look like?
Stuart_Armstrong
3y
25
34
Ultra-simplified research agenda
Stuart_Armstrong
3y
4
34
My AGI safety research—2022 review, ’23 plans
Steven Byrnes
6d
6
32
Immobile AI makes a move: anti-wireheading, ontology change, and model splintering
Stuart_Armstrong
1y
3