Go Back
You can't go any further
You can't go any further
meritocratic
regular
democratic
hot
top
alive
4 posts
The Pointers Problem
59 posts
Value Learning
115
The Pointers Problem: Human Values Are A Function Of Humans' Latent Variables
johnswentworth
2y
43
19
Stable Pointers to Value III: Recursive Quantilization
abramdemski
4y
4
13
Stable Pointers to Value II: Environmental Goals
abramdemski
4y
2
10
Stable Pointers to Value: An Agent Embedded in Its Own Utility Function
abramdemski
5y
9
80
The E-Coli Test for AI Alignment
johnswentworth
4y
24
77
Preface to the sequence on value learning
Rohin Shah
4y
6
66
Why we need a *theory* of human values
Stuart_Armstrong
4y
15
61
Humans can be assigned any values whatsoever…
Stuart_Armstrong
4y
26
60
Clarifying "AI Alignment"
paulfchristiano
4y
82
55
The Urgent Meta-Ethics of Friendly Artificial Intelligence
lukeprog
11y
252
53
Intuitions about goal-directed behavior
Rohin Shah
4y
15
52
AI Alignment Problem: “Human Values” don’t Actually Exist
avturchin
3y
29
50
Future directions for ambitious value learning
Rohin Shah
4y
9
50
The easy goal inference problem is still hard
paulfchristiano
4y
19
47
Using vector fields to visualise preferences and make them consistent
MichaelA
2y
32
46
What is ambitious value learning?
Rohin Shah
4y
28
44
Conclusion to the sequence on value learning
Rohin Shah
3y
20
42
Where do selfish values come from?
Wei_Dai
11y
62