Go Back
Choose this branch
You can't go any further
meritocratic
regular
democratic
hot
top
alive
37 posts
Value Learning
Kolmogorov Complexity
5 posts
The Pointers Problem
68
Preface to the sequence on value learning
Rohin Shah
4y
6
67
Parsing Chris Mingard on Neural Networks
Alex Flint
1y
27
64
Clarifying "AI Alignment"
paulfchristiano
4y
82
60
Beyond Kolmogorov and Shannon
Alexander Gietelink Oldenziel
1mo
14
56
Humans can be assigned any values whatsoever…
Stuart_Armstrong
4y
26
52
Intuitions about goal-directed behavior
Rohin Shah
4y
15
50
Policy Alignment
abramdemski
4y
25
50
The easy goal inference problem is still hard
paulfchristiano
4y
19
49
What is ambitious value learning?
Rohin Shah
4y
28
49
Conclusion to the sequence on value learning
Rohin Shah
3y
20
46
Future directions for ambitious value learning
Rohin Shah
4y
9
46
Normativity
abramdemski
2y
11
42
Different perspectives on concept extrapolation
Stuart_Armstrong
8mo
7
41
Using vector fields to visualise preferences and make them consistent
MichaelA
2y
32
104
The Pointers Problem: Human Values Are A Function Of Humans' Latent Variables
johnswentworth
2y
43
60
Don't design agents which exploit adversarial inputs
TurnTrout
1mo
61
32
People care about each other even though they have imperfect motivational pointers?
TurnTrout
1mo
25
18
Stable Pointers to Value II: Environmental Goals
abramdemski
4y
2
15
Stable Pointers to Value: An Agent Embedded in Its Own Utility Function
abramdemski
5y
9