Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
42 posts
Value Learning
The Pointers Problem
Kolmogorov Complexity
14 posts
Metaethics
Meta-Philosophy
Philosophy
Perceptual Control Theory
99
The Pointers Problem: Human Values Are A Function Of Humans' Latent Variables
johnswentworth
2y
43
71
Clarifying "AI Alignment"
paulfchristiano
4y
82
66
Parsing Chris Mingard on Neural Networks
Alex Flint
1y
27
62
Preface to the sequence on value learning
Rohin Shah
4y
6
56
Conclusion to the sequence on value learning
Rohin Shah
3y
20
54
What is ambitious value learning?
Rohin Shah
4y
28
54
Humans can be assigned any values whatsoever…
Stuart_Armstrong
4y
26
54
Policy Alignment
abramdemski
4y
25
54
Normativity
abramdemski
2y
11
53
Intuitions about goal-directed behavior
Rohin Shah
4y
15
53
Don't design agents which exploit adversarial inputs
TurnTrout
1mo
61
52
The easy goal inference problem is still hard
paulfchristiano
4y
19
50
Different perspectives on concept extrapolation
Stuart_Armstrong
8mo
7
45
Future directions for ambitious value learning
Rohin Shah
4y
9
76
Some Thoughts on Metaphilosophy
Wei_Dai
3y
27
40
Recursive Quantilizers II
abramdemski
2y
15
30
Deconfusing Human Values Research Agenda v1
Gordon Seidoh Worley
2y
12
29
A theory of human values
Stuart_Armstrong
3y
13
28
AI Alignment, Philosophical Pluralism, and the Relevance of Non-Western Philosophy
xuan
1y
21
26
Impossible moral problems and moral authority
Charlie Steiner
3y
8
25
My take on agent foundations: formalizing metaphilosophical competence
zhukeepa
4y
6
22
Meta-preferences two ways: generator vs. patch
Charlie Steiner
2y
0
21
Deliberation as a method to find the "actual preferences" of humans
riceissa
3y
5
21
Can we make peace with moral indeterminacy?
Charlie Steiner
3y
8
19
Gricean communication and meta-preferences
Charlie Steiner
2y
0
18
RFC: Philosophical Conservatism in AI Alignment Research
Gordon Seidoh Worley
4y
13
15
What Should AI Owe To Us? Accountable and Aligned AI Systems via Contractualist AI Alignment
xuan
3mo
15
11
The Value Definition Problem
Sammy Martin
3y
6