Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
67 posts
Value Learning
Inverse Reinforcement Learning
The Pointers Problem
Meta-Philosophy
Metaethics
Kolmogorov Complexity
Philosophy
Book Reviews
Perceptual Control Theory
59 posts
Community
Agent Foundations
Machine Intelligence Research Institute (MIRI)
Cognitive Reduction
Center for Human-Compatible AI (CHAI)
Regulation and AI Risk
Grants & Fundraising Opportunities
Future of Humanity Institute (FHI)
Population Ethics
Utilitarianism
Moral Uncertainty
The SF Bay Area
99
The Pointers Problem: Human Values Are A Function Of Humans' Latent Variables
johnswentworth
2y
43
76
Some Thoughts on Metaphilosophy
Wei_Dai
3y
27
71
Clarifying "AI Alignment"
paulfchristiano
4y
82
69
[Book Review] "The Alignment Problem" by Brian Christian
lsusr
1y
16
66
Parsing Chris Mingard on Neural Networks
Alex Flint
1y
27
62
Preface to the sequence on value learning
Rohin Shah
4y
6
59
Thoughts on "Human-Compatible"
TurnTrout
3y
35
56
Conclusion to the sequence on value learning
Rohin Shah
3y
20
54
What is ambitious value learning?
Rohin Shah
4y
28
54
Humans can be assigned any values whatsoever…
Stuart_Armstrong
4y
26
54
Policy Alignment
abramdemski
4y
25
54
Normativity
abramdemski
2y
11
53
Intuitions about goal-directed behavior
Rohin Shah
4y
15
53
Human-AI Collaboration
Rohin Shah
3y
7
197
Why Agent Foundations? An Overly Abstract Explanation
johnswentworth
9mo
54
177
2018 AI Alignment Literature Review and Charity Comparison
Larks
4y
26
146
The Rocket Alignment Problem
Eliezer Yudkowsky
4y
42
139
Full-time AGI Safety!
Steven Byrnes
1y
3
133
2019 AI Alignment Literature Review and Charity Comparison
Larks
3y
18
120
What I’ll be doing at MIRI
evhub
3y
6
116
Call for research on evaluating alignment (funding + advice available)
Beth Barnes
1y
11
105
Apply to the ML for Alignment Bootcamp (MLAB) in Berkeley [Jan 3 - Jan 22]
habryka
1y
4
82
Comparing Utilities
abramdemski
2y
31
71
AI Alignment Podcast: An Overview of Technical AI Alignment in 2018 and 2019 with Buck Shlegeris and Rohin Shah
Palus Astra
2y
27
70
Encultured AI Pre-planning, Part 1: Enabling New Benchmarks
Andrew_Critch
4mo
2
70
[AN #69] Stuart Russell's new book on why we need to replace the standard model of AI
Rohin Shah
3y
12
68
AGI Safety Fundamentals curriculum and application
Richard_Ngo
1y
0
64
Apply for research internships at ARC!
paulfchristiano
11mo
0