Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
48 posts
Community
Center for Human-Compatible AI (CHAI)
Moral Uncertainty
Regulation and AI Risk
Grants & Fundraising Opportunities
Future of Humanity Institute (FHI)
Population Ethics
Utilitarianism
The SF Bay Area
Future of Life Institute (FLI)
Disagreement
Events (Community)
11 posts
Agent Foundations
Machine Intelligence Research Institute (MIRI)
Cognitive Reduction
Dissolving the Question
69
Announcing AI Alignment Awards: $100k research contests about goal misgeneralization & corrigibility
Akash
28d
20
16
The Slippery Slope from DALLE-2 to Deepfake Anarchy
scasper
1mo
9
69
Announcing the Introduction to ML Safety course
Dan H
4mo
6
62
Encultured AI Pre-planning, Part 1: Enabling New Benchmarks
Andrew_Critch
4mo
2
74
AI Safety and Neighboring Communities: A Quick-Start Guide, as of Summer 2022
Sam Bowman
3mo
2
13
AI Safety Discussion Days
Linda Linsefors
2y
1
15
Looking for an alignment tutor
JanBrauner
3d
2
60
Jobs: Help scale up LM alignment research at NYU
Sam Bowman
7mo
1
30
A newcomer’s guide to the technical AI safety field
zeshen
1mo
1
58
AI Alignment Podcast: An Overview of Technical AI Alignment in 2018 and 2019 with Buck Shlegeris and Rohin Shah
Palus Astra
2y
27
105
Call for research on evaluating alignment (funding + advice available)
Beth Barnes
1y
11
43
Learning preferences by looking at the world
Rohin Shah
3y
10
61
Apply for research internships at ARC!
paulfchristiano
11mo
0
18
[Meta] Do you want AIS Webinars?
Linda Linsefors
2y
7
53
Clarifying the Agent-Like Structure Problem
johnswentworth
2mo
14
23
Bridging Expected Utility Maximization and Optimization
Whispermute
4mo
5
65
Prize and fast track to alignment research at ALTER
Vanessa Kosoy
3mo
4
247
Why Agent Foundations? An Overly Abstract Explanation
johnswentworth
9mo
54
21
My current take on the Paul-MIRI disagreement on alignability of messy AI
jessicata
5y
0
27
On motivations for MIRI's highly reliable agent design research
jessicata
5y
1
198
The Rocket Alignment Problem
Eliezer Yudkowsky
4y
42
72
Challenges with Breaking into MIRI-Style Research
Chris_Leong
11mo
15
41
Grokking the Intentional Stance
jbkjr
1y
20
110
What I’ll be doing at MIRI
evhub
3y
6
59
Another take on agent foundations: formalizing zero-shot reasoning
zhukeepa
4y
20