Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
37 posts
Community
Grants & Fundraising Opportunities
The SF Bay Area
Events (Community)
11 posts
Moral Uncertainty
Center for Human-Compatible AI (CHAI)
Regulation and AI Risk
Future of Humanity Institute (FHI)
Population Ethics
Utilitarianism
Summaries
Altruism
Disagreement
Future of Life Institute (FLI)
18
Event [Berkeley]: Alignment Collaborator Speed-Meeting
AlexMennen
1d
2
15
Looking for an alignment tutor
JanBrauner
3d
2
69
Announcing AI Alignment Awards: $100k research contests about goal misgeneralization & corrigibility
Akash
28d
20
30
A newcomer’s guide to the technical AI safety field
zeshen
1mo
1
74
AI Safety and Neighboring Communities: A Quick-Start Guide, as of Summer 2022
Sam Bowman
3mo
2
69
Announcing the Introduction to ML Safety course
Dan H
4mo
6
61
Seeking Interns/RAs for Mechanistic Interpretability Projects
Neel Nanda
4mo
0
62
Encultured AI Pre-planning, Part 1: Enabling New Benchmarks
Andrew_Critch
4mo
2
73
Introducing the ML Safety Scholars Program
Dan H
7mo
2
60
Jobs: Help scale up LM alignment research at NYU
Sam Bowman
7mo
1
95
Apply to the ML for Alignment Bootcamp (MLAB) in Berkeley [Jan 3 - Jan 22]
habryka
1y
4
21
*New* Canada AI Safety & Governance community
Wyatt Tessari L'Allié
3mo
0
105
Call for research on evaluating alignment (funding + advice available)
Beth Barnes
1y
11
61
Apply for research internships at ARC!
paulfchristiano
11mo
0
16
The Slippery Slope from DALLE-2 to Deepfake Anarchy
scasper
1mo
9
21
CHAI, Assistance Games, And Fully-Updated Deference [Scott Alexander]
berglund
2mo
1
190
2018 AI Alignment Literature Review and Charity Comparison
Larks
4y
26
130
2019 AI Alignment Literature Review and Charity Comparison
Larks
3y
18
68
Comparing Utilities
abramdemski
2y
31
58
AI Alignment Podcast: An Overview of Technical AI Alignment in 2018 and 2019 with Buck Shlegeris and Rohin Shah
Palus Astra
2y
27
60
[AN #69] Stuart Russell's new book on why we need to replace the standard model of AI
Rohin Shah
3y
12
43
Learning preferences by looking at the world
Rohin Shah
3y
10
19
Example population ethics: ordered discounted utility
Stuart_Armstrong
3y
16
16
RFC: Meta-ethical uncertainty in AGI alignment
Gordon Seidoh Worley
4y
6
12
Self-regulation of safety in AI research
Gordon Seidoh Worley
4y
6