Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
166 posts
AI Risk
Goodhart's Law
World Optimization
Threat Models
Instrumental Convergence
Corrigibility
Existential Risk
Coordination / Cooperation
Academic Papers
AI Safety Camp
Ethics & Morality
Treacherous Turn
689 posts
Newsletters
Logical Induction
Epistemology
SERI MATS
Logical Uncertainty
Intellectual Progress (Society-Level)
Practice & Philosophy of Science
AI Alignment Fieldbuilding
Distillation & Pedagogy
Bayes' Theorem
Postmortems & Retrospectives
Radical Probabilism
189
The next decades might be wild
Marius Hobbhahn
5d
21
40
AI Neorealism: a threat model & success criterion for existential safety
davidad
5d
0
82
Thoughts on AGI organizations and capabilities work
Rob Bensinger
13d
17
429
Counterarguments to the basic AI x-risk case
KatjaGrace
2mo
122
56
You can still fetch the coffee today if you're dead tomorrow
davidad
11d
15
69
Deconfusing Direct vs Amortised Optimization
beren
18d
6
108
AI will change the world, but won’t take it over by playing “3-dimensional chess”.
boazbarak
28d
86
986
AGI Ruin: A List of Lethalities
Eliezer Yudkowsky
6mo
653
517
It Looks Like You're Trying To Take Over The World
gwern
9mo
125
30
Refining the Sharp Left Turn threat model, part 2: applying alignment techniques
Vika
25d
4
292
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res
6mo
48
144
An Update on Academia vs. Industry (one year into my faculty job)
David Scott Krueger (formerly: capybaralet)
3mo
18
48
Some advice on independent research
Marius Hobbhahn
1mo
4
83
Niceness is unnatural
So8res
2mo
18
132
Logical induction for software engineers
Alex Flint
17d
2
42
Reflections on the PIBBSS Fellowship 2022
Nora_Ammann
9d
0
265
Lessons learned from talking to >100 academics about AI safety
Marius Hobbhahn
2mo
16
158
Most People Start With The Same Few Bad Ideas
johnswentworth
3mo
30
146
Quintin's alignment papers roundup - week 1
Quintin Pope
3mo
5
144
Your posts should be on arXiv
JanBrauner
3mo
39
167
Conjecture: Internal Infohazard Policy
Connor Leahy
4mo
6
76
SERI MATS Program - Winter 2022 Cohort
Ryan Kidd
2mo
12
72
QAPR 4: Inductive biases
Quintin Pope
2mo
2
40
Auditing games for high-level interpretability
Paul Colognese
1mo
1
69
Quintin's alignment papers roundup - week 2
Quintin Pope
3mo
2
221
Call For Distillers
johnswentworth
8mo
42
55
Methodological Therapy: An Agenda For Tackling Research Bottlenecks
adamShimi
2mo
6
73
How to do theoretical research, a personal perspective
Mark Xu
4mo
4