Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
3083 posts
AI
GPT
AI Timelines
Machine Learning (ML)
AI Takeoff
Interpretability (ML & AI)
Language Models
Conjecture (org)
Careers
Instrumental Convergence
Iterated Amplification
Art
763 posts
Anthropics
Existential Risk
Whole Brain Emulation
Sleeping Beauty Paradox
Threat Models
Academic Papers
Space Exploration & Colonization
Great Filter
Paradoxes
Extraterrestrial Life
Pascal's Mugging
Longtermism
472
Simulators
janus
3mo
103
364
chinchilla's wild implications
nostalgebraist
4mo
114
364
DeepMind alignment team opinions on AGI ruin arguments
Vika
4mo
34
351
What DALL-E 2 can and cannot do
Swimmer963
7mo
305
344
(My understanding of) What Everyone in Technical Alignment is Doing and Why
Thomas Larsen
3mo
83
338
A Mechanistic Interpretability Analysis of Grokking
Neel Nanda
4mo
39
314
How To Get Into Independent Research On Alignment/Agency
johnswentworth
1y
33
310
Without specific countermeasures, the easiest path to transformative AI likely leads to AI takeover
Ajeya Cotra
5mo
89
303
What should you change in response to an "emergency"? And AI risk
AnnaSalamon
5mo
60
287
Two-year update on my personal AI timelines
Ajeya Cotra
4mo
60
269
Why I think strong general AI is coming soon
porby
2mo
126
265
A challenge for AGI organizations, and a challenge for readers
Rob Bensinger
19d
30
259
We Choose To Align AI
johnswentworth
11mo
15
255
Are we in an AI overhang?
Andy Jones
2y
109
777
Where I agree and disagree with Eliezer
paulfchristiano
6mo
205
724
AGI Ruin: A List of Lethalities
Eliezer Yudkowsky
6mo
653
336
Counterarguments to the basic AI x-risk case
KatjaGrace
2mo
122
325
Discussion with Eliezer Yudkowsky on AGI interventions
Rob Bensinger
1y
257
319
What failure looks like
paulfchristiano
3y
49
253
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res
6mo
48
224
A Quick Guide to Confronting Doom
Ruby
8mo
36
210
Another (outer) alignment failure story
paulfchristiano
1y
38
209
Slow motion videos as AI risk intuition pumps
Andrew_Critch
6mo
36
203
What Multipolar Failure Looks Like, and Robust Agent-Agnostic Processes (RAAPs)
Andrew_Critch
1y
60
199
Some AI research areas and their relevance to existential safety
Andrew_Critch
2y
40
190
On saving one's world
Rob Bensinger
7mo
5
185
Whole Brain Emulation: No Progress on C. elgans After 10 Years
niconiconi
1y
77
168
AI Could Defeat All Of Us Combined
HoldenKarnofsky
6mo
29