Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
3083 posts
AI
GPT
AI Timelines
Machine Learning (ML)
AI Takeoff
Interpretability (ML & AI)
Language Models
Conjecture (org)
Careers
Instrumental Convergence
Iterated Amplification
Art
763 posts
Anthropics
Existential Risk
Whole Brain Emulation
Sleeping Beauty Paradox
Threat Models
Academic Papers
Space Exploration & Colonization
Great Filter
Paradoxes
Extraterrestrial Life
Pascal's Mugging
Longtermism
808
Simulators
janus
3mo
103
531
(My understanding of) What Everyone in Technical Alignment is Doing and Why
Thomas Larsen
3mo
83
521
chinchilla's wild implications
nostalgebraist
4mo
114
446
A Mechanistic Interpretability Analysis of Grokking
Neel Nanda
4mo
39
436
How To Get Into Independent Research On Alignment/Agency
johnswentworth
1y
33
432
DeepMind alignment team opinions on AGI ruin arguments
Vika
4mo
34
415
What DALL-E 2 can and cannot do
Swimmer963
7mo
305
404
Without specific countermeasures, the easiest path to transformative AI likely leads to AI takeover
Ajeya Cotra
5mo
89
394
Why I think strong general AI is coming soon
porby
2mo
126
373
We Choose To Align AI
johnswentworth
11mo
15
332
Two-year update on my personal AI timelines
Ajeya Cotra
4mo
60
331
What should you change in response to an "emergency"? And AI risk
AnnaSalamon
5mo
60
323
A challenge for AGI organizations, and a challenge for readers
Rob Bensinger
19d
30
314
Why Agent Foundations? An Overly Abstract Explanation
johnswentworth
9mo
54
1043
AGI Ruin: A List of Lethalities
Eliezer Yudkowsky
6mo
653
1039
Where I agree and disagree with Eliezer
paulfchristiano
6mo
205
455
Counterarguments to the basic AI x-risk case
KatjaGrace
2mo
122
437
What failure looks like
paulfchristiano
3y
49
432
Discussion with Eliezer Yudkowsky on AGI interventions
Rob Bensinger
1y
257
309
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res
6mo
48
266
What Multipolar Failure Looks Like, and Robust Agent-Agnostic Processes (RAAPs)
Andrew_Critch
1y
60
253
Another (outer) alignment failure story
paulfchristiano
1y
38
233
A Quick Guide to Confronting Doom
Ruby
8mo
36
219
Some AI research areas and their relevance to existential safety
Andrew_Critch
2y
40
217
Slow motion videos as AI risk intuition pumps
Andrew_Critch
6mo
36
205
On saving one's world
Rob Bensinger
7mo
5
205
[RETRACTED] It's time for EA leadership to pull the short-timelines fire alarm.
Not Relevant
8mo
165
192
Whole Brain Emulation: No Progress on C. elgans After 10 Years
niconiconi
1y
77