Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
166 posts
AI Risk
Missing Moods
Seed AI
29 posts
Threat Models
Sharp Left Turn
Multipolar Scenarios
Technological Unemployment
59
AI Safety Seems Hard to Measure
HoldenKarnofsky
12d
5
58
Who are some prominent reasonable people who are confident that AI won't kill everyone?
Optimization Process
15d
40
93
AI will change the world, but won’t take it over by playing “3-dimensional chess”.
boazbarak
28d
86
87
Meta AI announces Cicero: Human-Level Diplomacy play (with dialogue)
Jacy Reese Anthis
28d
64
217
Counterarguments to the basic AI x-risk case
KatjaGrace
2mo
122
515
Where I agree and disagree with Eliezer
paulfchristiano
6mo
205
405
AGI Ruin: A List of Lethalities
Eliezer Yudkowsky
6mo
653
79
Am I secretly excited for AI getting weird?
porby
1mo
4
109
Niceness is unnatural
So8res
2mo
18
22
Apply to attend winter AI alignment workshops (Dec 28-30 & Jan 3-5) near Berkeley
Akash
19d
1
17
Aligned Behavior is not Evidence of Alignment Past a Certain Level of Intelligence
Ronny Fernandez
15d
5
140
Worlds Where Iterative Design Fails
johnswentworth
3mo
26
80
What does it mean for an AGI to be 'safe'?
So8res
2mo
32
201
Slow motion videos as AI risk intuition pumps
Andrew_Critch
6mo
36
36
AI Neorealism: a threat model & success criterion for existential safety
davidad
5d
0
41
Refining the Sharp Left Turn threat model, part 2: applying alignment techniques
Vika
25d
4
12
How is the "sharp left turn defined"?
Chris_Leong
12d
3
56
Clarifying AI X-risk
zac_kenton
1mo
23
197
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res
6mo
48
42
Threat Model Literature Review
zac_kenton
1mo
4
171
AI Could Defeat All Of Us Combined
HoldenKarnofsky
6mo
29
25
AI X-risk >35% mostly based on a recent peer-reviewed argument
michaelcohen
1mo
31
37
It matters when the first sharp left turn happens
Adam Jermyn
2mo
9
55
Refining the Sharp Left Turn threat model, part 1: claims and mechanisms
Vika
4mo
3
30
We may be able to see sharp left turns coming
Ethan Perez
3mo
26
167
Another (outer) alignment failure story
paulfchristiano
1y
38
140
What Multipolar Failure Looks Like, and Robust Agent-Agnostic Processes (RAAPs)
Andrew_Critch
1y
60
104
Less Realistic Tales of Doom
Mark Xu
1y
13