Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
166 posts
AI Risk
Missing Moods
Seed AI
29 posts
Threat Models
Sharp Left Turn
Multipolar Scenarios
Technological Unemployment
68
AI Safety Seems Hard to Measure
HoldenKarnofsky
12d
5
336
Counterarguments to the basic AI x-risk case
KatjaGrace
2mo
122
61
Who are some prominent reasonable people who are confident that AI won't kill everyone?
Optimization Process
15d
40
103
AI will change the world, but won’t take it over by playing “3-dimensional chess”.
boazbarak
28d
86
95
Meta AI announces Cicero: Human-Level Diplomacy play (with dialogue)
Jacy Reese Anthis
28d
64
777
Where I agree and disagree with Eliezer
paulfchristiano
6mo
205
724
AGI Ruin: A List of Lethalities
Eliezer Yudkowsky
6mo
653
98
Am I secretly excited for AI getting weird?
porby
1mo
4
25
Apply to attend winter AI alignment workshops (Dec 28-30 & Jan 3-5) near Berkeley
Akash
19d
1
19
Aligned Behavior is not Evidence of Alignment Past a Certain Level of Intelligence
Ronny Fernandez
15d
5
98
Niceness is unnatural
So8res
2mo
18
144
Worlds Where Iterative Design Fails
johnswentworth
3mo
26
72
What does it mean for an AGI to be 'safe'?
So8res
2mo
32
209
Slow motion videos as AI risk intuition pumps
Andrew_Critch
6mo
36
39
AI Neorealism: a threat model & success criterion for existential safety
davidad
5d
0
102
Clarifying AI X-risk
zac_kenton
1mo
23
36
Refining the Sharp Left Turn threat model, part 2: applying alignment techniques
Vika
25d
4
13
How is the "sharp left turn defined"?
Chris_Leong
12d
3
55
Threat Model Literature Review
zac_kenton
1mo
4
253
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res
6mo
48
36
AI X-risk >35% mostly based on a recent peer-reviewed argument
michaelcohen
1mo
31
168
AI Could Defeat All Of Us Combined
HoldenKarnofsky
6mo
29
71
Refining the Sharp Left Turn threat model, part 1: claims and mechanisms
Vika
4mo
3
50
We may be able to see sharp left turns coming
Ethan Perez
3mo
26
35
It matters when the first sharp left turn happens
Adam Jermyn
2mo
9
210
Another (outer) alignment failure story
paulfchristiano
1y
38
203
What Multipolar Failure Looks Like, and Robust Agent-Agnostic Processes (RAAPs)
Andrew_Critch
1y
60
319
What failure looks like
paulfchristiano
3y
49