Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
13671 posts
Rationality
World Modeling
Practical
World Optimization
Covid-19
Community
Fiction
Site Meta
Scholarship & Learning
Politics
Book Reviews
Open Threads
18722 posts
AI
AI Risk
GPT
AI Timelines
Decision Theory
Interpretability (ML & AI)
Machine Learning (ML)
AI Takeoff
Inner Alignment
Anthropics
Research Agendas
Language Models
150
How to Convince my Son that Drugs are Bad
concerned_dad
3d
77
61
AGI Timelines in Governance: Different Strategies for Different Timeframes
simeon_c
1d
14
65
Shard Theory in Nine Theses: a Distillation and Critical Appraisal
LawrenceC
1d
9
12
Two Dogmas of LessWrong
omnizoid
5d
142
-20
Why Are Women Hot?
Jacob Falkovich
2d
10
8
Open & Welcome Thread - December 2022
niplav
16d
10
9
Who should write the definitive post on Ziz?
NicholasKross
5d
30
12
Boston Solstice 2022 Retrospective
jefftk
2d
2
22
Why I'm Sceptical of Foom
DragonGod
12d
26
40
Positive values seem more robust and lasting than prohibitions
TurnTrout
3d
9
4
Reflections: Bureaucratic Hell
Haris Rashid
4h
1
5
Can we, in principle, know the measure of counterfactual quantum branches?
sisyphus
2d
15
0
Where's the economic incentive for wokism coming from?
Valentine
12d
97
20
There have been 3 planes (billionaire donors) and 2 have crashed
Trevor1
3d
8
28
K-complexity is silly; use cross-entropy instead
So8res
1h
4
27
Discovering Language Model Behaviors with Model-Written Evaluations
evhub
4h
3
84
Towards Hodge-podge Alignment
Cleo Nardo
1d
20
16
An Open Agency Architecture for Safe Transformative AI
davidad
11h
11
198
The next decades might be wild
Marius Hobbhahn
5d
21
6
I believe some AI doomers are overconfident
FTPickle
6h
4
170
Can you control the past?
Joe Carlsmith
1y
93
41
The "Minimal Latents" Approach to Natural Abstractions
johnswentworth
22h
6
37
Reframing inner alignment
davidad
9d
13
7
Will research in AI risk jinx it? Consequences of training AI on AI risk arguments
Yann Dubois
1d
6
112
Bad at Arithmetic, Promising at Math
cohenmacaulay
2d
17
52
Existential AI Safety is NOT separate from near-term applications
scasper
7d
15
47
Next Level Seinfeld
Zvi
1d
6
26
Take 9: No, RLHF/IDA/debate doesn't solve outer alignment.
Charlie Steiner
8d
14