Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
13671 posts
Rationality
World Modeling
Practical
World Optimization
Covid-19
Community
Fiction
Site Meta
Scholarship & Learning
Politics
Book Reviews
Open Threads
18722 posts
AI
AI Risk
GPT
AI Timelines
Decision Theory
Interpretability (ML & AI)
Machine Learning (ML)
AI Takeoff
Inner Alignment
Anthropics
Research Agendas
Language Models
128
How to Convince my Son that Drugs are Bad
concerned_dad
3d
77
48
AGI Timelines in Governance: Different Strategies for Different Timeframes
simeon_c
1d
14
70
Shard Theory in Nine Theses: a Distillation and Critical Appraisal
LawrenceC
1d
9
-6
Two Dogmas of LessWrong
omnizoid
5d
142
17
Why Are Women Hot?
Jacob Falkovich
2d
10
8
Open & Welcome Thread - December 2022
niplav
16d
10
5
Who should write the definitive post on Ziz?
NicholasKross
5d
30
18
Boston Solstice 2022 Retrospective
jefftk
2d
2
16
Why I'm Sceptical of Foom
DragonGod
12d
26
42
Positive values seem more robust and lasting than prohibitions
TurnTrout
3d
9
-1
Reflections: Bureaucratic Hell
Haris Rashid
4h
1
1
Can we, in principle, know the measure of counterfactual quantum branches?
sisyphus
2d
15
-6
Where's the economic incentive for wokism coming from?
Valentine
12d
97
15
There have been 3 planes (billionaire donors) and 2 have crashed
Trevor1
3d
8
37
K-complexity is silly; use cross-entropy instead
So8res
1h
4
27
Discovering Language Model Behaviors with Model-Written Evaluations
evhub
4h
3
62
Towards Hodge-podge Alignment
Cleo Nardo
1d
20
13
An Open Agency Architecture for Safe Transformative AI
davidad
11h
11
153
The next decades might be wild
Marius Hobbhahn
5d
21
3
I believe some AI doomers are overconfident
FTPickle
6h
4
147
Can you control the past?
Joe Carlsmith
1y
93
37
The "Minimal Latents" Approach to Natural Abstractions
johnswentworth
22h
6
47
Reframing inner alignment
davidad
9d
13
5
Will research in AI risk jinx it? Consequences of training AI on AI risk arguments
Yann Dubois
1d
6
91
Bad at Arithmetic, Promising at Math
cohenmacaulay
2d
17
37
Existential AI Safety is NOT separate from near-term applications
scasper
7d
15
45
Next Level Seinfeld
Zvi
1d
6
36
Take 9: No, RLHF/IDA/debate doesn't solve outer alignment.
Charlie Steiner
8d
14