Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
13671 posts
Rationality
World Modeling
Practical
World Optimization
Covid-19
Community
Fiction
Site Meta
Scholarship & Learning
Politics
Book Reviews
Open Threads
18722 posts
AI
AI Risk
GPT
AI Timelines
Decision Theory
Interpretability (ML & AI)
Machine Learning (ML)
AI Takeoff
Inner Alignment
Anthropics
Research Agendas
Language Models
106
How to Convince my Son that Drugs are Bad
concerned_dad
3d
77
35
AGI Timelines in Governance: Different Strategies for Different Timeframes
simeon_c
1d
14
75
Shard Theory in Nine Theses: a Distillation and Critical Appraisal
LawrenceC
1d
9
-24
Two Dogmas of LessWrong
omnizoid
5d
142
54
Why Are Women Hot?
Jacob Falkovich
2d
10
8
Open & Welcome Thread - December 2022
niplav
16d
10
1
Who should write the definitive post on Ziz?
NicholasKross
5d
30
24
Boston Solstice 2022 Retrospective
jefftk
2d
2
10
Why I'm Sceptical of Foom
DragonGod
12d
26
44
Positive values seem more robust and lasting than prohibitions
TurnTrout
3d
9
-6
Reflections: Bureaucratic Hell
Haris Rashid
4h
1
-3
Can we, in principle, know the measure of counterfactual quantum branches?
sisyphus
2d
15
-12
Where's the economic incentive for wokism coming from?
Valentine
12d
97
10
There have been 3 planes (billionaire donors) and 2 have crashed
Trevor1
3d
8
46
K-complexity is silly; use cross-entropy instead
So8res
1h
4
27
Discovering Language Model Behaviors with Model-Written Evaluations
evhub
4h
3
40
Towards Hodge-podge Alignment
Cleo Nardo
1d
20
10
An Open Agency Architecture for Safe Transformative AI
davidad
11h
11
108
The next decades might be wild
Marius Hobbhahn
5d
21
0
I believe some AI doomers are overconfident
FTPickle
6h
4
124
Can you control the past?
Joe Carlsmith
1y
93
33
The "Minimal Latents" Approach to Natural Abstractions
johnswentworth
22h
6
57
Reframing inner alignment
davidad
9d
13
3
Will research in AI risk jinx it? Consequences of training AI on AI risk arguments
Yann Dubois
1d
6
70
Bad at Arithmetic, Promising at Math
cohenmacaulay
2d
17
22
Existential AI Safety is NOT separate from near-term applications
scasper
7d
15
43
Next Level Seinfeld
Zvi
1d
6
46
Take 9: No, RLHF/IDA/debate doesn't solve outer alignment.
Charlie Steiner
8d
14