Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
31 posts
SERI MATS
AI Alignment Fieldbuilding
Intellectual Progress (Society-Level)
Distillation & Pedagogy
Practice & Philosophy of Science
Information Hazards
PIBBSS
Intellectual Progress via LessWrong
Economic Consequences of AGI
Privacy
Superintelligence
Automation
532 posts
Epistemology
Intellectual Progress (Individual-Level)
Research Taste
Epistemic Review
Selection Effects
Social & Cultural Dynamics
Humility
20
Reflections on the PIBBSS Fellowship 2022
Nora_Ammann
9d
0
149
Lessons learned from talking to >100 academics about AI safety
Marius Hobbhahn
2mo
16
164
Most People Start With The Same Few Bad Ideas
johnswentworth
3mo
30
126
Your posts should be on arXiv
JanBrauner
3mo
39
66
SERI MATS Program - Winter 2022 Cohort
Ryan Kidd
2mo
12
163
Call For Distillers
johnswentworth
8mo
42
82
Principles of Privacy for Alignment Research
johnswentworth
4mo
30
71
Conjecture: Internal Infohazard Policy
Connor Leahy
4mo
6
39
[An email with a bunch of links I sent an experienced ML researcher interested in learning about Alignment / x-safety.]
David Scott Krueger (formerly: capybaralet)
3mo
1
16
Auditing games for high-level interpretability
Paul Colognese
1mo
1
87
Intuitions about solving hard problems
Richard_Ngo
7mo
23
90
Productive Mistakes, Not Perfect Answers
adamShimi
8mo
11
30
Behaviour Manifolds and the Hessian of the Total Loss - Notes and Criticism
Spencer Becker-Kahn
3mo
4
25
Epistemic Artefacts of (conceptual) AI alignment research
Nora_Ammann
4mo
1
95
How to do theoretical research, a personal perspective
Mark Xu
4mo
4
53
Methodological Therapy: An Agenda For Tackling Research Bottlenecks
adamShimi
2mo
6
17
Attempts at Forwarding Speed Priors
james.lucassen
2mo
2
63
How I Formed My Own Views About AI Safety
Neel Nanda
9mo
6
17
Shapes of Mind and Pluralism in Alignment
adamShimi
4mo
1
169
Alignment Research Field Guide
abramdemski
3y
9
45
David Wolpert on Knowledge
Alex Flint
1y
3
41
Epistemic Strategies of Selection Theorems
adamShimi
1y
1
20
Epistemology of HCH
adamShimi
1y
2
38
AI Alignment Open Thread August 2019
habryka
3y
96
18
Extended Picture Theory or Models inside Models inside Models
Chris_Leong
1y
3
30
On the falsifiability of hypercomputation
jessicata
2y
4
37
Uncertainty versus fuzziness versus extrapolation desiderata
Stuart_Armstrong
3y
8
27
Positive Feedback -> Optimization?
johnswentworth
2y
6