Go Back
Choose this branch
Choose this branch
meritocratic
regular
democratic
hot
top
alive
55 posts
Language Models
Agency
Deconfusion
Scaling Laws
Tool AI
Definitions
Simulation Hypothesis
PaLM
Prompt Engineering
Philosophy of Language
Carving / Clustering Reality
Astronomical Waste
33 posts
Conjecture (org)
Refine
Project Announcement
Encultured AI (org)
Analogy
28
Discovering Language Model Behaviors with Model-Written Evaluations
evhub
4h
3
27
Take 11: "Aligning language models" should be weirder.
Charlie Steiner
2d
0
185
Simulators
janus
3mo
103
234
chinchilla's wild implications
nostalgebraist
4mo
114
163
Language models seem to be much better than humans at next-token prediction
Buck
4mo
56
75
Inverse Scaling Prize: Round 1 Winners
Ethan Perez
2mo
16
28
Inverse scaling can become U-shaped
Edouard Harris
1mo
15
141
Announcing the Inverse Scaling Prize ($250k Prize Pool)
Ethan Perez
5mo
14
56
LOVE in a simbox is all you need
jacob_cannell
2mo
69
45
Smoke without fire is scary
Adam Jermyn
2mo
22
99
Help ARC evaluate capabilities of current language models (still need people)
Beth Barnes
5mo
6
63
Vingean Agency
abramdemski
3mo
13
40
Paper: Large Language Models Can Self-improve [Linkpost]
Evan R. Murphy
2mo
14
113
The case for becoming a black-box investigator of language models
Buck
7mo
19
64
[Interim research report] Taking features out of superposition with sparse autoencoders
Lee Sharkey
7d
10
143
Conjecture: a retrospective after 8 months of work
Connor Leahy
27d
9
96
The Singular Value Decompositions of Transformer Weight Matrices are Highly Interpretable
beren
22d
27
108
What I Learned Running Refine
adamShimi
26d
5
178
Mysteries of mode collapse
janus
1mo
35
52
Conjecture Second Hiring Round
Connor Leahy
27d
0
31
Searching for Search
NicholasKees
22d
6
41
Current themes in mechanistic interpretability research
Lee Sharkey
1mo
3
105
Announcing Encultured AI: Building a Video Game
Andrew_Critch
4mo
26
56
Interpreting Neural Networks through the Polytope Lens
Sid Black
2mo
26
42
My Thoughts on the ML Safety Course
zeshen
2mo
3
23
Embedding safety in ML development
zeshen
1mo
1
68
How to Diversify Conceptual Alignment: the Model Behind Refine
adamShimi
5mo
11
118
We Are Conjecture, A New Alignment Research Startup
Connor Leahy
8mo
24