Go Back
Choose this branch
You can't go any further
meritocratic
regular
democratic
hot
top
alive
1854 posts
AI
AI Sentience
Truthful AI
1 posts
62
Towards Hodge-podge Alignment
Cleo Nardo
1d
20
37
The "Minimal Latents" Approach to Natural Abstractions
johnswentworth
22h
6
21
Take 12: RLHF's use is evidence that orgs will jam RL at real-world problems.
Charlie Steiner
19h
0
153
The next decades might be wild
Marius Hobbhahn
5d
21
232
AI alignment is distinct from its near-term applications
paulfchristiano
7d
5
92
Trying to disambiguate different questions about whether RLHF is “good”
Buck
6d
39
3
I believe some AI doomers are overconfident
FTPickle
6h
4
265
A challenge for AGI organizations, and a challenge for readers
Rob Bensinger
19d
30
11
Solution to The Alignment Problem
Algon
1d
0
92
Revisiting algorithmic progress
Tamay
7d
6
18
Event [Berkeley]: Alignment Collaborator Speed-Meeting
AlexMennen
1d
2
83
Okay, I feel it now
g1
7d
14
159
Using GPT-Eliezer against ChatGPT Jailbreaking
Stuart_Armstrong
14d
77
59
Predicting GPU performance
Marius Hobbhahn
6d
24
12
Truthfulness, standards and credibility
Joe_Collman
8mo
2