AI Alignment Posts

Introducing the AI Alignment Forum (FAQ)

212mo6 min readShow Highlight
0

Three AI Safety Related Ideas

173d1 min readShow Highlight
11

Alignment Newsletter #36

55d10 min readShow Highlight
0

Figuring out what Alice wants: non-human Alice

45d1 min readShow Highlight
15

Assuming we've solved X, could we do Y...

125d2 min readShow Highlight
2

COEDT Equilibria in Games

710d3 min readShow Highlight
0

Why we need a *theory* of human values

1311d4 min readShow Highlight
3

Factored Cognition

812d16 min readShow Highlight
1

Alignment Newsletter #35

513d6 min readShow Highlight
0

Coherence arguments do not imply goal-directed behavior

1314d7 min readShow Highlight
9