AI Alignment Posts

Introducing the AI Alignment Forum (FAQ)

213mo6 min readShow Highlight
0

One-step hypothetical preferences

43h3 min readShow Highlight
0

Learning with catastrophes

515h3 min readShow Highlight
0

Alignment Newsletter #42

102d9 min readShow Highlight
0

Disentangling arguments for the importance of AI safety

162d7 min readShow Highlight
6

Following human norms

93d4 min readShow Highlight
0

Announcement: AI alignment prize round 4 winners

173d1 min readShow Highlight
0

Capability amplification

73d13 min readShow Highlight
1

[Question] Why not tool AI?

44d1 min readShow Highlight
1

Reward uncertainty

65d5 min readShow Highlight
0