AI Alignment Posts

0

Alignment Newsletter #46

81d9 min readShow Highlight
0

Thoughts on Human Models

172d9 min readShow Highlight
1

Pavlov Generalizes

153d7 min readShow Highlight
0

How the MtG Color Wheel Explains AI Safety

177d5 min readShow Highlight
2

Alignment Newsletter #45

99dShow Highlight
0

Humans interpreting humans

410dShow Highlight
0

Anchoring vs Taste: a model

410dShow Highlight
0

Nuances with ascription universality

810dShow Highlight
1

[Link] Learning preferences by looking at the world

1410d6 min readShow Highlight
4