AI Alignment Posts

0

Thoughts on Human Models

111d9 min readShow Highlight
1

Pavlov Generalizes

152d7 min readShow Highlight
0

How the MtG Color Wheel Explains AI Safety

177d5 min readShow Highlight
2

Alignment Newsletter #45

99dShow Highlight
0

Humans interpreting humans

49dShow Highlight
0

Anchoring vs Taste: a model

49dShow Highlight
0

Nuances with ascription universality

810dShow Highlight
1

[Link] Learning preferences by looking at the world

1410d6 min readShow Highlight
4

Would I think for ten thousand years?

911dShow Highlight
6