AI ALIGNMENT FORUM
AF

369
The Ethicophysics

The Ethicophysics

Nov 30, 2023 by MadHatter

In this sequence, we attempt to solve the alignment problem, rather than discussing it ad infinitum. Since the alignment problem is incredibly difficult to solve, this sequence is probably going to end up being pretty long, and many of the posts will be more complex and harder to read than they really have to be. We apologize to the reader for this situation, and promise to improve the individual posts and the overall flow of the sequence as quickly as our limited time permits.

44Moral Reality Check (a short story)
jessicata
2y
0
39Agent Boundaries Aren't Markov Blankets. [Unless they're non-causal; see comments.]
abramdemski
2y
6
-4My Alignment Research Agenda ("the Ethicophysics")
MadHatter
2y
0
1Some Intuitions for the Ethicophysics
MadHatter, mishka
2y
0
-7The Alignment Agenda THEY Don't Want You to Know About
MadHatter
2y
0
-2My Mental Model of Infohazards
MadHatter
2y
0
16Stupid Question: Why am I getting consistently downvoted?
Q
MadHatter, Shankar Sivarajan
2y
Q
0
35Trying to Make a Treacherous Mesa-Optimizer
MadHatter
3y
1
-8Homework Answer: Glicko Ratings for War
MadHatter
2y
0
-3Enkrateia: a safe model-based reinforcement learning algorithm
MadHatter
2y
0
-4A Formula for Violence (and Its Antidote)
MadHatter
2y
0