AI ALIGNMENT FORUM
AF

Ongoing project on moral AI

Jan 23, 2024 by Michele Campolo

For lack of a better name =)

The idea is to use current AI technologies, like language models, to get an impartial AI that understands ethics as humans do, possibly even better.

You heard me right: in the same way as an AI can be smarter than a human, we should also accept the fact that we are not morally perfect creatures, and that it's possible to create an AI which is better than us at, for example, spotting injustice. See Free agents for more details.

In case you are familiar with philosophical language, my objective is a philosopher AI that figures out epistemology and ethics on its own, and then communicates its beliefs.

In case you look at reality through the lenses of AI alignment, I am saying that going for ‘safe’ or ‘aligned’ is kind of lame, and that aiming for ‘moral’ is better. Instead of trying to limit the side effects of, or fix, agents which are morally clueless, I'd like to see more people working on agents which perceive and interpret the world from a human-like point of view.

If you are looking for a place to start, I suggest that you have a look at Free agents and decide where to go from there. Although it touches on some technical subjects, I tried to write that post for a relatively broad audience.

This sequence is just a collection of posts about the same topic, in chronological order. Later on, I expect that posts will become more algorithmic and, finally, about practical experiments run on hardware.

You can find this sequence also on the EA Forum.

5Naturalism and AI alignment
Michele Campolo
4y
10
2From language to ethics by automated reasoning
Michele Campolo
4y
0
7Criticism of the main framework in AI alignment
Michele Campolo
2y
0
3On value in humans, other animals, and AI
Michele Campolo
2y
16
0Free agents
Michele Campolo
2y
10
2Agents that act for reasons: a thought experiment
Michele Campolo
1y
0