AI ALIGNMENT FORUM
AF

1017
Formal Alignment

Formal Alignment

Dec 04, 2019 by Gordon Seidoh Worley

Alignment is typical defined loosely as "AI aligned with human intent, values, or preferences". This developing sequence of posts is part of an investigation into means for formally stating alignment in a precise enough way that we can use mathematics to formally verify if a proposed alignment mechanism would achieve alignment.

4Formally Stating the AI Alignment Problem
Gordon Seidoh Worley
8y
0
6Minimization of prediction error as a foundation for human values in AI alignment
Gordon Seidoh Worley
6y
0
5Values, Valence, and Alignment
Gordon Seidoh Worley
6y
2
4Towards deconfusing values
Gordon Seidoh Worley
6y
4
10Deconfusing Human Values Research Agenda v1
Gordon Seidoh Worley
6y
10