AI ALIGNMENT FORUM
Books of LessWrong
AF

Alignment & Agency

Apr 09, 2022 by Raymond Arnold
61An Orthodox Case Against Utility Functions
Abram Demski
5y
45
53The Pointers Problem: Human Values Are A Function Of Humans' Latent Variables
johnswentworth
5y
34
63Alignment By Default
johnswentworth
5y
72
72An overview of 11 proposals for building safe advanced AI
Evan Hubinger
5y
32
93The ground of optimization
Alex Flint
5y
50
34Search versus design
Alex Flint
5y
30
59Inner Alignment: Explain like I'm 12 Edition
Rafael Harth
5y
12
44Inaccessible information
Paul Christiano
5y
9
39AGI safety from first principles: Introduction
Richard Ngo
5y
15
40Is Success the Enemy of Freedom? (Full)
alkjash
5y
0