AI ALIGNMENT FORUMValue Learning
AF

Value Learning

Oct 29, 2018 by Rohin Shah

This is a sequence investigating the feasibility of one approach to AI alignment: value learning.

21Preface to the sequence on value learning
Rohin Shah
4y
5
Ambitious Value Learning
17What is ambitious value learning?
Rohin Shah
4y
12
11The easy goal inference problem is still hard
Paul Christiano
4y
1
11Humans can be assigned any values whatsoever…
Stuart Armstrong
4y
1
8Latent Variables and Model Mis-Specification
Jacob Steinhardt
4y
1
11Model Mis-specification and Inverse Reinforcement Learning
Owain Evans, Jacob Steinhardt
4y
0
13Future directions for ambitious value learning
Rohin Shah
4y
6
Goals vs Utility Functions

Ambitious value learning aims to give the AI the correct utility function to avoid catastrophe. Given its difficulty, we revisit the arguments for utility functions in the first place.

19Intuitions about goal-directed behavior
Rohin Shah
4y
5
40Coherence arguments do not entail goal-directed behavior
Rohin Shah
4y
49
21Will humans build goal-directed agents?
Rohin Shah
4y
21
25AI safety without goal-directed behavior
Rohin Shah
4y
6
Narrow Value Learning
8What is narrow value learning?
Rohin Shah
4y
3
8Ambitious vs. narrow value learning
Paul Christiano
4y
15
16Human-AI Interaction
Rohin Shah
4y
0
9Reward uncertainty
Rohin Shah
4y
0
9The human side of interaction
Rohin Shah
4y
1
14Following human norms
Rohin Shah
4y
4
6Future directions for narrow value learning
Rohin Shah
4y
2
20Conclusion to the sequence on value learning
Rohin Shah
4y
1