Top postsTop post
steven0461
8790
Ω
31
44
1575
157
This is where I'll put content that's too short for a whole post.
[Epistemic status: ¯\_(ツ)_/¯ ] Armstrong and Mindermann write about a no free lunch theorem for inverse reinforcement learning (IRL): the same action can reflect many different combinations of values and (irrational) planning algorithms. I think even assuming humans were fully rational expected utility maximizers, there would be an important underdetermination...