[AN #119]: AI safety when agents are shaped by environments, not rewards — AI Alignment Forum