The risk-reward tradeoff of interpretability research — AI Alignment Forum