maxnadeau

Wiki Contributions

Comments

Sorted by
maxnadeau4127

In the past, you've treated catching an AI escape attempt as a "win condition" for a control protocol, because such an attempt leads you to either 1) shut down your AI or 2) do few-shot catastrophe prevention.  You wrote that having these two tools would "substantially" change the situation w.r.t. the model's chance of success. 

But the arguments in this post suggest that catching an attempt may not change the overall strategic landscape/level of AI risk much. 

Is this a change in your views? Or have you always felt that catching an AI trying to escape substantially changes that AI's chance of success without substantially changing the overall level of AI risk?

Separately, another takeaway of this post: it seems valuable to work on research that allows you to provide evidence about the alignment of other labs' AIs, e.g. better technical work on searching for inputs that trigger competent misaligned behavior.

maxnadeau10

What are the considerations around whether to structure the debate to permit the judge to abstain (as Michael et al do, by allowing the judge to end the round with low credence) versus forcing the judge to pick an answer each time? Are there pros/cons to each approach? Any arguments about similarity of one or the other to the real AI debates that might be held in the future?
 

It's possible I'm misremembering/misunderstanding the protocols used for the debate here/in that other paper.