Strengthening Red Teams: A Modular Scaffold for Control Evaluations
Agentic sabotage occurs when an AI system intentionally takes actions that compromise safety in an environment while pursuing its own misaligned objectives. We evaluate this behavior in SHADE-Arena, a dataset for assessing sabotage and monitoring capabilities of LLM agents. We decompose agentic sabotage capability into five core components—knowing what overseers...
Nov 11, 20257