AI ALIGNMENT FORUM
AF

324
rusheb
Ω1000
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
Causal Scrubbing: a method for rigorously testing interpretability hypotheses [Redwood Research]
rusheb3y10

If your hypothesis predicts that model performance will be preserved if you swap the input to any other input which has a particular property, but no other inputs in the dataset have that property, causal scrubbing can’t test your hypothesis

 

Would it be possible to make interventions which we expect not to preserve the model's behaviour, and assert that the behaviour does in fact change? 

Reply
57Stress Testing Deliberative Alignment for Anti-Scheming Training
1mo
10
89Frontier Models are Capable of In-context Scheming
10mo
9
42Apollo Research 1-year update
1y
0
26A starter guide for evals
2y
0
11Scalable And Transferable Black-Box Jailbreaks For Language Models Via Persona Modulation
2y
1
21Understanding mesa-optimization using toy models
2y
0