Exploring out-of-context reasoning (OOCR) fine-tuning in LLMs to increase test-phase awareness — AI Alignment Forum