x
This website requires javascript to properly function. Consider activating javascript to get access to all site functionality.
AI ALIGNMENT FORUM
AF
Login
Robert Kirk — AI Alignment Forum
Robert Kirk
Posts
Sorted by New
Wikitag Contributions
Comments
Sorted by
Newest
5
Supervised finetuning on low-harm reward hacking generalises to high-harm reward hacking
1mo
0
6
Layered AI Defenses Have Holes: Vulnerabilities and Key Recommendations
7mo
0
Comments