AI ALIGNMENT FORUM
AF

Benjamin Hilton
Ω69100
Message
Dialogue
Subscribe

Head of Alignment at UK AI Security Institute (AISI). Previously 80,000 Hours, HM Treasury, Cabinet Office, Department for International Trade, Imperial College London.

Sequences

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No Comments Found
No wikitag contributions to display.
The Alignment Project Research Agenda
UK AISI Alignment Team: Debate Sequence
6Research Areas in Methods for Post-training and Elicitation (The Alignment Project by UK AISI)
1mo
0
7Research Areas in Benchmark Design and Evaluation (The Alignment Project by UK AISI)
1mo
0
3Research Areas in Probabilistic Methods (The Alignment Project by UK AISI)
1mo
0
8Research Areas in Evaluation and Guarantees in Reinforcement Learning (The Alignment Project by UK AISI)
1mo
0
13The Alignment Project by UK AISI
1mo
0
35An alignment safety case sketch based on debate
4mo
15
47UK AISI’s Alignment Team: Research Agenda
4mo
2
33A sketch of an AI control safety case
7mo
0
38Automation collapse
11mo
7