AI ALIGNMENT FORUM
AF

207
Wikitags

Research Agendas

Edited by plex last updated 16th Sep 2021

Research Agendas lay out the areas of research which individuals or groups are working on, or those that they believe would be valuable for others to work on. They help make research more legible and encourage discussion of priorities.

Subscribe
Discussion
1
Subscribe
Discussion
1
Posts tagged Research Agendas
38Research Agenda: Synthesizing Standalone World-Models
Thane Ruthenis
1mo
7
47UK AISI’s Alignment Team: Research Agenda
Benjamin Hilton, Jacob Pfau, Marie_DB, Geoffrey Irving
6mo
2
10How to Contribute to Theoretical Reward Learning Research
Joar Skalse
8mo
0
15The Theoretical Reward Learning Research Agenda: Introduction and Motivation
Joar Skalse
8mo
4
45My AGI safety research—2024 review, ’25 plans
Steven Byrnes
10mo
3
73Shallow review of technical AI safety, 2024
technicalities, Stag, Stephen McAleese, jordine, Dr. David Mathers
10mo
1
32Seeking Collaborators
abramdemski
1y
8
20Self-prediction acts as an emergent regularizer
Cameron Berg, Judd Rosenblatt, Mike Vaiana, Diogo de Lucena, florin_pop, Trent Hodgeson
1y
0
22Announcing Human-aligned AI Summer School
Jan_Kulveit, Tomáš Gavenčiak
1y
0
69EIS XIII: Reflections on Anthropic’s SAE Research Circa May 2024
scasper
1y
7
29Constructability: Plainly-coded AGIs may be feasible in the near future
Épiphanie Gédéon, Charbel-Raphaël
2y
1
43Sparsify: A mechanistic interpretability research agenda
Lee Sharkey
2y
17
26Gradient Descent on the Human Brain
Jozdien, gaspode
2y
0
20Natural abstractions are observer-dependent: a conversation with John Wentworth
Martín Soto
2y
0
46Four visions of Transformative AI success
Steven Byrnes
2y
11
Load More (15/107)
Add Posts