AI ALIGNMENT FORUM
AF

2239
Judd Rosenblatt2
000
Message
Dialogue
Subscribe

AE Studio is a team of 160+ programmers, product designers, and data scientists focused on increasing human agency through neglected high-impact approaches. Originally successful in BCI development and consulting, we're now applying our expertise to AI alignment research, believing that the space of plausible alignment solutions is vast and under-explored.

Our alignment work includes prosociality research on self-modeling in neural systems, with attention schema theory in particular, self-other overlap mechanisms, and various neglected technical and policy approaches. We maintain a profitable consulting business that allows us to fund and pursue promising but overlooked research directions without pressure to expedite AGI development.

Learn more about us and our mission here: 
https://ae.studio/ai-alignment

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No Comments Found
No wikitag contributions to display.
29Mistral Large 2 (123B) seems to exhibit alignment faking
7mo
0
32Reducing LLM deception at scale with self-other overlap fine-tuning
7mo
9
20Self-prediction acts as an emergent regularizer
1y
0
59Self-Other Overlap: A Neglected Approach to AI Alignment
1y
7
33Survey for alignment researchers!
2y
3