AI Alignment Posts

Standard ML Oracles vs Counterfactual ones

Stuart_Armstrong5d5 points5 min readShow Highlightsubdirectory_arrow_left
0

A Rationality Condition for CDT Is That It Equal EDT (Part 2)

abramdemski6d3 points7 min readShow Highlightsubdirectory_arrow_left
0

Alignment Newsletter #27

rohinmshahAlignment Newsletter7d1 point9 min readShow Highlightsubdirectory_arrow_left
0

A Rationality Condition for CDT Is That It Equal EDT (Part 1)

abramdemski11d4 points8 min readShow Highlightsubdirectory_arrow_left
3

The Rocket Alignment Problem

Eliezer_Yudkowsky12d19 points15 min readShow Highlightsubdirectory_arrow_left
0

EDT solves 5 and 10 with conditional oracles

jessicata15d17 points12 min readShow Highlightsubdirectory_arrow_left
5

[Link] New DeepMind AI Safety Research Blog

Vika18d13 points1 min readShow Highlightsubdirectory_arrow_left
0

Asymptotic Decision Theory (Improved Writeup)

Diffractor18d6 points13 min readShow Highlightsubdirectory_arrow_left
8

Wireheading as a potential problem with the new impact measure

Stuart_Armstrong20d8 points4 min readShow Highlightsubdirectory_arrow_left
19

Alignment Newsletter #25

rohinmshahAlignment Newsletter21d6 points9 min readShow Highlightsubdirectory_arrow_left
3