AI ALIGNMENT FORUM
AF

870
niplav
0032
Message
Dialogue
Subscribe

I operate by Crocker's rules. All LLM output is explicitely designated as such. I have made no self-hiding agreements.

Website.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No posts to display.
0shortplav
5y
0
A list of core AI safety problems and how I hope to solve them
niplav2y12

One issue I see with this plan is that it seems to rely on some mathematics that appear to me to not be fully worked out, e.g. infrabayesianism and «boundaries» (for which I haven't been able to find a full mathematical description), and it looks unclear to me whether they will actually be finished in time, and if they are, whether they lead to algorithms that are efficient enough to be scaled to such an ambitious project.

Reply
AI-Assisted Alignment
4 months ago
(+54)
AI-Assisted Alignment
4 months ago
(+127/-8)
Recursive Self-Improvement
4 months ago
(+68)
Alief
5 months ago
(+11/-11)
Old Less Wrong About Page
7 months ago
Successor alignment
8 months ago
(+26/-3)
Cooking
a year ago
(+26/-163)
Future of Humanity Institute (FHI)
a year ago
(+11)
Future of Humanity Institute (FHI)
a year ago
(+121/-49)
Axiom
2 years ago
(+112/-82)
Load More