AI ALIGNMENT FORUM
AF

918
Wikitags

Myopia

Edited by Dakara, abramdemski, et al. last updated 30th Dec 2024

Myopia refers to short-sightedness in planning and decision-making processes. It describes a tendency to prioritize immediate or short-term outcomes while disregarding longer-term consequences.

The most extreme form of myopia occurs when an agent considers only immediate rewards, completely disregarding future consequences. In artificial intelligence contexts, a perfectly myopic agent would optimize solely for the current query or task without attempting to influence future outcomes.

Myopic agents demonstrate several notable properties:

  • Limited temporal scope in decision-making
  • Focus on immediate reward optimization
  • Reduced instrumental incentives
Subscribe
Discussion
Subscribe
Discussion
Posts tagged Myopia
1
142Simulators
janus
3y
90
2
44MONA: Managed Myopia with Approval Feedback
Seb Farquhar, David Lindner, Rohin Shah
10mo
7
2
17MONA: Three Month Later - Updates and Steganography Without Optimization Pressure
David Lindner, Vikrant Varma
7mo
0
4
62How LLMs are and are not myopic
janus
2y
7
2
91The Parable of Predict-O-Matic
abramdemski
6y
16
2
8AXRP Episode 43 - David Lindner on Myopic Optimization with Non-myopic Approval
DanielFilan
5mo
0
1
72An overview of 11 proposals for building safe advanced AI
evhub
5y
32
2
41Thoughts on “Process-Based Supervision” / MONA
Steven Byrnes
2y
1
2
48You can still fetch the coffee today if you're dead tomorrow
davidad
3y
14
1
54Seeking Power is Often Convergently Instrumental in MDPs
TurnTrout, Logan Riggs
6y
34
3
40The Credit Assignment Problem
abramdemski
6y
32
1
23Interpretability’s Alignment-Solving Potential: Analysis of 7 Scenarios
Evan R. Murphy
4y
0
1
17Non-myopia stories
[anonymous]2y
5
0
18Acceptability Verification: A Research Agenda
David Udell, evhub
3y
0
2
24Steering Behaviour: Testing for (Non-)Myopia in Language Models
Evan R. Murphy, Megan Kinniment
3y
5
Load More (15/34)
Add Posts