AI ALIGNMENT FORUM
AF

109
Wikitags

Myopia

Edited by Dakara, abramdemski, et al. last updated 30th Dec 2024

Myopia refers to short-sightedness in planning and decision-making processes. It describes a tendency to prioritize immediate or short-term outcomes while disregarding longer-term consequences.

The most extreme form of myopia occurs when an agent considers only immediate rewards, completely disregarding future consequences. In artificial intelligence contexts, a perfectly myopic agent would optimize solely for the current query or task without attempting to influence future outcomes.

Myopic agents demonstrate several notable properties:

  • Limited temporal scope in decision-making
  • Focus on immediate reward optimization
  • Reduced instrumental incentives
Subscribe
Discussion
Subscribe
Discussion
Posts tagged Myopia
142Simulators
janus
3y
90
44MONA: Managed Myopia with Approval Feedback
Seb Farquhar, David Lindner, Rohin Shah
8mo
7
17MONA: Three Month Later - Updates and Steganography Without Optimization Pressure
David Lindner, Vikrant Varma
5mo
0
62How LLMs are and are not myopic
janus
2y
7
8AXRP Episode 43 - David Lindner on Myopic Optimization with Non-myopic Approval
DanielFilan
3mo
0
91The Parable of Predict-O-Matic
abramdemski
6y
16
72An overview of 11 proposals for building safe advanced AI
evhub
5y
32
39Thoughts on “Process-Based Supervision”
Steven Byrnes
2y
1
48You can still fetch the coffee today if you're dead tomorrow
davidad
3y
14
54Seeking Power is Often Convergently Instrumental in MDPs
TurnTrout, Logan Riggs
6y
34
40The Credit Assignment Problem
abramdemski
6y
32
23Interpretability’s Alignment-Solving Potential: Analysis of 7 Scenarios
Evan R. Murphy
3y
0
17Non-myopia stories
[anonymous]2y
5
18Acceptability Verification: A Research Agenda
David Udell, evhub
3y
0
24Steering Behaviour: Testing for (Non-)Myopia in Language Models
Evan R. Murphy, Megan Kinniment
3y
5
Load More (15/34)
Add Posts