Impact Regularization

Victoria Krakovna (+41/-31)
Alex Turner (+22) Softened epistemic status to not exaggerate intuitive implications of power-seeking theorems
Alex Turner (+95/-32)
Alex Turner (+7/-7)
Alex Turner (+57)
Ruben Bloom (+19/-17)
Ruben Bloom
Ruben Bloom (+10/-13)
Ruben Bloom (-1)
Alex Turner (+1507/-20)

Impact measuresregularizers penalize an AI for affecting us too much. To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. You reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side.

How do you rigorously define "low impact" in a way that a computer can understand – how do you measure impact? These questions are important for both prosaic and future AI systems: objective specification is hard; we don't want AI systems to rampantly disrupt their environment. In the limit of goal-directed intelligence, theorems suggest that seeking power tends to be optimal; we don't want highly capable AI systems to permanently wrench control of the future from us. 

Currently, impact measurementregularization research focuses on two approaches:

Sequences on impact measurement:regularization:

Impact measures penalize an AI for affecting us too much. To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. You reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side.

How do you rigorously define "low impact" in a way that a computer can understand – how do you measure impact? These questions are important for both prosaic and future AI systems: objective specification is hard; we don't want AI systems to rampantly disrupt their environment. In the limit of goal-directed intelligence, theorems suggest that seeking power tends to be optimal; we don't want highly capable AI systems to permanently wrench control of the future from us. 

Impact measures penalize an AI for affecting us too much. To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. You reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side.

How do you rigorously define "low impact" in a way that a computer can understand? Howunderstand – how do you measure impact? These questions are important for both prosaic and future AI systems: objective specification is hard, and; we don't want AI systems to rampantly disrupt their environment. Furthermore,In the limit of goal-directed intelligence, seeking power tends to be optimal; we don't want highly capable AI systems to permanently wrench control of the future from us. 

Impact measures penalize an AI for affecting us too much. To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. You reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side.

How do you rigorously define "low impact" in a way that a computer can understand? How do you measure impact?impact? These questions are important for both prosaic and future AI systems: objective specification is hard, and we don't want AI systems to rampantly disrupt their environment. Furthermore, we don't want highly capable AI systems to permanently wrench control of the future from us. 

Impact measures penalize an AI for affecting us too much.To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. You reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side.

How do you rigorously define "low impact" in a way that a computer can understand? How do you measure impact? These questions are important for both prosaic and future AI systems: objective specification is hard, and we don't want AI systems to rampantly disrupt their environment. Furthermore, we don't want highly capable AI systems to permanently wrench control of the future from us. 

To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. You reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side.

How do you rigorously define low impact"low impact" in a way that a computer can understand? How do you measure impact?impact? These questions are important for both prosaic and future AI systems: objective specification is hard, and we don't want AI systems to rampantly disrupt their environment. Furthermore, we don't want highly capable AI systems to permanently wrench control of the future from us. 

To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. You reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side. 


How do you rigorously define "low impact"low impact in a way that a computer can understand? How do you measure impact? These questions are important for both prosaic and future AI systems: objective specification is hard, and we don't want AI systems to rampantly disrupt their environment. Furthermore, we don't want highly capable AI systems to permanently wrench control of the future from us. 

  • Relative reachability: the AI preserves its ability to reach many kinds of world-states. The hope is that by staying able to reach many goal states, the AI stays able to reach the correct goal state.
  • Attainable utility preservation: the AI preserves its ability to achieve one or more auxiliary goals. The hope is that by penalizing gaining or losing control over the future, the AI doesn't take control away from us. 

To reduce the risk posed by a powerful AI, you might want to make it try accomplish its goals with as little impact on the world as possible. But howYou reward the AI for crossing a room; to maximize time-discounted total reward, the optimal policy makes a huge mess as it sprints to the other side. 

How do you rigorously define "low impact""low impact" in a way that a computer can understand? How do you measure impact? These questions are important for both prosaic and future AI systems: objective specification ishard, and we don't want AI systems to rampantly disrupt their environment. Furthermore, we don't want highly capable AI systems to permanently wrench control of the future from us. 

Currently, impact? measurement research focuses on two approaches:

  • Relative reachability: the AI preserves its ability to reach many kinds of world-states. The hope is that by staying able to reach many goal states, the AI stays able to reach the correct goal state.
  • Attainableutilitypreservation: the AI preserves its ability to achieve one or more auxiliary goals. The hope is that by penalizing gaining or losing control over the future, the AI doesn't take control away from us. 

For a review of earlier work, see A Survey of Early Impact Measures

Sequences on impact measurement:

  • Reframing Impact: we're impacted when we become more or less able to achieve our goals. Seemingly, goal-directed AI systems are only incentivized to catastrophically impact us in order to gain power to achieve their own goals. To avoid catastrophic impact, what if we penalize the AI for gaining power?
  • Subagents and Impact Measures explores how subagents can circumvent current impact measure formalizations.

Related tags: Instrumental Convergence, Corrigibility, Mild Optimization.

Load More (10/12)