The paper Optimal Policies Tend to Seek Power tries to justify the claim in the title with a mathematical argument, using a clever formal definition of "power". I like the general approach, but I think that parts of the formalisation do not correspond well to the intuitive understanding of the concepts involved. This wouldn't be an issue if the goal of the paper was to prove a mathematical theorem. Here, however, the goal is to be able to deduce something about the behaviour of future AI systems. Therefore, the interface between "math" and "the world" is much more important. This post attempts to take another route to formalisation, which I think is closer to the intuition, albeit more difficult to work with mathematically.


The power of a state , written , is defined as (roughly) the average optimal value of that state, over some chosen distribution of rewards . The paper quite convincingly argues for it:

Power is the ability to achieve a range of goals. In an MDP, the optimal value function captures the agent’s ability to “achieve the goal” . Therefore, average optimal value captures the agent’s ability to achieve a range of goals.

The problem with this approach is that it's very sensitive to the choice of the distribution of rewards . In particular, one might take a Dirac delta on a reward giving 1 in any chosen state, and 0 everywhere else, which would trivially assign maximal power to that particular state. To combat this difficulty, authors introduce some technical devices allowing them to say "for most distributions, the power of (one, obviously better, state) is greater than the power of (another, obviously worse, state)".

To me, " happens for most distributions of rewards" should mean that there is some reasonable, uninformative, broad prior over distributions of rewards, and that . The approach chosen in the paper not only doesn't correspond to this intuitive understanding, but seems to me to be (in certain sense) proving the thesis by assuming the conclusion.

More concretely, the approach taken in the paper is as follows. Given a reward and a permutation over states of the MDP, one can obtain another reward by permuting the states' rewards. This also works if we have not one reward , but a distribution over rewards, by pushing it forward by . Therefore, each such permutation induces an orbit in the space of (distributions of) rewards. The authors then say that the power of one state is greater than power of another state for most distributions, if, for each permutation , for the majority of distributions of rewards in the orbit , we have the inequality .

This is arguably quite far from how anyone would interpret the phrase "for most distributions" in real life. What we care about is the global behaviour in the whole space, not the the local behaviour in orbits. (Moreover, the permutations considered here are arbitrary in a sense of not respecting the structure of the MDP - but that is another issue).

The main problem of making the probabilistic meaning of "most" proposed above work here is that it is mathematically difficult to work in the infinite-dimensional space of distributions over rewards. There are many pathologies, such as the fact that there is no counterpart to the (Lebesgue-) uniform measure (after scouring the internet, I found that authors make this argument in a response to the review of the paper).

I still think it is possible to make progress on this. It should be possible to use something like Dirichlet/Pitman-Yor/Gaussian processes, or even something more geometric such as Wasserstein distance, based on the metric structure of the rewards in an MDP. However, regardless of the concrete distribution, the first step is to prove that the event is measurable in the space of distributions of rewards (that is, in ).

Thus, this post contains:

  • the proof of this fact, and
  • some (unfruitful) exploration of two simple cases of concrete distributions over distributions of rewards

The main technical device I use is the Giry monad, which gives a characterisation of the canonical -algebra on the space of probability distributions. I include the basic definition below, but those unfamiliar with the topic might enjoy this tutorial - which assumes only basic familiarity with measure theory and category theory.

Technical content

What the original paper did

To refresh the memory, the original paper proceeds as follows. We assign "power" to each state of a MDP (under a fixed distribution over reward functions), to be: For a set of distributions over rewards, we define the inequality between powers of states to be: if and only if, for all , for the majority of (which is a finite set) we have:

(The paper considers only rewards of a form and only deterministic policies .)

Giry monad

First, we note that if and are finite (countable case is the same), then the space of reward functions (in general, ) is equal to for some - therefore, it is a standard Borel space.

On the category of Borel spaces, we have the Giry monad , which assigns, to each measurable space , the space of the probability measures on . Therefore, points out a canonical -algebra on the , which we denote by .

We wiil describe . First, we say that for a fixed , the evaluation function is given by: Then, It is generated by all , that is, by sets all sets of the form , for any measuable set .

Lemma 1: For any measurable function , we can take the evaluation with respect to this function: and then is equivalently generated by all .

Measurability of the inequality

Let us fix some states . We can ask whether the set is measurable in (since ultimately, we want to say something about it's measure).

Unrolling the definitions gives us Let's now denote: Rewriting the definition of gives us: Now, we only have to argue that is measurable - then, by Lemma 1, we will know that is measurable. It is a sum of four terms - it is enough to show that and are measurable.

The first function is just a projection to the -th coordinate, therefore measurable.

The second function is just a maximum over a finite set of linear functionals (on a finite-dimensional space), therefore measurable as well - where is the visit distribution (doesn't really matter what it is, the important point is that it is independent of ).

Probabilities on

Ok, so now we can think about what is the probability mass of under different probability distributions over . To make our lives easier, we'll now switch to , . This doesn't change anything: we can rescale reward function by any factor and get an equivalent MDP (proof: just look at the Bellman equations).

Two (short) attempts I make below are unsuccessful. Both of them fail in a similar ways: by making the uniform distribution over rewards "too central", and by not having the full support over all probability distributions.

Lifting uniform distribution by Giry monad

We have one canonical distribution over distributions: take the uniform distribution : and apply the Giry monad: But unfortunately, is just the uniform distribution over the Dirac deltas in . Then, the situation is equivalent to calculating the measure of the set in one case of the uniform distribution over rewards (that is, a Diract delta on the uniform distribution of the rewards).

Dirichlet process

We can take Dirichlet process centered around the uniform distribution, for a fixed .

Definition: For a given distribution on and , the Dirichlet process on the space is characterized by the following property: for any partition of , we have that if , then:

Applying the definition to our and the partition , we get

where the measure of under uniform distribution .

Therefore, we have the mean:

Unfortunately, this is again quite uninteresting. All the choice was in centering the Dirichlet process at . (Another drawback is that Dirichlet process has support only on discrete distributions, which is at odds with our initial goal of having a broad prior over all distributions of rewards - in particular, continuous ones).


We proved that the event is measurable in the space of distibutions . We failed to point out any interesting concrete priors over distributions of reward functions, but I am optimistic that it should be possible with better understanding of the metric (topological, linear, differential, ...?) structure of the reward functions on an MDP.

New Comment