This updated version of my AGI 2025 paper with Marcus Hutter, "Value under ignorance in universal artificial intelligence," studies general utility functions for AIXI. Surprisingly, the (hyper)computability properties have connections to imprecise probability theory!
AIXI uses a defective Bayesian mixture called a semimeasure, which is often viewed as expressing a chance that the agent dies. I do not think that interpretation has been sufficiently justified. Recasting semimeasures as credal sets, we can recover the (recursive) value function from reinforcement learning for discounted nonnegative rewards. We can also obtain a wider class of lower semicomputable value functions, with optimal agents following a max min decision rule.
This is an early conference paper without complete proofs included. You should read it if:
Of course, this paper is largely motivated by AI safety (and my work was supported by the LTFF). However, any safety application would come at a much later stage, and I want to avoid the impression of making certain claims. This paper is one line of evidence from AIT potentially justifying the naturality of pessimism in the face of ignorance,[1] but the implications (for example, to rationality) need further study. Also, while I am hoping to place some constraints on reasonable utility/value functions for AIXI variants, I do not imagine it is easy to write down a safe but useful one (and this paper does not grapple with the problem directly).
(An earlier and much rougher version of these ideas appeared here)