AI ALIGNMENT FORUM
AF

Agent FoundationsInfra-BayesianismAI
Frontpage

15

An Introduction to Credal Sets and Infra-Bayes Learnability

by Brittany Gelb
22nd Aug 2025
15 min read
2

15

Agent FoundationsInfra-BayesianismAI
Frontpage
New Comment
Moderation Log
More from Brittany Gelb
View more
Curated and popular this week
0Comments

Introduction

Credal sets, a special case of infradistributions[1] in infra-Bayesianism and classical objects in imprecise probability theory, provide a means of describing uncertainty without assigning exact probabilities to events as in Bayesianism. This is significant because as argued in the introduction to this sequence, Bayesianism is inadequate as a framework for AI alignment research. We will focus on credal sets rather than general infradistributions for simplicity of the exposition.

Defining Credal Sets

Recall that the total-variation metric is one example of a metric on ΔX, the set of probability distributions over a finite set X. A set is closed with respect to a metric if it contains all of its limit points with respect to the metric. For example, let X0={0,1}. The set of probability distributions over X0 is given by

ΔX0={P:P(0)=a,P(1)=1−a,a∈[0,1]}.

There is a bijection between ΔX0 and the closed interval [0,1], which is the map that sends a distribution to the probability of zero. More generally, recall from the proof of Lemma 1 in the preceding post, that if X is a finite set with n elements, then there is a bijection between ΔX and the closed (n−1)−simplex 

Δn−1:={(x1,x2,…,xn)∈Rn:n∑i=1xi=1,xi≥0 for all 1≤i≤n}.

Consider the following subset of ΔX0:

S(0,1):={P:P(0)=a,P(1)=1−a,a∈(0,1)}⊂ΔX0.

The set S(0,1) is an open subset of ΔX0 in the same way that the open interval (0,1) is an open subset of the closed interval [0,1]. (See Figure 1.) In particular, S(0,1) does not contain the distributions P0 and P1 defined by P0(0)=0 and P1(0)=1, which are limit points of S. For example, given ϵ>0, the distribution Pϵ defined by Pϵ(0)=ϵ is an element S, and the total variation distance between P0 and Pϵ is dTV(P0,Pϵ)=maxx∈{0,1}{|P0(x)−Pϵ(x)|}=ϵ. 

In contrast,

S[0,1/4]:={P:P(0)=a,P(1)=1−a,a∈[0,1/4]}⊂ΔX0,S[3/4,1]:={P:P(0)=a,P(1)=1−a,a∈[3/4,1]}⊂ΔX0, andS[0,1/4]∪S[3/4,1]

are all examples of closed subsets of ΔX0.

We now consider the meaning of convexity in this context.

Definition: Convex set of probability distributions

A convex combination of probability distributions Pi is a linear combination ∑ni=1αiPi such that αi∈R,  αi≥0, and ∑ni=1αi=1. A set of probability distributions is convex if it closed under convex combinations.

If a distribution P is written as a convex combination of a set of distributions, then it is called a mixture distribution. Note that the definition of a convex combination ensures that mixture distributions are indeed probability distributions. Sampling from the mixture distribution ∑ni=1αiPi can be thought of as first determining an index i by sampling from the distribution τ over {1,...,n} defined by τ(i)=αi and then sampling from the corresponding Pi.

As defined above, S[0,1/4] and S[3/4,1] are convex sets, analogous to how [0,1/4] and [3/4,1] are convex subsets of R. However, S[0,1/4]∪S[3/4,1] is not a convex set. To see this, let P1/4(0)=1/4 and P3/4(0)=3/4. Note that P1/4 is an element of S[0,1/4] and P3/4 is an element of S[3/4,1]. The distribution defined by P1/2(0)=1/2 can be written as the mixture distribution P1/2=12P1/4+12P3/4. However, as shown in Figure 1, P1/2 is not an element of  S[0,1/4]∪S[3/4,1], so S[0,1/4]∪S[3/4,1] is not closed under convex combinations. 

Figure 1: (Left) The set S(0,1)⊂ΔX visualized as a subset of [0,1], which is not a closed set of distributions as it does not contain the limit points corresponding to 0 and 1. (Right) The sets S[0,1/4] and S[3/4,1], which are closed and convex, and thus credal sets. The union of S[0,1/4] and S[3/4,1] is not convex. 

If a set of probability distributions is both closed and convex, it is called a credal set. Among the examples we have seen, S[0,1/4], S[3/4,1], and ΔX are credal sets. See Figure 2 for another example. Since S[0,1/4]∪S[3/4,1] fails to satisfy convexity, it is not a credal set. 

Figure 2: A credal set visualized as a subset of the 2-dimensional simplex in R3.

Definition: Credal set

Let X be a set, and let ΔX denote the set of probability distributions over X. A credal set over X is a closed, convex subset of ΔX, where closedness is defined with respect to some metric on ΔX. The set of credal sets over X is denoted by □X.

Credal sets are a natural tool for describing situations in which uncertainty cannot be reduced to probabilities. These situations are said to have Knightian uncertainty[2]. Knightian uncertainty is notably distinct from situations in which probabilities can be assigned with low confidence. 

For example, in Risk, Ambiguity, and the Savage Axioms (1961), Daniel Ellsberg writes of the "Knightian urn," an urn with 100 balls in which an observer is ignorant of the proportion of black and red balls. We would say that an observer of this urn has Knightian uncertainty with respect to the events of drawing either a red or black ball. Corresponding red to 0 and black to 1, this Knightian uncertainty can be captured by the credal set ΔX0 as above. On the other hand, an observer may gain partial information while remaining ignorant about the remaining possibilities, such as learning that there are no more than 25 red balls. The credal set S[0,1/4] captures this partial information. In this case, we would say that the observer then has Knightian uncertainty over the elements of this credal set.

Convexity is a natural condition in the presence of Knightian uncertainty. If an observer considers some distributions as all possible and furthermore does not hold any beliefs about which distributions are more likely than others, then they should also consider any mixture distribution as possible. When the possible set of distributions is finite, this is equivalent to taking the convex hull of a finite set of points, which yields a convex polytope. 

Furthermore, the decision rule discussed below makes the same recommendation for a set of probability distributions and its closed convex hull. Therefore, we can always without loss of generality assume that a set of distributions is a credal set.

Infrakernels

In machine learning classification, neural networks learn functions from sets (e.g. sets representing images) to probability distributions over possible classes; this kind of function of type X→ΔY for sets X and Y is called a probability kernel. In this case, X is called the source of the kernel and Y is called the target of the kernel.

An infrakernel[3] is the infra-Bayesian analogue of a probability kernel. A special case of infrakernels in which the target is a set of credal sets is called a crisp infrakernel.

Definition: Crisp infrakernel
Let X be a topological space, and let □Y denote the set of credal sets over a set Y. A crisp infrakernel is a continuous function of type X→□Y.

For example, every continuous probability kernel is a crisp infrakernel since a single probability distribution is a credal set with one element.

An example of a crisp infrakernel that is not a probability kernel is shown in Figure 3. Let X={William,Carl,Amalie}, and let Y={0,1,2}, with the elements of Y referring to numbers of cats. We can define κ:X→□Y as follows: κ(William) is the set containing only the distribution that is one at two, κ(Carl) is the set of distributions that assign zero to two, and κ(Amalie) is the set of all distributions over Y. This infrakernel represents certainty that William has two cats and Carl does not have two cats, and complete uncertainty about how many cats Amalie has. Since X is finite, κ is automatically continuous.

Figure 3: An example of a crisp infrakernel that sends people to credal sets visualized respectively as a single point, the entire 2-dimensional simplex, and a line segment.

Deterministic Versus Stochastic Policies

In the preceding post, we discussed stochastic policies, maps of the form π:(A×O)∗→ΔA. Deterministic policies are a special case of stochastic policies, and can be written as partial functions of the form π:(A×O)∗⇀A. By partial function, we mean that the domain of the policy π is restricted to those histories consistent with it. We previously observed that under some mild assumptions in the classical setting, an optimal policy exists. Furthermore, the optimal policy may be chosen to be deterministic. As we argue below, deterministic policies can be represented by stochastic policies and thus from this point forward, we will assume that policies are deterministic.

To emulate a stochastic policy using a deterministic policy, an agent can choose an element of ΔA and use a random number generator (RNG) to sample it. The output of the RNG then dictates an action. An observation from the environment then follows as usual. To formalize this, if A and O are the sets of actions and observations respectively from the stochastic setting, we can define a new set of observations equal to A×O. This new set records both the action dictated by the RNG and the usual observation that follows. The new action space would be ΔA, with each element of ΔA corresponding to an RNG that encodes the distribution.[4]

Importantly, this allows the random number generator to be insecure, biased, or influenced by the environment. In comparison, assuming that the set of policies is stochastic implicitly invokes an assumption that the agent can access a source of true randomness that is inaccessible by the environment. The deterministic theory not only encompasses the stochastic theory, but it is also applicable to embedded agency and can be used to solve Newcombian problems.

Topologies on policies and destinies

In this section, we review several important topologies. These topologies are relevant to how we define credal sets over the set of destinies (A×O)ω, as we require a notion of closure in the abstract space that is Δ(A×O)ω.

The topology on the set of deterministic policies

Recall that the set of deterministic policies Π is a set of partial functions π:(A×O)∗⇀A. The set of deterministic policies Π can be written as a subset of the product space ∏i∈(A×O)∗Ai where Ai=A for all i∈(A×O)∗. 

We assume that A is finite, and thus the natural topology on A is the discrete topology. Then Π is naturally endowed with the (subspace topology of the) product topology that comes from the topology on A. This means that a basis element of the topology on Π has the form ∏i∈(A×O)∗Ui where Ui⊆A for all i∈(A×O)∗ and Ui⊂A for only finitely many i∈(A×O)∗. An open set is then an arbitrary union of basis elements.

Recall from Lemma 2 of the preceding post that Π is compact.

The topology on the set of destinies

The set of countably infinite histories (A×O)ω, referred to as destinies, is a metric space under the metric d(h,h′)=γt(h,h′) where γ∈(0,1) and t(h,h′) is the time (index) of first difference between h and h′. In the proof section, we show that (A×O)ω is compact under this metric. 

Lemma 1: Suppose A and O are finite. Then the set of destinies (A×O)ω is a compact metric space.

The topology on the set of credal sets over destinies

Recall that the total-variation distance between probability distributions P1 and P2 is given by dTV(P1,P2):=12∑x∈X|P1(x)−P2(x)|. A natural way to extend any metric on probability distributions to credal sets is to use the Hausdorff metric. The Hausdorff metric is a metric that can be defined on the closed, non-empty subsets of a metric space. When restricted to one-element sets, the Hausdorff metric is equivalent to the metric of the original space.

Definition: Hausdorff total-variation distance between credal sets
Let Θ1 and Θ2∈□X be two non-empty credal sets. For a credal set Θ∈□X and distribution θ∈ΔX, let d(θ,Θ):=infθ′∈ΘdTV(θ,θ′). The Hausdorff distance between Θ1 and Θ2 is given by 

dH(Θ1,Θ2):=max{supθ1∈Θ1d(θ1,Θ2),supθ2∈Θ2d(θ2,Θ1)}.

Because credal sets are closed by definition, the Hausdorff total variation distance between non-empty credal sets is a well-defined metric.

The topology of □(A×O)ω is then given by the metric topology induced by the Hausdorff distance between credal sets.

Crisp Causal Laws

Causal laws[5] are a special type of infrakernel used to analyze sets of environments.

Given a set of environments E and a policy π, we can consider the set of distributions over destinies that arise from the interaction of each environment and the policy, which we denote by {μπ:μ∈E} (for details see the preceding post).

Wrapping this into a function results in an infrakernel as follows.

Definition: The infrakernel generated from a set of environments
A set of environments E generates the infrakernel ΛE(π):Π→□(A×O)ω defined by ΛE(π)=¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯ch({μπ:μ∈E}), where ch denotes convex hull and ¯⋅ denotes closure.

This leads to the definition of crisp causal laws, which take the role of hypotheses in infra-Bayesianism, as opposed to environments in classical reinforcement learning theory.

Definition: Crisp causal law
A crisp causal law is an infrakernel Λ:Π→□(A×O)ω generated by a set of environments.

For example, let A={a0,a1} and O={o0,o1} denote a set of actions and a set of observations respectively. Let π be a policy such that π(ϵ)=a0 where ϵ denotes the empty history. Let μ1 denote an environment such that μ1(o0|a0)=112. Let μ2 denote a second environment such that μ2(o0|a0)=1. A calculation as in Figure 2 of the preceding post shows that μπ1 is given by (112,1112,0,0), with each index of the tuple corresponding to one of the four possible histories of length one. Furthermore, μπ2 is given by (1,0,0,0).

By disregarding the last two irrelevant coordinates, we can visualize μπ1 and μπ2 as points in the 1-dimensional simplex. Let E be the convex closure of {μ1,μ2}. Then ΛE(π) corresponds to a line segment on the 1-dimensional simplex as shown in Figure 4. 

Figure 4: The image of a crisp causal law, which is a credal set.

The next proposition states that crisp causal laws are continuous with respect to the topologies discussed in the previous section. Since we are working with topological spaces, a function σ:T1→T2 between topological spaces T1 and T2 is continuous if for every open set U in T2, f−1(U) is open in T1. It is well known that this definition is equivalent to the epsilon-delta definition of continuity when T1 and T2 are metric spaces. 

Proposition 1: Every crisp causal law Λ:Π→□(A×O)ω is continuous with respect to the product topology on Π and the Hausdorff topology on □(A×O)ω.

The Minimax Decision Rule

In the classical setting, given a set of policies and a prior over a class of environments, a natural rule for choosing a policy is expected loss minimization in which a policy is chosen to minimize expected loss with respect to the prior. Recall that crisp causal laws take the place of environments as hypotheses in infra-Bayesianism. Even if a prior over hypotheses (crisp causal laws) is given, it is impossible to carry out expected loss minimization because there is Knightian uncertainty over the distributions that make up a credal set.

Because expected loss minimization is impossible, it is necessary to use a different decision rule. In the context of AI alignment, it is of interest to have guarantees on agent behavior that hold even in the worst-case scenarios. Given a set of environments, worst-case reasoning describes a decision-making heuristic in which an agent assumes that the true environment is the environment that would be worst with respect to some measure. The minimax decision rule stipulates that an agent should choose a policy so that the maximum expected loss over the set of possible environments is minimized.[6] Other decision rules can be reduced to minimax[7], and furthermore, there are natural classes of hypotheses that are learnable with respect to minimax, a notion that we define below.[8] 

Worst-case reasoning is implicitly built into the following definition.

Definition: Expected loss of a policy with respect to a crisp causal law
Let L:(A×O)ω→[0,1] denote a continuous loss function. The expected loss of a policy π∈Π with respect to a crisp causal law Λ is defined by 

EΛ(π)[L]:=maxθ∈Λ(π)Eθ[L].

​Using the assumption that L is continuous, Lemma 1 and Lemma 4 of the proof section imply that EΛ(π)[L] is well defined.

Choosing a policy that minimizes expected loss with respect to a crisp causal law is equivalent to choosing a policy that minimizes the maximum possible loss over the distributions obtained from the interaction of the policy and each environment. Therefore, the infra-Bayesian version of expected loss minimization is exactly a minimax decision.

Infra-Regret

Now that we have established the definition of expected loss in this setting, we can define the infra-Bayesian analogue to the regret of a policy with respect to an environment.

Definition: Infra-regret of a policy with respect to a crisp causal law
Let L:(A×O)ω→[0,1] denote a continuous loss function. The infra-regret of a policy π∈Π with respect to a crisp causal law Λ is defined by 

Reg(π,Λ,L):=EΛ(π)[L]−minπ′∈ΠEΛ(π′)[L].

It is natural to ask whether this notion of regret is well-defined. A proof can be given using Lemma 2 of the preceding post, which states that Π is compact, and the following proposition.

Proposition 2[9]: Let π∈Π be a policy, Λ a crisp causal law, and L:(A×O)ω→[0,1] a continuous function. Then the map π↦EΛ(π)[L] is continuous.

In the classical theory, a prior over environments quantifies uncertainty about the true environment. The analogue in infra-Bayesianism is a prior over crisp causal laws. The expected regret with respect to a prior of this type is called the infra-regret.

Definition: Infra-regret of a policy with respect to a prior
Let ζ denote a prior over a set of crisp causal laws. The infra-regret of a policy π with respect to ζ is defined by 

IBReg(π,ζ,L):=EΛ∼ζ[Reg(π,Λ,L)].

Infra-Bayes Optimality and Learnability

In this section, we introduce some elementary results on learnability in infra-Bayesianism. The analogue to a Bayes optimal policy with respect to a prior over environments is an infra-Bayes optimal policy, defined as follows.

Definition: Infra-Bayes optimal policy
An infra-Bayes optimal policy with respect to a prior ζ over crisp causal laws is a policy 

π∗∈argminπ∈ΠEΛ∼ζ[EΛ(π)[L]].

A family of policies {πγ}γ∈[0,1) is said to be infra-Bayes optimal if for all γ∈[0,1), πγ is infra-Bayes optimal for the γ-dependent loss function Lγ.

A natural question is: Does an infra-Bayes optimal policy always exist? The answer is yes; this is a corollary of Proposition 2 and the compactness of Π. A standard continuity argument can also be used to show that the set of infra-Bayes optimal policies is closed, as stated in the following corollary.[10]

Corollary 1: Existence of the infra-Bayes optimal policy
Let π∈Π denote a policy, Λ a crisp causal law, and L:(A×O)ω→[0,1] a continuous function. Then argminπ∈ΠEΛ∼ζ[EΛ(π)[L]] is non-empty and closed.

The analogue of a learnable class of environments is a learnable class of causal laws.

Definition: Non-anytime learnable class of infra-Bayesian hypotheses
Let I denote an indexing set. A class of crisp causal laws (hypotheses) {Λi}i∈I is non-anytime learnable if there exists a set of policies {πγ}γ∈[0,1) such that for all i∈I, limγ→1Reg(πγ,Λi,Lγ)=0. In this case, we say that {πγ}γ∈[0,1) learns {Λi}i∈I.

Furthermore, we have the analogue of a non-anytime learnable prior over environments.

Definition: Non-anytime learnable prior over crisp causal laws
A prior ζ over a set of causal laws is non-anytime learnable if there exists a set of policies {πγ}γ∈[0,1) such that 

limγ→1(EΛ∼ζ[EΛ(πγ)[Lγ]−minπ∈ΠEΛ(π)[Lγ]])=0.

In this case, we say that {πγ}γ∈[0,1) learns ζ.

Proposition 2 of the preceding post states the classical result from learning theory that if a countable class of environments is learnable, given a non-dogmatic prior over the class, then any family of policies that is Bayes-optimal with respect to the prior must also learn the class. We end with the following corresponding result for crisp causal laws. The proof is contained in the proof section.

Proposition 3: For any non-dogmatic prior ζ over a learnable and countable collection of crisp causal laws {Λi}∞i=0,  if a family {π∗γ}γ∈[0,1) of policies is infra-Bayes optimal with respect to ζ, then {π∗γ}γ∈[0,1) learns {Λi}∞i=0.

Acknowledgements

I'm grateful to Vanessa Kosoy and Alexander Appel for insightful discussions, and Marcus Ogren and Mateusz Bagiński for their valuable feedback on the initial draft.

  1. ^

    In the original infra-Bayesianism sequence, credal sets are called crisp infradistributions.

  2. ^

    First called "unmeasurable uncertainty" by economist Frank Knight in his 1921 book Risk, Uncertainty, and Profit. 

  3. ^

    As defined in The Many Faces of Infra-Belief.

  4. ^

    This requires dealing with infinite action spaces, which is typically not a problem as long as the action space is compact Polish.

  5. ^

    Previously called belief functions, e.g. as in The Many Faces of Infra-Belief.

  6. ^

    If using reward rather than loss, this rule can be equivalently formulated as a maximin rule (c.f. Infra-Bayesianism Distillation: Realizability and Decision Theory).

  7. ^

    See Vanessa Kosoy's Shortform. 

  8. ^

    See Vanessa Kosoy's Shortform.

  9. ^

    This proposition is a special case of Proposition 5 from Belief Functions And Decision Theory.

  10. ^

    These two statements were written as Proposition 12 in Belief Functions And Decision Theory.

Mentioned in
15What is Inadequate about Bayesianism for AI Alignment: Motivating Infra-Bayesianism
5Proof Section to an Introduction to Credal Sets and Infra-Bayes Learnability