The aim of this post is to motivate the idea of representing probability spaces as states on a commutative algebra. We will consider how this abstract construction relates directly to classical probabilities.

In the standard axiomatization of probability theory, due to Kolmogorov, the central construct is a *probability space* . This consists of a state space , an event space , which is a sigma-algebra of subsets of , and a probability measure . The measure is defined as a map satisfying countable additivity and normalised as .

A measure space allows us to define integrals of real-valued measurable functions or, in the language of probability, expectations of random variables. We construct the set of all bounded measurable functions . This is a real vector space and, as it is closed under multiplication, is an algebra. Expectation, by definition, is the unique linear map , satisfying for and monotone convergence: if is a nonnegative sequence increasing to a bounded limit , then tends to .

In the opposite direction, any nonnegative linear map satisfying monotone convergence and defines a probability measure by . This is the unique measure with respect to which expectation agrees with the linear map, . So, probability measures are in one-to-one correspondence with such linear maps, and they can be viewed as one and the same thing. The Kolmogorov definition of a probability space can be thought of as representing the expectation on the subset of consisting of indicator functions . In practice, it is often more convenient to start with a different subset of . For example, probability measures on can be defined via their Laplace transform, , which represents the expectation on exponential functions . Generalising to complex-valued random variables, probability measures on are often represented by their characteristic function , which is just the expectation of the complex exponentials . In fact, by the monotone class theorem, we can uniquely represent probability measures on by the expectations on any subset which is closed under taking products and generates the sigma-algebra .

A simple corollary of the monotone class theorem states that there is a one-to-one correspondence between sigma-algebras on a set and algebras of bounded functions closed under monotone convergence, with the correspondence given by .

On the other hand, in quantum mechanics, we start with a Hilbert space , and observables are represented as self-adjoint operators. Restricting our consideration to bounded observables, these generate a subalgebra of , the space of bounded linear maps on . A *pure state* is represented by an element normalised so that , and the expectation of an observable is . This is a nonnegative linear map from a sub-algebra of to .

All of this suggests that it would be useful to consider an alternative approach to probability. Instead of a measurable space , we have an algebra . Instead of a probability measure , we have a positive linear map from to or . The underlying state space is not required at all — it is a *pointless* approach to probability, as we no longer include the points in the representation of the probability space. As multiplication of real (and complex) numbers is commutative, , algebras of the form are commutative. Hence, classical probability spaces will correspond to commutative algebras, with the generalisation to non-commutative algebras incorporating quantum probability.

As this post is primarily intended to motivate the algebraic approach to probability, rather than go into technical details, I will not give proofs of all theorems quoted here and, instead, will refer to the literature. We start with the definition of an algebra.

Definition 1Let be a field. Then, a -algebra, or algebra over , is a -vector space equipped with a binary product, , and identity element satisfying the following for all .

- Associativity: .
- Compatability with scalars: for all .
- Left-distributivity: .
- Right-distributivity:
- Identity: .

If, furthermore, for all then the algebra is said to be commutative.

Strictly speaking, this defines a *unitial associative algebra*. Sometimes, the axiom of associativity is dropped, although I do not look at such non-associative algebras here. Similarly, the existence of the identity is sometimes dropped along with its corresponding axiom. In this post, whenever the unqualified term `algebra’ is used, then it refers to a structure satisfying definition 1, so is unitial. Also, I will use the symbol to denote the identity element. This creates some ambiguity as to whether an expression of the form refers to multiplication by the identity element or by the scalar . However, as they both evaluate to , it should not cause any confusion.

A subset is called *commutative* if for all . In particular, the algebra itself is commutative if and only if is commutative as a set of elements. It is also easy to show that the sub-algebra of generated by a commutative set (i.e., the smallest subalgebra containing ) is itself commutative. Note that this means that the subalgebra generated by a single element is commutative.

Examples of algebras abound in mathematics. A small set of examples is:

- Polynomial rings are commutative -algebras.
- For a set , the collection of functions is a commutative -algebra, where the operations of addition, scalar multiplication, and multiplication are defined point-wise.
- For a measurable space , the collection of -measurable functions is an -algebra.
- For a normed real vector space , the collection of bounded linear maps is an -algebra.

We define the notion of a *state* on a commutative real algebra.

Definition 2Let be a commutative real algebra. Then, a linear map is

positiveif for all .- a
stateif it is positive and .

#### Correspondence with classical probabilities

As discussed above, a classical probability space determines a commutative real algebra, consisting of the bounded random variables, and a state on this algebra given by expectation. The question is, can this process be inverted? When can a state on a commutative real algebra be represented as an expectation on a set of random variables on some probability space? We start by considering a single element . This defines a map taking any polynomial to its evaluation , and the image is the sub-algebra generated by . For to be considered as a random variable on a probability space, then its distribution is a probability measure on satisfying

(1) |

for all polynomials . By linearity, (1) holds whenever it holds on monomials . That is, we require

for all positive integers and, for this to make sense, must have finite moments. This is the classical moment problem, to construct a probability measure from its moments. In the one factor case, it is known that the positivity of ensures existence of a solution.

Theorem 3Let be a state on commutative real algebra . Then, for any , there exists a probability measure on satisfying (1).

The existence of a measure with specified moments is known as the Hamburger problem. Unfortunately, uniqueness need not hold, as there do exist distinct probability measures on with the same moments. As an example, consider the log-normal distribution on the nonnegative reals, and a perturbation of this,

These measures have all the same moments,

and therefore generate the same state on the algebra . On the other hand, it is not difficult to show that the distribution of a bounded random variable is uniquely determined by its moments. This follows from the Stone–Weierstrass theorem, which states that the polynomials are dense in the space of continuous functions on any closed bounded interval. Furthermore, the distribution will be supported by an interval if for all positive . It is possible to relax this condition to bounds on the growth of the moments, such as Carleman’s condition (2).

Theorem 4Let be a state on commutative real algebra . If satisfies

(2)

then there exists auniqueprobability measure on satisfying (1).

This result goes back to T. Carleman, *Les fonctions quasi analytiques*, Gauthier–Villars, Paris, 1926. A proof of this result, and also of the Hamburger problem, is given in the lecture notes, The classical moment problem, by Sasha Sodin, 2019. See Theorem 3.1 and Corollary 2.12.

Moving to the mutifactor situation, where we have a sequence , the aim is to find a probability measure on satisfying

(3) |

for all polynomials and, as in the single factor case, must have finite moments for this to make sense. Unlike the single factor case, this is not always possible, so theorem 3 does not generalise to . The reason for this is that there exists multifactor polynomials which are positive on all of , yet cannot be expressed as a sum of squares. Consider

The AM-GM inequality shows that everywhere on . However, it is not possible to express as for a finite sequence of polynomials . This means that the definition of positivity for a state on is insufficient to ensure that , and the hyperplane separation theorem implies the existence of states with . No such state can arise from the expectation under a probability measure.

Fortunately, if sufficient bounds are imposed on the growth of the moments , then it is possible to show that a *unique* measure exists satisfying (3). Again, in the case that for all and some real , then the Stone–Weierstrass theorem can be used to show uniqueness of , which must be supported on , with the Riesz representation theorem providing existence. These conditions can be weakened considerably and, in fact, it is known that Carleman’s condition for each of the individual elements is sufficient to guarantee existence and uniqueness.

Theorem 5Let be a state on commutative real algebra . If each satisfy Carleman’s condition (2) then there is a unique probability measure on satisfying (3).

This result originates from Nussbaum, A. E., Quasi-analytic vectors, Arkiv för Matematik. 6 (1965), no. 2, 179–191.

Taking the idea a step further, we can consider infinite subsets of . Let be the space of functions and denote the coordinate map . Let be the sigma-algebra on generated by . That is, is the smallest sigma-algebra on with respect to which each is measurable. In particular, it is generated by the sets for Borel . The collection generates an algebra of random variables, which can be expressed as real polynomials in . Evaluating the polynomials at the values gives an algebra homomorphism . The aim is to find a probability measure on satisfying

(4) |

for all which, to make sense, requires to be integrable.

If we choose to be a generating set for , so that the smallest subalgebra of containing every is all of , then we obtain a representation of as an algebra of random variables on a probability space together with the expectation operator.

Theorem 6Let be a state on commutative real algebra . If each satisfy Carleman’s condition (2) then, there exists a unique probability measure on satisfying (4).

*Proof:* For each finite subset , theorem 5 uniquely determines a probability measure on satisfying

for all polynomials .

Define the projection map by for all and . For a probability measure on , the pushforward measure on is defined by . Condition (4) is then,

or, equivalently, . Existence and uniqueness of follows from Kolmogorov’s extension theorem. ⬜

This a great post, I can’t wait to read the continued notes. I have a question on the example ,we can indeed prove that using AM-GM inequality. But I didn’t not understand the sentence that cannot be expressed as , do you mean that each is a polynomial in one variable, otherwise we can write . Also, can you develop a little bit how do you use the hyperplane separation theorem to prove the existence of states with ? Thanks

I mean that each is in [I edited]. I might come back and add some detail on the use of the hyperplane separation theorem, but it is intended as a brief counterexample and was wanting to get some new posts up first. Alternatively, should be able to find reference giving more detail on this.

Thank you for your quick reply. I can wait , atfer all I really want to learn the quantum part note.