A random variable has the standard n-dimensional normal distribution if its components are independent normal with zero mean and unit variance. A well known fact of such distributions is that they are invariant under rotations, which has the following consequence. The distribution of is invariant under rotations of and, hence, is fully determined by the values of and . This is known as the noncentral chi-square distribution with n degrees of freedom and noncentrality parameter , and denoted by . The moment generating function can be computed,
which holds for all with real part bounded above by 1/2.
A consequence of this is that the norm of an n-dimensional Brownian motion B is Markov. More precisely, letting be its natural filtration, then has the following property. For times , conditional on , is distributed as . This is known as the `n-dimensional’ squared Bessel process, and denoted by .
Alternatively, the process X can be described by a stochastic differential equation (SDE). Applying integration by parts,
As the standard Brownian motions have quadratic variation , the final term on the right-hand-side is equal to . Also, the covarations are zero for from which it can be seen that
The standard existence and uniqueness results for stochastic differential equations do not apply here, since is not Lipschitz continuous. It is known that (3) does in fact have a unique solution, by the Yamada-Watanabe uniqueness theorem for 1-dimensional SDEs. However, I do not need and will not make use of this fact here. Actually, uniqueness in law follows from the explicit computation of the moment generating function in Theorem 4 below.
Although it is nonsensical to talk of an n-dimensional Brownian motion for non-integer n, Bessel processes can be extended to any real . This can be done either by specifying its distributions in terms of chi-square distributions or by the SDE (3). In this post I take the first approach, and then show that they are equivalent. Such processes appear in many situations in the theory of stochastic processes, and not just as the norm of Brownian motion. It also provides one of the relatively few interesting examples of stochastic differential equations whose distributions can be explicitly computed.
The distribution generalizes to all real , and can be defined as the unique distribution on with moment generating function given by equation (1). If and are independent, then has moment generating function and, therefore, has the distribution. That such distributions do indeed exist can be seen by constructing them. The distribution is a special case of the Gamma distribution and has probability density proportional to . If is a sequence of independent random variables with the standard normal distribution and T independently has the Poisson distribution of rate , then , which can be seen by computing its moment generating function. Adding an independent random variable Y to this produces the variable .
The definition of squared Bessel processes of any real dimension is as follows. We work with respect to a filtered probability space .
Definition 1 A process X is a squared Bessel process of dimension if it is continuous, adapted and, for any , conditional on , has the distribution.
Substituting in expression (1) for the moment generating function, this definition is equivalent to X being a continuous adapted process such that, for all times ,
This holds for all with nonnegative real part. Also, if the filtration is not specified, then a process X is a Bessel process if it satisfies Definition 1 with respect to its natural filtration .
Note that we have not yet shown that Bessel processes for arbitrary non-integer are well-defined. Definition 1 specifies the properties that such processes must satisfy, but this does not guarantee their existence. It is not difficult to show that (4) determines a Markov transition function, so that the Chapman-Kolmogorov identity is satisfied. In fact, as I show below, it is Feller. See Lemma 8 below, where the existence of continuous modifications is also proven.
For now, let us determine some of the properties of Bessel processes. There are some properties which can be stated directly from the definition. The fact that the sum of independent and distributed random variables has the distribution gives the following result for sums of Bessel processes.
Lemma 2 Suppose that X and Y are independent and processes respectively. Then, X+Y is a process.
Next, Definition 1 only referred to the ratio between the process X and the time increments. So, scaling the time axis and the process values by the same factor leaves the property unchanged.
Lemma 3 Let X be a process and be constant. Then, is also a process.
Taking the limit in expression (4) gives us the probability that is equal to 0.
So, a process has a positive probability of hitting 0 in any nontrivial time interval. Furthermore, since (5) gives , once it hits zero it remains there. That is, 0 is an absorbing boundary.
The case for is different. Equation (5) says that has zero probability of being equal to 0 at any given time. This does not mean that X cannot hit zero but, rather, that the total Lebesgue measure of its time spent there is zero,
In fact, as we will see, the process does hit zero for all values of n less than 2 so, for , 0 is a reflecting boundary.
We now show the equivalence of the definition of squared Bessel processes in terms of the skew chi-square distribution given in Definition 1 and in terms of the SDE (3). In particular, this demonstrates that (3) satisfies uniqueness in law, which we show by using Ito’s lemma to derive a partial differential equation for the moment generating function.
Theorem 4 For any nonnegative process X and real , the following are equivalent,
- X is a process.
- is a local martingale and .
- X satisfies the SDE
for a Brownian motion W (in the case n=0, it is necessary to assume the existence of at least one Brownian motion on the underlying filtration).
(1) implies (2): The moments of a random variable can be computed by expanding and and comparing powers of . From this, we see that Z has mean and variance . So, for the squared Bessel process, has mean and variance conditional on (). Therefore, is a local martingale. Note that it can fail to be a proper martingale, since is not required to be integrable. Also,
Comparing this with the following expression
(1) implies (3): By the argument above, for a continuous local martingale M with quadratic variation . One of the consequences of Lévy’s characterization is that, assuming that there is at least one Brownian motion defined on the underlying filtration,
for a Brownian motion W. It just needs to be shown that, if n is greater than zero, there does exist such a Brownian motion. Set
which is a local martingale. Its quadratic variation is
As we have already shown that X is nonzero almost everywhere, this gives and, again by Lévy’s characterization W is a Brownian motion.
(3) implies (2): First, is a local martingale. Then, using gives as required.
(2) implies (1): The idea is to derive a partial differential equation for the moment generating function of , and show that the solution is given by (4). Using the fact that has quadratic variation , Ito’s lemma gives
for constant and times . The final term is a local martingale and is bounded on finite time intervals (as all the other terms are). So, it is a proper martingale. Multiplying by a bounded -measurable random variable Z and taking expectations,
We now introduce the function and, noting that this has partial derivative ,
So, is continuously differentiable over and, by differentiating the above equation, it satisfies the following partial differential equation
This is a transport equation and can be simplified by replacing with a time dependent function satisfying ,
This is just an ordinary differential equation with the unique solution,
The ODE for is also easily solved, with the unique solution . Using this, the following expressions can be calculated,
It is natural to ask about the properties of Bessel process paths such as, does it ever hit zero? Also, what happens to in the limit as t goes to infinity? We show below, in Theorem 6, that for the process hits zero at arbitrarily large times and, for it never hits zero. Furthermore, for it tends to infinity. The idea is to transform the process into a local martingale , so that standard convergence results for continuous local martingales can be applied. The function f is called the scale function of X.
Lemma 5 Let X be a process satisfying the SDE (6) with and make the following substitution.
- If , let be the first time at which so that is the process stopped when it first hits zero, and set with
Then, Y satisfies the SDE
- if then satisfies
- if then with a, b, c as in (8) satisfies
Although the substitutions above for are not defined when X=0, X remains strictly positive in this case.
Proof: For any , let be the first time at which , so and the stopped process is strictly positive. As is a smooth function on , Ito’s lemma can be applied to on the interval ,
Here, the expression has been used. If then substituting expression (8) for b makes the last term on the right hand side equal to zero. So,
Using gives according to whether b is positive () or negative (). So,
on the interval . Letting n increase to infinity, we see that equations (9) and (11) are satisfied on the interval . If then , so is equal to zero over the interval , and (9) is satisfied. On the other hand, if then so Y explodes to infinity at time . However, is a nonnegative local martingale and cannot explode in a finite time. This is a consequence of Fatou’s lemma,
So, Y is almost surely bounded and .
Now, consider the case with and . As is smooth on , Ito’s lemma can be applied on the interval , in a similar way as above,
As above, this holds over the interval and it needs to be shown that almost surely. Letting be the first time at which for some constant K, Y is a local martingale bounded above over the interval . Applying Fatou’s lemma to –Y,
However diverges to at time , so for large enough n (almost surely). Letting K increase to infinity, goes to infinity, showing that is almost surely infinite, as required. ⬜
Using the transformation given by Lemma 5 together with the convergence of continuous local martingales, it is possible to say whether a process hits zero and to determine and , for each value of n.
Theorem 6 Let X be a process. With probability one,
- if then X hits 0 at some time, and remains there.
- if then X hits zero at arbitrarily large times and .
- if then X is strictly positive at all positive times and , .
- if then X is strictly positive at all positive times and as .
Recall that the convergence theorem for continuous local martingales states that the events on which a continuous local martingale Y converges to a finite limit, the event on which , and on which , are all identical (up to zero probability sets). This fact is used several times in the following proof.
Proof: The case with is easy. We have already shown that it hits zero with probability by time t conditional on and that, once it hits zero, it stays there. Letting t increase to infinity this converges to 1, so it hits zero almost surely at some positive time.
Let us show that for all . From the definition, has the distribution conditional on , which can be written as the sum of independent random variables and . So, for any constant ,
as , so .
For it was shown in Lemma 5 that X never hits 0 at any time conditional on . As it has zero probability of being equal to zero at a time , this shows that it never hits zero on and, letting t decrease to zero, it never hits zero at any positive time.
Now, consider . As shown in Lemma 5, is a nonnegative local martingale for some continuous and strictly increasing function satisfying , where is the process stopped when it first hits zero. As this trivially satisfies , martingale convergence implies that exists and is finite almost surely. However, as we have just shown, is infinite, so at large enough times. This can only happen if the process hits zero, after which Y is stopped at zero.
Alternatively, if , then is a local martingale for . As shown above, is infinite so, by martingale convergence, . Therefore, .
Finally, it only remains to show that for . Again, using Lemma 5, is a nonnegative local martingale for a strictly positive and continuous function f satisfying as . As above, martingale convergence implies that exists almost surely. However, since we have already shown that , it follows that . So, as , giving as required. ⬜
An immediate consequence of the preceding result is to n-dimensional standard Brownian motion. For positive integers n, the squared Bessel process was introduced above as the squared magnitude of an n-dimensional Brownian motion. In one and two dimensions, we see that Brownian motion is recurrent. That is, with probability one, it enters every nonempty set at arbitrarily large times. As has a countable base for the topology, it is enough to prove this for a countable collection of open sets and, by countable additivity of probability measures, it is equivalent to saying that X almost surely enters the open set U at arbitrarily large times, for each open individually.
In three or more dimensions, Brownian motion is not recurrent. In fact, it diverges to infinity with probability one.
Theorem 7 Let B be an n-dimensional Brownian motion. With probability one,
- if then B is recurrent.
- if then as .
Proof: For , let be a nonempty open set and choose . Then, is a process. By Theorem 6, , so at arbitrarily large times.
If then is a process so, by Theorem 6, as . ⬜
Finally for this post, I show that Bessel process are well defined Markov processes and that continuous modifications do indeed exist.
Lemma 8 Fix any and, for each , let be the transition probability on such that, for each , has moment generating function given by
Then, is a Feller transition function. Furthermore, any Markov process with this transition function has a continuous modification, which is then a process.
Proof: To show that is a transition function, it is only necessary to prove the Chapman-Kolmogorov equation . This can be verified by directly computing the moment generating functions. Setting ,
so as required.
We now move on to the proof that this is a Feller transition function. It needs to be shown that for any in the set of continuous functions vanishing at infinity, then and as . First, if for a constant then
which satisfies the required properties. This extends to all by uniformly approximating by linear combinations of such functions (see Lemma 9 below).
It only remains to show that a Markov process X with the transition function has a continuous modification. Any such process automatically satisfies (4) and, if it is continuous, is a squared Bessel process by definition. By the existence of cadlag modifications for Feller processes, we may assume that X is cadlag. It just needs to be shown that the jumps are almost surely equal to zero. If are a sequence of times with , then . From this, the following inequality is obtained, bounding the maximum jump of X over the interval .
We show that the left hand side converges to zero in probability. The moment generating function of can be computed from (4),
where . The expected value of conditional on can be calculated by expanding this out as a power series and looking at the coefficient of . This is a bit messy but, noting that the expression above expands in terms of and , we see that the coefficient of will be equal to multiplied by a polynomial in t–s, s and . Therefore,
over any bounded range for s and t. Taking the expected value conditional on , it follows that the left hand side of (13) goes to zero at rate . So, is almost surely zero, as required. ⬜
The following result for uniformly approximating continuous functions on was used in the proof that is a Feller transition function.
Lemma 9 The set of linear combinations of functions of the form for is dense in (in the uniform norm).
Proof: Any can be written as where is continuous with . By the Stone-Weierstrass approximation theorem there are polynomials converging uniformly to g on the unit interval. Replacing by if necessary, we can suppose that . Then, are linear combinations of functions of the form for and converges uniformly to . ⬜