The principal reason for introducing the concept of semimartingales in stochastic calculus is that they are precisely those processes with respect to which stochastic integration is well defined. Often, semimartingales are defined in terms of decompositions into martingale and finite variation components. Here, I have taken a different approach, and simply defined semimartingales to be processes with respect to which a stochastic integral exists satisfying some necessary properties. That is, integration must agree with the explicit form for piecewise constant elementary integrands, and must satisfy a bounded convergence condition. If it exists, then such an integral is uniquely defined. Furthermore, whatever method is used to actually construct the integral is unimportant to many applications. Only its elementary properties are required to develop a theory of stochastic calculus, as demonstrated in the previous posts on integration by parts, Ito’s lemma and stochastic differential equations.
The purpose of this post is to give an alternative characterization of semimartingales in terms of a simple and seemingly rather weak condition, stated in Theorem 1 below. The necessity of this condition follows from the requirement of integration to satisfy a bounded convergence property, as was commented on in the original post on stochastic integration. That it is also a sufficient condition is the main focus of this post. The aim is to show that the existence of the stochastic integral follows in a relatively direct way, requiring mainly just standard measure theory and no deep results on stochastic processes.
Recall that throughout these notes, we work with respect to a complete filtered probability space . To recap, elementary predictable processes are of the form
for an -measurable random variable , real numbers and -measurable random variables . The integral with respect to any other process X up to time t can be written out explicitly as,
The predictable sigma algebra, , on is generated by the set of left-continuous and adapted processes or, equivalently, by the elementary predictable process. The idea behind stochastic integration is to extend this to all bounded and predictable integrands . Other than agreeing with (2) for elementary integrands, the only other property required is bounded convergence in probability. That is, if is a sequence uniformly bounded by some constant K, so that , and converging to a limit then, in probability. Nothing else is required. Other properties, such as linearity of the integral with respect to the integrand follow from this, as was previously noted. Note that we are considering two random variables to be the same if they are almost surely equal. Similarly, uniqueness of the stochastic integral means that, for each integrand, the integral is uniquely defined up to probability one.
Using the definition of a semimartingale as a cadlag adapted process with respect to which the stochastic integral is well defined for bounded and predictable integrands, the main result is as follows. To be clear, in this post all stochastic processes are real-valued.
Theorem 1 A cadlag adapted process X is a semimartingale if and only if, for each , the set
is bounded in probability.
As was previously noted, the necessity of this condition follows from bounded convergence. In fact, boundedness in probability of the set in (3) is equivalent to the statement that, for any sequence of bounded predictable processes converging uniformly to zero, converges to zero in probability. It is interesting to note that this seemingly much weaker uniform convergence property is enough to imply the stronger property of bounded convergence in probability.
The proof of Theorem 1 is given below. However, it helps to take a step back at this point and ask the following question. What do we need to do, in general, to construct linear maps satisfying bounded convergence of sequences? This question is not restricted to stochastic calculus, and can be asked in a much more general situation. Given a measurable space and a topological vector space V, what does it take to construct a linear map which satisfies bounded convergence? In this general situation, bounded convergence means that if is a sequence uniformly bounded by a constant K, so that , and then, in the topology of V. Such maps will be referred to as V-valued measures. For stochastic integration, we are only really concerned with the case and where V is the space of random variables under the topology of convergence in probability. However, even in the general situation the following result is true. Here, a subalgebra of is a subset closed under linear combinations and pointwise multiplication, and containing the constant functions.
Theorem 2 Let be a measurable space, be a subalgebra of generating , and V be a complete vector space. Then, a linear map extends to a V-valued measure on if and only if it satisfies the following properties for sequences .
- If then .
- If , then .
The proof of this result will be given in the next post, as the details would cloud the main argument here and, in any case, it is just a statement involving pure measure theory with no stochastic calculus whatsoever. Both of the conditions of this theorem involve a uniformly bounded sequence tending to zero so that, if is to extend to a measure satisfying bounded convergence, it is necessary that . The main result is that these are also sufficient conditions.
It is instructive to pause here for a moment, and consider how these conditions are shown to be true in the case of Lebesgue integration and Stieltjes integration with respect to finite variation functions, and to compare with the stochastic case. The first condition, of monotone convergence, can be shown by making use of the compactness of the interval , and follows for stochastic integration in much the same way as for the Lebesgue integral. In fact, if the set of functions from which the integral is to be extended are continuous then, by compactness, any sequence decreasing monotonically to zero will also converge uniformly. This useful property is used in the proof below, first extending the integral to continuous integrands and, then, applying Theorem 2 to extend to all bounded predictable integrands.
The second condition of Theorem 2 is really where the difference between standard integration and the stochastic integral comes in. Consider, for example, the usual Lebesgue integral on the interval . This is a nonnegative map so that, as long as the absolute value of any is also in , the following inequality holds
It follows that so that as required by Theorem 2.
Similarly, if represents integration with respect to a function of finite variation V, then will be bounded by V whenever . Defining constants whenever and otherwise then, under the second condition of the theorem, the partial sums are all bounded by 1. This gives the inequality
So, once again, is finite.
This argument does not apply to stochastic integration, which takes values in the space of random variables. This is because as defined above would not be real numbers but, instead, are themselves random variables. So, need not be predictable processes and equation (4) just does not make sense. In fact, in the stochastic case it is not even true that is finite. For example, suppose that is integration with respect to standard Brownian motion on the unit interval. If then, are independent centered Gaussians with variance 1/n-1/(n+1). Their absolute values have mean going to zero at rate 1/n and, with probability 1, .
However, it turns out that a modified argument along similar lines does indeed work for stochastic integration and, more generally, for all vector-valued measures taking values in a space of random variables under convergence in probability. Instead of carefully choosing to make positive, as in equation (4), the total opposite is done. That is, are chosen completely at random in the set . This will be enough to show that is finite. The idea is to apply a variant of the Khintchine inequalities, as stated below in Lemma 3.
Suppose that is a sequence of independent random variables, each taking the values 1,-1 both with probability 1/2. If is a sequence of real numbers, define a random variable Z and by,
so that Z has mean zero and variance .
Lemma 3 Given a real number , there exists a such that for any of the form (5),
Here, only depends on K and not on the choice of .
Proof: We shall make use of the inequality
This holds for any integrable random variable with positive mean, and follows from taking expectations of and substituting for .
Now use the properties that are independent with zero mean and to get and,
If all of the are zero, the result is immediate for any , otherwise put into (6),
and the result follows by taking to be . ⬜
Lemma 3 can be used to give the following result concerning sums of random variables.
Lemma 4 Let be a sequence of random variables such that the set
is bounded in probability. Then, is almost surely finite. In particular, almost surely and, hence, also in probability.
Proof: Set . Then, if is the uniform probability measure on and are as in Lemma 3,
Now, choose a positive constant L, multiply both sides by and take expectations. Exchanging the order of integration with respect to and gives,
For any , the boundedness of S in probability means that L can be chosen large enough that for all Y in S. Substituting this inequality in the left hand side of (7),
Now, let n increase to infinity,
Finally, as can be chosen arbitrarily small, this implies that has zero probability. ⬜
This result shows that, in the case where V is a space of random variables under the topology of convergence in probability, the second condition in Theorem 2 can be dropped altogether.
Theorem 5 Let be a measurable space, be a subalgebra of generating , and be a probability space. Then, a linear map extends to an -valued measure on if and only if it satisfies the following property.
- For any sequence with , then (in probability).
Proof: It just needs to be shown then the second condition of Theorem 2 is satisfied. First, note that the condition of the theorem is strong enough to ensure that the set
is bounded in probability. This follows from the sequential characterization of boundedness — if are bounded by 1 then is decreasing to zero, so tends to zero.
Now consider a sequence satisfying . Then, for any , the partial sum is bounded by 1. So, the set of all such sums is in S and, hence, is bounded in probability. Then, Lemma 4 says that is almost surely finite, so that in probability. The result now follows from Theorem 2. ⬜
Proof of the existence of the stochastic integral
Throughout this section, it is assumed that X is a cadlag adapted process and that the set in (3) is bounded in probability. Fixing a time t, define for elementary predictable processes , which is a linear map from the elementary predictable processes to . The condition on X implies that is continuous under the topology of uniform convergence for . This can be strengthened to uniform convergence on compacts in probability.
Lemma 6 If is a sequence of elementary predictable processes converging ucp to zero, then in probability.
Proof: By definition of ucp convergence, tends to zero in probability as n goes to infinity. It follows that there is a sequence such that . Then, are elementary processes converging uniformly to zero, so in probability. From (1), whenever and agree on , which has probability , so in probability as required. ⬜
By extension of continuous linear functions, this allows us to uniquely extend to the closure of the elementary processes under the ucp topology, say, . So, is linear and continuous under the ucp topology on and convergence in probability on .
Every uniformly bounded, continuous and adapted process is in . In particular, the elementary processes
Lemma 7 extends to an -valued measure .
Proof: As noted above, can be extended to the closure of the elementary predictable processes under ucp convergence, which contains the set of uniformly bounded, continuous and adapted processes.
Suppose that a sequence of nonnegative processes decreases to zero. Then, for each and , the sets are closed, bounded and have empty intersection. By compactness, for large n, showing that decreases to zero. So, converges ucp to zero and, by continuity, goes to zero in probability.
Theorem 5 can now be applied to find an -valued measure agreeing with on . To complete the proof, it only remains to show that they agree on the elementary predictable processes.
Choosing , any elementary predictable process can be smoothed out as follows to obtain a continuous and adapted process,
Approximating uniformly by the elementary processes gives
where . Taking the limit as n goes to infinity,
where . Finally, as decreases to zero, by left-continuity of , and by right continuity of X. This gives as required. ⬜
The definition I gave for a semimartingale as being any cadlag adapted process with respect to which the stochastic integral is well-defined differs somewhat from the `classical’ definition. Historically, semimartingales have been defined as processes which can be decomposed into the sum of a local martingale and an FV process. Stochastic integration was developed with respect to such processes using generalizations of the Ito isometry applied to locally square integrable martingales. It was later shown, by the Bichteler-Dellacherie theorem, that any process satisfying the conditions of Theorem 1 above does have such a decomposition. So, the classical semimartingale definition does indeed agree with one given in these notes.
A different approach is used here, because it is possible to approach subjects such as Ito’s lemma and stochastic differential equations immediately in their full generality, without first developing a lot of technical machinery. Also, the proof given in this post that stochastic integration can be performed with respect to all processes satisfying the conditions of Theorem 1 is much, much more direct than the classical one. For example, it should be clear that the hypothesis that X is adapted can be dropped altogether, as it wasn’t even used anywhere above. Furthermore, right-continuity is only required right at the end to ensure that the constructed integral agrees with the formula for elementary processes.
Later authors have used a closer approach to the one in these notes. For example, Protter (Stochastic Integration and Differential Equations) gives two definitions of semimartingales. Both the classical definition, and the one given by Theorem 1 above are provided. Using this newer semimartingale definition, he is able to derive the stochastic integral for integrands which are L-processes (left continuous with right limits) using continuity under ucp convergence as stated in Lemma 6 above. This is enough for many applications, such as Ito’s lemma. However, he doesn’t extend the integral to arbitrary predictable integrands directly, instead relying on the Bichteler-Dellacherie theorem to show that the classical and newer semimartingale definitions agree. Then, for the full development of the integral, techniques closer to the standard ones were used.
The only author I am aware of who has developed stochastic calculus without relying on the classical definition is Bichteler (Stochastic Integration with Jumps), where a semimartingale was defined as any process satisfying the conditions of Theorem 1, and the integral was developed directly from this. This method seems to date back to a 1981 paper by the same author (Stochastic Integration and Lp-Theory of Semimartingales). It is from his treatment of stochastic integration where the idea of using the Khintchine inequality (Lemma 3) to simplify the construction of the integral was obtained. This allowed us to remove the second condition in Theorem 2. There are other methods of proving that the integral satisfies this property, but the use of the Khintchine inequality shows that it is much more general, applying to arbitrary -valued measures. In fact, Theorem 5 can be similarly proved for all spaces, , although that is not used in these notes.