In the previous post, I looked at optional projection. Given a non-adapted process X we construct a new, adapted, process Y by taking the expected value of conditional on the information available up until time t. I will now concentrate on predictable projection. This is a very similar concept, except that we now condition on the information available strictly before time t.
It will be assumed, throughout this post, that the underlying filtered probability space satisfies the usual conditions, meaning that it is complete and right-continuous. This is just for convenience, as most of the results stated here extend easily to non-right-continuous filtrations. The sigma-algebra
represents the collection of events which are observable before time t and, by convention, we take . Then, the conditional expectation of X is written as,
(1) |
By definition, Y is adapted. However, at each time, (1) only defines Y up to a zero probability set. It does not determine the paths of Y, which requires specifying its values simultaneously at the uncountable set of times in . So, (1) does not tell us the distribution of Y at random times, and it is necessary to specify an appropriate version for Y. Predictable projection gives a uniquely defined modification satisfying (1). The full theory of predictable projection for jointly measurable processes requires the predictable section theorem. However, as I demonstrate here, in the case where X is left-continuous, predictable projection can be done by more elementary methods. The statements and most of the proofs in this post will follow very closely those given previously for optional projection. The main difference is that left and right limits are exchanged, predictable stopping times are used in place of general stopping times, and the sigma algebra
is used in place of
.
Stochastic processes will be defined up to evanescence, so two processes are considered to be the same if they are equal up to evanescence. In order to apply (1), some integrability requirements need to imposed. I will use local integrability. Recall that, in these notes, a process X is locally integrable if there exists a sequence of stopping times increasing to infinity and such that
(2) |
is integrable. This is a strong enough condition for the conditional expectation (1) to exist, not just at each fixed time, but also whenever t is a stopping time. The main result of this post can now be stated.
Theorem 1 (Predictable Projection) Let X be a left-continuous and locally integrable process. Then, there exists a unique left-continuous process Y satisfying (1).
As it is left-continuous, the fact that Y is specified, almost surely, at any time t by (1) means that it is uniquely determined up to evanescence. The main content of Theorem 1 is the existence of Y, and the proof of this is left until later in this post.
The process defined by Theorem 1 is called the predictable projection of X, and is denoted by . So,
is the unique left-continuous process satisfying
(3) |
for all times t. In practice, X will usually not just be left-continuous, but will also have right limits everywhere. That is, it is caglad (“continu à gauche, limites à droite”).
Theorem 2 Let X be a caglad and locally integrable process. Then, its predictable projection is caglad.
The simplest non-trivial example of predictable projection is where is constant in t and equal to an integrable random variable U. Then,
is the left-limits of the cadlag martingale
, so
is easily seen to be a caglad process.
Existence of the predictable projection for caglad processes is much simpler than the general left-continuous case, and follows from the existence of the optional projection of cadlag processes. See Lemma 7 below. First, I prove the basic properties of the predictable projection under the assumption that it exists. Linearity of the projection is almost immediate.
Lemma 3 Let X and Y be left-continuous and locally integrable processes. Then, for
-measurable random variables
,
,
Proof: It is clear that is left-continuous and locally integrable. Setting
we have,
almost surely. So, by definition, Z is the predictable projection of . ⬜
Predictable projection commutes with multiplication by an adapted process.
Lemma 4 Let U and X be left-continuous processes such that U is adapted and X and UX are locally integrable. Then,
(4)
Proof: The process is left-continuous. As U is adapted and left-continuous,
is
-measurable, and
almost surely. So, by definition, Z is the predictable projection of UX. ⬜
An important property of predictable projection is the fact that equation (3) still holds when t is generalized to be any predictable stopping time. Compare with Theorem 5 from the post on optional projection. The main difference for predictable projection, other than the use of the sigma algebra of events observable strictly before
, is that we restrict to predictable stopping times.
Theorem 5 If X is a left-continuous and locally integrable process then, for all predictable stopping times
,
(5)
Proof: In order to avoid having to keep multiplying by the term in expressions such as (5), I will take all processes to be zero at infinity. Suppose, first, that
is integrable. Letting U be an adapted, bounded and left-continuous process, we need to prove
(6) |
As is
-measurable, and the sigma algebra
is generated by
as U ranges over the left-continuous and adapted processes, (6) will imply (5).
If is a simple predictable stopping time, taking values in a finite set S, then
is in
for each
. Using the fact that
is
-measurable,
This proves (5) for simple predictable stopping times. In particular, the set of random variables
over simple predictable stopping times is uniformly integrable.
Next, let be an arbitrary predictable stopping time. In the corresponding result for optional projection, we approximated
from the right by simple stopping times. For the current proof, we can proceed similarly if
is approximated from the left in the same way. However, this is not possible in general. Instead, we will construct simple stopping times approximating
and, almost surely, are eventually less than
.
Let be a sequence of stopping times announcing
. That is,
are stopping times increasing to
, and strictly less than
whenever
. Then, for each n, there will be a sequence of simple stopping times,
, decreasing to
as m goes to infinity. Replacing
by
if necessary, we can suppose that it is predictable. So, setting
for large enough m gives a simple predictable stopping time greater than
and satisfying
By the Borel-Cantelli lemma, this implies that, almost surely, for large n whenever
. We also set
when
. Then, by left-continuity,
,
and
converge, almost surely, to
,
and
respectively, as n goes to infinity.
So, using the above proof of (6) for simple predictable stopping times,
This proves (5) in the case where is integrable.
If X is locally integrable, then let be stopping times increasing to infinity such that (2) is integrable. From Lemma 4,
and the above shows that
(almost surely). Letting n go to infinity gives the result. ⬜
The restriction to predictable stopping times in Theorem 5 is necessary. For example, consider a compensated Poisson process M. This is a martingale with discrete jumps of size 1 and, if we let be the time of its first jump then the predictable projection of the constant process
(all t) is seen to be
As is
-measurable,
This contradicts (5). The reason is that is not a predictable stopping time and, in fact, is totally inaccessible.
I now show that predictable projection is a true projection operator. So, applying it twice gives the same result as applying it once.
Lemma 6 If X is a left-continuous and locally integrable process, then so is its predictable projection
and,
Proof: The fact that is adapted and left-continuous means that it is equal to its own predictable projection, just so long as we can show that it satisfies the requirement of local integrability. Supposing first that
is integrable, we have the bound,
where . As M is a martingale, its cadlag version is is locally integrable. So,
is left-continuous and locally integrable, showing that
is locally integrable.
Now, for arbitrary locally integrable X, let be a sequence of stopping times increasing to infinity such that (2) is integrable. Then, using Lemma (4),
is locally integrable for each n. So, is locally integrable as required. ⬜
Predictable Projection of Caglad Processes
I will now show that the predictable projection of a caglad process, X, exists and is itself caglad, proving Theorem 2 above. Taking the right limits gives a cadlag process and, using the results of the previous post, its optional projection exists. The predictable projection of X can be constructed from this simply by taking left limits.
Lemma 7 Suppose that X is a caglad locally integrable process. Then,
is a cadlag and prelocally integrable process. Furthermore, the predictable projection of X exists, is caglad, and satisfies
over
.
Proof: If is a sequence of stopping times increasing to infinity such that (2) is integrable then
is integrable. So, is prelocally integrable. Then, the optional projection of Y exists and is cadlag. So, the process
is caglad and adapted. It just needs to be shown that Z is the predictable projection of X. That is, the identity
(7) |
needs to be verified. This holds for by the definition of Z, so we suppose that t is strictly positive.
Let us first suppose that is integrable. Choosing any time
and bounded
-measurable random variable U, let
be a sequence of times increasing strictly to t.
(8) |
The second equality here uses the fact that
is a uniformly integrable sequence of random variables tending to . The third equality uses the fact that U is
-measurable, and the final equality uses dominated convergence. As the sigma algebras
over
generate
, (8) holds whenever U is bounded and
measurable, proving (7). ⬜
Predictable Projection of Left-Continuous Processes
I now give a proof that the predictable projection of left-continuous processes exists, proving Theorem 1. This is considerably more difficult to prove than the caglad case, and it is not possible to construct directly from the optional projection as we did above in Lemma 7. However, the proof does follow in a very similar way as for the optional projection of right-continuous processes. First, we show that it is possible to approximate X from above by caglad processes. As in the previous post, for processes X and Y, the inequality
is taken to mean that X is greater than Y up to evanescence. So, outside of a zero probability set,
for all t. Convergence of a sequence of processes
to a limit X is taken as pointwise convergence up to evanescence — outside of a zero probability set,
for all times t.
Lemma 8 Let X be a locally integrable left-continuous process. Then, there exists a sequence,
, of locally integrable caglad processes
which are decreasing in n and tending to X as n goes to infinity.
Proof: Choose a sequence of finite sets such that
for each n, and
is dense in
. For any given n we write
as
. Setting
and
, define the process
by
for all t in . We also set
equal to
. This is almost surely finite at all times, since X is locally integrable, and satisfies inequality
In particular, this implies that is locally integrable.
As is a refinement of
for
, the supremum in the definition of
is taken over a subinterval of
and, hence,
. Also, as
is dense in
, left-continuity ensures that
tends to
as n goes to infinity.
It is straightforward to see that is left-continuous and increasing across each interval
, so it is caglad. ⬜
The idea, then, is to construct the predictable projection of X by taking the limit of predictable projections of a sequence of caglad processes tending monotonically to X. This requires first proving that predictable projection commutes with taking limits of monotonic sequences of processes. To prove that, the following result will be used. This involves a rather tricky argument, and is very similar to Lemma 15 from the post on optional projection. The proof here is actually a bit more difficult, and makes use of the result for right-continuous processes. I will take all processes to be equal to 0 at time , in order to avoid having to keep multiplying by expressions such as
.
Lemma 9 Let
be a sequence of non-negative left-continuous adapted processes, and suppose that it is decreasing in n. If, for all predictable stopping times
,
as
, then
.
Proof: Given a random time , I will show that
tends to zero almost surely. Note that it is not required that
is a stopping time, much less a predictable stopping time. Fixing
, it is enough to show that
.
Letting be a sequence of strictly positive reals, define the times
,
It can be seen that these are stopping times. In fact, from left-continuity, if is a stopping time then
is -measurable, so
is also a stopping time. Furthermore, as
is strictly greater than
whenever it is finite, the limit
is a predictable stopping time and, by the hypothesis,
almost-surely as n goes to infinity. Fixing an m then, whenever
and
is finite, the construction of
implies that there exists times in the interval
where
and, hence,
exceeds
. So, whenever
is finite, left-continuity of X gives
. As this contradicts the limit
, we have shown that
and
increase to infinity with probability one.
So far, I have not stated how are chosen. Fixing
, the real numbers
are taken small enough so that
Summing over n, the probability that lies in the union of the intervals
is less than
.
Next, note that is bounded above by
on the intervals
. On the event
,
cannot lie in any of these intervals, so must be equal to 0 or lie in an interval
. Furthermore,
on this event since, by the hypothesis,
tends to 0 (a.s.). Hence,
So, as and
are arbitrary positive reals,
almost surely tends to 0.
Finally, for any positive real T, define the right-continuous processes
These are not adapted but, as they are measurable, they will be adapted with respect to the constant filtration . Then, for any
stopping time
,
and, as we have shown above, this tends almost surely to zero. By Lemma 15 from the previous post,
and, by taking T arbitrarily large,
. ⬜
Monotone convergence of the predictable projection follows easily from Lemma 9.
Lemma 10 (Monotone Convergence) Let
be a sequence of locally integrable left-continuous processes decreasing to 0 as n goes to infinity.
Then, supposing that their predictable projections
exist, they also decrease to 0.
Proof: For we have
almost surely. By left-continuity this implies that, outside of a zero probability set, is decreasing in n for all t. For any predictable stopping time
, Theorem 5 together with monotone convergence for the conditional expectation gives
almost surely. Lemma 9 then gives as required. ⬜
Finally, I give the proof of Theorem 1 and show that the predictable projection of all locally integrable and left-continuous processes exists. The predictable projection is constructed by taking limits of the predictable projections of caglad processes and applying monotone convergence. We follow along the same lines as the proof given for Lemma 17 in the post on optional projection.
Proof of Theorem 1: Applying Lemma 8 to both and
gives sequences of locally integrable caglad processes
decreasing in n, and
increasing in n, both tending to X as n goes to infinity. By Lemma 7, the predictable projections
exist. As and
are respectively decreasing and increasing in n, the same holds for their predictable projections. Lemma 10 says that
is decreasing to zero. So, we can define a process W up to evanescence by
By dominated convergence,
almost surely. To show that W is the predictable projection of X, only left-continuity remains. If is a sequence increasing to t then,
as m goes to infinity, for each fixed n. Then letting n go to infinity gives
Therefore, as required. ⬜