Here, I attempt to construct a counterexample to the hypotheses of the earlier post, Do convex and decreasing functions preserve the semimartingale property? There, it was asked, for any semimartingale X and function such that is convex in x and right-continuous and decreasing in t, is necessarily a semimartingale? It was explained how this is equivalent to the hypothesis: for any function such that is convex and Lipschitz continuous in x and decreasing in t, does it decompose as where and are convex in x and increasing in t. This is the form of the hypothesis which this post will be concerned with, so the example will only involve simple real analysis and no stochastic calculus. I will give some numerical calculations suggesting that the construction below is a counterexample, but do not have any proof of this. So, the hypothesis is still open.
Although the construction given here will be self-contained, it is worth noting that it is connected to the example of a martingale which moves along a deterministic path. If is the martingale constructed there, then
defines a function from to which is convex in x and increasing in t. The question is then whether C can be expressed as the difference of functions which are convex in x and decreasing in t. The example constructed in this post will be the same as C with the time direction reversed, and with a linear function of x added so that it is zero at .
I now move on to the construction of the example. For symmetry, I will consider x to take values in rather than the unit interval. So, we will construct a function where is the rectangle
such that is convex in x and decreasing in t. We start by defining f for ,
Next, extend to times () by successively linearly interpolating across intervals ,
Specifically, using the following intervals
gives the plots as in the first two iterations in Figure 2.
We now interpolate f between the times . For we have so, by monotonicity, f must be constant in t across the interval . On the other hand, for , we interpolate f as a scaled copy of itself on the whole of , plus a linear term in x in order to line up at times and . So, f will be self-similar on the rectangles . To express this self-similarity, define functions
We also let be the change in across the interval . To be explicit,
Then, f is defined on by,
This construction of f can be done iteratively. Let denote the space of functions such that is convex in x and continuous and decreasing in t, and satisfying (1). Then define by where, for ,
It can be seen that, with respect to the supremum norm, F is Lipschitz continuous on with Lipschitz constant 1/2. So, by the Banach fixed point theorem, there is a unique fixed point . The fixed point can be constructed iteratively. First choose any and, then, inductively define . By the Banach fixed point theorem this will converge uniformly to the required f. Furthermore, for each fixed n, the values of on the time slices () are fixed over all . So, equals at these time slices. Computing successive iterations generates f at these sequences of time slices, as in Figure 2. Figure 1 shows f computed in this way using 5 iterations.
Computing the decomposition
With the function f constructed as above, we ask if it is possible to decompose
for functions and convex in x and increasing in t. The hypotheses of the post Do convex and decreasing functions preserve the semimartingale property? state that this is possible. However, some numerical calculations which I will present here suggest that f is a counterexample, and decomposition (2) does not exist. I do not have any proof of this though, so the hypotheses are still open.
The method described in the earlier post can be used to construct the decomposition (2), if it exists. The idea is to construct to be convex in x such that is increasing in t and, then, (2) would follow by taking . This is done by approximating h along partitions of the unit interval. Using to denote the convex hull of a function u, the approximation is defined inductively at times by
As the function f was constructed on the n‘th iteration at times (), it is convenient to also compute the approximation to h along these partitions. Also, as is piecewise linear in x at these times, the same holds for h. This makes it relatively simple to compute the approximation to h for each successive iteration of f on a computer. The resulting time-slices, are shown in Figure 3.
It can be seen that the successive approximations to h are decreasing, initially having a minimum of -1 for the first two iterations, but a minimum of around -1.2 at the fourth iteration. The plot of h calculated at iteration 5 is given in Figure 4, showing that the minimum is slightly less than -1.3.
Now, by Lemma 2 of the previous post, one of the following statements holds.
- Decomposition (2) exists, and the successive approximations for h converge to a finite limit.
- Decomposition (2) does not exist, and the successive approximations for diverge to .
In an attempt to determine which of these cases is true, I computed h for a few iterations and looked at the successive and norms of h at time ,
Using convexity it can be shown that the norms are equivalent, , although I would expect the norm to be a bit better behaved as we compute successive iterations. The calculations for up to iteration 13 are given in the table below, and it can be seen that the norm is slowly but steadily increasing.
The function f constructed in this post is a counterexample to the hypothesis of the previous post if and only if the norm is increasing to infinity. Figure 5 displays the behaviour, and it does look like the norm is increasing roughly linearly in the number of iterations.
The successive differences in the norms are monotonically decreasing, but only very slowly. Certainly, they seem to be decreasing much more slowly than , in which case the norm would tend to infinity and f would be the counterexample to the hypotheses mentioned above. However, I do not have any proof of this. Also, the complexity of calculations used to obtain the numbers above grow exponentially in the number n of iterations. The number of time points in the partition grows like , and the number of x at which f and h are computed grows at rate . So, the time taken is of order and it quickly becomes infeasible to compute h as the number of iterations becomes large.