1. Introduction
Lévy processes are commonly employed in the study of asset returns and derivative pricing models, as well as in the prediction of high-frequency trading returns [Reference Barndorff-Nielsen and Shephard3, Reference Madan, Carr and Chang34, Reference Christensen, Murphy and Godsill24]. In derivative pricing in particular, the Black–Scholes model specifies the dynamics of a financial market which holds derivative instruments such as options, futures, and swaps [Reference Black and Scholes7]. Yet the assumption that option prices are modelled by a stationary log-Gaussian process often fails to be validated by empirical data [Reference Black and Scholes7], motivating research into the relaxation of this assumption. Lévy processes such as the normal tempered stable process [Reference Barndorff-Nielsen and Shephard4, Section 9], the generalised hyperbolic process [Reference Eberlein15], or the variance-gamma process [Reference Madan, Carr and Chang34, Reference Udoye and Ekhaguere43] can be used to model the variance, or volatility, of the security as a stochastic process.
In biology, the generalised inverse Gaussian (GIG) process has been used to model spike train activity from neurons [Reference Iyengar and Liao25], and to model the likelihood of extreme meteorological events with high economic and social costs [Reference El Adlouni, Chebana and Bobée17]. The tempered stable process has also been extensively studied in relation to Lévy flight models [Reference Rosinski40] in physics, and the variance-gamma process has been used to analyse continuous trait evolution such as population growth [Reference Landis, Schraiber and Liang35, Reference Russo41], and in the analysis of nitrate concentration in the eastern U.S. [Reference Woodard, Wolpert and O’Connell48]. More recently, the leptokurtic properties of non-Gaussian Lévy processes have been exploited in the field of object tracking [Reference Gan, Ahmad and Godsill21] in order to capture abrupt changes in the kinematic state of a moving object, which in turn has applications to animal tracking and the analysis of hunting patterns.
We are here concerned with the analysis of different representations of Lévy-driven stochastic processes of the form
where $\boldsymbol{A}$ is a $P\times P$ matrix, $\boldsymbol{h}\in\mathbb{R}^P$ , and (W(t)) is a one-dimensional non-Gaussian Lévy process.
Based on infinite-series representations for $\left(W(t)\right)$ , tractable and conditionally Gaussian models can be developed for simulation and inference purposes; see, for example, [Reference Lemke and Godsill33, Reference Godsill, Riabiz and Kontoyiannis23, Reference Gan, Ahmad and Godsill21], where such models and inference procedures were developed for $\alpha$ -stable Lévy processes driving linear stochastic differential equations (SDEs). Here, by contrast, we study variance-mean mixture Lévy processes [Reference Barndorff-Nielsen and Shephard3], including (but not limited to) the normal-gamma, normal tempered stable, and generalised hyperbolic processes. In particular, we provide functional central-limit-theorem-like results for the convergence of the residual terms when series representations of such processes are truncated at a finite level, and for the corresponding Gaussian convergence of the above SDE’s residuals when it too is truncated in a corresponding fashion. The results are linked in spirit to the study of [Reference Asmussen and Rosinski2], although posed in terms of a different truncation of small jumps.
Lévy processes have been studied extensively via the general theory of stochastic processes with stationary, independent increments. Samorodnitsky and Taqqu [Reference Samorodnitsky and Taqqu42], Küchler and Tappe [Reference Küchler and Tappe31], and Barndorff-Nielsen and Shephard [Reference Barndorff-Nielsen and Shephard3, Reference Barndorff-Nielsen and Shephard4, Reference Barndorff-Nielsen and Shephard5] consider important special cases, including the tempered stable, inverse Gaussian, and $\alpha$ -stable processes. The relevant properties established are then used in the definition and analysis of broader classes of Lévy processes, such as the normal variance-mean (NVM) mixture processes [Reference Barndorff-Nielsen and Shephard4], which are formed by the subordination of Brownian motion to a non-negative process. Samorodnitsky and Taqqu also examine stochastic integrals with respect to $\alpha$ -stable random noise, providing a series representation closely related to the series representation of the $\alpha$ -stable process. Barndorff-Nielsen extends the series representation of $\alpha$ -stable stochastic integrals to ones driven by a non-negative Lévy process [Reference Barndorff-Nielsen and Shephard3, Section 8], and Rosinski studies an analogous representation for symmetric Gaussian mixture processes [Reference Rosinski38].
The breadth and depth of the study of Lévy processes have allowed for numerous studies into their applications in modelling rare-event phenomena, e.g. in finance [Reference Barndorff-Nielsen and Shephard3, Reference Christensen, Murphy and Godsill24, Reference Cont and Tankov12], physics [Reference Rosinski40], and biology [Reference Iyengar and Liao25, Reference El Adlouni, Chebana and Bobée17, Reference Woodard, Wolpert and O’Connell48, Reference Gan, Ahmad and Godsill21].
Efficient methods for simulating Lévy processes have been critical in successfully bridging the gap between theory and application. Muller and Box outline a method of generating normal random variates [Reference Box and Muller9], and Chambers, Mallows and Stuck study the generation of $\alpha$ -stable random variables, both of which are summarised in work by Barndorff-Nielsen [Reference Barndorff-Nielsen and Shephard5, Sections 2.3–2.4] and Samorodnitsky and Taqqu [Reference Samorodnitsky and Taqqu42]. While simulation from finite-activity processes is straightforward [Reference Cont and Tankov12], exact simulation from infinite-activity processes, such as those considered in this work, is impossible, because of the presence of an infinite number of small jumps in any finite time interval. Khinchin [Reference Khintchine29], Bondesson [Reference Bondesson8], and Fergusson and Klass [Reference Ferguson and Klass18] outline the inverse Lévy method for simulating jumps from infinite-activity processes with non-negative increments, and Rosinski [Reference Rosinski39] develops generalised shot-noise methods with thinning and rejection sampling for processes with Lévy densities that cannot be simulated directly using approaches such as the inverse Lévy method. In [Reference Godsill and Kndap22], Godsill and Kindap propose novel algorithms for simulating from the infinite-activity GIG process by combining previous approaches, paving the way for direct simulation of generalised hyperbolic jumps.
While most simulation methods in principle require the computation of an infinite sum, truncation of the series is of course required in practice. Asmussen and Rosinski [Reference Asmussen and Rosinski2] analyse the effect of strict truncation of the process jumps on the representation of the ‘residual process’, namely, the truncation error. They provide necessary and sufficient conditions for the convergence of the residual to a diffusion process, and several authors [Reference Deligiannidis, Maurer and Tretyakov13, Reference Dia14] provide bounds on the rate of convergence of the residual for specific classes of Lévy processes. Further work [Reference Carpentier, Duval and Mariucci11, Reference Vlad and Yifeng44] has also examined the Gaussian representation of strictly small jumps when studying SDEs. In the setting of SDEs driven by Lévy processes, the work of [Reference Kohatsu-Higa and Tankov30] provides a promising approach in which the process is simulated by ODE solvers between large jumps of the process, leading to an alternative and general methodology for nonlinear SDEs. In the context of the generalised shot-noise representation of more complex processes, Rosinski studies instead random Poisson truncations of the epochs of the process. This methodology is applied by Godsill et al. [Reference Godsill, Riabiz and Kontoyiannis23] to the $\alpha$ -stable process, where they provide verification of the convergence of the $\alpha$ -stable residual to a Brownian motion. Yet the analysis there is limited to the $\alpha$ -stable normal mixture process where the mean and standard deviation are proportional to an underlying stable process, and it does not consider the much broader class of NVM mixture processes, where the mean and variance are proportional to the subordinator.
A framework for performing inference on state space models described by linear Lévy-driven SDEs is presented in [Reference Godsill, Riabiz and Kontoyiannis23]. The problem of inference on linear Gaussian state space models has largely been solved through the development of the Kalman filter [Reference Kalman27]. Variants such as the extended Kalman filter and the unscented Kalman filter have also been introduced in order to handle nonlinearities [Reference Alspach and Sorenson1]. Inference on non-Gaussian models has been accomplished primarily through Bayesian modelling combined with sequential Monte Carlo methods (outlined e.g. in [Reference Cappé, Godsill and Moulines10, Reference Kantas28]), and this is exploited in [Reference Godsill, Riabiz and Kontoyiannis23] through the use of the Rao–Blackwellised particle filters to perform inference for a conditionally Gaussian state space model driven by Lévy noise. The algorithm there uses the Kalman filter on the linear conditionally Gaussian part, and it employs particle filtering to infer the distribution of the hidden process driven by the $\alpha$ -stable noise. The main aim of this paper is to provide the theoretical foundations with which to extend the ‘Lévy state space model’ [Reference Godsill, Riabiz and Kontoyiannis23] to a broader class of processes.
2. Background: Lévy processes and series representations
2.1. Lévy processes
A Lévy process, $(X(t))=(X(t)\;;\;t\geq 0)$ , is a real-valued, infinitely divisible stochastic process with stationary and independent increments. The log-characteristic function, commonly known as the characteristic exponent (CE) or cumulant function, of any Lévy process is given by [Reference Eberlein16]
where $\mathbb{R}^{*} \,:\!=\, \mathbb{R}\backslash \{0\}$ . The term $\mathbb{1}(|x|<1)ix\theta$ is a centring term that ensures convergence of the CE for processes for which $\int_{|x|<1}xQ(dx)$ is divergent, though it can be omitted for processes with finite first absolute moment. The Lévy triplet $(a, b^{2}, Q)$ uniquely defines the Lévy process [Reference Eberlein16], with $a \in \mathbb{R}$ , $b \in [0,\infty)$ , and where the Lévy measure, Q(dx), is a Poisson-process intensity measure defining the distribution of jumps in the Lévy process, satisfying
where $(a\wedge b)$ denotes the minimum value of a and b. See e.g. [Reference Barndorff-Nielsen and Shephard5, Section 2.3] for more details.
2.2. Subordinators
A subordinator $(Z(t))_{t\geq0}$ is a particular case of a Lévy process with non-negative increments, such that its paths are almost surely (a.s.) increasing [Reference Wolpert and Ickstadt47]. The Lévy measure of any subordinator, $Q_{Z}(dz)$ , satisfies [Reference Barndorff-Nielsen and Shephard5]
Observe that this is a stricter condition than that required for general Lévy processes in (2.2). Consequently, the Lévy triplet of a subordinator has $b^{2} = 0$ and no centring term is required [Reference Wolpert and Ickstadt47]. The current work will consider only subordinators without drift; thus the CE for such a subordinator (Z(t)) will always be of the form
The mean and variance of Z(t), when they exist, may be obtained for all t from the Lévy measure:
Most but not all of the processes we consider here will have finite first and second moments, because their Lévy densities decay exponentially for large jump sizes; see Sections 3.1.4 and 3.3.2 of [Reference Barndorff-Nielsen and Shephard5]. We will also be concerned here with so-called infinite-activity subordinators, which exhibit infinitely many jumps in any finite time interval: $Q_Z((0,\infty)) = \infty$ . Combined with (2.2), this implies the presence of infinitely many small jumps ( $|x|<1$ ) within finite time intervals.
2.3. Normal variance-mean processes
A normal variance-mean (NVM) process is defined as time-deformed Brownian motion, where jumps in a subordinator process (Z(t)) drive random time deformations of an independent Brownian motion (B(t)) as follows [Reference Barndorff-Nielsen and Shephard5]:
Here (B(t)) is a standard one-dimensional Brownian motion and (Z(t)) is a subordinator process as in the previous section. We limit attention to the case $\mu=0$ without loss of generality. The parameter $\mu_{W}$ models the skewness of the jump distribution, with a fully symmetric process for $\mu_{W} = 0$ . The specification of $\left(Z(t)\right)$ , coupled with the choice of $\mu_{W}$ and $\sigma_{W}$ , allows for a broad family of heavy-tailed and skewed processes to be implemented. We can express the Lévy measure Q of any such process in terms of its subordinator’s Lévy measure $Q_Z$ :
where ${\mathcal N}(\cdot;\,\mu,\sigma^2)$ denotes the Gaussian law with mean $\mu$ and variance $\sigma^2$ . Finally, if $K_Z(t;\,\theta)$ is the Lévy–Khintchine exponent of Z(t) in (2.4), then the CE for the subordinated process is given by [Reference Cont and Tankov12, Section 4.4]
from which we obtain the mean and variance directly in terms of the moments of the subordinator. Specifically, for $t\geq 0,$
and
when these expectations exist.
We will consider examples based on the normal-gamma, normal tempered stable, and generalised hyperbolic processes, which are obtained via Brownian motion subordinated to the gamma, tempered stable, and GIG processes, respectively, with Lévy measures shown in (2.10), (2.11), (2.12), respectively:
Appropriate ranges for the different parameters are specified in Section 6.
2.4. Generalised shot-noise representation
The shot-noise representation of a Lévy process is fundamental to the simulation methods in [Reference Godsill, Riabiz and Kontoyiannis23], [Reference Rosinski39], and [Reference Godsill and Kndap22]. Adopting the perspective of viewing a Lévy process as a point process defined on $[0, T] \times \mathbb{R}^{d}$ , Rosinski [Reference Rosinski39] considers the equivalent representation
where $\delta_x$ denotes the Dirac measure at x, and $\left\{V_{i}\right\}$ is a sequence of independent and identically distributed (i.i.d.) uniforms, $V_{i} \sim U[0,T]$ , independent of $\left\{Z_{i}, U_{i}\right\}$ . Intuitively, $\{V_{i}\}$ represent the arrival times of the jumps $X_{i} = H(Z_{i}, U_{i})$ in the process, and $\{Z_i\}$ represent a non-increasing set of jump sizes drawn from the subordinator process.
For the NVM Lévy process, $H(\cdot)$ is related to the time-domain representation of the process in (2.5) via
The ordered jumps $\{Z_i\}$ are typically simulated through generation of the ordered epochs $\{\Gamma_i\}$ of a standard Poisson point process, obtained by the partial sum of exponential random variables, $e_{i} \overset{i.i.d.}{\sim} {\textrm{Exp}}(1)$ : $\Gamma_{i} = \Gamma_{i-1} + e_{i}.$ Then we obtain the ith ordered subordinator jump through a function $Z_{i} = h(\Gamma_{i})$ as
where the map $h(\cdot)$ can be expressed in terms of the upper tail of $Q_Z$ [Reference Barndorff-Nielsen and Shephard5, Section 3.4.1] as $h(\Gamma_{i}) = \inf \{z \in \mathbb{R} : Q_Z([z,\infty)) < \Gamma_i\},$ which may or may not be available in closed form, depending on the particular choice of $Q_Z$ . If it is not available, then rejection sampling or thinning methods can be employed, as in [Reference Godsill and Kndap22], which leads to suitable algorithms for all of the models considered here.
Substituting (2.13) into the Lévy–Khintchine representation (2.1), the Lévy process can be expressed as the convergent infinite series [Reference Rosinski39]
where $b_{i}$ is a compensator or centring term [Reference Rosinski39] that ensures the convergence of the series. A Lévy process need not be compensated if and only if
see [Reference Wolpert and Ickstadt47, Reference Winkel46]. In view of (2.3), any subordinator satisfies this condition, and hence so does the corresponding NVM process; see Appendix C. Therefore, none of the Lévy processes studied in this paper requires compensating terms, and we take $b_{i} = 0 $ for all i throughout.
3. Random truncation of subordinator
Consider the problem of simulating an NVM Lévy process via the shot-noise representation in the previous section. Suppose we have access to a simulation algorithm for subordinator generation, which is capable of producing a non-increasing set of random jumps $\{Z_i\}$ , $i=1,2,\ldots$ , and associated uniformly distributed random jump times $\{V_i\}$ . Exact simulation for infinite-activity processes via (2.15) is impossible with finite computational resources, so simulation of a truncated process is typically implemented. In this paper, we consider the effect of random Poisson truncations on the jumps of the subordinator process as in [Reference Godsill, Riabiz and Kontoyiannis23, Reference Rosinski39],
where, as before, H is defined by $H(Z_{i}, U_{i}) = \mu_{W}Z_i + \sigma_{W}\sqrt{Z_i}U_{i}$ , with $U_{i}$ being i.i.d. $\mathcal{N}(0,1).$ Then $\hat{X}_\epsilon(t)$ is the process X(t) with subordinator jumps truncated at level $\epsilon$ , and $X_{\epsilon}(t)$ is the remainder, corresponding to the small jumps. There intuitively exists a trade-off between the computational complexity of computing $\hat{X}_\epsilon(t)$ , how accurately it approximates the true process X(t), and whether a Gaussian approximation to $X_{\epsilon}(t)$ is valid.
The left-hand plots in Figures 1, 2, and 3 show sample paths from the truncated versions of the gamma (with parameters $\gamma=\sqrt{2}$ and $\nu=2$ ), tempered stable (with parameters $\kappa=1/2$ , $\gamma=1.35$ , and $\delta=1$ ), and GIG (with parameters $\delta=1/3$ , $\gamma=\sqrt{2}$ , and $\lambda=0.2$ ) subordinators, using the above methodology with $\epsilon=10^{-10}$ , as described in [Reference Godsill and Kndap22]. See Section 6 for the precise definitions of these processes. The corresponding right-hand plots compare the empirical distributions of $N=10^5$ truncated process values at time $t=1$ with random variates from the theoretical exact (not truncated) marginal distribution of each of these truncated processes at time $t=1$ . These are truncated at very low values of $\epsilon$ that might lead to infeasibly large computational burden in practical use. This motivates our justification of Gaussian approximations to the residuals in the following sections, when larger values of $\epsilon$ are used for computational reasons.
4. Gaussian process convergence
4.1. Preliminaries
Consider a subordinator (Z(t)) with Lévy measure $Q_Z$ . The corresponding process with jumps truncated at $\epsilon$ is denoted by $(Z_\epsilon(t))$ and has Lévy measure $Q_{Z_\epsilon}(B)=Q_Z(B\cap (0,\epsilon])$ for all Borel sets B. The corresponding residual NVM process, denoted by $(X_\epsilon(t))$ , has Lévy measure
and its moments can be obtained from (2.8) and (2.9):
and
where
Note that these moments are all well defined and finite since $Q_{Z_\epsilon}$ satisfies (2.3) and for all $n\geq 1$ we have $\lim_{\epsilon\rightarrow0}M_{Z_{\epsilon}}^{(n)} = 0$ .
Note also that, for any $0< \epsilon \leq 1$ and $m<n$ ,
since $x^n\leq \epsilon^{n-m} x^m$ for $0<x\leq \epsilon$ . In particular this implies that
and that
whenever the limit $\lim_{\epsilon\rightarrow 0}\frac{{M^{(1)}_{Z_\epsilon}}^2}{M^{(2)}_{Z_\epsilon}}$ exists.
4.2. Gaussianity of the limit of $X_{\epsilon}$
The following is our first main result; it gives conditions under which the residual process corresponding to an NVM Lévy process with jumps truncated at level $\epsilon$ converges to Brownian motion as $\epsilon\to 0$ . An analogous one-dimensional result for $\alpha$ -stable processes is established, along with corresponding convergence bounds, in [Reference Riabiz, Ardeshiri, Kontoyiannis and Godsill37].
Theorem 4.1. Consider a truncated NVM Lévy process $X_{\epsilon}=(X_{\epsilon}(t))$ with Lévy measure as in (4.1). Let the standardised process $Y_\epsilon$ be defined as $Y_\epsilon(t)=(X_\epsilon(t)-\mathbb{E}[X_\epsilon(t)])/\sigma_\epsilon$ , $t\geq 0$ . If
then $Y_\epsilon$ converges weakly to a standard Brownian motion $W=(W(t))$ in D[0,1] with the topology of uniform convergence.
Proof. The main content of the proof is in establishing the claim that $X_\epsilon(1)$ , properly standardised, converges to a Gaussian under (4.6). For that, it suffices to show that the CE of the standardised version of $X_\epsilon(1)$ converges pointwise to $-u^2/2$ . The CE of $X_\epsilon(1)$ is given as in (2.7), by
where $Q_X^\epsilon(dx)$ is the Lévy measure for $X_\epsilon$ , having the same structure as (4.1). The CE for $\tilde{X}_\epsilon(t)={X}_\epsilon(t)-\mathbb{E}[{X}_\epsilon(t)]$ is then given by
Note that $X_\epsilon$ has finite mean, $\int_{\mathbb{R}^*}x Q_X^\epsilon(dx)=\mu_{W}M_{Z\epsilon}^{(1)}<\infty$ , by the finiteness of $M_{Z\epsilon}^{(1)}$ in (4.3). Now, if we scale the centred process, $\tilde{X}_{\epsilon}$ , to have unit variance at $t=1$ , i.e., if we let $Y_\epsilon(t)=\tilde{X}_\epsilon(t)/\sigma_\epsilon,$ then the CE for $Y_\epsilon(1)$ becomes
where $v={iu\mu_{W}}{\sigma_\epsilon}-{u^2\sigma_{W}^2}/2$ , $\varphi_\epsilon(u)=\int_{0}^{\epsilon}\big[\exp\big(v\frac{z}{\sigma_\epsilon^2}\big)-1-v\frac{z}{\sigma_\epsilon^{2}}\big]Q_Z(dz)$ , and $\psi_\epsilon=\sigma_{W}^2\frac{M^{(1)}_{Z_\epsilon}}{\sigma_\epsilon^2}$ .
Consider the difference between $\exp(\phi^\epsilon_Y(u))$ and the CF of the standard normal:
First we establish bounds on $\varphi_\epsilon$ and $1-\psi_\epsilon$ . For $1-\psi_\epsilon$ , from (4.2) and (4.4) we have
therefore, since clearly $1-\psi_\epsilon\geq 0$ ,
For $\varphi_\epsilon$ , first recall that for any complex z with negative real part,
and so
The final integral is well defined by (4.3). Since $|v|^2 = {u^2\mu_{W}^2}{\sigma_\epsilon^2}+{u^4\sigma_{W}^4}/4$ , using (4.5) and (4.7), we have
where we used the fact that $M^{(2)}_{Z_\epsilon}/M^{(1)}_{Z_\epsilon} \leq \epsilon \rightarrow 0$ from (4.4). Combining the bounds in (4.7) and (4.10), we have
This bound is finite for any $|u|<\infty$ by the properties of $\sigma_\epsilon$ and $M^{(2)}_{Z_\epsilon}$ , and hence, under (4.6),
This proves the claimed pointwise convergence $\exp(\phi^\epsilon_Y(u))\rightarrow \exp({-}u^2/2)$ as $\epsilon\to 0$ .
Next we argue that the condition (4.6) is in fact sufficient for the process-level convergence claimed in the theorem. Note that the result of the claim also implies uniform convergence of the relevant characteristic functions on compact intervals [Reference Kallenberg26, Theorem 5.3, p. 86]. Now, it is straightforward, from the definition of a Lévy process, that the increments $Y_\epsilon(t)-Y_\epsilon(s)$ converge in distribution to the corresponding (Gaussian) increments of a Brownian motion, for all $s < t$ . We proceed to verify the requirement (III) in [Reference Pollard36, Theorem V.19], i.e. that given $\delta > 0$ , there are $\alpha > 0$ , $\beta>0$ , and $\epsilon_0 > 0$ such that ${\mathbb P}\{|Y_\epsilon(t) - Y_\epsilon(s)| \leq \delta\} \geq \beta$ whenever $|t - s| < \alpha$ and $\epsilon < \epsilon_0$ . To see this, note that the Lévy process $(Y_\epsilon(t) : t \in [0, 1])$ is centred, so for any $1 \geq t \geq s \geq 0$ , by Chebyshev,
By taking $\delta\in (0, 1)$ , we can choose any $\epsilon_0$ and $\alpha = \delta^3$ . Then, ${\mathbb P}\{|Y_\epsilon(t) - Y_\epsilon(s)| > \delta\}=1-{\mathbb P}\{|Y_\epsilon(t) - Y_\epsilon(s)| \leq \delta\}$ , so we can set $\beta=1-\delta$ , and we conclude that $(Y_{\epsilon}(t)) \xrightarrow{\epsilon \rightarrow 0}{} (W(t))$ in D[0,1], as claimed.
Next we show that some conditions are in fact necessary for the Gaussian limit in Theorem 4.1 to hold.
Theorem 4.2. Consider a truncated NVM Lévy process $X_{\epsilon}=(X_{\epsilon}(t))$ , and define the associated standardised process $Y_\epsilon$ as in Theorem 4.1. If the condition (4.6) does not hold and, moreover,
then $Y_\epsilon(1)$ does not converge to ${\mathcal N}(0,1)$ in distribution as $\epsilon\to 0$ .
Proof. Suppose the conditions in (4.11) hold. We will assume that $Y_\epsilon(1)$ converges to ${\mathcal N}(0,1)$ in distribution as $\epsilon\to 0$ , and derive a contradiction.
In the notation of the proof of Theorem 4.1, expanding the exponential series in (4.9) for one more term than before and using (4.8) yields
with $|D_\epsilon(v)|\leq |v|^3M^{(3)}_{Z_\epsilon}/(3!\sigma_\epsilon^6)$ . Hence, rearranging and integrating, we obtain
where $v={iu\mu_{W}}{\sigma_\epsilon}-{u^2\sigma_{W}^2}/2$ as before and $F_\epsilon(v)\,:\!=\,v^2 M^{(2)}_{Z_\epsilon}/\big(2\sigma_\epsilon^4\big).$ The assumption that $Y_\epsilon(1)$ converges to ${\mathcal N}(0,1)$ implies that $|\varphi_\epsilon(u)|\to 0$ for all u.
We consider several cases. First, we note that $F_\epsilon(v)$ cannot converge to zero for any $u\neq 0$ , since, by (4.11),
Second, we note that $D_\epsilon(v)$ also cannot converge to zero, because then $F_\epsilon(v)$ would need to converge to zero as well. The only remaining case is if, for every u, neither $F_\epsilon(v)$ nor $D_\epsilon(v)$ vanishes as $\epsilon\to0$ , while their sum does vanish. By (4.11) we have
On the other hand, by (4.12) and (4.13) we have
Therefore,
Since $L_1,L_2>0$ , this is clearly violated for u small enough, providing the desired contradiction and completing the proof.
Corollary 4.1. Suppose the subordinator Lévy measure $Q_{Z}(dz)$ admits a density $Q_{Z}(z)$ , which, for some $E>0$ , satisfies $Q_{Z}(z) > 0$ for all $z \in (0, E]$ . Then the condition (4.6) in Theorem 4.1 is equivalent to
Proof. A straightforward application of L’Hôpital’s rule twice gives
where the positivity of $Q_Z(z)$ in (0, E] was required to ensure that the derivative of the denominator is non-zero for all $\epsilon$ small enough.
5. Bounds on the marginal convergence rate
In addition to the asymptotic convergence of the process to a Brownian motion as in Theorems 4.1 and 4.2, it is possible to compute finite- $\epsilon$ bounds on the marginals' distance from Gaussianity. The following theorem, based on Berry–Esseen-style arguments, gives a general result. It may be observed once again (see Corollary 4.1) that the quantity $\epsilon/M^{(1)}_{Z_\epsilon}$ is of importance in determining the performance of different subordinators $Q_Z$ . In the following section we study specific NVM processes within this framework.
Theorem 5.1. Consider a truncated NVM Lévy process $X_{\epsilon}=(X_{\epsilon}(t))$ , and let the standardised process $Y_\epsilon$ be defined as in Theorem 4.1. Then the Kolmogorov distance $E_\epsilon$ between $Y_{\epsilon}(1)$ and $B\sim{\mathcal N}(0,1)$ satisfies
where $C=0.7975 \times 2\sqrt{2/\pi}$ , $\Phi(a, b;\, m)$ is the Kummer confluent hypergeometric function, and with the obvious extension of (4.3) to non-integer moments.
Furthermore, with the same constant C, $E_\epsilon$ may be bounded, for $\epsilon\in(0,1]$ , as
Proof. Arguing as in the proof of [Reference Godsill, Riabiz and Kontoyiannis23, Theorem 3.1], which was derived from [Reference Asmussen and Rosinski2, Theorem 2.1], the Kolmogorov distance $E_\epsilon$ between $Y^\epsilon(1)$ and $B\sim{\mathcal N}(0,1)$ is bounded above by
where $\sigma_{\epsilon}^{2}$ is the variance of the NVM process. From (2.9) it follows that
Using Fubini’s theorem, the third absolute moment of the residual process can be expressed as
and since the Kummer confluent hyper-geometric function is increasing for non-negative z [Reference Kummer32], we can bound ${\mathcal S}$ as
Substituting this bound into (5.5) and using (5.6), we obtain (5.1). Then, using the expansion
where $a^{(0)}=1$ and $a^{(n)}=a(a+1)(a+2)\cdots(a+n-1)$ , we obtain
from which the asymptotic expansion (5.2) is obtained.
Now, the term $M^{\big(\frac{3}{2}\big)}_{Z_\epsilon}/\sigma_{\epsilon}^{3}$ is bounded using (4.4) and (5.6), for $\epsilon\in(0,1]$ , as
which yields (5.3), and applying (5.7) once again leads to (5.4).
6. Examples
In this section, the validity of the conditions in Theorems 4.1 and 4.2 on the Gaussian convergence of the residual process is examined for several important cases of NVM Lévy processes. Simulation results validating the corresponding conclusions are also shown, and explicit bounds on the rate of convergence to the Gaussian are derived in some special cases, using the general framework of Theorem 5.1.
6.1. Normal-gamma process
The subordinator of the normal-gamma (NG) process is a gamma process, with parameters $\nu, \gamma>0$ and with Lévy density
Here, in view of Corollary 4.1,
and also,
Therefore, $L_1$ and $L_2$ in Theorem 4.2 are both non-zero, so we expect the residuals of the NG process not to be approximately normally distributed.
Furthermore, since $M^{(n)}_{Z_\epsilon}=\frac{\nu}{b^n}\gamma(n,b\epsilon)$ , where $b=\gamma^{2}/2$ , we have from (5.1) that
where we have obtained the asymptotic behaviour using $\gamma(s, x) = x^s\,\sum_{k=0}^\infty \frac{({-}x)^k}{k!(s+k)}$ (for positive, real s and x) and (5.7). Hence, as expected, the bound on the distance from Gaussianity, $E_\epsilon$ , does not tend to zero as $\epsilon\to 0$ .
In order to verify this empirically, we generate a random sample of $M=10^4$ residual NG paths with parameters $\mu=0$ , $\mu_w=1$ , $\sigma_W=\nu=2$ , and $\gamma=\sqrt{2}$ , and compare the empirical distribution of the values of the residual at time $t=1$ with a standard normal distribution by standardising the residual values to have zero mean and unit variance. As expected, the resulting histograms are not approximately Gaussian. Figures 4 and 5 show that the distribution of the residual (with truncation level $\epsilon=10^{-6}$ ) is in fact leptokurtic and heavier-tailed than the standard normal. Further simulations confirmed this empirical observation even for smaller truncation levels $\epsilon$ .
6.2. Normal tempered stable process
The subordinator for the normal tempered stable (NTS) process is the tempered stable (TS) process TS $(\kappa, \delta, \gamma)$ , for $ \kappa \in (0,1), \ \delta > 0, \ \gamma \geq 0$ , which has a Lévy density
where $A = \delta\kappa 2^{\kappa}\Gamma^{-1}(1-\kappa)$ and $\Gamma^{-1}(\cdot)$ is the reciprocal of the gamma function. Here,
since $\kappa\in(0,1)$ . Therefore, in view of Corollary 4.1 and Theorem 4.1, the residuals are expected to be approximately Gaussian. Moreover, in this case we can derive a bound on the corresponding marginal convergence rate.
Lemma 6.1. For $\epsilon\in(0,1)$ , let $(Y_{\epsilon}(t))$ denote the standardised truncated process associated to an NVM process subordinated to the residual TS process ${\textrm{TS}}(\kappa, \delta, \gamma)$ . If $B\sim\mathcal{N}(0,1)$ , then the Kolmogorov distance $E_\epsilon$ between $Y_{\epsilon}(1)$ and B satisfies
where
and $\Phi(a, b;\, m), \gamma(s,x)$ are the Kummer confluent hypergeometric function and the incomplete lower gamma function, respectively. Furthermore, as $\epsilon\rightarrow 0$ we have
Proof. From Theorem 5.1 we obtain $E_{\epsilon} \leq C \sigma_{W}^{3}\Phi_{\epsilon}M^{\big(\frac{3}{2}\big)}_{Z_\epsilon}/\sigma_{\epsilon}^{3}$ , and from (2.9) it follows that
where the residual first moment is given by
To find $M_{Z_{\epsilon}}^{\left(\frac{3}{2}\right)}$ , substitute the Lévy density of the TS process described earlier,
Substituting this in (5.5) and noting that $A = \delta \kappa 2^{\kappa}\Gamma^{-1}(1-\kappa)$ and $C = 0.7975 \times 2\sqrt{2/\pi}$ yields
as claimed. Finally, recalling the series expansion for $\gamma(s,x)$ from the previous section, the asymptotic expression in (5.2) leads to (6.2).
Next we examine the empirical accuracy of the Gaussian approximation in this case. Figures 6 and 7 show the empirical distribution of the residual NTS process at $t=1$ , with parameter values $\mu=0$ , $\mu_W=\delta=1$ , $\sigma_W=2$ , $\kappa=1/2$ , $\gamma=1.35$ , and truncation level $\epsilon=10^{-6}$ .
With the same parameter values, Figure 8 shows the behaviour of the bound in (6.1) and the first term in the asymptotic bound (6.2).
6.3. Generalised hyperbolic process
Finally, we consider the general class of generalised hyperbolic (GH) processes. The subordinator in this case is the generalised inverse Gaussian (GIG) process GIG( $\lambda, \delta, \gamma$ ), with constraints on parameter values as detailed in [Reference Godsill and Kndap22], and with Lévy density given by [Reference Godsill and Kndap22]
where $H_{v}(z)$ is the Hankel function of real order v. A direct verification of the sufficient condition in Corollary 4.1 is readily obtained as
since ${y\big|H_{\nu}(y)\big|^{2}}$ is non-zero for $z\in [z_1,\infty)$ where
see e.g. [Reference Godsill and Kndap22, Theorem 2] and [Reference Watson45]. Once again, in view of Corollary 4.1 and Theorem 4.1, the residuals are expected to be approximately Gaussian. Indeed, in this case we can derive the following bound on the corresponding marginal convergence rate.
Lemma 6.2. For $\epsilon\in(0,1)$ , let $Y_{\epsilon}(t)$ denote the standardised truncated process associated to an NVM process subordinated to the residual ${\textrm{GIG}}(\lambda, \delta, \gamma)$ process. If $B\sim\mathcal{N}(0,1)$ , then for any $z_0\in(0,\infty)$ , the Kolmogorov distance $E_\epsilon$ between $Y_{\epsilon}(1)$ and B can be bounded as
with $b = \frac{\gamma^{2}}{2}$ , $\tilde{\pi} = \sqrt{\frac{\pi}{2}}$ , $H_{0} = z_{0}\big|H_{|\lambda|}(z_{0})\big|^{2},$ and
where, using the standard definitions ${{\textrm{erf}}} (x) = \frac{2}{\sqrt\pi}\int_0^x e^{-t^2}\,\textrm{dt}$ and ${{\textrm{erfc}}} (x)=1-{{\textrm{erf}}} (x)$ , we denote by $\Phi(a, b;\, m)$ , $\gamma(s,x)$ , ${\textrm{erf}}(x)$ , and ${\textrm{erfc}}(x)$ the Kummer confluent hypergeometric function, the incomplete lower gamma function, the error function, and the complementary error function, respectively. Furthermore, as $\epsilon\rightarrow 0$ we have
Proof. Recalling the definition of the Jaeger integral as
we have from the bounds obtained in Appendix A, for $|\lambda|\leq 1/2$ ,
and for $|\lambda|>1/2$ ,
Noting that the variance of the GH process satisfies $\sigma_{\epsilon}^{2} = \mu_{W}^{2}M^{(2)}_{Z_{\epsilon}} + \sigma_{W}^{2}M^{(1)}_{Z_{\epsilon}} \geq \sigma_{W}^{2}M^{(1)}_{Z_{\epsilon}}$ , for $|\lambda|\leq \frac{1}{2}$ we obtain
and similarly for $|\lambda|>\frac{1}{2}$ we obtain
Therefore, we have the following bound on the variance:
Finding $M_{Z_{\epsilon}}^{\left(\frac{3}{2}\right)}$ in line with Theorem 5.1, we have
Writing $b = \frac{\gamma^{2}}{2}$ , we have
for $|\lambda| > \frac{1}{2}$ we have ${J(z)\leq \delta{\left(\frac{\pi}{2}\right)}^{3/2}z^{-\frac{1}{2}}}$ and
and for $|\lambda| \leq \frac{1}{2}$ ,
Combining the above bounds, for $|\lambda|\leq 1/2$ ,
and for $|\lambda|>1/2,$
Finally, substituting these bounds for $M_{Z_{\epsilon}}^{\left(\frac{3}{2}\right)}$ into (5.1), we obtain the bounds as stated in (6.3). The series expansions of the gamma and hypergeometric functions then lead to the asymptotic expansion (6.4).
Once again, we examine the validity of the Gaussian approximation empirically. Figures 9 and 10 show the empirical distribution of the residual GH process at time $t = 1$ , with parameter values $\mu = 0$ , $\mu_W =1$ , $\sigma_W = 2$ , $\delta=1.3$ , $\gamma=\sqrt{2}$ , $\lambda=0.2$ , and truncation level $\epsilon=10^{-6}$ .
With the same parameter values, Figure 11 shows behaviour of the bound in (6.3) and the first term in the asymptotic bound (6.4).
Note that the bounds (6.3) and (6.4) are discontinuous at $|\lambda| = \frac{1}{2}$ . This discrepancy is likely due to the upper bound for $M_{Z_{\epsilon}}^{(1)}$ in the case $|\lambda| > \frac{1}{2}$ . Although we do expect this could be improved, obtaining such refined bounds is beyond the scope of this paper.
7. Linear SDEs
The Lévy state space model [Reference Godsill, Riabiz and Kontoyiannis23] defines a stochastic process having the following dynamics:
where $\boldsymbol{A}$ is a $P\times P$ matrix, and $\boldsymbol{h}\in\mathbb{R}^P$ . In [Reference Godsill, Riabiz and Kontoyiannis23] (W(t)) is assumed to follow a stable law; here it is taken to be an NVM Lévy process. The solution of the state process takes the form
We first present a shot-noise representation of the stochastic integral in (7.1), and then prove the convergence of its small-jump residual to a Gaussian-driven SDE, under appropriate conditions.
7.1. Shot-noise representation of SDE
In order to apply the Lévy state space model to NVM Lévy processes, we first establish their representation as generalised shot-noise series. Theorem 7.1 gives the result for a general integrand.
Theorem 7.1. Let (X(u)) be an NVM process generating the filtration $(\mathcal{F}_{t})$ . Suppose $\boldsymbol{f}_{t} : [0, \infty) \rightarrow \mathbb{R}^{P}$ is an $L_{2}$ deterministic function, and let $\boldsymbol{I}({\textbf{f}_{t}})$ denote the integral
Then $\boldsymbol{I}({\boldsymbol{f}_{t}})$ admits the series representation
where
and $\mu_{W} \in \mathbb{R}, \ \sigma_{W} \in (0,\infty)$ are the variance-mean mixture parameters, $V_{i} \overset{i.i.d.}{\sim} \mathcal{U}[0,1]$ are normalised jump times, $Z_i$ are the jumps of the subordinator process arranged in non-increasing order, and $U_i \overset{i.i.d.}{\sim} \mathcal{N}(0, 1)$ .
Proof. Arguing as in [Reference Rosinski39, Section 7], we can extend (2.14) to any NVM process defined on $t\in[0, T]$ , with $\tilde{V}_i = TV_i\sim\mathcal{U}[0,T]$ , to obtain, for all $u\in[0,T]$ ,
where M(u) is a subordinator Lévy process and S(u) a symmetric Gaussian mixture process. Therefore,
and hence we obtain the following representation for $\boldsymbol{I}({\boldsymbol{f}_{t}})$ :
From [Reference Barndorff-Nielsen and Shephard3, Corollary 8.2], a stochastic integral with respect to a Lévy subordinator admits the a.s. generalised shot-noise representation
Similarly, Rosinski proves in [Reference Rosinski38, Section 4] the a.s. representation of stochastic integrals with respect to Lévy processes of type G. These are symmetric normal variance mixture processes such as the symmetric Student-t, symmetric $\alpha$ -stable, and Laplace which are special cases of NVM Lévy processes having $\mu_W=0$ , in the form
Combining the last three expressions proves the claimed result.
Applying the result of the theorem to X(t) in (7.1), with $\boldsymbol{f}_{t}(u) = e^{\boldsymbol{A}(t-u)}\boldsymbol{h}\mathbb{1}(s\leq u \leq t),$ yields
with $\tilde{V}_{i} \overset{i.i.d.}{\sim} \mathcal{U}(0,T]$ as before.
8. Convergence of residual SDE to a Gaussian SDE
In this section we prove that the residual series of the truncated shot-noise representation of the SDE in the previous section converges to Brownian-motion-driven SDE, as the truncation level $\epsilon\downarrow 0$ . Employing random Poisson truncations of the subordinator jumps $Z_i$ as before, we can write
where ${\boldsymbol{I}(\boldsymbol{f}_{\boldsymbol{t}})} = \boldsymbol{Z}_{\epsilon}(s,t) + \boldsymbol{R}_{\epsilon}(s,t)$ , with
and $\tilde{V}_{i} \sim {\mathcal{U}(0,T]} $ . Here, $\boldsymbol{R}_{\epsilon}(s,t)$ is the residual series driven by small jumps $Z_i < \epsilon$ , which will be approximated by a Brownian-driven SDE with matched moments. Theorem 4.1 and the results in Section 6 are the starting point of the proof of this approximation.
We first present a lemma concerning NVM Lévy processes.
Lemma 8.1. Let $(Y_{\epsilon}(t))$ denote the standardised residual NVM Lévy process as defined in Theorem 4.1, and let (B(t)) be a standard Brownian motion. Then there exists a coupling $(Y_{\epsilon}(t),B(t))$ such that, under the conditions of Theorem 4.1, we have
where $C_{T}$ is a constant independent of $\epsilon$ and
Proof. The process $(Y_{\epsilon}(t))$ is square-integrable, with drift parameter and diffusion coefficient $a=b=0$ . Write $Q_{X}$ for its Lévy measure, and define
The corresponding subordinator has Lévy measure $Q_{Z}$ , satisfying $Q_{Z}(\{z > \epsilon\}) = 0$ . Note that $m_{2}(Q_{X}) = 1$ . Following Theorem 4.1, the subordinator has no drift or diffusion component, so neither does $(Y_{\epsilon}(t))$ . From [Reference Fournier19, Theorem 3.1], if $m_{4}(Q_{X}) < \infty$ , we have directly
where $C_{T}$ depends on T only. Finally, writing $\sigma_{\epsilon}^{2}$ for the variance of the NVM residual process, we can express $m_{4}(Q_{X})$ as
Substituting this into the earlier bound yields the required result.
Theorem 8.1. Let $(X_{\epsilon}(t))$ denote the residual NVM Lévy process as in Theorem 4.1, and write $(\mathcal{F}_{t})$ for the filtration it generates. Let $\boldsymbol{f}_{t}(u)$ denote some deterministic, $L_{2}$ -integrable function, and define the process $(\boldsymbol{R}_{\epsilon}(t))$ via
If $(X_{\epsilon}(t))$ satisfies the conditions of Theorem 4.1, then $(\boldsymbol{Y}_{\epsilon}(t))$ , the $\epsilon$ -normalised version of $(\boldsymbol{R}_{\epsilon}(t))$ , satisfies
where, centring and normalising $(X_{\epsilon}(t))$ to obtain $(Y_{\epsilon}(t))$ as before, we define
and also a moment-matched Gaussian process $(\boldsymbol{B}(t))$ driven by standard Brownian motion B(t) that can be coupled with $(Y_{\epsilon}(t))$ in the sense of Lemma 8.1,
In particular, for each fixed time t, $\boldsymbol{Y}_{\epsilon}(t)$ converges in distribution to the Gaussian law
Proof. Consider the SDE
with solution given by
Since $\left\|\boldsymbol{f}_{t}\right\|_{L_{2}}<\infty$ and $(X_{\epsilon}(t))$ is a semi-martingale, its quadratic variation is well defined; see Appendix B. Therefore, we can compute the mean and variance of $\boldsymbol{R}_{\epsilon}(t)$ as
Centring and normalising $(X_{\epsilon}(t))$ to obtain $(Y_{\epsilon}(t))$ as before, we now consider the process
along with a matching process driven by standard Brownian motion,
where
Letting $T(t) = Y_{\epsilon}(t)-B(t)$ , we have
so that
and hence, applying Lemma 8.1, there exist coupled processes $(Y_{\epsilon}(t))$ and (B(t)) such that
By our assumptions on $\boldsymbol{f}_{t}(u), u \in [0, t]$ , and under the condition (4.6) of Theorem 4.1, we have $\lim_{\epsilon \rightarrow 0} S_{\epsilon} = 0$ , and since $C_{T}$ is independent of $\epsilon$ ,
which completes the proof.
Building on the results in Sections 6.1, 6.2, and 6.3, Theorem 8.1 proves the Gaussian representation of $\boldsymbol{R}_{\epsilon}(t)$ when the underlying process is known to converge to a diffusion.
In Sections 6.2 and 6.3 we justified the Gaussian convergence of the NTS and the GH process residuals, respectively, and therefore Theorem 8.1 justifies the Gaussian representation of $(\boldsymbol{R}_{\epsilon}(t))$ . While an exact expression for $M^{(1)}_{Z_{\epsilon}}$ exists in the NTS case, we can only provide small-argument bounds for the GIG case; see Appendix D.
In Section 6.1 we showed that the NG residual process fails to converge to Brownian motion as $\epsilon\to 0$ . Therefore, there is no mathematically justified reason to model $\boldsymbol{R}_{\epsilon}(t)$ as a diffusion process there. While the large concentration of residual jumps near 0 shown in Figure 4 suggests it might be reasonable to set $\boldsymbol{R}_{\epsilon}(t) = 0$ , we could alternatively model the residual process via its mean, so that ${\hat{\boldsymbol{R}}}_{\epsilon}(t) = \mathbb{E}[{\hat{\boldsymbol{R}}}_{\epsilon}(t)]$ , as suggested in [Reference Asmussen and Rosinski2].
9. Conclusions
In this work, new theoretical properties of NVM processes were established, motivated by the desire to investigate the application of some of the Bayesian inference algorithms introduced in [Reference Godsill, Riabiz and Kontoyiannis23] to state space models driven by more general classes of non-Gaussian noise. We identified natural sufficient conditions that guarantee the Gaussian convergence of the error process associated with an NVM process subjected to random Poisson truncations of their shot-noise series representations. We also showed that this Gaussian convergence does not always occur, and provided sufficient conditions for cases when it fails.
Moreover, under the same Gaussian-convergence conditions, we established the process-level convergence of a family of associated stochastic integrals, thus justifying the Gaussian representation of the residuals of these integrals. In Section 6 we showed that Brownian motion with drift subordinated to the residual processes of a TS or a GIG-type Lévy process converges to a Wiener process. Furthermore, in Section 7.1 we established the validity of the Gaussian representation of the residual of the stochastic integral with respect to the NTS and GH processes (excluding the NG edge case).
Subordination to a gamma process is shown not to converge to a Gaussian limit. Therefore, the residuals of stochastic integrals with respect to NG processes cannot be represented by Gaussians. One alternative direction would be to investigate whether fitting a Gaussian to the residual would still yield improved accuracy in Bayesian inference procedures such as particle filtering. A more interesting possibility would be to explore the distribution to which the NG residual converges, as, for any $\epsilon = h(\tilde{\epsilon}) > 0$ , in Section 6.1 we showed that the residual is non-zero and heavier-tailed than the Gaussian.
Finally, we note that the current analysis applies to one-dimensional Lévy processes and the corresponding linear SDEs. An extension of the methodology to multivariate Lévy processes would allow the generalisation of the ‘Lévy state space model’ [Reference Godsill, Riabiz and Kontoyiannis23] and the associated methodological tools to more general state space models, importantly allowing for more sophisticated modelling of spatial dependencies in high-dimensional data.
Appendix A. Upper and lower bounds on Jaeger integrals
Define first the Jaeger integral [Reference Freitas20] as parametrised in [Reference Godsill and Kndap22, Section 3]:
The integrand in the GIG Lévy measure in (2.12) depends on the value of $|\lambda|$ [Reference Godsill and Kndap22]. We first consider the region $|\lambda| \leq \frac{1}{2}$ . From [Reference Godsill and Kndap22, Theorem 3], a suitable upper bound on $\left[{z|H_{|\lambda|}(z)|^{2}}\right]^{-1}$ for $|\lambda| \leq \frac{1}{2}$ is given by
where
with $z_{0} \in (0, \infty)$ and $H_{0} = z_{0}|H_{|\lambda|}(z_{0})|^{2}$ . This leads to the upper bound
Recall the series expansion for $\gamma(s,x)$ from Section 6.1 and also that for $\Gamma(s,x)$ we have [Reference Bateman and Erdélyi6], when $|\lambda| \neq 0$ ,
Combining these with the previous bound, we have
Finally, we refer to [Reference Godsill and Kndap22, Theorem 1] for a lower bound on the Jaeger integral for $|\lambda|\leq \frac{1}{2}$ :
When $|\lambda| > \frac{1}{2}$ , $y|H_{|\lambda|}(y)|^{2}$ is non-increasing rather than non-decreasing, and the relevant bounds become
From [Reference Godsill and Kndap22, Theorem 3], a suitable lower bound on $z|H_{|\lambda|}(z)|^{2}]^{-1}$ for $|\lambda| \geq \frac{1}{2}$ is given by
where B(z) is defined as above, from which we deduce that for small $z \in [0, \epsilon]$ and $|\lambda| \leq \frac{1}{2}$ ,
Similarly, if $|\lambda|\geq \frac{1}{2}$ ,
Appendix B. Derivation of SDE moments
Recall the definition of $\boldsymbol{R}_{\epsilon}(t)$ as
Since $(X_{t}^{\epsilon})$ is a semi-martingale, it is clear that $\tilde{X}_{t}^{\epsilon} = X_{t}^{\epsilon} - t\mu_{W}M_{Z_{\epsilon}}^{(1)}$ is a martingale, and
Then, clearly, we have
and the Itô isometry yields
where $\big[\tilde{X}^{\epsilon}\big]_{t}$ is the quadratic variation of the compensated Lévy process, with expectation [Reference Cont and Tankov12]
Finally, the corresponding expression for the variance of $\boldsymbol{R}_{\epsilon}(t)$ is also easily obtained as
Appendix C. Centring for NVM processes
The following lemma is probably known, but as we could not easily locate a specific reference, we provide a proof for completeness. Recall the convergent generalised shot-noise representation of a Lévy process from Section 2.4:
Lemma C.1. For an NVM Lévy process (X(t)), the following generalised shot-noise representation converges a.s.:
Proof. Recall the form of the Lévy measure of NVM processes from (2.6). An NVM process need not be compensated, and hence we may take $b_i=0$ for all i in the shot-noise representation, provided
Since by the Lévy measure definition $\int_{-\infty}^{\infty}\left(1\wedge x^{2}\right)Q_{X}(dx)$ is finite, we must also have that $\int_{\left\{|x|>1\right\}}Q_{X}(dx)$ is finite. So, concentrating on the interval $|x|\leq1$ , we have
Note that the unimodal function $z^{-\frac{3}{2}}\exp\big\{-x^2/\big(2z\sigma_{W}^{2}\big)\big\}$ achieves its maximum at $z=\frac{x^{2}}{3\sigma_{W}^{2}}$ . Hence, the inner integrand may be bounded, for all $z\in (0,1]$ , by
with the $|x|\leq{3\sigma_{W}^{2}}$ case corresponding to the supremum lying to the right of $z=1$ . Therefore, the inner integral in (C.2) over $z\in(0,1)$ can be bounded, using (2.3), by
where $C=\int_0^1zQ_Z(dz)<\infty$ is a constant that does not depend on x or $\epsilon$ . Moreover, using (2.3) again, we obtain
Combining the above bounds yields
establishing (C.1) and confirming that compensation is not required for NVM Lévy processes.
Given that I is finite, the result of the lemma will follow from [Reference Rosinski39, Theorem 4.1], once we establish that, with
the limit $a\,:\!=\,\lim_{s\to\infty} A(s)$ exists and is finite. In the definition of A(s), the term $\sigma(\cdot;\,\cdot)$ denotes the kernel, $\sigma(h(r);\,F)={\mathbb P}(H(h(r),U)\in F)$ for all measurable F, which in the present setting is a collection of Gaussian measures. Since $r=Q_Z([z,+\infty ))$ , we have $dr=-Q_Z(dz)$ , and hence
Since we already showed that
we can exchange the order of integration, and since $h(s)\to 0$ as $s\to\infty$ by definition, it follows that the limit $a\,:\!=\,\lim_{s\to\infty} A(s)$ exists. Finally, we have
showing that a is finite and concluding the proof.
Appendix D. Moments for example processes
Recalling the discussion at the end of Section 8 in connection with the ‘Lévy state space model’ [Reference Godsill, Riabiz and Kontoyiannis23], we provide here expressions for $M_{Z_{\epsilon}}^{(1)}$ and $M_{Z_{\epsilon}}^{(2)}$ for our example processes.
For the NG process in Section 6.1, we have
and
Similarly, for the NTS process in Section 6.2,
and
The intractability of the Jaeger integral prohibits the derivation of an analytical expression for the moments of the GH process. However, for sufficiently small truncation levels $\epsilon$ , the use of asymptotic moment expansions provides useful approximations. For now, we restrict our analysis to the parameter range $|\lambda| \leq \frac{1}{2}$ ; the range $|\lambda| \geq \frac{1}{2}$ yields similar results.
To obtain a lower bound on the expected value of the subordinator jumps, $M_{Z_{\epsilon}}^{(1)}$ , we use the bound $\frac{1}{z|H_{|\lambda|}(z)|^{2}}\geq \frac{\pi}{2}$ :
Using the expansion for the erf function in Section 6.1 and the series expansion of the exponential function yields the lower bound
which, for $\epsilon\to 0$ , is equal to
A corresponding upper bound can be derived using the bound in Appendix A. We have
where, as $\epsilon\to 0$ , the last expression is equal to
Equations (D.1)–(D.2) imply that, for small $\epsilon$ , $M_{Z_\epsilon}^{(1)}$ is approximately bounded above and below by
and we can therefore conclude that $M_{Z_{\epsilon}}^{(1)} \sim \sqrt{\epsilon}$ as $\epsilon\to 0$ , for $\delta \neq 0$ and $|\lambda| \leq \frac{1}{2}$ .
To characterise the behaviour of $M_{Z_{\epsilon}}^{(2)}$ , we again use the bound $\frac{1}{z|H_{|\lambda|}(z)|^{2}}\geq \frac{\pi}{2}$ , which gives
where, as $\epsilon\to 0$ , the last expression above is
For the corresponding upper bound we similarly have
where the last expression above, for $\epsilon\to 0$ , is equal to
From (D.4) and (D.3) we have that, for small $\epsilon$ , $M_{Z_\epsilon}^{(2)}$ is approximately bounded above and below by
and therefore, $M_{Z_{\epsilon}}^{(2)} \sim \epsilon\sqrt{\epsilon}$ , for all $ \delta \neq 0, \left|\lambda\right|\leq \frac{1}{2}\ $ .
Finally, in the case $\left|\lambda\right|\geq \frac{1}{2}$ , the upper and lower bounds on $\frac{1}{z|H_{|\lambda|}(z)|^{2}}$ are reversed, and hence so are the bounds in (D.1), (D.2), (D.3), (D.4), so that $M_{Z_{\epsilon}}^{(1)} \sim \sqrt{\epsilon}$ and $M_{Z_{\epsilon}}^{(2)} \sim \epsilon\sqrt{\epsilon},$ for all $\delta \neq 0$ .
Acknowledgements
We wish to thank the anonymous reviewers for their careful reading of our original submission and their useful comments. In particular, we thank Reviewer I for suggesting the direct verification of the requirement (III) from [Reference Pollard36, Theorem V.19] in Theorem 4.1. We would also like to thank Professor Nicolas Fournier for some helpful discussion regarding Theorem 8.1.
Funding information
There are no funding bodies to thank in relation to the creation of this article.
Competing interests
There were no competing interests to declare which arose during the preparation or publication process of this article.
Data
The code used for all the simulations in this paper can be found at https://github.com/marcostapiac/PyLevy.