Theory and Modern Applications

# Weak order in averaging principle for stochastic differential equations with jumps

## Abstract

In this paper, we deal with the averaging principle for a two-time-scale system of jump-diffusion stochastic differential equations. Under suitable conditions, we expand the weak error in powers of timescale parameter. We prove that the rate of weak convergence to the averaged dynamics is of order 1. This reveals that the rate of weak convergence is essentially twice that of strong convergence.

## Introduction

We consider a two-time-scale system of jump-diffusion stochastic differential equation of the form

\begin{aligned}& dX^{\epsilon }_{t}=a \bigl(X_{t}^{\epsilon }, Y_{t}^{\epsilon } \bigr)\,dt+b \bigl(X_{t} ^{\epsilon } \bigr)\,d {B}_{t}+c \bigl(X_{t-}^{\epsilon } \bigr)\,d {P}_{t}, X_{0}^{ \epsilon }=x, \end{aligned}
(1.1)
\begin{aligned}& dY^{\epsilon }_{t}=\frac{1}{\epsilon }f \bigl(X_{t}^{\epsilon }, Y_{t}^{ \epsilon } \bigr)\,dt+\frac{1}{\sqrt{\epsilon }}g \bigl(X_{t}^{\epsilon }, Y_{t} ^{\epsilon } \bigr)\,d {W}_{t}+h \bigl(X_{t-}^{\epsilon },Y_{t-}^{\epsilon } \bigr)\,d {N}^{\epsilon }_{t},\qquad Y_{0}^{\epsilon }=y, \end{aligned}
(1.2)

where $$X_{t}^{\epsilon }\in \mathbb{R}^{n}$$, $$Y_{t}^{\epsilon }\in \mathbb{R}^{m}$$, the drift functions $$a(x, y)\in \mathbb{R}^{n}$$, $$f(x,y) \in \mathbb{R}^{m}$$, the diffusion functions $$b(x)\in \mathbb{R}^{n \times d_{1}}$$, $$c(x)\in \mathbb{R}^{n}$$, $$g(x,y)\in \mathbb{R}^{m\times d_{2}}$$, and $$h(x,y)\in \mathbb{R}^{m}$$, $$B_{t}$$ and $$W_{t}$$ are $$d_{1}$$- and $$d_{2}$$-dimensional independent Brownian motions on a complete stochastic base $$(\Omega, \mathcal{F},\mathcal{F}_{t}, \mathbb{P})$$, respectively, $$P_{t}$$ is a scalar Poisson process with intensity $$\lambda_{1}$$, and $$N_{t}^{\epsilon }$$ is a scalar Poisson process with intensity $$\frac{\lambda_{2}}{\epsilon }$$. The positive parameter ϵ is small and describes the ratio of time scales between $$X^{\epsilon }_{t}$$ and $$Y^{\epsilon }_{t}$$. Systems (1.1)–(1.2) with two time scales occur frequently in applications, including chemical kinetics, signal processing, complex fluids, and financial engineering.

With the separation of time scale, we can view the state variable of the system as being divided into two parts, the “slow” variable $$X^{\epsilon }_{t}$$ and the “fast” variable $$Y^{\epsilon }_{t}$$. It is often the case that we are interested only in the dynamics of the slow component. Then a simplified equation, which is independent of the fast variable and possesses the essential features of the system, is highly desirable. Such a simplified equation is often constructed by averaging procedure as in [2, 20] for deterministic ordinary differential equations and in the further development [7, 8, 1316, 18, 19, 25] for stochastic differential equations with continuous Gaussian processes. As far as averaging for stochastic dynamical systems in infinite-dimensional space is concerned, it is worth quoting the important works [46, 26] and also the works [9, 10, 21]. For related works on averaging for multivalued stochastic differential equations, we refer the reader to [12, 22].

To derive the averaged dynamics of system (1.1)–(1.2), we introduce the fast motion equation with a frozen slow component $$x\in \mathbb{R}^{n}$$ of the form

\begin{aligned}& dY_{t}^{x}= f \bigl(x, Y_{t}^{x} \bigr)\,dt+g \bigl(x, Y_{t}^{x} \bigr)\,d {W}_{t}+h \bigl(x,Y^{x}_{t-} \bigr)\,d {N}_{t},\qquad Y_{0}^{x}=y, \end{aligned}
(1.3)

and denote its solution by $$Y_{t}^{\epsilon }(y)$$. Under suitable conditions on f, g, and h, $$Y_{t}^{\epsilon }(y)$$ induces a unique invariant measure $$\mu^{x}(dy)$$ on $$\mathbb{R}^{m}$$, which is ergodic and ensures the averaged equation

\begin{aligned}& d\bar{X}_{t}=\bar{a}(\bar{X}_{t})\,dt+b( \bar{X}_{t})\,d{B}_{t}+c(\bar{X} _{t-})\,d{P}_{t},\qquad \bar{X}_{0} =x, \end{aligned}

where the averaging nonlinearity is defined by setting

\begin{aligned} \bar{a}(x) =& \int_{\mathbb{R}^{m}}a(x,y)\mu^{x}(dy) \\ =&\lim_{t\rightarrow +\infty } \mathbb{E} a \bigl(x, Y_{t}^{x}(y) \bigr). \end{aligned}

In , it was shown that, under the stated conditions, the slow motion $$X^{\epsilon }_{t}$$ converges strongly to the solution $$\bar{X}_{t}$$ of the averaged equation with jumps. The order of convergence $$\frac{1}{2}$$ in the strong sense was provided in . To our best knowledge, there is no literature addressing the weak order in averaging principle for jump-diffusion stochastic differential systems. In fact, it is fair to say that the weak convergence in stochastic averaging theory of systems driven by jump noise is not fully developed yet, although some strong approximation results on the rate of strong convergence were obtained [1, 23, 24].

Therefore, in this paper, we aim to study this problem. Here we are interested in the rate of weak convergence of the averaging dynamics to the true solution of slow motion $$X^{\epsilon }_{t}$$. In other word, we will determine the order, with respect to timescale parameter ϵ, of weak deviation between original solution to slow equation and the solution of the corresponding averaged equation. The main technique we adapted is finding an expansion with respect to ϵ of the solutions of the Kolmogorov equations associated with the jump diffusion system. The solvability of the Poisson equation associated with the generator of frozen equation provides an expression for the coefficients of the expansion. As a result, the boundedness for the coefficients of expansion can be proved by smoothing effect of the corresponding transition semigroup in the space of bounded and uniformly continuous functions, where some regular conditions on the drift and diffusion terms are needed.

Our result shows that the weak convergence rate is 1 even when there are jump components in the system. It is the main contribution of this work. We would like to stress that an asymptotic method was first applied by Bréhier  to an averaging result for stochastic reaction–diffusion equations in the case of Gaussian noise of additive type, which was included only in the fast motion. However, the extension of this argument is not straightforward. The method used in the proof of weak order in  is strictly related to the differentiability in time of averaged process. Therefore, once the noise is introduced in the slow equation, difficulties arise, and the procedure becomes more complicated. Our result in this paper bridges such a gap, in which the slow and fast motions are both perturbed by noise with jumps.

The rest of the paper is structured as follows. Section 2 is devoted to notations and assumptions and summarizes preliminary results. The ergodicity of a fast process and the averaged dynamics of system with jumps is introduced in Sect. 3. Then the main result of this article, which is derived via the asymptotic expansions and uniform error estimates, is presented in Sect. 4. Finally, we give the Appendix.

It should be pointed out that in the whole paper the letter C with or without subscripts denotes generic positive constants independent of ϵ.

## Assumptions and preliminary results

For any integer d, the scalar product and norm on the d-dimensional Euclidean space $$\mathbb{R}^{d}$$ are denoted by $$(\cdot,\cdot)_{ \mathbb{R}^{d}}$$ and $$\|\cdot \|_{\mathbb{R}^{d}}$$, respectively. For any integer k, we denote by $$C_{b}^{k}(\mathbb{R}^{d},\mathbb{R})$$ the space of all k-times differentiable functions on $$\mathbb{R}^{d}$$ with bounded uniformly continuous derivatives up to the kth order.

In what follows, we assume that the drift and diffusion coefficients arising in the system fulfill the following conditions.

1. (A1)

The mappings $$a(x,y)$$, $$b(x)$$, $$c(x)$$, $$f(x,y)$$, $$g(x,y)$$, and $$h(x,y)$$ are of class $$C^{2}$$ and have bounded first and second derivatives. Moreover, we assume that $$a(x,y)$$, $$b(x)$$, and $$c(x)$$ are bounded.

2. (A2)

There exists a constant $$\alpha >0$$ such that, for any $$x\in \mathbb{R}^{n}$$ and $$y\in \mathbb{R}^{m}$$,

\begin{aligned}& y^{T}g(x,y)g^{T}(x,y)y\geq \alpha \Vert y\Vert _{\mathbb{R}^{m}}. \end{aligned}
3. (A3)

There exists a constant $$\beta >0$$ such that, for any $$y_{1}, y_{2}\in \mathbb{R}^{m}$$ and $$x\in \mathbb{R}^{n}$$,

\begin{aligned}& \bigl(y_{1}-y_{2}, f(x,y_{1})-f(x,y_{2})+ \lambda_{2} \bigl(h(x,y_{1})-h(x,y _{2}) \bigr) \bigr)_{\mathbb{R}^{m}} \\& \qquad {}+ \bigl\Vert g(x,y_{1})-g(x,y_{2}) \bigr\Vert ^{2}_{\mathbb{R}^{m}}+\lambda_{2} \bigl\vert h(x,y_{1})-h(x,y_{2}) \bigr\vert ^{2} \\& \quad \leq -\beta \Vert y_{1}-y_{2}\Vert ^{2}_{\mathbb{R}^{m}}. \end{aligned}

### Remark 2.1

Notice that from (A1) it immediately follows that the following directional derivatives exist and are controlled:

\begin{aligned}& \bigl\Vert D_{x}a(x,y)\cdot k_{1} \bigr\Vert _{\mathbb{R}^{n}}\leq L\Vert k_{1}\Vert _{ \mathbb{R}^{n}}, \\& \bigl\Vert D_{y}a(x,y)\cdot l_{1} \bigr\Vert _{\mathbb{R}^{n}}\leq L\Vert l_{1}\Vert _{ \mathbb{R}^{m}}, \\& \bigl\Vert D_{xx}^{2}a(x,y)\cdot (k_{1},k_{2}) \bigr\Vert _{\mathbb{R}^{n}}\leq L\Vert k_{1}\Vert _{\mathbb{R}^{n}} \Vert k_{2}\Vert _{\mathbb{R}^{n}}, \\& \bigl\Vert D_{yy}^{2}a(x,y)\cdot (l_{1},l_{2}) \bigr\Vert _{\mathbb{R}^{n}}\leq L\Vert l_{1}\Vert _{\mathbb{R}^{m}} \Vert l_{2}\Vert _{\mathbb{R}^{m}}, \end{aligned}

where L is a constant independent of x, y, $$k_{1}$$, $$k_{2}$$, $$l_{1}$$, and $$l_{2}$$. For the differentiability of mappings b, c, f, g, and h, we possess similar results. For example, we have

\begin{aligned}& \bigl\Vert D^{2}_{xx}b(x)\cdot (k_{1},k_{2}) \bigr\Vert _{\mathbb{R}^{n}}\leq L \Vert k_{1}\Vert _{\mathbb{R}^{n}} \Vert k_{2}\Vert _{\mathbb{R}^{n}},\quad k_{1},k_{2} \in \mathbb{R}^{n}, \\& \bigl\Vert D^{2}_{yy}f(x,y)\cdot (l_{1},l_{2}) \bigr\Vert _{\mathbb{R}^{m}}\leq L \Vert l_{1}\Vert _{\mathbb{R}^{m}} \Vert l_{2}\Vert _{\mathbb{R}^{m}}\Vert,\quad l_{1},l_{2} \in \mathbb{R}^{m}. \end{aligned}

As far as assumption (A2) is concerned, it is a sort of nondegeneracy condition, which we assume in order to have the regularizing effect of the Markov transition semigroup associated with the fast dynamics. Assumption (A3) is the dissipative condition, which determines how the fast equation converges to its equilibrium state.

As assumption (A1) holds, for any $$\epsilon >0$$ and any initial conditions $$x\in \mathbb{R}^{n}$$ and $$y\in \mathbb{R}^{m}$$, system (1.1)–(1.2) admits a unique solution, which, to emphasize the dependence on the initial data, is denoted by $$(X_{t}^{\epsilon }(x,y), Y_{t}^{\epsilon }(x,y))$$. Moreover, we have the following lemma (for a proof, see, e.g., ).

### Lemma 2.1

Under assumptions (A1), (A2), and (A3), for any $$x\in \mathbb{R}^{n}$$, $$y\in \mathbb{R}^{m}$$, and $$\epsilon >0$$, we have

\begin{aligned}& \mathbb{E} \bigl\Vert X_{t}^{\epsilon }(x,y) \bigr\Vert ^{2}_{\mathbb{R}^{n}}\leq C_{T} \bigl(1+ \Vert x\Vert ^{2}_{\mathbb{R}^{n}}+\Vert y\Vert ^{2}_{\mathbb{R}^{m}} \bigr),\quad t\in [0, T], \end{aligned}
(2.1)

and

\begin{aligned}& \mathbb{E} \bigl\Vert Y_{t}^{\epsilon }(x,y) \bigr\Vert ^{2}_{\mathbb{R}^{n}}\leq C_{T} \bigl(1+ \Vert x\Vert ^{2}_{\mathbb{R}^{m}}+\Vert y\Vert ^{2}_{\mathbb{R}^{m}} \bigr),\quad t\in [0, T]. \end{aligned}
(2.2)

## Frozen equation and averaged equation

Fixing $$\epsilon =1$$, we consider the fast equation with frozen slow component $$x\in \mathbb{R}^{n}$$,

\begin{aligned}& \textstyle\begin{cases} dY_{t}^{x}(y)= f(x, Y_{t}^{x}(y))\,dt+g(x, Y_{t}^{x}(y))\,d {W}_{t}+h(x,Y ^{x}_{t-}(y))\,d {N}_{t}, \\ Y_{0}^{x}=y. \end{cases}\displaystyle \end{aligned}
(3.1)

Under assumptions (A1)–(A3), such a problem has a unique solution, which satisfies 

\begin{aligned}& \mathbb{E} \bigl\Vert Y_{t}^{x}(y) \bigr\Vert ^{2}_{\mathbb{R}^{m}}\leq C \bigl(1+\Vert x\Vert ^{2}_{ \mathbb{R}^{n}}+e^{-\beta t}\Vert y\Vert ^{2}_{\mathbb{R}^{m}} \bigr),\quad t\geq 0. \end{aligned}
(3.2)

Let $$Y_{t}^{x}( y')$$ be the solution of problem (3.1) with initial value $$Y_{0}^{x}=y'$$, the Itô formula implies that, for any $$t\geq 0$$,

\begin{aligned}& \mathbb{E} \bigl\Vert Y_{t}^{x}(y)-Y_{t}^{x} \bigl(y' \bigr) \bigr\Vert _{\mathbb{R}^{m}}^{2}\leq \bigl\Vert y-y' \bigr\Vert _{\mathbb{R}^{m}}^{2}e^{-\beta t}. \end{aligned}
(3.3)

Moreover, as discussed in  and , equation (3.1) admits a unique ergodic invariant measure $$\mu^{x}$$ satisfying

$$\int_{\mathbb{R}^{m}}\Vert y\Vert _{\mathbb{R}^{m}}^{2} \mu^{x}(dy)\leq C \bigl(1+\Vert x\Vert ^{2}_{\mathbb{R}^{n}} \bigr).$$
(3.4)

Then, by averaging the coefficient a with respect to the invariant measure $$\mu^{x}$$ we can define the $$\mathbb{R}^{n}$$-valued mapping

$$\bar{a}(x):= \int_{\mathbb{R}^{m}}a(x,y)\mu^{x}(dy), \quad x\in \mathbb{R} ^{n}.$$

Due to assumption (A1), it is easily to check that $$\bar{a}(x)$$ is twice differentiable with bounded derivatives, and hence it is Lipschitz-continuous:

$$\bigl\Vert \bar{a}(x_{1})-\bar{a}(x_{2}) \bigr\Vert _{\mathbb{R}^{n}}\leq C\Vert x_{1}-x_{2}\Vert _{\mathbb{R}^{n}},\quad x_{1}, x_{2}\in \mathbb{R}^{n}.$$

According to the invariance property of $$\mu^{x}$$, (3.4), and assumption (A1), we have

\begin{aligned} \bigl\Vert \mathbb{E}a \bigl(x, Y_{t}^{x}( y) \bigr)-\bar{a}(x) \bigr\Vert ^{2}_{ \mathbb{R}^{n}} =& \biggl\Vert \int_{\mathbb{R}^{m}}\mathbb{E} \bigl(a \bigl(x, Y_{t}^{x}(y) \bigr)-a \bigl(x, Y_{t}^{x}(z) \bigr) \bigr) \mu^{x}(dz) \biggr\Vert ^{2}_{\mathbb{R}^{n}} \\ \leq & \int_{\mathbb{R}^{m}}\mathbb{E} \bigl\Vert Y_{t}^{x} (y)-Y_{t}^{x}(z) \bigr\Vert _{\mathbb{R}^{m}}^{2} \mu^{x}(dz) \\ \leq &e^{-\beta t} \int_{\mathbb{R}^{m}}\Vert y-z\Vert ^{2}_{\mathbb{R}^{m}} \mu^{x}(dz) \\ \leq &Ce^{-\beta t} \bigl(1+\Vert x\Vert ^{2}_{\mathbb{R}^{n}}+ \Vert y\Vert ^{2}_{ \mathbb{R}^{m}} \bigr). \end{aligned}
(3.5)

Now we can introduce the effective dynamical system

\begin{aligned}& \textstyle\begin{cases} d\bar{X}_{t}(x)=\bar{a}(\bar{X}_{t}(x))\,dt+b(\bar{X}_{t}(x))\,d{B}_{t}+c( \bar{X}_{t-}(x))\,d {P}_{t}, \\ \bar{X}_{0} =x. \end{cases}\displaystyle \end{aligned}
(3.6)

As the coefficients ā, b, and c are Lipschitz-continuous, this equation admits a unique solution such that

\begin{aligned}& \mathbb{E} \bigl\Vert \bar{X}_{t}(x) \bigr\Vert ^{2}_{\mathbb{R}^{n}}\leq C_{T} \bigl(1+\Vert x\Vert ^{2}_{\mathbb{R}^{n}} \bigr),\quad t\in [0, T]. \end{aligned}
(3.7)

With these assumptions and notation, we have the following result, which is a direct consequence of Lemmas 4.1, 4.2, and 4.5.

### Theorem 3.1

Assume that $$x\in \mathbb{R}^{n}$$ and $$y\in \mathbb{R}^{m}$$. Then, under assumptions (A1), (A2), and (A3), for any $$T>0$$ and $$\phi \in C_{b}^{3}(\mathbb{R}^{n},\mathbb{R})$$, there exists a constant $$C_{T,\phi,x,y}$$ such that

\begin{aligned}& \bigl\vert \mathbb{E}\phi \bigl(X^{\epsilon }_{T}(x,y) \bigr)- \mathbb{E}\phi \bigl(\bar{X} _{T}(x) \bigr) \bigr\vert \leq C_{T,\phi,x,y}\epsilon. \end{aligned}

As a consequence, the weak order in the averaging principle for jump-diffusion stochastic systems is 1.

## Asymptotic expansion

Let $$\phi \in C_{b}^{3}(\mathbb{R}^{n}, \mathbb{R})$$ and define the function $$u^{\epsilon }(t, x,y):[0, T]\times \mathbb{R}^{n}\times \mathbb{R}^{m}\rightarrow \mathbb{R}$$ by

$$u^{\epsilon }(t, x,y)=\mathbb{E}\phi \bigl(X_{t}^{\epsilon }(x,y) \bigr).$$

We are now ready to seek an expansion formula for $$u^{\epsilon }(t, x,y)$$ with respect to ϵ of the form

\begin{aligned}& u^{\epsilon }(t,x,y)=u_{0}(t,x,y)+\epsilon u_{1}(t,x,y)+r^{\epsilon }(t,x,y), \end{aligned}
(4.1)

where $$u_{0}$$ and $$u_{1}$$ are smooth functions, which will be constructed further, and $$r^{\epsilon }$$ is the remainder term. To this end, let us recall the Kolmogorov operator corresponding to the slow motion equation, with a frozen fast component $$y\in \mathbb{R}^{m}$$, which is a second-order operator of the form

\begin{aligned} \mathcal{L}_{1}\Phi (x) =& \bigl(a(x,y),D_{x} \Phi (x) \bigr)_{\mathbb{R} ^{n}}+\frac{1}{2}\operatorname{Tr}\bigl[ D_{xx}^{2} \Phi (x) \cdot b(x) b^{T}(x) \bigr] \\ &{}+\lambda_{1} \bigl(\Phi \bigl(x+c(x) \bigr)-\Phi (x) \bigr),\quad \Phi \in C_{b}^{2} \bigl( \mathbb{R}^{n}, \mathbb{R} \bigr). \end{aligned}

For any frozen slow component $$x\in \mathbb{R}^{m}$$, the Kolmogorov operator for equation (3.1) is given by

\begin{aligned} \mathcal{L}_{2}\Psi (y) =& \bigl(f(x,y),D_{y} \Psi (y) \bigr)_{\mathbb{R} ^{m}}+\frac{1}{2}\operatorname{Tr}\bigl[ D_{yy}^{2} \Psi (y) \cdot g(x,y) g^{T}(x,y) \bigr] \\ &{}+\lambda_{2} \bigl(\Psi \bigl(y+h(x,y) \bigr)-\Psi (y) \bigr),\quad \Psi \in C_{b}^{2} \bigl( \mathbb{R}^{m}, \mathbb{R} \bigr). \end{aligned}

We set

$$\mathcal{L}^{\epsilon }:=\mathcal{L}_{1}+\frac{1}{\epsilon } \mathcal{L}_{2}.$$

It is known that $$u^{\epsilon }(t,x,y)$$ solves the equation

\begin{aligned}& \textstyle\begin{cases} \frac{\partial }{\partial t}u^{\epsilon }(t, x, y)=\mathcal{L}^{ \epsilon }u^{\epsilon }(t, x, y), \\ u^{\epsilon }(0, x,y)=\phi (x). \end{cases}\displaystyle \end{aligned}
(4.2)

Also, recall the Kolmogorov operator associated with the averaged equation (3.6) is defined as

\begin{aligned} \bar{\mathcal{L}} \Phi (x) =& \bigl(\bar{a}(x),D_{x} \Phi (x) \bigr)_{ \mathbb{R}^{n}}+\frac{1}{2}\operatorname{Tr}\bigl[ D_{xx}^{2} \Phi (x) \cdot b(x) b ^{T}(x) \bigr] \\ &{}+\lambda_{1} \bigl(\Phi \bigl(x+c(x) \bigr)-\Phi (x) \bigr),\quad \Phi \in C_{b}^{2} \bigl( \mathbb{R}^{n}, \mathbb{R} \bigr). \end{aligned}

Setting

$$\bar{u}(t, x)=\mathbb{E}\phi \bigl(\bar{X}_{t}(x) \bigr),$$

we have

\begin{aligned}& \textstyle\begin{cases} \frac{\partial }{\partial t}\bar{u}(t, x)=\bar{\mathcal{L}} \bar{u}(t, x), \\ \bar{u}(0, x)=\phi (x). \end{cases}\displaystyle \end{aligned}
(4.3)

Let us begin with constructing the leading term. By substituting expansion (4.1) into (4.2) we see that

\begin{aligned} \frac{\partial u_{0}}{\partial t}+\epsilon \frac{\partial u_{1}}{ \partial t}+\frac{\partial r^{\epsilon }}{\partial t} =& \mathcal{L} _{1}u_{0}+\epsilon \mathcal{L}_{1}u_{1}+ \mathcal{L}_{1}r^{\epsilon } \\ &{}+\frac{1}{\epsilon }\mathcal{L}_{2}u_{0}+\mathcal{ \mathcal{L}}_{2}u _{1}+\frac{1}{\epsilon } \mathcal{L}_{2}r^{\epsilon }. \end{aligned}

By equating the powers of ϵ, we obtain the following system of equations:

\begin{aligned}& \mathcal{L}_{2}u_{0}=0, \end{aligned}
(4.4)
\begin{aligned}& \frac{\partial u_{0}}{\partial t}=\mathcal{L}_{1}u_{0}+\mathcal{L} _{2}u_{1}. \end{aligned}
(4.5)

According to (4.4), we can conclude that $$u_{0}$$ does not depend on y, that is,

$$u_{0}(t,x, y)=u_{0}(t,x).$$

We also impose the initial condition $$u_{0}(0,x)=\phi (x)$$. Noting that $$\mathcal{L}_{2}$$ is the generator of a Markov process defined by equation (3.1), which admits a unique invariant measure $$\mu^{x}$$, we have

\begin{aligned}& \int_{\mathbb{R}^{m}}\mathcal{L}_{2}u_{1}(t,x,y) \mu^{x}(dy)=0. \end{aligned}
(4.6)

Thanks to (4.5), this yields

\begin{aligned} \frac{\partial u_{0}}{\partial t}(t,x) =& \int_{\mathbb{R}^{m}}\frac{ \partial u_{0}}{\partial t}(t,x)\mu^{x}(dy) \\ =& \int_{\mathbb{R}^{m}}\mathcal{L}_{1}u_{0}(t,x) \mu^{x}(dy) \\ =& \int_{\mathbb{R}^{m}} \bigl(a(x,y), D_{x} u_{0}(t,x) \bigr)_{ \mathbb{R}^{n}}\mu^{x}(dy) \\ &{}+\frac{1}{2}\operatorname{Tr}\bigl[ D_{xx}^{2}u_{0}(t,x) \cdot b(x) b^{T}(x) \bigr] \\ &{}+\lambda_{1} \bigl(u_{0} \bigl(x+c(x) \bigr)-u_{0}(x) \bigr) \\ =&\bar{\mathcal{L}}u_{0}(t,x), \end{aligned}

so that $$u_{0}$$ and ū are described by the same evolutionary equation. By a uniqueness argument, we easily have the following lemma.

### Lemma 4.1

Under assumptions (A1), (A2), and (A3), for any $$x\in \mathbb{R}^{n}$$, $$y\in \mathbb{R}^{m}$$, and $$T>0$$, we have $$u_{0}(T,x,y)=\bar{u}(T,x)$$.

### Construction of $${u_{1}}$$

According to Lemma 4.1, (4.3), and (4.5), we get

\begin{aligned}& \bar{\mathcal{L}}\bar{u}=\mathcal{L}_{1}\bar{u}+\mathcal{L}_{2}u_{1}, \end{aligned}

which means that

\begin{aligned} \mathcal{L}_{2}u_{1}(t,x,y) =& \bigl( \bar{a}(x)-a(x,y),D_{x}\bar{u}(t,x) \bigr)_{\mathbb{R}^{n}} \\ :=&-\rho (t,x,y), \end{aligned}
(4.7)

where ρ is of class $$C^{2}$$ with respect to y, with uniformly bounded derivatives. Moreover, for any $$t\geq 0$$ and $$x\in {\mathbb{R} ^{n}}$$, equality (4.6) guarantees that

\begin{aligned}& \int_{\mathbb{R}^{m}}\rho (t,x,y)\mu^{x}(dy)=0. \end{aligned}

For any $$y\in \mathbb{R}^{m}$$ and $$s>0$$, we have

\begin{aligned} \frac{\partial }{\partial s}\mathcal{P}_{s}\rho (t,x,y) =& \bigl(f(x,y),D _{y} \bigl[\mathcal{P}_{s}\rho (t,x,y) \bigr] \bigr)_{\mathbb{R}^{m}} +\frac{1}{2}\operatorname{Tr}\bigl[ D_{yy}^{2} \bigl[ \mathcal{P}_{s}\rho (t,x,y) \bigr]\cdot g(x,y) g^{T}(x,y) \bigr] \\ &{}+\lambda_{2} \bigl( \mathcal{P}_{s} \bigl[\rho \bigl(t,x,y+h(x,y) \bigr) \bigr]-\mathcal{P} _{s} \bigl[\rho (t,x,y) \bigr] \bigr), \end{aligned}
(4.8)

where

$$\mathcal{P}_{s} \bigl[\rho (t,x,y) \bigr]:=\mathbb{E}\rho \bigl(t, x,Y^{x}_{s}(y) \bigr).$$

Recalling that $$\mu^{x}$$ is the unique invariant measure corresponding to the Markov process $$Y^{x}_{t}(y)$$ defined by equation (3.1), from Lemma A.1 we infer that

\begin{aligned}& \biggl\vert \mathbb{E}\rho \bigl(t, x,Y^{x}_{s}(y) \bigr)- \int_{\mathbb{R}^{m}} \rho (t,x,z)\mu^{x}(dz) \biggr\vert \\& \quad = \biggl\vert \int_{\mathbb{R}^{m}}\mathbb{E} \bigl[\rho \bigl(t, x,Y^{x}_{s}(y) \bigr)- \rho \bigl(t, x,Y^{x}_{s}(z) \bigr) \bigr] \mu^{x}(dz) \biggr\vert \\& \quad \leq \int_{\mathbb{R}^{m}} \bigl\vert \mathbb{E} \bigl(a \bigl(x,Y^{x}_{s}(z) \bigr)- a \bigl(x,Y ^{x}_{s}(y) \bigr), D_{x} \bar{u}(t,x) \bigr)_{\mathbb{R}^{n}} \bigr\vert \mu^{x}(dz) \\& \quad \leq C \int_{\mathbb{R}^{m}} \mathbb{E} \bigl\Vert Y_{s}^{x}(z)-Y_{s}^{x}(y) \bigr\Vert _{\mathbb{R}^{n}} \mu^{x}(dz). \end{aligned}

Now it follows from (3.3) and (3.4) that

\begin{aligned}& \biggl\vert \mathbb{E}\rho \bigl(t, x,Y_{s}^{x}(y) \bigr)- \int_{\mathbb{R}^{m}} \rho (t,x,z)\mu^{x}(dz) \biggr\vert \leq C \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr)e^{-\frac{ \beta }{2}s}, \end{aligned}

which implies

$$\lim_{s\rightarrow {+\infty }}\mathbb{E}\rho \bigl(t, x,Y_{s}^{x}(y) \bigr)= \int_{\mathbb{R}^{m}}\rho (t,x,z)\mu^{x}(dz)=0.$$

With the aid of the last limit, we can deduce from (4.8) that

\begin{aligned}& \biggl(f(x,y),D_{y} \int_{0}^{+\infty } \bigl[\mathcal{P}_{s}\rho (t,x,y) \bigr]\,ds \biggr)_{\mathbb{R}^{m}} \\& \qquad {} +\frac{1}{2}\operatorname{Tr}\biggl[ D_{yy}^{2} \int_{0}^{+\infty } \bigl[\mathcal{P}_{s} \rho (t,x,y) \bigr]\cdot g(x,y) g^{T}(x,y)\,ds \biggr] \\& \qquad {} +\lambda_{2} \biggl( \int_{0}^{+\infty }\mathcal{P}_{s} \bigl[\rho \bigl(t,x,y+h(x,y) \bigr) \bigr]\,ds- \int_{0}^{+\infty }\mathcal{P}_{s} \bigl[\rho (t,x,y) \bigr]\,ds \biggr) \\& \quad = \int_{0}^{+\infty }\frac{\partial }{\partial s}\mathcal{P}_{s} \bigl[ \rho (t,x,y) \bigr]\,ds \\& \quad =\lim_{s\rightarrow {+\infty }}\mathbb{E}\rho \bigl(t, x,Y_{s} ^{x}(y) \bigr)-\rho (t,x,y) \\& \quad = \int_{\mathbb{R}^{m}}\rho (t,x,z)\mu^{x}(dz)-\rho (t,x,y) \\& \quad =-\rho (t,x,y), \end{aligned}

which implies

$$\mathcal{L}_{2} \biggl( \int_{0}^{{+\infty }}\mathcal{P}_{s}\rho (t,x,y) \,ds \biggr)=- \rho (t,x,y).$$

Therefore,

\begin{aligned}& u_{1}(t,x,y):= \int_{0}^{+\infty } \mathbb{E}\rho \bigl(t,x,Y^{x}_{s}(y) \bigr)\,ds \end{aligned}
(4.9)

is the solution to equation (4.7).

### Lemma 4.2

Under assumptions (A1), (A2), and (A3), for any $$x\in \mathbb{R}^{n}$$, $$y\in \mathbb{R}^{m}$$, and $$T>0$$, we have

\begin{aligned}& \bigl\vert u_{1}(t,x,y) \bigr\vert \leq C_{T} \bigl(1 + \Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R} ^{m}} \bigr),\quad t\in [0, T]. \end{aligned}
(4.10)

### Proof

By (4.9) we have

$$u_{1}(t,x,y)= \int_{0}^{{+\infty }}\mathbb{E} \bigl(\bar{a}(x)- a \bigl(x,Y _{s}^{x}(y) \bigr), D_{x}\bar{u}(t,x) \bigr)_{\mathbb{R}^{n} }\,ds,$$

so that

\begin{aligned}& \bigl\vert u_{1}(t,x,y) \bigr\vert \leq \int_{0}^{{+\infty }} \bigl\Vert \bar{a}(x)-\mathbb{E} \bigl[a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr] \bigr\Vert _{\mathbb{R}^{n} } \cdot \bigl\Vert D_{x}\bar{u}(t,x) \bigr\Vert _{ \mathbb{R}^{n} }\,ds. \end{aligned}

Therefore, from Lemma A.1 and (3.5) we get

\begin{aligned} \bigl\vert u_{1}(t,x,y) \bigr\vert \leq & C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n} } +\Vert y\Vert _{ \mathbb{R}^{m} } \bigr) \int_{0}^{{+\infty }}e^{-\frac{\beta }{2} s}\,ds \leq C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n} } +\Vert y \Vert _{\mathbb{R}^{m} } \bigr). \end{aligned}

□

### Determination of remainder $${r^{\epsilon }}$$

We now turn to the construction for the remainder term $$r^{\epsilon }$$. It is known that

\begin{aligned}& \bigl(\partial_{t}-\mathcal{L}^{\epsilon } \bigr)u^{\epsilon }=0, \end{aligned}

which, together with (4.4) and (4.5), implies

\begin{aligned} \bigl(\partial_{t}-\mathcal{L}^{\epsilon } \bigr)r^{\epsilon } =&- \bigl(\partial_{t}- \mathcal{L}^{\epsilon } \bigr) u_{0}-\epsilon \bigl(\partial_{t}-\mathcal{L}^{ \epsilon } \bigr)u_{1} \\ =&- \biggl(\partial_{t}-\frac{1}{\epsilon }\mathcal{L}_{2}- \mathcal{L}_{1} \biggr)u _{0}-\epsilon \biggl( \partial_{t}-\frac{1}{\epsilon }\mathcal{L}_{2}- \mathcal{L}_{1} \biggr)u_{1} \\ =&\epsilon (\mathcal{L}_{1}u_{1}-\partial_{t}u_{1}). \end{aligned}
(4.11)

To estimate the remainder term $$r^{\epsilon }$$, we need the following two lemmas.

### Lemma 4.3

Under assumptions (A1), (A2), and (A3), for any $$x\in \mathbb{R}^{n}$$, $$y\in \mathbb{R}^{m}$$, and $$T>0$$, we have

\begin{aligned}& \biggl\vert \frac{\partial u_{1}}{\partial t}(t,x,y) \biggr\vert \leq C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

### Proof

In view of (4.9), we get

$$\frac{\partial u_{1}}{\partial t}(t,x,y)= \int_{0}^{{+\infty }} \mathbb{E} \biggl( \bar{a}(x)- a \bigl(x,Y_{s}^{x}(y) \bigr), \frac{\partial }{ \partial t}D_{x} \bar{u}(t,x) \biggr) _{\mathbb{R}^{n}}\,ds.$$

By Lemma A.6 introduced in Sect. 4.3 we have

\begin{aligned} \biggl\vert \frac{\partial u_{1}}{\partial t}(t,x,y) \biggr\vert \leq & \int_{0} ^{{+\infty }}\mathbb{E} \biggl( \bigl\Vert \bar{a}(x)-a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr\Vert _{ \mathbb{R}^{n}}\cdot \biggl\Vert \frac{\partial }{\partial t}D_{x}\bar{u}(t,x) \biggr\Vert _{\mathbb{R}^{n}} \biggr)\,ds \\ \leq &C_{T} \int_{0}^{{+\infty }}\mathbb{E} \bigl\Vert \bar{a}(x)-a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr\Vert _{\mathbb{R}^{n}}\,ds, \end{aligned}

so that from (3.5) we have

\begin{aligned}& \biggl\vert \frac{\partial u_{1}}{\partial t}(t,x,y) \biggr\vert \leq C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

□

### Lemma 4.4

Under assumptions (A1), (A2), and (A3), for any $$x\in \mathbb{R}^{n}$$, $$y\in \mathbb{R}^{m}$$, and $$T>0$$, we have

\begin{aligned}& \bigl\vert \mathcal{L}_{1}u_{1}(t,x,y) \bigr\vert \leq C_{T} \bigl(1+\Vert x\Vert _{ \mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr), \quad t\in [0, T]. \end{aligned}

### Proof

Recalling that $$u_{1}(t,x,y)$$ is the solution of equation (4.7) and equality (4.9) holds, we have

\begin{aligned} \mathcal{L}_{1}u_{1}(t,x,y) =& \bigl(a(x,y),D_{x} u_{1}(t,x,y) \bigr)_{ \mathbb{R}^{n}} +\frac{1}{2}\operatorname{Tr}\bigl[ D^{2}_{xx}u_{1}(t,x,y) \cdot b(x) b^{T}(x) \bigr] \\ &{}+\lambda_{1} \bigl[u_{1} \bigl(t,x+c(x),y \bigr)-u_{1}(t,x,y) \bigr], \end{aligned}
(4.12)

and then, in order to prove the boundedness of $$\mathcal{L}_{1}u_{1}$$, we have to estimate the three terms arising in the right-hand side of (4.12).

Step 1: Estimate of $$(a(x,y),D_{x} u_{1}(t,x,y) )_{ \mathbb{R}^{n}}$$.

For any $$k\in \mathbb{R}^{n}$$, we have

\begin{aligned} D_{x}u_{1}(t,x,y)\cdot k =& \int_{0}^{{+\infty }} \bigl(D_{x} \bigl( \bar{ a}(x)- \mathbb{E}a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr) \cdot k,D_{x}\bar{u}(t,x) \bigr)_{\mathbb{R}^{n} }\,ds \\ &{}+ \int_{0}^{{+\infty }} \bigl(\bar{a}(x)-\mathbb{E} a \bigl(x,Y_{s}^{x}(y) \bigr),D ^{2}_{xx} \bar{u}(t,x)\cdot k \bigr)_{\mathbb{R}^{n} }\,ds \\ =:&I_{1}(t,x,y,k)+I_{2}(t,x,y,k). \end{aligned}

By Lemma A.1 and A.4 we infer that

\begin{aligned}& \bigl\vert I_{1}(t,x,y,k) \bigr\vert \\& \quad \leq \bigl\Vert D_{x}\bar{u}(t,x) \bigr\Vert _{\mathbb{R}^{n}} \int _{0}^{+\infty } \bigl\Vert D_{x} \bigl( \bar{ a}(x)-\mathbb{E}a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr) \cdot k \bigr\Vert _{\mathbb{R}^{n}}\,ds \\& \quad \leq C_{T}\Vert k\Vert _{\mathbb{R}^{n}} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{ \mathbb{R}^{m}} \bigr) \int_{0}^{+\infty }e^{-\frac{\beta }{2}s}\,ds \\& \quad \leq C_{T}\Vert k\Vert _{\mathbb{R}^{n}} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{ \mathbb{R}^{m}} \bigr). \end{aligned}
(4.13)

By Lemma A.2 and inequality (3.5) we obtain

\begin{aligned} \bigl\vert I_{2}(t,x,y,k) \bigr\vert \leq & C_{T} \Vert k\Vert _{\mathbb{R}^{n}} \int_{0}^{+\infty } \bigl\Vert \bar{a}(x)-\mathbb{E}a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr\Vert _{\mathbb{R}^{n}}\,ds \\ \leq &C_{T}\Vert k\Vert _{\mathbb{R}^{n}} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{ \mathbb{R}^{m}} \bigr) \int_{0}^{+\infty }e^{-\frac{\beta }{2}s}\,ds \\ \leq &C_{T}\Vert k\Vert _{\mathbb{R}^{n}} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{ \mathbb{R}^{m}} \bigr). \end{aligned}

This, together with (4.13), implies

\begin{aligned}& \bigl\Vert D_{x}u_{1}(t,x,y)\cdot k \bigr\Vert \leq C_{T}\Vert k\Vert _{\mathbb{R}^{n}} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr), \end{aligned}

and then, as $$a(x,y)$$ is bounded, it follows that

\begin{aligned} \bigl\vert \bigl(a(x,y),D_{x} u_{1}(t,x,y) \bigr)_{\mathbb{R}^{n}} \bigr\vert \leq & C_{T} \bigl\Vert a(x,y) \bigr\Vert _{\mathbb{R}^{n}} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr) \\ \leq & C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

Step 2: Estimate of $$\operatorname{Tr}[ D^{2}_{xx}u_{1}(t,x,y)\cdot b(x) b^{T}(x) ]$$.

Since $$u_{1}(t,x,y)$$ is given by the representation formula (4.9), for any $$k_{1},k_{2}\in \mathbb{R}^{n}$$, we have

\begin{aligned}& D^{2}_{xx}u_{1}(t,x,y)\cdot (k_{1}, k_{2}) \\& \quad = \int_{0}^{{+\infty }}\mathbb{E} \bigl(D^{2}_{xx} \bigl(\bar{a}(x)- a \bigl(x,Y _{s}^{x}(y) \bigr) \bigr)\cdot (k_{1}, k_{2}),D_{x}\bar{u}(t,x) \bigr)_{\mathbb{R} ^{n} }\,ds \\& \qquad {} + \int_{0}^{{+\infty }}\mathbb{E} \bigl(D_{x} \bigl( \bar{ {a}}(x)- a \bigl(x,Y_{s} ^{x}(y) \bigr) \bigr)\cdot k_{1},D^{2}_{xx}\bar{u}(t,x)\cdot k_{2} \bigr)_{ \mathbb{R}^{n} }\,ds \\& \qquad {}+ \int_{0}^{{+\infty }}\mathbb{E} \bigl(D_{x} \bigl( \bar{a}(x)- {a} \bigl(x,Y_{s} ^{x}(y) \bigr) \bigr)\cdot k_{2},D^{2}_{xx}\bar{u}(t,x)\cdot k_{1} \bigr)_{ \mathbb{R}^{n} }\,ds \\& \qquad {} + \int_{0}^{{+\infty }}\mathbb{E} \bigl(\bar{a}(x)- a \bigl(x,Y_{s}^{x}(y) \bigr), D^{3}_{xxx} \bar{u}(t,x)\cdot (k_{1},k_{2}) \bigr)_{\mathbb{R}^{n} }\,ds \\& \qquad :=\sum_{i=1}^{4}J_{i}(t,x,y,k_{1},k_{2}). \end{aligned}

Thanks to Lemma A.1 and Lemma A.5, we get

\begin{aligned}& \bigl\vert J_{1}(t,x,y,k_{1},k_{2})\bigr\vert \\& \quad \leq \int_{0}^{+\infty } \bigl\vert \mathbb{E} \bigl(D^{2}_{xx} \bigl(\bar{a}(x)- a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr)\cdot (k_{1},k_{2}),D_{x} \bar{u}(t,x) \bigr)_{ \mathbb{R}^{n} } \bigr\vert \,ds \\& \quad \leq C_{T} \bigl( 1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr) \Vert k_{1}\Vert _{\mathbb{R}^{n} } \Vert k_{2}\Vert _{\mathbb{R}^{n} } \int_{0}^{+ \infty }e^{-\frac{\beta }{2}s}\,ds \\& \quad \leq C_{T} \bigl( 1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr) \Vert k_{1}\Vert _{\mathbb{R}^{n} } \Vert k_{2}\Vert _{\mathbb{R}^{n} }. \end{aligned}
(4.14)

By Lemma A.4 and (3.5) we infer that

\begin{aligned}& \bigl\vert J_{2}(t,x,y,k_{1},k_{2})\bigr\vert \\& \quad \leq \int_{0}^{{+\infty }} \bigl\vert \mathbb{E} \bigl(D_{x} \bigl(\bar{ {a}}(x)- a \bigl(x,Y_{s}^{x}(y) \bigr) \bigr)\cdot k_{1},D^{2}_{xx}\bar{u}(t,x) \cdot k_{2} \bigr)_{ \mathbb{R}^{n} } \bigr\vert \,ds \\& \quad \leq C_{T} \bigl( 1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr) \Vert k_{1}\Vert _{\mathbb{R}^{n} } \Vert k_{2}\Vert _{\mathbb{R}^{n} } \int_{0}^{+ \infty }e^{-\frac{\beta }{2}s}\,ds \\& \quad \leq C_{T} \bigl( 1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr) \Vert k_{1}\Vert _{\mathbb{R}^{n} } \Vert k_{2}\Vert _{\mathbb{R}^{n} }. \end{aligned}
(4.15)

With a similar argument, we can also show that

\begin{aligned}& \bigl\vert J_{3}(t,x,y,k_{1},k_{2})\bigr\vert \\& \quad \leq C_{T} \bigl( 1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr) \Vert k_{1}\Vert _{\mathbb{R}^{n} } \Vert k_{2}\Vert _{\mathbb{R}^{n} }. \end{aligned}
(4.16)

Using Lemma A.3 and (3.5), we get

\begin{aligned}& \bigl\vert J_{4}(t,x,y,k_{1}, k_{2})\bigr\vert \\& \quad \leq C_{T}\Vert k_{1}\Vert _{\mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R}^{n}} \cdot \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}}\bigr) \int_{0}^{+ \infty } e^{-\frac{\beta }{2}s}\,ds \\& \quad \leq C_{T}\Vert k_{1}\Vert _{\mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R} ^{n}}\bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}}\bigr). \end{aligned}
(4.17)

In view of estimates (4.14), (4.15), (4.16), and (4.17), we can conclude that there exists a constant $$C_{T}$$ such that

\begin{aligned}& \bigl\vert D^{2}_{xx}u_{1}(t,x,y)\cdot (k_{1}, k_{2}) \bigr\vert \leq C_{T} \Vert k_{1}\Vert _{ \mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R}^{n}} \bigl(1+\Vert x\Vert _{\mathbb{R} ^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr),\quad t\in [0,T], \end{aligned}

which means that, for fixed $$y\in \mathbb{R}^{m}$$ and $$t\in [0, T]$$,

\begin{aligned}& \bigl\Vert D^{2}_{xx}u_{1}(t,x,y) \bigr\Vert _{L( {\mathbb{R}^{n}},\mathbb{R})}\leq C _{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+ \Vert y\Vert _{\mathbb{R}^{m}} \bigr), \end{aligned}

where $$\|\cdot \|_{L(\mathbb{R}^{n},\mathbb{R})}$$ denotes the usual operator norm on the Banach space consisting of bounded and linear operators from $$\mathbb{R}^{n}$$ to $$\mathbb{R}$$. As the diffusion function g is bounded, we get

\begin{aligned} \operatorname{Tr}\bigl(D^{2}_{xx}u_{1}(t,x,y)gg^{T} \bigr) \leq & C_{T} \bigl\Vert D^{2}_{xx}u_{1}(t,x,y) \bigr\Vert _{L( {\mathbb{R}^{n}},\mathbb{R})} \\ \leq &C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

Step 3: Estimate of $$\lambda_{1}[u_{1}(t,x+c(x),y)-u_{1}(t,x,y)]$$.

By Lemma 4.2 and the boundedness condition of $$c(x)$$, we directly have

\begin{aligned}& \bigl\vert \lambda_{1} \bigl[u_{1} \bigl(t,x+c(x),y \bigr)-u_{1}(t,x,y) \bigr] \bigr\vert \\& \lambda_{1} \bigl[ \bigl\vert u_{1} \bigl(t,x+c(x),y \bigr) \bigr\vert + \bigl\vert u_{1}(t,x,y) \bigr\vert \bigr] \\& \quad \leq C_{T} \bigl(1 +\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{\mathbb{R}^{m}} \bigr),\quad t \in [0, T]. \end{aligned}

Finally, it is now easy to gather all previous estimates for terms in (4.12) and conclude

\begin{aligned}& \bigl\vert \mathcal{L}_{1}u_{1}(t,x,y) \bigr\vert \leq C_{T} \bigl(1+\Vert x\Vert _{ \mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr),\quad t\in [0, T]. \end{aligned}

□

### Lemma 4.5

Under the conditions of Lemma 4.3, for any $$T>0$$, $$x\in {\mathbb{R} ^{n}}$$, and $$y \in {\mathbb{R}^{m}}$$, we have

\begin{aligned}& \bigl\vert r^{\epsilon }(T,x,y) \bigr\vert \leq C_{T}\epsilon \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

### Proof

By the variation-of-constant formula we write equation (4.11) in thes integral form

\begin{aligned} r^{\epsilon }(T,x,y) =&\mathbb{E} \bigl[r^{\epsilon } \bigl(0,X^{\epsilon }_{T}(x,y),Y ^{\epsilon }_{T}(x, y) \bigr) \bigr] \\ &{}+\epsilon \biggl[ \int_{0}^{T}\mathbb{E} \biggl(\mathcal{L}_{1}u_{1}- \frac{ \partial u_{1}}{\partial s} \biggr) \bigl(s, X^{\epsilon }_{T-s}(x,y),Y^{\epsilon }_{ {T-s} }(x, y) \bigr)\,ds \biggr] . \end{aligned}

Since $$u^{\epsilon }$$ and ū satisfy the same initial condition, we have

\begin{aligned} \bigl\vert r^{\epsilon }(0, x,y) \bigr\vert =& \bigl\vert u^{\epsilon }(0,x,y)-\bar{u}(0,x)-\epsilon u_{1}(0,x,y) \bigr\vert \\ =&\epsilon \bigl\vert u_{1}(0,x,y) \bigr\vert , \end{aligned}

so that, thanks to (4.10), (2.1), and (2.2), we have

\begin{aligned}& \mathbb{E} \bigl[r^{\epsilon }(0,X^{\epsilon }_{T}(x,y),Y^{\epsilon }_{T }(x,y)\bigr] \leq C\epsilon \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+ \Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}
(4.18)

Using Lemmas 4.3 and 4.4 yields

\begin{aligned}& \mathbb{E} \biggl[ \biggl(\mathcal{L}_{1}u_{1}- \frac{\partial u_{1}}{\partial s} \biggr) \bigl(s, X^{\epsilon }_{T-s}(x,y),Y^{\epsilon }_{T-s} (x,y) \bigr) \biggr] \\& \quad \leq C \mathbb{E} \bigl(1+ \bigl\Vert X^{\epsilon }_{T-s}(x,y) \bigr\Vert + \bigl\Vert Y^{\epsilon }_{T-s}(x,y) \bigr\Vert \bigr), \end{aligned}

and, according to (2.1) and (2.2), this implies that

\begin{aligned}& \mathbb{E} \biggl[ \int_{0}^{T} \biggl(\mathcal{L}_{1}u_{1}- \frac{\partial u _{1}}{\partial s} \biggr) \bigl(s, X^{\epsilon }_{T-s}(x,y), Y^{\epsilon }_{T-s}(x,y) \bigr)\,ds \biggr] \\& \quad \leq C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+ \Vert y \Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

The last inequality, together with (4.18), yields

\begin{aligned}& \bigl\vert r^{\epsilon }(T,x,y) \bigr\vert \leq \epsilon C_{T} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+ \Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

□

## References

1. Bao, J., Yin, G., Yuan, C.: Two-time-scale stochastic partial differential equations driven by α-stable noises: averaging principles. Bernoulli 23(1), 645–669 (2017)

2. Bogoliubov, N.N., Mitropolsky, Y.A.: Asymptotic Methods in the Theory of Non-linear Oscillations. Gordon & Breach, New York (1961)

3. Bréhier, C.E.: Strong and weak orders in averaging for SPDEs. Stoch. Process. Appl. 122, 2553–2593 (2012)

4. Cerrai, S., Freidlin, M.I.: Averaging principle for a class of stochastic reaction–diffusion equations. Probab. Theory Relat. Fields 144, 137–177 (2009)

5. Cerrai, S.: A Khasminskii type averaging principle for stochastic reaction–diffusion equations. Ann. Appl. Probab. 19(3), 899–948 (2009)

6. Cerrai, S.: Averaging principle for systems of reaction–diffusion equations with polynomial nonlinearities perturbed by multiplicative type noise. SIAM J. Math. Anal. 43(6), 2482–2518 (2011)

7. Freidlin, M.I., Wentzell, A.D.: Random Perturbation of Dynamical Systems, 2nd edn. Springer, New York (1998)

8. Freidlin, M.I., Wentzell, A.D.: Long-time behavior of weakly coupled oscillators. J. Stat. Phys. 123, 1311–1337 (2006)

9. Fu, H., Wan, L., Wang, Y., Liu, J.: Strong convergence rate in averaging principle for stochastic FitzHug–Nagumo system with two time-scales. J. Math. Anal. Appl. 416, 609–628 (2014)

10. Fu, H., Wan, L., Liu, J.: Strong convergence in averaging principle for stochastic hyperbolic–parabolic equations with two time-scales. Stoch. Process. Appl. 125, 3255–3279 (2015)

11. Givon, D.: Strong convergence rate for two-time-scale jump-diffusion stochastic differential systems. SIAM J. Multiscale Model. Simul. 6, 577–594 (2007)

12. Guo, Z.: $$L^{p}$$ ($$p\geq 2$$)-strong convergence in averaging principle for multivalued stochastic differential equation with non-Lipschitz coefficients. Adv. Differ. Equ. 2017, 386 (2017)

13. Khasminskii, R.Z.: On the principle of averaging the Itô stochastic differential equations. Kibernetika 4, 260–279 (1968) (Russian)

14. Kifer, Y.: Some recent advance in averaging. In: Modern Dynamical Systems and Applications, pp. 385–403. Cambridge University Press, Cambridge (2004)

15. Kifer, Y.: Diffusion approximation for slow motion in fully coupled averaging. Probab. Theory Relat. Fields 129, 157–181 (2004)

16. Kifer, Y.: Another proof of the averaging principle for fully coupled dynamical systems with hyperbolic fast motions. Discrete Contin. Dyn. Syst. 13, 1187–1201 (2005)

17. Liu, D.: Strong convergence rate of principle of averaging for jump-diffusion processes. Front. Math. China 7(2), 305–320 (2012)

18. Veretennikov, A.Y.: On the averaging principle for systems of stochastic differential equations. Math. USSR Sb. 69, 271–284 (1991)

19. Veretennikov, A.Y.: On large deviations in the averaging principle for SDEs with full dependence. Ann. Probab. 27, 284–296 (1999)

20. Volosov, V.M.: Averaging in systems of ordinary differential equations. Russ. Math. Surv. 17, 1–126 (1962)

21. Xu, J.: $$L^{p}$$-strong convergence of the averaging principle for slow–fast SPDEs with jumps. J. Math. Anal. Appl. 445, 342–373 (2017)

22. Xu, J., Liu, J.: An averaging principle for multivalued stochastic differential equations. Stoch. Anal. Appl. 32, 962–974 (2014)

23. Xu, Y., Duan, J., Xu, W.: An averaging principle for stochastic dynamical systems with Lévy noise. Physica D 240(17), 1395–1401 (2011)

24. Xu, Y., Pei, B., Li, Y.: Approximation properties for solutions to non-Lipschitz stochastic differential equations with Lévy noise. Math. Methods Appl. Sci. 11(38), 2120–2131 (2014)

25. Wainrib, G.: Double averaging principle for periodically forced slow–fast stochastic systems. Electron. Commun. Probab. 18(51), 1–12 (2013)

26. Wang, W., Roberts, A.J.: Average and deviation for slow–fast stochastic partial differential equations. J. Differ. Equ. 253, 1265–1286 (2012)

## Acknowledgements

We would like to thank Professor Jinqiao Duan for helpful discussions and comments. Hongbo Fu is supported by Natural Science Foundation of Hubei Province (No. 2018CFB688), NSF of China (No. 11301403), and Chinese Scholarship Council (No. 5104). Bengong Zhang is supported by NSF of China (No. 11401448). Li Wan is supported by NSF of China (No. 61573011). Jicheng Liu is supported by NSF of China (No. 11271013).

## Author information

Authors

### Contributions

The authors declare that the work was realized in collaboration with the same responsibility. All authors read and approved the final manuscript.

### Corresponding author

Correspondence to Hongbo Fu.

## Ethics declarations

### Competing interests

The authors declare that they have no competing interests.

### Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## Appendix

### Appendix

In this appendix, we collect some technical results to which we appeal in the proofs of the main results in Sect. 4.

### Lemma A.1

For any $$T>0$$, there exists a constant $$C_{T}>0$$ such that, for any $$x,k\in \mathbb{R}^{n}$$ and $$t\in [0, T]$$, we have

$$\bigl\vert D_{x}\bar{u}(t,x)\cdot k \bigr\vert \leq C_{T}\Vert k\Vert _{\mathbb{R}^{n}}.$$

### Proof

Observe that, for any $$k\in {\mathbb{R}^{n}}$$,

\begin{aligned} D_{x}\bar{u}(t,x)\cdot k =&\mathbb{E} \bigl[ D\phi \bigl( \bar{X}_{t}(x) \bigr) \cdot \eta^{k,x}_{t} \bigr] \\ =&\mathbb{E} \bigl( \phi ' \bigl(\bar{X}_{t}(x) \bigr), \eta^{k,x}_{t} \bigr) _{ \mathbb{R}^{n}}, \end{aligned}

where $$\eta^{k,x}_{t}$$ denotes the first mean-square derivative of $$\bar{X}_{t}(x)$$ with respect to $$x\in \mathbb{R}^{n}$$ along the direction $$k\in \mathbb{R}^{n}$$. Then we have

\begin{aligned}& \textstyle\begin{cases} d\eta^{k,x}_{t}=D_{x}\bar{a}(\bar{X}_{t}(x))\cdot \eta^{k,x}_{t}\,dt+D _{x}b(\bar{X}_{t}(x))\cdot \eta^{k,x}_{t}\,d {B}_{t} +D_{x}c(\bar{X}_{t-}(x))\cdot \eta^{k,x}_{t-}\,d {P}_{t}, \\ \eta^{k,x}_{0}=k. \end{cases}\displaystyle \end{aligned}

This means that $$\eta^{k, x}_{t}$$ is the solution of the integral equation

\begin{aligned} \eta^{k, x}_{t} =&k+ \int_{0}^{t}D_{x}\bar{a} \bigl( \bar{X}_{s}(x) \bigr)\cdot \eta^{k, x}_{s}\,ds+ \int_{0}^{t}D_{x}b \bigl( \bar{X}_{s}(x) \bigr)\cdot \eta^{k, x} _{s} \,dB_{s} \\ &{} + \int_{0}^{t}D_{x}c \bigl( \bar{X}_{s-}(x) \bigr)\cdot \eta_{s-}^{k, x} \,dP_{s}, \end{aligned}

and then, thanks to assumption (A1), we get

\begin{aligned}& \mathbb{E} \bigl\Vert \eta^{k,x}_{t} \bigr\Vert ^{2}_{\mathbb{R}^{n}} \leq C_{T}\Vert k\Vert ^{2} _{\mathbb{R}^{n}} +C_{T} \int_{0}^{t}\mathbb{E} \bigl\Vert \eta^{k,x}_{s} \bigr\Vert ^{2} _{\mathbb{R}^{n}} \,ds. \end{aligned}

Then by the Gronwall lemma it follows that

\begin{aligned}& \mathbb{E} \bigl\Vert \eta^{k,x}_{t} \bigr\Vert ^{2}_{\mathbb{R}^{n}} \leq C_{T}\Vert k\Vert ^{2}_{\mathbb{R}^{n}},\quad t\in [0, T], \end{aligned}
(A.1)

so that

\begin{aligned}& \bigl\vert D_{x}\bar{u}(t,x)\cdot k \bigr\vert \leq C_{T} \Vert k\Vert _{\mathbb{R}^{n}}. \end{aligned}

□

Next, we introduce a similar result for the second derivative of $$\bar{u}(t,x)$$.

### Lemma A.2

For any $$T>0$$, there exists a constant $$C_{T}>0$$ such that, for any $$x,k_{1},k_{2}\in \mathbb{R}^{n}$$ and $$t\in [0, T]$$, we have

$$\bigl\vert D^{2}_{xx}\bar{u}(t,x)\cdot (k_{1},k_{2}) \bigr\vert \leq C_{T}\Vert k_{1}\Vert _{ \mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R}^{n}}.$$

### Proof

For any $$k_{1}, k_{2} \in \mathbb{R}^{n}$$, we have

\begin{aligned} D^{2}_{xx}\bar{u}(t,x)\cdot (k_{1},k_{2}) =&\mathbb{E} \bigl[\phi '' \bigl( \bar{X}_{t}(x) \bigr)\cdot \bigl(\eta^{k_{1},x}_{t}, \eta^{k_{2},x}_{t} \bigr) \\ &{}+\phi ' \bigl(\bar{X}_{t}(x) \bigr)\cdot \xi^{k_{1},k_{2},x}_{t} \bigr], \end{aligned}
(A.2)

where $$\xi^{k_{1},k_{2},x}_{t}$$ is the solution of the second variation equation corresponding to the averaged equation, which may be rewritten in the following form:

\begin{aligned} \xi^{k_{1},k_{2},x}_{t} =& \int_{0}^{t} \bigl[D_{x}\bar{a} \bigl( \bar{X}_{s}(x) \bigr) \cdot \xi^{k_{1},k_{2},x}_{s}+D_{xx}^{2} \bar{a} \bigl(\bar{X}_{s}(x) \bigr) \cdot \bigl(\eta^{k_{1},x}_{s}, \eta^{k_{2},x}_{s} \bigr) \bigr]\,ds \\ &{}+ \int_{0}^{t} \bigl[D_{xx}^{2}b \bigl(\bar{X}_{s}(x) \bigr)\cdot \bigl(\eta^{k_{1},x} _{s},\eta^{k_{2},x}_{s} \bigr)+D_{x}b \bigl( \bar{X}_{s}(x) \bigr)\cdot \xi^{k_{1},k_{2},x} _{s} \bigr] \,dB_{s} \\ &{}+ \int_{0}^{t} \bigl[D_{xx}^{2}c \bigl(\bar{X}_{s-}(x) \bigr)\cdot \bigl(\eta^{k_{1},x} _{s-},\eta^{k_{2},x}_{s-} \bigr)+D_{x}c \bigl( \bar{X}_{s-}(x) \bigr)\cdot \xi^{k_{1},k_{2},x}_{s-} \bigr] \,dP_{s}. \end{aligned}

Thus, by assumption (A1) and (A.1) we have

\begin{aligned} \mathbb{E} \bigl\Vert \xi^{k_{1},k_{2},x}_{t} \bigr\Vert ^{2}_{\mathbb{R}^{n}} \leq & C _{T} \int_{0}^{t} \bigl( \bigl\{ \mathbb{E} \bigl\Vert \eta^{k_{1},x}_{s} \bigr\Vert ^{2}_{ \mathbb{R}^{n}} \bigr\} ^{\frac{1}{2}} \bigl\{ \mathbb{E} \bigl\Vert \eta^{k_{2},x}_{s} \bigr\Vert ^{2}_{\mathbb{R}^{n}} \bigr\} ^{\frac{1}{2}}+\mathbb{E} \bigl\Vert \xi^{k_{1},k_{2},x}_{s} \bigr\Vert ^{2}_{\mathbb{R}^{n}} \bigr)\,ds \\ \leq & C_{T}\Vert k_{1}\Vert _{\mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R} ^{n}} +C_{T} \int_{0}^{t}\mathbb{E} \bigl\Vert \xi^{k_{1},k_{2},x}_{s} \bigr\Vert ^{2}_{ \mathbb{R}^{n}}\,ds. \end{aligned}

By the Gronwall lemma we have

$$\mathbb{E} \bigl\Vert \zeta^{k_{1},k_{2},x}_{t} \bigr\Vert ^{2}_{\mathbb{R}^{n}}\leq C _{T}\Vert k_{1} \Vert _{\mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R}^{n}}.$$

Returning to (A.2), we get

\begin{aligned}& \bigl\vert D^{2}_{xx}\bar{u}(t,x)\cdot (k_{1},k_{2}) \bigr\vert \leq C_{T} \Vert h_{1}\Vert _{ \mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R}^{n}}. \end{aligned}

□

Using analogous arguments, we can prove the following estimate for the third-order derivative of $$\bar{u}(t,x)$$ with respect to x.

### Lemma A.3

For any $$T>0$$, there exists a constant $$C_{T}>0$$ such that, for any $$x,k_{1},k_{2},k_{3}\in \mathbb{R}^{n}$$ and $$t\in [0, T]$$, we have

$$\bigl\vert D^{3}_{xxx}\bar{u}(t,x)\cdot (k_{1},k_{2},k_{3}) \bigr\vert \leq C_{T }\Vert k_{1}\Vert _{\mathbb{R}^{n}}\cdot \Vert k_{2}\Vert _{\mathbb{R}^{n}}\cdot \Vert k_{3}\Vert _{\mathbb{R}^{n}}.$$

The following lemma states the boundedness for the first derivative of $$\bar{a}(x)-\mathbb{E}a(x, Y^{x}_{t}(y))$$ with respect to x.

### Lemma A.4

There exists a constant $$C>0$$ such that, for any $$x\in \mathbb{R}^{n}$$, $$y\in \mathbb{R}^{m}$$, $$k\in \mathbb{R}^{n}$$, and $$t>0$$,

\begin{aligned}& \bigl\Vert D_{x} \bigl(\bar{a}(x)-\mathbb{E}a \bigl(x, Y^{x}_{t}(y) \bigr) \bigr)\cdot k \bigr\Vert _{ \mathbb{R}^{n}} \leq Ce^{-\frac{\beta }{2}t}\Vert k\Vert _{\mathbb{R}^{n}} \bigl( 1+ \Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

### Proof

The proof is a modification of the proof of [3, Prop. C.2]. For any $$t_{0}>0$$, we set

\begin{aligned}& \tilde{a}_{t_{0}}(x,y,t)=\hat{a}(x,y,t)-\hat{a}(x,y,t+t_{0}), \end{aligned}

where

\begin{aligned}& \hat{a}(x,y,t):=\mathbb{E}a \bigl(x, Y^{x}_{t}(y) \bigr). \end{aligned}

Then we have

\begin{aligned}& \lim_{t_{0}\rightarrow +\infty }\tilde{a}_{t_{0}}(x,y,t)= \mathbb{E}a \bigl(x, Y^{x}_{t}(y) \bigr)-\bar{a}(x). \end{aligned}

By the Markov property we have

\begin{aligned} \tilde{a}_{t_{0}}(x,y,t) =&\hat{a}(x,y,t)-\mathbb{E} {a} \bigl(x,Y_{t+t_{0}} ^{x}(y) \bigr) \\ =&\hat{a}(x,y,t)-\mathbb{E}\hat{a} \bigl(x, Y_{t_{0}}^{x}(y),t \bigr). \end{aligned}

Due to assumption (A1), for any $$k\in \mathbb{R}^{n}$$, we have

\begin{aligned} D_{x}\tilde{a}_{t_{0}}(x,y,t)\cdot k =&D_{x} \hat{a}(x,y,t)\cdot k- \mathbb{E}D_{x} \bigl( \hat{a} \bigl(x, Y_{t_{0}}^{x }(y),t \bigr) \bigr) \cdot k \\ =&\hat{a}_{x}'(x,y,t)\cdot k-\mathbb{E} \hat{a}_{x}' \bigl(x, Y_{t_{0}}^{x}(y),t \bigr) \cdot k \\ &{} -\mathbb{E}\hat{a}_{y}' \bigl(x, Y_{t_{0}}^{x}(y),t \bigr)\cdot \bigl( D_{x}Y _{t_{0}}^{x}(y)\cdot k \bigr), \end{aligned}
(A.3)

where the symbols $$\hat{a}_{x}'$$ and $$\hat{a}_{y}'$$ denote the directional derivatives with respect to x and y, respectively. Note that the first derivative $$\zeta_{t}^{x,y, k}=D_{x}Y_{t}^{x}(y)\cdot k$$, at the point x and along the direction $$k\in \mathbb{R}^{n}$$, is the solution of the equation

\begin{aligned} d\zeta_{t}^{x, y,k} =& \bigl( f_{x}' \bigl(x, Y_{t}^{x}(y) \bigr)\cdot k+f_{y}' \bigl(x, Y_{t}^{x}(y) \bigr)\cdot \zeta_{t}^{x, y,k} \bigr)\,dt \\ &{}+ \bigl( g_{x}' \bigl(x, Y_{t}^{x}(y) \bigr)\cdot k+g_{y}' \bigl(x, Y_{t}^{x}(y) \bigr)\cdot \zeta_{t}^{x, y,k} \bigr)\,dW_{t} \\ &{}+ \bigl( h_{x}' \bigl(x, Y_{t-}^{x}(y) \bigr)\cdot k+h'_{y} \bigl(x, Y_{t-}^{x}(y) \bigr) \cdot \zeta_{t-}^{x, y,k} \bigr)\,dN_{t} \end{aligned}

with initial data $$\zeta_{0}^{x,y, k}=0$$. Hence, by assumption (A1), it is straightforward to check

\begin{aligned}& \mathbb{E} \bigl\Vert \zeta_{t}^{x,y, k} \bigr\Vert _{\mathbb{R}^{m}}\leq C\Vert k\Vert _{ \mathbb{R}^{n}} \end{aligned}
(A.4)

for any $$t\geq 0$$. Note that, for any $$y_{1}, y_{2}\in {\mathbb{R} ^{m}}$$, we have

\begin{aligned} \bigl\Vert \hat{a}(x,y_{1},t)-\hat{a}(x,y_{2},t) \bigr\Vert _{\mathbb{R}^{n}} =& \bigl\Vert \mathbb{E}a \bigl(x,Y_{t}^{x}(y_{1}) \bigr)-\mathbb{E}a \bigl(x,Y_{t}^{x}(y_{2}) \bigr) \bigr\Vert _{\mathbb{R}^{n}} \\ \leq &C\mathbb{E} \bigl\Vert Y_{t}^{x}(y_{1})-Y_{t}^{x}(y_{2}) \bigr\Vert _{\mathbb{R} ^{m}} \\ \leq & Ce^{-\frac{\beta }{2}t}\Vert y_{1}-y_{2}\Vert _{\mathbb{R}^{m}}, \end{aligned}

where (3.3) was used to obtain the last inequality. This means that

\begin{aligned}& \bigl\Vert \hat{a}_{y}'(x, y,t)\cdot l \bigr\Vert _{\mathbb{R}^{m}}\leq Ce^{-\frac{ \beta }{2}t}\Vert l\Vert _{\mathbb{R}^{m}},\quad l\in { \mathbb{R}^{m}}. \end{aligned}
(A.5)

From (A.4) and (A.5) we obtain

\begin{aligned}& \bigl\Vert \mathbb{E} \bigl[\hat{a}_{y}' \bigl(x,Y_{t_{0}}^{x}(y),t \bigr)\cdot \bigl( D_{x}Y_{t_{0}}^{x}(y) \cdot k \bigr) \bigr] \bigr\Vert _{\mathbb{R}^{m}} \\& \quad = \bigl\Vert \mathbb{E} \bigl[\hat{a}_{y}' \bigl(x,Y_{t_{0}}^{x}(y),t \bigr)\cdot \bigl( \zeta_{t_{0}}^{x,y, k} \bigr) \bigr] \bigr\Vert _{\mathbb{R}^{m}} \\& \quad \leq C e^{-\frac{\beta }{2}t}\Vert k\Vert _{\mathbb{R}^{n}}. \end{aligned}
(A.6)

Then by easy calculations we have

\begin{aligned}& \hat{a}_{x}'(x,y_{1},t)\cdot k- \hat{a}_{x}'(x,y_{2},t)\cdot k \\& \quad =\mathbb{E} \bigl( a_{x}' \bigl(x, Y_{t}^{x}(y_{1}) \bigr) \bigr) \cdot k- \mathbb{E} \bigl( a_{x}' \bigl(x, Y_{t}^{x}(y_{2}) \bigr) \bigr) \cdot k \\& \qquad {} +\mathbb{E} \bigl( a_{y}' \bigl(x, Y_{t}^{x}(y_{1}) \bigr)\cdot \zeta_{t}^{x,y_{1}, k}-a_{y}' \bigl(x, Y_{t}^{x}(y_{2}) \bigr)\cdot \zeta_{t}^{x,y_{2}, k} \bigr) \\& \quad = \mathbb{E} \bigl( a_{x}' \bigl(x, Y_{t}^{x}(y_{1}) \bigr) \bigr) \cdot k- \mathbb{E} \bigl( a_{x}' \bigl(x, Y_{t}^{x}(y_{2}) \bigr) \bigr) \cdot k \\& \qquad {} +\mathbb{E} \bigl( \bigl[a_{y}' \bigl(x, Y_{t}^{x}(y_{1}) \bigr)-a_{y}' \bigl(x, Y_{t}^{x}(y _{2}) \bigr) \bigr]\cdot \zeta_{t}^{x,y_{1}, k} \bigr) \\& \qquad {} +\mathbb{E} \bigl( a_{y}' \bigl(x, Y_{t}^{x}(y_{2}) \bigr)\cdot \bigl( \zeta_{t}^{x,y_{1}, k}-\zeta_{t}^{x,y_{2}, k} \bigr) \bigr) \\& \quad := \sum_{i=1}^{3}\mathcal{N}_{i}(t,x,y_{1},y_{2}, k). \end{aligned}
(A.7)

Now, we estimate the three terms in the right-hand side of the equality. Concerning $$\mathcal{N}_{1}(t,x,y_{1},y_{2}, k)$$, we have

\begin{aligned}& \bigl\Vert \mathcal{N}_{1}(t,x,y_{1},y_{2}, k) \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq \mathbb{E} \bigl\Vert \bigl( a_{x}' \bigl(x, Y_{t}^{x}(y_{1}) \bigr) \bigr) \cdot k- \bigl( a_{x}' \bigl(x, Y_{t}^{x}(y_{2}) \bigr) \bigr) \cdot k \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq C\mathbb{E} \bigl\Vert Y_{t}^{x}(y_{1})-Y_{t}^{x}(y_{2}) \bigr\Vert _{\mathbb{R} ^{m}}\cdot \Vert k\Vert _{\mathbb{R}^{n}} \\& \quad \leq Ce^{-\frac{\beta }{2}t}\Vert y_{1}-y_{2}\Vert _{\mathbb{R}^{m}}\cdot \Vert k\Vert _{\mathbb{R}^{n}}. \end{aligned}
(A.8)

Next, by assumption (A1) we get

\begin{aligned}& \bigl\Vert \mathcal{N}_{2}(t,x,y_{1},y_{2}, k) \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq \mathbb{E} \bigl\Vert \bigl[a_{y}' \bigl(x, Y_{t}^{x}(y_{1}) \bigr)-a_{y}' \bigl(x,Y_{t}^{x}(y_{2}) \bigr) \bigr]\cdot \zeta_{t}^{x,y_{1}, k} \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq C \bigl\{ \mathbb{E} \bigl\Vert \zeta_{t}^{x,y_{1},k} \bigr\Vert ^{2}_{\mathbb{R}^{m}} \bigr\} ^{\frac{1}{2}}\cdot \bigl\{ \mathbb{E} \bigl\Vert Y_{t}^{x}(y_{1})-Y_{t}^{x}(y_{2}) \bigr\Vert ^{2}_{\mathbb{R}^{m}} \bigr\} ^{\frac{1}{2}} \\& \quad \leq C e^{-\frac{\beta }{2}t}\Vert k\Vert _{\mathbb{R}^{n}}\cdot \Vert y_{1}-y_{2}\Vert _{\mathbb{R}^{m}}. \end{aligned}
(A.9)

For the third term, using assumption (A1) again, we can infer that

\begin{aligned}& \bigl\Vert \mathcal{N}_{3}(t,x,y_{1},y_{2}, k) \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq \mathbb{E} \bigl\Vert a_{y}' \bigl(x, Y_{t}^{x}(y_{2}) \bigr)\cdot \bigl( \zeta_{t}^{x,y_{1}, k}-\zeta_{t}^{x,y_{2}, k} \bigr) \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq C\mathbb{E} \bigl\Vert \zeta_{t}^{x,y_{1}, k}- \zeta_{t}^{x,y_{2}, k} \bigr\Vert _{\mathbb{R}^{m}} \\& \quad \leq C e^{-\frac{\beta }{2}t}\Vert y_{1}-y_{2}\Vert _{\mathbb{R}^{m}} \cdot \Vert k\Vert _{\mathbb{R}^{n}}. \end{aligned}
(A.10)

Now, returning to (A.7) and taking into account (A.8), (A.9), and (A.10), we get

\begin{aligned}& \bigl\Vert \hat{a}_{x}'(x,y_{1},t)\cdot k- \hat{a}_{x}'(x,y_{2},t)\cdot k \bigr\Vert \\& \quad \leq C e^{-\frac{\beta }{2}t}\Vert y_{1}-y_{2}\Vert _{\mathbb{R}^{m}}\cdot \Vert k\Vert _{\mathbb{R}^{n}}, \end{aligned}

\begin{aligned}& \bigl\Vert \hat{a}_{x}'(x,y,t)\cdot h- \mathbb{E} \hat{a}_{x}' \bigl(x,Y_{t_{0}}^{x}(y),t \bigr)\cdot k \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq C e^{-\frac{\beta }{2}t} \bigl(1+\Vert y\Vert _{\mathbb{R}^{m}}+ \bigl\Vert Y_{t_{0}}^{x}(y) \bigr\Vert _{\mathbb{R}^{m}} \bigr)\cdot \Vert k\Vert _{\mathbb{R}^{n}} \\& \quad \leq e^{-\frac{\beta }{2}t} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y \Vert _{ \mathbb{R}^{m}} \bigr)\cdot \Vert k\Vert _{\mathbb{R}^{n}}, \end{aligned}
(A.11)

where we used inequality (3.2). Returning to (A.3), by (A.6) and (A.11) we conclude that

\begin{aligned}& \bigl\Vert D_{x}\tilde{a}_{t_{0}}(x,y,t)\cdot k \bigr\Vert _{\mathbb{R}^{n}}\leq Ce^{-\frac{ \beta }{2}t} \bigl(1+\Vert x\Vert _{\mathbb{R}^{n}}+ \Vert y\Vert _{\mathbb{R}^{m}} \bigr)\Vert k\Vert _{\mathbb{R}^{n}}. \end{aligned}

Taking the limit as $$t_{0}\rightarrow +\infty$$, we obtain

\begin{aligned}& \bigl\Vert D_{x} \bigl(\bar{a}(x)-\mathbb{E}a \bigl(x, Y^{x}_{t}(y) \bigr) \bigr) \bigr\Vert _{\mathbb{R}^{n}} \leq Ce^{-\frac{\beta }{2}{t}}\Vert k\Vert _{\mathbb{R}^{n}} \bigl( 1+\Vert x\Vert _{ \mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

□

Proceeding with similar arguments, we obtain the following higher-order differentiability.

### Lemma A.5

There exists a constant $$C>0$$ such that, for any $$x, k_{1}, k_{2} \in \mathbb{R}^{n}$$, $$y\in \mathbb{R}^{m}$$, and $$t>0$$,

\begin{aligned}& \bigl\Vert D^{2}_{xx} \bigl(\bar{a}(x)-\mathbb{E}a \bigl(x,Y^{x}_{t}(y) \bigr) \bigr) (k_{1},k_{2}) \bigr\Vert _{\mathbb{R}^{n}} \\& \quad \leq Ce^{-\frac{\beta }{2}{t}}\Vert k_{1}\Vert _{\mathbb{R}^{n}}\Vert k_{2}\Vert _{\mathbb{R}^{n}} \bigl( 1+\Vert x\Vert _{\mathbb{R}^{n}}+\Vert y\Vert _{\mathbb{R}^{m}} \bigr). \end{aligned}

Finally, we introduce the following auxiliary result.

### Lemma A.6

There exists a constant $$C>0$$ such that, for any $$x, k\in \mathbb{R} ^{n}$$, $$y\in \mathbb{R}^{m}$$, and $$t>0$$,

$$\biggl\Vert \frac{\partial }{\partial t}D_{x}\bar{u}(t,x)\cdot k \biggr\Vert _{\mathbb{R} ^{n}}\leq C\Vert k\Vert _{\mathbb{R}^{n}}.$$

### Proof

For simplicity of presentation, we will prove it for the one-dimensional case. The multidimensional situation can be treated similarly, and only notations are somewhat involved. In this case, we only need to show that

\begin{aligned}& \biggl\vert \frac{\partial }{\partial t}\frac{\partial }{\partial x}\bar{u}(t,x) \biggr\vert \leq C. \end{aligned}
(A.12)

In fact, for any $$\phi \in C_{b}^{3}(\mathbb{R},\mathbb{R})$$, we have

\begin{aligned}& \frac{\partial }{\partial x}\bar{u}(t,x)=\frac{\partial }{\partial x} \mathbb{E}\phi \bigl( \bar{X}_{t}(x) \bigr)=\mathbb{E} \biggl( \phi ' \bigl( \bar{X}_{t}(x) \bigr) \cdot \frac{\partial }{\partial x}\bar{X}_{t}(x) \biggr). \end{aligned}

If we define

\begin{aligned}& \varsigma^{x}_{t}:=\frac{\partial }{\partial x}\bar{X}_{t}(x), \end{aligned}

then we have

\begin{aligned} \varsigma^{x}_{t} =&1+ \int_{0}^{t}\bar{a}' \bigl( \bar{X}_{s}(x) \bigr)\cdot \varsigma^{x}_{s}\,ds+ \int_{0}^{t}b' \bigl( \bar{X}_{s}(x) \bigr)\cdot \varsigma ^{x}_{s} \,dB_{s} \\ &{}+ \int_{0}^{t}c' \bigl( \bar{X}_{s-}(x) \bigr)\cdot \varsigma_{s-}^{x} \,dP_{s}. \end{aligned}

The boundedness of $$\bar{a}'$$, $$b'$$, and $$c'$$ guarantees that

\begin{aligned}& \mathbb{E} \bigl\vert \varsigma^{x}_{t} \bigr\vert ^{2}\leq C_{T},\quad t\in [0, T]. \end{aligned}
(A.13)

By Itô’s formula we have

\begin{aligned}& \mathbb{E} \bigl[\phi ' \bigl(\bar{X}_{t}(x) \bigr) \cdot \varsigma_{t}^{x} \bigr] \\& \quad =\phi '(x)+\mathbb{E} \int_{0}^{t} \bigl[\phi ' \bigl( \bar{X}_{s}(x) \bigr)\bar{a}' \bigl( \bar{X}_{s}(x) \bigr)\varsigma_{s}^{x} +\varsigma_{s}^{x} \phi '' \bigl(\bar{X} _{s}(x) \bigr)\bar{a} \bigl(\bar{X}_{s}(x) \bigr) \bigr]\,ds \\& \qquad {} +\mathbb{E} \int_{0}^{t}b' \bigl( \bar{X}_{s}(x) \bigr))\varsigma_{s}^{x}\phi '' \bigl( \bar{X}_{s}(x) \bigr)b \bigl( \bar{X}_{s}(x) \bigr)\,ds \\& \qquad {} +\frac{1}{2}\mathbb{E} \int_{0}^{t}\varsigma_{s}^{x} \phi ''' \bigl(\bar{X} _{s}(x) \bigr)b^{2} \bigl(\bar{X}_{s}(x) \bigr)\,ds \\& \qquad {} +\lambda_{1}\mathbb{E} \int_{0}^{t}\phi ' \bigl( \bar{X}_{s}(x) \bigr)c' \bigl(\bar{X} _{s-}(x) \bigr)\varsigma_{s}^{x}\,ds \\& \qquad {} +\lambda_{1}\mathbb{E} \int_{0}^{t}\varsigma_{s}^{x} \bigl[\phi ' \bigl(\bar{X} _{s-}(x)+c \bigl( \bar{X}_{s-}(x) \bigr) \bigr)-\phi ' \bigl( \bar{X}_{s-}(x) \bigr) \bigr]\,ds \\& \qquad {} +\lambda_{1}\mathbb{E} \int_{0}^{t}c' \bigl( \bar{X}_{s-}(x) \bigr)\varsigma_{s} ^{x} \bigl[\phi ' \bigl(\bar{X}_{s-}(x)+c \bigl(\bar{X}_{s-}(x) \bigr) \bigr)-\phi ' \bigl(\bar{X}_{s-}(x) \bigr) \bigr] \,ds. \end{aligned}

Since ϕ belongs to $$C_{b}^{3}(\mathbb{R}, \mathbb{R})$$, from assumption (A1) it follows that, for any $$t\in [0, T]$$,

\begin{aligned} \biggl\vert \frac{\partial }{\partial t} \biggl[\frac{\partial }{\partial x} \bar{u}(t,x) \biggr] \biggr\vert =& \biggl\vert \frac{\partial }{\partial t}\mathbb{E} \bigl[ \phi ' \bigl(\bar{X}_{t}(x) \bigr)\cdot \varsigma_{t}^{x} \bigr] \biggr\vert \\ \leq & C \bigl\vert \mathbb{E}\varsigma_{t}^{x} \bigr\vert . \end{aligned}

Then, taking (A.13) into account, we easily arrive at (A.12). □

## Rights and permissions 