1. Introduction
In a recent paper [Reference Bakhtin and Pajor-Gyulai4] we studied tails of diffusion exit times from neighborhoods of unstable critical points, in the limit of vanishing noise, in one dimension. The typical exit time
$\kappa_{\varepsilon}$
in this setting (see the detailed description of the setting below) is of the order of
$\frac{1}{\lambda}\log\frac{1}{{\varepsilon}}$
, where
$\lambda>0$
is the local expansion coefficient of the linearization of the system near the critical point, and
${\varepsilon}\downarrow 0$
is the noise magnitude. The main result of [Reference Bakhtin and Pajor-Gyulai4] is that the following polynomial asymptotics holds for a class of initial conditions near the critical point:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn1.png?pub-status=live)
for
$\alpha>1$
, with an explicit dependence of the factor c on the initial condition and the parameters of the model; see (1.7) below for details.
This result is part of an ongoing effort to understand the long-term properties of multi-dimensional diffusions in the context of noisy heteroclinic networks, including the limiting behavior of invariant distributions associated with such systems in the vanishing noise limit. The typical behavior in such settings is understood for timescales logarithmic in
${\varepsilon}^{-1}$
; see [Reference Almada and Bakhtin1, Reference Bakhtin2, Reference Bakhtin3]. To see what happens in the long run, though, one has to quantify rare events responsible for transitions that are atypical at the logarithmic time scale. Our work in progress shows that these rare events play a crucial role in the long-term dynamics near noisy heteroclinic networks. Moreover, we argue that they occur exactly due to atypically long stays near unstable critical points. The resulting picture is similar to that of metastability but with polynomial transition rates in place of exponential ones. We give more details on this picture in Section 6, while here we only reiterate that the result of the form (1.1) and its ramifications will be crucial for that program. However, the technique we used in [Reference Bakhtin and Pajor-Gyulai4] to analyze densities of auxiliary random variables was based on heavy tools from Malliavin calculus. That approach somewhat obscures the reason why this result is true and does not seem to be tractable when applied to the study of the analogous exit problem in the neighborhood of a hyperbolic saddle in
${\mathbb R}^d$
,
$d>1$
, i.e. when both attracting and repelling directions are present.
In the present note, our goal is to give a new proof of this result that (a) is based on a more precise description of the dynamics at small scales, (b) uses more elementary tools of stochastic calculus, and (c) has a strong potential to be applicable in higher dimensions. In fact, we prove a slightly more general result on probabilities of the form
$\mathbb{P}\big(\kappa_{\varepsilon}>\frac{\alpha}{\lambda}\log \frac{1}{{\varepsilon}}+t\big)$
,
$\alpha>1$
, for all
$t\in{\mathbb R}$
instead of
$t=0$
as considered in [Reference Bakhtin and Pajor-Gyulai4]. It turns out that, asymptotically, the dependence on t is exponential, which implies that for any
$T\in{\mathbb R}$
,
$\kappa_{\varepsilon}-\frac{\alpha}{\lambda}\log \frac{1}{{\varepsilon}}-T$
conditioned on
$\kappa_{\varepsilon}-\frac{\alpha}{\lambda}\log \frac{1}{{\varepsilon}}>T$
converges in distribution to an exponential random variable. This phenomenon is a manifestation of loss of memory in the system under conditioning, and it is consistent with the fact that
$\kappa_{\varepsilon} - \frac{1}{\lambda}\log \frac{1}{{\varepsilon}}$
converges in distribution to a random variable with exponentially decaying right tails; see, e.g., [Reference Bakhtin2].
An important ingredient in this note is Lemma 3.2, a conditional equidistribution result that states that the distribution of the diffusion, conditioned on no exit from a small interval, converges to the uniform distribution. Thus our new approach is closer in spirit to the one based on quasi-stationary distributions; see [Reference Champagnat and Villemonais6]. However, the existing general theory does not provide answers for us since in our situation both the system and the domain depend on
${\varepsilon}$
. Moreover, the timescales we are interested in are too short for the
$t\to\infty$
limit to be a good approximation while taking
$\varepsilon\downarrow 0$
.
Let us be more precise now. We consider the family of stochastic differential equations
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn2.png?pub-status=live)
on a bounded interval
$\mathcal{I}=[q_-,q_+]\subseteq \mathbb{R}$
with the origin in its interior. The drift is given by a vector field
$b\in\mathcal{C}^{2}({\mathbb R})$
and the random perturbation is given via a standard Brownian motion W with respect to a filtration
$(\mathcal{F}_t)_{t\ge 0}$
defined on some probability space
$(\Omega,\mathcal{F},\mathbb{P})$
under the usual conditions. The noise magnitude is given by a small parameter
${\varepsilon}>0$
in front of the diffusion coefficient
$\sigma$
, which is assumed to be Lipschitz and satisfy
$\sigma(0)>0$
. Although we are interested only in the evolution within
$\mathcal{I}$
, we can assume that b and
$\sigma$
are globally Lipschitz without changing the setting.
Standard results on stochastic differential equations (see, e.g., [8, Chapter 5]) imply that, for any starting location
$X^{{\varepsilon}}(0)\in\mathcal{I}$
, (1.2) has a unique strong solution up to the exit time from
$\mathcal{I}$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU1.png?pub-status=live)
Let
$(S^t)_{t\in{\mathbb R}}$
be the flow generated by the vector field b, i.e.
$x(t)=S^tx_0$
is the solution of the autonomous ordinary differential equation
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU2.png?pub-status=live)
(we recall that solving this equation for negative times is equivalent to solving
$\dot y(t)=-b(y(t))$
for
$y(t)=x(\!-t)$
). We assume that there is a unique repelling zero of the vector field b on
${\mathbb R}$
, which, without loss of generality, we place at the origin. In other words, we assume that
$b(0)=0$
and, for some
$\lambda>0$
and
$\eta\in\mathcal{C}^2(\mathcal{I})$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn3.png?pub-status=live)
Note that since the origin is the only zero of b in the closed interval
$\mathcal{I}$
, this assumption implies that, for all
$x\neq 0$
, there is a uniquely defined finite time T(x) such that
$S^{T(x)}\in\partial\mathcal{I}$
.
Under the condition (1.3), the map
$f\,{:}\,\mathcal{I}\to{\mathbb R}$
defined by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn4.png?pub-status=live)
is an order-preserving
$\mathcal{C}^2$
-diffeomorphism (see [Reference Eizenberg7]). In particular,
$f(q_-)<0<f(q_+)$
. This map linearizes the flow
$(S^t)$
, see (2.1), and helps to state the main result concisely; see (1.7).
Under the above assumptions, a version of the following theorem was proved in [Reference Bakhtin and Pajor-Gyulai4]. In its statement and throughout the paper we use
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn5.png?pub-status=live)
Theorem 1.1. Consider
$X_\varepsilon$
defined by (1.2) with initial condition
$X_\varepsilon(0)=\varepsilon x$
, and let
$K({\varepsilon})$
be any function that satisfies
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn6.png?pub-status=live)
Then, for all
$\alpha>1$
and all
$t\in{\mathbb R}$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn7.png?pub-status=live)
where
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU3.png?pub-status=live)
In particular, for any
$T\in{\mathbb R}$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU4.png?pub-status=live)
where
$\Rightarrow$
stands for weak convergence, and
$\exp_{\lambda}$
is the exponential distribution with rate
$\lambda>0$
, i.e.
$\exp_\lambda[t,\infty)={\rm e}^{-\lambda t}$
for
$t\ge 0$
.
Remark 1.1. We say that a function satisfying (1.6) grows at most subpolynomially at 0. We say that a function
$c(\varepsilon)$
decays at most subpolynomially at 0 if
$1/c(\varepsilon)$
grows subpolynomially at 0. For brevity, we will usually omit the reference to 0 and simply say grows/decays subpolynomially even when the function might not actually grow.
Remark 1.2. The theorem is stated for initial conditions that are at most of the order of
${\varepsilon}$
away from the origin up to a subpolynomial factor. The case of initial conditions of the order of
${\varepsilon}^\beta$
for
$\beta<1$
is less interesting since then the tails of exit times decay as stretched exponentials of
${\varepsilon}^{-1}$
instead of the power decay given by (1.7) (see Proposition 2.1).
Remark 1.3. In the statement of Theorem 1.1 and in the following we adopt the usual convention that each relation involving
$\pm$
and
$\mp$
stands for two relations, one with all top signs and one with all bottom signs.
The brief outline of our approach to the proof of this theorem is as follows. It is convenient to work in coordinates given by the function f defined in (1.4) where the drift is linear. We study the dynamics of the linear process in two separate phases: (1) in a neighborhood of the critical point of radius
$\varepsilon^{\beta}$
for
$\beta\in(0,1)$
; (2) between leaving this small neighborhood and reaching the boundary of
$f(\mathcal{I})$
.
In the second stage, the drift dominates the noise, and the process closely follows the corresponding deterministic trajectory one obtains by setting
$\varepsilon=0$
. The outcome of the first stage, though, i.e. the exit from
$[\!-\varepsilon^{\beta},\varepsilon^{\beta}]$
, is determined by a delicate interplay between the noise and the drift in an even smaller neighborhood of the origin (
$\beta$
can be chosen arbitrarily close to one). We study this regime by introducing an auxiliary process
$Z_\varepsilon(t)$
with constant diffusion coefficient approximating
$Y_\varepsilon(t)=f(X_\varepsilon(t))$
pathwise, at least over time intervals that are not too large and for which Theorem 1.1 is easier to establish. Since
$Z_\varepsilon(t)$
and
$Y_\varepsilon(t)$
do not, in general, stay close on longer timescales, we introduce an iterative scheme to tackle this problem. Namely, we split the longer time interval into shorter ones and show that a useful approximation result, which holds under conditioning on the process not having exited the spatial interval, can be applied sequentially.
The plan of the paper is as follows. In Section 2 we perform the aforementioned change of variables to linearize the drift and prove Theorem 1.1 using an intermediate result on the exit from a small neighborhood of the origin. In Section 3 we introduce an auxiliary process, which is fully linear and thus allows us to derive certain properties of the exit problem through explicit calculations. In Section 4 we prove an approximation result which allows us to transfer these properties from the fully linear process to the case where only the drift is linear, as long as the timescales involved are not too large. Finally, in Section 5 we use an iterative scheme to lift this limitation, thereby finishing the proof of the intermediate result. In Section 6 we explain how the result of this paper fits our program on long-term behavior of diffusions near heteroclinic networks.
2. Proof of Theorem 1.1
As outlined above, we study the system first in a small neighborhood of the origin and then after the process has escaped this small neighborhood.
Let us start with the first part. The diffeomorphism
$f\,{:}\,\mathcal{I}\to{\mathbb R}$
introduced in (1.4) and its inverse
$g=f^{-1}$
provide a conjugation between the flow
$(S^t)$
and a linear flow:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn8.png?pub-status=live)
Note that the integrand in (1.4) is quadratic when x is close to zero and thus we have
$f(0)=0$
and
$f'(0)=1$
. Outside of
$\mathcal{I}$
, we define f so that f and f are bounded.
Let
$Y_\varepsilon(t)=f(X_\varepsilon(t))$
for times prior to the escape from
$\mathcal{I}$
. Itô’s formula and (2.1) then imply that this process satisfies the stochastic differential equation
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn9.png?pub-status=live)
for
$t<\tau_{\mathcal{I}}^\varepsilon$
, where
$\tilde\sigma(y)=f'(g(y))\sigma(g(y))$
and
$h(y)=f''(g(y))\sigma^2(g(y))$
. Due to the boundedness of f and f,
$\tilde\sigma$
and h are also bounded.
By Duhamel’s formula,
$Y_\varepsilon$
satisfies the integral equation
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn10.png?pub-status=live)
where
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU5.png?pub-status=live)
Due to our conventions on f, f outside of
$\mathcal{I}$
, the processes
$U_{\varepsilon}(t)$
and
$V_{\varepsilon}(t)$
are defined for all
$t\ge 0$
. Moreover, the boundedness of h immediately implies the boundedness of
$V_\varepsilon(t)$
:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn11.png?pub-status=live)
where
$\|\cdot\|_\infty$
is the sup-norm on
$[0,\infty)$
. The boundedness of
$\tilde\sigma$
yields a similar conclusion about the quadratic variation of
$U_\varepsilon(t)$
. Hence, the existence of constants
$c_1,c_2, N_0>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn12.png?pub-status=live)
is implied by the following exponential martingale inequality (see, e.g., Problem 12.10 in [Reference Bass5]):
Lemma 2.1. Let M(t) be a centered martingale with quadratic variation process
$\langle M\rangle_t$
. Then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU6.png?pub-status=live)
Let us take
$\beta\in(0,1)$
and set
$\mathcal{V}=g\left([\!-\varepsilon^{\beta},\varepsilon^{\beta}]\right)\subseteq \mathcal{I}$
. The following result describes the tail behavior of
$\tau_\mathcal{V}^\varepsilon$
, the exit time from
$\mathcal{V}$
. In particular, it says that, in the
$\varepsilon\downarrow 0$
asymptotics, the choice of the exit direction is distributed symmetrically independently of the exit time.
Theorem 2.1. Let
$Y_\varepsilon(0)=\varepsilon y$
, where
$|y|\leq K(\varepsilon)$
with
$K(\varepsilon)$
growing subpolynomially at 0. Then, for all
$\alpha>1$
,
$C\in{\mathbb R}$
, and any function
$c({\varepsilon})$
satisfying
$\lim_{{\varepsilon}\to 0}c({\varepsilon})=0$
, there is
$\beta_0\in(0,1)$
such that, for
$\beta\in(\beta_0,1)$
, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU7.png?pub-status=live)
We give the proof of Theorem 2.1 in Section 5.
After exit from
$\mathcal{V}$
, the deterministic dynamics dominates the evolution, which means that the exit time will be close to
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn13.png?pub-status=live)
the time it takes for
$X_0(t)$
to exit
$\mathcal{I}$
starting at
$g\left(\pm \varepsilon^{\beta}\right)$
. This is captured by the following standard large deviation estimates.
Proposition 2.1. Let
$X_{\varepsilon}(0)=g\left(\pm \varepsilon^{\beta}\right)$
. Then, for every
$\beta'\in(0,\beta)$
and subpolynomially decaying function
$c(\varepsilon)>0$
, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn14.png?pub-status=live)
and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn15.png?pub-status=live)
Proof. We start by showing that with overwhelming probability the exit happens through the endpoint that is on the same side as the starting point. Indeed, (2.3), (2.4), and (2.5) imply
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU8.png?pub-status=live)
and (2.8) follows.
To prove (2.7), let us introduce
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU9.png?pub-status=live)
and note that (2.3) implies
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU10.png?pub-status=live)
where we used (2.6), (2.8), (2.4), (2.5), and the subpolynomial decay of
$c(\varepsilon)$
. Similarly,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU11.png?pub-status=live)
□
Proof of Theorem 1.1. We have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU12.png?pub-status=live)
and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU13.png?pub-status=live)
for small enough
$\varepsilon$
, so the right-hand side is a function subpolynomially growing at 0. Let us define
$\theta^{\pm}_\varepsilon=T^\pm_{\varepsilon}-(\tau_\mathcal{I}^\varepsilon-\tau_\mathcal{V}^\varepsilon)$
. Due to (2.8),
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU14.png?pub-status=live)
where the error term (along with all subsequent error terms) is uniform in the starting points
$|x|\leq K(\varepsilon)$
. Note that the strong Markov property implies the conditional independence of
$\theta^{\pm}_\varepsilon$
and
$\tau_\mathcal{V}^\varepsilon$
given
$Y_\varepsilon(\tau_\mathcal{V}^\varepsilon)=\varepsilon^{\beta}$
. This, along with (2.7), allows us to give upper and lower estimates of the first term on the right-hand side:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU15.png?pub-status=live)
where
$c(\varepsilon)$
is an arbitrary positive function that decays subpolynomially. Now we may apply Theorem 2.1 with
$C={-t} +\lambda^{-1}\log|f(q_\pm)|$
to both sides, which concludes the proof.□
3. Linear system with additive noise
In this section we introduce an auxiliary process, which is a simpler special case of (2.2). Namely, we consider
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn16.png?pub-status=live)
where
$\sigma_0=\tilde{\sigma}(0)=\sigma(0)>0$
,
$|z|<K({\varepsilon})$
, and
$K({\varepsilon})$
grows subpolynomially at 0.
We will need a precise description of the exit of
$Z_\varepsilon$
from
$\mathcal{V}_Z=[\!-\varepsilon^{\beta}(1+\delta_\varepsilon),\varepsilon^{\beta}(1+\delta_\varepsilon)]$
for
$\beta\in (0,1)$
and any
$\delta_{\varepsilon}>0$
satisfying
$\delta_\varepsilon\downarrow 0$
as
$\varepsilon\downarrow 0 $
. Let us introduce a stopping time
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU16.png?pub-status=live)
and, for a constant C and a subpolynomially decaying at 0 function
$c({\varepsilon})$
, a deterministic time
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn17.png?pub-status=live)
We will often use
$C_\varepsilon=C-c(\varepsilon)$
. The first result of this section is a version of Theorem 2.1 for the process
$Z_{\varepsilon}$
with stronger control of the dependence on the initial point.
Lemma 3.1. For any
$\alpha>1$
and any subpolynomially decaying function
$c(\varepsilon)$
in the definition of
$t_{\varepsilon}$
, there is
$c>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU17.png?pub-status=live)
Proof. Duhamel’s formula gives an explicit solution to (3.1):
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn18.png?pub-status=live)
where
$N(t)=\int_0^t {\rm e}^{-\lambda s} \, {\rm d} W(s)$
. Plugging in
$\tau_\varepsilon$
for t, we obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU18.png?pub-status=live)
which is equivalent to
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU19.png?pub-status=live)
Therefore,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU20.png?pub-status=live)
By the martingale convergence theorem, N(t) converges to a random variable
$N_{\infty}$
as
$t\to\infty$
almost surely and in
$L^1$
. We claim that, in addition, there are
$c_1,c_2>0$
such that, for any
$L>0$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn19.png?pub-status=live)
Indeed, Lemma 2.1 implies
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU21.png?pub-status=live)
for some
$c_1,c_2>0$
, which proves the claim since
$\varepsilon^{\alpha-\beta}{\rm e}^{\lambda t_\varepsilon}={\rm e}^{-\lambda C_{\varepsilon}}$
.
Let us fix
$\gamma\in (\alpha-1,\alpha-\beta)$
and write
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU22.png?pub-status=live)
where
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU23.png?pub-status=live)
The random variables
$N(\tau_\varepsilon)$
and
$N(\tau_\varepsilon)-N_\infty$
are independent due to the strong Markov property. So the Gaussian tail of the maximum of the Brownian motion and (3.4) imply
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU24.png?pub-status=live)
The desired asymptotics of
$H_1(z,{\varepsilon})$
follows from the explicit form of the Gaussian density of the random variable
$z+\sigma_0 N_\infty$
.□
The next result is based on the fact that the distribution of
$Z_\varepsilon(t_\varepsilon)$
conditioned on non-exit is approximately uniform over
$[\!-\varepsilon^{\beta}(1+\delta_\varepsilon),\varepsilon^{\beta}(1+\delta_\varepsilon)]$
. In fact, this stronger statement is proved as an intermediate step. We first note that the density of any absolutely continuous random variable conditioned on a positive probability event is well defined.
Lemma 3.2. If
$f_{\varepsilon}^c(u)$
is the probability density of
$Z_{\varepsilon}(t_{\varepsilon})$
conditioned on
$\{\tau_{\varepsilon} > t_{\varepsilon}\}$
, then, for any
$\delta>0$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn20.png?pub-status=live)
Moreover, for any integrable function h with exponentially decaying tails and any subpolynomially growing function
$K(\varepsilon)$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn21.png?pub-status=live)
Proof. By (3.3) and the Dambis–Dubins–Schwartz theorem (see, e.g., [8, Section 3.4B]),
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU25.png?pub-status=live)
where
$r(t)=\sigma_0^2\left(1-{\rm e}^{-2\lambda t}\right)/(2\lambda)$
, and B is an auxiliary standard Brownian motion. Since
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU26.png?pub-status=live)
we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU27.png?pub-status=live)
This is a Gaussian random variable. Its density at a point
$u\in{\mathbb R}$
is given by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU28.png?pub-status=live)
If
$|u|\le{\varepsilon}^\beta$
, then
$|u{\varepsilon}^{\alpha-1-\beta}|\le {\varepsilon}^{\alpha-1}$
, so
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn22.png?pub-status=live)
uniformly over u and z satisfying
$|u|\le{\varepsilon}^\beta$
,
$|z|\leq K(\varepsilon)$
.
Therefore, (3.5) will follow from
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU29.png?pub-status=live)
where
$f_{\varepsilon}(z,u)$
is the sub-probability density of
$Z_\varepsilon(t_\varepsilon)$
on the event
$\{\tau_\varepsilon>t_\varepsilon\}$
. For this, it suffices to see that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn23.png?pub-status=live)
decays exponentially fast as
${\varepsilon}\downarrow 0$
, where
$g_{\varepsilon}(z,u)$
is the sub-probability density of
$Z_{\varepsilon}(t_{\varepsilon})$
on the event
$\{\tau_{\varepsilon} \le t_{\varepsilon}\}$
. Given that
$Z_{\varepsilon}(\tau_\varepsilon)=\varepsilon^{\beta}(1+\delta_\varepsilon)$
(similarly for
$-\varepsilon^{\beta}(1+\delta_\varepsilon)$
), we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU30.png?pub-status=live)
and thus
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn24.png?pub-status=live)
where
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU31.png?pub-status=live)
with
$\psi(\cdot,\cdot)$
as introduced in (1.5) and
$w={\rm e}^{\lambda(t_{\varepsilon}-t)}$
.
We claim that there is
$c>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn25.png?pub-status=live)
This, along with (3.9) and the fact that
$\psi_0(z)>\lambda^{1/2}\pi^{-1/2}\sigma_0^{-1}{\rm e}^{-\lambda K^2({\varepsilon})/\sigma_0^2}$
for all
${\varepsilon}$
and z satisfying
$|z|<K({\varepsilon})$
, will imply the desired exponential decay in (3.8).
Let us fix any
$w_0$
and find
$c_0>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU32.png?pub-status=live)
for
$w> w_0$
and all
${\varepsilon}>0$
. Then there is a constant
$c_0$
such that, for all
$|u|\le (1-\delta){\varepsilon}^\beta$
and
${\varepsilon}>0$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn26.png?pub-status=live)
If
$1 \le w \le w_0$
, then
$(w\varepsilon^{\beta}(1+\delta_\varepsilon)-u)^2\ge\Delta\coloneqq \varepsilon^{2\beta}\delta^2$
. So, denoting
$D={\varepsilon}^2\sigma_0^2(w^2-1)/(2\lambda)$
, we obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn27.png?pub-status=live)
The restriction on w implies
$0\le D \le c_1{\varepsilon}^2$
for some
$c_1$
. To maximize
$\psi(D,\Delta)$
, we compute
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU33.png?pub-status=live)
so
$\psi(D,\Delta)$
grows in
$D\in [0,\Delta]$
, and we obtain, from (3.12),
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn28.png?pub-status=live)
Combining (3.11) and (3.13), we obtain (3.10) and hence (3.8). Thus, (3.5) is proved.
To prove (3.6), we write
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU34.png?pub-status=live)
and notice that the first term on the right-hand side equals
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU35.png?pub-status=live)
while the second term decays much faster than
${\varepsilon}^{1-\beta}$
due to the decay assumption applied to h.
4. Proof of Theorem 2 for
$\alpha\in(1,1+\beta)$
We start by studying the deviations
$\Delta_{\varepsilon}(t)=Y_{\varepsilon}(t)-Z_{\varepsilon}(t)$
as long as both processes
$Y_{\varepsilon}(t)$
and
$Z_{\varepsilon}(t)$
are close to the origin. Let us fix
$\beta\in(0,1)$
and introduce the stopping times
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU36.png?pub-status=live)
Lemma 4.1. Suppose
$\alpha\in(1,1+\beta)$
,
$\beta'\in(\alpha-1,\beta)$
,
$L(\varepsilon)>0$
is a bounded function, and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn29.png?pub-status=live)
Then, for sufficiently small
${\varepsilon}>0$
, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn30.png?pub-status=live)
provided
$Y_\varepsilon(0)=Z_\varepsilon(0)={\varepsilon} y$
for every
$y\in(\!-2{\varepsilon}^\beta, 2{\varepsilon}^\beta)$
.
Proof. Recalling (2.3), we obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU37.png?pub-status=live)
where
$I_{\varepsilon}^{(1)}(t)=\int_0^t {\rm e}^{-\lambda s}(\tilde\sigma(Y_{\varepsilon}(s))-\sigma_0)\,{\rm d} W(s)$
and
$I^{(2)}_\varepsilon=\mathcal{O}(\varepsilon)$
.
Clearly,
$I_\varepsilon^{(1)}$
is a martingale satisfying
$\langle I_{\varepsilon}^{(1)}\rangle_t=\mathcal{O}\left({\varepsilon}^{2\beta}\right)$
for
$t\le \bar{\bar\tau}_{\varepsilon}$
, so for any
$c_0>0$
, there is
$c>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU38.png?pub-status=live)
by Lemma 2.1. Therefore
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU39.png?pub-status=live)
and on the complementary event we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU40.png?pub-status=live)
if
$c_0$
is chosen sufficiently small, which finishes the proof.□
Based on this approximation result and the calculation for
$Z_\varepsilon$
in the previous section, the following theorem proves Theorem 2.1 for
$\alpha$
not too large.
Theorem 4.1. Let
$\alpha\in(1,1+\beta)$
and let
$t_{\varepsilon}$
be as in (3.2). Then there is
$c>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn31.png?pub-status=live)
Proof. We start with an upper bound on
$\mathbb{P}(\bar\tau_\varepsilon>t_\varepsilon)$
in terms of
$Z_{\varepsilon}$
. Let us fix any
$\beta''\in(\beta,2\beta -(\alpha-1))$
, so that
$\beta'=\beta'' -\beta+\alpha -1\in(\alpha-1,\beta)$
, which will allow us to apply Lemma 4.1 several times in this proof. Let us take any family of events
$(B_{\varepsilon})_{{\varepsilon}>0}$
and estimate
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU41.png?pub-status=live)
Note that
$t_\varepsilon$
is of the form (4.1) with
$L({\varepsilon})={\rm e}^{-\frac{(\alpha-\beta)C_{\varepsilon}}{\lambda}}$
and thus (4.2) implies
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn32.png?pub-status=live)
where, for any
$\gamma>0$
, we use
$o_{\exp}(1)$
as a shorthand for
$o({\rm e}^{-{\varepsilon}^{-\gamma}})$
. Also,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU42.png?pub-status=live)
We need to approximate this in terms of the exit time of
$Z_{\varepsilon}$
instead of
$\bar\tau_{\varepsilon}$
. We do not have control over the difference of these two times in general as we can only control the difference of the processes until
$t_{\varepsilon}$
. Instead, we are going to set a different threshold for
$Z_{\varepsilon}$
to reach. Let
$\gamma\in (\beta,\beta'')$
,
$l_\varepsilon^1=\varepsilon^{\beta}$
, and
$l_\varepsilon^2=l_\varepsilon^1+{\varepsilon}^\gamma$
. This implies
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU43.png?pub-status=live)
where
$\tau^Z_{\varepsilon}$
is the exit time from
$[\!-l_\varepsilon^2,l_\varepsilon^2]$
for
$Z_{\varepsilon}$
, and thus
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU44.png?pub-status=live)
Combining this with (4.4), we obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn33.png?pub-status=live)
Next, we set
$l_\varepsilon^3=l_\varepsilon^{1}-\varepsilon^{\gamma}$
and define
$\eta_\varepsilon^Z$
to be the exit time of
$Z_{\varepsilon}$
from
$[\!-l_\varepsilon^3,l_\varepsilon^3]$
. Lemma 4.1 and the fact that
$\{\bar\tau_\varepsilon\leq t_\varepsilon;\ \sup_{0\leq t\leq t_\varepsilon\wedge\bar{\bar\tau}_\varepsilon}|\Delta_\varepsilon(t)|\leq\varepsilon^{\beta''}\}\subset \{\eta_\varepsilon^Z\le t_\varepsilon\}$
imply
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn34.png?pub-status=live)
Combining (4.5) and (4.6), we obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn35.png?pub-status=live)
This, Lemma 3.1, and our choice of
$l_\varepsilon^i$
,
$i=1,2,3$
, imply
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn36.png?pub-status=live)
To finish the proof, we will need
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn37.png?pub-status=live)
which holds since, due to (2.3),
$Y_{\varepsilon}(\bar\tau_\varepsilon)=\varepsilon^{\beta}$
is equivalent to
$\varepsilon^{-1}u+U_\varepsilon(\bar\tau_\varepsilon)+\varepsilon V_\varepsilon(\bar\tau_\varepsilon)>0,$
and so
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU45.png?pub-status=live)
due to the boundedness of
$V(\bar\tau_\varepsilon)$
and (2.5).
Using (4.9), we can write
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn38.png?pub-status=live)
Due to (4.2),
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn39.png?pub-status=live)
In the last convergence, we used (3.7) to compute
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU46.png?pub-status=live)
and we used (4.7) with
$B_{\varepsilon}\equiv \Omega$
, along with Lemma 3.1, to compute
$\mathbb{P}(\bar\tau_{\varepsilon}>t_{\varepsilon})=2 {\rm e}^{\lambda C}\psi_0(y){\varepsilon}^{\alpha-1}(1+o(1))$
, so
$\mathbb{P}(|Z_\varepsilon(t_\varepsilon)|<2\varepsilon^{\beta''})/\mathbb{P}(\bar\tau_{\varepsilon}>t_{\varepsilon})\to 0$
follows from our assumptions on
$\alpha$
,
$\beta$
, and
$\beta''$
.
Also, due to (4.2),
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU47.png?pub-status=live)
Using (4.7) with
$B_{\varepsilon}=\{Z_\varepsilon(t_\varepsilon)>2\varepsilon^{\beta''}\}$
,
$B_{\varepsilon}=\{Z_\varepsilon(t_\varepsilon)>0\}$
, and
$B_{\varepsilon}=\Omega$
, we can switch conditioning to that in terms of
$Z_{\varepsilon}$
:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU48.png?pub-status=live)
where both the left- and right-hand sides converge to
$1/2$
as
$\varepsilon\downarrow 0$
due to Lemma 3.2. Combining this with (4.10) and (4.11), noticing that all the o(1) terms in these estimates are independent of the starting point y, and using (4.8), we obtain (4.3), which completes the proof.
5. Extension to arbitrary timescales
The goal of this section is to extend Theorem 4.1 for arbitrary
$\alpha>1$
and thus prove Theorem 2.1. We set
$\theta=\alpha-\beta$
,
$L(\varepsilon)={\rm e}^{-\frac{\lambda}{\theta} C_\varepsilon}$
, and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU49.png?pub-status=live)
When
$\theta\in(1-\beta,1)$
, Theorem 4.1 applies and there is nothing new to prove. Here we study the case
$\theta\geq 1$
. Up to this point the only restriction on
$\beta$
was
$\beta\in(0,1)$
. Let us now set
$N=[\theta]+1\ge 2$
,
$\beta_0=\frac{1}{2}\left(1+\frac{\theta}{N}\right)$
, and assume
$\beta\in (\beta_0,1)$
throughout this section. We have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn40.png?pub-status=live)
We also define
$t_\varepsilon'=t_\varepsilon/N$
and
$t_{\varepsilon,k} = kt_\varepsilon'$
,
$k=0,1,\dots, N$
. Our plan is to track
$Y_{\varepsilon,k}=Y_{\varepsilon}(t_{\varepsilon,k})$
,
$k=0,1,\dots, N$
, using the results of the previous section on the short intervals
$[t_{\varepsilon,k},t_{\varepsilon,k+1}]$
.
The first step is the following lemma, which establishes that the process needs to stay close to the origin to delay the exit.
Lemma 5.1. We have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn41.png?pub-status=live)
for some
$C_1,C_2>0$
. In particular,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU50.png?pub-status=live)
Proof. Using the strong Markov property and applying Duhamel’s principle (2.3) and (2.4) to the initial condition y with
$|y|>{\varepsilon} K({\varepsilon})$
, we reduce the lemma to the estimate
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU51.png?pub-status=live)
and the desired inequality follows by (2.5) since
$\beta+\frac{\theta}{N}-1>0$
due to (5.1).□
We now collect some results needed for our iteration scheme.
Lemma 5.2. Let
$Y_{\varepsilon}(0) = \varepsilon y$
. Then there is
$c>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn42.png?pub-status=live)
Moreover, for any Lipschitz function h on
${\mathbb R}$
, exponentially decaying at
$\infty$
, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn43.png?pub-status=live)
Proof. The first claim follows from Theorem 4.1 with
$C/N$
in place of C and
$\alpha=\beta+\theta/N$
. Note that this value of
$\alpha$
belongs to
$(1,1+\beta)$
due to (5.1).
The second claim is a direct consequence of the first one and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn44.png?pub-status=live)
Once again, to prove this, we would like to use the result for the linear process. However, the estimate in (4.2) is insufficient when applied directly. Instead, let us note that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU52.png?pub-status=live)
where
$I_{\varepsilon}(t_\varepsilon')={\rm e}^{-\lambda C_{\varepsilon}/N}\int_0^{t_\varepsilon'}{\rm e}^{-\lambda t}\left(\sigma(Y_\varepsilon(t))-\sigma_0\right){\rm d} W(t)$
and choose any
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU53.png?pub-status=live)
The exponential martingale inequality (Lemma 2.1) and the Lipschitz continuity of
$\sigma$
imply
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU54.png?pub-status=live)
and thus
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU55.png?pub-status=live)
Note that
$\beta'-\theta/N-1+\beta>0$
due to the choice of
$\beta$
and
$\beta'$
. Now we can use the last display and the Lipschitz continuity of h to obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU56.png?pub-status=live)
as
$\varepsilon\downarrow 0$
. This and (3.6) imply (5.5), which completes the proof of the lemma.□
Finally, the next theorem implies Theorem 2.1 with
$\beta_0=\frac{1}{2}\left(1+\frac{\theta}{N}\right)$
.
Theorem 5.1. There is
$c>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn45.png?pub-status=live)
and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn46.png?pub-status=live)
Proof. We will use
$o_K(1)$
to denote any function that decays faster than any power of
$\varepsilon$
as
$\varepsilon\downarrow 0$
. It suffices to prove the theorem in the case where the function
$K({\varepsilon})$
grows fast enough as
${\varepsilon}\to0$
to guarantee that the right-hand side of (5.2) is
$o_K(1)$
. To see that the theorem will then follow in full generality, we just notice that enlarging the set of initial conditions y from
$\{|y|\le K({\varepsilon})\}$
to
$\{|y|\le K({\varepsilon})\vee |\log {\varepsilon}|\}$
reduces the situation to that special case.
We will prove by induction that, for every
$k=1,\dots, N$
, there is
$c>0$
such that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn47.png?pub-status=live)
where we explicitly indicate the dependence on the starting point
$Y_\varepsilon(0)=\varepsilon y$
as a subscript in
$\mathbb{P}_y$
for clarity. The case
$k=N$
is the desired result (5.6). The base of induction, the case
$k=1$
, is the first claim of Lemma 5.2. Let us make the induction step assuming that (5.8) holds for some k.
Lemma 5.1 and the Markov property allows us to write
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn48.png?pub-status=live)
Using the induction hypothesis (5.8), we obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU57.png?pub-status=live)
where the error term is uniform over
$|y'|\leq K(\varepsilon)$
. This means that the first term on the right-hand side of (5.9) can be written as
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn49.png?pub-status=live)
where the error term satisfies
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn50.png?pub-status=live)
where we used (5.4) in the last step with
$h(y)={\rm e}^{-cy^2}$
. The main term on the right-hand side of (5.10) can be estimated using Lemma 5.1 and (5.4) with
$h(y)=\psi_0(y)$
(so
$\int_{\mathbb R} h(y)\, {\rm d} y=1$
):
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn51.png?pub-status=live)
and this expansion holds uniformly over
$|y|\leq K(\varepsilon)$
.
Putting together (5.9), (5.10), (5.11) and (5.12) yields, for sufficiently small
$c'>0$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU58.png?pub-status=live)
This completes the induction step and finishes the proof of (5.6).
To prove (5.7), note first that (5.2) and the strong Markov property imply
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU59.png?pub-status=live)
Using (5.3), the integrand can be written as
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU60.png?pub-status=live)
where the error terms are uniform in y, and thus another application of (5.2) finishes the proof of (5.7).□
6. Rare transitions in heteroclinic networks
In this section we discuss, briefly and nonrigorously, the questions that led us to study the tails of exit times in detail. These questions originate in the long-term behavior of diffusions near heteroclinic networks in the vanishing noise limit. A heteroclinic network is a feature of the phase portrait associated with a vector field composed of multiple hyperbolic critical points (‘saddles’) connected to each other by heteroclinic orbits; see an example of a phase portrait with a heteroclinic network for a cellular flow in Figure 1.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_fig1.png?pub-status=live)
Figure 1: A heteroclinic network is the backbone of this phase portrait associated with a cellular flow.
Let us consider a diffusion process solving the Itô equation (1.2) in
${\mathbb R}^2$
(although, with minor modifications, the discussion below applies to higher dimensions as well) with drift b giving rise to a heteroclinic network. The typical behavior of such processes for times that are of the order of
$\log {\varepsilon}^{-1}$
was studied in [Reference Almada and Bakhtin1, Reference Bakhtin2, Reference Bakhtin3]. Its main features depend mainly on the linearization of the drift b near the saddle points and can be described as follows. Upon reaching a small neighborhood of a saddle, the process spends a long (logarithmic in
${\varepsilon}$
) time in that neighborhood where the drift is weak and eventually exits along the unstable manifold associated with the positive eigenvalue
$\lambda$
of the linearization. This manifold is composed of two outgoing heteroclinic orbits, so the dynamics chooses one of them and follows it for a time of the order of 1 until it reaches the next saddle where it will also eventually decide between two outgoing directions, etc. This description may seem to imply the picture where the limiting (as
${\varepsilon}\to 0$
) process is essentially a random walk on the directed graph of heteroclinic connections. However, the character of the limiting process is often not Markovian and depends on the linearizations of b near the saddles.
To see what is going on, let us consider a two-dimensional diffusion
$(X_{\varepsilon},Y_{\varepsilon})$
near a model saddle described by equations
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU61.png?pub-status=live)
driven by independent standard Wiener processes W and B. Here,
$\lambda>0$
and
$\mu>0$
can be viewed as coefficients of expansion and contraction, respectively. Assuming that
$(X_{\varepsilon}(0),Y_{\varepsilon}(0))=(0,1)$
, i.e. starting the process on the stable manifold of the saddle located at the origin, we are interested in the distribution of
$(X_{\varepsilon}(\tau_{\varepsilon}),Y_{\varepsilon}(\tau_{\varepsilon}))$
, where
$\tau_{\varepsilon}=\inf\{t\ge 0\,{:}\,|X_{\varepsilon}(t)|=1\}$
is the exit time from the strip
$[\!-1,1]\times{\mathbb R}$
. Since
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU62.png?pub-status=live)
where
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU63.png?pub-status=live)
we find that, for small
${\varepsilon}$
,
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn52.png?pub-status=live)
and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn53.png?pub-status=live)
where
$\rho=\mu/\lambda$
. This allows us to conclude that, as
${\varepsilon}\to 0$
, the distribution of the exit point
$Y_{\varepsilon}(\tau_{\varepsilon})$
depends cruciially on how
$\rho$
compares to 1. In particular, if
$\rho<1$
(i.e. the contraction is not as strong as expansion:
$\mu<\lambda$
) then the first term
${\varepsilon}^{\rho} |N(\infty)|^\rho$
dominates. It is positive and scales as
${\varepsilon}^\rho\gg {\varepsilon}$
, i.e. it is stronger than the noise magnitude
${\varepsilon}$
. Therefore, in this situation, at the next saddle point the system is most likely to stay on the same side of the heteroclinic network. If
$\rho> 1$
, then the probability of choosing either of the two outgoing connections at the next saddle approaches
$1/2$
as
${\varepsilon}\to0$
.
This analysis can be extended to more general initial conditions, to nonlinear drift and diffusion coefficients, to higher dimensions, and to sequences of saddles. The result is that for each sequence of saddles one can iteratively determine the asymptotic probability of realization of each next step along that sequence and the scaling asymptotics of the associated exit distributions. This was done in [Reference Almada and Bakhtin1, Reference Bakhtin2, Reference Bakhtin3]. The result is that at logarithmic timescales (the saddle exit times are typically logarithmic in
${\varepsilon}$
, see (6.1)), certain pathways in the network are typical but many pathways are not realized due to the largely one-sided exit distributions scaling as
${\varepsilon}^\alpha$
with
$\alpha<1$
, which in turn are due to insufficient contraction at saddles.
This is interesting per se, and among other applications gives an explanation of the poor vocabulary of excitation patterns in neural networks and similar dynamics modeled by Lotka–Volterra-type systems with small noise. However, this information is not sufficient to address questions about timescales that are longer than logarithmic, such as the limiting behavior of the invariant distribution. To answer these questions, one must quantify the probabilities of rare events corresponding to atypical exits from saddle points. This means that one needs to study probabilities of events like
$\mathbb{P}(Y_{\varepsilon}(\tau_{\varepsilon})\sim {\varepsilon})$
for
$Y_{\varepsilon}(\tau_{\varepsilon})$
given in (6.2). The second term in (6.2) is of the order of
${\varepsilon}$
, so ignoring many technical details we reduce this question to estimating
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn54.png?pub-status=live)
with
$\gamma=\frac{1}{\rho}-1$
. The last relation holds since
$N(\infty)$
has continuous Lebesgue density at 0.
This means that the probability of an atypical exit from the saddle is asymptotically polynomial in
${\varepsilon}$
, of the order of
${\varepsilon}^\gamma$
, which in turn means that one typically has to wait for time of the order of
${\varepsilon}^{-\gamma}\log {\varepsilon}^{-1}$
before one sees such a rare event happen. Ordering all exponents emerging in such calculations for all rare transitions,
$\gamma_1<\gamma_2<\cdots<\gamma_N$
, and introducing
$T_{k,{\varepsilon}}={\varepsilon}^{-\gamma_k}\log {\varepsilon}^{-1}$
, we see that, for
$t_{\varepsilon}$
satisfying
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqnU64.png?pub-status=live)
transitions can be classified into admissible (that typically occur many times up to
$t_{\varepsilon}$
) and rare (that typically do not occur at all up to
$t_{\varepsilon}$
). As
$t_{\varepsilon}$
crosses a level
$T_{k,{\varepsilon}}$
from below, some new transitions become available. Increasing
$t_{\varepsilon}$
gradually from 0 to values beyond
$T_{N,{\varepsilon}}$
creates a hierarchical structure of merging clusters and associated timescales such that, at each timescale, the system explores one cluster, making no transitions between different clusters.
This picture containing the description of the limit of invariant distribution, homogenization results, etc., is similar to the Freidlin–Wentzell picture of metastability and the associated hierarchy of cycles. The important difference is that in our picture the probabilities of rare events decay polynomially and the associated timescales grow polynomially, while the large deviation estimates in the Freidlin–Wentzell theory lead to exponentially decaying probabilities and exponentially growing transition times between the metastable states.
We do not have a rigorous derivation of a general precise version of the asymptotic relation (6.3). The difficulties that emerge are related to handling nonlinearities in the drift and diffusion terms, and to the fact that
$N(\infty)$
and
$M(\infty)$
are only approximations to the (mutually dependent) random variables
$N_{\varepsilon}(\tau_{\varepsilon})$
and
$M_{\varepsilon}(\tau_{\varepsilon})$
, where the
${\varepsilon}$
subscript of
$N_{\varepsilon}$
and
$M_{\varepsilon}$
refers to the fact that for the case of non-additive noise, these stochastic processes do depend on
${\varepsilon}$
. So, to realize this program, among other things we must either prove that the density of
$N_{\varepsilon}(\tau_{\varepsilon})$
uniformly converges to the Gaussian density in a small neighborhood of zero, or find other means to compare the distribution of
$N_{\varepsilon}(\tau_{\varepsilon})$
to the Gaussian at small scales, which requires going beyond the known weak convergence of distributions.
According to (6.1), if
$N(\infty)$
takes an atypically small value of the order of
${\varepsilon}^{\gamma}$
, then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200716080621857-0518:S0021900220000169:S0021900220000169_eqn55.png?pub-status=live)
i.e. exit takes an abnormally long time (the typical exit time corresponds to
$\gamma=0$
). In other words, the rare transitions determining the long-term behavior of diffusions near heteroclinic networks occur due to atypically long stays in the neighborhood of saddle points withstanding the repulsion in the unstable direction.
In the present paper (as well as in [Reference Bakhtin and Pajor-Gyulai4]), we study the polynomial decay of the distribution of exit times at scales described by (6.4). We believe that the method we propose here is applicable in the multi-dimensional situation, and we plan to give a rigorous treatment of it in upcoming publications.
Acknowledgement
Yuri Bakhtin gratefully acknowledges partial support from the NSF via grant DMS-1811444.