1 Introduction
The surplus process of an insurance company is given by the following process:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU1.gif?pub-status=live)
where u≥0 is the initial surplus and c>0 the constant premium rate. Here the aggregate claims process J={J t , t≥0} is a subordinator with Laplace exponent
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU2.gif?pub-status=live)
where ν is the Lévy measure of J supported on (0, ∞) satisfying the condition
$$\mu \,{\equals}\,{\bf E}[J_{1} ]\,{\equals}\,{\int}_0^\infty {x\nu (dx)} \,\lt\,\infty$$
. Here we suppose the safety loading condition holds, i.e. c>μ. The infinite-time horizon ruin probability Ψ(u) is defined by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU4.gif?pub-status=live)
The corresponding survival probability Φ(u) is defined by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU5.gif?pub-status=live)
For simplicity, we will suppose that Φ(u) has the first derivative g(u) which is of the polynomial growth, i.e.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU6.gif?pub-status=live)
for some k, p and integer-valued q.
The statistical inference for the ruin probability has been studied by many authors (see e.g. Frees, Reference Frees1986; Hipp, Reference Hipp1989; Croux & Veraverbeke, Reference Croux and Veraverbeke1990; Pitts, Reference Pitts1994; Bening & Korolev, Reference Bening and Korolev2002; Politis, Reference Politis2003; Mnatsakanov et al., Reference Mnatsakanov, Ruymgaart and Ruymgaart2008; Zhang & Yang, Reference Zhang and Yang2013). Statistical methodology has some advantages over analytic and probabilistic methods. On the one hand, the model can be more general. For example, no specific structure on the claim size distribution is assumed. On the other hand, in practical situations, instead of knowing the specific model one can only obtain the data on the surplus. Thus, statistical methodology can be directly used to analyse the insurances risk from the data. For more recent contributions on statistical inference of the ruin probability we refer the reader to Mnatsakanov et al. (Reference Mnatsakanov, Ruymgaart and Ruymgaart2008), Shimizu (Reference Shimizu2012) and Zhang & Yang (Reference Zhang and Yang2013).
In Zhang & Yang (Reference Zhang and Yang2013), the ruin probability for the pure-jump Lévy risk model is estimated and the key tool for estimation is the Pollaczek–Khinchin formula. The authors apply the Fourier method to transform the infinite sum of convolutions to a single integral and then construct the estimator of the ruin probability. In Mnatsakanov et al. (Reference Mnatsakanov, Ruymgaart and Ruymgaart2008), the authors consider an empirical-type estimator of the Laplace transform of the ruin probability, recovered it by a regularised Laplace inversion technique, and show the weak consistency in a sense of the integrated squared error (ISE). In Shimizu (Reference Shimizu2012), the author constructs an estimator of the Gerber–Shiu function for the Wiener–Poisson risk model in a manner similar to Mnatsakanov et al. (Reference Mnatsakanov, Ruymgaart and Ruymgaart2008) and shows the consistency in the ISE sense.
In this paper, we will estimate the ruin probability in the pure-jump Lévy risk model. Note that in the Lévy risk model there may exist an infinite number of jumps of small size in the finite time interval. For an insurance company, if the surplus has lots of small fluctuations, it is not easy to identify the inter-claim times. One feasible way of dealing with this problem is to observe the surplus process.
The rest of this paper is organised as follows. In sections 2 and 3 we will present the exact estimator of Φ(u) and study its properties. In section 4, we will do some simulations to show the finite sample size performance of the estimator. Finally, some conclusions are given in section 5. All the technical proofs will be presented in the Appendix.
2 Construction of the Estimator
2.1 Preliminaries: general notation
Throughout the paper, we use the primary notations and assumptions.
∙ A≲B signifies that there exists an universal constant k>0 such that A≤kB.
∙ Symbols
$$\buildrel {\rm P} \over \longrightarrow $$ and
$$\buildrel {\rm D} \over \longrightarrow $$ stand for the convergence in probability, and in distribution, respectively.
∙ L M is the Laplace transform for a function M: for s>0
$$L_{M} (s)\,{\equals}\,{\int}_0^\infty {e^{{{\minus}su}} M(u)du} $$
∙
$\left\Vert f \right\Vert {K} \,{\equals}\,\left( {{\int}_0^K {\left| {f(t)} \right|^{2} dt} } \right)^{{{1 \over 2}}} $ for any function f and K>0. In particular,
$$\left\Vert f \right\Vert\,{\equals}\,\left\Vert f \right\Vert_{\infty} $$ . We say that f∈L 2 (0, K) if
$$\left\Vert f \right\Vert_{K} \,\lt\,\infty$$ . In particular, f∈L 2(0, ∞) if
$$\left\Vert f \right\Vert\,\lt\,\infty$$ .
∙ For a stochastic sequence {X n , n≥1} and any real-valued sequence {R n , n≥1}, X n =O P (R n ) if, for any ϵ>0, there exists a constant d such that
$$P\left( {\left| {{{X_{n} } \over {R_{n} }}} \right|\,\gt\,d} \right)\leq {\epsilon}$$ for all n and X n =o P (R n ) if, for any ϵ>0,
$$P\left( {\left| {{{X_{n} } \over {R_{n} }}} \right|\,\gt\,{\epsilon}} \right)\to0$$ as n→∞.
∙
$${\mib {\cal N}} (a,b)$$ indicates the Gaussian distribution with mean a and variance b.
2.2 Estimator
As we know, the exact closed-form expression for Φ(u) is difficult to obtain, but the corresponding Laplace transform of Φ(u) can be obtained easily by Morales (Reference Morales2007). The Laplace transform of Φ(u) is
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU17.gif?pub-status=live)
where ρ=(μ)/(c).
By the Pollaczek–Khinchin formula for Φ(u) in Huzak et al. (Reference Huzak, Perman, Sikic and Vondracek2004), we can obtain Φ(0)=1−ρ.
Now we consider to construct an estimator of L
Φ(s). It follows from (2.1) that we have to estimate the parameter ρ and the function ψ
J
(s). Suppose that the observations consist of a discrete sample
$$\left\{ {J_{{t_{i}^{n} }} \!\mid\!t_{i}^{n} \,{\equals}\,ih;\,i\,{\equals}\,0,1,2,...,n} \right\}$$
for some fixed h>0. Let
$$\Delta _{i}^{n} J\,{\equals}\,J_{{t_{i}^{n} }} {\minus}J_{{t_{{i{\minus}1}}^{n} }} $$
, i=1, 2, … , n.
Immediately, an unbiased estimator for ρ is given by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU20.gif?pub-status=live)
Obviously, we can obtain an estimator of Φ(0) which is given by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU21.gif?pub-status=live)
Because J is a spectrally positive Lévy process, the Laplace exponent is denoted by ψ J (s) via
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU22.gif?pub-status=live)
Let us consider the empirical Laplace transform function of the increments of J, at each stage n (below s>0):
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU23.gif?pub-status=live)
Then we set
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU24.gif?pub-status=live)
Let
$$\widetilde{{\psi _{J} }}_{n} (s)$$
be an estimator of ψ
J
(s).
Therefore, we can define
$$\widetilde{{L_{\Phi } }}(s)$$
as an estimator of L
Φ(s) as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU27.gif?pub-status=live)
In order to estimate the original functions Φ(u), we will apply the L
2-inversion method proposed by Chauveau et al. (Reference Chauveau, Vanrooij and Ruymgaart1994) to
$$\widetilde{{L_{\Phi } }}(s)$$
. The L
2-inversion method is defined as follows, which is available for any L
2(0, ∞) functions.
Definition 2.1
Let m>0 be a constant. The regularised Laplace inversion
$$L_{m}^{{{\minus}1}} \,\colon\,L^{2} (0,\infty)\toL^{2} (0,\infty)$$
is given by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU30.gif?pub-status=live)
for a function g∈L 2(0, ∞) and t∈(0, ∞), where
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU31.gif?pub-status=live)
and a m =π −1cosh −1(πm).
Remark 2.2
It is well known that the norm L
−1 is generally unbounded, which causes the ill-posedness of the Laplace inversion. However,
$$L_{m}^{{{\minus}1}} $$
is bounded for each m>0, in particular
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU33.gif?pub-status=live)
For details and further information see Chauveau et al. (Reference Chauveau, Vanrooij and Ruymgaart1994).
Proposition 2.3
For s>0, we have
$$\widetilde{{L_{\Phi } }}(s) \,\notin\, L^{2} (0,\infty)$$
.
From Proposition 2.3, we know
$$\widetilde{{L_{\Phi } }}(s) \,\notin\, L^{2} (0,\infty)$$
for s>0. Hence the L
2-inversion method in Definition 2.1 cannot be applied directly. Below, let us define
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU36.gif?pub-status=live)
for arbitrary fixed θ>0.
It is obvious that Φ θ (u)∈L 2(0, ∞) and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU37.gif?pub-status=live)
Obviously,
$$L^{2}_{{\Phi _{\theta } }} \in L (0,\infty)$$
.
We define an estimator of
$$L_{{\Phi _{\theta } }} $$
as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU40.gif?pub-status=live)
We have
$$\widetilde{{L_{{\Phi _{\theta } }} }} \in L^{2} (0,\infty)$$
by (2.6) and (A.6).
For suitable m(n)>0, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU42.gif?pub-status=live)
Finally, we construct an estimator of Φ(u) as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU43.gif?pub-status=live)
3 Asymptotic Properties of Estimators
We first consider the asymptotic properties of
$$\widetilde{\rho }_{n} $$
and
$$\widetilde{{\psi _{J} }}_{n} (s)$$
.
Theorem 3.1
As mentioned in section 1, we suppose that the safety loading condition holds and
$${\int}_0^\infty {z^{2} \nu (dz)\,\lt\,\infty} $$
. Then, for s>0
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU47.gif?pub-status=live)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU48.gif?pub-status=live)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU49.gif?pub-status=live)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU50.gif?pub-status=live)
as
$$n\to\infty$$
.
Remark 3.2 By (3.1) and (3.3), it is easy to obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU52.gif?pub-status=live)
Now, we will present our main result which states a convergence in probability of the ISE.
Theorem 3.3
Suppose that the conditions of Theorem 3.1 are satisfied and Φ(u) has first derivative g(u) with polynomial growth. Then, for
$$m\left( n \right)\,{\equals}\,\sqrt {\left( {{{\left( n \right)} \over {\left( {{\rm log}\,n} \right)}}} \right)} $$
, B>0 and u>0, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU54.gif?pub-status=live)
Remark 3.4 The explicit integral expression for the estimator of the survival probability is
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU55.gif?pub-status=live)
where
$$\Psi _{{m(n)}} (y)\,{\equals}\,{\int}_0^{a_{{m(n)}} } {\cos h\,(\pi x){\rm cos}\,\left( {x\,{\rm log}\,(y)} \right)dx} $$
,
$a_{{m(n)}} \,{\equals}\,\pi ^{{{\minus}1}} \,{{\rm cos}h}^{{\minus}1} \left( {\pi m(n)} \right)$
and
$m\left( n \right)\,{\equals}\,\sqrt {\left( {{{\left( n \right)} \over {\left( {{\rm log}\,n} \right)}}} \right)} .$
4 Simulation Studies
In this part we provide some simulation results to illustrate the behaviour of our estimator. We assume that the Lévy measure is given by
$\nu (dx)\,{\equals}\,\lambda {1 \over {\mu _{0} }}e^{\!{{\minus}{1 \over {\mu _{0} }}x}} dx$
. Then (J
t
)
t≥0 is a compound Poisson process where the Poisson intensity is λ and the individual claim sizes are exponentially distributed with mean μ
0. Let the premium rate be c and c>λμ
0. In this case, the survival probability is given by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU60.gif?pub-status=live)
Because
$g(u)\,{\equals}\,\Phi '(u)\,{\equals}\,{{\mu _{0} \lambda } \over c}\left( {{1 \over {\mu _{0} }}{\minus}{\lambda \over c}} \right)e^{{{\minus}\left( {{1 \over {\mu _{0} }}{\minus}{\lambda \over c}} \right)u}} $
and c >λμ
0, we have
$g(u)\,\lt\,{{\mu _{0} \lambda } \over c}\left( {{1 \over {\mu _{0} }}{\minus}{\lambda \over c}} \right)$
, we have that g(u) satisfies the condition of the polynomial growth.
First, we give the simulation of survival probability when the initial surplus u=0. Let us take c=λ=10,
$\mu _{0} \,{\equals}\,{1 \over 2}$
and h=0.1. Obviously, we have Φ(0)=0.5 by (4.1). On Figure 1, we plot the true survival probability point and some mean points with sample sizes n=1,000; 8,000; 10,000; 50,000; 100,000; 500,000 as u=0, which are computed based on 500 simulation experiments.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_fig1g.jpeg?pub-status=live)
Figure 1 True point and mean points.
Now, Figure 1 shows that the results improve as the sample size increases. In order to give the better depiction, we use the following tables to compare data.
Table 1 shows some true value data and some mean data with sample sizes n=1,000; 8,000; 10,000; 50,000; 100,000; 500,000 as u=0, which are computed based on 500 simulation experiments. Table 2 shows the errors of the mean data and true value.
Table 1 True value and some estimation values of survival probability at u=0.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_tab1.gif?pub-status=live)
Table 2 Error of true value and some estimation values in Table 1.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_tab2.gif?pub-status=live)
On Table 2, we can find that the errors are very small when the sample size n≥8,000.
If the initial surplus u>0, we plot the true survival probability curve and some mean points with sample sizes n=1,000; 8,000; 10,000; 50,000; 100,000; 500,000. Let us take c=λ=10,
$\mu \,{\equals}\,{1 \over 2}$
, h=0.1 and θ=0.075.
Tables 3 and 4 give some data to show the behaviour of our estimator as the initial surplus u>0. Table 3 shows some true values and some simulation results when the initial surplus u>0. Table 4 shows the errors of some simulation results and true value.
Table 3 True value and some estimation values of survival probability at u>0.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_tab3.gif?pub-status=live)
Table 4 Error of true value and some estimation values in Table 3.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_tab4.gif?pub-status=live)
Figure 2 shows the results improve as the sample size increases. However, when the sample size n<10,000, we can find that the simulation results are not good. This result may explain the slow rate of the estimator. Therefore, our simulations need to choose a large sample size (e.g. n≥100,000). On Table 4, the errors is very small when the sample size n≥100,000.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_fig2g.jpeg?pub-status=live)
Figure 2 True curve and mean points.
5 Conclusions
In this paper, we adopted the regularised Laplace inversion technique since it was an easy way to construct an estimator of the Laplace transform of the survival probability. The rate of convergence for the estimator is logarithmic. Although the convergence rate is slower than that in Croux & Veraverbeke (Reference Croux and Veraverbeke1990), Bening & Korolev (Reference Bening and Korolev2002), Zhang & Yang (Reference Zhang and Yang2013), the procedure is easy to manipulate in practice.
Acknowledgements
This research is supported by the Chinese NFSC (11571189, 11501304).
Appendix
Proof of Proposition 2.3: The first derivative of
$$\widetilde{{\psi _{J} }}_{n} (s)$$
is denoted by
$$\widetilde{\psi }_{{J_{{}} n}}^{{}} '(s)$$
as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU67.gif?pub-status=live)
The second derivative of
$$\widetilde{{\psi _{J} }}_{n} (s)$$
is denoted by
$$\widetilde{{\psi _{J} }}_{n} ''(s)$$
as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU70.gif?pub-status=live)
By Taylor expansion we obtain
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU71.gif?pub-status=live)
By (A.1), we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU72.gif?pub-status=live)
By (A.2) and (A.4), we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU73.gif?pub-status=live)
Thanks to c μ, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU74.gif?pub-status=live)
almost surely.
Therefore, by (2.6) and (A.6), we know that
$$\widetilde{{L_{\Phi } }}(s)\notinL^{2} (0,\infty)$$
.
Proof of Theorem 3.1: First, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU76.gif?pub-status=live)
Because J has independent and stationary increments, we know that the random variables
$$\left\{ {\Delta _{i}^{n} J,i\,{\equals}\,1,2,...,n} \right\}$$
are independent and identically distributed.
Thanks to c μ,
$${\int}_0^\infty {z^{2} \nu (dz)\,\lt\,\infty} $$
and (2.4), we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU79.gif?pub-status=live)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU80.gif?pub-status=live)
By law of large numbers and central limit theorem it follows that the (3.1) and (3.3) are right.
Next, we consider the (3.2) and (3.5).
Because of
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU81.gif?pub-status=live)
and the properties of
$$\left\{ {\Delta _{i}^{n} J,i\,{\equals}\,1,2,...,n} \right\}$$
, we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU83.gif?pub-status=live)
and
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU84.gif?pub-status=live)
as
$$n\to\infty$$
.
Since
$${\minus}{1 \over h}{\rm log}\,(x)$$
is a continuous function defined on (0, ∞) and its first derivative is not equal 0 on (0, ∞), we may make use of Continuous Mapping Theorem and Delta Theorem to obtain (3.2) and (3.5).
For the proof of Theorem 3.3 we need the following Lemma A.1.
Lemma A.1 Suppose that, for a function f∈L
2(0,∞) with the derivative f′,
${\int}_0^\infty {\left[ {t\left( {t^{{{1 \over 2}}} f(t)} \right)^{\prime} } \right]^{2} t^{{{\minus}1}} dt\,\lt\,\infty} $
, then
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU88.gif?pub-status=live)
Lemma A.1 can be essentially obtained by the proof of Theorem 3.2 in Chauveau et al. (Reference Chauveau, Vanrooij and Ruymgaart1994) and it shows that
$$L_{n}^{{{\minus}1}} $$
can be a Laplace inversion asymptotically in ISE sense.
Proof of Theorem 3.3: Let us first observe that (see (2.13))
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU90.gif?pub-status=live)
In order to deal with the bias part, i.e. the second term on the right-hand side of (A.9), let us write
$$\Phi _{\theta }^{\prime} \,{\equals}\,g_{\theta } $$
and note that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU92.gif?pub-status=live)
It is obvious that
$$g^{2} (x){\plus}\Phi ^{2} (x)\,{\lesssim}\,1{\plus}\,\mid\!x\!\mid\,^{C} $$
for some integer-valued C. Thus
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU94.gif?pub-status=live)
By Lemma A.1, we may conclude that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU95.gif?pub-status=live)
Next, let us consider the first term on the right-hand side of (A.9), the variance part. It is immediate from (2.1), (2.6), (2.10) and (2.11) that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU96.gif?pub-status=live)
Exploiting (A.6) and the fact that
${\bf P}\left( {\left\{ {\omega \in\Omega ;\,\widetilde{\rho }_{n} \,{\equals}\,1} \right\}} \right)\,{\equals}\,0$
, it follows after some algebra that, almost surely, the right-hand side of (A.11) is bounded by
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU98.gif?pub-status=live)
where
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU99.gif?pub-status=live)
By Theorem 3.1, it follows that
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU100.gif?pub-status=live)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU101.gif?pub-status=live)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU102.gif?pub-status=live)
as n→∞.
Combining (A.12), (A.13), (A.14) and (A.15) yields
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU103.gif?pub-status=live)
Combining (2.8), (A.10) and (A.16), we have
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20180813112737215-0225:S1748499517000100:S1748499517000100_eqnU104.gif?pub-status=live)
With an optimal
$m(n)\,{\equals}\,\sqrt {{n \over {{\rm log}\,n}}} $
balancing the two terms on the right-hand side of (A.17), the order becomes O
P
((log n)−1).