1. Introduction
Insurance is an efficient and popular risk hedging tool. An insurance contract is usually composed of an indemnity function and a premium, where the premium needs to be paid ex ante by the insured to the insurer while the insurance indemnity is provided ex post from the insurer to the insured. Traditional indemnity functions include, for example, the stop-loss and proportional functions. Other forms of indemnity function are also possible and their optimality has been studied since the seminal works of Borch (Reference Borch1960) and Arrow (Reference Arrow1973). Under various objective functions and premium principles and taking into account of more sophisticated economic factors, considerable advancements have been achieved in the literature. We refer to Chi and Tan (Reference Chi and Tan2011), Chi (Reference Chi2012), Cheung et al. (Reference Cheung, Chong and Lo2019), Ghossoub (Reference Ghossoub2019b), Boonen and Ghossoub (Reference Boonen and Ghossoub2019) for recent developments.
It is realistic that in addition to the insurable loss, the insured is also exposed to uninsurable risk such as inflation or catastrophe risk. These risk classes are jointly denoted as background risk and have received considerable attention since the work of Doherty and Schlesinger (Reference Doherty and Schlesinger1983). They show that the presence of background risk may affect the insured’s demand of insurance, which is particularly the case if the background risk is correlated with the insurable loss. Several directions of introducing background risk in the optimal insurance problem have been proposed in the literature. One direction is on additive background risk, where the insured’s total risk is the sum of insurable risk and background risk. In this way, Gollier (Reference Gollier1996) shows that when the uninsurable loss increases with respect to the insurable loss, the optimal indemnity function is of disappearing deductible form. Dana and Scarsini (Reference Dana and Scarsini2007) study the Pareto-optimal insurance contract in the presence of background risk, where the qualitative properties of the optimal contract were derived under the assumption of stochastic increasingness. Chi and Wei (Reference Chi and Wei2018) investigate the optimal insurance under background risk and with higher order risk attitude. They also establish the optimality of the stop-loss insurance under specific dependence structures between the insurable risk and background risk. An extension is given by Chi and Wei (Reference Chi and Wei2020) for general dependence structures where the authors impose a constraint to avoid the moral hazard and present a very general result for the optimal insurance. Although the general solution to their model is implicit, many specific cases are analyzed and the corresponding optimal indemnity functions are derived. The aforementioned studies are all within the expected utility framework. Recently, Chi and Tan (Reference Chi and Tan2021) extend the study into the mean–variance framework. Their methodology allows the derivation of the optimal indemnity function for a very general dependence structure between the insurable risk and background risk.
The main focus of this article is on multiplicative background risk. To gain a comprehensive view of this type of problem, we refer the interested readers to Franke et al. (Reference Franke, Schlesinger and Stapleton2006). Multiplicative background risk may arise from the possibility of the insurer to default. The coverage is then the product of the promised indemnity and some random variable distributed on [0,1]. This type of background risk is then interpreted as counterparty risk or default risk.
To model counterparty risk, there are roughly two streams in the literature. First, if there is only one insured in the market, default of the insurer can be modelled explicitly as a function of the indemnity function. In this way, optimal indemnities are derived by Cai et al. (Reference Cai, Lemieux and Liu2014), and the pricing is studied by Filipović et al. (Reference Filipović, Kremslehner and Muermann2015). The latter study is extended by Boonen (Reference Boonen2019) to the case with multiple policyholders, but under the assumption of exchangeable multivariate risk. Second, one assumes that the market with policyholders is “large”, so that individual insurance transactions do not impact the likelihood of default. This assumption is imposed by Cummins and Mahul (Reference Cummins and Mahul2003), Bernard and Ludkovski (Reference Bernard and Ludkovski2012), and Li and Li (Reference Li and Li2018), and in this paper we also impose this assumption. Cummins and Mahul (Reference Cummins and Mahul2003) study the optimal insurance problem when the insurer has a positive probability to default and the insured and insurer have divergent beliefs about this probability. Their study is probably closest to Bernard and Ludkovski (Reference Bernard and Ludkovski2012), who investigate the impact of counterparty risk on optimal insurance when the insurer’s default probability depends on the loss incurred by the insured. The optimal indemnity function is derived in an implicit way for the case where the insurer is risk neutral. Finally, Li and Li (Reference Li and Li2018) derive the optimal indemnity function in a Pareto-optimal insurance problem when the loss and recovery rate are negatively correlated for the cases where the information is symmetric and where the information is asymmetric. All these papers focus primarily on expected utility preferences of the insured, while our focus is on a mean–variance objective of the insured.
In this article, we explicitly focus on the set of indemnity functions that are incentive compatible, which plays key role in alleviating the ex post moral hazard. For example, if the slope of indemnity function is larger than 1 at some point or the indemnity function has a discontinuous upward jump, the insured would be incentivized to create an increase in the insurable loss. Such behavior is called moral hazard. To remove the moral hazard issue, Ghossoub (Reference Ghossoub2019a) introduces a state-verification cost such that the insurer could verify the state of real world by paying to a third party some extra cost. Another popular way to mitigate the moral hazard issue, as proposed by Huberman et al. (Reference Huberman, Mayers and Smith1983), is by requiring both the insured and insurer to pay more when the loss becomes larger. This is also named as the no-sabotage condition by Carlier and Dana (Reference Carlier and Dana2003). Note that such condition is also considered by Chi and Wei (Reference Chi and Wei2020) and Chi and Tan (Reference Chi and Tan2021) where the additive background risk is the focus.
We summarize the main contributions of this article as follows:
-
First, to the authors’ best knowledge, this article is the first one that studies the counterparty-risk-based optimal insurance problem under mean–variance preferences of the insured and the incentive-compatibility condition. Through the marginal indemnity function formulation, we present a general result which characterizes the optimal indemnity function implicitly to the problem under counterparty risk within a mean–variance framework. It is shown that the optimal indemnity function depends on the first and second order expectations of the counterparty risk conditioned on the realized insurable loss.
-
Second, without assuming any specific dependence structure between the counterparty risk and the insurable loss, we derive the explicit structure of the optimal indemnity function based on the element-wise minimizer to our problem under some mild assumptions on the counterparty risk. We point out that the problem in Chi and Tan (Reference Chi and Tan2021) could also be solved using the marginal indemnity function approach, which provides an alternative approach to study the additive background risk model within the mean–variance framework.
The rest of this article is structured as follows. Section 2 sets up the problem. Section 3 first characterizes the solution to the main problem in an implicit way, and then unveils the explicit structure of the optimal indemnity function based on its implicit characterization. Section 4 studies two special cases of the main problem. Section 5 presents some numerical examples illustrating the main result of this paper. Section 6 concludes the paper and gives directions for future research. All the proofs are delegated to Appendix A.
2. Problem formulation
We confine ourselves to a one-period economy. Suppose there is a decision maker (DM, also called insured) who is faced with a non-negative, bounded random loss X whose support is [0,M] (i.e., the set of numbers which have non-zero probability densities). The cumulative distribution function and density function of X are given by F(x) and f(x), respectively. The DM would like to purchase an insurance contract $(I,\pi)$ where I is the indemnity function and $\pi(I)$ is the premium principle that is used by the insurer.
There exists counterparty risk in the sense that the insurer may fail to pay its promised indemnity as per the contract at the end of period. We assume that the coverage received by the DM is given by $I(X)\cdot Y$ , where Y is a random variable distributed over [0,1] and may be correlated with X. In this paper, we follow Cummins and Mahul (Reference Cummins and Mahul2003) and Bernard and Ludkovski (Reference Bernard and Ludkovski2012) by assuming that the default event is exogenous in the sense that it is not affected by the DM’s transactions. To avoid some trivial cases of the default event, we adopt the following assumption throughout the paper.
Assumption 1
-
(i) $\mathbb{P}(Y=1)<1$ .
-
(ii) $\mathbb{P}(Y=0|X=x)<1$ for all $x\in[0,M]$ .
Assumption 1(ii) states that regardless of the DM’s loss, the probability of that the insurer recovers some partial liability (e.g., through selling its remaining assets) is positive. Under this setting, the end-of-period loss of the DM is
In insurance or reinsurance, some policies provide the DM an incentive to misreport the loss. For example, a franchise deductible indemnity function, that is $I(x)=x\cdot \unicode{x1d7d9}_{[d,\infty)}(x)$ for some $d\ge0$ , may incentivize the DM to over-report the loss or create an increment in the loss; a truncated excess-of-loss indemnity function, that is $I(x)=(x-d_1)_+\cdot\unicode{x1d7d9}_{[0,d_2]}(x)$ for some $d_2\ge d_1\ge0$ , may incentivize the DM to under-report the loss if it exceeds $d_2$ . Such behavior is called ex post moral hazard and should be seriously treated. A popular way in the literature to handle this issue is to restrict the indemnity functions to the following class:
The indemnity functions belonging to the set $\mathcal{I}$ are said to satisfy the incentive-compatibility or no-sabotage condition (Huberman et al. Reference Huberman, Mayers and Smith1983; Carlier and Dana, Reference Carlier and Dana2003). The advantages of limiting the indemnity function to the class $\mathcal{I}$ are twofold. First, the indemnity function is always non-decreasing, that is $I'(x)\ge 0$ , and satisfies $0\le I(x)\le x$ . That means, the indemnity is increasing with respect to the loss and can never be negative nor exceed the loss x generated by the DM’s loss. Second, one unit increment of loss cannot be compensated by more than one unit of indemnity. For any $I\in\mathcal{I}$ , I(X) and $X-I(X)$ are comonotonic with respect to the loss X. In particular, if the insurable loss X increases, then both the indemnity I(X) and the retained loss $X-I(X)$ increase.
For the premium, we assume that it is based on the coverage $I(X)\cdot Y$ rather than the promised indemnity I(X). We further assume that the premium is determined as a function of the expected coverage:
where $h(\!\cdot\!)$ is some differentiable function satisfying $h(0)=0$ and $h'(x)>1$ for $x\ge 0$ . A special case of h is $h(x)=(1+\theta)x$ with $\theta>0$ , which leads to the expectation premium principle.
We follow Chi and Tan (Reference Chi and Tan2021) and study the optimal insurance problem within a mean-variance framework. We focus on the following general problem with counterparty risk.
Problem 1 (Main problem)
where $\mathcal{M}(z_1,z_2)$ is increasing with respect to both $z_1$ and $z_2$ .
Problem 1 accommodates a wide range of mean-variance problems. For example, if let $\mathcal{M}(z_1,z_2)=z_1+\frac{\mathcal{B}}{2}z_2$ for some $\mathcal{B}\ge 0$ , the traditional mean-variance criterion is recovered, where $\mathcal{B}$ measures the DM’s aversion towards volatility.
We remark that the optimal indemnity function within the mean–variance framework for additive background risk model has been derived by Chi and Tan (Reference Chi and Tan2021) through a constructive approach with stochastic ordering techniques. The marginal indemnity function formulation, which is obtained through the calculus of variations, not only helps us to solve the multiplicative background risk (or counterparty risk) model but also provides an alternative way to solve the additive background risk model.
3. Optimal indemnity function
Solving Problem 1 usually takes a two-step procedure. In the first step, we fix the mean (i.e., the first argument of $\mathcal{M}(\cdot,\cdot)$ ) and minimize the variance (i.e., the second argument of $\mathcal{M}(\cdot,\cdot)$ ):
where c is a constant. As problem (3.1) depends on c, we denote its solution by $I_c$ . In the second step, we search for the optimal c such that $\mathcal{M}(c,\text{Var}(L(I_c,\pi)))$ reaches its minimum. The second step is a one-dimensional problem, and can be solved using standard techniques. The first step plays a vital role as it identifies the optimal indemnity function with c as its parameter. In the sequel, we focus on this first step, and thus on the problem (3.1).
Let $\psi_1(x)=\mathbb{E}[Y|X=x]$ and $\psi_2(x)=\mathbb{E}[Y^2|X=x]$ . Under Assumption 1, it follows that $\psi_1(x)>0$ and $\psi_2(x)>0$ for all $x\in[0,M]$ . Apply the conditional variance formula to the objective function of (3.1) gives
The constraint of (3.1) reduces to
Since $\left(h(x)-x\right)'=h'(x)-1>0$ and $h(0)=0$ , for any $c\ge \mathbb{E}[X]$ , the equation $h(x)-x=c-\mathbb{E}[X]$ has only one solution, which is denoted as $x^{\ast}$ . Based on the above simplifications, problem (3.1) reduces to
Solving problem (3.2) is equivalent to solving its Lagrangian dual problem:
where $\lambda\in\mathbb{R}$ is the Lagrangian coefficient.
The following lemma characterizes the optimal indemnity function to problem (3.3). In this lemma, $\unicode{x1d7d9}_{A}(t)$ is defined as the indicator function: $\unicode{x1d7d9}_{A}(t)=1$ if $t\in A$ and $\unicode{x1d7d9}_{A}(t)=0$ otherwise.
Lemma 3.1 Let Assumption 1 hold, and
Then, $I^{\ast}(x)=\int_{0}^{x}\eta^{\ast}(t)dt$ is an optimal solution to problem (3.3) if and only if
where
and $\xi(t)\in[0,1]$ is such that $I^{\ast}\in\mathcal{I}$ .
Lemma 3.1 characterizes the optimal marginal indemnity functions. However, the optimal indemnity functions given by Lemma 3.1 are implicit since $I^{\ast}$ also appears in $L(t;I^{\ast},\lambda)$ . Nevertheless, Lemma 3.1 provides insights about the explicit structure of $I^{\ast}$ , which will be derived in detail in the rest of this section. We remark that in deriving the above result, the calculus of variations plays important role. In recent years, such technique has been widely applied to obtain an implicit characterization of the optimal indemnity function under other preference functionals. See, for example, Chi and Wei (Reference Chi and Wei2020) and Chi and Zhuang (Reference Chi and Zhuang2020).
For now, let $\lambda$ be fixed and
It is easy to verify that $\phi_{\lambda}(x)$ is the element-wise minimizer to problem (3.3).Footnote 1 The following assumption is needed to proceed.
Assumption 2 The mapping $x\mapsto \frac{\psi_1(x)}{\psi_2(x)}$ is continuously differentiable.
Under Assumption 2, $\phi_{\lambda}(x)$ is also continuously differentiable over [0,M]. As such, the whole domain [0,M] could be partitioned as per the first order derivative of $\phi_{\lambda}(x)$ such that
where
and m is the smallest positive integer required for such a partition. Throughout the rest of this paper, m is assumed to be finite. Under this partition rule, we have $j_{i+1}\neq j_i$ and $|j_{i+1}-j_i|=1$ for $i=1,2,\dots,m-1$ . Let $x_{i-1}=\inf\left\{x\;:\; x\in S_{i,j_i}\right\}$ for $i=1,2,\dots,m$ , we have $0=x_0\le x_1\le\cdots\le x_{m-1}\le x_m=M$ . Intuitively, $x_i=\sup\left\{x\;:\;x\in S_{i,j_i}\right\}$ for $i=1,2,\dots,m$ . These points $\{x_i\}_{i=1,2,\dots,m-1}$ are referred to as the change points (see also Chi and Tan, Reference Chi and Tan2021). Our goal is to obtain the explicit structure of the optimal indemnity function over each piece $S_{i,j_i}$ . For the ease of presentation, the following layer-type indemnity function is defined:
The following theorem gives the optimal parametric indemnity function over each $S_{i,j_i}$ for $i=1,2,\dots,m$ .
Theorem 3.1 Let Assumptions 1 and 2 hold. For problem (3.3), the optimal indemnity function is given by $I^{\ast}(x)$ such that, for $x\in S_{m,j_m}$ ,
-
(1). if $j_m=1$ , then $I^{\ast}(x)=I^{\ast}(x_{m-1})+(x-\gamma_{m,1})_+$ for some $\gamma_{m,1}\in[x_{m-1},M]$ ,
-
(2). if $j_m=2$ , then $I^{\ast}(x)=\min\left\{\max\left\{\phi_{\lambda}(x),I^{\ast}(x_{m-1})\right\},I^{\ast}(x_{m-1})+x-x_{m-1}\right\}$ ,
-
(3). if $j_m=3$ , then $I^{\ast}(x)=I^{\ast}(x_{m-1})+I_{(x_{m-1},\gamma_{m,3}]}(x)$ for some $\gamma_{m,3}\in[x_{m-1},M]$ ,
and for $x\in S_{i,j_i}$ , $i=1,2,\dots,m-1$ ,
-
(4). if $j_i=1$ , then $I^{\ast}(x)=I^{\ast}(x_{i-1})+I_{(\gamma_{i,1},\gamma_{i,1}+I^{\ast}(x_i)-I^{\ast}(x_{i-1})]}(x)$ for some $\gamma_{i,1}\in[x_{i-1},x_i]$ ,
-
(5). if $j_i=2$ , then $I^{\ast}(x)=\min\{\max\{\phi_{\lambda}(x),I^{\ast}(x_i)+x-x_i,I^{\ast}(x_{i-1})\},$ $I^{\ast}(x_{i-1})+x-x_{i-1},I^{\ast}(x_i) \}$
-
(6). if $j_i=\!3$ , then $I^{\ast}(x)=\!I^{\ast}(x_{i-1})\!+\!x\!-\!x_{i-1}\!-\! I_{(\gamma_{i,3},\gamma_{i,3}+x_i-x_{i-1}-(I^{\ast}(x_i)-I^{\ast}(x_{i-1}))]}(\!x\!)$ for some $\gamma_{i,3}\in[x_{i-1},x_{i}]$ .
An illustration of the optimal indemnity function $I^{\ast}$ is shown in Figure 1, where the red line denotes $I^{\ast}$ and the green dashed line denotes the element-wise minimizer $\phi_{\lambda}$ to problem (3.3). We can see that $\phi_{\lambda}$ does not satisfy the incentive compatibility condition, so the optimal indemnity function $I^{\ast}$ is obtained based on the slope of $\phi_{\lambda}$ . In Figure 1, there are two change points, that is $x_1$ and $x_2$ , and the domain of loss is partitioned into three pieces. We obtain the parametric form of the optimal indemnity function on each piece by using Theorem 3.1.
-
Over $[0,x_1]$ , $\phi'_{\!\!\lambda}(x)<0$ , so $j_1=3$ , and
\begin{equation*} I^{\ast}(x)=x-I_{(\gamma_{1,3},\gamma_{1,3}+x_1-I^{\ast}(x_1)]}(x). \end{equation*} -
Over $[x_1,x_2]$ , $0\le \phi'_{\!\!\lambda}(x)\le 1$ , so $j_2=2$ , and
\begin{equation*} I^{\ast}(x)=\min\left\{\max\left\{\phi_{\lambda}(x),I^{\ast}(x_2)+x-x_2,I^{\ast}(x_{1})\right\},I^{\ast}(x_{1})+x-x_{1},I^{\ast}(x_2) \right\}. \end{equation*} -
Over $[x_2,M]$ , $\phi'_{\!\!\lambda}(x)>1$ , so $j_3=1$ , and
\begin{equation*} I^{\ast}(x)=I^{\ast}(x_2)+(x-\gamma_{3,1})_+. \end{equation*}
Theorem 3.1 shows the applicability of Lemma 3.1 in practice and gives the explicit structure of the optimal indemnity function. It reduces the dimension of the original optimization problem from $\infty$ to at most $2m-1$ (i.e., $I^{\ast}(x_i)$ for $i=1,2,\dots,m-1$ , $\gamma_{i,1}$ and $\gamma_{i,3}$ for $i=1,2,\dots,m$ , and $\lambda$ ).
We remark that the partition (3.4) varies with respect to the Lagrangian coefficient $\lambda$ , which is a key parameter pertaining to the premium $\pi$ . Therefore, if the number of change points (i.e., $m-1$ ) is large, numerically optimizing the parameters in Theorem 3.1 is still computationally expensive. To slightly simplify the computation, the following proposition is given.
Proposition 3.1 Let Assumption 1 hold. If $\lambda>0$ , then for problem (3.3) the optimal indemnity function over $[0,\frac{\lambda}{2}]$ is given by
for some $d\in[0,\frac{\lambda}{2}]$ .
Proposition 3.1 shows that the optimal indemnity function is of the stop-loss form in a neighborhood of 0 if $\lambda>0$ . To derive the general solution, only the interval $[\frac{\lambda}{2},M]$ needs to be partitioned as per (3.4) and (3.5).
Remark 3.1 The marginal indemnity function approach applied in this paper is also applicable to the additive-background-risk-based optimal insurance problem within the mean-variance framework. The only difference is that for the additive-background-risk-based problem the partition does not rely on the Lagrangian coefficient. In Chi and Tan (2021), a constructive approach, together with some stochastic ordering technique, was applied to identify the parametric form of the optimal indemnity function on each piece of the domain, which yields different Lagrangian coefficients to be optimized over different pieces. Through Theorem 3.1, we are able to show that at the optimum all the Lagrangian coefficients in their result are equal, which is a supplementary finding to their study.
4. Two special dependence structures
In this section, we study two special cases of our problem, that is when the counterparty risk Y is independent of X and when Y is a decreasing function of X. In both special cases, we will derive a much simpler solution.
4.1. Y is independent of X
In this section, we study the indemnity function for the case where Y is independent of X. This happens when the DM’s loss does not affect the solvency status of the insurer. In such a case, $\frac{\psi_1(x)}{\psi_2(x)}=\frac{\mathbb{E}[Y|X=x]}{\mathbb{E}[Y^2|X=x]}=\frac{\mathbb{E}[Y]}{\mathbb{E}[Y^2]}$ is a constant. Then $\phi'_{\!\!\lambda}(x)=\frac{\mathbb{E}[Y]}{\mathbb{E}[Y^2]}>1$ for any $x\in[0,M]$ . This implies that $m=1$ and there is no change point. Applying Theorem 3.1 leads to the following corollary.
Corollary 4.1 If Y is independent of X, then the optimal indemnity function to Problem 1 is given by
for some $d\ge 0$ .
The optimality of stop-loss indemnity function is also verified in the expected utility framework. For example, Cummins and Mahul (Reference Cummins and Mahul2003) show the optimality of a stop-loss function in a situation where the recovery rate can only take 1 or 0 and both the DM and insurer have the same belief about the default probability. Bernard and Ludkovski (Reference Bernard and Ludkovski2012) extend the result of Cummins and Mahul (Reference Cummins and Mahul2003) by considering a budget constraint, where the stop-loss function is proved to be optimal again. In the above-mentioned works, the recovery rate does not need to be independent of the insurable loss.
4.2. Y is a decreasing function of X
In the literature, the recovery rate Y is generally assumed to be negatively correlated with X (see Bernard and Ludkovski Reference Bernard and Ludkovski2012; Li and Li Reference Li and Li2018). This is intuitive, as a larger loss would make the insurer more likely to default, which results in a smaller recovery rate. In this section, we analyze a special case where Y is a decreasing function X, that is $Y=g(X)$ where g is a decreasing function.Footnote 2
To simplify our discussion, we focus on the following situation.
Assumption 3
-
(i) there exists an $x_1\in[0,M]$ such that $g(x)=1$ for all $x\in(0,x_1]$ , and $g(x)<1$ for all $x\in(x_1,M]$ ;
-
(ii) $g(M)>0$ .
Assumption 3(i) states that the insurer will not default when the DM’s loss is less than some threshold. In the rare case where the realized loss is the largest, the insurer may default, but is able to sell its remaining assets and recover part of the indemnity to the DM. This leads Assumption 3(ii) to hold. Assumption 3 is for instance related to the model of Cai et al. (Reference Cai, Lemieux and Liu2014), when the insurer sells an insurance contract to only one DM. In such case, default happens if and only if the insurable loss exceeds a certain threshold.
Note that $\frac{\psi_1(x)}{\psi_2(x)}=\frac{\mathbb{E}[g(X)|X=x]}{\mathbb{E}[g(X)^2|X=x]}=\frac{g(x)}{g(x)^2}=\frac{1}{g(x)}$ . Thus, under Assumption 3
For $x>\max\left\{x_1,\frac{\lambda}{2}\right\}$ , we have
Depending on the value of $\lambda$ , we have the following two sub-cases.
-
Case 1: $\lambda\le 2x_1$ In this case, we have $\phi'_{\!\!\lambda}(x)=1$ for $x\in[0,x_1]$ and $\phi'_{\!\!\lambda}(x)>1$ for $x\in(x_1,M]$ . Therefore, $x_1$ is the only change point. Applying Theorem 3.1 and Proposition 3.1 gives
\begin{equation*} I^{\ast}(x)=(\min\left\{x,d_2\right\}-d_1)_++(x-d_3)_+, \end{equation*}where $0\le d_1\le\frac{\lambda}{2}\le d_2\le x_1\le d_2\le M$ . -
Case 2: $\lambda>2x_1$ In this case, as per Proposition 3.1, for any $x\in[0,\frac{\lambda}{2}]$ we have $I^{\ast}(x)=(x-d_1)_+$ for some $d_1\in[0,\frac{\lambda}{2}]$ . For $x\in(\frac{\lambda}{2},M]$ , applying Theorem 3.1 leads to $I^{\ast}(x)=I^{\ast}(\frac{\lambda}{2})+(x-d_2)_+$ for some $d_2\in[\frac{\lambda}{2},M]$ . In conclusion, the optimal indemnity function for this case is given by
\begin{equation*} I^{\ast}(x)=(\min\left\{x,\frac{\lambda}{2}\right\}-d_1)_++(x-d_2)_+, \end{equation*}where $0\le d_1\le\frac{\lambda}{2}\le d_2$ .
Note that the solutions for Case 1 and 2 are of similar formats. The above discussions are summarized in the next corollary.
5. Numerical illustrations
In this section, we present a numerical example illustrating the main result of this paper. For the ease of discussion, we assume that the DM uses the traditional mean–variance criterion: $\mathcal{M}(z_1,z_2)=z_1+\frac{\mathcal{B}}{2}z_2$ . The premium is determined by the expectation premium principle, that is $\pi(I)=(1+\theta)\mathbb{E}[I(X)Y]$ for some $\theta>0$ . Moreover, we assume the following structure of $Y|X=x$ to depict the dependence between the recovery rate and loss:
where x is the realization of X and U(0,1) denotes the uniform distribution over [0,1). In this case, the random variable $Y|Y<1$ is independent of X. Such structure is inspired by Bernard and Ludkovski (Reference Bernard and Ludkovski2012), who study the case where Y takes one out of two values: $Y\in\left\{y_0,1\right\}$ for some $y_0\in[0,1]$ . As Y is usually negatively correlated with X, $p(\!\cdot\!)$ is usually a decreasing function. Then, the larger the loss is, the smaller is the probability of full recovery. Moreover, for any increasing function g, it holds that
which is decreasing with respect to x. This implies that Y is stochastically decreasing with respect to X. For simplicity, we assume that p(x) is subject to exponential decay, that is $p(x)=e^{-ax}$ for some $a>0$ .
Under these assumptions, we can easily get
Furthermore,
Therefore $\frac{\psi_1(x)}{\psi_2(x)}\ge\frac{\psi_1(0)}{\psi_2(0)}=1$ .
Based on the value of $\lambda$ , we have the following cases.
-
If $\lambda>0$ , applying Proposition 3.1 leads to
\begin{equation*} I^{\ast}(x)=(x-d_1)_+ \end{equation*}on $[0,\frac{\lambda}{2}]$ for some $d_1\in[0,\frac{\lambda}{2}]$ . When $x>\frac{\lambda}{2}$ ,\begin{equation*} \phi'_{\!\!\lambda}(x)=\left(\frac{\psi_1(x)}{\psi_2(x)}\right)\!{}^{^{'}}\left(x-\frac{\lambda}{2}\right)+\frac{\psi_1(x)}{\psi_2(x)}>1. \end{equation*}Applying Theorem 3.1 leads to\begin{equation*} I^{\ast}(x)=I^{\ast}\left(\frac{\lambda}{2}\right)+(x-d_2)_+ \end{equation*}for some $d_2\ge\frac{\lambda}{2}$ . -
If $\lambda\le 0$ , then since $\phi'_{\!\!\lambda}(x)>1$ on [0,M], applying Theorem 3.1 leads to
\begin{equation*} I^{\ast}(x)=(x-d_3)_+ \end{equation*}over the whole domain, where $d_3\in[0,M]$ .
Summarizing the above discussions leads to the general solution
where $0\le a_1\le a_2\le a_3\le M$ . Note that if $a_1=a_2$ or $a_2=a_3$ , this solution reduces to a stop-loss function. An illustrative $I^{\ast}$ is given by Figure 2.
Under the traditional mean-variance criterion, our goal is to minimize
With the optimal indemnity function $I_{a_{1\kern-1pt},a_2,a_3}$ , the above problem reduces to
To optimize $a_{1\kern-1pt},a_2$ and $a_3$ , we use the “fmincon” or “patternsearch” function in MATLAB. Under different loss distributions and volatility aversion parameters, the optimal $a_{1\kern-1pt},a_2$ and $a_3$ are presented in Tables 1 and 2.
Interestingly, in most of the cases in Tables 1 and 2, we get either $a_1=a_2$ or $a_2=a_3$ . This implies that the optimal indemnity function for the considered model settings are of the stop-loss type. For the sensitivity of the retention point with respect to model parameters (e.g., the mean loss, $\mathcal{B}$ , $\theta$ and a), we have the following observations.
-
The retention point increases with respect to the mean loss. Usually a larger mean loss results in a larger premium. The DM may choose to increase the retention point to keep the premium affordable.
-
A larger volatility aversion level leads to a smaller retention point. This is as expected since a smaller retention point implies that the DM cedes more risk to the insurer so that its retained risk is subject to less uncertainty.
-
A larger safety loading parameter leads to a larger retention point. Relatively speaking, a larger safety loading parameter corresponds to a smaller volatility aversion level by looking at problem (5.2). Then as per the second bullet point, a smaller volatility aversion level leads to a larger retention point. From the perspective of premium, larger safety loading increases the premium. So the DM increases the retention point to maintain the affordable premium.
-
A larger a leads to a smaller retention point. It is straightforward that a larger a leads to a smaller expected recovery rate, then the DM tends to cede out more risk to reduce the uncertainty of its retained risk.
6. Conclusions and future research
In this article, we re-visit the optimal insurance problem from an insured’s perspective under counterparty risk but within a mean–variance framework. Compared with the existing literature on this topic, the incentive compatibility is imposed in this article to alleviate the possible ex post moral hazard issue. We assume that the insured is informative of the risk. As such the premium is calculated based on the coverage instead of the promised indemnity. Under incentive compatibility, the problem could be re-formulated as the one in terms of the marginal indemnity function. By applying the calculus of variations, the optimal marginal indemnity function, or the optimal indemnity function, could be characterized in an implicit manner. It is shown that the optimal indemnity function depends on both the first and second-order conditional expectations of the random recovery rate. To make the implicit characterization applicable in practice, we deeply analyze the format of representation and unveil the explicit structure of optimal indemnity function implied by it. Two special cases are studied in detail: the case that the recovery rate is independent of the underlying loss and the case that the recovery rate is a decreasing function of the insurable loss. For both cases, we derive the optimal indemnity functions explicitly.
However, a major drawback of our main result is that the computation cost may be very high if the number of change points is large. This situation may happen when the dependence structure between the counterparty risk Y and the insurable risk X becomes rather complex. Future research is needed to further reduce the complexity of the optimization.
Our study could be extended by replacing the mean–variance criterion with other preference measures, such as expected utility or distortion risk measures. However, switching to the expected utility framework would make the implicit characterization of the optimal indemnity function more complicated, which brings non-trivial technical difficulties to the derivation of the explicit optimal indemnity function. In view of this, we decide to leave such extension for future research.
Besides the specific cases studied in this article, another realistic case is that the default of insurer is driven by a portfolio of policies, and the decision maker in this paper is one of the claimants. In such a case, the recovery rate Y can be modelled via an ex post proportional bankruptcy rule (Ibragimov et al. Reference Ibragimov, Jaffee and Walden2010; Boonen, Reference Boonen2019). The recovery rate Y then depends on the indemnity functions of all policyholders. In this article, we assume that Y is independent of the indemnity function, and so we do not study this case. We leave this problem as a suggestion for future research.
Acknowledgment
We are indebted to two anonymous reviewers for their comments which help improving the paper substantially. W.J. Jiang acknowledges the financial support received from the Natural Sciences and Engineering Research Council (RGPIN-2020-04204, DGECR-2020-00332) of Canada. W.J. Jiang is also grateful to the start-up grant received from the University of Calgary.
A. Proofs of the main results
A.1 Proof of Lemma 3.1
The function $I\in \mathcal{I}$ is 1-Lipschitz continuous, and thus admits the following integral representation
where $\eta$ is called the marginal indemnity function (MIF) as per, for example, Assa (Reference Assa2015) and Zhuang et al. (Reference Zhuang, Weng, Tan and Assa2016). It is easily seen that seeking an optimal I within $\mathcal{I}$ is equivalent to seeking an optimal $\eta$ within the class
Denote by J(I) the objective function of problem (3.3). If $I^{\ast}$ is an optimal indemnity function, then given any $I\in\mathcal{I}$ , we have $\epsilon I^{\ast}+(1-\epsilon)I\in\mathcal{I}$ for any $\epsilon\in[0,1]$ . The first and second order derivatives of $J(\epsilon I^{\ast}+(1-\epsilon)I)$ with respect to $\epsilon$ are
Therefore, $J(\epsilon I^{\ast}+(1-\epsilon)I)$ is convex with respect to $\epsilon$ . It reaches its minimum at $\epsilon=1$ , and thus
This implies
Note that
where the third equation holds due to the Fubini’s theorem. Now let
it is straightforward that (A.1) gets minimized if its integrand function $2L(t\;;\;I^{\ast},\lambda)\eta(t)$ gets minimized for each $t\in[0,M]$ . Since $I\in\mathcal{I}\Longleftrightarrow \eta\in\tilde{\mathcal{I}}$ , we have
where $\xi(t)\in[0,1]$ such that $\eta^{\ast}\in\tilde{\mathcal{I}}$ . This ends the proof.
A.2 Proof of Theorem 3.1
To prove (1), we first show by contradiction that there does not exist a point $t^{\ast}\in S_{m,j_m}$ such that $L(t^{\ast};I^{\ast},\lambda)>0$ and $L'(t^{\ast};I^{\ast},\lambda)\ge 0$ . Note that
and thus $L'(t^{\ast};I^{\ast},\lambda)\ge 0$ is equivalent to $I^{\ast}(t^{\ast})-\phi_{\lambda}(t^{\ast})\le 0$ . Since $\phi'_{\!\!\lambda}(x)>1$ over $S_{m,j_m}$ if $j_m=1$ and ${I^{\ast}}'(x)$ is always bounded by 0 and 1, ${I^{\ast}}'(t)-\phi'_{\!\!\lambda}(t)<0$ . Therefore $I^{\ast}(t)-\phi_{\lambda}(t)\le 0$ for all $t\in[t^{\ast},M]$ , which implies that $L'(t;I^{\ast},\lambda)\ge 0$ over $[t^{\ast},M]$ . However, this also implies that
which contradicts with $L(t^{\ast};I^{\ast},\lambda)>0$ . Therefore, such $t^{\ast}$ does not exist. This implies that $L(t;I^{\ast},\lambda)$ cannot up-cross the t-axis on $S_{m,j_m}$ . Furthermore, $L(t;I^{\ast},\lambda)=0$ cannot hold over any subintervals of $S_{m,j_m}$ as otherwise
over these subintervals. However, this contradicts with ${I^{\ast}}'(x)\in[0,1]$ since $\phi'_{\!\!\lambda}(x)>1$ . Now define $t_0=\inf\left\{t\in S_{m,j_m}:L(t;I^{\ast},\lambda)\le 0\right\}$ , we have $L(t;I^{\ast},\lambda)>0$ for $t\in[x_{m-1},t_0)$ and $L(t;I^{\ast},\lambda)\le 0$ for $t\in[t_0,M]$ . As per Lemma 3.1, we get $\eta^{\ast}(x)=\unicode{x1d7d9}_{[t_0,M]}(x)$ . The optimal indemnity function over $S_{m,j_m}$ in this case is given by
To prove (2), we show by contradiction that there does not exist points $t^{\ast}, t^{**}\in S_{m,j_m}$ such that
If such $t^{\ast}$ exists, then from $L'(t^{\ast};I^{\ast}, \lambda)\ge 0$ we get $I^{\ast}(t^{\ast})- \phi_{\lambda}(t^{\ast})\le 0$ . since $L(t^{\ast};I^{\ast},\lambda)>0$ , ${I^{\ast}}'(t^{\ast})=0$ as per Lemma 3.1. As such, ${I^{\ast}}'(t^{\ast})-\phi'_{\!\!\lambda}(t^{\ast})\le 0$ . This implies that $I^{\ast}(t)-\phi_{\lambda}(t)\le 0$ for any $t\in[t^{\ast},M]$ . Therefore $L'(t;I^{\ast},\lambda)\ge 0$ for any $t\in[t^{\ast},M]$ . This leads to
which contradicts with $L(t^{\ast};I^{\ast},\lambda)>0$ .
Similarly, if such $t^{**}$ exists, then from $L(t^{**};I^{\ast},\lambda)\le 0$ we get $I^{\ast}(t^{**})-\phi_{\lambda}(t^{**})\ge 0$ . Since $L(t^{**};I^{\ast},\lambda)<0$ , ${I^{\ast}}'(t^{**})=1$ as per Lemma 3.1. As such, ${I^{\ast}}'(t^{**})-\phi'_{\!\!\lambda}(t^{**})\ge 0$ . This implies that $I^{\ast}(t)-\phi_{\lambda}(t)\ge 0$ for any $t\in[t^{**},M]$ . Therefore $L'(t;I^{\ast},\lambda)\le 0$ for any $t\in[t^{**},M]$ . This leads to
which contradicts with $L(t^{**};I^{\ast},\lambda)<0$ .
Based on the above findings, $L(t;I^{\ast},\lambda)$ cannot cross the t-axis on $S_{m,j_m}$ . Note that when $L(t;I^{\ast},\lambda)=0$ over any sub-intervals of $S_{m,j_m}$ , we have $I^{\ast}(x)=\phi_{\lambda}(x)$ over those intervals. Now define $t_1=\inf\left\{t\in S_{m,j_m}\;:\; L(t;I^{\ast},\lambda)=0\right\}$ , then we have the following situations:
-
(i) $L(t;I^{\ast},\lambda)>0$ over $[x_{m-1},t_1)$ and $L(t;I^{\ast},\lambda)=0$ over $[t_1,M]$ . This leads to $\eta^{\ast}(x)=\phi'_{\!\!\lambda}(x)\unicode{x1d7d9}_{[t_1,M]}(x)$ .
-
(ii) $L(t;I^{\ast},\lambda)<0$ over $[x_{m-1},t_1)$ and $L(t;I^{\ast},\lambda)=0$ over $[t_1,M]$ . This leads to $\eta^{\ast}(x)=\unicode{x1d7d9}_{[x_{m-1},t_1)}(x)+\phi'_{\!\!\lambda}(x)\unicode{x1d7d9}_{[t_1,M]}(x)$ .
Applying the basic formula $I^{\ast}(x)=I^{\ast}(x_{m-1})+\int_{x_{m-1}}^{x}\eta^{\ast}(t)dt$ leads to the result in (2).
To prove (3), we can show similarly that does not exist a point $t^{**}\in S_{m,j_m}$ such that $L(t^{**};I^{\ast},\lambda)<0$ and $L'(t^{**};I^{\ast},\lambda)\le 0$ . As such, $L(t;I^{\ast},\lambda)$ cannot down-cross the t-axis on $S_{m,j_m}$ . Furthermore, $L(t;I^{\ast},\lambda)=0$ cannot hold on any subintervals of $S_{m,j_m}$ as otherwise $I^{\ast}(x)=\phi_{\lambda}(x)$ on those intervals, which is a contradiction since ${I^{\ast}}'(x)\in[0,1]$ but $\phi'_{\!\!\lambda}(x)<0$ in this case. Now define $t_2=\inf\left\{t\in S_{m,j_m}\;:\; L(t;I^{\ast},\lambda)\ge 0\right\}$ , we have $L(t;I^{\ast},\lambda)<0$ for $t\in[x_{m-1},t_2)$ and $L(t;I^{\ast},\lambda)\ge0$ for $t\in[t_2,M]$ . According to Lemma 3.1, we have $\eta^{\ast}(x)=\unicode{x1d7d9}_{[x_{m-1},t_2)}(x)$ , for which the corresponding $I^{\ast}$ is given by (3).
To prove (4), we note that for any $t\in[x_{i-1},x_i]$
where $L'(t;I^{\ast},\lambda)=-\psi_2(t)(I^{\ast}(t)-\phi_{\lambda}(t))f(t)$ . If $j_i=1$ , then for any $t\in[x_{i-1},x_i]$ , ${I^{\ast}}'(t)-\phi'_{\!\!\lambda}(t)<0$ . We next focus on the case where the root of $I^{\ast}(t)=\phi_{\lambda}(t)$ exists on $(x_{i-1},x_i)$ . Other cases could be studied in a similar way and are thus omitted.
Denote by $t_{r_1}$ the root of $I^{\ast}(t)=\phi_{\lambda}(t)$ on $(x_{i-1},x_i)$ , then $L'(t;I^{\ast},\lambda)<0$ for $t\in[x_{i-1},t_{r_1})$ and $L'(t;I^{\ast},\lambda)>0$ for $t\in(t_{r_1},x_i]$ . That means, $L(t;I^{\ast},\lambda)$ can cross the t-axis at most twice and on $(x_{i-1},t_{r_1})$ and $(t_{r_1},x_i)$ respectively. Let
then $L(t;I^{\ast},\lambda)>0$ over $[x_{i-1},t_2)$ , $L(t;I^{\ast},\lambda)<0$ over $(t_2,t_3)$ and $L(t;I^{\ast},\lambda)>0$ over $(t_3,x_i]$ . According to Lemma 3.1, we have $\eta^{\ast}(x)=\unicode{x1d7d9}_{(t_2,t_3)}(x)$ . Applying the equation $I^{\ast}(x)=I^{\ast}(x_{i-1})+\int_{x_{i-1}}^{x}\eta^{\ast}(t)dt$ leads to the result in (4).
To prove (5), we show by contradiction that there cannot exist two or more than two sub-intervals of $[x_{i-1},x_i]$ such that $L(t;I^{\ast},\lambda)=0$ . If there exist two sub-intervals, for example [a,b] and [c,d] where $x_{i-1}\le a<b<c\le x_i$ , such that $L(t;I^{\ast},\lambda)=0$ for $t\in[a,b]\cup[c,d]$ , then there must exist a point $t^{\ast}$ or $t^{**}$ in (b,c) as described in the proof of (2). However, if $t^{\ast}$ exists, then similar to the proof of (2) we get $L'(t;I^{\ast},\lambda)\ge 0$ over $[t^{\ast},x_i]$ . As such $L(t;I^{\ast},\lambda)>0$ for $t\in[t^{\ast},x_i]$ , which contradicts with $L(t;I^{\ast},\lambda)=0$ over $[c,d]\subseteq[t^{\ast},x_i]$ . If $t^{**}$ exists, then similar to the proof of (2) we get $L'(t;I^{\ast},\lambda)\le0$ over $[t^{**},x_i]$ . As such $L(t;I^{\ast},\lambda)<0$ for $t\in[t^{**},x_i]$ , which also contradicts with $L(t;I^{\ast},\lambda)=0$ over $[c,d]\subseteq[t^{**},x_i]$ . Therefore, there exists at most one subinterval, for example $[t_4,t_5]\subseteq [x_{i-1},x_i]$ , on which $L(t;I^{\ast},\lambda)=0$ . We have four situations based on the sign of $L(t;I^{\ast},\lambda)$ on $[x_{i-1},t_4)$ and $(t_5,x_i]$ :
-
(i) $L(t;I^{\ast},\lambda)>0$ over $[x_{i-1},t_4)$ and $L(t;I^{\ast},\lambda)>0$ over $(t_5,x_i]$ . This leads to $\eta^{\ast}(x)=\phi'_{\!\!\lambda}(x)\unicode{x1d7d9}_{[t_4,t_5]}(x)$ .
-
(ii) $L(t;I^{\ast},\lambda)<0$ over $[x_{i-1},t_4)$ and $L(t;I^{\ast},\lambda)>0$ over $(t_5,x_i]$ . This leads to $\eta^{\ast}(x)=\unicode{x1d7d9}_{[x_{i-1},t_4)}(x)+\phi'_{\!\!\lambda}(x)\unicode{x1d7d9}_{[t_4,t_5]}(x)$ .
-
(iii) $L(t;I^{\ast},\lambda)>0$ over $[x_{i-1},t_4)$ and $L(t;I^{\ast},\lambda)<0$ over $(t_5,x_i]$ . This leads to $\eta^{\ast}(x)=\phi'_{\!\!\lambda}(x)\unicode{x1d7d9}_{[t_4,t_5]}(x)+\unicode{x1d7d9}_{(t_5,x_i]}(x)$ .
-
(iv) $L(t;I^{\ast},\lambda)<0$ over $[x_{i-1},t_4)$ and $L(t;I^{\ast},\lambda)<0$ over $(t_5,x_i]$ . This leads to $\eta^{\ast}(x)=\unicode{x1d7d9}_{[x_{i-1},t_4)}(x)+\phi'_{\!\!\lambda}(x)\unicode{x1d7d9}_{[t_4,t_5]}(x)+\unicode{x1d7d9}_{(t_5,x_i]}(x)$ .
Applying the formula, $I^{\ast}(x)=I^{\ast}(x_{i-1})+\int_{x_{i-1}}^{x}\eta^{\ast}(t)dt$ leads to the result in (5).
To prove (6), we note that ${I^{\ast}}'(t)-\phi'_{\!\!\lambda}(t)>0$ in this case. We next focus on the case where the root of $I^{\ast}(t)=\phi_{\lambda}(t)$ exists on $(x_{i-1},x_i)$ . Other cases could be studied in a similar way and are thus omitted.
Denote by $t_{r_2}$ the root of $I^{\ast}(t)=\phi_{\lambda}(t)$ on $(x_{i-1},x_i)$ , then $L'(t;I^{\ast},\lambda)>0$ for $t\in[x_{i-1},t_{r_2})$ and $L'(t;I^{\ast},\lambda)<0$ for $t\in(t_{r_2},x_i]$ . That means $L(t;I^{\ast},\lambda)$ can cross the t-axis at most twice and on $(x_{i-1},t_{r_2})$ and $(t_{r_2},x_i)$ respectively. Similar to the proof of (4), let
then $L(t;I^{\ast},\lambda)<0$ over $[x_{i-1},t_6)$ , $L(t;I^{\ast},\lambda)>0$ over $(t_6,t_7)$ and $L(t;I^{\ast},\lambda)<0$ over $(t_7,x_i]$ . According to Lemma 3.1, we have $\eta^{\ast}(x)=\unicode{x1d7d9}_{[x_{i-1},t_6)}(x)+\unicode{x1d7d9}_{(t_7,x_i]}(x)$ . Applying the basic formula $I^{\ast}(x)=I^{\ast}(x_{i-1})+\int_{x_{i-1}}^{x}\eta^{\ast}(t)dt$ leads to the result in (6). This ends the proof.
A.3 Proof of Proposition 3.1
Under the conditions of this proposition, for any $t\in[0,\frac{\lambda}{2})$ ,
As such, for $t\in[0,\frac{\lambda}{2})$ ,
Let $t_1=\inf\left\{t\in[0,\frac{\lambda}{2}]\;:\; L(t;I^{\ast},\lambda)\le0\right\}$ , then as per the monotonicity of $L(t;I^{\ast},\lambda)$ over $[0,\frac{\lambda}{2})$ , we have $L(t;I^{\ast},\lambda)>0$ for $t\in[0,t_1)$ and $L(t;I^{\ast},\lambda)<0$ for $t\in(t_1,\frac{\lambda}{2})$ . Applying Lemma 3.1 gives ${I^{\ast}}'(x)=\unicode{x1d7d9}_{(t_1,\frac{\lambda}{2})}(x)$ . As such, for any $x\in[0,\frac{\lambda}{2}]$
This finishes the proof.