Hostname: page-component-745bb68f8f-s22k5 Total loading time: 0 Render date: 2025-02-07T08:01:56.883Z Has data issue: false hasContentIssue false

Moment bounds of PH distributions with infinite or finite support based on the steepest increase property

Published online by Cambridge University Press:  22 July 2019

Qi-Ming He*
Affiliation:
University of Waterloo
Gábor Horváth*
Affiliation:
Budapest University of Technology and Economics
Illés Horváth*
Affiliation:
MTA-BME Information Systems Research Group
Miklós Telek*
Affiliation:
Budapest University of Technology and Economics
*
*Postal address: Department of Management Sciences, University of Waterloo, Waterloo, Ontario N2L 3G1, Canada. Email address: q7he@uwaterloo.ca
**Postal address: Department of Networked Systems and Services, Budapest University of Technology and Economics, PO Box 91, 1521 Budapest, Hungary.
****Postal address: MTA-BME Information Systems Research Group, PO Box 91, 1521 Budapest, Hungary. Email address: horvath.illes.antal@gmail.com
**Postal address: Department of Networked Systems and Services, Budapest University of Technology and Economics, PO Box 91, 1521 Budapest, Hungary.
Rights & Permissions [Opens in a new window]

Abstract

The steepest increase property of phase-type (PH) distributions was first proposed in O’Cinneide (1999) and proved in O’Cinneide (1999) and Yao (2002), but since then has received little attention in the research community. In this work we demonstrate that the steepest increase property can be applied for proving previously unknown moment bounds of PH distributions with infinite or finite support. Of special interest are moment bounds free of specific PH representations except the size of the representation. For PH distributions with infinite support, it is shown that such a PH distribution is stochastically smaller than or equal to an Erlang distribution of the same size. For PH distributions with finite support, a class of distributions which was introduced and investigated in Ramaswami and Viswanath (2014), it is shown that the squared coefficient of variation of a PH distribution with finite support is greater than or equal to 1/(m(m + 2)), where m is the size of its PH representation.

Type
Original Article
Copyright
© Applied Probability Trust 2019 

1. Introduction

Phase-type (PH) distributions were introduced in 1975 by Neuts [Reference Neuts3] for the study of queueing systems. Since then, PH distributions have been a subject of research, and have found applications in many areas of applied probability. For example, a succession of papers by O’Cinneide ([Reference O’cinneide5]–[Reference O’cinneide8]) revealed some fundamental properties of PH distributions. In [Reference Aldous and Shepp1] it was shown that the squared coefficient of variation (SCV) of a PH distribution with a PH representation of size m is greater than or equal to 1/m. In [Reference Telek, Latouche and G. Taylor11] the minimal SCV of discrete PH distributions was found. In [Reference He, Zhang, Vera and Latouche2] stochastic comparison was utilized in the study of PH distributions, which also led to moment bounds of PH distributions. These results not only deepen our understanding of PH distributions and PH representations, but also facilitate the applications of PH distributions significantly.

The steepest increase property of PH distributions was first proposed in O’Cinneide [Reference O’cinneide8] and proved by O’Cinneide [Reference O’cinneide8] and Yao [Reference Yao12]. We think that it is a very interesting property of PH distributions, which received little attention in the literature. In this paper we apply the property to find moment bounds of PH distributions, which demonstrates the usefulness of the property.

PH distributions with finite support were introduced and investigated in [Reference Ramaswami and Viswanath9]. This generalization extends the applications of PH distributions significantly.

In this paper we find a number of stochastic and moment bounds for PH distributions with finite and infinite support. Many of the moment bounds depend only on the size of PH representations and the eigenvalue with the largest real part of PH generators. A highlight of this paper is that the SCV of a bounded PH distribution with a PH representation of size m is greater than or equal to 1/(m(m + 2)). The moment bounds of PH distributions reveal fundamental properties of such probability distributions, e.g. they indicate which type of general distributions can be closely approximated. The results are useful in, but not limited to,

  • finding moment bounds of performance measures for stochastic models, and

  • selecting PH representations in the parameter estimation of PH distributions.

The rest of the paper is organized as follows. In Section 2 we introduce PH distributions with infinite support together with the steepest increase property and with some of their moment bounds. In Section 3 we introduce PH distributions with finite support and study their moment bounds. Section 4 concludes the paper.

2. PH distributions with infinite support

In this section we first define PH distributions and their corresponding PH representations. Then we review the so-called ‘steepest increase lemma’ (see [Reference Yao12]) on the density function of (ordinary) PH distributions, since this turned out to be useful to derive the moment bounds. We further extend the ‘steepest increase lemma’ and derive new moment bounds for ordinary PH distributions.

A nonnegative random variable 𝒴 has a PH distribution if it is the absorption time in a finite-state, continuous-time Markov chain (see [Reference Neuts4]). We assume that 𝒴 has a PH representation (α, A) of size m, where α is the initial distribution of the underlying continuous-time Markov chain and A contains the transition rates among the transient states of the underlying continuous-time Markov chain (referred to as a PH generator or sub-intensity matrix). That is, α is a nonnegative probability vector and A has nonnegative off-diagonal and negative diagonal elements such that the diagonal element dominates each row, A1 ≤ 0 (elementwise), where 1 is a column vector of 1s with the appropriate size. Let us denote the density function of 𝒴 by f (t) = αeA t( − A1) and the cumulative distribution function (CDF) by F(t) = 1αeAt1 (both for t ≥ 0). To avoid the trivial case 𝒴≡0, we assume that 0 < α1 ≤ 1. We also note that, by [Reference O’CINNEIDE6], f (t) > 0 for all t > 0.

The steepest increase conjecture was first published and partially proven by O’Cinneide [Reference O’cinneide8]. Its complete proof was given by Yao [Reference Yao12]. The steepest increase lemma states that f′(t)/f (t) ≤ (m − 1)/t for t > 0, or, equivalently, f (t)/t m−1 is decreasing in t for t > 0. Next, we present and show a ‘sharp’ form of the steepest increase lemma.

Lemma 1

For a PH distribution with representation (α, A) of size m and with density function f (t), we have f (t)

(1)\begin{equation} \frac{f'(t)}{f(t)}\leq \frac{m-1}{t} - \lambda \quad for \quad t > 0, \label{eq:sil1} \end{equation}

where λ is the absolute value of the eigenvalue of A with the largest real part (which is real and negative). Hereλ is also referred to as the dominant eigenvalue of matrix A. In (1) the equality holds when 𝒴 is Erlang(m, λ)-distributed (i.e. 𝒴 is the sum of m independent exponential random variables with the same parameter λ).

Proof. The inequality just before Equation (12) of [Reference O’cinneide8], combined with the proof in [Reference Yao12], leads to ((m − 1 − λ)IA)eA ≥ 0 for PH generator A, where I is the identity matrix. For PH generator A, At is also a PH generator for t > 0. The eigenvalue with the largest real part of At is −λt for t > 0. Setting A =: At in the inequality, we obtain, for t > 0,

\begin{equation} ((m-1-\lambda t){\boldsymbol I}-{{\boldsymbol A}}t)\text {e}^{{\boldsymbol A}t}\geq 0, \end{equation}

which leads to

(2)\begin{equation}\label{ieq:sil} \text {e}^{{{\boldsymbol A}}t}{{\boldsymbol A}}t \leq (m-1-\lambda t)\text {e}^{{{\boldsymbol A}}t}. \end{equation}

Pre-multiplying and post-multiplying both sides of (2) by α and −A1, respectively, we obtain f′(t)t ≤ (m − 1 − λt)f (t), which proves the lemma.

Similar to Lemma 1, −λ stands for the dominant eigenvalue of A in the sequel.

Equation (1) can be written in several alternative forms. For t ≥ 0,

(3)\begin{align} &t f'(t) \leq (m-1) f(t) - \lambda t f(t), \\ &\quad\frac{{\text {d}}}{{\text {d}}t}(t f(t)) \leq (m- \lambda t) f(t), \label{eq:diffsil1} \end{align}

and utilizing the fact that λ > 0, we also have, for t > 0,

(4)\begin{align} \frac{{\text {d}}}{{\text {d}}t}(t f(t)) < m f(t). \label{eq:diffsil} \end{align}

Lemma 2

For 𝒴 with PH generator A of size m and dominant eigenvalue −λ, we have, for 0 < t 1 < t 2,

\begin{equation}\frac{f(t_2)}{f(t_1)} \leq \frac{f_{(m,\lambda)}(t_2)}{f_{(m,\lambda)} (t_1)}, \end{equation}

where f (m, λ)(t) = λmt m−1eλt/(m − 1)! is the density function of the Erlang random variable with parameters (m, λ).

Proof. For t > 0, (1) can be written as,

(5)\begin{equation} \label{eq:ln-ieq} (m-1)(\ln(t))' - (\lambda t)' \geq (\ln(f(t)))'. \end{equation}

Integrating both sides of (5) from t 1 > 0 to t 2 > t 1 yields

(6)\begin{equation} \label{eq:ln-ieq1} \ln\biggl(\frac{(t_2/t_1)^{m-1}f(t_1)}{f(t_2)}\biggr) \geq \lambda(t_2-t_1), \end{equation}

where f (t) > 0 for all t > 0 ensures that the integration can be done properly. Taking the exponent of both sides of (6), we obtain

\begin{equation} f(t_2) \leq \frac{f(t_1)}{t_1^{m-1}}\text {e}^{\lambda t_1}t_2^{m-1}\text {e}^{-\lambda t_2} = f(t_1)\frac{f_{(m,\lambda)}(t_2)}{f_{(m,\lambda)}(t_1)} \quad\mbox{for } t_2>t_1>0,\end{equation}

which leads to the desired result.

Lemma 2 implies that f (t)/f (m,λ)(t) is decreasing in t, which is a generalization of the monotonicity of the function f (t)/t m−1. Furthermore, Lemma 2 leads to the following stochastic comparison result between 𝒴 and the Erlang random variable 𝒳(m,λ) with parameters (m, λ). A random variable Y is stochastically smaller than or equal to a random variable X, denoted as Yd X, if F Y(t) ≥ F X(t) holds for all real t (see [Reference Stoyan10]).

Corollary 2.1

The PH-distributed random variable 𝒴 (of size m and with dominant eigenvalue −λ) is stochastically smaller than or equal to 𝒳(m,λ). Consequently, we have, for n ≥ 1,

\begin{equation} \mathbb{E}[{\mathcal{Y}^n}] \leq \mathbb{E}[{\mathcal{X}_{(m,\lambda)}^n}] = \frac{(m+n-1)!}{(m-1)!\,\lambda^n}. \end{equation}

Proof. Since both f (t) and f (m,λ)(t) are density functions on [0, ∞), there must be at least one intersection in (0, ∞). If t* is an intersection (i.e. f (t*) = f (m,λ)(t*)), by Lemma 2, we must have f (t) ≤ f (m,λ) (t) for t > t* and f (t) ≥ f (m,λ)(t) for t < t*. Thus, there are only three possible cases:

  • f (t) and f (m,λ)(t) are identical;

  • f (t) and f (m,λ)(t) have exactly two intersections, t = 0 and t = t*;

  • f (t) and f (m,λ)(t) have exactly one intersection, t = t*.

Then we must have f (t) ≥ f (m,λ)(t) for 0 < tt* and f (t) ≤ f (m,λ)(t) for tt* (> 0), which leads to F(t) ≥ F 𝒳(m,λ) (t), where F 𝒳(m,λ) (t) is the CDF of 𝒳(m,λ) for 0 < tt*, and 1 − F(t) ≤ 1 − F 𝒳(m,λ) (t) for t > t*. Consequently, we obtain F(t) ≥ F 𝒳(m,λ)(t) for t > 0, which leads to the first result. All the moment bounds can be obtained from 𝒴≤d 𝒳(m,λ) directly.

A random variable X is smaller in the mean residual life than a random variable Y, denoted as Xc Y, if 𝔼 [max{0, Xt}] ≤ 𝔼 [max{0, Yt}] holds for all real t (see [Reference Stoyan10]). By [Reference O’cinneide7], it is known that the Erlang distribution with parameters (m, m/𝔼[𝒴]) is smaller in the mean residual life than 𝒴, which has the same mean. Then the moments of 𝒴 are bounded from above and below as follows:

\begin{equation} \frac{(m+n-1)!}{(m-1)!} \left(\frac{\mathbb{E}\left[{\mathcal{Y}}\right]}{m}\right)^n \leq \mathbb{E}[{\mathcal{Y}^n}] \leq \frac{(m+n-1)!}{(m-1)!}\frac{1}{\lambda^n}. \end{equation}

A by-product of the above inequalities is that 𝔼[𝒴] = m/λ if and only if 𝒴 has an Erlang distribution, which can also be obtained from inequality (3) directly.

Based on Lemma 1, the next lemma refines the upper bound for the (n + 1)th moment based on the nth moment, which gives Corollary 2.1 an alternative proof.

Lemma 3

For n = 0, 1, …, the (n + 1)th moment of 𝒴 (of size m and with dominant eigenvalue −λ) is bounded by

(7)\begin{equation} \mathbb{E}[{\mathcal{Y}^{n+1}}] \leq \frac{m+n}{\lambda} \mathbb{E}[{\mathcal{Y}^n}], \label{eq:silXn} \end{equation}

and the equality holds when 𝒴 = 𝒳(m,λ).

FIGURE 1: Bounds of the SCV for ordinary PH distributions. The lower bound is the well known 1/m bound provided in [Reference Aldous and Shepp1] and [Reference O’cinneide7], while the upper bound is provided based on the steepest increase property.

Proof. Multiplying both sides of (3) by t n and integrating from 0 to ∞ gives the following identities for the left-hand side (LHS) and the right-hand side (RHS):

\begin{align} {\rm LHS} =\int_{t=0}^{\infty} t^n {\text {d}}(t f(t)) &\, = [t^{n+1} f(t)]^{\infty}_0 - \int_{t=0}^{\infty} t f(t) {\text {d}}t^n = - n \int_{t=0}^{\infty} t f(t) t^{n-1} {\text {d}} t = - n \mathbb{E}[{\mathcal{Y}^n}]; \\ &\ \,{\rm RHS} =\int_{t=0}^{\infty} t^n (m-\lambda t) f(t) {\text {d}}t \\ &\ \kern1.5pt\quad\quad= m \int_{t=0}^{\infty} t^n f(t) {\text {d}}t - \lambda \int_{t=0}^{\infty} t^{n+1} f(t) {\text {d}}t \\ &\ \kern1.5pt\quad\quad= m \mathbb{E}[{\mathcal{Y}^n}] - \lambda \mathbb{E}[{\mathcal{Y}^{n+1}}]. \end{align}

Thus

$$- n \mathbb{E}[{\mathcal{Y}^n}] \leq m \mathbb{E}[{\mathcal{Y}^n}] - \lambda \mathbb{E}[{\mathcal{Y}^{n+1}}].$$

When 𝒴 is Erlang(λ, m) distributed, the equality in (7) comes from the fact that Lemma 1 gives equality for an Erlang distribution for all t > 0.

Applying Lemma 3 for n = 0 and n = 1 enables us to derive the following upper bounds on the mean 𝔼[𝒴] and the squared coefficient of variation SCV𝒴 = 𝔼[𝒴2]/𝔼[𝒴]2 – 1:

(8)\begin{gather} \mathbb{E}\left[{\mathcal{Y}}\right] \leq \frac{m}{\lambda}, \end{gather}
(9)$${\rm SCV}_\mathcal{Y} \leq \frac{m+1-\lambda \mathbb{E}\left[{\mathcal{Y}}\right]}{\lambda \mathbb{E}\left[{\mathcal{Y}}\right]}. $$

Interestingly, (9) gives an upper bound for SCV𝒴, while the lower bound for SCV 𝒴 is much more widely known as SCV𝒴 ≥ 1/m. Hence, we have (see Figure 1)

\begin{equation} \frac{1}{m} \leq {\rm SCV}_\mathcal{Y} \leq \frac{m+1}{\lambda \mathbb{E}\left[{\mathcal{Y}}\right]} - 1. \end{equation}

3. PH distributions with finite support

PH distributions with finite support were introduced in [Reference Ramaswami and Viswanath9], where three classes of finite support distributions were considered (matrix exponential densities from the lower bound to the upper bound, and from the upper bound to the lower bound, and a convex combination of the two). Instead, we define $\mathcal{Z}$ to have distribution b + (𝒴 | 𝒴 < T). The support of $\mathcal{Z}$ is thus [b, B) with b < B and B = b + T. Recall that 𝒴 is the ordinary (or infinite support) PH distribution with density function f (t) = αeAt(− A)1. Then the density function of $\mathcal{Z}$ is given by

\begin{align} f_{\mathcal{Z}}(t) &= \frac{{\boldsymbol \alpha} \rm {e}^{{\boldsymbol A}(t-b)}(-{{\boldsymbol A}){\bf 1} }} { 1-{\boldsymbol \alpha} \text {e}^{{\boldsymbol A} T}{\bf 1}} \end{align}

for t ∈ [b, B), and $f_{\mathcal{Z}}(t)=0$ for t ∉ [b, B).

We denote this class of finite PH distributions by FTPH1 (with some further similar classes FTPH2 and FTPH3 in mind: FTPH2 is defined as $\mathcal{Z}_2=B-(\mathcal{Y}\mid \mathcal{Y}<T)$ and FTPH3 as the convex combination of $\mathcal{Z}$ and $\mathcal{Z}_2$; these are subject to future investigations). In this paper we primarily focus on the moments of FTPH1 distributions.

Although FTPH1 is obtained by just a simple truncation of an ordinary PH distribution, it has some very interesting members. A truncated exponential distribution with a very small intensity parameter leads to a uniform distribution (Figure Figure 2(a)), since

\begin{align} \lim_{\lambda\rightarrow 0} f_{\mathcal{Z}}^{\rm Exp}(t) = \lim_{\lambda\rightarrow 0} \frac{\lambda \text {e}^{\lambda t}}{1-\text {e}^{-\lambda}} = 1. \end{align}

FIGURE 2: Some interesting members of the FTPH1 class.

Similarly, a truncated Erlang-N distribution with very small intensity gives

\begin{align} \lim_{\lambda\rightarrow 0} f_{\mathcal{Z}}^{{\rm Erl}-N}(t) = N\,t^{N-1}, \end{align}

yielding linear, quadratic, and cubic distributions (see Figure 2(b) and 2(c)).

The importance of these special FTPH1 members is that density functions of such shapes are notoriously difficult to capture by ordinary PH distributions. Ordinary PH distributions with a limited number of phases purely approximate them. In practical computations the λ → 0 limit also causes difficulties, but easily computable small positive λ values give reasonably good approximations. These approximation issues and parameter estimations of PH distributions with finite support will be addressed in a separate paper.

A general formula for computing the moments for PH distributions with finite support can be obtained as follows.

Theorem 1

The nth moment of an FTPH1 random variable, $\mathcal{Z}$ with parameters (α, A) over interval [b, B), is expressed by

\begin{equation} \mathbb{E}[{\mathcal{Z}^n}] = \frac{ b^n(1-{\boldsymbol \alpha}{\bf 1})+\sum_{d=0}^n \binom{n}{d}d!\,\boldsymbol{\alpha}(-{\boldsymbol A})^{-d} (b^{n-d}{\boldsymbol I}-T^{n-d}\text {e}^{{\boldsymbol A}T} ){\bf 1}}{1-\boldsymbol{\alpha} \text {e}^{{\boldsymbol A}T}{\bf 1}}, \end{equation}

where T = Bb.

Proof. The theorem is proved by routine calculations.

3.1. Moment bounds for the case with b = 0

In this subsection we assume that b = 0, which is extended to b > 0 in the next subsection. A random variable in FTPH1 is denoted as 𝒲 = 𝒴 | 𝒴 < T. We derive and prove the lower and upper bounds for the moments of 𝒲. For i = 0, 1, … , we introduce the notation $E_i(T)=\mathbb{E}[{\mathbb{I}_{\{\mathcal{Y}< T\}}}{\mathcal{Y}^i}]=\int_{t=0}^{T} t^i f(t) {\text {d}}t$, where 𝕀{a} denotes the indicator of a. Then 𝔼[𝒲i] can be written as, for i = 0, 1, …,

\begin{equation} \mathbb{E}[{\mathcal{W}^i}] = \frac{E_i(T)}{E_0(T)}. \end{equation}

The next lemma is similar to Lemma 3 (for ordinary PH distributions), but it does not depend on the dominant eigenvalue −λ.

Lemma 4

The moments of an FTPH1 random variable, 𝒲, with support on (0, T), satisfies

(10)\begin{equation} \mathbb{E}[{\mathcal{W}^n}] \leq \frac{(m+n-1) T }{m+n} \mathbb{E}[{\mathcal{W}^{n-1}}]. \label{eq:silZn0} \end{equation}

Proof. Multiplying both sides of (4) by t n−1(Tt) and integrating from 0 to T gives the following identities for the left-hand side:

\begin{align} {\rm LHS} &= \int_{t=0}^{T} t^{n-1}(T-t) \frac{{\text {d}}}{{\text {d}}t}(t f(t)) {\text {d}}t \\ &= - \int_{t=0}^{T} t f(t) {\text {d}}(t^{n-1}(T-t)) \\ & =\int_{t=0}^{T} (n t^n -(n-1)T t^{n-1})\kern1pt f(t) {\text {d}}t \\ &=n E_n(T)- (n-1) T E_{n-1}(T). \end{align}

We use Stieltjes integration by parts in the first step,

$\int_{a}^{b} f(x) {\text {d}}g(x)=f(b)g(b)-f(a)g(a)-\int_{a}^{b} g(x) {\text {d}}f(x),$

and

$\int_{a}^{b} g(x) {\text {d}}f(x) = \int_{a}^{b} g(x) f'(x) {\text {d}}x$

in the second step. For the right-hand side, we have

\begin{align} {\rm RHS}&=\int_{t=0}^{T} t^{n-1} (T-t) m f(t) {\text {d}}t \\ &= \int_{t=0}^{T} m T t^{n-1} f(t) {\text {d}}t - m \int_{t=0}^{T} t^{n} f(t) {\text {d}}t \\ & = m T E_{n-1}(T) - m E_n(T), \end{align}

from which

(11)\begin{align} n E_n(T)&- (n-1) T E_{n-1}(T)\leq m T E_{n-1}(T) - m E_n(T), \\ &\!(m+n) E_n(T) \leq (m+n-1) T E_{n-1}(T), \\ &\quad\frac{E_n(T)}{E_0(T)} \leq \frac{(m+n-1) T}{m+n}\frac{E_{n-1}(T)}{E_0(T)}, \label{eq:eqnn} \end{align}

which leads to the desired result.

In the following corollary the upper bound for the nth moment is provided, independent from the lower-order moments.

Corollary 1

The nth moment of an FTPH1 random variable, 𝒲, with support on (0, T), is bounded by

(12)\begin{equation} \mathbb{E}[{\mathcal{W}^n}] \leq \frac{m T^n }{m+n}, \label{eq:silZn0s} \end{equation}

and the upper bound is strict except when 𝒴 is Erlang(λ, m) distributed and λ tends to 0. In particular, we have 𝔼 [𝒲] ≤ mT/(m + 1), which indicates that no PH distribution with finite support can have a mean close to the upper bound T = B.

Proof. Recursively applying (10) for moments 1, …, n gives the upper bound. The statement on the equality comes from the fact that the right-hand side of (12) gives equality only when (11) gives equality for 1, …, n, and it occurs only when 𝒴 is Erlang(λ, m) distributed and λ tends to 0, because (4) gives equality only for an Erlang(λ, m) distribution when λ tends to 0.

Having derived these moment bounds, we now consider how to reach the extreme values, and the FTPH1 structure which realizes the moment bounds. According to the next lemma, Erlang distributions play an important role in this respect.

Lemma 5

The upper bound in Lemma 4 is strict when 𝒲equals 0 with probability p =1 − (m + n − 1)μ n−1/mT n−1 and is truncated Erlang(λ, m) distributed with probability 1 − p such that λ tends to 0, where μ n−1 = 𝔼[𝒲n−1].

Proof. To prove the statement, we show that compared to the extreme distribution of Lemma 5 any valid change in the probability of the mass at 0, p, decreases the nth moment. First we note that increasing p is not a valid change, because to maintain 𝔼[𝒲n−1] the (n − 1)th moment of a strictly positive FTPH1 needs to be increased above mT n−1/(m + n − 1) (which is not possible according to Corollary 1).

Let us now try to decrease rather than increase p. Consider a distribution whose mass at 0 has probability = p − Δμ n−1 , where Δ is a small positive number. In this case, the (n − 1)th moment of the strictly positive part, $\mu_{n-1}^+$, is

\begin{align} \mu_{n-1}^+ &= \frac{\mu_{n-1}}{1-\hat{p}} \\ &= \frac{\mu_{n-1}}{1-p+\Delta\mu_{n-1} } \\ &= \frac{\mu_{n-1}}{{(m+n-1)\mu_{n-1}}/{m T^{n-1}}+\Delta\mu_{n-1} } \\ &= \frac{m T^{n-1} }{(m+n-1)+\Delta m T^{n-1}} \\ &<\frac{m T^{n-1} }{m+n-1}. \end{align}

When the (n − 1)th moment of the strictly positive part, 𝒲+, is $\mu_{n-1}^+$, its nth moment is bounded by (10), and using that, we can write

\begin{align} &(1-\hat{p}) \mathbb{E}[{{\mathcal{W}^+}^n}] \\ &\quad\quad < \underbrace{\mu_{n-1} \left(\frac{(m+n-1)+ \Delta m T^{n-1}}{m T^{n-1}}\right)}_{1-\hat{p}}\frac{(m+n-1) T }{m+n} \underbrace{\left(\frac{m T^{n-1} }{(m+n-1)+\Delta m T^{n-1}}\right)}_{\mu_{n-1}^+} \\ &\quad\quad= \frac{(m+n-1) T \mu_{n-1}}{m+n}, \end{align}

where the inequality is strict, because the distribution of 𝒲+ is different from an Erlang(λ, m) → 0, since its (n − 1)th moment is less than mT n−1/(m + n − 1).

While the moment bounds derived in Lemma 4 are independent of the dominant eigenvalue, the following results provide moment bounds as a function of λ.

Lemma 6

For n = 1, 2, …, the (n + 1)th moment of 𝒲 is bounded by

\begin{align}\frac{m+n+ \lambda T}{\lambda} \mathbb{E}[{\mathcal{W}^{n}}] - \frac{(m + n-1) T}{\lambda}\mathbb{E}[{\mathcal{W}^{n-1}}] &\leq \mathbb{E}[{\mathcal{W}^{n+1}}] \\ &\leq \frac{m+n}{\lambda} \mathbb{E}[{\mathcal{W}^{n}}] - \frac{T^{n+1} f(T)}{\lambda E_0(T)}. \end{align}

Proof. Multiplying both sides of (3) by (Tt)t n−1 and integrating from 0 to T gives

\begin{align} {\rm LHS} &=\int_{t=0}^{T} (T-t) t^{n-1} \frac{{\text {d}}}{{\text {d}}t}(t f(t)) {\text {d}}t \\ &= - \int_{t=0}^{T} t f(t) {\text {d}}(T t^{n-1}-t^n) \\ &= - \int_{t=0}^{T} t f(t) (T(n-1)t^{n-2} - n t^{n-1} ) {\text {d}}t \\ &= - T(n-1)E_{n-1}(T)+ n E_{n}(T), \\ {\rm RHS} &=\int_{t=0}^{T} (T t^{n-1}-t^n) (m- \lambda t) f(t) {\text {d}}t \\ &= m T E_{n-1}(T) - m E_n(T) - \lambda T E_{n}(T) + \lambda E_{n+1}(T), \end{align}

from which it follows that

\begin{align} &\,- T(n-1)E_{n-1}(T)+ n E_{n}(T) \leq m T E_{n-1}(T) - (m + \lambda T) E_{n}(T) + \lambda E_{n+1}(T), \\ &\quad\quad\quad\quad(m+n+ \lambda T) E_n(T) - (m + n-1) T E_{n-1}(T) \leq \lambda E_{n+1}(T), \\ &\quad\quad\quad\quad\frac{m+n+ \lambda T}{\lambda} \frac{E_n(T)}{E_0(T)} - \frac{(m + n-1) T}{\lambda}\frac{E_{n-1}(T)}{E_0(T)} \leq \frac{E_{n+1}(T)}{E_0(T)}. \end{align}

On the other hand, multiplying both sides of (3) by t n and integrating from 0 to T gives

\begin{align} {\rm LHS}&=\int_{t=0}^{T} t^n \frac{{\text {d}}}{{\text {d}}t}(t f(t)) {\text {d}}t \\ &= T^{n+1} f(T) - \int_{t=0}^{T} t f(t) {\text {d}}t^n \\ &= T^{n+1} f(T) - \int_{t=0}^{T} t f(t) n t^{n-1} {\text {d}}t \\ &= T^{n+1} f(T) - n E_n(T), \\ {\rm RHS}&=\int_{t=0}^{T} t^n (m- \lambda t) f(t) {\text {d}}t \\ &= m \int_{t=0}^{T} t^n f(t) {\text {d}}t - \lambda \int_{t=0}^{T} t^{n+1} f(t) {\text {d}}t \\ &= m E_n(T) - \lambda E_{n+1}(T), \end{align}

from which we obtain

\begin{align} T^{n+1} f(T) \,{-}\, &n E_n(T) \leq m E_n(T) - \lambda E_{n+1}(T), \\ \lambda E_{n+1}(T) \,&{\leq}\, (m+n) E_n(T) - T^{n+1} f(T) , \\ \frac{E_{n+1}(T)}{E_0(T)} \,&{\leq}\, \frac{m+n}{\lambda} \frac{E_n(T)}{E_0(T)} - \frac{T^{n+1} f(T)}{\lambda E_0(T)}, \end{align}

which leads to the desired results.

Lemma 6 gives a tight moment bound, for which the upper and the lower limits are identical if 𝒴 is Erlang distributed. To get rid of the density function in the boundary, a loose version of Lemma 6 is

\begin{align} \frac{m+n+ \lambda T}{\lambda} \mathbb{E}[{\mathcal{W}^{n}}] - \frac{(m + n-1) T}{\lambda}\mathbb{E}[{\mathcal{W}^{n-1}}] & \leq \mathbb{E}[{\mathcal{W}^{n+1}}] \\ &\leq \frac{m+n}{\lambda} \mathbb{E}[{\mathcal{W}^{n}}] - \frac{T^{n+1} f(T)}{\lambda E_0(T)} \\ & < \frac{m+n}{\lambda} \mathbb{E}[{\mathcal{W}^{n}}], \end{align}

where the strict inequality indicates the loose boundary.

Now we look at the lower-order moments by focusing on moment bounds for the mean and SCV for FTPH1 distributions with b = 0.

Corollary 2

The SCV of 𝒲, SCV𝒲 = 𝔼[𝒲2]/𝔼[𝒲]2 − 1, is bounded by

\begin{equation} \frac{m+1+ \lambda T}{\lambda \mathbb{E}\left[{\mathcal{W}}\right]} - \frac{m T}{\lambda (\mathbb{E}\left[{\mathcal{W}}\right])^2} - 1\leq {\rm SCV}_\mathcal{W} < \frac{m+1}{\lambda \mathbb{E}\left[{\mathcal{W}}\right]} - 1. \end{equation}

Proof. For n = 1, Lemma 6 gives m

\begin{equation} \frac{m+1+ \lambda T}{\lambda} \mathbb{E}\left[{\mathcal{W}}\right] - \frac{m T}{\lambda} \leq \mathbb{E}[{\mathcal{W}^{2}}] \leq \frac{m+1}{\lambda} \mathbb{E}\left[{\mathcal{W}}\right] - \frac{T^{2} f(T)}{\lambda E_0(T)} , \end{equation}

whose right-hand side can be upper bounded by (m + 1)𝔼[𝒲]/λ, from which the corollary follows by dividing with (𝔼 [𝒲] )2 and subtracting 1.

We note that the difference between the upper and the lower limits of SCV𝒲 in Corollary 2 is

\begin{equation} \frac{T}{(\mathbb{E}\left[{\mathcal{W}}\right])^2}\left(\frac{m}{\lambda} - \mathbb{E}\left[{\mathcal{W}}\right]\right)\!, \end{equation}

for which, according to (8) and the definition of 𝒲, we have

$$\mathbb{E}\left[{\mathcal{W}}\right] = \mathbb{E}\left[{\mathcal{Y}\mid \mathcal{Y} < T}\right] < \mathbb{E}\left[{\mathcal{Y}}\right] \leq \frac{m}{\lambda}.$$

That is, when λ tends to 0, the upper bound converges to ∞. In this case, the λ independent upper limit mT/(m + 1) from Corollary 1 can be applied. Combining the results, we obtain 𝔼[𝒲] ≤ min{m/λ, mT/(m + 1)}.

Our final result of this subsection gives a lower bound of SCV𝒲 in terms of m only, which generalizes the result of Aldous and Shepp [Reference Aldous and Shepp1] mentioned in the introduction.

Theorem 2

The SCV of the FTPH1 random variable 𝒲 with support on (0, B) is bounded by SCV𝒲 ≥ 1/(m(m + 2)).

Proof. Recall that T = Bb = B. Define

\begin{equation} g(t) = \frac{(m-1)f(t) - tf'(t)}{mF(T) - Tf(T)} \quad\mbox{for } 0\leq t\leq T. \end{equation}

By Lemma 1, (m − 1 − λt)f (t) − tf′(t) ≥ 0 for t > 0, which leads to (m − 1)f (t) − tf′(t) > 0 for t > 0. By integrating from 0 to T, we obtain

(13)\begin{equation} \label{eq:integralSIL} mF(T)-Tf(T) =\int_0^T((m-1)f(t)-tf'(t)){\text {d}}t > 0. \end{equation}

Consequently, g(t) is a density function of a random variable, to be called Y T, with support [0, T). Note that $\int_0^T t^n{\text {d}}F(t) = \mathbb{E}[{\mathcal{Y}^n{\mathbb{I}_{\{\mathcal{Y} < T\}}}}]$ for n = 0, 1, 2, …. By routine calculations, we obtain

\begin{equation} \mathbb{E}[{Y_T}] = \frac{(m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}] - T^2f(T)}{mF(T)-Tf(T)}, \quad\quad \mathbb{E}[{Y_T^2}] = \frac{(m+2)\mathbb{E}[{\mathcal{Y}^2{\mathbb{I}_{\{\mathcal{Y} < T\}}}}] - T^3f(T)}{mF(T)-Tf(T)}. \end{equation}

It is well known that $\mathbb{E}\left[{Y_T^2}\right] /(\mathbb{E}\left[{Y_T}\right])^2 \geq 1$. Using the above expressions, we obtain

\begin{equation} (m+2)\mathbb{E}[{\mathcal{Y}^2{\mathbb{I}_{\{\mathcal{Y} < T\}}}}] \geq T^3f(T) + \frac{((m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}]-T^2f(T))^2}{mF(T)-Tf(T)}. \end{equation}

Recall that 𝒲 = 𝒴 | 𝒴 < T. We also note that 𝔼[𝒲2] = 𝔼[𝒴2𝕀{𝒴<T]/F(T) and 𝔼[𝒲] = 𝔼[𝒴𝕀{𝒴<T]/F(T). The above equation leads to

\begin{align} &\frac{\mathbb{E}[{\mathcal{W}^2}]} {(\mathbb{E}[{\mathcal{W}}])^2} \\ &\quad= \frac{\mathbb{E}[{\mathcal{Y}^2{\mathbb{I}_{\{\mathcal{Y}< T\}}}}]F(T)}{(\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])^2} \\ &\quad\geq \frac{F(T)(T^3f(T) + {((m+1)\mathbb{E}[{\mathcal{Y} {\mathbb{I}_{\{\mathcal{Y}< T\}}}}]-T^2f(T))^2}/{(mF(T)-Tf(T))})}{(m+2) (\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])^2} \\ &\quad= \frac{F(T)(((m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])^2-2(m+1)T^2f(T)\mathbb{E} [{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}]+mT^3f(T)F(T))} {(m+2)(mF(T)-Tf(T)) (\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])^2} \\ &\quad= \frac{(m+1)^2}{m(m+2)} \\ &\quad\times\biggl(\frac{1 - {2T^2f(T)}/{((m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])} + {mT^3f(T)F(T)}/{((m+1)\mathbb{E} [{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])^2}}{1 - {Tf(T)}/{(mF(T))}}\biggr) \\ &\quad= \frac{(m+1)^2}{m(m+2)} \Theta(T). \end{align}

We want to show that Θ(T) ≥ 1 for all T > 0. Since mF(T) > Tf (T) according to (13), Θ(T) ≥ 1 is equivalent to

\begin{equation} \frac{Tf(T)}{mF(T)} + \frac{mT^3f(T)F(T)}{((m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])^2} \geq 2\frac{T^2f(T)}{(m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}]}, \end{equation}

which is equivalent to

\begin{equation} \frac{1}{mF(T)} + \frac{T^2mF(T)}{((m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}])^2} \geq 2\frac{T}{(m+1)\mathbb{E}[{\mathcal{Y}{\mathbb{I}_{\{\mathcal{Y}< T\}}}}]}. \end{equation}

The last equation holds by applying the well-known inequality a 2 + b 2 ≥ 2ab for any real numbers a and b. Thus, we have shown that Θ(T) ≥ 1 for all T > 0. Consequently, we have shown that

\begin{equation} \frac{\mathbb{E}[{\mathcal{W}^2}]}{(\mathbb{E}\left[{\mathcal{W}}\right])^2} \geq \frac{(m+1)^2}{m(m+2)}, \end{equation}

which is equivalent to SCV𝒲 ≥ 1/(m(m + 2)).

By Corollary 1, the lower bound of SCV𝒲 is strict for all PH distributions with finite support. The following lemma and corollary show how the lower bound of SCV𝒲 can be attained approximately by bounded Erlang distributions. For t ≥ 0, denote by

\begin{align} F_m(t) =\,&1-\text {e}^{-\lambda t}-\lambda t\text {e}^{-\lambda t} - \cdots - \frac{(\lambda t)^{m-1}}{(m-1)!}\text {e}^{-\lambda t}, \\ &f_m(t) = f_{(m,\lambda)}(t) = \frac{\lambda^mt^{m-1}}{(m-1)!}\text {e}^{-\lambda t}, \end{align}

the distribution function and density function of an Erlang random variable 𝒳(m,λ), respectively. By routine calculations, we obtain

\begin{align} \mathbb{E}[\mathcal{X}_{(m, \lambda)}\mid \mathcal{X}_{(m, \lambda)} < T] = \frac{\int_0^T tf_m(t){\text {d}}t}{F_m(T)} &= \frac{m}{\lambda}\frac{F_{m+1}(T)}{F_m(T)}, \\ \mathbb{E}[\mathcal{X}_{(m, \lambda)}^2\mid \mathcal{X}_{(m, \lambda)} < T] = \frac{\int_0^Tt^2f_m(t){\text {d}}t}{F_m(T)} =\,& \frac{m(m+1)}{\lambda^2}\frac{F_{m+2}(T)}{F_m(T)}, \\ {\rm SCV}_{\{\mathcal{X}_{(m, \lambda)}\mid \mathcal{X}_{(m, \lambda)}< T\}} = \frac{ \mathbb{E}[{\mathcal{X}_{(m, \lambda)}^2\mid \mathcal{X}_{(m, \lambda)} < T}]}{ (\mathbb{E}[{\mathcal{X}_{(m, \lambda)}\mid \mathcal{X}_{(m, \lambda)} < T}])^2} -1 &= \biggl(\frac{ m+1}{ m}\biggr)\frac{ F_m(T)F_{m+2}(T)}{ (F_{m+1}(T))^2} -1. \end{align}

Lemma 7

For all T > 0, the following bounds apply for the distribution functions of Erlang random variables:

\begin{equation} \frac{m+1}{m+2} \leq \frac{F_m(T)F_{m+2}(T)}{(F_{m+1}(T))^2} \leq 1. \end{equation}

In addition, we have limT→0 F m(T)F m+2(T)/(F m+1(T))2 = (m + 1)/(m + 2) and limT→∞ F m(T)F m+2(T)/(F m+1(T))2 = 1.

Proof. For convenience, we denote λT as t in this proof. First, we prove the upper bound. Since $\text {e}^{\lambda t} = \sum_{k=0}^\infty (\lambda t)^k/k!$, we need to show that

\begin{equation} \bigg(\sum_{k=m+2}^\infty\frac{t^k}{k!}\bigg)\bigg(\frac{t^m}{m!} + \frac{t^{m+1}}{(m+1)!}+\sum_{k=m+2}^\infty\frac{t^k}{k!}\bigg) \leq \bigg(\frac{t^{m+1}}{(m+1)!}+\sum_{k=m+2}^\infty\frac{t^k}{k!}\bigg)^2, \end{equation}

which can be reduced to

\begin{equation} \bigg(\sum_{k=m+2}^\infty\frac{t^k}{k!}\bigg)\frac{t^m}{m!} \leq \bigg(\sum_{k=m+2}^\infty\frac{t^k}{k!}\bigg)\frac{t^{m+1}}{(m+1)!} + \bigg(\frac{t^{m+1}}{(m+1)!}\bigg)^2. \end{equation}

Next, we compare the coefficients of t k on both sides. For k = 2m + 2, the left-hand side is 1/((m !(m + 2)!) and the right-hand side is 1/((m + 1)!)2. From

\begin{equation} \frac{1}{m! (m+2)! } = \frac{m+1}{m+2} \left(\frac{1}{(m+1)!}\right)^2\!, \end{equation}

it follows that the left-hand side is smaller than the right-hand side. For k ≥ 2m + 3, we have, for k = j + m,

\begin{equation} \frac{1}{j!\,m!} \leq \frac{1}{(j-1)!\,(m+1)!}, \end{equation}

which is equivalent to m + 1 ≤ j, and holds since j = kmm + 3. Consequently, we have shown the upper bound.

The proof of the lower bound is similar but tedious. The lower-bound expression can be rewritten as $(m+1)F^2_{m+1}(t/\lambda)\leq (m+2)F_m(t/\lambda)F_{m+2}(t/\lambda)$, which can be rewritten explicitly as

\begin{equation} (m+1)\bigg(\sum_{k=m}^\infty\frac{t^k}{k!}-\frac{t^{m}}{m!}\bigg)^2 \leq (m+2)\bigg(\sum_{k=m}^\infty\frac{t^k}{k!}\bigg)\bigg(\sum_{k=m}^\infty\frac{t^k}{k!}-\frac{t^m}{m!}-\frac{t^{m+1}}{(m+1)!}\bigg), \end{equation}

which leads to

\begin{equation} (m+2)\bigg(\sum_{k=m}^\infty\frac{t^k}{k!}\bigg)\frac{t^{m+1}}{(m+1)!} + (m+1)\bigg(\frac{t^{m}}{m!}\bigg)^2 \leq \bigg(\sum_{k=m}^\infty\frac{t^k}{k!}\bigg)^2 + m\bigg(\sum_{k=m}^\infty\frac{t^k}{k!}\bigg)\frac{t^{m}}{m!}. \end{equation}

To prove the above inequality, we compare the coefficients of t n on both sides. For n = 2m, we have

\begin{equation} \frac{m+1}{m!\,m!} \leq \frac{1}{m!\,m!} + \frac{m}{m!\,m!}, \end{equation}

which holds. For n ≥ 2m + 1, we need to prove that

\begin{equation} \frac{m+2}{(n - m - 1)!\,(m + 1)!} \leq \frac{m}{(n - m)!\,m!} + \sum_{i=m}^{n-m}\frac{1}{i!(n - i)!}. \end{equation}

Separating the first and last terms of the summation and applying k ! = k(k − 1)!, we obtain

\begin{equation} \frac{(m+2)(n-m)}{(n - m)!\,(m + 1)!} \leq \frac{(m+2)(m+1)}{(n - m)!\,(m+1)!} + \sum_{i=m+1}^{n - m - 1}\frac{1}{i!\,(n - i)!}, \end{equation}

which leads to

(14)\begin{equation} \label{eq:factineq} \frac{(m+2)(n-2m-1)}{(n-m)!\,(m+1)!} \leq \sum_{i=m+1}^{n-m-1}\frac{1}{i!\,(n-i)!}. \end{equation}

For any i ∈ {m + 1, … , nm − 1}, we have inm and m + 1 ≤ ni, from which we can write

\begin{align}\\[-24pt] &\quad\ \ \quad\quad\biggl(\frac{i}{n-m} \frac{i-1}{n-m-1}\cdots \frac{m+2}{n-i+2}\biggr) \frac{m+2}{n-i+1} \leq 1, \\ &(i(i-1)\cdots(m+2)) (m+2) \leq (n-m) (n-m-1) \cdots (n-i+1), \\ &\ \frac{i!}{(m+1)!} (m+2) \leq \frac{(n-m)!}{(n-i)!}, \quad\quad \frac{m+2}{(n-m)!(m+1)!} \leq \frac{1}{i!\,(n-i)!}, \end{align}

Considering that nm − 1 − (m + 1) + 1 = n − 2m − 1 terms are summed on the right-hand side of (14), each of which is greater than or equal to (m + 2)/((nm)! (m + 1)!), inequality (14) as well as the lower bound of Lemma 7 are proved. This completes the proof of the lemma.

Immediate consequences of Lemma 7 are a lower bound and an upper bound of the SCV for bounded 𝒳(m,λ).

Corollary 3

Assume that 𝒳 has an Erlang distribution with parameters (m, λ). For all T > 0, we have

\begin{equation} \frac{1}{m(m+2)} \leq {\rm SCV}_{\{\mathcal{X}_{(m, \lambda)}\mid \mathcal{X}_{(m, \lambda)} < T\}} \leq \frac {1}{m}. \end{equation}

3.2. The b > 0 case

Let $\mathcal{Z}=b+\mathcal{W}=b+(\mathcal{Y}mid \mathcal{Y} < T)$. Then, for $\mathbb{E}\left[{\mathcal{Z}^n}\right]$, we have

$$\mathbb{E}[{\mathcal{Z}^n}] = \sum_{i=0}^{n} \binom{n}{i} b^{n-i} \frac{E_i(T)}{E_0(T)} = \sum_{i=0}^{n} \binom{n}{i} b^{n-i} \mathbb{E}[{\mathcal{W}^i}],$$

where E i(T) is defined as before. That is, for n = 1, 2, we have

$$\mathbb{E}\left[{\mathcal{Z}}\right] = b + \mathbb{E}\left[{\mathcal{W}}\right] \quad\text{and}\quad \mathbb{E}[{\mathcal{Z}^2}] = b^2 + 2 b \mathbb{E}\left[{\mathcal{W}}\right] + \mathbb{E}[{\mathcal{W}^2}].$$

Corollary 4

The nth moment of an FTPH1 random variable, $\mathcal{Z}$, with support on (b, B) is bounded by

(15)\begin{equation} \label{eq:bound_mean} b^n \leq \mathbb{E}[{\mathcal{Z}^n}] \leq \sum_{i=0}^{n} \binom{n}{i} b^{n-i} \frac{m T^i}{m+i}. \end{equation}

Proof. Equation (15) directly follows from Corollary 1.

For lower-order moments, according to (15), the mean of $\mathcal{Z}$ is bounded by

\begin{equation} b \leq \mathbb{E}[{\mathcal{Z}}] \leq b+ \frac{m T}{m+1} = \frac{b +m B}{m+1} < B, \end{equation}

where both moment bounds are tight. The lower boundary is reached when 𝔼[𝒴] tends to 0 and the upper boundary is reached when 𝒴 is Erlang(λ, m) distributed and λ tends to 0.

For the SCV, we have the following corollary.

Corollary 5

The ${\rm SCV}_\mathcal{Z}$ is bounded by the following λ independent and dependent moment bounds:

\begin{align} &\quad\quad{\rm SCV}_\mathcal{Z} = \frac{\mathbb{E}[{\mathcal{W}^2}]- \mathbb{E}\left[{\mathcal{W}}\right]^2}{(b + \mathbb{E}\left[{\mathcal{W}}\right])^2} \leq \frac{{(m+1)}\mathbb{E}\left[{\mathcal{W}}\right] T/{(m+2)} - \mathbb{E}\left[{\mathcal{W}}\right]^2}{(b + \mathbb{E}\left[{\mathcal{W}}\right])^2}, \\ &\frac{- m T + (m+1+ \lambda T) \mathbb{E}\left[{\mathcal{W}}\right]- \lambda \mathbb{E}\left[{\mathcal{W}}\right]^2}{\lambda(b + \mathbb{E}\left[{\mathcal{W}}\right])^2} \leq {\rm SCV}_\mathcal{Z} < \frac{(m+1) \mathbb{E}\left[{\mathcal{W}}\right]-\lambda \mathbb{E}\left[{\mathcal{W}}\right]^2}{\lambda(b + \mathbb{E}\left[{\mathcal{W}}\right])^2}. \end{align}

Proof. From Lemmas 4 and 6 we have 𝔼[𝒲2] ≤ (m + 1)T𝔼 [𝒲]/(m + 2) and

(16)\begin{align} &\frac{m+1+ \lambda T}{\lambda} \mathbb{E}\left[{\mathcal{W}}\right] - \frac{m T}{\lambda} \leq \mathbb{E}[{\mathcal{W}^2}] < \frac{m+1}{\lambda} \mathbb{E}\left[{\mathcal{W}}\right], \label{eq:utcso} \end{align}

respectively. Subtracting 𝔼[𝒲]2 and then dividing by (b + 𝔼[𝒲])2 in (16) gives the corollary.

Different from the b = 0 case, the ${\rm SCV}_{\mathcal{Z}}$ can reach 0 in the b > 0 case.

4. Discussion and conclusion

In this paper we presented new moment bounds on PH distributions with infinite and finite supports by using the steepest increase property. For PH distributions with infinite support and a PH representation (α, A) of size m, denoted as 𝒴, we have

  • shown that any PH distribution is stochastically smaller than or equal to an Erlang distribution 𝒳(m,λ) with λ the absolute value of the dominant eigenvalue of A; and

  • obtained upper bounds of moments in terms of m and λ (e.g. 𝔼[𝒴] ≤ m/λ).

For PH distributions with finite support (for the set FTPH1), denoted as 𝒲 = 𝒴 | 𝒴 < T, we have

  • obtained upper bounds of moments in terms of m and T;

  • obtained lower and upper bounds of moments depending on λ;

  • shown that 𝔼[𝒲]≤ min{mT/(m + 1), m/λ}; and

  • shown that SCV𝒲 ≥ 1/(m(m + 2)).

For the finite support case, we focused on the distribution set FTPH1. Results for the set FTPH2 can be obtained similarly. The set FTPH3 is a convex mixture of FTPH1 and FTPH2. Moment bounds can also be obtained as a convex mixture of the moment bounds obtained for FTPH1 and FTPH2, but this is outside the scope of the current work.

Acknowledgements

This work was supported by the Hungarian research project OTKA K123914, a NSERC

discovery grant (Canada), and by the János Bolyai Research Scholarship of the Hungarian Academy of Sciences. The authors would like to thank two anonymous referees and an Associate Editor for their exceptionally careful reviews of the paper.

References

Aldous, D. and Shepp, L. (1987). The least variable phase type distribution is Erlang. Commun. Statist. Stoch. Models 3, 467473.Google Scholar
He, Q.-M., Zhang, H., and Vera, J. (2013). Majorization and extremal PH-distributions. In Matrix-Analytic Methods in Stochastic Models (Springer Proc. Math. Statist. 27), eds Latouche, G. et al., Springer, New York, pp. 107121.CrossRefGoogle Scholar
Neuts, M. F. (1975). Probability distributions of phase type. In Liber Amicorum Prof. Emeritus H. Florin, University of Louvain, Belgium, pp. 173206.Google Scholar
Neuts, M. F. (1981). Matrix-Geometric Solutions in Stochastic Models: An Algorithmic Approach. Johns Hopkins University Press, Baltimore.Google Scholar
O’cinneide, C. A. (1989). On non-uniqueness of representations of phase-type distributions. Commun. Statist. Stoch. Models 5, 247259.CrossRefGoogle Scholar
O’CINNEIDE, C. A. (1990). Characterization of phase-type distributions. Commun. Statist. Stoch. Models 6, 157.Google Scholar
O’cinneide, C. A. (1991). Phase-type distributions and majorization. Ann. Appl. Prob. 1, 219227.Google Scholar
O’cinneide, C. A. (1999). Phase-type distributions: open problems and a few properties. Commun. Statist. Stoch. Models, 15, 731757,Google Scholar
Ramaswami, V. and Viswanath, N. C. (2014). Phase type distributions with finite support. Stoch. Models, 30, 576597.Google Scholar
Stoyan, D. (1983). Comparison Methods for Queues and Other Stochastic Models. John Wiley, Chichester.Google Scholar
Telek, M. (2000). The minimal coefficient of variation of discrete phase type distributions. Proc. 3rd Internat. Conf. on Matrix Analytic Methods in Stochastic Models, eds Latouche, G. and G. Taylor, P., Notable Publications, New Jersey.Google Scholar
Yao, R. (2002). A proof of the steepest increase conjecture of a phase-type density. Stoch. Models, 18, 16.Google Scholar
Figure 0

FIGURE 1: Bounds of the SCV for ordinary PH distributions. The lower bound is the well known 1/m bound provided in [1] and [7], while the upper bound is provided based on the steepest increase property.

Figure 1

FIGURE 2: Some interesting members of the FTPH1 class.