Hostname: page-component-745bb68f8f-5r2nc Total loading time: 0 Render date: 2025-02-11T08:38:01.422Z Has data issue: false hasContentIssue false

On component failure in coherent systems with applications to maintenance strategies

Published online by Cambridge University Press:  03 December 2020

M. Hashemi*
Affiliation:
University of Isfahan
M. Asadi*
Affiliation:
University of Isfahan and IPM
*
*Postal address: Department of Statistics, Faculty of Mathematics and Statistics, University of Isfahan, Isfahan 81746-73441, Iran.
*Postal address: Department of Statistics, Faculty of Mathematics and Statistics, University of Isfahan, Isfahan 81746-73441, Iran.
Rights & Permissions [Opens in a new window]

Abstract

Providing optimal strategies for maintaining technical systems in good working condition is an important goal in reliability engineering. The main aim of this paper is to propose some optimal maintenance policies for coherent systems based on some partial information about the status of components in the system. For this purpose, in the first part of the paper, we propose two criteria under which we compute the probability of the number of failed components in a coherent system with independent and identically distributed components. The first proposed criterion utilizes partial information about the status of the components with a single inspection of the system, and the second one uses partial information about the status of component failure under double monitoring of the system. In the computation of both criteria, we use the notion of the signature vector associated with the system. Some stochastic comparisons between two coherent systems have been made based on the proposed concepts. Then, by imposing some cost functions, we introduce new approaches to the optimal corrective and preventive maintenance of coherent systems. To illustrate the results, some examples are examined numerically and graphically.

Type
Original Article
Copyright
© Applied Probability Trust 2020

1. Introduction

Nowadays, coherent systems are used in many areas of human life, such as industrial manufacturing lines, airplane systems, power supply systems, and telecommunication systems. In reliability engineering, an n-component system is called coherent if the structure function of the system is nondecreasing and the system has no irrelevant components (see Barlow and Proschan [Reference Barlow and Proschan7]). A well-known subclass of the class of coherent systems is that of k-out-of-n systems. Recall that an n-component system is said to be a k-out-of-n system if it operates when at least k components out of n operate.

In the last two decades, an extensive number of research works have been reported assessing the reliability and stochastic properties of coherent systems using various approaches. An approach which has recently received great attention is to use the notion of signature. Let ${X_1,X_2,\dots,X_n}$ denote the lifetimes of an n-component coherent system and let ${T=T(X_1,\dots,X_n)}$ be the system lifetime. Under the assumption that the component lifetimes are independent and identically distributed (i.i.d.), Samaniego [Reference Samaniego31] defined the concept of signature to express the reliability function of the system lifetime as a mixture representation of the reliability function of ordered component lifetimes. To be more precise, let ${X_{1:n},X_{2:n},\dots,X_{n:n}}$ denote the order statistics corresponding to the lifetimes ${X_1,\dots,X_n}$ . Then the reliability function of the system’s lifetime, at time t, can be expressed as

(1) \begin{equation} \mathbb{P}(T>t)=\sum_{i=1}^{n} s_i \mathbb{P}(X_{i:n}>t),\end{equation}

where ${s_i=\mathbb{P}(T=X_{i:n}),}$ ${i=1,2,\dots,n}$ . The probability vector ${\textbf{{s}}=(s_1,s_2,...,s_n)}$ is called the signature vector of the system. The ith element of the vector ${\textbf{{s}}}$ is calculated as ${s_i=n_i/n!}$ , where ${n_i}$ denotes the number of permutations of components under which the ith component failure causes the system failure (see Samaniego [Reference Samaniego31]). It is known that the vector ${\textbf{{s}}}$ depends only on the structure of the system. The representation (1) is valid under the weaker condition that the component lifetimes are exchangeable (see Navarro and Rychlik [Reference Navarro and Rychlik23]). For references on the signature-based properties of system lifetime, we refer the reader to [Reference Balakrishnan, Navarro and Samaniego5], [Reference Eryilmaz11], [Reference Kochar, Mukerjee and Samaniego17], [Reference Mahmoudi and Asadi20], [Reference Navarro, Samaniego and Balakrishnan24], [Reference Navarro, Samaniego and Balakrishnan25], [Reference Samaniego32], [Reference Spizzichino36], [Reference Tavangar38], and [Reference Zhang41]. For a recent work on different methods and algorithms for computing system signature, see Reed [Reference Reed29] and references therein.

In assessing the reliability and stochastic characteristics of systems, a problem of interest for engineers and system designers is to maintain the system in optimum working condition and to determine the number of spares that should be available in the depot for this purpose. The importance of this problem arises from the fact that the failure and unavailability of the system may cause high unexpected costs to the users. In many complex coherent systems, such as k-out-of-n systems, the design of the structure of the system is such that the system operates even if a number of components have already failed. However, if the number of failed components passes a certain threshold, then the system does fail. Hence, the computation of the probability of the number of failed components in the system, under various conditions, is important for the system operators. These probabilities provide crucial information for preventing the system’s failure and maintaining the system in optimal operating condition. The aim of maintenance schedules is mainly to diminish the occurrence of system failure or to change the status of a failed system to the working state. For this purpose, operators try to restore a failed component to an operative state. In the literature, this maintenance action is called corrective maintenance (CM). In a CM action, the failed components may undergo repair or may be replaced. Two other important actions in maintenance theory are (a) minimal repair, which eliminates the failure but does not change the failure rate, and (b) preventive maintenance (PM), which means performing a maintenance policy for an operating system (component) to bring the system (component) back to better working condition. Throughout the paper, we assume that the PM (CM) is perfect in the sense that an unfailed (failed) component is returned to ‘as-good-as-new’ condition.

In the literature, many research papers and books have been devoted to various maintenance schedules. We refer the reader to [Reference Bada, Berrade, Cha and Lee4], [Reference Barlow and Hunter6], [Reference Cha, Finkelstein and Levitin9], [Reference Finkelstein and Gertsbakh12], [Reference Finkelstein and Gertsbakh13], [Reference Gertsbakh14], [Reference Levitin, Finkelstein and Dai18], [Reference Morse21], [Reference Pham and Wang27], [Reference Sheu34], [Reference Wang and Pham39], and [Reference Zhang, Fouladirad and Barros42]. Recently, some comparisons of policies for minimal repair of systems have been studied in Belzunce et al. [Reference Belzunce, Martnez-Riquelme and Ruiz8] and Arriaza et al. [Reference Arriaza, Navarro and Suárez-Llorens1].

The main objective of the present research is to propose some maintenance policies for a coherent system under some partial information on the number of failures in the system. So far, only a small portion of the literature has considered maintenance of a multi-component system. Most of the works on this topic have been limited to a one-unit system, or to an entire system treated as a single-unit system. Finkelstein and Gertsbakh [Reference Finkelstein and Gertsbakh12], [Reference Finkelstein and Gertsbakh13] studied PM for networks (systems) where the components fail based on shock models. Cha et al. [Reference Cha, Finkelstein and Levitin9] considered PM of items operating in a random environment. Zarezadeh and Asadi [Reference Zarezadeh and Asadi40] studied PM scheduling for systems under multiple external shocks. We introduce two new optimal strategies for the maintenance of an n-component coherent system, with i.i.d. component lifetimes and signature vector ${\textbf{{s}}}$ , under the condition that there is some partial information on the number of destroyed components in the system. The information is collected under two scenarios: single inspection and double inspection of the system. In the first strategy, before a predetermined time ${\tau}$ the system undergoes minimal repair, and after ${\tau}$ the system is equipped with a warning light that turns on at the time of the kth component failure. Then the system is inspected at time t, ${t>\tau}$ . The operator decides to perform CM on the entire system when the system fails, or to perform distinct PM actions (depending on whether the light turns on or not) when the total operating time reaches t, whichever occurs first. In the second strategy, the system is inspected at two times ${t_1}$ and ${t_2}$ , ${t_1<t_2}$ , and depending on the information obtained at ${t_1}$ , the operator performs different maintenance actions at ${t_2}$ .

Suppose the system starts to operate at time ${t=0}$ and each component may fail over time. Assume that the system is functioning at time t and at least k components have failed before t. Under these assumptions, in Section 2, we compute the probability of the number of failed components in the system. In other words, if ${N_t}$ is the number of failed components up to time t, then we calculate the following conditional probabilities:

(2) \begin{eqnarray} p_{k,n}^{t}(i)=\mathbb{P}(N_t=i\mid X_{k:n}\leq t<T),\ \ \ \ i=k,..., n-1.\end{eqnarray}

The second scenario considers the condition that exactly k components have failed by time ${t_1}$ , and at time ${t_2}$ ${(t_2 > t_1)}$ the system is still operating. Under this condition, we calculate the probability of the number of failed components ${N_{t_2}}$ ; i.e., we calculate

(3) \begin{eqnarray}p_{k,n}^{t_1,t_2}(i)=\mathbb{P}(N_{t_2}=i\mid X_{k:n}\leq t_1 < X_{k+1:n}, {T >t_2}),\ \ \ \ i=k,..., n-1.\end{eqnarray}

We investigate the properties of ${p_{k,n}^{t}(i)}$ and ${p_{k,n}^{t_1,t_2}(i)}$ in Section 2. In particular, some stochastic ordering results for these conditional probabilities are established. In Section 3, we propose some optimal PM policies for coherent systems with n components as applications of ${p_{k,n}^{t}(i)}$ and ${p_{k,n}^{t_1,t_2}(i)}$ . The criteria which will be employed to obtain the optimal PM time for the system are the minimal long-run expected cost per unit of time and stationary availability of the system. We examine the results of the paper by considering the well-known bridge system consisting of 5 components. Using the presented bridge system, the robustness of the proposed approaches is also analyzed numerically. The graphical and computational results of the paper are obtained using ${\textrm{MATHEMATICA}^{\circledR}}$ , Version 10.

The following auxiliary concepts and definitions are useful in our derivations.

Definition 1.1. Assume that X and Y are nonnegative random variables with cumulative distribution functions (CDFs) F and G, probability density functions f and g, and reliability functions ${\bar{F}=1-F}$ and ${\bar{G}=1-G}$ , respectively.

  1. (a) If ${\frac{g(t)}{f(t)}}$ is increasing in ${t\geq 0}$ , then X is said to be less than Y in the likelihood ratio ordering (denoted by ${X\leq_{lr}Y}$ ).

  2. (b) If ${\frac{\bar{G}(t)}{\bar{F}(t)}}$ is increasing in ${t\geq 0}$ , then X is said to be less than Y in the hazard rate ordering (denoted by ${X\leq_{hr}Y}$ ).

  3. (c) If ${\bar{G}(t)\geq \bar{F}(t)}$ for all ${t\geq 0}$ , then X is said to be less than Y in the usual stochastic ordering (denoted by ${X\leq_{st}Y}$ ).

The implications among these orderings are as follows:

\[ X\leq_{lr}Y \ \ \Rightarrow \ \ X\leq_{hr}Y \ \ \Rightarrow \ \ X\leq_{st}Y.\]

Definition 1.2.

  1. (a) A probability vector ${\textbf{{p}}=(p_1,p_2,...,p_n)}$ is said to have increasing failure rate (IFR) if ${p_k/ \sum_{i=k}^{n}p_i}$ is increasing in ${k=1,2,...,n}$ .

  2. (b) For two probability vectors ${\textbf{{p}}=(p_1,p_2,...,p_n)}$ and ${\textbf{{q}}=(q_1,q_2,...,q_n)}$ , if ${\sum_{j=i}^{n}q_j/\sum_{j=i}^{n}p_j}$ is increasing in i, then ${\textbf{{p}}}$ is said to be less than ${\textbf{{q}}}$ in the hazard rate order (denoted by ${\textbf{{p}}\leq_{hr}\textbf{{q}}}$ ).

For more details on various notions of partial orderings and their applications, we refer to Shaked and Shanthikumar [Reference Shaked and Shanthikumar33] and Lucia [Reference Lucia19].

Definition 1.3. (Karlin [Reference Karlin15].) A nonnegative function h(x,y) is totally positive of order 2 (TP ${_2}$ ) if ${h(x_1,y_1)h(x_2,y_2)-h(x_1,y_2)h(x_2,y_1)\geq 0}$ whenever ${x_1<x_2}$ and ${y_1<y_2}$ . The function h(x, y) is said to be reverse regular of order 2 (RR ${_2}$ ) if ${h(x_1,y_1)h(x_2,y_2)-h(x_1,y_2)h(x_2,y_1)\leq 0}$ whenever ${x_1<x_2}$ and ${y_1<y_2}$ .

2. The probability of the number of failed components of the system

In the present section, first we obtain the functional forms of the conditional probabilities ${p_{k,n}^{t}(i)}$ and ${p_{k,n}^{t_1,t_2}(i)}$ . These conditional probabilities are useful in our derivations in Section 3 to establish the new optimal maintenance strategies on the coherent systems.

Single inspection

Consider a coherent system with lifetime T, as described in the introduction. The system begins to operate at time ${t=0}$ and each component is subject to failure over time. Assume that the operator inspects the system at time t and he/she observes that at least k components have already failed before t, but the system is still working. As we mentioned in the previous section, the probability of the number of failed components up to time t is given by ${p_{k,n}^{t}(i)}$ in (2). Asadi and Berred [Reference Asadi and Berred2] explored several properties of the above conditional probability ${p^t_{k,n}(i)}$ in the case that ${k=0}$ . Eryilmaz [Reference Eryilmaz10] considered the number of failed components for a coherent system whose component lifetimes are exchangeable. Ashrafi and Asadi [Reference Ashrafi and Asadi3] studied the number of failed components in three-state networks under different conditions and applied their results for age-replacement of three-state networks. The conditional probability ${p^t_{k,n}(i)}$ can be represented as follows:

(4) \begin{eqnarray}p^t_{k,n}(i)&\,=\,&\mathbb{P}(N_t=i\mid X_{k:n}\leq t<T)\nonumber\\[3pt]&\,=\,&\frac{\bar{S}_{i}\binom{n}{i}\phi^i(t)}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\phi^j(t)}, \ \ \ i=k,...,n-1,\end{eqnarray}

where ${\bar{S}_i=\sum_{j=i+1}^{n}s_j}$ and ${\phi(t)=F(t)/\bar{F}(t)}$ , provided that ${\bar{F}(t)=1-F(t)>0}$ ; see Appendix A for the proof. The expectation of ${(N_t\mid X_{k:n}\leq t<T)}$ can be calculated as follows:

(5) \begin{eqnarray}\mathbb{E}(N_t\mid X_{k:n}\leq t<T)=\frac{n\sum_{i=k}^{n-1}\bar{S}_{i}\binom{n-1}{i-1}\phi^i(t)}{\sum_{j=k}^{n-1}\bar{S}_{j}\binom{n}{j}\phi^j(t)}, \ \ \ k=0,1,..., n-1.\end{eqnarray}

One can easily verify that the common reliability function ${\bar{F}(t)}$ of the components can be recovered from two successive values ${p^t_{k,n}(i)}$ and ${p^t_{k,n}(i+1)}$ as follows:

\[ \bar{F}(t)=\left(1+\frac{i+1}{n-i}\left(\frac{s_{i+1}}{\bar{S}_{i+1}}+1\right) \frac{p^t_{k,n}(i+1)}{p^t_{k,n}(i)}\right)^{-1}, \quad t>0.\]

Remark 2.1. The probability ${p^t_{k,n}(i)}$ may be interpreted from a Bayesian viewpoint. Assume that the system designer is interested in the failure probability of the system components at time t, denoted by ${\mathbb{P}(N_t=i)}$ , ${i=1,2,...,n}$ , as his/her prior belief (prior distribution) when the system has not yet been put into operation. Now suppose that the system starts to perform its mission at time ${t=0}$ , and the designer has information that at least k components have failed before t, while the system is functioning. Then the probability ${p^t_{k,n}(i)}$ can be viewed as the designer’s posterior belief on ${N_t}$ , given the information that is provided for the designer.

Remark 2.2. Asadi and Berred [Reference Asadi and Berred2] investigated the time-dependent behavior of ${p^t_{0,n}(i)}$ for different values of i. The following results can be established regarding the time-dependent behavior of ${p^t_{k,n}(i)}$ ; these are similar to Theorem 2.3 of [Reference Asadi and Berred2]. Let ${i^*=\max\{i:\ s_i>0\}}$ . Then (a) ${p^t_{k,n}(k)}$ is decreasing in t; (b) for ${i=k+1,...,i^*-2}$ , ${p^t_{k,n}(i)}$ first increases with respect to t until it attains its maximum, then declines; and (c) ${p^t_{k,n}(i^*-1)}$ is increasing in t. We omit the proof, which is analogous to that of Theorem 2.3 in Asadi and Berred [Reference Asadi and Berred2]. The next example gives applications of these results.

Figure 1: The bridge system.

Example 2.1. In Figure 1, a bridge system consisting of five identical components is pictured. We assume that the component lifetimes are i.i.d. random variables having Weibull distribution with CDF ${F(t)=1-\exp\{-t^{2}\}}$ , ${t\geq0}$ . The system signature can be computed as ${\textbf{{s}}=(0,0.2,0.6,0.2,0)}$ . Then we have ${\phi(t)=e^{t^2}-1}$ and

(6) \begin{eqnarray} p^t_{k,5}(i)=\frac{\bar{S}_{i}\binom{5}{i}(e^{t^2}-1)^i}{\sum_{j=k}^{4}\bar{S}_j\binom{5}{j}(e^{t^2}-1)^j}, \ \ \ 0\leq k\leq i\leq 4. \end{eqnarray}

As can be seen in Figure 2(a), ${p^t_{1,5}(1)}$ is decreasing in t, ${p^t_{1,5}(2)}$ increases for a period of time and then decreases, and ${p^t_{1,5}(3)}$ is increasing in t (see Remark 2.2). It should be mentioned that in this system, ${p^t_{1,5}(4)=0}$ , since for the component with lifetime ${X_{5:5}}$ , ${\mathbb{P}(T=X_{5:5})=s_5=0}$ ; i.e., the component with lifetime ${X_{5:5}}$ will never entail the failure of the system. Also, using Equation (5), we obtain

(7) \begin{eqnarray}H_{k,n}^t\,:\!=\mathbb{E}(N_t\mid X_{k:n}\leq t<T)=\frac{5\sum_{i=k}^{4}\bar{S}_{i}\binom{4}{i-1}(e^{t^2}-1)^i}{\sum_{j=k}^{4}\bar{S}_{j}\binom{5}{j}(e^{t^2}-1)^j},\ \ \ 0\leq k\leq 4.\end{eqnarray}

Figure 2(b) shows the graph of ${H_{k,n}^t}$ as a function of t for ${k=0,1,2}$ . It is seen that ${H_{k,n}^t}$ is increasing in k and t. We have proved the same result for a general system structure and an arbitrary baseline distribution in Corollary 2.1.

Figure 2: (a) The plots of ${p^t_{k,n}(i)}$ for ${i=1, 2,3}$ and ${k=1}$ in Example 2.1. (b) The plots of ${H_{k,n}^t}$ for ${k=0, 1,2}$ in Example 2.1.

The following theorem gives some stochastic properties of ${(N_t\mid X_{k:n}\leq t<T)}$ . Before giving the theorem, we mention the well-known result that, if X and Y are two nonnegative random variables with probability density functions ${f_1}$ and ${f_2}$ , respectively, then the order ${X\leq_{lr}Y}$ is equivalent to saying that ${f_m(t)}$ is TP ${_2}$ in ${(m,t)\in \{1,2\}\times [0,\infty)}$ (see Shaked and Shanthikumar [Reference Shaked and Shanthikumar33]).

Theorem 2.1. Assume that the distribution function F is absolutely continuous. Then the conditional random variable ${(N_t\mid X_{k:n}\leq t<T)}$ in (2) satisfies in the following orderings:

  1. (a) for ${0\leq k\leq n-1}$ , ${(N_t\mid X_{k:n}\leq t<T)\leq_{\mathrm{lr}}(N_t\mid X_{k+1:n}\leq t<T)}$ ;

  2. (b) for each ${0<t_1\leq t_2}$ , ${(N_{t_1}\mid X_{k:n}\leq t_1<T)\leq_{\mathrm{lr}}(N_{t_2}\mid X_{k:n}\leq t_2<T)}$ .

Proof. See Appendix B. □

It is known that the likelihood ratio order between two random variables implies the usual stochastic order between them. Using this fact and Theorem 2.1, one can verify the behavior of ${\mathbb{E}(N_t\mid X_{k:n}\leq t<T)}$ in terms of k and t, as shown in the following corollary.

Corollary 2.1. ${\mathbb{E}(N_t\mid X_{k:n}\leq t<T)}$ is increasing in terms of k and t.

We now provide a comparison between two different coherent systems based on the number of failed components in each system.

Theorem 2.2. Suppose that two coherent systems with, respectively, orders n and ${n+1}$ have signature vectors ${\textbf{{s}}^{(1)}=(s_1, s_2,... , s_n)}$ and ${\textbf{{s}}^{(2)}=(p_1, p_2,... , p_{n+1})}$ . Let the component lifetimes of the first system be ${X_1,X_2,...,X_n}$ , and let those of the second system be ${Y_1,Y_2,...,Y_{n+1}}$ , where in the two systems the components are independent and have a common distribution function F. Denote by ${N_t}$ and ${N_t^*}$ the number of failed components of the first and second systems, respectively, at time t. If ${\textbf{{s}}^{(1)}\leq _{\rm hr}\textbf{{s}}^{(2)}}$ , then

\[(N_t^*\mid Y_{k:n+1}\leq t<T_2)\geq_{\rm lr}(N_t\mid X_{k:n}\leq t<T_1),\ \ \ t\geq 0,\]

where ${T_1}$ and ${T_2}$ denote the lifetimes of the system with order n and the system with order ${n+1}$ , respectively.

Proof. See Appendix C. □

The next result gives a likelihood ratio order comparison on the failed components between two coherent systems.

Theorem 2.3. Assume that ${\mathcal{S}_{1}}$ and ${\mathcal{S}_{2}}$ are two coherent systems with i.i.d. component lifetimes ${X_1,X_2,...,X_n}$ and ${Y_1,Y_2,...,Y_n}$ whose CDFs are F and G, respectively. Let ${N_t}$ and ${N_t^*}$ be the numbers of failed components in ${\mathcal{S}_{1}}$ and ${\mathcal{S}_{2}}$ , respectively, on [0, t]. Further, suppose that the lifetime of the system ${\mathcal{S}_{1}}$ ( ${\mathcal{S}_{2}}$ ) is denoted by ${T_1}$ ( ${T_2}$ ) and the corresponding signature vector is denoted by ${\textbf{{s}}^{(1)}}$ ( ${\textbf{{s}}^{(2)}}$ ). If ${{X_1}\leq _{\rm st}{Y_1}}$ and ${\textbf{{s}}^{(1)}\geq _{\rm hr}\textbf{{s}}^{(2)}}$ , then

\[(N_t\mid X_{k:n}\leq t<T_1)\geq_{\rm lr}(N_t^*\mid Y_{k:n}\leq t<T_2),\ \ \ t\geq 0.\]

Proof. See Appendix D. □

Double inspection

Consider again a coherent system with n i.i.d. components and suppose that the system starts working at time ${t=0}$ . We assume that the system is monitored by the operator at two time instances ${t_1}$ and ${t_2}$ (with ${t_1<t_2}$ ). This method of inspection is known in the literature as double monitoring. Some recent references in this regard are Zhang and Meeker [Reference Zhang and Meeker43], Parvardeh et al. [Reference Parvardeh, Balakrishnan and Arshadipour26], and Navarro and Calì [Reference Navarro and Calì22]. Suppose that the number of failed components up to time ${t_1}$ is k, and at time ${t_2}$ the system is still operating. Under these circumstances, we intend to study the probability of the number of failed components in the system at time ${t_2}$ . The probability mass function of this random variable, ${p^{t_1,t_2}_{k,n}(i)}$ , can be written as

(8) \begin{eqnarray} p^{t_1,t_2}_{k,n}(i)&\,=\,&\mathbb{P}(N_{t_2}=i\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)\nonumber\\[3pt]&\,=\,&\dfrac{\bar{S}_i\binom{n}{i}\binom{i}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{i-k}}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\binom{j}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{j-k}}, \ \ i=k,..., n-1.\end{eqnarray}

For the proof, see Appendix E. From (8), the mean number of failed components up to time ${t_2}$ can be represented as

(9) \begin{eqnarray}\varphi(t_1,t_2)&\,=\,&\mathbb{E}(N_{t_2}\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)\nonumber\\[3pt]&\,=\,&\dfrac{n\sum_{i=k}^{n-1}\bar{S}_i\binom{n-1}{i-1}\binom{i}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{i}}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\binom{j}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{j}}.\end{eqnarray}

The next theorem reveals some stochastic properties of ${(N_{t_2}\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)}$ in terms of k, ${t_1}$ , and ${t_2}$ .

Theorem 2.4. Assume that the common baseline CDF F is absolutely continuous. Then

  1. (a) for ${0\leq k\leq n-2}$ ,

    \[(N_{t_2}\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)\leq_{\mathrm{lr}}(N_{t_2}\mid X_{k+1:n}\leq t_1 <X_{k+2:n}, T>t_2);\]
  2. (b) for each ${0<t_1\leq t^{*}_1< t_2\leq t^{*}_2}$ ,

    \[(N_{t_2}\mid X_{k:n}\leq t^{*}_1<X_{k+1:n}, T>t_2)\leq_{\mathrm{lr}}(N_{t_2}\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t^{*}_2).\]

Proof. See Appendix F. □

As the likelihood ratio order is a subclass of the usual stochastic order, we get the following corollary from Theorem 2.4.

Corollary 2.2. ${\mathbb{E}(N_{t_2}\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)}$ is a decreasing function of ${t_1}$ and an increasing function of k and ${t_2}$ .

The next theorem provides a comparison between the failed components in two coherent systems in terms of likelihood ratio order.

Theorem 2.5. Let ${\mathcal{S}_{1}}$ and ${\mathcal{S}_{2}}$ denote two coherent systems with i.i.d. component lifetimes ${X_1,X_2,...,X_n}$ and ${Y_1,Y_2,...,Y_n}$ whose CDFs are F and G, respectively. Let ${N_t}$ and ${N_t^*}$ be the number of failed components of ${\mathcal{S}_{1}}$ and ${\mathcal{S}_{2}}$ , respectively, on [0, t]. Further, suppose the lifetime of the system ${\mathcal{S}_{1}}$ ( ${\mathcal{S}_{2}}$ ) is ${T_1}$ ( ${T_2}$ ) with signature vector ${\textbf{{s}}^{(1)}}$ ( ${\textbf{{s}}^{(2)}}$ ). If ${{X_1}\leq _{\rm hr}{Y_1}}$ and ${\textbf{{s}}^{(2)}\leq _{\rm hr}\textbf{{s}}^{(1)}}$ , then

\[(N_{t_2}\mid X_{k:n}\leq t_1<X_{k+1:n}, T_1>t_2)\geq_{\rm lr}(N_{t_2}^*\mid Y_{k:n}\leq t_1<Y_{k+1:n}, T_2>t_2),\ \ \ 0\leq t_1 < t_2.\]

Proof. See Appendix G. □

3. Optimal corrective and preventive maintenance models

In this section, we develop two maintenance strategies for n-component coherent systems based on the conditional probabilities introduced in the previous section. The following notation is used in our strategies:

  1. ${c_{cm}}$ : cost of CM for each component;

  2. ${c_{pm}}$ : cost of PM for each component;

  3. ${c_{cms}}$ : cost of CM for the whole system;

  4. ${c_{pms}}$ : cost of PM for the whole system;

  5. ${c^*_{pms}}$ : cost of rigid PM for the whole system;

  6. ${c_{min}}$ : cost of minimal repair for each component;

  7. ${w_{1}}$ : time to perform CM together with PM;

  8. ${w_{2}}$ : time to perform CM on the system;

  9. ${w_{3}}$ : time to perform PM on the system;

  10. ${w_{4}}$ : time to perform rigid PM on the system.

Strategy I

Assume that a coherent system begins to operate at time 0. A minimal repair has been performed on each component of the system that fails in the interval ${(0,\tau)}$ . Thus, we can assume that the system, consisting of n unfailed components with age ${\tau}$ , is alive at ${\tau}$ . Here, ${\tau}$ is a predetermined constant which may be considered, for example, as a guarantee time of the system. We assume that, after ${\tau}$ , a warning lamp is installed on the system that turns on at the time of the kth component failure, where k is predetermined (for a realistic example where warning lamps are employed in a system, we refer to Shimizu and Kawai [Reference Shimizu and Kawai35], where the authors consider a warning lamp in a vehicle’s electronic power steering system (EPS) which will operate in case of failure of components of the EPS; see also Khaledi and Shaked [Reference Khaledi and Shaked16]).

The operator decides to perform CM on the whole system at a cost of ${c_{cms}}$ once the system fails in the interval ${(\tau,t)}$ , or he/she decides to perform a PM action when the total operating time reaches t if the lamp turns on, whichever occurs first. More precisely, the operator decides to perform PM on all operating components together with CM on the failed ones at a cost of ${c_{pm}}$ for PM and a cost of ${c_{cm}}$ for CM, when the total operating time reaches t, provided that the warning light has lit up before or at time t. On the other hand, if the system is alive at t and the lamp does not turn on, he/she performs a perfect PM on the entire system with a cost of ${c_{pms}}$ . It should be mentioned here that, in the above situation, the operator knows only whether the light is on or off. In other words, the warning light might have been turned on at a time before t; the operator does not know the exact time of the kth failure in the interval ${(\tau,t)}$ . Two different cases of renewal cycle for Strategy ${\textbf{I}}$ are shown in Figure 3. The upper axis in Figure 3 shows the case where system failure has not occurred up to time t; that is, the system age reaches t. The lower axis depicts the case where the system fails before the age reaches t.

Figure 3: Maintenance Strategy ${\textbf{I}}$ .

It is evident that in the above policy, the inspection of the system is done in the interval ${(0,\tau)}$ , where the components of the system have been monitored continuously, and at the single time instant t (see Pham and Wang [Reference Pham and Wang27]). As mentioned in [Reference Pham and Wang27], a justification for the first part is that in the interval ${(0,\tau)}$ the components are ‘young’ and hence a minor repair is adequate. Thus, before ${\tau}$ , only minimal repairs, which require little time and have low cost, are carried out. After the system reaches age ${\tau}$ , as the cost of continuous monitoring may be substantial and minimal repair may not be reasonable because of the increased failure rate of the components, the operator does not monitor the system; instead, an alarm is installed which operates at the time of the kth failure. Using this information we obtain the optimal time of PM under the following settings on the cost function.

To evaluate the cost in interval ${(0,\tau)}$ , we shall utilize a cost function used by Sheu [Reference Sheu34]; see also Pham and Wang [Reference Pham and Wang27]. For each component, let the cost of the ith minimal repair depend on the deterministic part ${a_1(t,i)}$ and the age-dependent random part ${a_2(t)}$ . Note that ${a_1(t,i)}$ depends on the number i of minimal repairs performed on the component and the age t of that component. Two parts are now linked by a positive function h. In fact, the required cost of the ith minimal repair at age t for each component is ${h(a_1(t,i),a_2(t))}$ , where h is a continuous nondecreasing function of t, and is a nondecreasing function of i. Thus, the expected cost of minimal repairs for the whole system in a renewal period is

\[c^*_{min}=n \mathbb{E}\left[\sum_{i=1}^{N(\tau)}h(a_1(S_i,i),a_2(S_i))\right]\!,\]

where ${N(\tau)}$ denotes the total number of minimal repairs during the time interval ${(0,\tau)}$ , and ${S_1,S_2,...}$ are the successive failure times of each component at which minimal repairs have been performed. It is known that for the minimal repair the failure times follow a nonhomogeneous Poisson process with rate r(t) (see Barlow and Hunter [Reference Barlow and Hunter6] or Gertsbakh [Reference Gertsbakh14]). Note that r(t) is the failure rate of a component lifetime. It has been proved by Sheu [Reference Sheu34] that

(10) \begin{eqnarray}c^*_{min}=n\int_{0}^{\tau}\nu(y)r(y)dy,\end{eqnarray}

where ${\nu(y)=\mathbb{E}_{N(y)}\mathbb{E}_{a_2(y)}[h(a_1(y,N(y)+1),a_2(y))]}$ ; see also Pham and Wang [Reference Pham and Wang27]. A special case considered in the literature is that in which ${h(a_1(t,i),a_2(t))}$ is assumed to be a constant ${c_{min}}$ (see Barlow and Hunter [Reference Barlow and Hunter6] and Tahara and Nishida [Reference Tahara and Nishida37]). We assume that this cost includes the cost of monitoring and the cost of repair. Thus ${c^*_{min}}$ reduces to

\[c^*_{\min}=n c_{\min}H(\tau),\]

where ${H(\tau)=\int_{0}^{\tau}r(t)dt}$ is known as the mean value function of the failure process.

In this strategy, we consider t as a decision variable and ${\tau}$ as a predetermined time instant. By using the renewal reward theorem (see, e.g., Ross [Reference Ross30], p. 52), the average cost of system maintenance per unit time is then defined as the ratio of the average cost of the system maintenance per renewal cycle to the expected duration of a renewal cycle. In other words,

(11) \begin{eqnarray}\eta_I(t)&\,=\,&\frac{n \int_{0}^{\tau}\nu(y)r(y)dy+F_{\tau}(t-\tau)c_{cms}+c_{pms}P_{1,k}(\tau, t)}{\tau+\mathbb{E}(\!\min\!(t-\tau,T_\tau))}\nonumber\\[3pt]&&+\,\frac{P_{2,k}(\tau ,t)\left[(c_{cm}-c_{pm})\mathbb{E}(N_{t,\tau}\mid (X_{\tau})_{k:n}\leq t<T_{\tau})+nc_{pm}\right]}{\tau+\mathbb{E}(\!\min\!(t-\tau,T_\tau))},\end{eqnarray}

where ${F_{\tau}(\!\cdot\!)}$ denotes the CDF of the lifetime of an n-component system where each component’s age is ${\tau}$ , ${\mathbb{E}(N_{t,\tau}\mid (X_{\tau})_{k:n}\leq t<T_{\tau})}$ is the expectation of the number of failed components of the live system at time t with at least k failed components when all components are functioning at time ${\tau}$ (for ${\tau<t}$ ),

\[P_{1,k}(\tau, t)=\mathbb{P}(X_{k:n}>t, T>t\mid X_{1:n}>\tau),\]

and

\[P_{2,k}(\tau ,t)=\mathbb{P}(X_{k:n}\leq t<T\mid X_{1:n}>\tau).\]

Note that ${(X_{\tau})_{k:n}}$ is the kth order statistic from the CDF ${F(t|\tau)=1-\bar{F}(t)/\bar{F}(\tau)}$ , ${t>\tau}$ . We can easily show that

(12) \begin{eqnarray}\quad \ \ F_{\tau}(t-\tau)=1-\sum_{j=0}^{n-1}\bar{S}_{j}\binom{n}{j}\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{j}\left(\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{n-j},\end{eqnarray}
\begin{eqnarray*}P_{1,k}(\tau, t)=\sum_{i=0}^{k-1}\bar{S}_{i}\binom{n}{i}\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^i\left(\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{n-i},\end{eqnarray*}

and

\begin{eqnarray*}P_{2,k}(\tau, t)=\sum_{i=k}^{n-1}\bar{S}_{i}\binom{n}{i}\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^i\left(\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{n-i}.\end{eqnarray*}

Also, from Equation (5), we can obtain

\begin{eqnarray*}\mathbb{E}(N_t\mid X_{k:n}\leq t<T)=\frac{n\sum_{i=k}^{n-1}\bar{S}_{i}\binom{n-1}{i-1}\phi^i(t)}{\sum_{j=k}^{n-1}\bar{S}_{j}\binom{n}{j}\phi^j(t)}.\end{eqnarray*}

By replacing ${\phi(t)}$ with ${\left(\frac{\bar{F}(\tau)}{\bar{F}(t)}-1\right)}$ , we may obtain the corresponding formula for ${\mathbb{E}(N_{t,\tau}\mid (X_{\tau})_{k:n}\leq t<T_{\tau})}$ . Also, it can easily be shown that

\begin{eqnarray*}\mathbb{E}(\!\min (t-\tau, T_{\tau}))&\,=\,&\int_{0}^{t-\tau}[1-F_{\tau}(x)]dx\nonumber\\[3pt]&\,=\,&\frac{1}{\bar{F}^{n}(\tau)}\sum_{j=0}^{n-1}\bar{S}_{j}\binom{n}{j}\int_{0}^{t-\tau}(\bar{F}(\tau)-\bar{F}(x+\tau))^j(\bar{F}(x+\tau))^{n-j}dx.\end{eqnarray*}

In the following proposition, in the case that the decision variable in ${\eta_I(t)}$ is t, we verify the existence of the optimal value ${t^*}$ minimizing ${\eta_I(t)}$ .

Proposition 3.1. Let r(t) be the failure rate of components and ${\eta_I(t)}$ be as given in Equation (11). If ${\lim_{t\rightarrow \infty}r(t)}$ is finite and

(13) \begin{eqnarray} \lim_{t\rightarrow \infty}r(t)>\frac{c_{cms}+n\int_{0}^{\tau}\nu(y)r(y)dy}{(n-i^{*}+1)(c_{cms}-n c_{pm}-(i^{*}-1)(c_{cm}-c_{pm}))(\tau+\mu_{\tau})},\end{eqnarray}

then there exists a finite ${t^*}$ which satisfies ${\frac{d}{dt}\eta_I(t)\mid_{t=t^*}=0}$ and minimizes ${\eta_I(t)}$ , where ${\mu_{\tau}}$ is the expectation of the lifetime of an n-component system whose components have age ${\tau}$ , and ${i^{*}}$ is defined in Remark 2.2.

Proof. See Appendix H. □

Remark 3.1. Although, in Strategy I, we have considered t as a decision variable and ${\tau}$ as a predetermined time instant, one can instead consider ${\tau}$ as a decision variable (or even consider both t and ${\tau}$ as decision variables). In this case, the cost function (11) has to be minimized with respect to ${\tau}$ (or with respect to both ${\tau}$ and t simultaneously). In Example 3.1, we have numerically and graphically illustrated Strategy I for the bridge system in the cases that either t or both ${\tau}$ and t are considered as the decision variables.

It should be pointed out that, in the policy described above, we assume that all maintenance actions take negligible times. Now, suppose that minimal repair also takes negligible time, but PM combined with CM takes ${w_1}$ time units, CM on the whole system at time t takes ${w_2}$ time units, and PM on the whole system at time t takes ${w_3}$ time units. In the literature, a well-known criterion in the maintenance of systems is stationary availability. The stationary availability is defined as the ratio of the average time that the system is in a functioning state to the average length of a cycle. The stationary availability for Strategy ${\textbf{I}}$ is then given by

\begin{eqnarray*}A_I(t)=\frac{\tau+\mathbb{E}(\!\min\!(t-\tau,T_\tau))}{\tau+\mathbb{E}(\!\min\!(t-\tau,T_\tau))+w_1 P_{2,k}(\tau, t)+w_2 F_{\tau}(t-\tau)+w_3P_{1,k}(\tau, t)}.\end{eqnarray*}

Remark 3.2. We should mention here that the maintenance type depends not only on the action that is performed (replacement or repair of the failed component, a major overhaul of the system, and so on) but also on the complexity of the system structure. For example, the replacement of a failed component of a complex system does not generally improve the system’s performance, and hence can be considered a minimal repair. By contrast, if the system is not complex, then the same replacement may produce a noticeable improvement and therefore cannot be considered a minimal repair; see Pulcini [Reference Pulcini28].

Now let us see the following example.

Example 3.1. Assume that the bridge system in Figure 1 has component lifetimes which are independent Weibull random variables with CDF ${{F}(t)=1-\exp\{-t^{2}\}}$ , ${t\geq0}$ . It is known that the system signature is ${(0, 0.2, 0.6, 0.2, 0)}$ (see, e.g., Samaniego [Reference Samaniego32]). Let ${c_{min}=0.5}$ , ${c_{cms}=25}$ , ${c_{pms}=4}$ , ${c_{cm}=2}$ , and ${c_{pm}=1}$ . In Figure 4, the graphs of ${\eta_I(t)}$ are presented for different values of k and ${\tau}$ .

Figure 4: The expected cost of the system maintenance per unit time in Example 3.1: (a) ${\tau=0.1}$ , ${k=1,2,3}$ from top to bottom; (b) ${k=0}$ , ${\tau=0.1,0.3,0.5}$ from top to bottom.

In order to investigate the robustness of Strategy ${\textbf{I}}$ with respect to the model parameters ${c_{pms}}$ , ${c_{cms}}$ , ${{c}_{min}}$ , ${{c}_{cm}}$ , and ${{c}_{pm}}$ , we have provided some numerical results in Tables 1 and 2. As the tables show, when ${c_{cms}}$ increases, ${t^*}$ decreases and the operator should perform preventive action sooner. On the other hand, when ${c_{pms}}$ gets larger, ${t^*}$ increases too; i.e., the larger the cost of system PM, the later the time of performing system PM. When ${{c}_{min}}$ gets larger, the optimal time of PM gets larger too, and when ${{c}_{cm}}$ increases, as expected, ${t^*}$ decreases. Hence, according to this example, it is evident that the model is robust in terms of ${c_{cms}}$ , ${c_{pms}}$ , ${{c}_{min}}$ , and ${{c}_{cm}}$ . In the same manner one can see that the model is robust in terms of the other parameters ${{c}_{pm}}$ and ${\tau}$ .

Table 1: Optimal maintenance time for Strategy ${\textbf{I}}$ .

Table 2: Optimal maintenance time for Strategy ${\textbf{I}}$ .

The optimal times ${t^*}$ that minimize the average cost per unit of time and ${\eta_I(t^*)}$ are presented in Table 3 for several time instants ${\tau}$ . Note that if we consider ${\tau}$ and t as two decision variables, then the optimal value for the pair ${(\tau, t)}$ is ${(1.25, 1.25787)}$ , which results in the minimum maintenance cost ${6.30534}$ . The optimal values of ${(\tau,t)}$ are also tabulated in Table 4 for different values of ${c_{cms}}$ and ${c_{pms}}$ .

Table 3: Optimal maintenance times for Strategy ${\textbf{I}}$ with ${k=2}$ , ${c_{min}=0.5}$ , ${c_{cms}=25}$ , ${c_{pms}=4}$ , ${c_{cm}=2}$ , and ${c_{pm}=1}$ .

Table 4: Bivariate optimal maintenance times for Strategy ${\textbf{I}}$ .

Figure 5 depicts the three-dimensional plots of the cost function in terms of ${(\tau,t)}$ for ${k=0, 2}$ , ${c_{min}=0.5}$ , ${c_{cms}=25}$ , ${c_{pms}=4}$ , ${c_{cm}=2}$ , and ${c_{pm}=1}$ .

Figure 5: The three-dimensional plot of cost function in Example 3.1: (a) ${k=0}$ , (b) ${k=2}$ .

Figure 6 depicts the plots of ${A_I(t)}$ for ${w_{1}=0.08}$ , ${w_{2}=0.2}$ , and ${w_{3}=0.02}$ , and for several values of k and ${\tau}$ . As the plots show, the system availability first increases to attain its maximum and then decreases.

Figure 6: The stationary availability in Example 3.1: (a) ${\tau=0.1}$ , ${k=1,2,3}$ from top to bottom; (b) ${k=0}$ , ${\tau=0.1,0.3,0.5}$ from top to bottom.

Now, assume that ${h(a_1(t,i),a_2(t))=a_1(t)+a_2(t)}$ , where ${a_1(t)=3t}$ and ${a_2(t)}$ follows the normal distribution with mean 1. Then, by (10), ${c^*_{min}=n(2\tau^3+\tau^2)}$ . The expected maintenance cost ${\eta_I(t)}$ is presented in Figure 7 for several values of k and ${\tau}$ .

Figure 7: The expected cost of the system maintenance per unit time in Example 3.1 with ${h(a_1(t,i),a_2(t))=3t+a_2(t)}$ where ${a_2(t)}$ follows the normal distribution with mean 1: (a) ${\tau=0.1}$ , ${k=1,2,3}$ from top to bottom; (b) ${k=0}$ , ${\tau=0.1,0.3,0.5}$ from top to bottom.

The special case of this policy in which there is no minimal repair, i.e. ${\tau=0}$ , may be of interest. In this particular case, the average cost of the system maintenance per unit of time can be reduced to

\begin{eqnarray*}\eta_{I}(t)&\,=\,&\frac{c_{cms}F_T(t)+c_{pms}\mathbb{P}(T>t,X_{k:n}>t)}{\mathbb{E}(\!\min\!(t,T))}\\[4pt]&&+\, \frac{\mathbb{P}(X_{k:n}\leq t<T)[(c_{cm}-c_{pm})\mathbb{E}(N_t\mid X_{k:n}\leq t<T)+n c_{pm}]}{\mathbb{E}(\!\min\!(t,T))},\end{eqnarray*}

where

\begin{eqnarray*}\mathbb{P}(X_{k:n}\leq t<T)=\sum_{i=k}^{n-1}\bar{S}_i\binom{n}{i}F^i(t)\bar{F}^{n-i}(t),\end{eqnarray*}
\begin{eqnarray*}\mathbb{P}(T>t,X_{k:n}>t)=\sum_{i=0}^{k-1}\bar{S}_{i}\binom{n}{i}F^i(t)\bar{F}^{n-i}(t),\end{eqnarray*}

and ${\mathbb{E}(N_t\mid X_{k:n}\leq t<T)}$ is defined in (5).

Also, for this particular case where ${\tau=0}$ , the stationary availability may be written as

\begin{eqnarray*}A_{I}(t)=\frac{\mathbb{E}(\!\min\!(t,T))}{\mathbb{E}(\!\min\!(t,T))+w_1\mathbb{P}(X_{k:n}\leq t<T)+w_{2}F_T(t)+w_3 \mathbb{P}(T>t,X_{k:n}>t)}.\end{eqnarray*}

Now we consider the following example.

Example 3.2. Consider again the bridge system in Example 2.1 whose component lifetimes are independent Weibull random variables with CDF ${F(t)=1-\exp\{-t^{2}\}}$ , ${t\geq0}$ . Figure 8(a) depicts the plots of ${\eta_{I}(t)}$ for ${c_{cms}=20}$ , ${c_{pms}=4}$ , ${c_{cm}=2}$ , and ${c_{pm}=1}$ and the values ${k = 1,2,3}$ . Figure 8(b) depicts the plots of ${A_{I}(t)}$ for ${k=2}$ , ${w_{1}=0.08}$ , ${w_{2}=0.2}$ , and ${w_{3}=0.02}$ . As the plot shows, the system availability first increases to attain its maximum and then decreases. The availability attains its maximum at ${t^*=0.32645}$ and ${A_{I}(t^*)=0.921662}$ .

An analysis of the results of Table 5 indicates that the maintenance strategy ${\textbf{I}}$ with ${\tau=0}$ is robust. It is seen that when ${{c}_{cms}}$ increases, the optimal time of PM decreases, as expected. Also, when ${{c}_{pms}}$ gets larger, the optimal time of PM gets larger, too. One can easily see that, based on this example, the model is robust in terms of the costs ${{c}_{pm}}$ and ${{c}_{cm}}$ .

Table 5: Optimal maintenance time for Strategy ${\textbf{I}}$ with ${\tau=0}$ .

Figure 8: (a) The average maintenance cost per unit time in Example 3.2; (b) the stationary availability in Example 3.2.

We should mention here that the only difference between the maintenance schedules given in this example and those of Example 3.1 is that, in Example 3.1, the operator performs minimal repair in a time interval at the beginning of the system operation, while in the present example there is no minimal repair. By considering the same initial values for these two maintenance policies, one can make a comparison based on the expected cost or the availability criterion. For example, let ${c_{min}=0.5}$ , ${c_{cms}=25}$ , ${c_{pms}=4}$ , ${c_{cm}=2}$ , ${c_{pm}=1}$ , ${w_{1}=0.08}$ , ${w_{2}=0.2}$ , and ${w_{3}=0.02}$ . If ${k=1}$ and ${\tau=0.1}$ , the optimal values of ${\eta_I}$ and ${A_I}$ are 15.0488 and 0.8790, respectively, whereas if we do not perform minimal repair, the optimal values of ${\eta_I}$ and ${A_I}$ are 15.4200 and 0.8727, respectively. Therefore, based on these observations, the operator prefers to perform minimal repair at the starting point of the maintenance. One can show that the situation is reversed if ${k=3}$ , in which case no minimal repair is preferred.

Strategy II

Assume that a new coherent system begins to operate at time 0. Suppose that the system has been inspected at two times ${t_1}$ and ${t_2}$ , with ${t_1<t_2}$ . If the system fails before ${t_1}$ , then the operator performs CM on the entire system with a cost of ${c_{cms}}$ at the time of the system failure. He/she performs the same action if the system fails during the time interval ${(t_1,t_2)}$ . On the other hand, if the system is functioning at ${t_2}$ , the operator decides among three different actions:

  1. (a) If the number of components that have failed by ${t_1}$ , namely ${N_{t_1}}$ , is at most ${(k_1-1)}$ , the operator performs PM on the whole system with a cost of ${c_{pms}}$ .

  2. (b) If ${k_1\leq N_{t_1}\leq k_2}$ , then he/she decides to perform PM on all operating components of the system together with CM on all failed ones, at a cost of ${c_{pm}}$ for PM and a cost of ${c_{cm}}$ for CM.

  3. (c) If ${N_{t_1}}$ is at least ${(k_2+1)}$ , then the operator decides to perform a more rigid PM on the system (than in Case (a)) at a cost of ${c^*_{pms}}$ .

In this strategy, we assume that ${t_2}$ is the decision variable, while ${t_1}$ , ${k_1}$ , and ${k_2}$ are fixed constants. Three different cases of renewal cycle for Strategy ${\textbf{II}}$ are shown in Figure 9. The upper axis in Figure 9 depicts the case in which system failure has not occurred up to time ${t_2}$ ; that is, the age of the system reaches ${t_2}$ . The middle axis shows the case in which the system is alive at the inspection time ${t_1}$ and fails before attaining the age ${t_2}$ . The lower axis depicts the case where the system fails before its age reaches the time of inspection ${t_1}$ .

Figure 9: Maintenance Strategy ${\textbf{II}}$ .

The average cost of the system maintenance per unit of time is

(14) \begin{eqnarray}\eta_{II}(t_2)&\,=\,&\frac{D(t_2)}{\mathbb{E}(\!\min\!(t_2,T))},\end{eqnarray}

where

\begin{eqnarray*}D(t_2)&\,=\,&c_{cms}\mathbb{P}(T\leq t_2)+c_{pms}\mathbb{P}(T>t_2,N_{t_1}\leq k_1-1)\\[3pt]&&+\, [(c_{cm}-c_{pm})\mathbb{E}(N_{t_2}\mid k_1\leq N_{t_1}\leq k_2,T>t_2)+n c_{pm}]\\[3pt]& &\times\ \mathbb{P}(T>t_2,k_1\leq N_{t_1}\leq k_2)+c^*_{pms}\mathbb{P}(T>t_2,N_{t_1}\geq k_2+1),\end{eqnarray*}

and

\begin{eqnarray*}\mathbb{E}(\!\min\!(t_2,T))&\,=\,&\int_{0}^{t_2}\sum_{j=0}^{n-1}\bar{S}_j\binom{n}{j}F^j(t)\bar{F}^{n-j}(t) dt.\end{eqnarray*}

In the special case where ${k_1=k_2=k}$ , ${D(t_2)}$ may be reduced to

\begin{eqnarray*}D(t_2)&\,=\,&c_{cms}\mathbb{P}(T\leq t_2)+c_{pms}\mathbb{P}(T>t_2,N_{t_1}\leq k-1)\\[3pt]&&+\, [(c_{cm}-c_{pm})\varphi(t_1,t_2)+n c_{pm}]\mathbb{P}(T>t_2, N_{t_1}= k)+c^*_{pms}\mathbb{P}(T>t_2,N_{t_1}\geq k+1),\end{eqnarray*}

where, from (9),

\begin{eqnarray*}\varphi(t_1,t_2)=\frac{n\sum_{i=k}^{n-1}\bar{S}_i\binom{n-1}{i-1}\binom{i}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{i}}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\binom{j}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{j}}.\end{eqnarray*}

Also,

\begin{eqnarray*}&&\mathbb{P}(T>t_2,N_{t_1}\leq k-1)=\sum_{i=1}^{k}s_i\sum_{j=0}^{i-1}\binom{n}{j}F^j(t_2)\bar{F}^{n-j}(t_2)\\[3pt]&&\ \ \ \ \ \ +\sum_{i=k+1}^{n}s_i\sum_{m=n-i+1}^{n}\sum_{l=\max(m,n-k+1)}^{n}\binom{n}{l}\binom{ l}{m}F^{n-l}(t_1)(\bar{F}(t_1)-\bar{F}(t_2))^{l-m}\bar{F}^{m}(t_2)\end{eqnarray*}

and

\begin{eqnarray*}\mathbb{P}(T>t_2,N_{t_1}\geq k+1)&\,=\,&\sum_{i=k+2}^{n}s_i\sum_{j=k+1}^{i-1}\sum_{m=n-i+1}^{n-k-1}\binom{n}{m}\binom{ n-m}{j}F^{j}(t_1)\\[3pt]&\times&(\bar{F}(t_1)-\bar{F}(t_2))^{n-j-m}\bar{F}^{m}(t_2).\end{eqnarray*}

On the other hand,

\begin{eqnarray*}\mathbb{P}(T>t_2,N_{t_1}=k)&\,=\,&\sum_{i=k+1}^{n}s_i\sum_{j=k}^{i-1}\binom{n}{j}\binom{j}{k}F^k(t_1)\bar{F}^{n-j}(t_2)(\bar{F}(t_1)-\bar{F}(t_2))^{j-k}\\[3pt]&\,=\,&\sum_{j=k}^{n-1}\bar{S}_j \binom{n}{j}\binom{j}{k}F^k(t_1)\bar{F}^{n-j}(t_2)(\bar{F}(t_1)-\bar{F}(t_2))^{j-k}.\end{eqnarray*}

Also, we may obtain

\begin{eqnarray*}\mathbb{P}(T<t_2)=1-\sum_{j=0}^{n-1}\bar{S}_j\binom{n}{j}F^{j}(t_2)\bar{F}^{n-j}(t_2).\end{eqnarray*}

The aim here is to minimize ${\eta_{II}(t_2)}$ with respect to the decision variable ${t_2}$ ; that is, we should find the possible value ${t_2}$ , if it exists, such that

\[\eta_{II}(t_2^*)=\min_{t_2>t_1}\eta_{II}(t_2).\]

Now, let us assume that PM combined with CM takes ${w_1}$ time units, CM on the whole system takes ${w_2}$ time units, PM on the whole system takes ${w_3}$ time units, and the rigid PM on the system takes ${w_4}$ time units. The stationary availability for Strategy ${\textbf{II}}$ is given by

\begin{eqnarray*}A_{II}(t_2)=\frac{\mathbb{E}(\!\min\!(t_2,T))}{\mathbb{E}(\!\min\!(t_2,T))+B(t_2)},\nonumber\end{eqnarray*}

where

\begin{eqnarray*}B(t_2)&\,=\,&w_1 \mathbb{P}(T>t_2,k_1\leq N_{t_1}\leq k_2)+w_{2}\mathbb{P}(T\leq t_2)\\[3pt]&&+\,w_{3}\mathbb{P}(T>t_2,N_{t_1}\leq k_1-1)+w_{4}\mathbb{P}(T>t_2,N_{t_1}\geq k_2+1).\end{eqnarray*}

Remark 3.3. In Strategy ${\textbf{II}}$ , we assumed that ${t_2}$ is the only decision variable and ${t_1}$ is a fixed constant. However, in this strategy, one can instead assume that ${t_1}$ is the decision variable (or even that both ${t_1}$ and ${t_2}$ are decision variables) and minimize the cost function (14) based on that. The other point we should mention here is that of the conditions under which there exists an optimum value which minimizes the cost function (14). Since the functional form of ${\eta_{II}(t_2)}$ is rather complicated in the general setting, we can verify the existence of a possible optimum value for a given lifetime distribution of the components numerically (or graphically) using mathematical software. In the next example, we illustrate this in more detail.

Example 3.3. Let us look at again the bridge system in Example 2.1, where the component lifetimes are i.i.d. with Weibull distribution, with reliability function ${\bar{F}(t)=\exp\{-t^{2}\}}$ , ${t\geq0}$ . Figure 10(a) shows the plot of ${\eta_{II}(t_2)}$ for ${t_1=0.5}$ , ${c^*_{pms}=20}$ , ${c_{cms}=20}$ , ${c_{pms}=5}$ , ${c_{cm}=2}$ , and ${c_{pm}=1}$ , and different values ${k = 1,2,3}$ . Figure 10(b) depicts the plots of ${A_{II}(t)}$ for ${k=2}$ , ${t_1=0.5}$ , ${w_{1}=0.04}$ , ${w_{2}=0.05}$ , ${w_{3}=0.02}$ , and ${w_{4}=0.06}$ . It can be seen that the system availability first increases to arrives at its maximum and then decreases. The availability attains its maximum at ${t_2^*=0.719917}$ and ${A_{II}(t_2^*)=0.952746}$ .

Figure 10: (a) The average maintenance cost per unit time in Example 3.3; (b) the stationary availability for Strategy ${\textbf{II}}$ in Example 3.3.

An analysis of the results in Table 6 indicates that the maintenance strategy ${\textbf{II}}$ is robust. It is seen that when ${{c}_{cms}}$ increases (or ${c_{pm}}$ decreases), the optimal time of PM decreases, as expected. This robustness is true also in terms of ${{c}^*_{pms}}$ . That is, when ${{c}^*_{pms}}$ gets larger, the optimal time of PM gets larger, too. One can easily see that, based on this example, the model is robust in terms of the costs ${{c}_{pms}}$ and ${{c}_{cm}}$ .

Table 6: Optimal maintenance time for Strategy ${\textbf{II}}$ with ${t_1=0.5}$ , ${k=1}$ .

Table 7: Bivariate optimal maintenance times for Strategy ${\textbf{II}}$ .

Considering ${t_1}$ and ${t_2}$ as two decision variables, the three-dimensional plot of the cost function is depicted in Figure 11. The optimal values of ${(t_1,t_2)}$ are also tabulated in Table 7 for different values of ${c_{cms}}$ and ${c^*_{pms}}$ .

Figure 11: The three-dimensional plot of the cost function in Example 3.3.

Recalling Example 3.2, it is interesting to compare Strategies I and II for the bridge system. To do so, let ${\tau=0.1}$ , ${t_1=0.5}$ , ${c_{min}=0.5}$ , ${c_{cms}=25}$ , ${c_{pms}=4}$ , ${c_{cm}=2}$ , ${c_{pm}=1}$ , ${c^*_{pms}=20}$ , ${w_{1}=0.08}$ , ${w_{2}=0.2}$ , ${w_{3}=0.02}$ , and ${w_{4}=0.06}$ . If ${k=2}$ in Strategy ${\textbf{I}}$ and ${k_1=k_2=1}$ in Strategy ${\textbf{II}}$ , then the optimal values of ${\eta_I}$ and ${\eta_{II}}$ are 13.1557 and 20.6165, and the optimal values of ${A_I}$ and ${A_{II}}$ are 0.9250 and 0.8728, respectively. This means that based on both the expected maintenance cost and the stationary availability criteria, Strategy ${\textbf{I}}$ is preferred.

4. Conclusions

In reliability engineering, although an extensive number of research works have been devoted to optimal maintenance of one-unit systems, only a small portion of the literature has considered maintenance of multi-component systems. This paper aimed to propose some optimal strategies for the maintenance of a multi-component coherent system using some partial information on the number of destroyed components in the system. For this purpose, first, we proposed two criteria for evaluating the conditional probability function of the number of destroyed components in the system. In the computation of the proposed measures, we utilized some partial information on the status of the components of the system based on some inspection strategies. The derivations of both criteria rely on the notion of the signature associated with a coherent system. Using these criteria, we then introduced two different approaches to the optimal maintenance of the system. In the first approach, before a predetermined time ${\tau}$ , the system undergoes minimal repair, and after ${\tau}$ , the system is equipped with a warning light that turns on at the time of the kth component failure. Then the system is inspected at time t, where ${t>\tau}$ , and the operator decides to perform CM on the entire system once the system fails or to perform a PM action when the total operating time reaches t if the warning light turns on, whichever occurs soonest. In the second approach, the system is inspected at two times ${t_1}$ and ${t_2}$ , where ${t_1<t_2}$ , and depending on the information obtained at ${t_1}$ , the operator performs different maintenance actions at ${t_2}$ . In the proposed approaches, optimality criteria were defined based on the long-run expected costs of maintenance and availability of the system. The results of the paper were applied to the bridge system, for which several illustrative plots were presented. In this paper, we assumed that the component lifetimes were i.i.d. One may also consider other scenarios for component failure to propose optimal strategies for maintaining complex systems. The extension of the results of this paper to systems with dependent and/or non-identical components would be an interesting area for future research.

Appendix A. Proof of Equation (4)

The probability mass function of ${(N_t\mid X_{k:n}\leq t<T)}$ can be computed as

(15) \begin{eqnarray}p^t_{k,n}(i)&\,=\,&\mathbb{P}(N_t=i\mid X_{k:n}\leq t<T)\nonumber\\[5pt]&\,=\,&\frac{\sum_{m=i+1}^{n}\mathbb{P}(X_{i:n}\leq t<X_{i+1:n}, X_{k:n}\leq t<T, T=X_{m:n})}{\mathbb{P}(X_{k:n}\leq t<T)}\nonumber\\[5pt]&\,=\,&\frac{\sum_{m=i+1}^{n}\mathbb{P}(T=X_{m:n})\mathbb{P}(X_{i:n}\leq t<X_{i+1:n}, X_{k:n}\leq t<X_{m:n} )}{\mathbb{P}(X_{k:n}\leq t<T)}\nonumber\\[5pt]&\,=\,&\frac{\bar{S}_{i}\binom{n}{i}F^i(t)\bar{F}^{n-i}(t)}{\mathbb{P}(X_{k:n}\leq t<T)}, \qquad i=k,...,n-1,\end{eqnarray}

where the second equality follows from the law of total probability; the third and fourth equalities follow from the facts that the order statistics are independent from their ranks (see Kochar et al. [Reference Kochar, Mukerjee and Samaniego17]), and that ${[X_{i:n}\leq t<X_{i+1:n}]\subseteq[X_{k:n}\leq t<X_{m:n}]}$ for ${k\leq i<m}$ , respectively; and ${\bar{S}_i=\sum_{j=i+1}^{n}s_j}$ . On the other hand, using the law of total probability,

(16) \begin{eqnarray}\mathbb{P}(X_{k:n}\leq t<T)&\,=\,&\sum_{m=k+1}^{n}\mathbb{P}(X_{k:n}\leq t<T\mid T=X_{m:n})\mathbb{P}(T=X_{m:n})\nonumber\\[3pt]&\,=\,&\sum_{m=k+1}^{n}s_{m}\sum_{j=k}^{m-1}\binom{n}{j}F^j(t)\bar{F}^{n-j}(t)\nonumber\\[3pt]&\,=\,&\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}F^j(t)\bar{F}^{n-j}(t).\end{eqnarray}

Therefore

\begin{eqnarray*}p^t_{k,n}(i)=\frac{\bar{S}_{i}\binom{n}{i}\phi^i(t)}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\phi^j(t)}, \qquad i=k,...,n-1.\end{eqnarray*}

Appendix B. Proof of Theorem 2.1

To prove Part (a), first note that

\begin{eqnarray*}p^t_{k,n}(i)=\frac{\bar{S}_{i}\binom{n}{i}\phi^i(t)}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\phi^j(t)}I_{\{k,k+1,...,n-1\}}(i),\end{eqnarray*}

where ${I_{\{k,k+1,...,n-1\}}}$ denotes the indicator function on the set ${\{k,k+1,..., n-1\}}$ . It is easy to show that ${I_{\{k,k+1,...,n-1\}}(i)}$ is TP ${_2}$ in ${(i,k)\in\{k,...,n-1\}\times\{0,...,n-1\}}$ . This, in turn, implies that ${p^t_{k,n}(i)}$ is TP ${_{2}}$ in ${(i,k)\in\{k,...,n-1\}\times\{0,...,n-1\}}$ .

Part (b) follows from the result that was stated right before the theorem and the fact that ${\phi^i(t)}$ and hence ${p^t_{k,n}(i)}$ are TP ${_{2}}$ in ${(i,t)\in\{k,...,n-1\}\times[0,\infty)}$ .

Appendix C. Proof of Theorem 2.2

The probability mass function of ${(N_t\mid X_{k:n}\leq t<T_1)}$ is given in (4). Similarly, the probability mass function of ${(N_t^*\mid Y_{k:n+1}\leq t<T_2)}$ can be expressed as

\[q^t_{k,n+1}(i)=\frac{\bar{S}^{(2)}_{i}\binom{n+1}{i}\phi^i(t)}{\sum_{j=k}^{n}\bar{S}^{(2)}_{j}\binom{n+1}{j}\phi^j(t)}, \ \ \ i=k,...,n,\]

where ${\bar{S}^{(2)}_{i}=\sum_{j=i+1}^{n+1}p_j}$ . Therefore, the fraction ${q^t_{k,n+1}(i)/ p^t_{k,n}(i)}$ is proportional to

\[\frac{\bar{S}^{(2)}_{i}}{\bar{S}^{(1)}_{i}}\frac{\binom{n+1}{i}}{\binom{n}{i}},\]

which is increasing in ${i=0,1,...,n}$ . This completes the proof.

Appendix D. Proof of Theorem 2.3

Denote by ${p^{t,r}_{k,n}(i)}$ the conditional probability in (4) for the system ${\mathcal{S}_{r}}$ , ${r=1,2}$ ; that is

(17) \begin{eqnarray}p^{t,r}_{k,n}(i)&\,=\,&\frac{\bar{S}^{(r)}_{i}\binom{n}{i}\phi_r^i(t)}{\sum_{j=k}^{n-1}\bar{S}^{(r)}_{j}\binom{n}{j}\phi_r^j(t)}, \qquad i=k,...,n-1,\end{eqnarray}

where ${\phi_1(t)=F(t)/\bar{F}(t)}$ , ${\phi_2(t)=G(t)/\bar{G}(t)}$ , and ${\bar{S}^{(r)}_{i}=\sum_{j=i+1}^{n}s_j^{(r)}}$ , ${r=1,2}$ . It follows from the assumptions of the theorem that both ${\bar{S}^{(r)}_{i}}$ and ${\phi_r^i(t)}$ are RR ${_{2}}$ in ${(i,r)\in\{k,...,n-1\}\times\{1,2\}}$ . Since the product of two RR ${_2}$ functions is again an RR ${_2}$ function, we conclude that ${p^{t,r}_{k,n}(i)}$ is RR ${_{2}}$ in ${(i,r)\in\{k,...,n-1\}\times\{1,2\}}$ . This completes the proof.

Appendix E. Proof of Equation (8)

The probability mass function of ${(N_{t_2}\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)}$ can be computed as

(18) \begin{eqnarray}&&p^{t_1,t_2}_{k,n}(i)=\mathbb{P}(N_{t_2}=i\mid X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)\nonumber\\[3pt]&&\ \ = \frac{\sum_{m=i+1}^{n}\mathbb{P}(X_{i:n}\leq t_2<X_{i+1:n}, X_{k:n}\leq t_1<X_{k+1:n}, T>t_2,T=X_{m:n})}{\mathbb{P}( X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)}\nonumber\\[3pt]&&\ \ = \frac{\sum_{m=i+1}^{n}\mathbb{P}(T=X_{m:n})\mathbb{P}(X_{i:n}\leq t_2<X_{i+1:n}, X_{k:n}\leq t_1<X_{k+1:n}, X_{m:n}>t_2)}{\mathbb{P}( X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)}\nonumber\\[3pt]&&\ \ =\frac{\sum_{m=i+1}^{n}s_m\binom{n}{i}\binom{i}{k}F^{k}(t_1)(\bar{F}(t_1)-\bar{F}(t_2))^{i-k}\bar{F}^{n-i}(t_2)}{\mathbb{P}( X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)}.\end{eqnarray}

But

(19) \begin{eqnarray}&&\mathbb{P}( X_{k:n}\leq t_1<X_{k+1:n}, T>t_2)\nonumber\\[3pt]&&\ \ \ \ \ \ =\sum_{m=k+1}^{n}\mathbb{P}(X_{k:n}\leq t_1<X_{k+1:n}, T>t_2\mid T=X_{m:n})\mathbb{P}(T=X_{m:n})\nonumber\\[3pt]&&\ \ \ \ \ \ =\sum_{m=k+1}^{n}\mathbb{P}(X_{k:n}\leq t_1<X_{k+1:n}, X_{m:n}>t_2)\mathbb{P}(T=X_{m:n})\nonumber\\[3pt]&&\ \ \ \ \ \ =\sum_{m=k+1}^{n}s_m\sum_{j=k}^{m-1}\binom{n}{j}\binom{j}{k}F^{k}(t_1)(\bar{F}(t_1)-\bar{F}(t_2))^{j-k}\bar{F}^{n-j}(t_2)\nonumber\\[3pt]&&\ \ \ \ \ \ =\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\binom{j}{k}F^{k}(t_1)(\bar{F}(t_1)-\bar{F}(t_2))^{j-k}\bar{F}^{n-j}(t_2).\end{eqnarray}

Therefore

(20) \begin{eqnarray}p^{t_1,t_2}_{k,n}(i)=\frac{\bar{S}_i\binom{n}{i}\binom{i}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{i-k}}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\binom{j}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{j-k}}, \ \ i=k,..., n-1.\end{eqnarray}

Appendix F. Proof of Theorem 2.4

The probability mass function in (8) may be written as

\begin{eqnarray}p^{t_1,t_2}_{k,n}(i)=\frac{\bar{S}_i\binom{n}{i}\binom{i}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{i-k}}{\sum_{j=k}^{n-1}\bar{S}_j\binom{n}{j}\binom{j}{k}\left(\frac{\bar{F}(t_1)}{\bar{F}(t_2)}-1\right)^{j-k}}I_{\{k,..., n-1\}}(i)\nonumber.\end{eqnarray}

To prove Part (a), first note that ${\binom{i}{k}}$ and ${I_{\{k,..., n-1\}}(i)}$ are TP ${_{2}}$ in ${(i,k)\in\{k,...,n-1\}\times\{0,...,n-1\}}$ . Since the product of two TP ${_2}$ functions is a TP ${_2}$ function, we conclude that ${p_{k,n}^{t_1 ,t_2}(i)}$ is TP ${_{2}}$ in ${(i,k)\in\{k,...,n-1\}\times\{0,...,n-1\}}$ , which implies that ${p_{k+1,n}^{t_1 ,t_2}(i)/p_{k,n}^{t_1 ,t_2}(i)}$ is increasing in i. Hence the proof of Part (a) is complete.

To prove Part (b), it can be shown that ${p_{k ,n}^{t_1 ,t_2}(i)}$ is RR ${_{2}}$ in ${(i,t_1)\in\{k,...,n-1\}\times [0,t_2)}$ and TP ${_{2}}$ in ${(i, t_2)\in\{k,...,n-1\}\times ( t_1,\infty)}$ , and hence the result follows.

Appendix G. Proof of Theorem 2.5

Denote by ${p^{t_1,t_2,r}_{k,n}(i)}$ the conditional probability in (8) for the system ${\mathcal{S}_{r}}$ , ${r=1,2}$ ; that is,

(21) \begin{eqnarray}p^{t_1,t_2,r}_{k,n}(i)&\,=\,&\frac{\bar{S}^{(r)}_{i}\binom{n}{i}\binom{i}{k}(\psi_{r}(t_1, t_2)-1)^{i-k}}{\sum_{j=k}^{n-1}\bar{S}^{(r)}_{j}\binom{n}{j}\binom{j}{k}(\psi_{r}(t_1, t_2)-1)^{j-k}}, \qquad i=k,...,n-1,\end{eqnarray}

where ${\bar{S}^{(r)}_{i}=\sum_{j=i+1}^{n}s_j^{(r)}}$ , ${r=1,2}$ , and

(22) \begin{eqnarray}\psi_{r}(t_1, t_2)=\left\{ \begin{array}{ll} \frac{\bar{F}(t_1)}{\bar{F}(t_2)}, & \hbox{$r=1$;} \\ \\[-7pt] \frac{\bar{G}(t_1)}{\bar{G}(t_2)}, & \hbox{$r=2$.} \end{array} \right.\end{eqnarray}

By the assumption ${{\textbf{{s}}}^{(1)}\geq _{\rm hr}{\textbf{{s}}}^{(2)}}$ , ${\bar{S}^{(2)}_{i}/\bar{S}^{(1)}_{i}}$ is non-increasing in i, and hence ${\bar{S}^{(r)}_{i}}$ is RR ${_{2}}$ in ${(i,r)\in\{k,...,n-1\}\times\{1,2\}}$ . On the other hand, it can be concluded from ${{X_1}\leq _{\rm hr}{Y_1}}$ that ${(\psi_{r}(t_1, t_2)-1)^{i}}$ is also RR ${_{2}}$ in ${(i,r)\in\{k,...,n-1\}\times\{1,2\}}$ . From the fact that product of two RR ${_{2}}$ functions is an RR ${_{2}}$ function, we conclude that ${p^{t_1,t_2,r}_{k,n}(i)}$ is RR ${_{2}}$ in ${(i,r)\in\{k,...,n-1\}\times\{1,2\}}$ . This completes the proof of the theorem.

Appendix H. Proof of Proposition 3.1

Let ${f_\tau}$ denote the density function corresponding to ${\bar{F}_\tau}$ . On differentiating the cost function (11) with respect to t, we obtain

(23) \begin{eqnarray}\frac{d}{dt}\eta_I(t)&\stackrel{sgn}{=}&\left\{\left(c_{cms}\frac{f_{\tau}(t-\tau)}{\bar{F}_{\tau}(t-\tau)}+\frac{A_1(t)}{\bar{F}_{\tau}(t-\tau)}+\frac{A_2(t)}{\bar{F}_{\tau}(t-\tau)}+\frac{A_3(t)}{\bar{F}_{\tau}(t-\tau)}\right)\left(\tau+\int_{0}^{t-\tau}\bar{F}_{\tau}(x)dx\right)\right.\nonumber\\\nonumber\\[-7pt] &-&P_{2,k}(\tau ,t)\left[(c_{cm}-c_{pm})\mathbb{E}(N_{t,\tau}\mid (X_{\tau})_{k:n}\leq t<T_{\tau})+nc_{pm}\right]\nonumber\\ \nonumber\\[-7pt] &-&\left. n \int_{0}^{\tau}\nu(y)r(y)dy-F_{\tau}(t-\tau)c_{cms}-c_{pms}P_{1,k}(\tau, t)\right\},\end{eqnarray}

where ‘ ${\stackrel{sgn}{=}}$ ’ means to have the same sign,

\begin{eqnarray*}A_1(t)=c_{pms}\sum_{i=0}^{k-1}\bar{S}_i\binom{n}{i}\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{i-1}\left(\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{n-i-1}\frac{f(t)}{\bar{F}(\tau)}\left\{i\frac{\bar{F}(t)}{\bar{F}(\tau)}-(n-i)\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)\right\},\end{eqnarray*}

\begin{align*}A_2(t)=&n\left(c_{cm}-c_{pm}\right)\sum_{i=k}^{n-1}\bar{S}_i\binom{n-1}{i-1}\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{i-1}\left(\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{n-i-1}\frac{f(t)}{\bar{F}(\tau)}\\&\qquad \qquad\qquad\qquad \times\left\{i\frac{\bar{F}(t)}{\bar{F}(\tau)}-(n-i)\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)\right\},\end{align*}

and

\begin{eqnarray*}A_3(t)=nc_{pm}\sum_{i=k}^{n-1}\bar{S}_i\binom{n}{i}\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{i-1}\left(\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)^{n-i-1}\frac{f(t)}{\bar{F}(\tau)}\left\{i\frac{\bar{F}(t)}{\bar{F}(\tau)}-(n-i)\left(1-\frac{\bar{F}(t)}{\bar{F}(\tau)}\right)\right\}.\end{eqnarray*}

By using Equation (12), it can be observed that

\[\lim_{t\rightarrow\infty}\frac{f_\tau(t-\tau)}{\bar{F}_\tau(t-\tau)}=\lim_{t\rightarrow\infty}r(t),\]
\[\lim_{t\rightarrow\infty}\frac{A_1(t)}{\bar{F}_\tau(t-\tau)}=0,\]
\[\lim_{t\rightarrow\infty}\frac{A_2(t)}{\bar{F}_\tau(t-\tau)} = -(i^{*}-1) (n-i^{*}+1) (c_{cm}-c_{pm})\lim_{t\rightarrow\infty}r(t),\]
\[\lim_{t\rightarrow\infty}\frac{A_3(t)}{\bar{F}_\tau(t-\tau)} = -n(n-i^{*}+1) c_{pm} \lim_{t\rightarrow\infty}r(t).\]

One can show that if

\[\lim_{t\rightarrow \infty}r(t)>\frac{c_{cms}+n\int_{0}^{\tau}\nu(y)r(y)dy}{(c_{cms}-nc_{pm}-(i^*-1)(c_{cm}-c_{pm}))(n-i^*+1)(\tau+\mu_\tau)},\]

then the right-hand side of Equation (23) is positive. This means that, under the condition (13), ${\eta_I(t)}$ is eventually strictly increasing. On the other hand, one can obtain

\[\lim_{t\rightarrow\tau}\frac{d}{dt}\eta_I(t)=-\frac{n \int_{0}^{\tau}\nu(y)r(y)dy+c_{pms}}{\tau^2}.\]

Therefore, ${\eta_I(t)}$ is initially decreasing. We conclude that ${\eta_I(t)}$ has at least a finite minimum.

Acknowledgements

We would like to express our sincere thanks to the associate editor and two anonymous referees for their constructive comments and suggestions which improved the presentation of the paper. M. Asadi’s research work was performed at IPM-Isfahan and was in part supported by a grant from IPM (No. 98620215).

References

Arriaza, A., Navarro, J. and Suárez-Llorens, A. (2018). Stochastic comparisons of replacement policies in coherent systems under minimal repair. Naval Res. Logistics 65, 550565.CrossRefGoogle Scholar
Asadi, M. and Berred, A. (2012). On the number of failed components in a coherent operating system. Statist. Prob. Lett. 82, 21562163.CrossRefGoogle Scholar
Ashrafi, S. and Asadi, M. (2017). The failure probability of components in three-state networks with applications to age replacement policy. J. Appl. Prob. 54, 10511070.CrossRefGoogle Scholar
Bada, F. G., Berrade, M. D., Cha, J. H. and Lee, H. (2018). Optimal replacement policy under a general failure and repair model: minimal versus worse than old repair. Reliab. Eng. System Safety 180, 362372.CrossRefGoogle Scholar
Balakrishnan, N., Navarro, J. and Samaniego, F. J. (2012). Signature representation and preservation results for engineered systems and applications to statistical inference. In Recent Advances in System Reliability: Signatures, Multi-state Systems and Statistical Inference, eds. A. Lisnianski and I. Frenkel, Springer, London, pp. 122.CrossRefGoogle Scholar
Barlow, R. E. and Hunter, L. C. (1960). Optimum preventive maintenance policies. Operat. Res. 8, 90100.CrossRefGoogle Scholar
Barlow, R. E. and Proschan, F. (1975). Statistical Theory of Reliability and Life Testing. Holt, Rinehart and Winston, New York.Google Scholar
Belzunce, F., Martnez-Riquelme, C. and Ruiz, J. M. (2018). Allocation of a relevation in redundancy problems. To appear in Appl. Stoch. Models Business Industry.Google Scholar
Cha, J. H., Finkelstein, M. and Levitin, G. (2017). On preventive maintenence of systems with lifetimes dependent on a random shock process. Reliab. Eng. System Safety 168, 9097.CrossRefGoogle Scholar
Eryilmaz, S. (2012). The number of failed components in a coherent system with exchangeable components. IEEE Trans. Reliab. 61, 203207.CrossRefGoogle Scholar
Eryilmaz, S. (2018). The number of failed components in a k-out-of-n system consisting of multiple types of components. Reliab. Eng. System Safety 175, 246250.CrossRefGoogle Scholar
Finkelstein, M. and Gertsbakh, I. (2015). ‘Time-free’ preventive maintenance of systems with structures described by signatures. Appl. Stoch. Models Business Industry 31, 836845.CrossRefGoogle Scholar
Finkelstein, M. and Gertsbakh, I. (2016). On preventive maintenance of systems subject to shocks. P. I. Mech. Eng. O-J. Ris. 230, 220227.Google Scholar
Gertsbakh, I. (2000). Reliability Theory with Applications to Preventive Maintenance. Springer, London.Google Scholar
Karlin, S. (1968). Total Positivity, Vol. I. Stanford University Press.Google Scholar
Khaledi, B-E. and Shaked, M. (2007). Ordering conditional lifetimes of coherent systems. J. Statist. Planning Infer. 137, 11731184.CrossRefGoogle Scholar
Kochar, S. C., Mukerjee, H., and Samaniego, F. J. (1999). The ‘signature’ of a coherent system and its application to comparisons among systems. Naval Res. Logistics 46, 507523.3.0.CO;2-D>CrossRefGoogle Scholar
Levitin, G., Finkelstein, M. and Dai, Y. (2018). Optimizing availability of heterogeneous standby systems exposed to shocks. Reliab. Eng. System Safety 170, 137145.CrossRefGoogle Scholar
Lucia, U. (2013). Thermodynamic paths and stochastic order in open systems. Physica A 392, 39123919.CrossRefGoogle Scholar
Mahmoudi, M. and Asadi, M. (2011). The dynamic signature of coherent systems. IEEE Trans. Reliab. 60, 817822.CrossRefGoogle Scholar
Morse, P. M. (1958). Queues, Inventories, and Maintenance. Wiley, New York.CrossRefGoogle Scholar
Navarro, J. and Calì, C. (2018). Inactivity times of coherent systems with dependent components under periodical inspection. To appear in Appl. Stoch. Models Business Industry.Google Scholar
Navarro, J. and Rychlik, T. (2007). Reliability and expectation bounds for coherent systems with exchangeable components. J. Multivariate Anal. 98, 102113.CrossRefGoogle Scholar
Navarro, J., Samaniego, F. J. and Balakrishnan, N. (2011). Signature-based representations for the reliability of systems with heterogeneous components. J. Appl. Prob. 48, 856867.CrossRefGoogle Scholar
Navarro, J., Samaniego, F. J. and Balakrishnan, N. (2013). Mixture representations for the joint distribution of lifetimes of two coherent systems with shared components. Adv. Appl. Prob. 45, 10111027.CrossRefGoogle Scholar
Parvardeh, A., Balakrishnan, N. and Arshadipour, A. (2018). A note on the conditional residual lifetime of a coherent system under double monitoring. Commun. Statist. Theory Meth. 47, 23732378.CrossRefGoogle Scholar
Pham, H. and Wang, H. (2000). Optimal ${(\tau,T)}$ opportunistic maintenance of a k-out-of- ${n\,:\,G}$ system with imperfect PM and partial failure. Naval Res. Logistics 47, 223239.3.0.CO;2-A>CrossRefGoogle Scholar
Pulcini, G. (2003). Mechanical reliability and maintenance models. In Handbook of reliability engineering, ed. H. Pham, Springer, London, pp. 317348.CrossRefGoogle Scholar
Reed, S. (2017). An efficient algorithm for exact computation of system and survival signatures using binary decision diagrams. Reliab. Eng. System Safety 165, 257267.CrossRefGoogle Scholar
Ross, S. M. (1970). Applied Probability Models with Optimisation Applications. Holden-Day, San Francisco.Google Scholar
Samaniego, F. J. (1985). On closure of the IFR class under formation of coherent systems. IEEE Trans. Reliab. 34, 6972.CrossRefGoogle Scholar
Samaniego, F. J. (2007). System Signatures and their Applications in Engineering Reliability. Springer, New York.CrossRefGoogle Scholar
Shaked, M. and Shanthikumar, J. G (2007). Stochastic Orders. Springer, New York.CrossRefGoogle Scholar
Sheu, S. (1991). Generalized block replacement policy with minimal repair and general random repair costs for a multi-unit system. J. Operat. Res. Soc. 42, 331341.CrossRefGoogle Scholar
Shimizu, Y. and Kawai, T. (1991). Development of electric power steering. SAE Technical Paper 910014. Available at https://doi.org/10.4271/910014. CrossRefGoogle Scholar
Spizzichino, F. (2018). Reliability, signature, and relative quality functions of systems under time-homogeneous load-sharing models. Appl. Stoch. Models Business Industry 35, 158176.CrossRefGoogle Scholar
Tahara, A. and Nishida, T. (1975). Optimal replacement policy for minimal repair model. J. Operat. Res. Soc. Japan 18, 113124.Google Scholar
Tavangar, M. (2016). Conditional inactivity time of components in a coherent operating system. IEEE Trans. Reliab. 65, 359369.CrossRefGoogle Scholar
Wang, H. and Pham, H. (2006). Reliability and Optimal Maintenence. Springer, London.Google Scholar
Zarezadeh, S. and Asadi, M. (2019). Coherent systems subject to multiple shocks with applications to preventative maintenance. To appear in Reliab. Eng. System Safety.CrossRefGoogle Scholar
Zhang, Z. (2009). Ordering conditional general coherent systems with exchangeable components. J. Statist. Planning Infer. 140, 454460.CrossRefGoogle Scholar
Zhang, N., Fouladirad, M. and Barros, A. (2017). Maintenance analysis of a two-component load-sharing system. Reliab. Eng. System Safety 177, 2434.CrossRefGoogle Scholar
Zhang, Z. and Meeker, W. Q. (2013). Mixture representations of reliability in coherent systems and preservation results under double monitoring. Commun. Statist. Theory Meth. 42, 385397.CrossRefGoogle Scholar
Figure 0

Figure 1: The bridge system.

Figure 1

Figure 2: (a) The plots of ${p^t_{k,n}(i)}$ for ${i=1, 2,3}$ and ${k=1}$ in Example 2.1. (b) The plots of ${H_{k,n}^t}$ for ${k=0, 1,2}$ in Example 2.1.

Figure 2

Figure 3: Maintenance Strategy ${\textbf{I}}$.

Figure 3

Figure 4: The expected cost of the system maintenance per unit time in Example 3.1: (a) ${\tau=0.1}$, ${k=1,2,3}$ from top to bottom; (b) ${k=0}$, ${\tau=0.1,0.3,0.5}$ from top to bottom.

Figure 4

Table 1: Optimal maintenance time for Strategy ${\textbf{I}}$.

Figure 5

Table 2: Optimal maintenance time for Strategy ${\textbf{I}}$.

Figure 6

Table 3: Optimal maintenance times for Strategy ${\textbf{I}}$ with ${k=2}$, ${c_{min}=0.5}$, ${c_{cms}=25}$, ${c_{pms}=4}$, ${c_{cm}=2}$, and ${c_{pm}=1}$.

Figure 7

Table 4: Bivariate optimal maintenance times for Strategy ${\textbf{I}}$.

Figure 8

Figure 5: The three-dimensional plot of cost function in Example 3.1: (a) ${k=0}$, (b) ${k=2}$.

Figure 9

Figure 6: The stationary availability in Example 3.1: (a) ${\tau=0.1}$, ${k=1,2,3}$ from top to bottom; (b) ${k=0}$, ${\tau=0.1,0.3,0.5}$ from top to bottom.

Figure 10

Figure 7: The expected cost of the system maintenance per unit time in Example 3.1 with ${h(a_1(t,i),a_2(t))=3t+a_2(t)}$ where ${a_2(t)}$ follows the normal distribution with mean 1: (a) ${\tau=0.1}$, ${k=1,2,3}$ from top to bottom; (b) ${k=0}$, ${\tau=0.1,0.3,0.5}$ from top to bottom.

Figure 11

Table 5: Optimal maintenance time for Strategy ${\textbf{I}}$ with ${\tau=0}$.

Figure 12

Figure 8: (a) The average maintenance cost per unit time in Example 3.2; (b) the stationary availability in Example 3.2.

Figure 13

Figure 9: Maintenance Strategy ${\textbf{II}}$.

Figure 14

Figure 10: (a) The average maintenance cost per unit time in Example 3.3; (b) the stationary availability for Strategy ${\textbf{II}}$ in Example 3.3.

Figure 15

Table 6: Optimal maintenance time for Strategy ${\textbf{II}}$ with ${t_1=0.5}$, ${k=1}$.

Figure 16

Table 7: Bivariate optimal maintenance times for Strategy ${\textbf{II}}$.

Figure 17

Figure 11: The three-dimensional plot of the cost function in Example 3.3.