Hostname: page-component-745bb68f8f-f46jp Total loading time: 0 Render date: 2025-02-11T07:26:14.621Z Has data issue: false hasContentIssue false

A robust system reliability analysis using partitioning and parallel processing of Markov chain

Published online by Cambridge University Press:  30 September 2014

Po Ting Lin
Affiliation:
Department of Mechanical Engineering, Research and Development Center for Microsystem Reliability, Center for Biomedical Technology, Chung Yuan Christian University, Chungli City, Taiwan
Yu-Cheng Chou*
Affiliation:
Institute of Undersea Technology, National Sun Yat-sen University, Kaohsiung City, Taiwan
Yung Ting
Affiliation:
Department of Mechanical Engineering, Chung Yuan Christian University, Chungli City, Taiwan
Shian-Shing Shyu
Affiliation:
Institute of Nuclear Energy Research, Atomic Energy Council, Chungli City, Taiwan
Chang-Kuo Chen
Affiliation:
Institute of Nuclear Energy Research, Atomic Energy Council, Chungli City, Taiwan
*
Reprint requests to: Yu-Cheng Chou, Institute of Undersea Technology, National Sun Yat-sen University, 70 Lienhai Road, Kaohsiung City 80424, Taiwan. E-mail: ycchou@mail.nsysu.edu.tw
Rights & Permissions [Opens in a new window]

Abstract

This paper presents a robust reliability analysis method for systems of multimodular redundant (MMR) controllers using the method of partitioning and parallel processing of a Markov chain (PPMC). A Markov chain is formulated to represent the N distinct states of the MMR controllers. Such a Markov chain has N2 directed edges, and each edge corresponds to a transition probability between a pair of start and end states. Because N can be easily increased substantially, the system reliability analysis may require large computational resources, such as the central processing unit usage and memory occupation. By the PPMC, a Markov chain's transition probability matrix can be partitioned and reordered, such that the system reliability can be evaluated through only the diagonal submatrices of the transition probability matrix. In addition, calculations regarding the submatrices are independent of each other and thus can be conducted in parallel to assure the efficiency. The simulation results show that, compared with the sequential method applied to an intact Markov chain, the proposed PPMC can improve the performance and produce allowable accuracy for the reliability analysis on large-scale systems of MMR controllers.

Type
Special Issue Articles
Copyright
Copyright © Cambridge University Press 2014 

1. INTRODUCTION

For the purpose of reliability evaluation in system designs, different representative methods have been developed over the years, including the fault tree, reliability block diagram, reliability graph, event sequence diagram (Mutha et al., Reference Mutha, Jensen, Tumer and Smidts2013), Bayesian network, and Markov chain. Some literature has indicated that, mainly due to its flexibility, the Markov chain is the most suitable technique to analyze the reliability of redundant systems (Liu & Rausand, Reference Liu and Rausand2011).

Two configurations of subsea blowout preventer control systems, including triple modular redundancy and double dual modular redundancy control systems, were evaluated in terms of the probability of failure on demand, availability, and reliability using the Markov method (Cai et al., Reference Cai, Liu, Liu, Tian, Li and Ren2012). A Markov chain based methodology was proposed to perform the reliability, availability, maintainability, and safety for a hybrid redundancy system modeled as a stochastic Petri net (Liu et al., Reference Liu, Liu, Cai, Liu, Li, Tian and Ji2013). A Markov modeling approach was applied to the reliability analysis for a patented modular converter system that has multiple identical and interchangeable power converter modules in a wind turbine (Zhang et al., Reference Zhang, Zhang, Chen and Jin2013). The comparative reliability study concerning fault coverage on the all-voting triple modular redundancy system, dual-duplex system, and double 2-out-of-2 system was conducted using the discrete-time Markov chain (Kim et al., Reference Kim, Lee and Lee2005; Wang et al., Reference Wang, Ji, Dong and Yang2007). The availability, production rate, and reliability of multistate degraded systems subjected to minimal repairs and imperfect preventive maintenance were evaluated via the continuous-time Markov chain (Soro et al., Reference Soro, Nourelfath and Ait-Kadi2010). The reliability and benefits for a programmable logic controller hot standby system, which has the manager–slave architecture and two types of repair mechanisms, were evaluated using a semi-Markov approach (Parashar & Taneja, Reference Parashar and Taneja2007). A scalable technique based on the discrete-time Markov chain was developed to evaluate the reliability for both nonredundant and structural redundancy-based large circuit designs (Bhaduri et al., Reference Bhaduri, Shukla, Graham and Gokhale2007). A redundant system, designed with a special failure dependency (i.e., the redundant dependency), was evaluated with respect to the system reliability through a continuous-time Markov model (Yu et al., Reference Yu, Chu, Châtelet and Yalaoui2007). Markov chain models and bifurcation analysis were used to compare the degree of redundancy and system reliability for several logic redundancy schemes, including von Neumann's multiplexing logic, N-tuple modular redundancy, and interwoven redundant logic (Han et al., Reference Han, Gao, Jonker, Qi and Fortes2005). In addition to redundant systems, different Markov models, including continuous-time/discrete-time and full/semi/quasi Markov chains, have also been applied to study issues on the reliability, safety, availability, and diagnosis of different other systems (Zhang et al., Reference Zhang, Long and Sato2003; Dominguez-Garcia et al., Reference Dominguez-Garcia, Kassakian and Schindall2006; Guo & Yang, Reference Guo and Yang2008; Liu et al., Reference Liu, Ni, Liu, Song and Wang2011; Lisnianski et al., Reference Lisnianski, Elmakias, Laredo and Ben Haim2012; Matthews & Philip, Reference Matthews and Philip2012). However, no literature has been reported with respect to partitioning and reordering a Markov chain to enable seamless incorporation of parallel computing techniques.

This paper presents a robust reliability analysis method, called the partitioning and parallel processing of a Markov chain (PPMC), for systems of multimodular redundant (MMR) controllers. A Markov chain (MC) is formulated to represent the N distinct states of the MMR controllers. Such an MC graph has N 2 directed edges. Each edge weight represents the transition probability from the start state to the end state. A system of β modular controllers is shown in Figure 1. Each of the β controllers processes signals from α sensors. Each of the α sensors has γ states. Thus, the number of distinct states for this system of MMR controllers is calculated as N = αβγ. Because N can be drastically increased by any of the three factors, the system reliability evaluation, which is based on the N 2 edge weights, may require large computational resources, such as the central processing unit (CPU) usage and memory occupation.

Fig. 1. N-state Markov chain for multimodular redundant controllers.

To this end, techniques of partitioning and reordering an MC have been conducted to reduce the complexity of an MC graph. Lightly weighted edges are suspended based on the threshold parameter, and the MC graph is partitioned and reordered to produce multiple weakly connected subgraphs, which can be processed independently in parallel. Therefore, parallel computing techniques can be applied to improve or ensure the calculation efficiency. In this paper, the task-farming paradigm is adopted for parallel processing. The task-farming paradigm has a manager–workers pattern. The manager is responsible for decomposing a task, distributing subtasks among workers, and gathering partial results from workers to coordinate the final calculation for the reliability evaluation. The worker processes execute in a simple cycle: get a subtask, process the subtask, and send the result to the manager.

The efficiency and accuracy of reliability analysis depend on the suspension threshold, the partition level, and the availability of parallel processors. The worst-case reliability is evaluated to compensate for inaccuracy due to the suspended edge weights. Higher suspension thresholds produce simpler MC models and faster calculations but more conservative reliabilities. The performance of PPMC is maximized to find the optimal suspension and partition levels. The proposed methodology has been successfully integrated with the detection and diagnosis processes in the fault-tolerant system. The simulation results show that, compared to the sequential method applied to an intact MC, the proposed PPMC can improve the performance and produce allowable accuracy for the reliability analysis on large-scale systems of MMR controllers.

2. RELIABILITY ANALYSIS OF MMR CONTROLLERS

Suppose the MMR controllers have N distinct states S = {1, 2, … , N} and the transitions between each state are represented in a MC. Figure 2 illustrates the graph G(S, E) of the MC where the edge set E is composed of N 2 directed edges: {E 1,1, E 1,2, … , E i,j, … , E N,N}. Each directed edge has two edge weights μi,j and Q i,j. The prior one is the transition probability that the state i moves to the state j while the later one represents the rate of change of the transition probability. The total transition probability that the state i moves to every state, including itself, equals 1, namely,

(1)$$\sum\limits_{\,j = 1}^N {{\rm \mu} _{i\comma j} = 1}\quad \forall i.$$

Therefore, the total rate of change of probability is zero, as in

(2)$$\sum\limits_{\,j = 1}^N {Q_{i\comma j} = 0} \quad \forall i.$$

Given a vector of the initial state probabilities,

$${{\bf p}\lpar 0\rpar = 1{\bf e}_1} + \sum\limits_{i = 2}^N {0{{\bf e}_i}\comma \; }$$

at the time t = 0, the vector of the state probabilities,

$${{\bf p}\lpar t\rpar = \sum\limits_{i = 1}^N {\,p_i}\lpar t\rpar {{\bf e}_i}\comma \; }$$

is desirable for the evaluation of the system reliability.

Fig. 2. N-state Markov chain for multimodular redundant controllers.

For most MCs, the time-dependent state probabilities can be evaluated using the described method. However, the computational complexity increases dramatically when the number of states increases in the MMR controllers. It is very difficult to solve the large-scale eigenvalue problem and obtain the state probabilities in allowable working time. To this end, we will focus on the discrete-time approach, which requires less calculation.

2.1. Discrete-time MC

With the Markov property, the state variable X(n + 1) at the time n + 1 depends only upon its state at the time n, X(n) That is, given the present state of the system X(n), the future state of the discrete-time MC (DTMC) X(n + 1) is independent of its previous discrete-time states X(0), X(1), … , X(n − 1). A DTMC with the finite state space S is time homogeneous if Equation (3) holds.

(3)$${{\rm \mu}_{i\comma j}} \equiv P\lsqb X\lpar n + 1\rpar = j\vert X\lpar n\rpar = i\rsqb = P\lsqb X\lpar 1\rpar = j\vert X\lpar 0\rpar = i\rsqb$$

for n ≥ 0 and i, j = 1 …  N. For the DTMC, μi,j is called the one-step transition probability at the time n. In this paper, we assume the one-step transition probability remains constant for all times n. The state probabilities will be calculated using the given one-step transition probabilities.

We are interested in the probabilities

$$P\left[X\lpar n\rpar = j \right]\quad \forall j \in S \quad \hbox{and} \quad n \ge 0.$$

Statistically, the probability

$$P\left[X\lpar n\rpar = j \right]$$

is calculated by

(4)$$P\left[{X\lpar n\rpar = j} \right]= \sum\limits_{i = 1}^N {P\left[{X\lpar 0\rpar = i} \right]P\left[{X\lpar n\rpar = j\vert X\lpar 0\rpar = i} \right].}$$

Equation (4) is further rewritten in the matrix form as in

(5)$${\bf p}\lpar n\rpar = {\bf A}\lpar n\rpar \times {\bf p}\lpar 0\rpar \comma \;$$

where

$${\bf p}\lpar n\rpar = \sum\limits_{\,j = 1}^N {P\lsqb X\lpar n\rpar = j\rsqb {{\bf e}_j}}$$

indicates the state probabilities at n;

$${\bf p}\lpar 0\rpar = \sum\limits_{i = 1}^N {P\lsqb X\lpar 0\rpar = i\rsqb {{\bf e}_i}}$$

is for the initial probabilities; and

$${\bf A}\lpar n\rpar = \sum\limits_{i = 1}^N {\sum\limits_{\,j = 1}^N {P\left[{X\lpar n\rpar = j\vert X\lpar 0\rpar = i} \right]{{\bf e}_j}{{\bf e}_i}} }$$

is the matrix of the n-step transition probabilities. Because the DTMC is assumed to be time homogeneous, the matrix A(n) can be calculated by the nth power of A (Stewart, Reference Stewart2009), as in

(6)$${\bf p}\lpar n\rpar = {\bf A}^{\prime\prime} \times {\bf p}\lpar 0\rpar.$$

To sum up, the transition probabilities of the DTMC are calculated by the matrix operations in Equation (6). The matrix multiplication algorithm has message-passing characteristics and is more applicable in message-passing systems, such as distributed memory based cluster of computers. In Section 3, the MC is partitioned into smaller systems in order to distribute the loads to several computers. In Section 4, the parallel processing of the MC is introduced.

3. PARTITIONING OF A MC

This section presents the partitioning of a large-scale MC. The lightly weighted edges are suspended based on the threshold parameters, and the MC is partitioned into multiple weakly connected subgraphs using the sparse matrix partitioning. The created subgraphs can be processed individually to produce partial results, which are combined to form the complete state probability distribution at a desired time instant.

3.1. Reduction of a MC

In this paper, we want to first reduce the complexity of the MC in terms of eliminating the connectivity of the weakly linked edges. Chou and Lin (Reference Chou and Lin2014) have eliminated the edges of minimal probabilities in the MC and greatly reduced the required calculations. The evaluated network probability of the reduced MC (RMC) was slightly underestimated and can be utilized as a worst-case probability, that is, the true probability of a MC is at least higher than or equal to the worst-case probability in the RMC.

A binary dependency matrix D represents the connectivity of the MC, as in

(7)$${\bf D} = \sum\limits_{i = 1}^N {\sum\limits_{\,j = 1}^N {{D_{i\comma j}}{{\bf e}_j}{{\bf e}_i}}\comma \; }$$

where D i,j = 1 when the ith state is directed to the jth one; otherwise, D i,j = 0. It is noted that D i,j = D j,i only when there exist a reversible transition between the nodes i and j; however, μi,j and μj,i do not need to be the same. Given a threshold parameter τ, the connectivity D i,j of the RMC is reduced to zero if μi,j < τ. As a result, the dependency matrix becomes sparse. We want to reorder the sparse dependency matrix and partition them into smaller matrices in order to distribute the computational workloads to multiple computers.

3.2. M-partitioning

Figure 3a shows an example of the partition of a 60 × 60 dependency matrix. Each blue dot represents a nonzero element in the matrix. The four red lines divide the matrix into 9 smaller matrices, including a M 1 × M 1 submatrix, a M 1 × (M 2M 1) submatrix, … , and a (NM 2) × (NM 2) submatrix. Each submatrix is required for the calculations of the state probabilities in Equation (6). Given the initial permutation of the sparse matrices

$${{\bf \prod}} =\sum\limits_{i = 1}^N {i{{\bf e}_i}}\comma \;$$

Fig. 3. Partitions of (a) an original and (b) a reordered dependency matrices.

we focus on finding the optimal permutation to minimize the connectivity between the partitioned systems. In other words, we want to reorder the dependency matrix such that some submatrices are closer to zero matrices, as shown in Figure 3b. In this way, the required matrix multiplications are reduced.

Define a M-partition that partitions the dependency matrix of the RMC, D, into two subsystems:

(8)$${\bf D} = \left[{\matrix{{{\bf D}_1} \hfill &{{{\bf \Phi} _1}} \cr {{{\bf \Phi} _2}} \hfill &{{{\bf D}_2}} \cr } } \right]\comma \;$$

where D1 and D2 are M × M and (NM) × (NM) submatrices, respectively. The sizes of the off-diagonal matrices Φ1 and Φ2 are M × (NM) and (NM) × M, respectively. Meanwhile, the permutation vector is partitioned into two vectors:

$${{\bf \prod}}_{1} = \sum\limits_{i = 1}^M {i{{\bf e}_i}} \quad \hbox{and} \quad {{\bf \prod}}_{2} = \sum\limits_{i = M + 1}^N {i{{\bf e}_i}}.$$

The partition parameter M is controlled to balance the computational loads for each computer. When M < N/2, the computational complexity of the first subproblem is lower; when M > N/2, the second subproblem is more complex. Multiple matrix partitions can be accomplished using multiple M-partitions.

Based on the M-partition, the calculation in Equation (9) is partitioned as in

(9)$$\left[\matrix{{{\bf p}_1}\lpar n\rpar \hfill \cr {{\bf p}_2}\lpar n\rpar \hfill} \right]= {\left[{\matrix{ {{{\bf A}_1}} &{{{\bf \Psi} _1}} \cr {{\bf \Psi} 2} &{{{\bf A}_2}} \cr } } \right]^n}\left[\matrix{{{\bf p}_1}\lpar 0\rpar \hfill \cr {{\bf p}_2}\lpar 0\rpar \hfill} \right]\comma \;$$

where

$${{\bf A}_k} = \sum\limits_{i \in {{\bf P}_k}} {\sum\limits_{\,j \in {{\bf P}_k}} {{D_{i\comma j}}{{\rm \mu} _{i\comma j}}{{\bf e}_j}{{\bf e}_i}} }$$

is calculated based on the permutation Πk. Furthermore, the off-diagonal matrices Ψ1 and Ψ2 are calculated by

$$\sum\limits_{i \in {{\bf P}_1}} {\sum\limits_{\,j \in {{\bf P}_2}} {{D_{i\comma j}}{{\rm \mu} _{i\comma j}}{{\bf e}_j}{{\bf e}_i}} }$$

and

$$\sum\limits_{i \in {{\bf P}_2}} {\sum\limits_{\,j \in {{\bf P}_1}} {{D_{i\comma j}}{{\rm \mu} _{i\comma j}}{{\bf e}_j}{{\bf e}_i}} }\comma \;$$

respectively. The state probabilities in the first and second subproblems are then determined by

(10)$${{\bf p}_1}\lpar n\rpar = {\bf A}_1^n {{\bf p}_1}\lpar 0\rpar + {\,f_1}\lpar {{\bf \Psi} _1}\comma \; {{\bf \Psi} _2}\rpar \comma \;$$
(11)$${{\bf p}_2}\lpar n\rpar = {\bf A}_2^n {{\bf p}_2}\lpar 0\rpar + {f_2}\lpar {{\bf \Psi} _1}\comma \; {{\bf \Psi} _2}\rpar.$$

If the permutations of the matrices and vectors are reordered such that the total probability in the off-diagonal matrices Ψ1 and Ψ2 are close to zero, the functions f 1 and f 2 in Equations (10) and (11) become negligible. Thus, the state probabilities can be approximated by the following equation:

(12)$${{\bf p}_i}\lpar n\rpar \cong {\bf A}_i^n {{\bf p}_i}\lpar 0\rpar.$$

3.3. Iteration process of finding the minimal coupling between subproblems

The objective of our approach is to find the permutation order that minimizes the total transition probability between the two subsystems:

(13)$${\,p_C} = \sum\limits_{i = 1}^M {\sum\limits_{\,j = 1}^{N - M} {{\Psi _{1\comma ij}} + } } \sum\limits_{i = 1}^{N - M} {\sum\limits_{\,j = 1}^M {{\Psi _{2\comma ij}}}. }$$

One of the possible methods is to diagonalize the dependency matrix D; however, it does not ensure the coupling probability p C is minimized for the specific M-partition. In this paper, an effective iteration process is proposed as in the following steps to determine the optimal permutation order:

  • Step 1. Given the initial permutation Π(0) and dependency matrix D(0), calculate the initial coupling probability p C(0) using Equation (13).

  • Step 2. Begin with Ψ1,N−M(0) in the upper-right off-diagonal dependency matrix as the pivot element. The iteration number k equals one.

  • Step 3. Suppose the location of the pivot element is (i, j). If the (i, j)th element in Ψ1(k−1), or the (j,i)th element in Ψ2(k−1), is not zero, the column number of the pivot element is denoted as c p(k); go to step 5. Otherwise, go to step 4.

  • Step 4. If the pivot element is the last element ΨM,1(k−1), the process cannot advance further; go to step 10. Otherwise, move to the adjacent element along the zig-zag trajectory, shown in Figure 4, and consider it as the new pivot element. Go to step 3.

  • Step 5. Begin with the first column, that is, c = 1. Let Π(k) = Π(k−1) and n C(k) = n C(k−1).

  • Step 6. If c = c p(k), move to the adjacent column by c = c + 1. If cN, go to step 7; otherwise, go to step 9.

  • Step 7. Interchange c and c p(k) in the permutation Π(k−1) and calculate the coupling probability p 1 of the interchanged off-diagonal dependency matrix.

  • Step 8. If n 1 < n C(k), n C(k) = n 1 and assign the interchanged permutation as Π(k). Advance to the adjacent column by c = c + 1 and go to step 6.

  • Step 9. If n C(k) = n C(k−1), the iteration process has been converged. Go to step 10. Otherwise, update the interchanged matrix D(k) using the determined permutation Π(k) and advance to the next iteration by k = k + 1. Go to step 4.

  • Step 10. The optimal permutation is Π* = Π(k) and the optimal reordered dependency matrix is D* = D(k).

Fig. 4. Determination of pivot elements along the zig-zag trajectory in the off-diagonal dependency matrix.

4. PPMC

As illustrated in Section 3.2, the partitioned MC has multiple transition probability submatrices. As also mentioned in Section 3, these square matrices can be processed individually to produce partial results, which are combined to form the complete state probability distribution at a desired time instant. Thus, in this paper, the task-farming paradigm is adopted for parallel-processing purposes. The task-farming paradigm has a manager–workers pattern, as shown in Figure 5. The manager is responsible for decomposing a task, distributing subtasks among workers, and gathering partial results from workers to coordinate the final calculation for the reliability evaluation. The worker processes execute in a simple cycle: get a subtask, process the subtask, and send the result to the manager. In addition, due to the task-farming paradigm's dynamic load balancing characteristic, the proposed methodology can respond well to the failure of some processors. This feature facilitates the creation of robust applications that are able to survive from the loss of workers or even the manager. In this paper, for simplicity purposes, with a static load balancing feature, the task-farming paradigm is employed to obtain the state probability vector at any desired time instant.

Fig. 5. Basic flowcharts for the manager and worker programs in the task-farming paradigm.

For a large-scale MC that has been partitioned, each of its transition probability submatrices might still need a large memory space for processing on a single processor. Therefore, another level of parallelism might be required to release or alleviate this memory burden. In addition, the MC-based reliability evaluation is essentially a series of matrix multiplications. Moreover, the message-passing programming paradigm is one of the earliest and most widely used approaches for programming parallel computers, mainly because it imposes minimal requirements on the underlying hardware (Grama et al., Reference Grama, Gupta, Karypis and Kumar2004). Furthermore, the most natural message-passing architecture for matrix operations is a two-dimensional mesh, where each node in the mesh computes one element, or one submatrix, of the resultant array. The mesh connections allow messages to pass between adjacent nodes in the mesh simultaneously. Thus, in this paper, the Cannon algorithm (Cannon, Reference Cannon1969), a memory efficient algorithm based on the two-dimensional mesh framework, is adopted for the matrix multiplication. The Cannon algorithm uses a mesh of nodes with wraparound connections to shift submatrices. In this paper, a node in a mesh represents a processor in a cluster computer. An important point that has to be mentioned is that the Cannon algorithm can also be applied to a nonpartitioned MC for reliability evaluation purposes, which is how this algorithm is used in this paper.

For clarity, elements, instead of submatrices, of the arrays a and b are used to illustrate the Cannon algorithm as follows:

  1. 1. Initially, processor πi,j has elements a i,j and b i,j (0 ≤ i < n, 0 ≤ j < n).

  2. 2. Elements are moved from their initial position to an aligned position. In other words, the complete ith row of a is left-shifted i positions, and then the complete jth column of b is upshifted j positions. This step places the element a i,j+i and the element b i+j,i in processor πi,j. This pair of elements are required to calculate the element c i,j.

  3. 3. Each processor πi,j multiplies its elements.

  4. 4. The ith row of a is shifted one position left, and the jth column of b is shifted one position upward. This step brings together the adjacent elements of a and b, which are also required in the computation of c i,j.

  5. 5. Each processor πi,j multiplies the elements brought to it and adds the result to the accumulating sum.

  6. 6. Repeat Steps 4 and 5 until the final result of c i,j is obtained. In other words, given n rows and n columns of elements, a total of n − 1 shifts need to be conducted.

Next, given that both a and b have s 2 submatrices, each submatrix has m × m elements. Thus, based on the above Cannon algorithm, for both a and b, the initial alignment requires a maximum of s − 1 shift operations. Afterward, there are other s − 1 shift operations required for computation purposes. Each shift operation performs a communication involving m × m elements. Therefore, the communication time t comm can be determined using Equation (14):

(14)$$t_{\rm{comm}} = 4\left({s - 1} \right)\left({t_{\rm{startup}} + m^2 t_{\rm{data}} } \right).$$

Thus, the Cannon algorithm has a communication time complexity of O(sm 2) = O(mn), where s = n/m is assumed in this paper.

As for the computation aspect, each submatrix multiplication requires m 3 multiplications and m 3 additions. Therefore, with s − 1 shifts as mentioned above, the computation time t comp can be determined using Equation (15):

(15)$$t_{\rm{comp}}= 2sm^3 = 2m^2 n.$$

Hence, the Cannon algorithm has a computation time complexity of O(m 2n).

In this paper, the Cannon algorithm is implemented through the message-passing interface (MPI; Gropp et al., Reference Gropp, Huss-Lederman, Lumsdaine, Lusk, Nitzberg, Saphir and Snir1998; Snir et al., Reference Snir, Otto, Huss-Lederman, Walker and Dongarra1998). MPI is a specification for building a library that provides standard functions for writing portable and efficient message-passing programs to be run on a variety of parallel computers. In this paper, the selected MPI library is a widely used open source library called MPICH2 (Gropp, Reference Gropp2002).

5. Numerical Examples

In this section, we study the MCs for three different examples to validate the proposed PPMC method. The inputs and outputs of the proposed PPMC reliability evaluation mechanism are shown in Figure 6. The inputs include all distinct states of a system, all transition probabilities, the threshold parameter, initial state probabilities, the number of time steps, the index of the initial state, and the index (or indices) of the failure state (or states). The outputs include the probability (or probabilities) of the failure state (or states), and the overall system reliability.

Fig. 6. Inputs and outputs of the parallel processing of a Markov chain reliability evaluation mechanism.

The PPMC reliability evaluation mechanism can be applied to all kinds of systems whose states can be modeled as a MC. Particularly, it is suitable for systems that potentially contain a vast number of distinct states, such as nuclear power systems (Aldemir et al., Reference Aldemir, Miller, Stovsky, Kirschenbaum and Buccim2006). As shown in Figure 7, it is assumed that there are 14 sensors located at different components of a nuclear steam system. In addition, each sensor is assumed to have two states indicating that a component is either functioning well or not functioning at all. Thus, the number of distinct states of the system is 214 = 16384. Among these states, some of them can be defined as the failure states. The PPMC reliability evaluation mechanism can be used to efficiently generate, for a specified number of time steps, the failure states' probabilities and the overall system reliability. Therefore, if desired, engineers can perform necessary maintenance activities, in advance, on the components related to critical failure states. Afterward, the transition probability matrix can be changed by increasing or decreasing probability values on corresponding elements. Eventually, the PPMC reliability evaluation mechanism can be used to generate new results for necessary further improvements on the system reliability.

Fig. 7. A nuclear steam system containing a large number of different states.

5.1. Example 1: 60-state MC for MMR controllers

The first example considers a randomly generated 60-state MC, which is composed of 277 distinct directed edges. Figure 8a shows the dependency matrix of the 60-state MC while each blue dot represents one of the directed connections. The blank areas stand for no connections between the states. Figure 8a can also be used to represent the transition matrix; in this case, each blue dot is a nonzero transition probability μi,j.

Fig. 8. The (a) original and (b) partitioned dependency matrices for the 60-state Markov chain.

When the size of MC is in the order of tens of states, it is not necessary to reduce the MC using the threshold parameter. In the latter case, when the MC size is larger, the reduction can help improve the numerical efficiency. To uniformly distribute the computational workloads to five computers, we partition the dependency matrix at the positions of M = {12, 24, 36, 48}. The red lines represent the desirable locations for matrix partitioning. According to the desirable partitioning locations, the dependency matrix is reordered such that the total probability in the off-diagonal submatrices, shown in the gray areas in Figure 8b, is minimized. The ratio of the total coupling probability and the total probability is 9.66/60 = 0.16. In other words, around 84% of workloads are uniformly distributed in the five diagonal submatrices. The discrete-time state probabilities are calculated using Equation (7).

Suppose the initial condition is given as p 15(0) = 1 and p i(0) = 0 for i = 1 … 14, 16, … 60, the state probabilities at time n = 10 are calculated using the proposed method. Furthermore, the numerical performances of the PPMC method are compared with the calculations using the original MC in Equation (5). Three different simulation results are listed in Table 1. The first case considers the second state as the fail state. The original matrix multiplication shows the failure probability is 0.07%, while the proposed method overestimates the failure rate as 12.5%. A worst-case reliability, 87.50%, is then found, that is, the true probability, 99.93%, is at least larger than or equal to the underestimation, 87.50%.

Table 1. Results of the 60-state Markov chain

Note: PPMC, Partitioning and parallel processing of Markov chain.

The second case shows a higher failure probability of 3.06% when the 24th state is considered as the fail state. In this case, the proposed method is capable of finding the overestimated failure probability of 12.5% and the worst-case reliability of 87.5%. In the last case, the 30th state is considered as the fail state. In this case, the PPMC method obtains the worst-case reliability of 99.87%, which is just slightly lower than the true probability of 99.88%.

5.2. Example 2: 200-state MC for MMR controllers

The second case focuses on a 200-state MC, which contains 605 directed edges. Figure 9a shows the dependency matrix of the given MC. In this case, the MC reduction is also unnecessary. To distribute the workloads to five computers, the desirable partitioning locations are M = {40, 80, 120, 160}, indicated by the red lines. The off-diagonal probabilities are minimized to reorder the dependency matrix. As a result, the off-diagonal probability, the sum of probabilities in the gray areas, equals 34.33, which is around 17.16% over the total probability. The values in the off-diagonal submatrices will be neglected in Equation (7), yielding the underestimated measure of system reliability. Therefore, the worst-case reliability can be determined.

Fig. 9. The (a) original and (b) partitioned dependency matrices for the 200-state Markov chain.

We now want to demonstrate different types of simulation results from the first example. Suppose the 5th state is considered as the fail state, then three different initial states are studied. The first case starts with p(0) = 1e32. The original matrix multiplication shows the fail probability of 6.63% as the PPMC overestimates the fail probability of 7.05%; therefore, a worst-case reliability of 92.95% is determined. Another initial condition considers p(0) = 1e18. The proposed method underestimates the reliability, that is, 82.83% < true probability = 83.25%, and utilizes the worst-case measure as a robust estimator of the system probability. Finally, when the 99th state is considered as the initial state, a lower reliability of 77.01% is found using the original calculations. Even for the special situation of low reliability, the proposed method is capable of finding the worst-case reliability, that is, 75.32%. Table 2 lists the detailed information about the simulation results.

Table 2. Results of the 200-state Markov chain

Note: PPMC, Partitioning and parallel processing of Markov chain.

5.3. Example 3: 480-state MC for MMR controllers

The final example considers a large-scale MC that contains 480 states. Figure 10a shows the dependency matrix with 1899 blue dots, that is, there are 1899 distinct directed edges in the MC. In a large problem like this, the matrix reduction can effectively improve the numerical performances. We assume the threshold parameter is τ = 0.1. Figure 10b shows the dependency matrix of the reduced MC, which now only contains 710 directed edges. The matrix permutation is then reordered and the resultant dependency matrix is shown in Figure 10c. The total off-diagonal probability equals 110.37, which is around 23% of the total probability. However, 77% of the workloads are uniformly distributed to eight different computers. The red lines represent the given partitioning locations M = {60, 120, 180, 240, 300, 360, 420}.

Fig. 10. The (a) original, (b) reduced, and (c) partitioned dependency matrices for the 480-state Markov chain.

Table 3 lists two different cases of the simulations. The first case considers the 10th and 406th states as the initial and fail states. The proposed method is able to find the worst-case reliability, 99.04%, which is slightly lower than the true measure, 99.75%. In the other case, in which the 432nd and 31st states are the initial and fail states, the PPMC method obtains an underestimation of the system reliability, that is, 70.79% << true probability = 99.90%, due to the errors from the matrix reduction and approximated calculation in Equation (7). However, the proposed method still guarantees that the true system reliability is at least larger than or equal to the underestimated measure.

Table 3. Results of the 480-state Markov chain

Note: PPMC, Partitioning and parallel processing of Markov chain.

5.4. Numerical performances in Example 3

A 25-node cluster computer is used to run example 3 in the following scenarios:

  1. 1. Evaluate the system reliability of the original MC at the 10th time step by serial matrix multiplication method using one CPU.

  2. 2. Evaluate the system reliability of the original MC at the 10th time step by the Cannon algorithm using 16 CPUs.

  3. 3. Evaluate the system reliability of the partitioned MC at the 10th time step by applying serial matrix multiplication method on eight diagonal submatrices using eight CPUs.

All of the nodes of the cluster computer have identical key specifications, including a dual core CPU, 1-GHz core CPU frequency, and 1.837 GB of memory. Due to homogeneous hardware features, the time spent by each CPU to finish its tasks is almost the same even though the computation time is still defined as the longest time spent by any CPU. Moreover, for each scenario, the computation time only changes very slightly in every execution. Thus, the average of 10 computation times is provided for each scenario. As shown in Table 4, in scenario 1, a single CPU needs to process 230,400 data points; in scenario 2, each of the 16 CPUs needs to process 14,400 data points; in scenario 3, each of the 8 CPUs needs to process 3600 data points. The average computation times for scenarios 1, 2, and 3 are 25501, 2397, and 74 ms, respectively. The speedup ratios for scenarios 1, 2, and 3 are 1, 10.6, and 344.6, respectively.

Table 4. Computation times for three scenarios

Note: PPMC, Partitioning and parallel processing of Markov chain.

There are two factors that affect the accuracy of reliability evaluation: the reductions of directed edges with transition probabilities lower than the selected threshold parameter and the unconsidered edges in the off-diagonal matrix. Lin et al. (Reference Lin, Chou, Manuel and Hsu2014) have further investigated the numerical performance of PPMCs with various level of threshold parameter τ. In Example 3, 1899 distinct directed edges in Figure 10a deduced to 710 edges when τ = 0.1 was considered. Fewer edges would be reduced as a lower threshold parameter is considered but the complexity of MC increases. More edges would be reduced as a higher threshold parameter is utilized but the accuracy of reliability evaluation may be jeopardized. A proper selection of threshold parameter has been found related to the distribution of transition probabilities in the MC. For more information, please refer to Lin et al. (Reference Lin, Chou, Manuel and Hsu2014).

6. Conclusions

In this paper, the method of PPMC has been presented to perform a robust reliability analysis for complex systems, such as MMR controllers. The number of different states N can be easily increased by three factors, including the number of modular redundant controllers, the number of sensors each controller manages, and the number of states each sensor indicates. The proposed method partitioned the complex MC and solved for a worst-case reliability using parallel computational processes. In PPMC, a threshold parameter has been chosen to reduce the dependency of MC in terms of eliminating the edges of negligible probabilities between different states. A M-partitioning procedure has been developed to gather together M states that are mutually related and decomposed the complex problem into multiple subproblems, which are lowly dependent from each other. In addition, calculations regarding the submatrices can be performed independently in parallel. This research is the first attempt to reduce, partition, and reorder a MC, in order to enable seamless incorporation of parallel computing techniques for resource-efficient calculations. Compared with the sequential method applied to a nonpartitioned MC, the proposed PPMC reduces the overall computation time, lowers the amount of memory required on a single processor, and produces allowable accuracy for the reliability analysis on large-scale systems of MMR controllers in the three demonstrated examples.

The main advantage of the proposed method is its capability of efficiently performing the reliability analysis of a MC in various engineering problems. It also assumes a worst-case scenario when compared to evaluation of the true system reliability, which is conservative in its nature. The proposed method is also anticipative in the advances in the field of computing sciences, which could further reduce the computational parameters. However, PPMC currently assumes the transition probabilities between different states are known and remain constant. Furthermore, the decision of threshold parameter for the reduction of MC is not automatically determined. It is expected that the numerical performances of PPMC depend on proper selection of the threshold parameter. The methodology developments include the adaptiveness to the dynamic transition probabilities between different states, the optimal selection of threshold parameter for MC reduction with the consideration of the distribution of the probability density, the enhancement of the numerical efficiency of the decomposition procedure, and the improvement of the numerical accuracy of the reliability analysis. These developments are expected to make the proposed method competitive in the implementation of evaluating and analyzing MCs in complex engineering systems.

ACKNOWLEDGMENTS

This research was supported by the Atomic Energy Council of Taiwan (Grant 1012001INER010), the Ministry of Science and Technology of Taiwan (Grants 101-2628-E-033-001-MY3, 102-2218-E-033-002-MY2, and 102-2221-E-033-020), the Research and Development Center for Microsystem Reliability at Chung Yuan Christian University, and the Center for Biomedical Technology at Chung Yuan Christian University.

NOMENCLATURE

a

An array in the Cannon algorithm

A

Transition matrix

Ai

i-th partitioned transition matrix

b

Another array in the Cannon algorithm

c

Column number

c p

Column number of the pivot element in dependency matrix

C i

ith coefficient

D

N × N dependency matrix

Di

ith partitioned dependency matrix

D ij

Connectivity parameter, where D ij = 1 when the ith and jth states are connected; otherwise, it is zero.

ei

ith normal basis

m

Dimensional parameter of submatrices for multiple processors

M

Dimensional parameter for the partitioned matrix

n

Time parameter in the computer processing

N

Number of states in Markov chain

p

Vector of transition probabilities

p C

Total probability of the partitioned Markov chains couple with each other

pi

ith partitioned vector of transition probabilities

p I

Total coupling probability for the interchanged dependency matrix

P(A)

Probability of A

P(A | B)

Conditional probability of A given B

Q

Infinitesimal generator of continuous-time Markov chain

s

A dimensional parameter

S

Finite state space

t

Time

X(t)

Variable with Markov property at time t

λi

ith eigenvalue

μi,j

Transition probability (failure/recovery rate) of the state i moves to the state j

vi

ith eigenvector

δ

A small number

πi,j

A processor with elements a i,j and b i,j

Π

N × 1 permutation vector

Πi

ith partitioned permutation vector

τ

Threshold parameter

Φi

ith off-diagonal dependency matrix

Ψi

ith off-diagonal transition matrix

Superscripts

Derivative with respect to time

*

Optimal solution

(k)

kth iteration in finding the optimal partition

(n)

nth step in computer processing

Po Ting Lin is a Professor in the Department of Mechanical Engineering, Research and Development Center for Microsystem Reliability and Center for Biomedical Technology at Chung Yuan Christian University.

Yu-Cheng Chou is a Professor in the Institute of Undersea Technology at National Sun Yat-sen University.

Yung Ting is a Professor in the Department of Mechanical Engineering at Chung Yuan Christian University.

Shian-Shing Shyu is a Scientist at the Institute of Nuclear Energy Research of the Atomic Energy Council in Taiwan.

Chang-Kuo Chen is a Scientist at the Institute of Nuclear Energy Research of the Atomic Energy Council in Taiwan.

References

REFERENCES

Aldemir, T., Miller, D., Stovsky, M., Kirschenbaum, J., & Buccim, P. (2006). Current State of Reliability Modeling Methodologies for Digital Systems and Their Acceptance Criteria for Nuclear Power Plant Assessments. Washington, DC: US Nuclear Regulatory Commission.Google Scholar
Bhaduri, D., Shukla, S.K., Graham, P.S., & Gokhale, M.B. (2007). Reliability analysis of large circuits using scalable techniques and tools. IEEE Transactions on Circuits and Systems I 54(11), 24472460.Google Scholar
Cai, B., Liu, Y., Liu, Z., Tian, X., Li, H., & Ren, C. (2012). Reliability analysis of subsea blowout preventer control systems subjected to multiple error shocks. Journal of Loss Prevention in the Process Industries 25, 10441054.Google Scholar
Cannon, L.E. (1969). A cellular computer to implement the Kalman filter algorithm. PhD Thesis. Montana State University.Google Scholar
Chou, Y.-C., & Lin, P.T. (2014). An efficient and robust design optimization of multi-state flow network for multiple commodities using generalized reliability evaluation algorithm and edge reduction method. International Journal of Systems Science. Advance online publication. doi:10.1080/00207721.2013.879228Google Scholar
Dominguez-Garcia, A.D., Kassakian, J.G., & Schindall, J.E. (2006). Reliability evaluation of the power supply of an electrical power net for safety-relevant applications. Reliability Engineering & System Safety 91(5), 505514.Google Scholar
Grama, A., Gupta, A., Karypis, G., & Kumar, V. (2004). Introduction to Parallel Computing. Essex: Pearson Education.Google Scholar
Gropp, W. (2002). MPICH2: A new start for MPI implementations. Proc. 9th European PVM/MPI Users' Group Meeting on Recent Advances in Parallel Virtual Machine and Message Passing Interface.Google Scholar
Gropp, W., Huss-Lederman, S., Lumsdaine, A., Lusk, E., Nitzberg, B., Saphir, W., & Snir, M. (1998). MPI: The Complete Reference—The MPI-2 Extensions. Cambridge, MA: MIT Press.Google Scholar
Guo, H., & Yang, X. (2008). Automatic creation of Markov models for reliability assessment of safety instrumented systems. Reliability Engineering & System Safety 93(6), 829837.Google Scholar
Han, J., Gao, J., Jonker, P., Qi, Y., & Fortes, J.A. (2005). Toward hardware-redundant, fault-tolerant logic for nanoelectronics. IEEE Design & Test of Computers 22(4), 328339.Google Scholar
Kim, H., Lee, H., & Lee, K. (2005). The design and analysis of AVTMR (all voting triple modular redundancy) and dual-duplex system. Reliability Engineering & System Safety 88(3), 291300.Google Scholar
Lin, P.T., Chou, Y.-C., Manuel, M.C.E., & Hsu, K.S. (2014). Investigation of numerical performance of partitioning and parallel processing of Markov chain (PPMC) for complex design problems. Proc. ASME 2014 Int. Design & Engineering Technical Confs. and Computers & Information in Engineering Conf., IDETC/CIE 2014, Paper No. DETC2014-34652, Buffalo, NY.Google Scholar
Lisnianski, A., Elmakias, D., Laredo, D., & Ben Haim, H. (2012). A multi-state Markov model for a short-term reliability analysis of a power generating unit. Reliability Engineering & System Safety 98(1), 16.Google Scholar
Liu, Y., & Rausand, M. (2011). Reliability assessment of safety instrumented systems subject to different demand modes. Journal of Loss Prevention in the Process Industries 24(1), 4956.Google Scholar
Liu, Z., Liu, Y., Cai, B., Liu, X., Li, J., Tian, X., & Ji, R. (2013). RAMS analysis of hybrid redundancy system of subsea blowout preventer based on stochastic Petri nets. International Journal of Security and Its Applications 7(4), 159166.Google Scholar
Liu, Z., Ni, X., Liu, Y., Song, Q., & Wang, Y. (2011). Gastric esophageal surgery risk analysis with a fault tree and Markov integrated model. Reliability Engineering & System Safety 96(12), 15911600.Google Scholar
Matthews, P., & Philip, A. (2012). Bayesian project diagnosis for the construction design process. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 26(4), 375391.Google Scholar
Mutha, C., Jensen, D., Tumer, I., & Smidts, C. (2013). An integrated multidomain functional failure and propagation analysis approach for safe system design. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 27(4), 317347.Google Scholar
Parashar, B., & Taneja, G. (2007). Reliability and profit evaluation of a PLC hot standby system based on a master-slave concept and two types of repair facilities. IEEE Transactions on Reliability 56(3), 534539.Google Scholar
Snir, M., Otto, S., Huss-Lederman, S., Walker, D., & Dongarra, J. (1998). MPI: The Complete Reference—The MPI Core. Cambridge, MA: MIT Press.Google Scholar
Soro, I.W., Nourelfath, M., & Ait-Kadi, D. (2010). Performance evaluation of multi-state degraded systems with minimal repairs and imperfect preventive maintenance. Reliability Engineering & System Safety 95(2), 6569.Google Scholar
Stewart, W.J. (2009). Probability, Markov Chains, Queues, and Simulation: The Mathematical Basis of Performance Modeling. Princeton, NJ: Princeton University Press.Google Scholar
Wang, S., Ji, Y., Dong, W., & Yang, S. (2007). Design and RAMS analysis of a fault-tolerant computer control system. Tsinghua Science & Technology 12(Suppl. 1), 116121.Google Scholar
Yu, H., Chu, C., Châtelet, Ė., & Yalaoui, F. (2007). Reliability optimization of a redundant system with failure dependencies. Reliability Engineering & System Safety 92(12), 16271634.Google Scholar
Zhang, C.W., Zhang, T., Chen, N., & Jin, T. (2013). Reliability modeling and analysis for a novel design of modular converter system of wind turbines. Reliability Engineering & System Safety 111, 8694.CrossRefGoogle Scholar
Zhang, T., Long, W., & Sato, Y. (2003). Availability of systems with self-diagnostic components-applying Markov model to IEC 61508-6. Reliability Engineering & System Safety 80(2), 133141.Google Scholar
Figure 0

Fig. 1. N-state Markov chain for multimodular redundant controllers.

Figure 1

Fig. 2. N-state Markov chain for multimodular redundant controllers.

Figure 2

Fig. 3. Partitions of (a) an original and (b) a reordered dependency matrices.

Figure 3

Fig. 4. Determination of pivot elements along the zig-zag trajectory in the off-diagonal dependency matrix.

Figure 4

Fig. 5. Basic flowcharts for the manager and worker programs in the task-farming paradigm.

Figure 5

Fig. 6. Inputs and outputs of the parallel processing of a Markov chain reliability evaluation mechanism.

Figure 6

Fig. 7. A nuclear steam system containing a large number of different states.

Figure 7

Fig. 8. The (a) original and (b) partitioned dependency matrices for the 60-state Markov chain.

Figure 8

Table 1. Results of the 60-state Markov chain

Figure 9

Fig. 9. The (a) original and (b) partitioned dependency matrices for the 200-state Markov chain.

Figure 10

Table 2. Results of the 200-state Markov chain

Figure 11

Fig. 10. The (a) original, (b) reduced, and (c) partitioned dependency matrices for the 480-state Markov chain.

Figure 12

Table 3. Results of the 480-state Markov chain

Figure 13

Table 4. Computation times for three scenarios