1 Introduction

In some real world situations, we have to make important decisions based on less information, because obtaining more information would cost resources like time, effort and money. For instance, consider an expensive coherent system with known structure. If we had some information about the lifetimes of coherent systems with the same structure, we could easily find exact and efficient prediction intervals. But the system is expensive, and it would be costly to obtain the information. Only some informations about the component lifetimes of the system are available and we are interested in finding prediction intervals for the future lifetime of the system. In this paper, we intend to follow such a valuable plan and construct some prediction intervals for the lifetime of a coherent system with known structure or signature vector. With this in mind, let us briefly review some relevant results on coherent systems.

Consider the space \(\{0, 1\}^n\) of all possible state vectors for an \(n\)-component system. The structure function \(\varphi : \{0,1\}^n\rightarrow \{0, 1\}\) is a mapping that associates those state vectors \(\mathbf x \) for which the system works with value one and those state vectors \(\mathbf x \) for which the system fails with the value zero. A system is said to be coherent if each of its components is relevant and if its structure function is monotone. A set of components \(P\) is said to be a path set if the system works whenever all the components in the set \(P\) work. A path set is minimal if it has no proper subset that is also path set and the algebraic union of all minimal path sets is the set of all the system’s components. A set of components \(C\) is said to be a cut set if the system fails whenever all the components in the set \(C\) fail. A minimal cut set is a cut set that contains no proper subset that is also cut set. For more details on the coherent system and its relevant concepts, see Barlow and Proschan (1981).

Let \(Y_1, Y_2, \ldots , Y_n\) be independent and identically distributed \((i. i. d.)\) random variables with cumulative distribution function (cdf) \(F(y)\) and probability density function (pdf) \(f(y)\), denoting the component lifetimes of a coherent system and \(Y_{r:n}, \ r=1, \ldots , n\) denote the \(r\)th smallest lifetime. Samaniego (1985) defined the signature \(\mathbf s \) of a coherent system of order \(n\) with the \(n\)-dimensional probability vector whose \(i\)th element is \(s_i=P(T=Y_{i:n})\), where \(T\) is the system lifetime. System signatures have been found to be quite useful tools in the study and comparison of engineered systems. Samaniego (1985) also proved that for a coherent system with \(i. i.d.\) components \(Y_1, Y_2,\ldots , Y_n\) , the signature vector \(\mathbf s \) only depends on the structure function of the system. Moreover, the reliability function of \(T\) is given by

$$\begin{aligned} \bar{F}_T (t) = \sum \limits _{i = 1}^n {s_i \bar{F}_{i:n} (t)}, \end{aligned}$$
(1)

where \(\bar{F}_{i:n} (t) = P(X_{i:n} > t)\), for \(i=1, 2, \ldots , n\) and \(\sum \nolimits _{i = 1}^n {s_i}=1\). From (1), the density function of \(T\) in terms of signature vector \(\mathbf s \) is given by

$$\begin{aligned} f_T(t)= \sum \limits _{i = 1}^n {i s_i {n \atopwithdelims ()i}(F(t))^{i - 1} (\bar{F}(t))^{n - i} f(t)}. \end{aligned}$$
(2)

Navarro and Rychlik (2007) proved that the identity (1) also holds for coherent systems with component lifetimes having an absolutely continuous exchangeable joint distribution. We recall that the vector \((X_1, X_2, \ldots , X_n)\) has a joint exchangeable probability density function \(f\), if \(f(x_1, x_2,\ldots , x_n)=f(x_{\pi _1}, x_{\pi _2}, \ldots , x_{\pi _n})\) for any permutation \(\mathbf{{\pi }}=(\pi _1, \ldots ,\pi _n)\) of \(\{1, 2, \ldots , n\}\). It should be mentioned that for an absolutely continuous exchangeable joint distribution, \(\bar{F}_{i:n} (t)\) is given by

$$\begin{aligned} \bar{F}_{i:n} (t) = \sum \limits _{j = n - i + 1}^n {( - 1)^{j + i - n - 1} {n \atopwithdelims ()j} {j-1 \atopwithdelims ()n-i} \bar{F}_{1:j}(t)}, \end{aligned}$$
(3)

see for example David and Nagaraja (2003, p. 46).

Navarro et al. (2007) proved that for a coherent system with exchangeable components, \(\bar{F}_T\) can be expressed based on the lifetimes of series and parallel systems as

$$\begin{aligned} \bar{F}_T (t) = \sum \limits _{i = 1}^n {a_i } \bar{F}_{1:i} (t) = \sum \limits _{i = 1}^n {b_i } \bar{F}_{i:i} (t), \end{aligned}$$
(4)

where \(\sum \nolimits _{i=1}^{n}a_i=1\) and \(\sum \nolimits _{i=1}^{n}b_i=1\). The vectors of coefficients \(\mathbf a =(a_1, a_2, \ldots ,a_n)\) and \(\mathbf b =(b_1, b_2, \ldots , b_n)\) only depend on the structure function of the system and called minimal signature and maximal signature, respectively. They can be obtained from the representation of reliability function based on minimal path set or minimal cut set (see Navarro et al. 2007).

In recent years, several authors have studied the reliability properties of coherent systems by using the signature concept. We refer, among the others, to Navarro et al. (2010a, b), Khaledi and Shaked (2007), Li and Zhang (2008a, b), Triantafyllou and Koutras (2008), Samaniego et al. (2009), Eryilmaz (2009, 2010, 2011, 2013), Eryilmaz and Zuo (2010), Bhattacharya and Samaniego (2010), Balakrishnan et al. (2011) and Ng et al. (2012). For a comprehensive discussion on the applications of system signature in engineering reliability, one can see Samaniego (2007). In this paper, we will obtain prediction intervals for the future lifetime of a coherent system with \(i. i. d.\) and exchangeable components by using signature, minimal signature and maximal signature vectors. For this purpose the rest of the paper is organized as follows: In Sect. 2, it would be supposed that the failure times of \(m\) components with identical distribution \(F\), which are sampled from the production line, are observed. A prediction interval for the future lifetime of a coherent system, composed of the same components and put into operation in the future, is constructed based on the observed failure times. Clearly the observed failure times contain valuable information about the system’s lifetime. Also, in Sect. 2, we will find a prediction interval for the lifetime of a coherent system based on order statistics. In Sect. 3, we consider a sequence of minimal repair times of a component and determine a prediction interval for the lifetime of a coherent system based on this partial information. These results are extended to the case in which the \(m\) components begin to operate separately at time zero. If each of the components fail, then it undergoes minimal repair and begins to operate again. The repair times and the lifetime of the components are assumed to be independent of each other. Each component can be repaired \(\tau -1\) times. Thus, we observe an \(m \times \tau \) matrix of failure times of \(m\) components. This information applies to construct some nonparametric prediction intervals for the lifetime of a coherent system with known structure.

2 Prediction intervals based on order statistics

Consider a coherent system composed of \(n\) independent identical components with cdf \(F\), where its signature vector is known. It would be interesting to predict the system’s lifetime based on systems with a simple structure. For this purpose, let \(X_i, i=1, \ldots , m\) be positive independent random variables with common distribution \(F\) and \(X_{k:m}\) be the \(k\)th order statistic. If \(X_i, i=1, \ldots , m\) are the lifetimes of the components of a \(k\)-out-of-\(m\) system (the systems that fail upon the \(k\)th component failure), then \(X_{k:m}\) is the lifetime of the system. Many properties and applications of this system have been studied by several authors (see, e.g., Barlow and Proschan 1981; Meeker and Escobar 1998). For the recent results on the lifetime of \(k\)-out-of-\(m\) systems, we refer to Gurler (2012) and references therein. Here, we obtain prediction intervals for the future system’s lifetime based on the observed order statistics, \(X_{1:m}, \ldots , X_{m:m}\). This is stated in the next theorem.

Theorem 1

Let \(X_1, X_2, \ldots , X_m\) be a sample of size \(m\) of \(i. i. d.\) positive continuous random variables with cdf \(F(x)\) and pdf \(f(x)\), and \(X_{1:m}, X_{2:m}, \ldots , X_{m:m}\) be the corresponding order statistics. Let \(T\) denote the lifetime of a coherent system based on component failure times \(Y_1, Y_2, \ldots , Y_n\) with the same cdf and pdf. Then, \((X_{i:m}, X_{j:m}),\;j> i \ge 1\), is a two-sided prediction interval for \(T\) whose coverage probability is free of \(F\) and is given by

$$\begin{aligned} \alpha _1(i, j; m, n, \mathbf s )=\frac{n}{m+n}\sum \limits _{h = i}^{j-1} {\sum \limits _{\ell = 1}^{n} {{s_{\ell }\frac{{m \atopwithdelims ()h} {{n-1} \atopwithdelims (){\ell -1}}}{{{m+n-1} \atopwithdelims (){\ell +h-1}}}} }}, \end{aligned}$$
(5)

where \(\mathbf s =(s_1, \ldots , s_n)\) with \(s_{\ell }=P(T=Y_{\ell :n})\) and \(\sum \nolimits _{\ell =1}^n s_{\ell }=1\).

Proof

By conditional arguments, we have

$$\begin{aligned} \alpha _1\big (i, j; m, n, \mathbf s )&= P(X_{i:m} \le T \le X_{j:m}\big )\\&= \int \limits _{0}^{\infty } {P\big (X_{i:m} \le T \le X_{j:m} |T = t\big )\,f_T (t)\,dt}. \end{aligned}$$

By independence of \(T\) and \(\{X_i; 1\le i \le m \}\) and using (2) it is easy to show that

$$\begin{aligned} \alpha _1(i, j; m, n, \mathbf s )&= \int \limits _{0}^{\infty } {\sum \limits _{h = i}^{j - 1} {m \atopwithdelims ()h} \big [F(t)\big ]^h \big [\bar{F}(t)\big ]^{m - h} f_T (t)dt} \\&= {\sum \limits _{h=i}^{j-1}{\sum \limits _{\ell =1}^{n}{\ell }{m \atopwithdelims ()h}{n \atopwithdelims ()\ell }s_\ell \int \limits _{0}^{1} u^{\ell +h-1}{(1-u)}^{m+n-\ell -h} du}} \\&= \frac{n}{m+n} {\sum \limits _{h = i}^{j - 1}\sum \limits _{\ell =1}^{n}s_{\ell }{{m \atopwithdelims ()h}{{n-1} \atopwithdelims (){\ell -1}}}\bigg /{{{m+n-1} \atopwithdelims (){\ell +h-1}}}}. \end{aligned}$$

\(\square \)

From Theorem 1, we immediately deduce the following special cases.

Corollary 1

Under the assumptions of Theorem 1, we have:

  1. (i)

    for a \(k\)-out-of-\(n\) system, i.e., \(s_k=P(T=Y_{k:n})=1\), the prediction coefficient (5) reduces to

    $$\begin{aligned} \alpha _1(i, j; m, n, \mathbf s )=\frac{n}{m+n}{\sum \limits _{h=i}^{j-1}{m \atopwithdelims (){h}}{{n-1} \atopwithdelims (){k-1}} \bigg /{{m+n-1} \atopwithdelims (){k+h-1}}}; \end{aligned}$$
    (6)
  2. (ii)

    \(X_{i:m}\) is a lower prediction bound with prediction coefficient

    $$\begin{aligned} P(T \ge X_{i: m} ) =\frac{n}{m+n}{\sum \limits _{h=i}^{m}\sum \limits _{\ell =1}^{n}s_{\ell }{{m \atopwithdelims ()h}{{n-1} \atopwithdelims (){\ell -1}}}\bigg /{{{m+n-1} \atopwithdelims (){\ell +h-1}}}}; \end{aligned}$$
    (7)
  3. (iii)

    similarly, \(X_{j:m}\) is an upper prediction bound for \(T\), where the coverage probability for prediction interval \((0, X_{j:m})\) is given by \( \alpha _1(0, j; m, n, \mathbf s )\) as in (5).

It may be noted that the probability elements of (6) are identical with Eq. (5.24) in Samaniego (2007). Also, when \(T\) is a \(j\)-out-of-\(n\) system, that is \(s_j=P(T\!=\!Y_{j:n})=1\), then the expression in (7) coincides with the expression (5.24) in Samaniego (2007, p. 70).

For the case of minimal signature or maximal signature, the corresponding prediction coefficient can be obtained by using (4), directly. If the system’s components are \(i. i. d.\) the coverage probability in terms of minimal signature is given by

$$\begin{aligned} P(T \ge X_{r: m} )&= \int \limits _{0}^{\infty } {\sum \limits _{i = 1}^n {a_i } \bar{F}_{1:i} (t) m} {{m-1}\atopwithdelims (){r-1}}\big [F(t)\big ]^{r - 1} \big [\bar{F}(t)\big ]^{m - r} f(t)dt\\&= \sum \limits _{i = 1}^n {a_i } \frac{{m \atopwithdelims ()r}}{{{i+m} \atopwithdelims ()r}}, \quad {\mathrm{where}} \quad \sum \limits _{i=1}^{n}a_i=1. \end{aligned}$$

2.1 Optimal prediction interval

For a given \(\alpha _0\), the two-sided prediction interval \((X_{i:m}, X_{j:m}), \;1 \le i <j \le m, \) exists if and only if, \(P(X_{1:m}\le T \le X_{m:m})\ge \alpha _0\). In other words, for a given \(\alpha _0, ~\mathbf s \) and \(n\), the sample size \(m\) should be satisfied in the following inequality

$$\begin{aligned} 1-\frac{n}{m+n}\sum \limits _{\ell =1}^{n}s_{\ell }\left[ \frac{{{n-1} \atopwithdelims (){\ell -1}}}{{{m+n-1}\atopwithdelims (){n-\ell }}}+\frac{{{n-1}\atopwithdelims (){\ell -1}}}{{{m+n-1}\atopwithdelims (){\ell -1}}} \right] \ge \alpha _0. \end{aligned}$$
(8)

For a \(k\)-out-of-\(n\) system, \(m\) should be satisfied in the following inequality

$$\begin{aligned} \mathop {max }\limits _{i,j} \alpha _1 (i,j; m, n, \mathbf s ) =1- \frac{n}{m+n}\left[ \frac{{{n-1} \atopwithdelims (){k-1}}}{{{m+n-1}\atopwithdelims (){n-k}}}+\frac{{{n-1}\atopwithdelims (){k-1}}}{{{m+n-1}\atopwithdelims (){k-1}}} \right] \ge \alpha _0. \end{aligned}$$

The prediction coefficient and expected width of the prediction interval are decreasing in \(i\) and increasing in \(j\). Hence, if \(\mathbf s \) and \(\alpha _0\) are prefixed and \(m\) satisfies in (8), then we can find \(i_{opt}\) and \(j_{opt}\) such that the expected width of the prediction interval \((X_{i_{opt}:m},\, X_{j_{opt}:m})\) be less than any other prediction interval \((X_{i:m}, X_{j:m})\). For this purpose, we present the following algorithm.

Algorithm 1

For a given \(\alpha _0\), \(\mathbf s \), \(m\) and \(n\), the optimal prediction interval is determined through the following steps:

  • Step 1: Take \(i_{max}= max\{i: \alpha _1(i, m; m, n, \mathbf s ) \ge \alpha _0\}\)

  • Step 2: Set \(i=i_0=1\) and \(j=i_0+1\).

  • Step 3: Gradually increase \(j\) until \(\alpha _1(i_0, j; m, n, \mathbf s )\) becomes greater than or equal to \(\alpha _0\).

  • Step 4: Calculate the expected width of the prediction interval resulting from step 3.

  • Step 5: Set \(i=i_0+1\) and start with \(j=i_0+2\) and follow the above procedure until \(1 \le i \le i_{max}\).

  • Step 6: By this procedure, we find all the pairs of \((i, j)\) such that \(\alpha _1(i, j; m, n, \mathbf s ) \ge \alpha _0\). With comparing the equivalent expected width of these prediction intervals we can find a prediction interval with prediction coefficient at least \(\alpha _0\) and minimum width.

It should be noted that the prediction coefficient \(\alpha _1(i, j; m, n, \mathbf s )\) is distribution-free, and hence in step 4 of Algorithm 1, the expected width of the interval \((X_{i:m}, X_{j:m})\) can be calculated for uniform distribution.

For \(k\)-out-of-\(n\) systems, the optimal prediction interval can be found easier. In this case, let us to take \(\varphi (\ell )\) as the form

$$\begin{aligned} \varphi (\ell )=\frac{n}{n+m} {{{m} \atopwithdelims (){\ell }} {{n-1} \atopwithdelims (){k-1}} \bigg / {{m+n-1} \atopwithdelims (){\ell +k-1}}}. \end{aligned}$$

Then, from the Eq. (6), the prediction coefficient of the interval \((X_{i:m}, X_{j:m})\) for a \(k\)-out-of-\(n\) system can be written as

$$\begin{aligned} \alpha _1(i, j; m, n, \mathbf s )=\sum \limits _{\ell =i}^{j-1} \varphi (\ell ). \end{aligned}$$

Assume that there exist \(\ell _0\) such that \(\varphi (\ell _0) \ge \varphi (\ell )\), for \(\ell = 1, \ldots , m\), then we can find the point \(\ell _0\) by solving the following inequalities:

$$\begin{aligned} \varphi (\ell _0) > \varphi (\ell _0+1)\quad {\mathrm{and }}\quad \varphi (\ell _0)>\varphi (\ell _0-1). \end{aligned}$$

After some algebraic calculations, we finally obtain

$$\begin{aligned} \ell _0 = \left[ \frac{(m+1)(k-1)}{n-1} \right] , \end{aligned}$$
(9)

where \([u]\) stands for the integer part of \(u\). Consequently, the steps of Algorithm 1 can be modified as:

Algorithm 2

For a given \(\alpha _0\), \(\mathbf s \), \(m\) and \(n\), the optimal prediction interval for \(k\)-out-of-\(n\) systems can be derived through the following steps:

  • Step 1: First, using (9) find \(\ell _0\).

  • Step 2: Set \(i=\ell _0\) and \(j= \ell _0+1\). If \(\alpha (\ell _0, \ell _0+1, m, n, \mathbf s ) \ge \alpha _0\), then consider the interval \((X_{\ell _0:m}, X_{\ell _0+1:m})\) as the first candidate for optimal prediction interval.

  • Step 3: If \(\varphi (\ell _0-1) > \varphi (\ell _0+1)\), then consider the interval \((X_{\ell _0-1:m}, X_{\ell _0+1:m})\), otherwise take the interval \((X_{\ell _0:m},\, X_{\ell _0+2:m})\).

  • Step 4: This procedure should be followed until for a fixed \((i_0=i-1, j= j_0)\) or \((i_0=i, j_0= j+1)\), the following inequalities are satisfied

    $$\begin{aligned} \alpha (i, j, m, n, \mathbf s ) < \alpha _0, \quad \alpha (i_0, j_0, m, n, \mathbf s ) \ge \alpha _0. \end{aligned}$$

By appealing Algorithm 2, we have obtained \((i_{opt}, j_{opt})\) and \(\alpha _{1}(i_{opt}, j_{opt}; m, n,\mathbf s )\), for some given selected values of \(\alpha _0\), \(m\), \(n\) and \(\mathbf s \). Table 1 contains the optimal prediction interval indices for some selected \(k\)-out-of-\(n\) systems based on the ordered failure times of \(m\) components. From Table 1, it is observed that the indices, \(i_{opt}\) and \(j_{opt}\) are increasing in \(m\).

Table 1 The optimal prediction intervals for \(k\)-out-of-\(n\) systems

In Fig. 1, we plot \(\varphi (\ell )\) for \(n=10\), \(k=4\) and \(m=10,15\) and 20. As shown in the Fig. 1, with increasing \(m\), while \(n\) and \(k\) are prefix, the maximum value of \(\varphi (\ell )\) decreases.

Fig. 1
figure 1

The values of \(\varphi (\ell )\), for \(n=10\), \(k\)=\(4\) and \(m= 10\), 15 and 20

Now, we find some optimal prediction intervals for a coherent system with five components with some selected signature vectors. Navarro and Rubio (2010) obtained all coherent systems with five components and computed their signature vectors. Table 2 shows the optimal prediction intervals for some coherent systems with five components based on order statistics.

Table 2 The optimal prediction intervals for coherent systems with five components

In Table 2, we consider three systems with structure functions \(\varphi _i(\mathbf x ), \; i=1, 2, 3\) as follows:

$$\begin{aligned} \varphi _1(\mathbf x )&= max\big \{min(x_1,x_2,x_3),\, min(x_1,x_2, x_4),\, min(x_1,x_2,x_5),\, \\&min(x_1,x_3, x_4, x_5)\big \},\\ \varphi _2(\mathbf x )&= max\big \{min(x_1,x_2),\, min(x_1, x_3),\, min(x_1, x_4),\, min(x_2, x_3, x_5),\,\\&min(x_2, x_4, x_5)\big \},\\ \varphi _3(\mathbf x )&= max\{x_1, min(x_2, x_3), min(x_2, x_4), min(x_2, x_5), min(x_3, x_4, x_5)\}, \end{aligned}$$

where for each \(i\), \(x_i=1\), if the \(i\)th component is working and \(x_i=0\), if it is not working. The signature vectors for these systems can be found as \(\mathbf s _1=(\frac{1}{5},\frac{1}{2},\frac{3}{10},0,0)\), \(\mathbf s _2=(0,\frac{1}{5},\frac{1}{2},\frac{3}{10},0)\) and \(\mathbf s _3=(0,0,\frac{3}{10},\frac{1}{2},\frac{1}{5})\), respectively.

Figure 2 also gives a graph of \(\psi (m, n, h, \mathbf s )\) for the signature vectors \(\mathbf s _1, \mathbf s _2\) and \(\mathbf s _3\) in the case \(m=15\) and \(n=5\), where

$$\begin{aligned} \psi (m, n, h, \mathbf s ) = \frac{n}{n+m}\sum \limits _{\ell =1}^{n}s_{\ell }{{m \atopwithdelims ()h}{{n-1} \atopwithdelims (){\ell -1}}}\bigg / {{{m+n-1} \atopwithdelims (){\ell +h-1}}}. \end{aligned}$$

Then, from (5), \(\alpha _1(i, j, m, n, \mathbf s )\) can be re-expressed as \(\alpha _1(i, j, m, n, \mathbf s )=\sum \nolimits _{h=i}^{j-1} \psi (m, n, h, \mathbf s ).\)

Fig. 2
figure 2

The values of \(\psi (m, n, h, \mathbf s _i)\), for \(i=1{,} 2{,} 3\)

Figure 2 shows that if \(\mathbf s \le _{st} \mathbf s ^*\) (here, ‘\(\le _{st}\)’ stands for stochastic ordering, we refer the reader to Shaked and Shanthikumar (2007), for more details on the stochastic orders), then \(h_0 \le h_0^*\), where \(h_0\) and \(h_0^*\) are the points that maximize \(\psi (m, n, h, \mathbf s )\) and \(\psi (m, n, h, \mathbf s ^*)\). Hence, it would be expected that for predicting the lifetime of a system with signature \(\mathbf s ^*\), we need the larger order statistics versus the case that the system signature is \(\mathbf s \). This is supported by the results of Kochar et al. (1999) in which, they proved that if \(\mathbf s \le _{st} \mathbf s ^*\), then \(T \le _{st} T^*\), where \(T\) and \(T^*\) are the lifetimes of the systems with signature vectors \(\mathbf s \) and \(\mathbf s ^*\), respectively. It should be noted that in Table 2 and also Fig. 2, \(\mathbf s _1 \le _{st} \mathbf s _2\) and \(\mathbf s _2\le _{st} \mathbf s _3\).

3 Prediction interval based on minimal repair times

The notion of minimal repair was introduced in reliability by Barlow and Hunter (1960). Its intuitive meaning is putting the system back to operation when it fails in such a way that the situation immediately preceding the failure is restored. Let \(X\) be the lifetime of an original system with continuous cdf \(F(x)\), when the system fails, minimal repair is done. Let \(T_{(i)}\) denote the lifetime of the system that \(i-1 (i\ge 1)\) minimal repairs are allowed, then

$$\begin{aligned} P\big (T_{(i)} > t\big ) = \big (1- F(t)\big )\sum \limits _{h = 0}^{i - 1} {\frac{{\big [ - \log ( 1- F(x))\big ]^h }}{{h!}}}, \end{aligned}$$
(10)

see for example Shaked and Shanthikumar (1994, p. 496), or Theorem 1 of Nakagawa and Kowada (1983).

In this section, we intend to construct prediction intervals for the lifetime of a future system with an arbitrary structure based on the observed minimal repair times. With this in mind, we consider two cases. In the first case, we just consider a sequence of minimal repair times of a component and a prediction interval for the lifetime of a future system with \(n\) components would be obtained. In the second case, a prediction interval for the lifetime of a future system would be found based on the observed minimal repair times of \(m\) components. To obtain this sample, \(m\) identical components begin to operate at time zero, separately. If each of the components fails, then it undergoes minimal repair and begins to operate again. For second scheme, we assume the components can be repaired only one time.

3.1 Case I

Let \(T\) be the lifetime of a system with component lifetimes \(Y_1, \ldots ,Y_n\) and arbitrary signature vector \(\mathbf s =(s_1, s_2, \dots , s_n)\). Also, suppose that \(X\) and \(Y_1, \ldots , Y_n\) have the same continuous distribution \(F\), then we have the next result.

Theorem 2

Let \(T^X_{(i)},\; i\ge 1 \) be the sequence of minimal repair times of component \(X\). Then \((T^X_{(i)},\, T^X_{(j)}), \, j>i \ge 1,\) is a two-sided prediction interval for \(T\) whose coverage probability is given by

$$\begin{aligned} \alpha _2(i, j; n, \mathbf s )&= \sum \limits _{h = 1}^n \sum \limits _{\ell = 0}^{h - 1} \frac{{ ( - 1)^\ell h {n \atopwithdelims ()h} {{h-1} \atopwithdelims ()\ell }}}{n - h + \ell + 1}s_h \nonumber \\&\times \left[ {\frac{1}{{(n - h + \ell + 2)^i }} - \frac{1}{{(n - h + \ell + 2)^j }}} \right] , \end{aligned}$$
(11)

where \(\mathbf s =(s_1, \ldots , s_n)\) with \(s_{h}= P(T=Y_{h:n})\) and \(\sum \nolimits _{h=1}^n s_{h}=1\).

Proof

Using (2), (10) and independence of \(T\) and \(\{T^X_{(i)}; i\ge 1\}\), we have

$$\begin{aligned}&P\big (T^X_{(i)} \le T \le T^X_{(j)} \big )\nonumber \\&\quad = \sum \limits _{r = i}^{j - 1} \sum \limits _{h = 1}^n h \Bigg (\begin{array}{c}n\\ h\end{array}\Bigg ) s_h \int \limits _{0}^{1} \frac{{[-\log y]}^r}{r!}y^{n - h + 1} \big [{1-y}\big ]^{h - 1} dy \nonumber \\&\quad = \sum \limits _{r = i}^{j-1} \sum \limits _{h = 1}^n \sum \limits _{\ell = 0}^{h - 1} h \Bigg (\begin{array}{c} n\\ h\end{array}\Bigg ) s_h \Bigg (\begin{array}{c} {h-1} \\ \ell \end{array}\Bigg ) (-1)^\ell \int \limits _{0}^{1} \frac{{(-\log y)}^r}{r!} y^{n-h+\ell +1}dy\nonumber \\&\quad = \sum \limits _{r = i}^{j - 1} \sum \limits _{h = 1}^n \sum \limits _{\ell = 0}^{h - 1} h \Bigg (\begin{array}{c} n\\ h\end{array}\Bigg ) s_h \Bigg (\begin{array}{c} {h-1}\\ \ell \end{array}\Bigg ) \frac{(-1)^\ell }{\big (n-h+\ell +2\big )^{r+1}}. \end{aligned}$$
(12)

By simplifying the expression in the right-hand side of (12), the required result follows. \(\square \)

For a \(k\)-out-of-\(n\) system, the expression in the right-hand side of (11) reduces to

$$\begin{aligned} k {n \atopwithdelims ()k}\sum \limits _{\ell = 0}^{k-1} \frac{( - 1)^\ell {{k-1} \atopwithdelims ()\ell } }{n-k+\ell +1} \left[ {\frac{1}{{\big (n - k + \ell + 2 \big )^i }} - \frac{1}{{\big (n - k + \ell + 2 \big )^j }}} \right] . \end{aligned}$$

For a given prediction level \(\alpha _0\), signature vector \(\mathbf s \) and \(n\), we can choose \(i\) and \(j\) such that \(\alpha _2(i, j; n, \mathbf s )\) exceeds \(\alpha _0\). Notice that, the two-sided prediction interval \((T^X_{(i)},\,T^X_{(j)}),\, 1 \le i< j \le m,\) exists if and only if, for large values of \(m\), \(P(T^X_{(1)} \le T \le T^X_{(m)}) \ge \alpha _0\). In other word, we should have

$$\begin{aligned} \mathop {\max }\limits _{i,j} \alpha _2 (i, j; n, \mathbf s )=\sum \limits _{h=1}^n\frac{h}{n+1} s_h\ge \alpha _0. \end{aligned}$$
(13)

For a \(k\)-out-of-\(n\) system, (13) reduces to \(\frac{k}{n+1} \ge \alpha _0 \).

3.2 Case II

Consider the situation that we have some information about a sample of components with size \(m\), produced by a factory. It is assumed that the components begin to operate separately. If each of the components fails, it undergoes minimal repair and begins to operate again. The components can be repaired \(\tau -1\) times (for simplicity we consider the case \(\tau =2\)). Thus, we have a sequence of \(\tau \) failure times for each components (i. e., \(m \times \tau \) observed failure times for \(m\) components). We use these information to construct a prediction interval for the future lifetime of a coherent system composed of the \(n\) components with the same distribution. First, we present the following theorem that will be used to prove the new results in this section.

Theorem 3

Let \(\mathbf X _1, \ldots , \mathbf X _m\) be \(m\) \(i. i. d.\) samples of multivariate continuous random variables such that \(\mathbf X _i=(X_{i,1}, \ldots , X_{i,\tau }), \; 1 \le i \le m\) and \(X_{i,j} \le X_{i, \ell }\) with probability 1, for \(i= 1, \ldots , m,\; 1\le j < \ell \le \tau \), also \(X_{s,0}\equiv -\infty \) and \(X_{s,\tau +1}\equiv +\infty \). Suppose that the ordered values of \(X_{i,j}, \; 1 \le i \le m, \; 1 \le j \le \tau \), are denoted by \(Z_{1:m\tau },\, Z_{2:m\tau }, \ldots , Z_{m\tau :m\tau }\). Then, the marginal cdf of \(Z_{i:m\tau }\), the \(i\)th order statistic of the pooled sample, is given by

$$\begin{aligned} Pr\big (Z_{i:m\tau } \le x \big )&= \sum \limits _{r = i}^{m\tau } \sum \limits _{{h_0} = \max \big \{ 0, r-(\tau - 1)m \big \} }^{[r/\tau ]} \cdots \sum \limits _{{h_i} = \max \big \{ 0,r - \sum \nolimits _{\ell = 0}^{i - 1} {\big (i + 1 - \ell \big ){h_\ell } - m\big (\tau - i - 1 \big )} \big \} }^{\big [(r - \sum \nolimits _{\ell = 0}^{i - 1} {(\tau - \ell ){h_\ell }} )/(\tau - i)\big ]}\nonumber \\&\cdots \sum \limits _{{h_{\tau - 2}} = \max \big \{ 0,r - m - \sum \nolimits _{\ell = 0}^{\tau - 3} {(\tau - \ell - 1){h_\ell }} \big \} }^{\big [(r - \sum \nolimits _{\ell = 0}^{\tau - 3} {(\tau - \ell ){h_\ell }} )/2\big ]} {\sum \limits _{{A_{{h_0},{h_1}, \ldots , h_{\tau -2}, r - \sum \nolimits _{\ell = 0}^{\tau - 2} {(\tau - \ell ){h_\ell },m} }}} { \mathop \prod \limits _{s = 0}^{\tau } {P_s}}},\nonumber \\ \end{aligned}$$
(14)

with

$$\begin{aligned} {P_s}&= \mathop \prod \limits _{j = \sum \nolimits _{\ell = 0}^{s - 1} {{h_\ell } + 1} }^{\sum \nolimits _{\ell = 0}^s {{h_\ell }} } \Pr \big ({X_{{t_j},\tau - s}} \le x,{X_{{t_j},\tau - s + 1}} > x \big ), \quad 0 \le s \le \tau - 2, \\ {P_{\tau - 1}}&= \mathop \prod \limits _{j = \sum \nolimits _{\ell = 0}^{\tau - 2} {{h_\ell } + 1} }^{r - \sum \nolimits _{\ell = 0}^{\tau - 2} {(\tau - \ell - 1){h_\ell }} } \Pr \big ({X_{{t_j},1}} \le x,{X_{{t_j},2}} > x \big ),\\ {P_\tau }&= \mathop \prod \limits _{j = r - \sum \nolimits _{\ell = 0}^{\tau - 2} {(\tau - \ell - 1){h_\ell }} + 1}^m \Pr \big ({X_{{t_j},1}} > x \big ), \end{aligned}$$

where \([u]\) stands for the integer part of \(u\) and \(A_{i_1,\ldots ,i_h,m}\) extends over all permutations of \((t_1,\ldots ,t_m)\) from \(\{1, \ldots , m\}\) such that \(t_1< \cdots <t_{i_1},\) \(t_{i_1+1}< \cdots < t_{i_1+i_2}, \ldots ,t_{\sum \nolimits _{j=1}^{h}i_j+1} <\cdots <t_m\).

Proof

We present the proof for the case \(\tau =3\) and the other cases can be treated in analogous way. For simplicity, let \(\mathbf X _i =(X_i, Y_i, W_i), \ i=1,\ldots , m\). Thus, there are \(3m\) statistics as \(X_1\le Y_1\le W_1, X_2\le Y_2\le W_2, \ldots , X_m\le Y_m\le W_m\) which are extracted from \(m\) independent random samples. Let \(Z_{i:3m}\) denote the \(i\)th order statistic of the pooled sample. The marginal cdf of \(Z_{i:3m}\) can be expressed as

$$\begin{aligned} Pr\big (Z_{i:3m} \le x \big )= \sum \limits _{r=i}^{3m} {\eta }_{m} (r, x), \end{aligned}$$
(15)

where \(\eta _m (r, x)= Pr(\)exactly \(r\) elements of the pooled sample are at most \(x)\).

Now, we derive an explicit expression for \(\eta _m (r, x)\). Consider four events \(A= \{ W_{t_s} \le x\}, \ B= \{ Y_{t_s} \le x, W_{t_s} > x \}, C=\{X_{t_s} \le x, Y_{t_s}> x \}\) and \(D=\{ X_{t_s} > x\}\), \(s=1, \ldots , m\). For arranging the pooled sample such that exactly \(r\) elements of the sample be at most \(x\), we should determine the number of times that the events \(A, B, C\) and \(D\) occur. Let \(j\) and \(h\) be the number of times that the events \(A\) and \(B\) occur, respectively, such that \(max\{0, r-2m\} \le j \le [r/3]\) and \(max\{0, r-2j-m\} \le h \le [(r-3j)/2]\) (Notice that if \(j\) events of \(A\) are occurred, then the number of statistics that are less than \(x\) is \(3j\), because \(X_{t_s}<Y_{t_s}<W_{t_s}\)). Therefore \(3j+2h\) elements of the pooled sample are at most \(x\) and \(h\) elements are at least \(x\). Thus, we need exactly \(r-3j-2h\) of \(X_i, \; i=1, \ldots , m\) to be at most \(x\) and exactly \(m-r+h+2j\) of them should be at least \(x\). With these assumptions, we will have exactly \(r\) elements less than \(x\) and \(m-r\) elements greater than \(x\). The number of cases that we can select \(j\) of \(A\), \(h\) of \(B\), \(r-3j-2h\) of \(C\) and \(m-r+h+2j\) of \(D\) is \(A_{j, h, r-3j-2h, m}\). Hence, the marginal cdf of \(Z_{i:3m}\) can be expressed as

$$\begin{aligned} Pr\big (Z_{i:3m} \le x \big )= \sum \limits _{r = i}^{3m} {\sum \limits _{j = \max \{ 0,r - 2m\} }^{[r/3]} {\sum \limits _{h = \max \{ 0,r - 2j - m\} }^{[(r - 3j)/2]} {\sum \limits _{{A_{j,h,r - 2h - 3j,m}}} {\prod \limits _{s = 0}^3 {{P_s}} } } } }, \end{aligned}$$

where \(P_0 = \prod \limits _{s = 1}^j Pr(W_{t_s} \le x)\), \(P_1=\prod \limits _{s=j+1}^{h+j}Pr(Y_{t_s} \le x, W_{t_s} > x)\), \(P_2 = \prod \limits _{s=h+j+1}^{r-h-2j} Pr(X_{t_s} \le x, Y_{t_s} > x)\) and \(P_3=\prod \limits _{s= r-h-2j+1}^{m} Pr(X_{t_s} > x)\).

For \(m>3\), the proof is similar and we will begin the sorting of the pooled sample based on the largest order statistic in each independent sample. \(\square \)

Theorem 3 is useful for constructing prediction intervals for the lifetime of a future coherent system based on the observed failure times, when minimal repair at failures is considered. In order to simplify the calculations, we consider the case in which every component is allowed to have one minimal repair. Suppose that obtaining process of minimal repair times from cdf \(F\) is repeated for \(m\) independent and identical components such that for each components, we are allowed to do one minimal repair. Thus, the observed data set is as follows:

$$\begin{aligned} \begin{array}{rcl} {\mathrm{Sample}} 1: &{} {T_{1,(1)}},&{}{T_{1,(2)}}\\ {\mathrm{Sample}} 2: &{}{T_{2,(1)}},&{}{T_{2,(2)}}\\ \vdots &{} \vdots &{} \\ {\mathrm{Sample}} ~m: &{}{T_{m,(1)}},&{}{T_{m,(2)}}, \end{array} \end{aligned}$$

where \(T_{i,(j)},~i=1,\ldots , m;~j=1, 2 \) is the \(j\)th failure time of the \(i\)th component. One can construct a prediction interval for the lifetime of a future coherent system based on \(T_{i,(j)},~i=1,\ldots , m;~j=1, 2 \) which is stated in the next result.

Theorem 4

Let \((T_{1,(1)}, T_{1,(2)}), \ldots , (T_{m,(1)}, T_{m,(2)})\) be corresponding \(2m\) failure times of \(m\) components with cdf \(F\), and \(T_{(1)}^*,\ldots , T_{(2m)}^*\) be the order statistics of the pooled sample. Let us denote by \(T\) the lifetime of a coherent system with \(n\) components from cdf \(F\). Then, \((T^*_{(i)}, T^*_{(j)}), \; 1 \le i < j \le 2m\) is a two-sided prediction interval that its coverage probability is given by

$$\begin{aligned} \alpha _3\big (i, j; m, n, \mathbf s )&= \sum \limits _{r = i}^{j - 1} {\sum \limits _{t =\max \{0, r-m\}}^{[ {\frac{{r }}{2}}] } {\sum \limits _{h = 0}^{t} {\sum \limits _{v = 1}^n {\sum \limits _{\ell = 0}^{t - h + v - 1}v }}}}{\ s_v}{n \atopwithdelims ()v}{{t} \atopwithdelims ()h}{{t-h+v-1} \atopwithdelims ()\ell }\nonumber \\&\quad \times \frac{ {C_m}(r,r-t){{( - 1)^{\ell + h}} (r-2t + h)!}}{{{{(\ell + h - t + m + n - v + 1)}^{r-2t + h + 1}}}}, \end{aligned}$$
(16)

where \({C_m}(r,r-t)=\frac{m!}{t!(r-2t)!(m-r+t)!}\), and \(\mathbf s =(s_1, \ldots , s_n)\) with \(s_{v}= P(T=Y_{v:n})\) and \(\sum _{v=1}^n s_{v}=1\).

Proof

By using Theorem 3 for \(\tau =2\), the cdf of \(T^*_{(i)}\) is given by

$$\begin{aligned} Pr \big (T_{(i)}^* \le x \big ) = \sum \limits _{r =i}^{2m} {\sum \limits _{j = \max \{ 0,r - m\} }^{[r/2]}{\sum \limits _{{A_{j,r - 2j,m}}}}} \prod \limits _{s=0}^{2} P_s, \end{aligned}$$
(17)

where \(P_0=\prod \limits _{s = 1}^j Pr (T_{t_s,(2)} \le x) \), \(P_1=\prod \limits _{s = j + 1}^{r - j} Pr (T_{t_s,(1)} \le x,T_{t_s,(2)}> x)\) and \(P_2=\prod \limits _{s = r - j + 1}^m Pr(T_{t_s,(1)}> x)\). It should be mentioned that with slight modification, (17) deduces to the expression (8) in Ahmadi and Razmkhah (2007).

Here, \(T_{t_s,(1)}\) and \(T_{t_{s},(2)}\) for \(s=1, \ldots , m\) have the same distribution with \(T_{1,(1)}\) and \(T_{1,(2)}\) (the first and second upper records), respectively. Thus, by using (10) and the joint pdf of the first and second upper records (for more details on the theory and applications of record values, we refer the reader to Arnold et al. (1988)) we have

$$\begin{aligned} Pr\big (T_{1,(1)} \le x \big ) = F(x), \end{aligned}$$
(18)
$$\begin{aligned} Pr\big (T_{1, (2)} \le x \big ) = F(x)+\bar{F}(x) \log (\bar{F}(x)), \end{aligned}$$
(19)
$$\begin{aligned} P\big (T_{1, (1)} \le x, T_{1, (2)} > x\big )= -\bar{F}(x) \log (\bar{F}(x)). \end{aligned}$$
(20)

Upon substitution the Eqs. (18), (19) and (20) into (17), the marginal cdf of the \(i\)th order statistic from the pooled sample can be found. After some manipulations the proof would be completed. \(\square \)

For given \(m\), \(n\) and signature vector \(\mathbf s \), the coverage probability \(\alpha _3(i, j; m, n, \mathbf {s}) \) is decreasing in \(i\) and increasing in \(j\). Consequently, we have

$$\begin{aligned} \mathop {\max }\limits _{i,j} \alpha _3 (i, j; m, n, \mathbf s )&= Pr\left( \mathop {\min }\limits _{1 \le i \le m} \big \{T_{i,(1)}\big \} \le T \le \mathop {\max }\limits _{1 \le i \le m}\big \{T_{i,(2)}\big \}\right) \\&= 1- \int \limits _{0}^{\infty }\left[ {\big (\bar{F}_{T_{1,(1)}}(t)\big )^m }+{\big (\bar{F}_{T_{1,(2)}}(t)\big )^m } \right] f_T(t) dt, \end{aligned}$$

where \(\bar{F}_{T_{1,(1)}}(t)\) and \(\bar{F}_{T_{1,(2)}}(t)\) (the survival functions of \(T_{1,(1)}\) and \(T_{1,(2)}\)) can be derived from (18) and (19), respectively.

For \(m=10\), \(n=5\) and some selected signature vectors, we have computed the maximum coverage probabilities of the prediction intervals \([T^X_{(i)},\, T^X_{(j)}]\) and \([T^*_{(i)},\,T^*_{(j)}]\), \(1\le i < j \le 10\), for the lifetime of a future coherent system with five components. These are presented in Table 3.

Table 3 The maximum coverage probabilities of the prediction intervals \([T^X_{(i)},\, T^X_{(j)}]\) and \([T^*_{(i)},\,T^*_{(j)}]\), \(1 \le i < j \le 10\), for a coherent system with five components

From Table 3, it is observed that in the most cases \(\mathop {\max }\limits _{i,j} \alpha _3 (i, j; m, n, \mathbf s )\) is greater than \(\mathop {\max }\limits _{i,j} \alpha _2 (i, j; n,\mathbf s )\) and they are almost close to each other, when the lifetime of a coherent system is equal to a larger order statistic.

It may be noted that by using Algorithm 1, we can find the optimal prediction intervals by similar way as in Sect. 2. It would be enough to compute the length of the prediction intervals from uniform distribution and compare them for different \(i\) and \(j\).