Abstract
The study deals with the analysis of Type-II hybrid censored data from the modified Weibull distribution. We provide maximum-likelihood estimates of the parameters, reliability, and hazard rate functions along with their standard errors. The confidence intervals along with their widths have also been obtained. Assuming gamma and Jeffrey’s invariant priors for the unknown parameters, Bayes estimates along with its posterior errors and highest posterior density credible intervals are obtained. The Markov Chain Monte Carlo technique has been used to simulate draws from the complicated posterior densities of the parameters. A simulation study is conducted to compare the performances of classical and Bayesian methods of estimation. Finally, a real data analysis is performed for illustrative purpose.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
In reliability and life testing experiments, the equipments are put on test and their failure times are recorded. These failure time observations are then used to draw inferences on various system reliability characteristics. However, due to long life times of today’s products, obtaining times-to-failure data is very time consuming which increases the cost of experimentation as well. To overcome this difficulty, censoring is used in life testing experiments to save the time and cost. Failure censoring schemes are broadly classified as Type-I (time censoring) and Type-II (failure censoring). Suppose, n items are put on test. In Type-I censoring, we fix the time T to terminate the experiment in advanced and observe the life times of those items which fail up to time T, whereas in Type-II censoring scheme, time to terminate the experiment is a random variable and required number of failed items to stop the test is fixed, say m (\(m \le n\)), so that at the time of the mth failure, the test terminates leaving n − m partially observed failure times. Type-I and Type-II censoring schemes have been extensively studied by numerous authors, including Mann et al. (1974), Lawless (1982), Balakrishnan and Cohen (1991), and Harter et al. (1996).
Since major constraints in life testing experiments are time and cost, therefore, we need a censoring scheme which can make a trade-off between the number of units used and the time required to stop the experimentation without sacrificing the desired efficiency of the statistical inference. Hybrid censoring is one such censoring scheme introduced by Epstein (1954, 1960). The hybrid censoring is also categorized as Type-I and Type-II. If the test is terminated at the random time T * = min{X R:n , T}, where R and T are prefixed numbers, and \(X_{R:n}\) indicates the time of Rth failure in a sample of size n. Then, it is called Type-I hybrid censoring scheme. Many studies, including Kundu (2007), Kundu and Pradhan (2009), Dube et al. (2011), Ganguly et al. (2012), Rastogi and Tripathi (2013), and Gupta and Singh (2013), dealt with statistical inference for different life time distributions using Type-I hybrid censoring scheme. On the other hand, if the test is stopped at time T ** = max{X R:n , T}, it is known as Type-II hybrid censoring scheme (Childs et al. 2003). With Type-II hybrid censoring scheme, one has the advantage to record the complete life times of at least R units before the experiment is terminated. For more details regarding statistical inferences under Type-II hybrid censoring scheme, one may refer Banerjee and Kundu (2008), Al-Zahrani and Gindwan (2014), and Singh et al. (2014). The detail review of various hybrid censoring schemes is given in Balakrishnan and Kundu (2013). Though many distributions have been considered for drawing inferences with Type-I and Type-II hybrid censored data, to the best of our knowledge, none of the study has reported the inferential statistics on the modified Weibull (MW) distribution under Type-II hybrid censoring scheme. Initially, this distribution was proposed by Lai et al. (2003), and has the following probability density function (PDF):
The distribution function (DF) of MW distribution is written as:
The corresponding reliability and hazard rate functions are as follows:
The shape of hazard rate function h(x) of MW distribution depends on the parameter \(\nu\). For \(\nu \ge 1\), h(x) is increasing in x, whereas for \(0 < \nu < 1\), it has bathtub shape. Due to the flexible shape of its hazard rate function, MW distribution has become important life time model for reliability engineers.
In lieu of above considerations, we explore the inferential properties of the modified Weibull distribution with Type-II hybrid censoring data. We provide maximum-likelihood estimators of the parameters, reliability, and hazard functions of MW distribution with Type-II hybrid censored data. Fisher information matrix has been given to construct confidence intervals of the parameters as well as reliability and hazard functions. Bayes estimates and highest posterior density (HPD) intervals of the parameters have been obtained using gamma and Jeffrey’s priors. Markov Chain Monte Carlo (MCMC) technique has been used to generate draws from complex posterior densities of the parameters. The coverage probabilities of confidence and HPD intervals have also been provided. To illustrate the application of MW distribution, a real data analysis is presented. A simulation study is also carried out to access the performances of ML and Bayes methods of estimation. At the end, some concluding remarks are given.
2 Estimation and Confidence Intervals under Type-II Hybrid Censoring Scheme
2.1 Maximum-Likelihood Estimators
Let n units are put on test. Then, under the Type-II hybrid censoring scheme, we have one of the following three types of observations:
Case I: \(\left\{ {x_{1:n} < x_{2:n} < \cdots < x_{R:n}} \right\}\quad if\;x_{R:n}>T.\)
Case II: \(\left\{ {x_{1:\,n} < x_{2:\,n} < \cdots < x_{R:\,n} < x_{R + 1:\,n} < \cdots < x_{k:\,n} < T < x_{k + 1:\,n} } \right\}\;\) if \(R \le k < n\), and \(x_{k:n} < T < x_{k + 1:n}\)
Case III: \(\left\{ {x_{1:\,n} < x_{2:\,n} < \cdots < x_{{n:\text{ }n}} < T} \right\}.\)
Graphically, it can be presented as shown in Fig. 1:
The likelihood functions for the above three different cases are as follows:
Case I:
Case II:
Case III:
On combining three likelihood functions, one gets
Here, D denotes the number of failure, that is
and
The log-likelihood function is
The first derivatives of (8) with respect to β, ν, and λ are as follows:
Therefore, for fixed ν and λ, the MLE of β say \(\hat{\beta }\) can be obtained as:
The MLE of ν and λ can be obtained by solving the following non-linear equations:
and
The MLE of β say \(\hat{\beta }\) can be obtained by Eq. (12), but Eqs. (13) and (14) are very complicated, so they cannot be expressed explicitly. Thus, some suitable iterative method is required to get \((\hat{\nu },\,\hat{\lambda })\). Now, using the general theory of MLEs, the asymptotic distribution of \(\left( {\hat{\beta } - \beta \,\,\,\hat{\nu } - \nu \,\,\,\hat{\lambda } - \lambda } \right)^{\prime }\) is \(N_{3} (0,\,\varSigma^{ - 1} )\). Where \(\Delta\) is the Fisher’s information matrix whose elements are as follows:
The second derivative with respect to β, ν, and λ are as:
Using the invariance property of maximum-likelihood estimates, we get the MLEs of the reliability and hazard rate functions as follows:
The asymptotic sampling distributions of \(\left[ {\hat{R}(x) - R(x)} \right]\) and \(\left[ {\hat{h}(x) - h(x)} \right]\) are \(N(0,R^{\prime}\sum^{ - 1} R)\) and \(N(0,h^{\prime}\sum^{ - 1} h)\), where \(R^{\prime} = \left( {\frac{\partial R(x)}{\partial \beta },\,\,\frac{\partial R(x)}{\partial \nu },\,\,\frac{\partial R(x)}{\partial \lambda }} \right)\) and \(h^{\prime} = \left( {\frac{\partial h(x)}{\partial \beta },\,\,\frac{\partial h(x)}{\partial \nu },\,\,\frac{\partial h(x)}{\partial \lambda }} \right)\).
2.2 Bayesian Estimation
In Bayesian inference, one of the tedious tasks is how to construct prior models for the unknown parameters as there is no unique approach of choosing a priori, and that the choice of a wrong priori may lead to the inappropriate inference. The prior distribution is the key to Bayesian inference and its determination is, therefore, the most important step in drawing the inference (Robert 2007). In practice, informative and non-informative priors are used to represent uncertainties about the model parameters. Berger (1985) pointed out that when there is no information or very difficult to gather regarding the prior variations in the parameters, it is better to use non-informative prior distribution. However, non-informative priors generally lack invariance property under one-to-one transformation, thereby leading to incoherent analysis. On the other hand, informative priors are based on the investigator’s experience about the random behavior of the process under consideration. In lieu of this, we consider the Bayesian method of estimation with both informative and non-informative priors. First, we assume that the parameters β, ν, and λ have \({\text{Gamma}}(a_{1} ,b_{1} )\), \({\text{Gamma}}(a_{2} ,b_{2} )\), and \({\text{Gamma}}(a_{3} ,b_{3} )\) priors, respectively, with PDFs:
Based on the above priors, the joint distribution of data and parameters β, ν, and λ is given by,
For drawing Bayesian inference, we need joint posterior distribution of the parameters β, ν, and λ which are very difficult to compute analytically due to multidimensional parameter space. Therefore, Gibbs sampler proposed by Geman and Geman (1984) is used for this purpose. One important advantage with Gibbs sampler is that we only require full conditional posterior distributions of each of the parameters. The full conditional posterior of parameters β, ν, and λ are given by the following:
2.2.1 Gibbs Algorithm
-
1.
Set starting values for \(\nu\) and \(\lambda\), and generate \(\beta\) from the conditional density \(W(\beta \left| {\underset{\raise0.3em\hbox{$\smash{\scriptscriptstyle\thicksim}$}}{x} ,\nu ,\lambda } \right.)\) given in (22).
-
2.
Generate \(\nu\) from the conditional density \(W(\nu \left| {\underset{\raise0.3em\hbox{$\smash{\scriptscriptstyle\thicksim}$}}{x} ,\beta ,\lambda } \right.)\) in (23) for the above given simulated value of \(\beta\) and starting value of \(\lambda\).
-
3.
Generate \(\lambda\) from the conditional density \(W(\lambda \left| {\underset{\raise0.3em\hbox{$\smash{\scriptscriptstyle\thicksim}$}}{x} ,\nu ,\beta } \right.)\) in (24) for the above simulated values of \(\beta\) and \(\nu\).
-
4.
Repeat steps 1–3 N times and stored the generated draws of \(\,\beta ,\,\nu ,\,{\text{and}}\;\lambda\) after the first M-iterations to nullify the effect of the starting values.
-
5.
Bayes estimates of the parameters \(\,\beta ,\,\nu ,\,{\text{and}}\;\lambda\) are then given by the following:
$$\,\,\beta^{*} = \frac{1}{N - M}\sum\limits_{i = M + 1}^{N} {\beta_{i} ,\,\,\nu^{*} = \frac{1}{N - M}\sum\limits_{i = M + 1}^{N} {\nu_{i} } \,\,} {\text{and}}\,\,\lambda^{*} = \frac{1}{N - M}\sum\limits_{i = M + 1}^{N} {\lambda_{i} } .$$ -
6.
HPD intervals for \(\,\beta ,\,\nu \;{\text{and}}\;\lambda\) are obtained using the method proposed by Chen and Shao (1999).
-
7.
For computing Bayes estimates and HPD intervals of reliability and hazard functions, we first substituted the generated values of \(\,\beta ,\,\nu ,\,{\text{and}}\;\lambda\) in (2) and (3) and then used the procedures as in steps 5 and 6, respectively.
The posterior results under Jeffrey’s priors can be obtained in the same way by setting all gamma priors’ parameters equal to zero i.e., \(a_{1} = b_{1} = 0,\) \(a_{2} = b_{2} = 0\) and \(a_{3} = b_{3} = 0\). Here, it is to be noted that the simulation in steps 2 and 3 is not easy as the inverse distribution function method of generating draws from the posterior densities of \(\nu\) and \(\lambda\) is not applicable here. Therefore, for simulating parametric draws from these densities, one needs an advanced simulating algorithm. Therefore, we utilize Metropolis–Hastings (MH) algorithm (Metropolis and Ulam 1949; Hastings 1970), one of the MCMC techniques to generate \(\nu\) and \(\lambda\).
2.2.2 MH Algorithm
Suppose we want to simulate draws \((\theta_{1} ,\,\theta_{2} , \ldots \theta_{N} )\) from the distribution \(g\left( \theta \right)\). Given the target density \(g\left( \theta \right)\) and an arbitrary proposal density or jumping distribution \(q(\theta |\theta^{\prime}) = P(\theta^{\prime} \to \theta )\), i.e., the probability of returning a new sample value \(\theta\) given a previous sample value \(\theta^{'}\), the algorithm is as follows:
-
1.
Set initial value θ° satisfying target density \(g(\theta^{ \circ } ) > 0\).
-
2.
For t = 1,2,…N, repeat the following steps.
-
3.
Set \(\theta = \theta^{(t - 1)} .\)
-
4.
Using current \(\theta\) value, sample a point \(\theta^{*} \sim \,\,q(\theta^{*} |\theta^{(t - 1)} )\). Here, we assume normal distribution as the proposal density that is \(q(\theta^{*} |\theta^{(t - 1)} ) \equiv N(\theta^{(t - 1)} ,\sigma_{\theta } )\). The standard deviation \(\sigma_{\theta }\) is chosen, so that the chain explores the whole area of the target density with sufficient acceptance probability.
-
5.
Compute the acceptance probability \(\rho (\theta^{*} ,\theta^{(t - 1)} ) = \hbox{min} \left\{ {1,\,\log \left( {\frac{{g(\theta^{*} )q(\theta^{(t - 1)} |\theta^{*} )}}{{g(\theta^{t - 1} )q(\theta^{*} |\theta^{(t - 1)} )}}} \right)} \right\}.\)
-
6.
Generate u from \(U(0, 1)\) and take \(z = { \log }u\).
-
7.
If \(z < \rho (\theta^{*} ,\theta^{t - 1} )\), accept \(\theta^{*}\) and set \(\theta^{(t)} = \theta^{*}\) with probability \(\rho (\theta^{*} ,\theta^{t - 1} )\). Otherwise, reject \(\theta^{*}\) and set \(\theta^{(t)} = \theta^{(t - 1)}\).
3 Real Data Analysis
In this section, an analysis of a real data set is performed. The data set is taken from Lawless (2003), which contains 60 observations on electrical appliance failure times (1000s of cycles) as follows:
0.014, 0.034, 0.059, 0.061, 0.069, 0.08, 0.123, 0.142, 0.165, 0.21, 0.381, 0.464, 0.479, 0.556, 0.574, 0.839, 0.917, 0.969, 0.991, 1.064, 1.088, 1.091, 1.174, 1.270, 1.275, 1.355, 1.397, 1.477, 1.578, 1.649, 1.707, 1.893, 1.932, 2.001, 2.161, 2.292, 2.326, 2.337, 2.628, 2.785, 2.811, 2.886, 2.993, 3.122, 3.248, 3.715, 3.79, 3.857, 3.912, 4.1, 4.106, 4.116, 4.315, 4.510, 4.584, 5.267, 5.299, 5.583, 6.065, 9.701.
First, we compare the fitting of three-parameter MW distribution with some two-parameter distributions, such as Weibull, gamma, and log-normal to this data set. The Kolmogorov–Smirnov (K–S) test, log-likelihood criterion, and Akaike information criterion (AIC) are applied for this purpose. For complete data set, the MLEs of the parameters and fitting summaries of considered models are given in Table 1. While obtaining the MLEs of the three-parameter MW distribution, we found the MLE of \(\beta\) in closed form, whereas MLEs of the other two parameters \(\nu\) and \(\lambda\) are obtained using the optimization function maxLik() of R-software. To check the convergence of this algorithm, the contour plot of the two-dimensional log-likelihood surface for \(\nu\) and \(\lambda\) corresponding to the considered data set has been drawn in Fig. 2a. The surface is well behaved with a unique maximum. The MLEs of \(\nu\) and \(\lambda\) are shown in the contour plot. The difference in curvature in the log-likelihood surface along with \(\nu\) and \(\lambda\) directions can be easily observed. We also draw the individual profile plots of log-likelihood function for \(\nu\) and \(\lambda\) in Fig. 2b, c, respectively. The values of K–S statistics and associated P values clearly indicate that the MW distribution is a better model for the given data. The values of the log-likelihood and AIC also suggested the same. The fitted density plots and the P–P plots of Kaplan–Meier estimator (KME) (Kaplan and Meier 1958) versus fitted survival functions of the considered models are displayed in Fig. 3a and b. From these plots, it can be seen that the MW distribution is superior to the other distributions in terms of model fitting. More so, the plots of the cumulative hazard function superimposed on nonparametric counterpart given in Fig. 3c confirm that MW distribution is an adequate model for considered data set.
Now for analyzing this data with MW distribution under Type-II hybrid censoring scheme, three artificially hybrid censored data sets are formed from the complete data with the following censoring schemes:
Scheme 1: R = 48, T = 4 (20 % censored data).
Scheme 2: R = 36, T = 4 (40 % censored data).
Scheme 3: R = 24, T = 1.5 (60 % censored data).
In all the cases, the unknown parameters are estimated using the ML and Bayes methods of estimation. Bayes estimates of \(\beta ,\,\nu ,\;{\text{and}}\;\lambda\), and HPD intervals are obtained using gamma and Jeffrey priors. The different estimates of the parameters along with their standard errors and 95 % confidence/HPD intervals are summarized in Table 2.
4 Comparison Study
Here, some simulation results for accessing the performances of the classical and Bayesian methods of estimation under various Type-II hybrid censoring schemes for different choices of (R, T) are presented. The comparisons are made on the basis of the average standard errors/posterior errors (ASEs/APEs) of the estimates, average interval lengths (AILs) of the confidence/HPD intervals, and coverage probabilities (CPs). Assuming \(\beta = 0.5,\,\nu = 2,\,\,\lambda = 0.2\), two sets of data containing, respectively, n = 30 and 100 observations were generated from (1). We replicate the process 1000 times and obtain average estimates, ASEs/APEs, average confidence/HPD intervals, average interval lengths, and coverage probabilities with different combinations of R and T. For Bayesian estimation, 5000 realizations of the parameters \(\beta ,\,\,\nu ,\,\,{\text{and}}\;\lambda\) from the posterior densities in (22), (23), and (24) are generated using Gibbs sampler. While simulating posterior densities of \(\nu\) and \(\lambda\), we use MH algorithm as discussed in Sect. 2.2. In this algorithm, the initial guess values of the parameters \(\nu\) and \(\lambda\) are taken as their MLEs. The first 1000 burn-in iterations have been discarded to eliminate the effect of the stating values of the parameters. The graphical diagnostic tools, such as trace and autocorrelation plots, have been employed to check the convergence of the chains. The trace plots of \(\beta ,\,\,\nu ,\,\,\text{and}\;\lambda\) in Figs. 4a, 5a, and 6a, respectively show well mixing of the chains. More so, the chains are scattered around their mean values. The autocorrelation plots of \(\beta ,\,\,\nu ,\,\,\text{and}\;\lambda\) in Figs. 4b, 5b, and 6b clearly revel that the chains have low autocorrelations. Finally, we have drawn the simulated posterior densities of \(\beta ,\,\,\nu ,\,\,\text{and}\;\lambda\) in Figs. 4c, 5c, and 6c, and we found that all posterior distributions are unimodal. The results of the comparison study have been summarized in Tables 3, 4, 5, 6, 7, and 8. The results for reliability and hazard functions for varying values of the mission time x are listed in Tables 9 and 10, respectively. Note that, in Tables 2, 3, 4, 5, 6, 7, and 8, the entries in the bracket [] represent ASEs/APEs and those in the brackets () and {}, respectively, represent average confidence/HPD intervals and their lengths. For all the numerical computations, the programs are developed in the R-software. From the results given in Tables 3, 4, 5, 6, 7, 8, 9, and 10, the following is observed:
-
The ML estimates in comparison to Bayes are not performing well for small values of n, R, and T. However, for large sample, both the methods of estimation are precisely estimating the parameters, reliability, and hazard functions in terms of ASEs/APEs and AILs and associated CPs of the confidence/HPD intervals.
-
The performance of Bayes estimation under gamma priors is better in comparison with those under Jeffrey’s priors as well as ML estimates in terms of AILs and associated CPs of HPD intervals. In all the cases, the HPD intervals under gamma priors cover the true parameters’ values with probability one. Over all, the coverage probabilities of the confidence intervals are observed to be smaller than those of HPD intervals.
-
In most of the cases, the CPs of confidence and HPD intervals tend to increase with n, R and T.
-
As expected, the average estimated errors and AILs of the various intervals tend to decrease with increase sample size n.
-
Bayes estimation with gamma prior provides more efficient estimates as compared with the Jeffrey’s prior as well as the ML method of estimation for all considered combinations of n, R, and T.
-
The APEs and the AILs of the HPD intervals based on gamma priors are smaller than those with Jeffrey’s priors.
-
As x increases, the reliability function (hazard rate function) decreases (increases). The same trend is observed in the case of their ML and Bayes estimates.
5 Concluding Remarks
In this article, the estimation of the parameters of the modified Weibull distribution under Type-II hybrid censoring scheme is presented. The ML and Bayesian estimation (with gamma and Jeffrey’s priors) methods have been used for this purpose. The performances of different estimators are examined based on small and large samples with different combinations of Type-II hybrid censoring parameters R and T. A real data set analysis is carried out under Type-II hybrid censored scheme to show the applicability of the modified Weibull distribution. Based on the censoring results, it is proposed that the modified Weibull distribution may be a better lifetime model for analyzing the reliability characteristics of real-life systems with Type-II hybrid censoring scheme.
References
Al-Zahrani B, Gindwan M (2014) Parameter estimation of a two-parameter Lindley distribution under hybrid censoring. Int J Syst Assur Eng Manag. doi:10.1007/s13198-013-0213-2
Balakrishnan N, Cohen AC (1991) Order statistics and inference: estimation methods. Academic Press, San Diego
Balakrishnan N, Kundu D (2013) Hybrid censoring: models, inferential results and applications. Comput Stat Data Anal 57(1):166–209
Banerjee A, Kundu D (2008) Inference based on type-II hybrid censored data from a Weibull distribution. IEEE Trans Reliab 57:369–378
Berger J (1985) Statistical decision theory and bayesian analysis. Springer, New York
Chen MH, Shao QM (1999) Monte Carlo estimation of Bayesian credible and HPD intervals. J Comput Graph Stat 8:69–92
Childs A, Chandrasekhar B, Balakrishnan N, Kundu D (2003) Exactlikelihood inference based on type-I and type-II hybrid censored samples from the exponential distribution. Ann Inst Stat Math 338(55):319–330
Dube S, Pradhan B, Kundu D (2011) Parameter estimation for the hybrid censored log-normal distribution. J Stat Comput Simul 81(3):275–282 (343)
Epstein B (1954) Truncated life tests in the exponential case. Ann Math Stat 25:555–564
Epstein B (1960) Estimation from life-test data. Technometrics 2:447–454
Ganguly et al (2012) Exact inference for the two-parameter exponential distribution under type-II hybrid censoring scheme. J Stat Plan Inference 142(3):613–625
Geman S, Geman A (1984) Stochastic relaxation, Gibbs distributions and the Bayesian restoration of images. IEEE Trans Pattern Anal Mach Intell 6:721–740
Gupta PK, Singh B (2013) Parameter estimation of hybrid censored data with Lindley distribution. Int J Syst Assur Eng Manag 4(4):378–385
Harter H, Leon, Balakrishnan N (1996) CRC handbook of tables for the use of order statistics in estimation. CRC Press, Boca Raton
Hastings WK (1970) Monte Carlo sampling methods using Markov Chains and their applications. Biometrika 57:97–109
Kaplan EL, Meier P (1958) Nonparametric estimation from incomplete observations. J Am Stat Assoc 53:457–481
Kundu D (2007) On hybrid censoring Weibull distribution. J Stat Plan Inference 137:2127–2142
Kundu D, Pradhan B (2009) Estimating the parameters of the generalized exponential distribution in presence of the hybrid censoring. Commun Stat Theory Methods 38(12):2030–2041
Lai CD, Xie M, Murthy DN (2003) A modified Weibull distribution. IEEE Trans Reliab 52:33–37
Lawless JF (1982) Statistical model and methods for lifetime data. Wiley, NewYork
Lawless JF (2003) Statistical models and methods for lifetime data, 2nd edn. Wiley, New York
Mann NR, Schafer RE, Singpurwala ND (1974) Methods for statistical analysis of reliability and life data. Wiley, New York
Metropolis N, Ulam S (1949) The Monte Carlo method. J Am Stat Assoc 44:335–341
Rastogi MK, Tripathi YM (2013) Inference on unknown parameters of Burr distribution under hybrid censoring. Stat Pap 54:619–643
Robert C (2007) The Bayesian Choice, 2nd edn. Springer, New York
Singh B, Gupta PK, Sharma VK (2014) On Type-II hybrid censored Lindley distribution. Stat Res Lett 3(2):58–62
Acknowledgments
The authors thankfully acknowledge the critical suggestions from the learned referees which greatly helped in the improvement of the paper.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Singh, B., Goel, R. Reliability Estimation of Modified Weibull Distribution with Type-II Hybrid Censored Data. Iran J Sci Technol Trans Sci 42, 1395–1407 (2018). https://doi.org/10.1007/s40995-016-0124-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40995-016-0124-6