Abstract
A new extension of the exponential distribution, proposed by Nadarajah and Haghighi (Statistics 45, 543–558 (2011)), is an alternative to the gamma, Weibull and generalized-exponential models, it is also known as NH distribution. The maximum likelihood and Bayes inferential approaches for estimating the unknown two-parameters and some lifetime parameters such as survival and hazard rate functions of the NH distribution in presence of progressive first-failure censored sampling are considered. Based on observed Fisher’s information matrix, the approximate confidence intervals for the two-parameters, and any function of them, are constructed. Using Lindley’s approximation and Markov chain Monte Carlo methods under the assumption of conjugate gamma priors, the Bayes estimates and associate highest posterior density credible intervals for the unknown parameters and reliability characteristics are developed based on squared error loss function. Although the proposed estimators cannot be expressed in explicit forms, these can be easily obtained through the use of appropriate numerical techniques. A Monte Carlo simulation study is carried out to examine the performance of proposed methods. Using different optimality criteria, an optimal censoring scheme has been suggested. Finally, a real data set is analyzed for illustration purposes.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
A new generalization of the exponential distribution as an alternative to the gamma, Weibull and generalized-exponential lifetime models has been introduced by Nadarajah and Haghighi (2011). The extension of the exponential distribution was named NH distribution (NHD) by Lemonte (2013) as an abbreviation for the name authors Nadarajah and Haghighi. Also, many properties of NHD are discussed by Nadarajah and Haghighi (2011). Suppose that the lifetime X of a testing unit follows two-parameter NHD(α, λ). The probability density function (PDF), f(⋅), cumulative distribution function (CDF), F(⋅), survival function (SF), S(⋅), hazard rate function (HRF), H(⋅), at mission time t, are given, respectively, by
and
where α and λ are the shape and scale parameters, respectively. Putting α = 1 in Eq. 1.1, exponential distribution is introduced as a special case.
Recently, many studies on estimating the unknown parameters of NHD based on different life-testing experiments have been carried out by many authors, including: Singh et al. (2015a) obtained the maximum likelihood estimators (MLEs) and Bayes estimators (BEs) of the NHD under Type-II progressive censoring scheme (PCS). Singh et al. (2015b) discussed the MLEs and BEs of the unknown parameters and reliability characteristics of the NHD based on complete sampling. Dey et al. (2017) described different methods of estimation for the unknown parameters of NHD.
Progressively first-failure censored sampling (PFFCS) proposed by Wu and Kuş (2009). This censoring scheme have become quite popular in reliability and lifetime testing studies. However, the drawback of the first-failure censoring is that it cannot be to allow for sets to be removed from the test at the points other than the final termination point. For this reason, they suggested the PFFCS by allowing to remove some of survival sets from the life-test. They, defined a PFFCS as a combination between the concepts of first-failure censoring and Type-II PCS. Hence, an extend of Type-II PCS is referred to as PFFCS. The PFFCS can be describe as follows: Suppose that n independent groups with k items within each group are put on a life-testing experiment at time zero, m is a pre-fixed number of failures, and the PCS \({\mathbf {R}=(R_{1},R_{2},\dots ,R_{m})}\) is pre-fixed. At the time of first-failure observed (say X(1)), R1 groups and the group in which the first-failure are randomly removed from the remaining live groups n − R1 − 1. Following, at the time of second-failure observed (say X(2)), R2 groups and the group in which the second-failure are randomly removed from the remaining live groups n − R1 − R2 − 2, and so on. This procedure continues until all remaining live Rm, (m ≤ n), groups and the group in which the m − th failure (say X(m)) has occurred are randomly removed at the time of the observed failure m − th. Then \( X_{(1)},X_{(2)},\dots ,X_{(m)} \) represents the independent lifetimes PFFCS order statistics with pre-fixed progressive censoring R. If the failure times of the n × k items originally in the life-test are from a continuous population, with PDF f(x(i);Θ) and CDF F(x(i);Θ), then the likelihood function for \( X_{(i)},\ i=1,2,\dots ,m \), where Θ is parameter vector, is given by
where \({C}={n({n}-R_{1}-1)\dots ({n}-{\sum }_{i=1}^{m}(R_{i}+1))}\) and \({n = m+{\sum }_{i=1}^{m}R_{i}}\).
From Eq. 1.5, some sampling schemes can be obtained as a special cases, such as: Type-II PCS by putting k = 1, the joint PDF of the first-failure-censored order statistics by putting \({\mathbf {R}=(0,0,\dots ,0)}\), Type-II censoring by putting \({\mathbf {R}=(0,0,\dots ,n-m)}\) and k = 1, if putting \({\mathbf {R}=(0,0,\dots ,0)}\) and k = 1, then n = m, which yields the complete sample.
To our best knowledge, statistical inference for unknown parameters and reliability characteristics of NHD has not yet been studied under progressively first-failure censored data. In this paper, our main purpose is to use the maximum likelihood and Bayesian approaches to derive both point and interval estimates of the unknown parameters, as well as some lifetime parameters such as SF and HRF, of the NHD when the lifetime data are collected under PFFCS. It is observed that the MLEs of the unknown parameters cannot be obtained in closed form, as expected, and they have to obtained by solving two non-linear equations simultaneously. Independent conjugate gamma priors of the unknown parameters are considered to develop the BEs under squared error loss (SEL) function. Also, approximate confidence intervals (ACIs) and highest posterior density (HPD) credible intervals of the unknown parameters α and λ, and any function of them, are constructed. It is observed that the BEs have not been obtained in explicit form. Thus, we propose to use the Lindley’s approximation and Markov chain Monte Carlo (MCMC) method such as Metropolis-Hastings (M-H) algorithm to generate MCMC samples from the posterior distribution and then to compute BEs and also to construct the associated HPD credible intervals. A Monte Carlo simulation study is performed to compare the effects of group sizes k, number of groups n, effective sample sizes m, and different censoring schemes R to compare the performances among various estimates in terms of their mean square error (MSE) and relative absolute bias (RAB). Furthermore, a real data set is discussed based on different PCSs to illustrate our proposed estimators. To choose the optimum censoring scheme from a class of possible schemes, four commonly-used criteria were considered to compare between different schemes. The rest of the paper is organized as follows: In Section 2, we derive the MLEs and associated two-sided ACIs. In Section 3, we develop the Bayesian inference based on SEL function. Monte Carlo simulation results are presented in Section 4. In Section 5, optimal censoring plans are discussed. Section 6 deals with a real-life data set for illustration purposes. Finally, concluding remarks are provided in Section 7.
2 Maximum Likelihood Estimation
Let \( X_{(i)},\ i=1,2,\dots ,m \), (1 ≤ m ≤ n), be a progressive first-failure censored sample obtained from NHD(α, λ) with pre-fixed PCS R. Then, with substituting Eqs. 1.1 and 1.2 into 1.5, we get
where \(\xi (x_{(i)};\lambda ){}={}{(1{}+{}\lambda {x_{(i)}})}\) and \({\omega (x_{(i)},R_{i};\alpha ,\lambda ){}=(R_{i}+1)(1{}-{}(\xi (x_{(i)};\lambda ))^{\alpha })}\).
By dropping terms that do not involve α and λ, the corresponding log-likelihood function, ℓ(.), of Eq. 2.1 can be written with proportional, \( {\ell (.)}\propto {\log {L(.)}} \), as
Differentiating (2.2) partially with respect to α and λ, we get two likelihood equations must be simultaneously satisfied to obtain the MLEs \( \hat {\alpha } \) and \( \hat {\lambda } \) of α and λ, respectively, as
and
Clearly, from Eqs. 2.3 and 2.4, we have a system of two nonlinear equations in the unknown parameters α and λ. It is obvious that a closed form solution is quite difficult to obtain. So, we need to apply a suitable iterative procedure like the Newton-Raphson method to obtain the desired MLEs of the unknown two-parameter for given values of (k, n, m, R, x).
Also, using the invariance property of MLEs \( \hat {\alpha } \) and \( \hat {\lambda } \), the MLEs \( \hat {S}(t) \) and \( \hat {H}(t) \) of S(t) and H(t), as in Eqs. 1.3 and 1.4, respectively, for given mission time t can be derived by replacing α and λ by their MLEs \( \hat {\alpha } \) and \( \hat {\lambda } \), respectively. Asymptotic variance-covariance (V-C) matrix of the MLEs \( \hat {\Theta }=(\hat {\alpha },\hat {\lambda })^{\mathbf {T}} \) can be obtained by the inverting of the Fisher information matrix, I0(Θ), which is given by
Unfortunately, the exact expressions of the expectation (2.5) is difficult to get. By dropping the expectation operator E and replacing Θ by \( \hat {\Theta } \), see Cohen (1965), we get the approximate asymptotic V-C matrix for the MLEs as
From (2.6), the Fisher’s elements will be
and
The asymptotic normality of MLEs \({\hat {\Theta }}\sim {N({\Theta },\mathrm {I}_{0}^{-1}(\hat {\Theta }))} \) can be used to construct 100(1 − γ)% two-sided ACIs for the unknown parameters α and λ, respectively, as
where zγ/2 is the percentile of the standard normal distribution with right-tail probability (γ/2) − th. To construct the ACIs of S(t) and H(t), we need to find the variances of them. Therefore, the delta method is considered to obtain the approximate estimates of the variance of \( {\hat {S}(t)} \) and \( \hat {H}(t) \). Delta method is a general approach for computing ACIs for any function of the MLEs \( \hat {\alpha } \) and \( \hat {\lambda } \), see Greene (2012). Hence, according to this method, the variance of \( {\hat {S}(t)} \) and \( \hat {H}(t) \), can be approximated, respectively by
where \( \nabla {\hat {S}(t)} \) and \( \nabla {\hat {H}(t)} \) are, respectively, the gradient (vector of first partial derivatives) of S(t) and H(t) with respect to α and λ obtained at \( {\alpha =\hat {\alpha }} \) and \( {\lambda =\hat {\lambda }} \), i.e.,
and
Hence, the 100(1 − γ)% two-sided ACIs of S(t) and H(t), are given, respectively, by
where zγ/2 is the percentile of the standard normal distribution with right-tail probability (γ/2) − th.
3 Bayesian Inference
In this section, the BEs and associated HPD credible intervals of the unknown model parameters α and λ, as well as, S(t) and H(t) under SEL function are developed. A very well-known symmetric loss function is the SEL function, which is defined as
where \( \tilde {\Theta } \) being an estimate of Θ. From Eq. 3.1, the BE is given by the posterior mean of Θ. However, any other loss function can be easily incorporated.
In order to obtain the BEs, we assume that the both unknown parameters α and λ are independent random variables. Following Singh et al. (2015a, b) and Dey et al. (2017), we assumed that α and λ having conjugate gamma priors as \( \alpha \sim {Gamma(a_{1},b_{1})} \) and \( \lambda \sim {Gamma(a_{2},b_{2})} \), respectively. The hyper-parameters ai,bi, i = 1,2, are chosen to reflect prior know ledge about the two unknown parameters α and λ. Here, the gamma conjugate prior of the NH parameters are chosen because this prior is flexible enough to cover a large variety of prior beliefs of the experimenter. Consequently, the joint prior density of α and λ can be written with proportional as
In the case of non-informative priors ai,bi = 0, i = 1,2, the joint prior density (3.2) becomes π(α, λ) ∝ (αλ)− 1. In continuous Bayes’ theorem, the joint posterior PDF of α and λ, up to proportionality, is given by
Combining (3.2) with (2.1), the joint posterior distribution (3.3) can be written as
The normalizing constant, κ, of Eq. 3.4 is given by
Hence, the BE of any function of α and λ, say φ(α, λ), under SEL function is the posterior expectation of φ(α, λ) is given by
Unfortunately, based on Eq. 3.5, the BEs are obtained in the form of a ratio of two-dimensional integrals for which a closed-form solution is not possible. Hence, the BEs and corresponding HPD credible intervals can be computed numerically. Because of that we propose to use two approaches to approximate (3.5), namely, Lindley’s approximation and importance sampling procedure.
3.1 Lindley’s Approximation
Lindley (1980) presented procedure to evaluate the posterior expectation, this procedure used when the posterior distribution take a ratio form that involves an integration in the denominator and cannot be reduced to a closed form. Thus, we use Lindley’s method to approximate all of the BEs. The posterior expectation of φ(𝜃) is given by
where φ(𝜃) = ℓ(𝜃) + ρ(𝜃) is function of 𝜃 only, ℓ(𝜃) is the log-likelihood function and \( \rho (\theta )=\log \pi (\theta ) \) is the log-prior density function. For the two-parameters case φ(𝜃1,𝜃2), with sufficiently large sample size n, the approximate BE, \( \tilde \varphi _{L}(\theta _{1},\theta _{2}) \), of any function of 𝜃1 and 𝜃2 can be written as
where \( A={\sum }_{i=1}^{2}{\sum }_{j=1}^{2}u_{ij}\sigma _{ij} \), Aij = uiσii + ujσji, Bij = (uiσii + ujσij)σii, \( C_{ij}=3u_{i}\sigma _{ii}\sigma _{ij} +u_{j}(\sigma _{ii}\sigma _{jj}+2\sigma _{ij}^{2}) \), ui = ∂φ(𝜃1,𝜃2)/∂𝜃i, uij = ∂2φ(𝜃1,𝜃2)/∂𝜃i∂ 𝜃j, \( \ell _{ij}=\partial ^{i+j}\ell (\theta _{1},\theta _{2})/{\partial \theta _{1}^{i}}{\partial \theta _{2}^{j}} \), i, j = 0,1,2,3, i + j = 3, \( \rho =\log \pi (\theta _{1},\theta _{2}) \), ρi = ∂ρ/∂𝜃i, and σij is the (i, j) − th element of the asymptotic V-C matrix \( \mathrm {I}_{0}^{-1}(\hat \theta _{1},\hat \theta _{2}) \). All terms in (17) are evaluated at the MLEs \( \hat \theta _{1} \) and \( \hat \theta _{2} \) of the unknown parameters 𝜃1 and 𝜃2, respectively.
Taking φ(α, λ) is a function of the unknown parameters α and λ or them such as S(t) and H(t), then the approximate BE \( {\tilde \alpha _{L}} \) and \( {\tilde \lambda _{L}} \) of α and λ, under the SEL function, are given, respectively, by
and
Now, to compute the BEs based on Lindley’s approximation, we must have
Using Eq. 3.6, the approximate BE \( {\tilde {S}_{L}(t;\alpha ,\lambda )} \) of S(t;α, λ) is given by
where
and
Similarly, the approximate BE \( {\tilde {H}_{L}(t;\alpha ,\lambda )} \) of H(t;α, λ) is given by
where
and
Although the BEs, using the Lindley’s approximation method, of the unknown parameters of α and λ as well as S(t) and H(t) can be obtained easily, but unfortunately, we cannot construct the HPD credible intervals. For this purpose, the M-H algorithm can be used to generate MCMC samples from the posterior distribution and then to compute BEs and also to construct the associated HPD credible intervals.
3.2 M-H Algorithm
Importance sampling technique such as the M-H sampler algorithm is required for the implementations of MCMC methodology to generate a samples from Eq. 3.4. One may refer to Metropolis et al. (1953) and Hastings (1970) for various applications of this algorithm. The M-H algorithm is mainly used to simulate samples from probability distribution by making use of full joint density function and (independent) proposal distribution. To implement the M-H algorithm the full conditional posterior distributions of α and λ, from Eq. 3.4, are given, respectively, by
and
The conditional posterior distributions (3.7) and (3.8) of the unknown parameters α and λ, respectively, cannot be reduced analytically to well-known distributions. Therefore, the M-H algorithm with normal proposal distribution is used to generate random samples from these distributions, in turn obtain the BEs and corresponding HPD credible intervals. The steps of M-H algorithm are carried out as follows:
Step 1: Start with an initial guess α(0) and λ(0).
Step 2: Set J = 1.Step 3: Generate α(J) and λ(J) from Eqs. 3.7 and 3.8 with normal proposal distributions \( N{(\alpha ^{(J-1)},\sigma _{\alpha }^{2})} \) and \( N{(\lambda ^{(J-1)},\sigma _{\lambda }^{2})} \), respectively, as
(a) Generate α∗ from \( N{(\alpha ^{(J-1)},\sigma _{\alpha }^{2})} \) and λ∗ from \( N{(\lambda ^{(J-1)},\sigma _{\lambda }^{2})} \).
(b) Evaluate the acceptance probabilities by \( \varphi _{\alpha }=\min \limits \left \{1, \frac {\pi _{1}^{*}{(\alpha ^{*}|\lambda ^{(J-1)},\underline {x})}} {\pi _{1}^{*}{(\alpha ^{(J-1)}|\lambda ^{(J-1)},\underline {x})}} \right \} \) and \( \varphi _{\lambda }=\min \limits \left \{1, \frac {\pi _{2}^{*}{(\lambda ^{*}|\alpha ^{(J)},\underline {x})}} {\pi _{2}^{*}{(\lambda ^{(J-1)}|\alpha ^{(J)},\underline {x})}} \right \} \).
(c) Generate a samples u1 and u2 from uniform distribution.
(d) If u1 ≤ φα, accept the proposal and set α(J) = α∗, else set α(J) = α(J− 1).
(e) If u2 ≤ φλ, accept the proposal and set λ(J) = λ∗, else set λ(J) = λ(J− 1).
Step 4: For a given time t, the BEs of SF and HRF, are given, respectively, by \( S^{(J)}(t)=\exp ({1-(1+\lambda ^{(J)}{t})^{\alpha ^{(J)}}}) \) and \( H^{(J)}(t)=\alpha ^{(J)}\lambda ^{(J)}(1+\lambda ^{(J)}{t})^{\alpha ^{(J)}-1} \).
Step 5: Set J = J + 1.
Step 6: Repeat Steps 3-5 for M times and obtain
Step 7: To compute the HPD credible intervals of φ = (α, λ, S(t),H(t)) order the MCMC sample of \( \varphi ^{(i)},\ i=1,2,\dots ,M \), as \( (\varphi _{(1)},\varphi _{(2)},\dots ,\varphi _{(M)}) \). Using the algorithm proposed by Chen and Shao (1999), the 100(1 − γ)% HPD credible interval for φ is given by \( (\varphi _{(J^{*})}, \varphi _{(J^{*}+(1-\gamma )M)}) \), where J∗ is chosen such that
Here [x] denotes the largest integer less than or equal to x. Hence, the HPD credible interval of φ is that interval which has the shortest length.
The first simulated varieties, M0, of the algorithm may be biased by the initial value, therefore, usually discarded in the beginning of the analysis implementation (burn-in period) in order to guarantee the convergence and to remove the affection of the selection of initial. Then the selected samples φ(i) for \( i=M_{0}+1,\dots ,M \), are sufficiently large can be used to develop the Bayesian inferences. Thus, the MCMC BEs of φ(α, λ), of any function of α and λ based on SEL function, as in Eq. 3.1, are given by \( {\tilde {\varphi }}_{M-H}={\sum }_{i=M_{0}+1}^{M}{{\varphi }}^{(i)}/(M-M_{0}) \), where M0 is burn-in.
4 Monte Carlo Simulation Study
In this section, a Monte Carlo simulation study is carried out in order to compare the proposed estimates of α, λ, S(t) and H(t). A large number N = 1000 of progressively first-failure censored samples for a true value of parameters α and λ, and different combinations of n(number of groups), m(progressively first-failure-censoring data), k(number of items within each group) are generated from the NHD by using the algorithm described in Balakrishnan and Sandhu (1995). In each case, the MLEs and BEs of the unknown parameters and reliability characteristics are computed. The Lindley’s approximation and MCMC methods based on SEL function are used for Bayesian computation purposes. Comparison between different estimators is made with respect to their MSE and RAB values. Also, the performances of the two-sided 95% ACI/HPD credible intervals are compared by the average lengths (ALs). The simulation study was performed for different values of (k, n, m) such as: n = 20(small), 50(medium) and 80(large) for each group size k = 2 and 5. The test is terminated when the number of failed subjects achieves or exceeds a certain value m, where the failure proportion m/n = 30, 60 and 90%. For each test, we compute the average estimates with their MSEs and RABs, which are given, respectively, by \( \bar {\hat \varphi }=\frac {1}{N}{\sum }_{i=1}^{N}{\hat \varphi _{i}} \), \( \text {MSE}{(\hat \varphi )}=\frac {1}{N}{\sum }_{i=1}^{N}{(\hat \varphi _{i}-\varphi )^{2}} \) and \( \text {RAB}{(\hat \varphi )}=\frac {1}{N}{\sum }_{i=1}^{N}{\frac {|\hat \varphi _{i}-\varphi |}{\varphi }} \), where \( \hat \varphi \) is the MLE or BE of the parametric function φ.
Based on the non-informative priors ai,bi for i = 1,2, the joint posterior distribution of the unknown two parameters is proportional to the likelihood function. Hence, we have used informative prior of α and λ with a1 = a2 = 0.1 and b1 = b2 = 1 when (α, λ) = (0.1,0.1). Here, the values of hyper-parameters are chosen to satisfy the prior mean become the expected value of the corresponding parameter. Here, the values of hyper-parameters are chosen to satisfy the prior mean become the expected value of the corresponding parameter. Further, we have also obtained the MLEs and BEs of the SF and HRF where the true values of S(t) and H(t), for specified time t = 5, are taken as S(5) = 0.959464 and H(5) = 0.0069425.
Using the M-H sampler algorithm proposed in Section 3, the different BEs based on 12,000 MCMC samples and discard the first 2000 values as “burn-in” are developed. For interval estimation, we compute the ALs of 95% ACI/HPD credible intervals of α, λ, S(t) and H(t). Now, we consider the following different sampling schemes (CSs) as
Extensive computations were performed using \(\mathcal {R}\) statistical programming language software with mainly two useful statistical packages are ‘coda’ package proposed by Plummer et al. (2006), and ‘maxLik’ package, which using Newton-Raphson method of maximization in the computations, proposed by Henningsen and Toomet (2011). The average MLEs and BEs of α, λ, S(t) and H(t) are presented in Tables 1, 2, 3, 4, 5 and 6. In addition, the ALs of α, λ, S(t) and H(t) are listed in Tables 7 and 8. From Tables 1, 2, 3, 4, 5 and 6, it can be seen that the MLEs and BEs of unknown parameters and the reliability characteristics of NHD are very good in terms of MSEs and RABs. As n and m increases, the MSEs and RABs of all estimates decrease as expected. Moreover, as group size k increases, the MSEs and RABs associated with shape parameter α increase while that associated with scale parameter λ decrease. It is also observed that as the failure proportion m/n increases, the point estimates become even better. Also, the BEs using gamma informative prior are better as they include prior information than MLEs in terms of MSEs and RABs. Furthermore, the MCMC method using M-H algorithm is better than the Lindley’s approximation method in respect of both MSEs and RABs. For interval estimation, the HPD credible intervals are better than ACIs in respect of ALs. In each case, as the failure proportion m/n increases, the interval estimates become even better in respect of ALs. Also, as k increases, the ALs of ACIs increases while the ALs of HPD credible intervals narrows down. Therefore, we recommend the Bayesian point and interval estimation of the unknown parameters and reliability characteristics using M-H algorithm. Furthermore, in most cases, comparing the CS-I and CS-III, it is clear that the MSEs and RABs of the MLEs and BEs for the unknown parameters and reliability characteristics are greater for the CS-III than CS-I. This may not be surprising, because the expected duration of the experiments for CS-I is greater than the CS-III. Thus the data obtained by the CS-I would be expected to provide more information about the unknown parameters and reliability characteristics than the data obtained by CS-III.
5 Optimal Censoring Plans
In the context of life-testing experiment, the choice of optimum censoring scheme from a class of all possible schemes is an important concern to the experimenter. Recently, choosing the optimal censoring scheme in different problems has received considerable attention in the statistical literature. See for example, Pradhan and Kundu (2009, 2013), Sultan et al. (2014), Dube et al. (2016) and Sen et al. (2018). Some of these criteria is the variance optimality used for one parameter case, but, the trace and determinant optimality are used for the multi-parameter case, intend to minimize the variance or the determinant of the V-C matrix of estimators under consideration. It is desirable for practical purposes to choose the optimum PFFCS from a class of possible censoring schemes. Thus, for given values of k, n, m and R, where \( {\sum }_{i=1}^{m}{R_{i}}=n-m \), we consider four criteria are commonly-used to compare between different censoring schemes, are reported in Table 9.
According to Criterion-I and Criterion-II, our goal is minimize the determinant and the trace, respectively, of V-C matrix (2.6) of the MLEs \( \hat {\alpha } \) and \( \hat {\lambda } \). In the case of one-parameter distributions, the comparison between different criteria can be easily made. But, if both the parameters are unknown, then the comparison of the two Fisher information matrices is not a trivial task, see Pradhan and Kundu (2013). Unfortunately, in presence of the shape and scale parameters, the Criterion-I and Criterion-II are not scale invariant, Gupta and Kundu (2006). However, in the case of multi-parameter distributions, there are other criteria which are scale invariant. Criterion-III depends on the choice of p, and it is tends to minimize the variance of logarithmic of MLE of the p − th quantile \( (\ln (\hat {T}_{p})) \), where 0 < p < 1. Hence, the logarithmic for Tp of the NHD is given by
From Eq. 5.1, using the delta method, the variance of \( (\ln (\hat {T}_{p})) \) can be approximated by
where \({[\nabla {\ln (\hat {T}_{p})}]}^{\mathbf {T}} =[\partial \nabla {\ln ({T}_{p})}/\partial {\alpha }, \partial \nabla {\ln ({T}_{p})}/\partial {\lambda }]_{({\alpha =\hat {\alpha }},{\lambda =\hat {\lambda }})} \) is the gradient of \( \ln ({T}_{p}) \) with respect to the unknown parameters α and λ.
From Table 9, the weight function w(p) ≥ 0 is a non-negative function satisfying \( {{\int \limits }_{0}^{1}}{w(p)}dp=1 \). Also, \( \ln (\hat {T}_{p}) \) the same as in Criterion-III. Without loss of generality, we take w(p) = 1, 0 < p < 1.
6 Real-life Data Analysis
In this section, a real data set is analyzed to illustrate how the proposed methodology can be applied in real life phenomenon. The considered data set is taken form Linhart and Zucchini (1986), which represents the failure times of the air conditioning system of an air-plane. The ordered data with are as follows: 1, 3, 5, 7, 11, 11, 11, 12, 14, 14, 14, 16, 16, 20, 21, 23, 42, 47, 52, 62, 71, 71, 87, 90, 95, 120, 120, 225, 246 and 261. Recently, this real data set was analyzed by Singh et al. (2015a, b). Before progressing further, we first fit the NHD to the complete data set and compare it’s fitting with three lifetime distributions namely, GD, WD and GED with PDFs, respectively, as
and
where α and λ are the shape and scale parameters, respectively.
One question arises about whether the data fit the GD, WD, GED and NHD or not. To check the validity of proposed model, the Kolmogorov-Smirnov (K-S) and Anderson-Darling (A-D) goodness-of-fit test statistics with associated p-value are considered. Also, the negative log-likelihood function \( -\hat {\ell } \) evaluated at the MLEs, Akaike information criterion (AIC), \( \text {AIC}=-2\hat {\ell }+2S \), corrected Akaike information criterion (AICc), AICc = AIC + (2nS/n − S − 1), consistent Akaike information criterion (CAIC), \( \text {CAIC}=-2\hat {\ell }+(2nS/n-S-1) \), Bayesian information criteria (BIC), \( \text {BIC}=-2\hat {\ell }+S\log (n) \), and Hannan-Quinn information criterion (HQIC), \( \text {HQIC}=-2\hat {\ell }+2S\log (\log (n)) \), where n and S are the sample size and the number of model parameters, respectively, are considered. In general, the best distribution corresponds to the lowest of \( -\hat {\ell } \), AIC, AICc, CAIC, BIC, HQIC, K-S and A-D statistic values and highest p-values. The values of MLEs of the model parameters and corresponding goodness-of-fit measures are reported in Table 10. The statistic value of K-S and A-D goodness-of-fit tests with its p-values are listed in Table 11. Also, we use a graphical method for goodness-of-fit of distributions. We draw quantile-quantile (Q-Q) plots of the GD, WD, GED and NHD, which are shown in Fig. 1. A Q-Q plot depicts the points \( \{F^{-1}((i-0.5)/n;\hat {\theta }),x_{(i)}\},\ i=1,2,\dots ,n, \) where \( \hat {\theta } \) is the MLE of 𝜃.
From Tables 10 and 11, it can be seen that the NHD is the best choice among the competing GD, WD and GED in the literature for fitting lifetime real data, since it has the smallest goodness of statistic values and highest p-values. Furthermore, the Q–Q plots support the above findings. For more fitting illustration, we have also provided two plots computed at the estimated parameters. First plot is the empirical cumulative distribution function plot and fitted the CDF of GD, WD, GED and NHD. Second plot is the histogram of the real data and the fitted PDF of GD, WD, GED and NHD, as in Fig. 2. To illustrate the inferential methods developed in the preceding sections, we generate a first-failure censored sample after randomly grouping this data set into 15 groups with k = 2 items within each group and report it in Table 12.
The calculated values of the four criteria are reported in Table 16. From Table 16, we observe that scheme R1 is the optimal scheme than schemes R2 and R3 for criteria I, II and IV. Moreover, censoring scheme R2 is the optimal for Criterion-II. Then, the first-failure censored sample is obtained in order as: 1, 3, 5, 7, 11, 12, 14, 14, 14, 16, 16, 23, 42, 90 and 120. Using the first-failure censored data, three different progressive first-failure censored samples, using three different schemes with k = 2, n = 15 and m = 8 are generated and reported in Table 13. For brevity, the censoring scheme R = (2,0,0,0,2) is denoted by R = (2,0∗3,2). Using Table 13, the MLEs and the BEs of the unknown parameters α and λ, as well as, the reliability characteristics S(t) and H(t) at given mission time t = 10, are computed and listed in Table 14. The BEs are developed using Lindley’s approximation and MCMC methods under with a non-informative prior. Now, we generate 11000 MCMC samples and then first 1000 iterations (burn-in period) have been discarded from the generated sequence as has been suggested in Section 3. The initial values for the unknown parameters α, λ, S(t) and H(t) for running the MCMC sampler algorithm were taken to be their maximum likelihood estimates. Moreover, two-sided 95% ACIs and HPD credible intervals are listed in Table 15. Now, we illustrate the concept of optimal censoring based on the four criteria are considered in Table 9. Using the three data sets in Table 13, Criterion-I and Criterion-II can be obtained by compute the determinant and trace of the observed V-C matrix, respectively. Based on four different quantiles namely: p = 0.50 and p = 0.99, the Criterion-III and Criterion-IV are computed.
7 Concluding Remarks
In this paper, the problem of estimating the unknown parameters and reliability characteristics of the NHD when the data are collected under the PFFCS is discussed. The MLEs and BEs of the unknown parameters and reliability characteristics are obtained. These estimates cannot be obtained in closed forms, but can be derived numerically. Using gamma informative priors under SEL function, the BEs using the Lindley’s approximation and MCMC methods are developed. Based on the asymptotic normality of the MLEs and delta method, the 95% ACIs of the parameters and reliability characteristics are constructed. Since the Lindley’s approximation method cannot construct the HPD credible intervals, we made use of the M-H algorithm to obtain point estimates and associated HPD credible intervals. Moreover, the results of Singh et al. (2015a) if putting k = 1 and the results of Singh et al. (2015b) if putting \( \mathbf {R}=(0,0,\dots ,0) \) and k = 1 may be obtained as a special cases from the new results. It is important for a reliability practitioner to know which censoring scheme would be optimum for the life-testing experiment, we have provided the optimal censoring scheme based on four criteria measures. We hope that the results and methodology discussed in this paper will be beneficial to data analyst and reliability practitioners.
Change history
References
Balakrishnan, N. and Sandhu, R.A. (1995). A simple simulational algorithm for generating progressive type-II censored samples. The American Statistician49, 229–230.
Chen, M.H. and Shao, Q.M. (1999). Monte Carlo estimation of Bayesian credible and HPD intervals. Journal of Computational and Graphical Statistics 8, 69–92.
Cohen, A.C. (1965). Maximum likelihood estimation in the Weibull distribution based on complete and censored samples. Technometrics 7, 579–588.
Dey, S., Zhang, C., Asgharzadeh, A. and Ghorbannezhad, M. (2017). Comparisons of methods of estimation for the NH distribution. Annal. Data Sci. 4, 1095–1114.
Dube, M., Krishna, H. and Garg, R. (2016). Generalized inverted exponential distribution under progressive first-failure censoring. Journal of Statistical Computation and Simulation 86, 1095–1114.
Greene, W.H. (2012). Econometric Analysis, 7th edn. Pearson Prentice-Hall, Upper Saddle River.
Gupta, R.D. and Kundu, D. (2006). On the comparison of Fisher information of the Weibull and GE distributions. Journal of Statistical Planning and Inference136, 9, 3130–3144.
Hastings, W.K. (1970). Monte Carlo sampling methods using Markov chains and their applications. Biometrika 57, 97–109.
Henningsen, A. and Toomet, O. (2011). maxLik: a package for maximum likelihood estimation in R. Comput. Stat. 26, 443–458.
Lemonte, A.J. (2013). A new exponential-type distribution with constant, decreasing, increasing, upside-down bathtub and bathtub-shaped failure rate function. Computational Statistics and Data Analysis 62, 149–170.
Linhart, H. and Zucchini, W. (1986). Model Selection. John Wiley, New York.
Lindley, D.V. (1980). Approximate Bayesian methods. Trab Estadistica.31, 223–245.
Metropolis, N., Rosenbluth, A.W., Rosenbluth, M.N., Teller, A.H. and Teller, E. (1953). Equation of state calculations by fast computing machines. J. Chem. Phys. 21, 1087–1091.
Nadarajah, S. and Haghighi, F. (2011). An extension of the exponential distribution. Statistics 45, 543–558.
Plummer, M., Best, N., Cowles, K. and Vines, K. (2006). CODA: convergence diagnosis and output analysis for MCMC. R News 6, 7–11 .
Pradhan, B. and Kundu, D. (2009). On progressively censored generalized exponential distribution. Test 18, 497–515.
Pradhan, B. and Kundu, D. (2013). Inference and optimal censoring schemes for progressively censored Birnbaum–Saunders distribution. Journal of Statistical Planning and Inference 143, 1098–1108.
Sen, T., Tripathi, Y.M. and Bhattacharya, R. (2018). Statistical inference and optimum life testing plans under type-II hybrid censoring scheme. Annals of Data Science 5, 679–708.
Singh, U., Singh, S.K. and Yadav, A.S. (2015a). Bayesian estimation for extension of exponential distribution under progressive type-II censored data using Markov chain Monte Carlo method. Journal of Statistics Applications and Probability4, 275–283.
Singh, S.K., Singh, U. and Yadav, A.S. (2015b). Reliability estimation and prediction for extension of exponential distribution using informative and non-informative priors. International Journal of System Assurance Engineering and Management 6, 466–478.
Sultan, K.S., Alsadat, N.H. and Kundu, D. (2014). Bayesian and maximum likelihood estimations of the inverse Weibull parameters under progressive type-II censoring. Journal of Statistical Computation and Simulation 84, 2248–2265.
Wu, S.J. and Kuş, C. (2009). On estimation based on progressive first-failure-censored sampling. Computational Statistics and Data Analysis53, 3659–3670.
Acknowledgements
The authors would like to thank the Editor-in-Chief, the Associate Editor and the two anonymous referees for carefully reading the paper and for their comments which greatly improved the paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original version of this article was revised: This article was originally published without an Open Access.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ashour, S.K., El-Sheikh, A.A. & Elshahhat, A. Inferences and Optimal Censoring Schemes for Progressively First-Failure Censored Nadarajah-Haghighi Distribution. Sankhya A 84, 885–923 (2022). https://doi.org/10.1007/s13171-019-00175-2
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13171-019-00175-2
Keywords
- Bayes procedure
- Lindley’s approximation method
- Maximum likelihood estimation
- Nadarajah-Haghighi distribution
- Optimum censoring scheme
- Progressively first-failure censored sampling