1. Introduction
The inference of stress-strength reliability in statistics is an important topic of interest. It has many applications in practical areas. In the stress-strength modeling,
is a measure of component reliability. If
, the component fails or the component may malfunction, where
X is subject to
Y.
R can also be considered in electrical and electronic systems. Many authors have studied its properties for many statistical models including double exponential, Weibull, generalized Pareto and Lindley distribution (see [
1,
2,
3,
4]). Classical and Bayesian estimation of reliability in a multicomponent stress-strength model under a general class of inverse exponentiated distributions was researched by Ref. [
5]. Ref. [
6] studied classical and Bayesian estimation of reliability in multicomponent stress-strength model under Weibull distribution. Furthermore, Refs. [
7,
8,
9] also considered the problem of the stress-strength reliability.
The inverse Weibull (IW) distribution has attracted much attention recently. If
T denotes the random variable from Weibull model, we define
X as follows
, then the random variable
X is said to have inverse Weibull distribution. It is a lifetime probability distribution that can be used in the reliability engineering discipline. The inverse Weibull distribution has the ability to model failure rates, which is quite common in reliability and biological studies. The inverse Weibull model was referred to with many different names like “Frechet-type” ([
10]) and “Complementary Weibull” ([
11]). Ref. [
11] discussed a graphical plotting technique to settle the suitability of the model. Ref. [
12] presented the IW distribution for modeling reliability data, this model was further discussed by researching the failures of mechanical components subject to degradation. Ref. [
13] proposed a discrete inverse Weibull distribution and its parameters were estimated. The mixture model of two IW distributions and its identifiability properties were studied by [
14]. For the theoretical analysis of IW distribution, we can refer to [
15]. Ref. [
16] proposed the generalized IW distribution and several properties of this model. For more details on the inverse Weibull distribution, see [
17].
In this paper, we focus on the estimation of the stress-strength reliability , where X and Y follow the inverse Weibull distribution. As far as we know, this model has not been previously studied, although, we believe it plays an important role in reliability analysis.
We obtain the maximum likelihood estimator (MLE), approximate maximum likelihood estimator (AMLE) and the asymptotic distribution of the estimator. The asymptotic distribution is used to construct an asymptotic confidence interval. We also present two bootstrap confidence intervals of R. By using Gibbs sampling technique, we obtain Bayes estimator of R and its corresponding credible interval. Finally, we present a real data example to illustrate the performance of different methods.
The layout of this paper is organized as follows: in
Section 2, we introduce the distribution of the inverse Weibull. In
Section 3, we obtain the MLE of
R. In
Section 4, we derive the estimator of
R by approximating maximum likelihood equations. Different confidence intervals are presented in
Section 5. In
Section 6, Bayesian solutions are introduced. In
Section 7, we compare different proposed methods using Monte Carlo simulation. A numerical example is also provided. Finally, in
Section 8, we conclude the paper.
2. Inverse Weibull Distribution
The probability density function of the known Weibull distribution is given by
where
is the shape parameter and
is the scale parameter.
Let
T denote the random variable from Weibull model, namely,
. Define
X as follows:
The random variable
X is said to have inverse Weibull distribution, and its probability density function (pdf) is given by
The cumulative distribution function(cdf) is given by
where
and
. The inverse Weibull distribution will be denoted by
.
3. Maximum Likelihood Estimator of
In this section, we consider the problem of estimating
under the assumption that
and
. Then, it can be easily calculated that
To computer the MLE of
R, first we obtain the MLEs of
and
. Suppose
is a random sample from
and
is a random sample from
. The joint likelihood function is:
Then, the log-likelihood function is
,
and
, the MLEs of the parameters
,
and
, can be numerically obtained by solving the following equations:
From (8) and (9), we obtain
Putting the expressions of
and
into (
7), we obtain
Therefore,
can be obtained as a fixed point solution of the non-linear equation of the form
where
Using a simple iterative procedure
, where
is the j-th iterate of
, we stop the iterative procedure when
is adequately small less than a specified level. Once we obtain
, then
and
can be calculated from (
10). Therefore, we obtain the MLE of
as
4. Approximate Maximum Likelihood Estimator of
The MLEs do not take explicit forms; therefore, we approximate the likelihood equation and derive explicit estimators of the parameters.
Since the random variable
X follows
, then
has the extreme value distribution with pdf as
where
and
. The
and
are location and scale parameters, respectively. The pdf and cdf of the standard extreme value distribution can be obtained as
Suppose and are the ordered and , we assume the following notations: , and , , where , and .
The log-likelihood function of the data
and
is
Differentiating (
17) in regard to
, the score equations are obtained as
We note that the function
makes the score Equation (
18) nonlinear and intricate. Thus, we approximate the function
by expanding it in a Taylor series around
. Furthermore, we also approximate the function
by expanding it in a Taylor series around
. From [
18], it is known that
where
is the
i-th order statistic from the uniform
distribution. Therefore,
and
We use the following notations, , ; therefore, ,
Expanding the function
and
and keeping the first two terms, we have
where
Therefore, (
18)–(20) can be represented as
The estimator of
and
can be obtained from Equations (
21) and (22), as
where
The estimator of
can be determined as the unique solution of the quadratic equation
where
Once
is obtained,
and
can be derived immediately. Hence, the AMLE of
R is given by
where
5. Confidence Intervals of R
In this section, we present an asymptotic confidence interval (C.I.) of R and two C.I.s based on the non-parametric bootstrap methods.
5.1. Asymptotic Confidence Interval of R
In this subsection, we derive the asymptotic distribution of the MLE
and
. Based on the asymptotic distribution of
, the corresponding asymptotic confidence interval of
R is obtained. We denote the exact Fisher information matrix of
as
, where
,
and
L is given in (
6):
Moreover,
where
is the Gamma function.
Theorem 1. As and and , thenwhereand Proof. We can use the asymptotic properties of MLEs to prove it. ☐
Theorem 2. As and and , thenwhereand . Proof. By using Theorem 1 and the delta method, we immediately derive the asymptotic distribution of
as follows:
where
with
and
We can derive the
confidence interval for
R using Theorem 2 as
where
is 100
th percentile of
N (0, 1). The confidence interval of
R can be derived by using the estimate of
B in (
30). To estimate
B, we use the MLEs of
,
and
and the following:
5.2. Bootstrap Confidence Intervals
In this subsection, two confidence intervals based on the non-parametric bootstrap methods are proposed: (i) the percentile bootstrap method (Boot-
p) (see [
19]) and (ii) the bootstrap-t method (Boot-
t) (see [
20]).
The algorithms for conducting the confidence intervals of R are presented as follows:
- (i)
Boot-p method:
Step 1: From the sample and compute , and .
Step 2: Generate a bootstrap sample using and and generate a bootstrap sample using and . Based on and , compute the estimate of R, say .
Step 3: Repeat Step 2 NBOOT times.
Step 4: Let
be the cumulative distribution function(cdf) of
. Define
for a given
x. Thus, the approximate
C.I. of
R is given by:
Note: In this paper,
can be computed using (
14) in Step 2.
- (ii)
Boot-t method:
Step 1: From the sample and , compute , and .
Step 2: Generate a bootstrap sample
using
and
and generate a bootstrap sample
using
and
. Based on
and
, compute the estimate of
R as
and the statistic is
Step 3: Repeat Step 2 NBOOT times.
Step 4: Let
be the cumulative distribution function(cdf) of
. Define
for a given x. Thus, the approximate
C.I. of
R is given by:
Note: In this paper, can be obtained using Theorem 2 in Step 2.
6. Bayesian Inference on
In this section, the Bayes estimate of
R can be obtained under assumption that the shape parameter
and scale parameters
and
are random variables. According to the likelihood function in
Section 3, we note that
has a positive exponent and
and
have negative exponents, respectively, so we can assume that
and
have independent Inverse Gamma pdf and
follows Gamma distribution. We choose this family such that prior-to-posterior updating yields a posterior that is also in the family:
where all the hyper-parameters
and
are assumed to be known and non-negative.
The prior density function of is denoted as , and we assume that it has a Gamma (0, 1) distribution.
We have the likelihood function based on the above assumptions as
The joint density of the data,
,
and
becomes
Therefore, the joint posterior density of
,
and
given the data is
Since the expression (
35) cannot be written in a closed form, the Bayes estimate of
R and the corresponding credible interval of
R are derived adopting the Gibbs sampling technique:
The posterior pdfs of
,
and
can be obtained based on the expression
as the following:
The posterior pdf of is not known. We use the Metropolis–Hastings method with normal proposal distribution to generate random numbers from this distribution.
The algorithm of Gibbs sampling is described as follows:
Step 1: Start with an initial guess .
Step 2: Set .
Step 3: Generate form .
Step 4: Generate form .
Step 5: Using the Metropolis–Hastings method, generate from .
Generate a new candidate value from .
Set .
calculate .
Update with probability p; otherwise set .
Step 6: Compute
from (
14).
Step 7: Set .
Step 8: Repeat step 3–7, M times.
The approximate posterior mean of
R is
and the approximate posterior variance of
R is
Using the method proposed by [
21], we immediately construct the
highest posterior density (HPD) credible interval as
where
and
are the
-th smallest integer and the
-th smallest integer of
.
7. Numerical Simulations and Data Analysis
In this section, we present a Monte Carlo simulation study and a real data set to illustrate different estimation methods proposed in the preceding sections.
7.1. Numerical Simulations Study
Since we cannot compare the performances of the different methods theoretically, some simulation results to compare the performances of the different methods are presented. We mainly compute the biases and mean square errors (MSEs) of the MLEs, AMLEs and Bayes estimates. The asymptotic C.I. of R and two C.I.s based on the non-parametric bootstrap methods are obtained. We also conduct the Bayes estimates and HPD credible intervals of R. Here, we assume that . We consider sample sizes For parameter values, we assume that , and . All the results are based on the 1000 replications. For the bootstrap methods, we compute the confidence intervals based on 300 resampling. The Bayes estimates are based on 1000 sampling, namely, M = 1000. In each case, the nominal level for the C.I.s or the credible intervals is 0.95.
We also obtain the average biases and MSEs of the MLEs, AMLEs and Bayes estimates over 1000 replications in
Table 1. From
Table 1, we can find that the Bayes estimates are almost as efficient as the MLEs and the AMLEs for all sample sizes. Interestingly, in most of the cases, the MSEs of the Bayes estimate are smaller than the MSEs of the MLEs or AMLEs. We can find that the biases and MSEs of the MLEs and AMLEs are very close. As the sample size
increases, the MSEs of the estimates decrease as expected.
Table 2 reports the results of 95% asymptotic C.I. of
R, we also obtain C.I.s based on the bootstrap methods and the HPD credible interval. We represent the results of the average confidence/credible lengths and the coverage probabilities. From
Table 2, the coverage probabilities reach the nominal level 95% with the increase of the sample sizes. We observe that the MLE method is the most valid procedure to obtain the confidence intervals. The AMLEs and the Bayes estimates are the second best confidence intervals. Interestingly, we find that the HPD credible intervals provide the most highest coverage probabilities. The Boot-
p confidence intervals perform better than Boot-
t confidence intervals, in terms of coverage probabilities. One point we should know is that the bootstrap method depends on the number of resampling. For small sample sizes
, the coverage probabilities for the MLEs and AMLEs are less than the nominal levels, with the increase of sample sizes
, they perform well.
7.2. Data Analysis
We consider a real data set to illustrate the methods of inference discussed in this article. These strength data sets in
Table 3 and
Table 4 were analyzed previously by [
3,
22]. We know that if the random variable
X follows
, the random variable
has the
. Hence, we have the following data sets from the inverse Weibull distribution. These data are presented in
Table 5 and
Table 6. We analyze the data by adding 0.5 from both data sets. We fit the inverse Weibull models to the two data sets separately. The estimated shape and scale parameters, log-likelihood values, Kolmogorov-Smirnov (K-S) distances and corresponding
p-values are presented in
Table 7. The expected frequencies and the observed frequencies based on the fitted models are also presented in the
Table 8 and
Table 9. We also obtain the chi-square values of 5.9914 and 5.9915. Obviously, the inverse Weibull model fits very well to Data Set 1 and Data Set 2.
The K-S values and the corresponding
p-values in
Table 10 show that the inverse Weibull models with equal shape parameters fit reasonably well to the modified data sets. It is clear that we cannot reject the null hypothesis that the two shape parameters are equal.
Based on Equations (
14) and (
29), we provide that the MLE and AMLE of
R are
and 0.7571. The 95% confidence intervals of MLE, AMLE, Boot-
p method and Boot-
t method are obtained as (0.6917, 0.8235), (0.6911, 0.8231), (0.6993, 0.8197), (0.7015, 0.8421), respectively.
The Bayesian estimate of
R is also presented based on Equation (
36). In the previous sections, we assume that
and
have independent IG priors,
has a Gamma (0, 1) prior and
. Under certain assumptions, we can conduct the Bayesian estimate of
R as 0.7437 and the 95% HPD credible interval of
R can be obtained as (0.6690, 0.8102).
8. Inference on R with All Different Parameters
In the sections above, we assume that the shape parameters are taken to be equal. In order to expand the paper, in this section, we study the inference of
R with all different parameters. We consider the problem of estimating
under the assumption that
and
. Then, it can be easily calculated that
To computer the MLE of R, we first obtain the MLEs of
and
. Suppose
is a random sample from
and
is a random sample from
. The joint likelihood function is:
Then, the log-likelihood function is
Then, similar to the previous approaches, we can obtain the point estimates and interval estimates of R by using the MLE, AMLE and Bayesian method. Furthermore, we also can get bootstrap confidence intervals of R.
9. Conclusions
In this paper, we have addressed the estimator of for the inverse Weibull distribution. We assume independent inverse Weibull random variables with equal shape parameters but different scale parameters.
We obtain the maximum likelihood estimator of R and its asymptotic distribution. Note that MLEs do not have explicit forms, and we propose the approximate maximum likelihood estimator of R. The confidence interval of R is obtained using the asymptotic distribution. Two bootstrap confidence intervals are also obtained. By using the Gibbs sampling technique, we present the Bayesian estimator of R and the corresponding credible interval. The Metropolis–Hastings algorithm with the normal proposal distribution is used to generate random numbers from a given density function. Monte Carlo simulations are conducted to compare the proposed methods. Analysis of a real dataset is performed. In the future, we will consider the MLE, AMLE, asymptotic C.I., bootstrap C.I.s and Bayesian inference of for inverse Weibull distribution based on incomplete data such as progressively type-II censored samples.