Statistical literature contains a large number of probability distributions for modeling real lifetime data, the most popular probability distributions are gamma (Ga), lognormal (Log-N), Weibull (W), and exponentiated exponential (Exp-E) distributions. However, these probability models suffer from some drawbacks. First, none of them exhibit bathtub shapes for their hazard rate functions (hrfs), the four models exhibit only monotonically decreasing, monotonically increasing or constant hrfs and this is a major weakness since most real life systems exhibit bathtub shapes for their hrfs. Second, at least three of these distributions exhibit constant hazard rates and this is an unrealistic feature since few real life systems have constant hrf. This work introduces a new three parameter lifetime distribution as an alternative to the Ga, Log-N, W, and the Exp-E probability models that does not have the above mentioned drawbacks. The new model, called the Topp Leone Generated Lindley (TLGLi), is constructed based on the Topp Leone Generated (TLG) family introduced by Rezaei
The cumulative distribution function (cdf) and the probability density function (pdf) of a random variable (r.v.)
and
respectively. The scale parameter
and the corresponding pdf is
Equations (1.1) and (1.2) are established based on the TLG distribution.
Suppose
and represent the failure times of the components of a series system, assumed to be independent. Then the probability that the system will fail before time
So, Equation (1.1) is the distribution of the failure of a series system with independent components if
where
and
is the cdf of the Exp-Li distribution with power parameter (
where
is the Exp-Li density with power parameter (
We provide some plots of the pdf and hrf of the TLGLi model to show its flexibility. Figure 1 displays some plots of the TLGLi density for selected values of
In the literature, certain generalizations on the Li distribution are proposed and studied (Ghitany
The rest of the paper is outlined as follows: Certain characterizations of the TLGLi model are proposed in Section 2. In Section 3, we discuss some properties of this distribution. In Section 4, we describe four methods of estimation. In Section 5, the usefulness of the new distribution is illustrated by means of two real data sets. A modified goodness-of-fit test using a Nikulin-Rao-Robson (NRR) statistic test is presented in Section 6. In Section 7, a simulation study is carried out to compare the performance of the four methods of estimation. Section 8 offers some concluding remarks.
In this Section, we present a number of characterizations of the TLGLi distribution in the following directions: (i) in terms of the ratio of two truncated moments; (ii) in terms of the hazard function; (iii) based on the reverse hazard function; and (iv) in terms of the conditional expectation of certain function of the random variable. These characterizations are presented in four subsections.
Certain characterizations of TLGLi distribution based on a simple relationship between two truncated moments are presented. The first characterization employs a theorem due to Glänzel (1987), see Hamedani
Suppose the random variable
and
Further,
Conversely, if
and
Now, according to Theorem 1,
The hazard function,
For many univariate continuous distributions, the above equation is the only differential equation available in terms of the hazard function. In this subsection we present a non-trivial characterization of TLGLi distribution, for
Is straightforward and hence omitted.
The reverse hazard function,
In this subsection we present a characterization of TLGLi distribution, for
Is straightforward and hence omitted.
Here, we employ a function
For (
The
where
and
The first four moments of
The skewness and kurtosis measures can be calculated from the ordinary moments using well-known relationships.
The
The moment generating function (mgf)
The characteristic function (cf) of
and
respectively, where
The
where
and
where 1
The
The
Therefore
The
or
Then, the
The reliability, say
where
Let
where
where
and
where
Let
where
The maximum likelihood estimators (MLEs) of
where
Setting them equal to zero and solving the system simultaneously yields the MLE
The theory of ordinary least square (OLS) estimation and weighted least square (WLS) estimation was originally proposed by Swain
Suppose that
then
The OLSE of the parameters (
where
The OLSE of the parameters (
The WLSE of the parameters are obtained by solving the following non-linear equations
where
The Cramer-Von- Mises estimation method of the parameters is based on the theory of minimum distance estimation (MacDonald, 1971). The Crammer-Von Mises estimates (CVME) of the parameter
and
The CVME of the parameters are obtained by solving the following non-linear equations
and
where
Using the squared error loss function (SELF), the Bayes estimators are computed under informative gamma priors for all
where the hyperparameters
and
Under this loss function, posterior mean is the Bayes estimate of the respective parameter. Thus, the Bayes estimators under SELF are obtained as
where
respectively. The above equations cannot be solved analytically, thus we use Markov chain Monte Carlo (MCMC) technique to generate the posterior sample from the full conditional posterior distribution. The full conditional posterior distributions are given by
The following steps are used to extract the posterior samples from full-conditional posterior density
Starts with
Generate posterior samples from full conditional distribution using normal distribution as a proposal density;
Repeat the above step for
Under SELF, the Bayes estimates of
where,
In this section, real data sets are used to demonstrate the real life applicability of the proposed model using MLE, LS, WLS, and CVM. We consider the Cramér-Von Mises (
and
where
where
The first data with size 63 shows the strength measured in GPa for single carbon fibers and impregnated at gaugelengths of 20mm. The data are:
1.9010, 2.132, 2.203, 2.2280, 2.257, 2.350, 2.361, 2.3960, 2.397, 2.445, 2.4540, 2.474, 2.518, 2.5220, 2.525, 2.5320, 2.575, 2.614, 2.616, 2.6180, 2.624, 2.659, 2.6750, 2.738, 2.740, 2.8560, 2.917, 2.928, 2.9370, 2.937, 2.9770, 2.9960, 3.030, 3.125, 3.139, 3.1450, 3.220, 3.223, 3.2350, 3.243, 3.264, 3.272, 3.2940, 3.332, 3.346, 3.377, 3.408, 3.4350, 3.493, 3.501, 3.537, 3.554, 3.5620, 3.628, 3.852, 3.871, 3.886, 3.971, 4.024, 4.027, 4.225, 4.395, 5.020.
In general, the smaller value of KS, the better fit to the data. Table 2 gives the values of estimators of
From Table 2 we conclude that the LS method is the best method for modelling the carbon fibers with KS = 0.06427,
The second data set represents on the relief times of twenty patients receiving an analgesic. The data are:
1.10, 1.4, 1.30, 1.7, 1.90, 1.8, 1.6, 2.20, 1.7, 2.70, 4.1, 1.80, 1.5, 1.20, 1.4, 30, 1.7, 2.30, 1.6, 2
From Table 3 we conclude that the CVM method is the best method for modelling the relief times data with KS = 0.09269,
The fitted density plot, the relative histogram plot with the fitted density of the proposed model and fitted survival function for data-II are given in Figure 3. Afterall, from Figures 3 and 4, it has been noticed that the proposed model fitted well to the considered data set.
The NRR statistic (
and
The boundaries of the intervals
The NRR statistic, with
The Pearson’s statistic
where vector of probabilities
The
Under the null hypothesis
We test the following null hypothesis
where
Data was reported by Badar and Priest (1982) with size 63 shows the strength measured in GPa for single carbon fibers and impregnated at gaugelengths of 20mm. We choose
We can aver that
since
thus, the strength measured in GPa for single carbon fibers follows a TLGLi distribution with
The data constitutes twenty observations of the lifetime data relating to relief times (in minutes) of patients receiving an analgesic. This data was reported by Gross and Clark (1975). We take
The NRR statistic tests
We conclude that we have a concordance of the analgesic failure time data and our TLGLi model.
A MCMC simulation study is conducted in this section, to compare the performance of the different estimators of the unknown parameters of the TLGLi distribution. This performance is evaluated regarding their mean squared errors (MSEs). Computations in this section are done by ‘Mathcad program Version 15.0’. We generate 1,000 samples of the TLGLi distribution, where
This paper introduces a new extension of the Lindley model. The estimation of the parameters is carried out via different methods. Bayes estimation is computed under gamma informative prior under the squared error loss function. The performances of the proposed estimation methods are studied through Monte Carlo simulations. The potentiality of the proposed model is analyzed through two data sets. A modified goodness-of-fit test using the NRR statistic test is investigated via two examples. Certain characterizations of the proposed distribution are presented. A modified goodness-of-fit test for the new model in complete data case is investigated via two examples. We propose the construction of a modified chi-squared goodness of fit statistic test for the new TLGLi model in complete data case. The new test is based on the NRR statistic separately proposed by Nikulin (1973) and Rao and Robson (1974). As a second step, an application to real data has been proposed to show the applicability of the proposed test and the new TLGLi model for modeling different data sets.
Let (Ω,ℱ ,
is defined with some real function
where the function
We like to mention that this kind of characterization based on the ratio of truncated moments is stable in the sense of weak convergence (Glänzel, 1990), in particular, let us assume that there is a sequence {
This stability theorem makes sure that the convergence of distribution functions is reflected by corresponding convergence of the functions
A further consequence of the stability property of Theorem 1 is the application of this theorem to special tasks in statistical practice such as the estimation of the parameters of discrete distributions. For such purpose, the functions
In some cases, one can take
We, however, believe that employing three functions
Plots of the probability density function of the Topp Leone Generated Lindley for selected parameter values.
Plots of the hazard rate function of the Topp Leone Generated Lindley for selected parameter values.
Fitted density, the histograms with the fitted density of the TLGLi distribution for various methods and fitted survival function for data I. TLGLi = Topp Leone Generated Lindley; MLE = maximum likelihood estimator; LS = Least square; WLS = weighted least square; CVM = Crammer-Von Mises.
Fitted density, the histograms with the fitted density of the TLGLi distribution for various methods and fitted survival function for data II. TLGLi = Topp Leone Generated Lindley. TLGLi = Topp Leone Generated Lindley; MLE = maximum likelihood estimator; LS = least square; WLS = weighted least square; CVM = Crammer-Von Mises.
Mean, variance, skewness and kurtosis of the Topp Leone Generated Lindley distribution with different values of parameters
| Variance | Skewness | Kurtosis | ||||
|---|---|---|---|---|---|---|
| 0.5 | 0.5 | 0.5 | 0.5283 | 0.9050 | ||
| 1.5 | 1.7810 | 2.6030 | 1.4835 | 5.9315 | ||
| 3.0 | 3.0484 | 3.5869 | 1.0063 | 4.3890 | ||
| 5.0 | 4.1345 | 4.0290 | 0.8156 | 3.9783 | ||
| 10 | 5.7107 | 4.2990 | ||||
| 1.0 | 1.5 | 0.5 | 0.5165 | 0.3515 | 2.1602 | 9.6757 |
| 1.5 | 1.4047 | 0.6927 | 1.1384 | 4.9824 | ||
| 3.0 | 2.1364 | 0.8069 | 0.8925 | 4.3499 | ||
| 5.0 | 2.7128 | 0.8435 | 0.8038 | 4.1739 | ||
| 10 | 3.5140 | 0.8562 | 0.7464 | 4.0793 | ||
| 2.0 | 2.5 | 0.5 | 0.3065 | 0.0882 | 1.9651 | 8.7477 |
| 1.5 | 0.7665 | 0.1568 | 1.1332 | 5.0869 | ||
| 3.0 | 1.1286 | 0.1785 | 0.9343 | 4.5435 | ||
| 5.0 | 1.4109 | 0.1858 | 0.8612 | 4.3793 | ||
| 10 | 1.8025 | 0.1887 | 0.8127 | 4.2841 | ||
| 4.0 | 5.0 | 0.5 | 0.2006 | 0.0229 | 1.6979 | 7.5324 |
| 1.5 | 0.4420 | 0.0351 | 1.1204 | 5.1709 | ||
| 3.0 | 0.6206 | 0.0386 | 0.9823 | 4.7674 | ||
| 5.0 | 0.7580 | 0.0399 | 0.9301 | 4.6324 | ||
| 10 | 0.9478 | 0.0406 | 0.8939 | 4.5461 | ||
| 10 | 10 | 0.5 | 0.1006 | 0.0036 | 1.5163 | 6.8066 |
| 1.5 | 0.1982 | 0.0049 | 1.1258 | 5.2763 | ||
| 3.0 | 0.2674 | 0.0053 | 1.0308 | 4.9781 | ||
| 5.0 | 0.3201 | 0.0055 | 0.9936 | 4.8694 | ||
| 10 | 0.3927 | 0.0056 | 0.9663 | 4.7934 | ||
The values of estimators, KS,
| Method | KS | ||||||
|---|---|---|---|---|---|---|---|
| MLE | 19.04243 | 3.126965 | 1.249401 | 0.08472 | 0.7564 | 0.0658 | 0.3444 |
| LS | 18.06670 | 2.310553 | 1.128924 | ||||
| WLS | 21.02794 | 2.369555 | 1.168441 | 0.07204 | 0.8993 | 0.0640 | 0.3372 |
| CVM | 21.01143 | 2.227783 | 1.145041 | 0.06687 | 0.9408 | 0.0635 | 0.3353 |
KS = Kolmogorov-Smirnov;
The values of estimators, KS,
| Method | KS | ||||||
|---|---|---|---|---|---|---|---|
| MLE | 83.98383 | 0.450189 | 1.322298 | 0.13861 | 0.8369 | 0.0578 | 0.3399 |
| LS | 87.57996 | 0.45128 | 1.389133 | 0.11138 | 0.9651 | 0.0577 | 0.3396 |
| WLS | 89.28296 | 0.41184 | 1.316787 | 0.10808 | 0.9736 | 0.0591 | 0.3477 |
| CVM | 101.4514 | 0.52226 | 1.499851 |
KS = Kolmogorov-Smirnov;
Average values of estimates and mean squared errors (in parentheses) for
| Parameters | MLE | LS | WLS | CVM | Bayesian |
|---|---|---|---|---|---|
| 2.3532 (0.38887) | 2.1393 (2.08396) | 1.8615 (1.17179) | 1.8403 (0.78797) | 1.9825 (0.36291) | |
| 1.0345 (0.05332) | 1.0082 (0.28217) | 1.0695 (0.37182) | 0.9998 (0.27709) | 0.9725 (0.05102) | |
| 1.4189 (0.04491) | 1.6744 (0.49578) | 1.3668 (0.56632) | 1.8278 (0.75370) | 1.4062 (0.02231) | |
| 1.3733 (0.14894) | 1.7486 (1.73004) | 1.6431 (1.47016) | 1.6111 (0.89955) | 1.0090 (0.11238) | |
| 1.3593 (0.07589) | 1.5477 (0.49841) | 1.4652 (0.53282) | 1.5500 (0.44179) | 1.1924 (0.06207) | |
| 0.4919 (0.00380) | 0.5382 (0.14299) | 0.6003 (0.31093) | 0.5259 (0.14304) | 0.4726 (0.00175) | |
| 2.6570 (0.56914) | 2.4663 (1.45153) | 2.2088 (1.86197) | 2.4313 (1.59231) | 2.5121 (0.53294) | |
| 0.8241 (0.02121) | 0.9339 (0.10929) | 0.9314 (0.21013) | 0.9520 (0.14371) | 0.8321 (0.02019) | |
| 0.3012 (0.00128) | 0.3745 (1.19557) | 0.3345 (0.06211) | 0.5054 (3.00787) | 0.2752 (0.00089) | |
MLE = maximum likelihood estimator; LS = least square; WLS = weighted least square; CVM = Crammer-Von Mises.
Average values of estimates and mean squared errors for
| Parameters | MLE | LS | WLS | CVM | Bayesian |
|---|---|---|---|---|---|
| 2.2743 (0.15772) | 1.8969 (0.39803) | 1.8156 (0.93618) | 1.8654 (0.38163) | 1.9625 (0.13194) | |
| 1.0101 (0.02351) | 0.8766 (0.06023) | 1.0299 (0.32983) | 0.9116 (0.18209) | 0.8562 (0.02127) | |
| 1.4205 (0.02041) | 1.6965 (0.32712) | 1.4223 (0.52627) | 1.8211 (0.59441) | 1.2213 (0.01749) | |
| 1.3246 (0.05230) | 1.6139 (0.54912) | 1.6080 (1.11703) | 1.6155 (0.65038) | 1.1230 (0.05067) | |
| 1.3331 (0.02870) | 1.4953 (0.18328) | 1.4126 (0.51303) | 1.5109 (0.29059) | 1.1960 (0.02513) | |
| 0.4908 (0.00147) | 0.4684 (0.01874) | 0.5942 (0.29852) | 0.4873 (0.09689) | 0.4238 (0.00071) | |
| 2.5587 (0.17512) | 3.1008 (1.41519) | 2.3333 (1.73189) | 2.7694 (0.97663) | 2.5670 (0.16327) | |
| 0.8084 (0.00710) | 0.9115 (0.04340) | 0.9317 (0.14571) | 0.9154 (0.05181) | 0.7865 (0.00591) | |
| 0.3009 (0.00048) | 0.3104 (0.36087) | 0.3183 (0.06067) | 0.3284 (1.04234) | 0.3421 (0.00029) | |
MLE = maximum likelihood estimator; LS = least square; WLS = weighted least square; CVM = Crammer-Von Mises.
Average values of estimates and mean squared errors for
| Parameters | MLE | LS | WLS | CVM | Bayesian |
|---|---|---|---|---|---|
| 2.2659 (0.09170) | 1.7649 (0.17525) | 1.8554 (0.79055) | 1.8139 (0.17631) | 1.9920 (0.08923) | |
| 1.0084 (0.01832) | 0.8492 (0.05150) | 0.9972 (0.27431) | 0.8493 (0.05137) | 0.9321 (0.01652) | |
| 1.4113 (0.01541) | 1.6787 (0.17699) | 1.4944 (0.47237) | 1.6295 (0.19440) | 1.4621 (0.01469) | |
| 1.3222 (0.02832) | 1.5396 (0.27328) | 1.6031 (0.93027) | 1.5404 (0.27639) | 1.0951 (0.02343) | |
| 1.3334 (0.01804) | 1.4637 (0.09777) | 1.5482 (0.48893) | 1.4641 (0.09843) | 1.1970 (0.01601) | |
| 0.4885 (0.00082) | 0.4587 (0.00348) | 0.5638 (0.27453) | 0.4631 (0.01789) | 0.4235 (0.00063) | |
| 2.5538 (0.09009) | 2.9853 (0.67365) | 2.4582 (1.57146) | 2.9866 (0.68129) | 2.6120 (0.07321) | |
| 0.8083 (0.00367) | 0.8962 (0.02395) | 0.9102 (0.10320) | 0.8963 (0.02412) | 0.7126 (0.00296) | |
| 0.2997 (0.00024) | 0.2826 (0.00084) | 0.3148 (0.05267) | 0.2826 (0.00084) | 0.3712 (0.00018) | |
MLE = maximum likelihood estimator; LS = least square; WLS = weighted least square; CVM = Crammer-Von Mises.
Average values of estimates and mean squared errors for
| Parameters | MLE | LS | WLS | CVM | Bayesian |
|---|---|---|---|---|---|
| 2.0022 (0.02182) | 1.7879 (0.10016) | 2.1824 (0.61904) | 1.7845 (0.09545) | 1.8921 (0.00861) | |
| 1.0113 (0.01584) | 0.8332 (0.01154) | 0.9739 (0.07393) | 0.8316 (0.01086) | 0.7829 (0.01329) | |
| 1.4067 (0.01229) | 1.6824 (0.09175) | 1.6873 (0.18729) | 1.6813 (0.08527) | 1.5700 (0.01017) | |
| 1.3189 (0.02400) | 1.4896 (0.13249) | 1.4998 (0.61342) | 1.4914 (0.13387) | 1.1921 (0.02091) | |
| 1.3331 (0.01279) | 1.4420 (0.05639) | 1.5302 (0.37676) | 1.4431 (0.05649) | 1.3624 (0.00972) | |
| 0.4875 (0.00051) | 0.4576 (0.00268) | 0.4958 (0.07906) | 0.4573 (0.00264) | 0.3905 (0.00047) | |
| 2.5496 (0.04815) | 2.9140 (0.34917) | 2.4723 (1.41766) | 2.9138 (0.32176) | 2.6412 (0.02938) | |
| 0.8080 (0.00195) | 0.8860 (0.08602) | 0.9093 (0.08668) | 0.8863 (0.01300) | 0.9214 (0.00072) | |
| 0.2991 (0.00012) | 0.2819 (0.00059) | 0.2911 (0.01006) | 0.2819 (0.00057) | 0.2764 (0.00011) | |
MLE = maximum likelihood estimator; LS = least square; WLS = weighted least square; CVM = Crammer-Von Mises.