Bayesian Estimation for the Doubly Censored Topp Leone Distribution using Approximate Methods and Fuzzy Type of Priors

Department of Statistics, The University of Azad Jammu and Kashmir, Muzaffarabad, Pakistan Department of Mathematics and Statistics, College of Science, King Faisal University, P.O. Box 400, Al-Ahsa 31982, Saudi Arabia Department of Statistics, COMSATS University Islamabad-Lahore Campus, Pakistan Department of Mathematics, Islamic University, Kushtia-7003, Bangladesh Department of Applied Statistics and Insurance, Faculty of commerce, Mansoura University, Mansoura 35516, Egypt


Introduction
Topp and Leone [1] introduced a lifetime distribution and named it TLD. Nadarajah and Kotz [2] have discussed that the hazard rate function of the TLD is U-shaped. The distributions with U-shaped hazard rate are used to model the human populations where the death rate is high at the infant age owing to infant diseases and birth defects; thereafter, the death rate remains constant up to thirties, and then, it increases again. Some of the manufactured items also have similar life patterns. The advantage of TLD is that it has only two parameters, while most of the distributions with U-shaped hazard rate have at least three parameters which increases the computational difficulties. The regularity conditions are satisfied by the TLD. The distribution function of the TLD is in a compact form, due to which it can be very easily applied when the data are censored, unlike the gamma and lognormal distributions. Another feature of the distribution, which distinguishes it from the other lifetime distributions, is that it has a finite support. Due to these distinguishing aspects, the TLD has received much attention from the researchers recently. Al-Zahrani and Alshomrani [3] analyzed the stress strength reliability for the TLD. Genc [4] derived the expressions for single and product moments for the order statistics from the TLD. Genc [5] considered the estimation of P ðX > YÞ from the TLD. Mir Mostafaee et al. [6] discussed the estimation for moments of order statistics from the TLD. Bayoud [7] estimated the shape parameter of the TLD using the Bayesian and classical methods under type-I censored samples. Bayoud [8] used progressive type-II censored samples to obtain the Bayesian and classical estimates for the shape parameter of the TLD. The approximate maximum likelihood method, LA, and importance sampling method have been used for the estimation. Mir Mostafaee et al. [9] studied the lower krecord values from the TLD using the Bayesian approach. Reyad and Othman [10] used the Topp Leone genesis to produce and generalization of the Burr-XII distribution and called it Topp-Leone Burr-XII distribution. Different properties and estimation of the proposed model have been discussed. The effectiveness of the proposed model has been demonstrated using three real-life datasets. Rezaei et al. [11] introduced a new family of distributions from the TLD and discussed the mathematical properties of this family of distributions.
The U-shaped hazard rate with a fewer number of parameters and finite support for the TLD distinguishes it from repeatedly used lifetime models such as Weibull, Gamma, and Log-normal distributions. Motivated by these features of the TLD, we planned to estimate this distribution under a Bayesian framework. The reason for the choice of Bayesian estimation is that the results under Bayesian inference are often better than the classical methods even if we do not have sufficient prior information; for example, please see Kundu and Joarder [12]. Further, the lifetime data are often censored. Therefore, we have considered doubly censored samples from the TLD for the estimation. Doubly censoring is used to analyze the duration times between two events. For example, let the ages of certain equipment are not known when the monitoring of the equipment starts and tracking of the equipment is stopped after the predetermined observational period. Then, the ages of the equipment which are still functioning will be doubly censored. The doubly censored samples are very useful in reliability analysis for the products which are already under use. In medical applications, the doubly censored samples usually deal with time between infection and onset of a disease; for example, the infectious diseases like COVID-19 often produce doubly censored data. Although Feroze and Aslam [13] and Feroze and Aslam [14] have considered the analysis of doubly censored samples from the single and mixture of TLD, respectively, these contributions considered only single parametric TLD which restricted the flexibility of the TLD to a great extent. In addition, these contributions did not include any credible intervals.
In addition, the Bayesian methods are very much linked to the Fuzzy sets. The Fuzzy priors can be used as prior information in the Bayesian models. The fuzzy approximations enable users to incorporate more flexible sets of priors and likelihood functions in the Bayesian inference. This method allows authors to bypass the closed form conjugacy link between likelihood function and prior [15]. Stegmaier and Mikut [16] suggested that the fuzzy priors can improve the performance of the Bayesian inference. This paper considers the posterior analysis of TLD, when the samples are doubly censored. The independent informative priors (IPs) which are very close to the Fuzzy priors have been proposed for the analysis. It has been noticed that the expressions for the posterior estimators are not available in the closed form, so we have used four approximation techniques, namely, QM, LA, TKA, and GS, for the analysis. The assumption of independent gamma priors has been made for both of the parameters; this assumption is not rare; for example, see Kundu and Howlder [17]. Further symmetric and asymmetric loss functions have been considered for the estimation.
The rest of the paper is placed in the following sections. The model and likelihood function have been introduced in Section 2. The Bayesian estimation has been considered in Section 3. The results regarding the simulation study have been reported in Section 4. The real-life example has been presented in Section 5. The concluding remarks have been given in Section 6.
In the Bayesian QM, we choose a set of points between the finite integral in an order to ensure the stability of our uncertainty. Consider the posterior densitygðθ 1 , θ 2 jxÞ, where θ 1 and θ 2 are the parameters. We evaluate this density over a number of the points in the entire range as where w i are the increments. The Mathematica software has been used to obtain BEs and associated PRs for the parameters θ 1 and θ 2 using SELF and PLF under IPs.

Lindley's Approximation (LA).
The QM is not suitable in many situations, for example, in the case of functions with singularities; the method is not well suited. In such situations, some other approximation methods, such as LA, can be used. The LA has an important role in Bayesian inference. Using this method, we can compute BEs without performing any complex numerical integration. Hence, if we need to obtain only the BEs, the LA can be efficiently used to serve this purpose. Bayoud [8] used LA to estimate the shape parameter of the TLD considering uncensored data. We have considered the more general and complex case by performing the LA for both parameters of the TLD using censored data. Assuming a sufficiently large sample, Lindley [21] proposed that any ratio of the integral of the form where hðθ 1 , θ 2 Þ is any function of θ 1 or θ 2 , lðθ 1 , θ 2 jxÞ is the log-likelihood function, and Gðθ 1 , θ 2 Þ is the logarithmic of joint prior for the parameters θ 1 and θ 2 , can be evaluated as 3 Journal of Function Spaces where b θ 1 and b θ 2 are MLEs of the parameters θ 1 and θ 2 , respectively, and σ ij is the ði, jÞ th element of the inverse of the matrix fL ij g, all evaluated at the MLEs of the parameters. Now, the log-likelihood function from (4) can be obtained as The maximum likelihood estimates (MLEs) of the parameters θ 1 and θ 2 can be obtained by differentiating (13) with respect to θ 1 and θ 2 and equating to zero, respectively, as Let Based on (14) and (15), the approximate MLEs have been obtained using numerical methods. The second order derivatives of the log-likelihood function are presented in the following: Now, ((16)-(18)) have been evaluated at the MLEs of θ 1 and θ 2 .
As the third order derivatives with respect to θ 1 and θ 2 contain long expressions, therefore, they have not been presented in the paper.
Based on the second order derivatives, the matrix fL ij g is Now, using LA, the BEs for the parameters θ 1 and θ 2 under IP using SELF are, respectively, presented as Journal of Function Spaces Again, using LA, the BEs for the parameters θ 1 and θ 2 under IP using PLF are, respectively, presented as 3.5. Tierney and Kadane's Approximation (TKA). The LA requires the evaluation of the third derivatives from the log-likelihood function which at times gets tedious, especially in case of problems with several parameters. This issue can be resolved by considering rather easily computable approximation called TKA. The additional advantage of this approximation is that it has less error as compared to LA. Therefore, in this subsection, we have considered TKA for the approximate Bayes estimation of the parameters from the doubly censored TLD. Consider Qðθ 1 ,
3.6. Gibbs Sampler. Unfortunately, by using QM, Lindley's method, and TKA, it is impossible to obtain the highest posterior density (HPD) credible intervals. The HPD credible intervals can be obtained by using the GS. Hence, in this subsection, we have used the GS to obtain the approximate BEs along with HPD credible intervals.
Consider a PD with two parameters θ 1 and θ 2 . Assuming that the full densitiesgðθ 1 jθ 2 , xÞ and gðθ 2 jθ 1 , xÞ are extractable, we need to obtain gðθ 1 jxÞ and gðθ 2 jxÞ. Using GS, we start with choosing initial guess values for parameters θ 1 and θ 2 . The said initial values are denoted by θ 10 and θ 20 .
Then, we generate random samples from the conditional distributions in the following sequence: In order to employ the GS for the PD (7), we extract the conditional distribution for θ 1 given θ 2 under IP as  The conditional distribution of the parameter θ 2 given θ 1 under IP is Now, using the conditional distributions from (29) and (30), the GS can be implemented following the methodology proposed by Pandey and Bandyopadhyay [23] using Winbugs software. The generated samples for the parameters θ 1 and θ 2 can be used for the estimation of the parameters using SELF and PLF. The BE and the posterior risks (PRs) for the parameter θ 1 using SELF and IP can be obtained by using the formulae θ 1,S = ∑ m i=1 θ 1,i /m and ρðθ 1,S Þ = ∑ m i=1 ðθ 1,i − θ 1,S Þ 2 , respectively. Similarly, the BE and PR for the parameter θ 1 using PLF and IP can be calculated by using the ,i /mÞg, respectively. A similar process can be followed to implement GS for the PD under IP and for the PDs for the parameter θ 2 .

Simulation Study
As the performance of the BEs cannot be compared analytically, we have carried out a simulation study to compare different estimators numerically. The comparison among the BEs is based on the level of the convergence of the estimates and the amounts of the PRs. The following parametric space has been used for the estimation: ðθ 1 , θ 2 Þ = fð0:50, 0:50Þ, ð1, 1Þ, ð1:50, 1:50Þ, ð2, 2Þg and ðθ 1 , θ 2 Þ = fð0:50, 1Þ, ð1 , 0:5Þ, ð1, 1:50Þ, ð1:50, 1Þg. The samples of sizes n = 15, 30, 50, and 100 have been considered. The values for the hyperparameters have been considered in such a way that, for these values, the mean of the concerned prior is equal to the true parametric value. The samples have been assumed to be 20% censored: 10% from the left and 10% from the right. The test termination point in each sample is assumed to be so that observed the said censoring rate. The results for the simulated samples have been presented in Tables 1-8. The results using parametric space ðθ 1 , θ 2 Þ = fð0:50, 0:50Þ, ð1, 1, 1:50, 1:50, 2, 2Þg, using samples of sizes 15, 30, 50, and 100, have been reported in Tables 1-4, respectively. Similarly, the results for sample sizes 15, 30, 50, and 100, using parametric space ðθ 1 , θ 2 Þ = fð0:50, 1, 1, 0:5, 1, 1:50, 1:50, 1Þg, have been given in Tables 5-8, respectively. The results using simulated samples have been presented in Tables 1-10. The said results advocate that the proposed estimates are consistent in nature because the estimated values of the parameters converged to the true parametric values with increase in the sample size. The performance of the BEs has been better as compared to ML counterparts. The convergence under BEs was better as compared to MLEs. In additions, the amounts of PRs associated with each estimate become smaller as sample size increases. The performance of loss functions and different approximation methods has been compared on the basis of amounts of PRs associated with each estimate.
It is interesting to note that for smaller choice of true parametric values, the SELF performs better than PLF. Conversely, the performance of PLF seems better for estimation. There is a sense of competition among the performance of different approximation methods; however, the TKA provides the minimum amounts of the PRs among all the approximation methods with few exceptions. The amounts of the PRs are directly proportional with the magnitude of the true parametric values which is in accordance with the theory. In general, the parameter (θ 1 ) has been estimated more efficiently than the parameter (s).

Real-Life Example
In this section, the data regarding the failure times (in minutes) for a specific type of electrical insulation has been used for the illustration of the results obtained in the previous section. The data has been used by Fernandez [24] and is reported by Lawless [25]. The twelve items have been  11 Journal of Function Spaces considered for the experiment out of which the experimenter was unable to observe the smallest two failure times and experiment was terminated after the ninth failure. The data is as follows: -, -, 24.4, 28.6, 43.2, 46.9, 70.7, 75.3, 95.5, -, -, and -. Therefore, we have n = 12, r = 3, s = 9, and m = 7. The descriptive statistics for the data are mean = 54:94, variance = 689:50, standard error = 9:92, skewness = 0:39, and kurtosis = −1:17. The data has been censored in such a way that same number of observations has been censored from left and right. We have named this dataset as dataset-1. The other dataset used in the study has been named as dataset-2. The dataset-2 represents the failure times of the air condition system of an airplane, reported by Linhart and Zucchini [26]. The dataset-2 consists of observations 1, 3, 5, 7, 11(3), 12, 14(3), 16(2), 20, 21, 23, 42, 47, 52, 62, 71(2), 87, 90, 95, 120(2), 225, 246, and 261. The number in the parenthesis represents the repeated observations. The chi-square and the Kolmogorov-Smirnov tests have been used to confirm that the data can be modeled using TLD. The BEs and corresponding PRs, based on these data, have been presented in Tables 11 and 12, respectively. For convenience, the numerical values for the hyperparameters have been assumed to be a = b = c = d = 1 for the estimation.
Tables 11 and 12 contain the BEs and associated PRs using different approximation methods. The amounts for the PRs under PLF are on average smaller to those under SELF. As far as the approximation is concerned, the performance of the TKA seems superior to its counterparts, as the least magnitudes of the PRs are associated with these estimates. Hence, the results obtained by analyzing the real data are in agreement to those obtained from simulated samples. The marginal density plots for model parameters under different approximation methods have been presented in  For brevity, we have presented the density plots only for dataset-1. The density plots for parameter θ 1 are positively skewed. However, the skewness is not that high for density curves of parameter θ 2 . The density plots under PLF are representing slightly more skewness as compared to those for SELF. From these graphs, it can also be seen that the graphs for marginal densities under SELF and PLF are relatively closer under TKA for both parameters. However, the graphs for marginal densities using other approximation techniques, such as QM, LA, and GS, are quite different under SELF and PLF. This simply suggests that the estimates under TKA are relatively robust with respect to choice of loss functions.

Conclusion
This paper considers the posterior analysis for the parameters of the TLD assuming doubly censoring. As the expressions for the BEs cannot be obtained explicitly, we considered different approximation methods to obtain the numerical solutions. The performance of the BEs is compared based on amounts of PRs. The analysis of simulated and real datasets suggests that there is a sense of competition among different approximation methods; however, the performance of the TKA is the best among the approximate methods used in the study. The study can be extended for application of TLD in modeling doubly censored heterogeneous data.

Data Availability
The data used in the paper are available in the paper.

Conflicts of Interest
The authors declare that they have no conflicts of interest. 14 Journal of Function Spaces