Preprint
Article

This version is not peer-reviewed.

Generalized Bayes Prediction Study Based on Jointly Type-II Censored Sample from K-Exponential Populations

A peer-reviewed article of this preprint also exists.

Submitted:

17 June 2023

Posted:

19 June 2023

You are already at the latest version

Abstract
In this paper, the problem of predicting future failure times based on a jointly type-II censored sample from k exponential populations is considered. The Bayesian prediction intervals and point predictors were then obtained. Generalized Bayes is a Bayesian study based on a learning-rate parameter. This study investigated the effects of the learning rate parameters on the prediction results. The loss functions of squared error, Linex, and general entropy were used as point predictors. Monte Carlo simulations were performed to show the effectiveness of the learning rate parameter in improving the results of prediction intervals and point predictors.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  

1. Introduction and motivations

Generalized Bayes is a Bayesian study based on a learning rate parameter ( η > 0 ) as a power of the likelihood function. The traditional Bayes framework is obtained for η = 1 , and we demonstrate the effect of the learning rate parameter on the prediction results. That is, if the prior distribution of the parameter θ is π ( θ ) then the generalized Bayes posterior distribution for θ is
π * θ     d a t a ) L η θ ; d a t a   π ( θ ) ,         θ Θ ,         η > 0 .
For more details on the generalized Bayes method and the choice of the value of the rate parameter we refer readers to [1-11]. An exact inference method based on maximum likelihood estimates (MLEs) was developed in [12], and its performance was compared with that of approximate, Bayesian, and bootstrap methods. The joint progressive censoring type II and the expected number of failures for two populations under the joint progressive censoring type II were introduced and studied by [13]. In contrast, exact likelihood inference for two exponential populations under joint progressive censoring of type II was studied in [14] and some precise results were obtained based on the maximum likelihood estimates developed by [15]. Exact likelihood inference for two populations of two-parameter exponential distributions under type II joint censoring was studied by [16].
One might be interested in predicting future failures using a joint type II censored sample. To accomplish this, prediction points or intervals should be determined. Bayesian prediction bounds for future observations based on certain distributions have been discussed by several authors. A study of Bayesian estimation and prediction based on a joint censored sample of type II from two exponential populations was presented by [17]. Prediction (various classical and Bayesian point predictors) for future failures in the Weibull distribution under hybrid censoring was studied by [18]. A Bayesian prediction based on generalized order statistics with multiple censoring of type II was developed by [19].
The main objective of this study is to predict future failures based on a joint type-II censoring scheme for k-exponential populations when censoring is performed on k-samples in a combined manner. Suppose that products from k different lines are produced in the same factory, and k independent samples of size n h , 1 h k are selected from these k lines and simultaneously placed in a lifetime experiment. To reduce the cost and time of the experiment, the experimenter may decide to stop the lifetime test when a certain number ( r ) of failures occurs. The nature of the problem and the distributions used in our study are presented below.
Suppose { X h n h , h = 1 , . . . , k } are k -samples, where X h n h = { X h 1 , X h 2 , . . . , X h n h } are the lifetimes of n h samples of product line A h and are assumed to be independent and identically distributed (iid) random variables from a population with a a probability density function (pdf) f h ( x ) and a cumulative distribution function (cdf) F h ( x ) .
Furthermore, let N = i = 1 k n i be the total sample size and let r be the total number of observed failures. Let W 1 . . . W N denotes the order statistics of N random variables, { X h n h , h = 1 , . . . , k } . Under the joint type- II censoring scheme for the k -samples, the observable data then consist of ( δ , W ) , where W = ( W 1 , . . . , W r ) ,         W i { X h i n h i , h i = 1 , . . . , k } , with r < N being a pre-fixed integer and δ = ( δ 1 ( h ) , . . . , δ r ( h ) ) associated to ( h 1 , . . . , h r ) is defined by
δ i ( h ) = 1 ,         i f         h = h i 0 ,         o t h e r w i s e .
Letting M r ( h ) = i = 1 r δ i ( h ) denote the number of X h -failures in W and r = h = 1 k M r ( h ) , then the joint density function of ( δ , w ) is given by
f ( δ , w ) = h = 1 k c r ( F ¯ h ( w r ) ) n h M r ( h ) . i = 1 r h = 1 k ( f h ( w i ) ) δ i ( h ) .
where F ¯ h = 1 F h , is the survival functions of h t h population and c r = n h ! ( n h M r ( h ) ) ! .
The joint density function of ( δ , δ s , w , w s ) , r < s N is given by
f ( δ , δ s , w , w s ) = Q s 1 h = 1 k c 1 h { F ¯ h ( w r )                                                                                                                                               F ¯ h ( w s ) } M r s ( h ) ( F ¯ h ( w s ) ) n ¯ h s ( f h ( w s ) ) δ s ( h )
  × i = 1 r h = 1 k ( f h ( w i ) ) δ i ( h )
where,
w r < w s w N ,     M r s h = M s 1 h M r h , n ¯ h s = n h M s h , c 1 h = n h ! M r s h ! n ¯ h s ! ,   Q s = δ r + 1 = 0 1 . . . δ s = 0 1 with Q s = { δ h = δ r + 1 , . . . , δ s ,   f o r     1 h k } .
The conditional density function of W s given ( δ , W ) , is given by
f w s δ , w ) = Q s h = 1 k c h ( f h ( w s ) ) δ s h { F ¯ h ( w r )                                                                                                                                                         F ¯ h ( w s ) } M r s ( h ) ( F ¯ h ( w s ) ) n ¯ h s ( F ¯ h ( w r ) ) n ¯ h r
= Q s h = 1 k c h ( f h ( w s ) ) δ s ( h ) l h = 0 M r s ( h ) a l h ( F ¯ h ( w s ) ) n ¯ h s + l h ( F ¯ h ( w r ) ) n ¯ h ( s 1 ) + l h
where,
c h = n ¯ h r ! M r s ( h ) ! n ¯ h s ! ,   a l h = ( 1 ) l h M r s ( h )           l h .
When the k   populations are exponential with pdf and cdf respectively,
f h = θ h e x p θ h x ,         F h = 1 e x p θ h x ,         x > 0 , θ h > 0 ;     1 h k .
Then the likelihood function in (3) becomes
f ( Θ , δ , w ) = h = 1 k c r { e x p ( θ h w r ) } n ¯ h r i = 1 r h = 1 k { θ h e x p ( θ h w i ) } δ i ( h )
= h = 1 k c r θ h M r ( h ) e x p θ h u h
where, Θ = ( θ 1 , . . . , θ k ) and u h = i = 1 r w i δ i ( h ) + w r n ¯ h r .
Substituting (6) into (5), we obtain the conditional density function of W s , given ( δ , W )
f w s δ , w ) = Q s h = 1 k θ h δ s ( h ) l h = 0 M r s ( h ) C h e x p { θ h D h ( w s w r ) }
where, C h = c h a l h , D h = n ¯ h ( s 1 ) + l h and n ¯ h s + δ s ( h ) = n ¯ h ( s 1 ) .
Some special cases of the conditional density function are described as follows:
Case 1:
Suppose that k 1 is the number of samples satisfy   M r ( h ) = n h and only one sample satisfies M r ( q ) < n q for h = 1 , . . . , k , h q equivalently W r > m a x { X h n h , h = 1 , . . . , k ; h q } but W r { X q n q } ; w N = x q n q .
In Case 1, the conditional density function of W s given ( δ , W ) is given by
f 1 w s δ , w ) = n ¯ q r ! ( n ¯ q r s + r ) ! l q = 0 s r 1 a l q θ q e x p { θ q D q ( w s w r ) }
where, a l q = ( 1 ) l q s r 1             l q , D q = n ¯ q r s + r + l q + 1 , w r < w s x q n q .
Case 2:
Suppose that k R is the number of samples satisfy M r ( h i ) = n h i for h i = 1 , . . . , k ; and R < k is the number of samples satisfy M r ( h j ) < n h j for h j = 1 , . . . , k ; h i h j , equivalently W r > m a x { X h i n h i , h i = 1 , . . . , k } but W r { X h j n h j , h j = 1 , . . . , k ; h i h j } .
Under case 2, let’s just consider the R samples, where q = 1 , . . . , R , then the conditional density function of W s given ( δ , W ) is given by
f 2 ( w s | δ , w ) = Q ' s q = 1 R θ q δ s ( q ) l q = 0 M r s ( q ) C q e x p { θ q D q ( w s w r ) }       (1.10)
where, Q ' s = { δ ( q ) = ( δ r + 1 , . . . , δ s )     f o r     1 q R } , C q = c q a l q , D q = n ¯ q ( s 1 ) + l q and n ¯ q s + δ s ( q ) = n ¯ q ( s 1 ) .
The remainder of this article is organized as follows. Section 2 presents the generalized Bayesian and Bayesian prediction points and intervals using squared error, Linex, and general entropy loss functions in the point predictor. A numerical study of the results from Section 2 is presented in Section 3. Finally, we conclude the paper in Section 4.

2. Generalized Bayes prediction

In this section, we introduce the concept of generalized Bayesian prediction, which is an investigation of Bayesian prediction under the influence of a learning rate parameter η > 0 . To apply the concept of generalized Bayesian prediction to a prediction study, we give a brief description of generalized Bayesian prediction based on a learning rate parameter η > 0 . A scheme for predicting a sample based on joint censoring of type II samples from k exponential distributions is presented. The main goal is to obtain the point predictors and prediction intervals given at the end of this section.

2.1. Generalized Bayes 

The parameters Θ are assumed to be unknown, we may consider the conjugate prior distributions of Θ as independent gamma prior distributions, i.e. θ h G a m ( a h , b h ) . Hence, the joint prior distribution of Θ is given by
π ( Θ ) = h = 1 k π h ( θ h ) ,
where
π h ( θ h ) = b h a h Γ ( a h )     θ h a h 1 e x p { b h θ h } ,
and Γ ( ) denotes the complete gamma function.
Combining (7) and (10) after raising (7) to the power η , the posterior joint density function of Θ is then
π * Θ   d a t a ) = h = 1 k ( u h η + b h ) η M r ( h ) + a h θ h η M r ( h ) + a h 1 Γ ( η M r ( h ) + a h ) e x p { θ h ( u h η + b h ) } ,
= h = 1 k ξ h μ h θ h μ h 1 Γ ( μ h ) e x p { θ h ξ h } ,
where, ξ h = u h η + b h ,         μ h = η M r ( h ) + a h .
Since π h is a conjugate prior, where θ h G a m ( a h , b h ) then it follows that the posterior density function of ( θ h | d a t a ) is G a m ( η M r ( h ) + a h , u h η + b h ) .

2.2. One sample prediction 

A sample prediction scheme for the case of joint censoring of samples from two exponential distributions was studied in [17] and then three cases for the future failures were derived, where in the first case the future predicted failure surly belongs to X 1 n 1 failures if M r 1 < n 1 , M r 2 = n 2 , second case, the future predicted failure surly belongs to X 2 n 2 failures if M r 2 < n 2 , M r 1 = n 1 , third case, it is unknown to which sample the future predicted failure belongs. Here, we generalize the results reported in [17] and examine two special cases in addition to the general case.
In the general case, the size of any sample is greater than the number of observed failures; that is, M r h < n h w r < x h n h for h = 1 , . . . , k . The first special case arises when all future values (predictors) belong to only one sample and the observations of the remaining k 1 samples are less than w r . The second special case arises when all future values (predictors) belong to some samples and all observations of the other samples are less than w r . The forms of all functions related to the second special case are similar to those related to the general case; therefore, we will introduce only the general case and the first special case.
For the general case, to predict w r for r < s N based on the observed data ( δ , w ) , we use the conditional density function ( 9 ) . Let us define the following integral
I h δ s ( h ) = ξ h μ h Γ ( μ h ) 0 θ h μ h 1 + δ s ( h ) e x p θ h ξ h + D h ( ω s ω r ) d θ h
= μ h ξ h 1 + D h ω s ω r ξ h μ h + 1 ,     f o r   δ s h = 1 1 + D h ω s ω r ξ h μ h                     ,       f o r   δ s h = 0
Using (8), (12), and (13) the Bayesian predictive density function of w s , given ( δ , w ) , is given by
f B ω s   δ , w = Θ f w s δ , w π * Θ δ , w d Θ = Q s h = 1 k l h = 0 M r s C h ξ h μ h Γ μ h 0 θ h μ h 1 + δ s h e x p θ h ξ h + D h ω s ω r d θ h = Q s h = 1 k C h υ = 1 k I υ 1   q = 1 , q υ k I q 0 ,
Where
h = 1 k 0 f ( θ h ) d θ h = 0 . . . 0 f ( θ 1 ) . . . f ( θ k ) d θ 1 . . . d θ k .
Under case 1, the Bayesian predictive density function of w s , given ( δ , w ) , is given by
f 1 B w s δ , w = n ¯ q r ! μ q n ¯ q r s + r ! ξ q l q = 0 s r 1 a l q 1 + D q ( w s w r ) ξ q ( μ q + 1 ) ,
where, ξ q = u q η + b q , μ q = η M r ( q ) + a q , w r < w s x q n q .

2.3. Bayesian point predictors 

For the point predictor, we considered three types of loss functions:
(i). The squared error loss function (SE), which is classified as a symmetric function, is given by
L S E ( φ * , φ ) ( φ * φ ) 2 ,
where φ * is an estimate of φ .
(ii). The Linex loss function, which is asymmetric is given by
L L φ * , φ e τ φ * φ τ φ * φ 1 ,   τ 0 .
(iii). The generalization of the entropy (GE) loss function is
L G E φ * , φ φ * φ c c ln φ * φ 1 ,     c 0 .
It is worth noting that the Bayes estimates under the GE loss function coincide with those under the SE loss function when c = 1 . However, when c = 1 , 2 ,     the Bayes estimates under GE become those under the weighted squared error loss function and the precautionary loss function, respectively.
Now, the Bayesian point predictors W s ,     r < s N , under different loss functions (SE, Linex, GE) can be obtained using the predictive density function (14), which are denoted respectively by W S P , W L P , W E P and given as follows:
Preprints 76955 i001
Under case 1, W S P , W L P , W E P are respectively given by
W S P = n ¯ q r ! μ q ( n ¯ q r s + r ) ! ξ q l q = 0 s r 1 a l q t w s ( 1 + D q ( ω s ω r ) ξ q ) ( μ q + 1 ) d ω s
W L P = 1 τ l n n ¯ q r ! μ q ( n ¯ q r s + r ) ! ξ q l q = 0 s r 1 a l q t e τ w s ( 1 + D q ( ω s ω r ) ξ q ) ( μ q + 1 ) d ω s
W E P = n ¯ q r ! μ q ( n ¯ q r s + r ) ! ξ q l q = 0 s r 1 a l q t w s c ( 1 + D q ( ω s ω r ) ξ q ) ( μ q + 1 ) d ω s 1 / c
Numerical integration is required to obtain the predictors W S P , W L P , W E P .

2.4. Prediction Interval 

The predictive survival function of W s is given by
F ¯ B t = P W s > t   δ , w = t f B w s δ , w d w s     = Q s 1 h = 1 k l h = 0 M r s C h v = 1 k t I υ 1   q = 1 , q υ k I q 0 d w s
Numerical integration is required to obtain the predictive distribution function in Equation (22). In Case 1, the predictive survival function of W s is given by
F ¯ 1 B t = P W s > t δ , w = t f 1 B w s δ , w d w s
= n ¯ q r ! ( n ¯ q r s + r ) ! l q = 0 s r 1 a l q D q 1 + D q ( t w r ) ξ q μ q .
The Bayesian predictive bounds of a two-sided equi-tailed 100 ( 1 γ ) % interval for W s ,     r < s N , can be obtained by solving the following two equations,
F ¯ ( L | δ , w ) = 1 γ 2 ,             F ¯ ( U | δ , w ) = γ 2

3. Numerical study

In this section, the results of the Monte Carlo simulation study are conducted to evaluate the performance of the prediction study derived in the previous section, and an example is presented to illustrate the prediction methods discussed here.

3.1. Simulation study 

We considered three samples from three populations with ( n 1 , n 2 , n 3 , r ) for choices ( 10 , 10 , 10 , 25 ) and ( 15 , 15 , 15 , 40 ) . In case 1, we choose the exponential parameters ( θ 1 , θ 2 , θ 3 ) as (2, 1, 0.1) based on the hyperparameters represented by Δ = ( a 1 , b 1 , a 2 , b 2 , a 3 , b 3 ) , where Δ = Δ 1 = ( 2,1 , 2,2 , 1,10 ) .
In the general case we choose the exponential parameters ( θ 1 ,   θ 2 , θ 3 ) as (2, 2.5, 3 ) based on the hyperparameters Δ = Δ 2 = ( 2 , 1 , 5 , 2 , 3 , 1 ) .
For the generalized Bayesian study, three values are chosen for the learning rate parameter η = 1 , 2 , 5 and 10,000 repetitions are used for the Monte Carlo simulations. For ( n 1 , n 2 , n 3 , r ) = ( 10 , 10 , 10 , 25 ) , under case 1 we use (19), (20) and (21), to calculate the mean squared prediction errors (MSPEs) of the point predictors ( W S P , W L P , W E P ) for s = 26 , . . . , 30 , where τ = 0.1,0.5 ; c = 0.1,0.5 and the results are presented in Table 1.
Preprints 76955 i002
The results of the MSPEs in the general case are calculated using (16), (17), and (18) and shown in Table 2.
Preprints 76955 i003
For ( n 1 , n 2 , n 3 , r ) = ( 10 , 10 , 10 , 25 ) and ( n 1 , n 2 , n 3 , r ) = ( 15,15,15,40 ) the results of the prediction bounds of W s ,     s = 26 , . . . , 30 and s = 41 , . . . , 45 , respectively are calculated using (23), (24) in Case 1 then are presented in Table 3.
Preprints 76955 i004
Table 4 presents the prediction bounds using (22), (24) to show the results of the general case.
Preprints 76955 i005

3.2. Illustrative example 

To illustrate the usefulness of the results developed in the previous sections, we consider three samples of size n 1 = n 2 = n 3 = 10 from Nelson's data (groups 1, 4, and 5) corresponding to the breakdown of an insulating fluid subjected to a high-stress load (see [20] p. 462), These breakdown times, referred to here as samples X i ,     i = 1,2 , 3 , are jointly type-II censored data in the form of ( w , h i ) obtained from these three samples with r = 24 and are shown in Table 5.
Preprints 76955 i006
Using (16), (17), (18), (22), and (24), the MSPEs of the point predictors and prediction intervals of ws, s = 25,..., 30 are calculated and presented in Table 6 using 𝜂 = 1, 2, 5 and Δ = Δ3 = (1,2.6,1,2,1,3); 𝜏 = 0.1, 0.5 and c = 0.1, 0.5.
Preprints 76955 i007

4. Conclusion

In this study, we considered a joint type- II censoring scheme in which the lifetimes of the three populations have exponential distributions. We determined the MSPEs of the point predictors and prediction intervals using different values for the learning rate parameter η and different values for the parameters of the losses in both the simulation study and the illustrative example. From all tables in this prediction study, it can be seen that the results improve with increasing c, τ and η . In the simulation study, a comparison of the results in Tables 1 and 2 shows that the results in Table 1 are better, and the length of the prediction intervals in Table 4 is smaller than those in Table 3 because the observed values used in Table 3 are larger than those used in Table 4. The results of the illustrative example improve with larger values of c, τ and η . However, in both studies, the lengths of the prediction intervals increased as s increased. It may be interesting to examine this work using a different type of censoring.

Acknowledgement

The authors extend their sincere appreciation to Princess Nourah bint Abdulrahman University Researchers Supporting Project Number (PNURSP2023R226), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

References

  1. Bissiri, P. G., Holmes, C. C., and Walker, S. G. General framework for updating belief distributions. Journal of the Royal Statistical Society. Series B, Statistical Methodology, 2016, 78(5), 1103-1130. [CrossRef]
  2. Miller, J. W. and Dunson, D. B. Robust Bayesian inference via coarsening. Journal of the American Statistical Association, 2019, 114(527): 1113-1125. [CrossRef]
  3. Grünwald, P. (2012). The safe Bayesian: learning the learning rate via the mixability gap. In Algorithmic Learning Theory, 2012, volume 7568 of Lecture Notes in Computer Science, 169-183. Springer, Heidelberg. MR3042889.
  4. Grünwald, P. and van Ommen, T. Inconsistency of Bayesian inference for misspecified linear models, and a proposal for repairing it. Bayesian Analysis, 2017, 12(4): 1069-1103. [CrossRef]
  5. Grünwald, P. Safe probability. Journal of Statistical Planning and Inference, 2018, 47-63. MR3760837. [CrossRef]
  6. De Heide, R., Kirichenko, A., Grünwald, P., and Mehta, N. Safe-Bayesian generalized linear regression. In International Conference on Artificial Intelligence and Statistics, 2020, 2623-2633. PMLR. 106, 113.
  7. Holmes, C. C. and Walker, S. G. Assigning a value to a power likelihood in a general Bayesian model. Biometrika, 2017, 497-503. [CrossRef]
  8. Lyddon, S. P., Holmes, C. C., and Walker, S. G. General Bayesian updating and the loss-likelihood bootstrap. Biometrika, 2019, 465-478. [CrossRef]
  9. Martin, R. Invited comment on the article by van der Pas, Szabó, and van der Vaart. Bayesian Analysis, 2017, 1254-1258.
  10. Martin, R. and Ning, B. Empirical priors and coverage of posterior credible sets in a sparse normal mean model. Sankhy a ¯ . Series A, 2020, 477-498. Special issue in memory of Jayanta K. Ghosh. [CrossRef]
  11. Abdel-Aty, Y., Kayid, M., and Alomani, G. Generalized Bayes estimation based on a joint type-II censored sample from k-exponential populations. Mathematics, 2023, 1-11. [CrossRef]
  12. Balakrishnana, N. and Rasouli, A. Exact likelihood inference for two exponential populations under joint Type-II censoring, Computational Statistics & Data Analysis, 2008. 2725-2738. [CrossRef]
  13. Parsi, S. Bairamov, I. Expected values of the number of failures for two populations under joint Type-II progressive censoring. Computational Statistics and Data Analysis, 2009, 3560-3570. [CrossRef]
  14. Rasouli, A., Balakrishnan, N. Exact likelihood inference for two exponential populations under joint progressive type-II censoring. Communications in Statistics - Theory and Methods, 2010, 2172-2191. [CrossRef]
  15. Su, F. Exact likelihood inference for multiple exponential populations under joint censoring. Open Access Dissertations and Theses, 2013, Ph.D Thesis, McMaster University.
  16. Abdel-Aty, Y. Exact likelihood inference for two populations from two-parameter exponential distributions under joint Type-II censoring, Communications in Statistics - Theory and Methods, 2017, 9026-9041. [CrossRef]
  17. Shafay, A.R., Balakrishnan, N. Y. Abdel-Aty, Y. Bayesian inference based on a jointly type-II censored sample from two exponential populations, Journal of Statistical Computation and Simulation, 2014, 2427-2440. [CrossRef]
  18. Asgharzadeh, A., Valiollahi, R. and Kundu, D. Prediction for future failures in Weibull distribution under hybrid censoring. Journal of Statistical Computation and Simulation, 2013, 824-838. [CrossRef]
  19. Abdel-Aty, Y., Franz, J., Mahmoud, M.A.W. Bayesian prediction based on generalized order statistics using multiply type-II censoring. Statistics, 2007, 495-504. [CrossRef]
  20. Nelson, W. Applied Life Data Analysis; Wiley: New York, NY, USA, 1982.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated