Preprint
Article

This version is not peer-reviewed.

Median Based Unit Weibull Distribution (MBUW): Do the Higher Order Probability Weighted Moments (PWM) Add More Information over the Lower Order PWM in Parameter Estimation

Submitted:

11 December 2025

Posted:

15 December 2025

You are already at the latest version

Abstract
In the present paper, Probability weighted moments (PWMs) method for parameter estimation of the median based unit weibull (MBUW) distribution is discussed. The most widely used first order PWMs is compared with the higher order PWMs for parameter estimation of (MBUW) distribution. Asymptotic distribution of this PWM estimator is derived. This comparison is illustrated using real data analysis.
Keywords: 
;  ;  ;  

1. Introduction

Imam Attia was the first to introduce Median Based Unit Weibull distribution (MBUW) [1]. Given a random variable y that is distributed as Median Based Unit Weibull distribution (MBUW), the PDF, CDF, and quantile functions are as follows:
f ( y ) = 6 α β 1 y 1 α β y 2 α β 1 , 0 < y < 1 , α > 0 , β > 0
F ( y ) = 3 y 2 α β 2 y 3 α β , 0 < y < 1 , α > 0 , β > 0
y = F 1 ( y ) = . 5 cos cos 1 ( 1 2 u ) 3 3 sin cos 1 ( 1 2 u ) 3 + . 5 α β
Various methods are employed to estimate the parameters of a distribution. Maximum Likelihood Estimation (MLE) is popular because it leads to efficient, asymptotically minimum variance estimators, although these estimators may not necessarily be unbiased. The method of moments is straightforward to apply and can serve as starting values for the numerical procedures associated with MLE.
[2] advocated for Probability Weighted Moments (PWMs) and initiated their use in hydrology for small sample sizes, as MLE does not always perform well with limited data. PWMs are a leading alternative to method of moments (MOMs) and MLE for fitting statistical distributions to data, especially those expressed in inverse form. For instance, if y is a random variable and F represents the cumulative distribution function (CDF) for y, the value of y can be expressed as a function of F; y = y ( F ) . These distributions include the Gumbel, Weibull, and logistic distributions, as well as lesser-known types such as Tukey’s symmetric lambda, Thomas’ Wakeby, and Mielke’s kappa.
[3] advocated for PWMs, demonstrating their superior performance over other estimators through Monte Carlo simulations. For most of these distributions, relatively simple expressions for the parameters can be derived, as noted by [4], including several for which parameter estimates are not easily obtained using MLE or conventional moments These distributions relate to various fields, including hydrology, resource management, and the forecasting of weather parameters such as temperature, precipitation, wind velocity, floods, droughts, and rainfall.
Many researchers have utilized PWMs, including [5] for estimating parameters of the type II extreme value distribution, and [6] who applied the generalized form of PWMs to estimate the two-parameter Weibull distribution. [7] introduced a new class of generalized PWMs for estimating parameters of the Pareto distribution, while [8] implemented partial PWMs for censored data from the generalized extreme value distribution (GEV). Many authors like [9,10,11,12,13,14,15,16] have applied PWMs to various distributions, mainly focusing on the GEV.
PWMs are more robust than conventional moments in the presence of outliers in the data and yield more efficient estimators from small samples regarding the underlying probability distribution. They are computed as expectations of certain functions of a random variable. A significant advantage over conventional moments is that PWMs are, by definition, linear functions of the ordered data, resulting in reduced sensitivity to sampling variability. Various methods are employed to estimate the parameters of a distribution.
This paper is organized as follows: Section 2 introduces the methodology and the definition of Probability Weighted Moments (PWMs). It explicates the classic PWMs method for parameter estimation. In this section the author derives the PWMs for the Median Biased Unit Weibull (MBUW) and clarifies how to utilize these moments for parameter estimation. The author also illustrates the derivation of the asymptotic variance for different estimators. Section 3 demonstrates Monte Carlo simulation study to validate the premise of the paper. Section 4 explains the results of the simulation study. Section 5 explores the application of this method in real data analysis. Section 6 comprehends the conclusions. Finally, Section 7 suggests the future work.

2. Materials and Methods:

2.1. Definitions of Probability Weighted Moments (PWMs)

The population PWM, as shown in equations (4-5), is defined as M p , r , s , where p , r , and s are real numbers.
M p , r , s = E ( y p [ F ( y ) ] r [ 1 F ( y ) ] s ) = y p [ F ( y ) ] r [ 1 F ( y ) ] s f ( y ) d y
M p , r , s = E ( y p [ F ( y ) ] r [ 1 F ( y ) ] s ) = 0 1 [ y ] p [ F ( y ) ] r [ 1 F ( y ] s d F
In practice, the value of p is chosen to be 1, and so M 1 , r , s are used to estimate the parameter. Here, M 1 , 0 , 0 represents the mean. If this mean exists, then M 1 , r , s exist for any real positive values of r and s. These values are often restricted to small positive integers. Choosing p = 1 has the dual advantage of not unduly overweighting sample values while also leading to a class of linear L-moments that exhibit asymptotic normality, [17,18]. While only small positive integers are needed to estimate the parameters of distributions, using real numbers, regardless of their size, can offer significant advantages. This concept was explored by [19], who extended PWMs into generalized PWMs to include those with p 1 . PWMs is a generalization of the classic method of moments when r = s = 0 , M 1 , 0 , 0 , are the non-central moments of order p. [20] modified the method to accommodate the models without an analytic CDF and quantile function. [2] and [18] strongly advised employing M 1 , r , s due to the significantly clearer relationships that exist between parameters and moments. This approach simplifies analysis and enhances our understanding of these critical relationships. The empirical estimate of M 1 , r , s is typically less sensitive to outliers and exhibits good properties, especially when the sample size is small. For convenience, several authors chose to use p = 1 and non-negative integer values for r and s. This approach is referred to as the classic PWM method. When p = 1 , r and s are non-negative, the following equations (6-7) are defined:
α s = M 1 , 0 , s = E y ( 1 F ( y ) ) s , s = 0 , 1 , . . . , & r = 0
β r = M 1 , r , 0 = E y [ F ( y ) ] r , r = 0 , 1 , . . . & s = 0
Both α s and β r are related by the following equation (8):
α s = i = 0 s ( 1 ) i s i β i a n d β r = i = 0 r ( 1 ) i r i α i
For non-negative integers values of r and s that are as small as possible, both α s and β r are equivalent. [21] defined the sample unbiased estimators for PWMs α s and β r as the followings in Equation (9):
α ^ s = 1 n i = 1 n s n i s n 1 s y i : n and β ^ r = 1 n i = 1 + r n i 1 r n 1 r y i : n
The biased sample estimator for the PWM is defined as
M ^ 1 , r , s = 1 n i = 1 n y i : n ( p i : n ) r ( 1 p i : n ) s , so it can be defined in Equations (10-11) for:
α ^ s = M 1 , 0 , s = 1 n i = 1 n y i : n ( p i : n ) 0 ( 1 p i : n ) s
β ^ r = M 1 , r , 0 = 1 n i = 1 n y i : n ( p i : n ) r ( 1 p i : n ) 0
Where p i : n = i b n , 0 b 1 or p i : n = i b n + 1 2 b , 0.5 b 0.5
[22] empirically concluded that moderated biased estimates of the PWMs could produce more accurate estimates of upper quantiles. In this paper, M 1 , 0 , 1 and M 1 , 1 , 0 are used as a system of equations to estimate the parameters of the Median Based Unit Weibull (MBUW) distribution. Higher moments M 1 , 0 , 2 and M 1 , 2 , 0 are also used to estimate the parameters and compare the results with lower moments M 1 , 0 , 1 and M 1 , 1 , 0 . Parameter estimation using PWMs is carried out by equating the analytic expression of the population PWMs by the corresponding sample estimates of PWMs and solving the resulting systems of equations in terms of the parameters.

2.2. Calculating PWMs for MBUW

2.2.1. Calculating M_(1,0,1): Equations (1220)

Substitute in equation 4 to get: M p , r , s = M 1 , 0 , 1 = E ( y [ 1 F ( y ) ] s ) = 0 1 y [ 1 F ( y ) ] s f ( y ) d y
M 1 , 0 , 1 = 0 1 y 1 3 y 2 α β 2 y 3 α β s 6 α β 1 y 1 α β y 2 α β 1 d y
= 0 1 1 3 y 2 α β 2 y 3 α β s 6 α β 1 y 1 α β y 2 α β d y
= 0 1 6 α β y 2 α β 1 3 y 2 α β 2 y 3 α β s d y 0 1 6 α β y 3 α β 1 3 y 2 α β 2 y 3 α β s d y
= 0 1 A ( y ) d y 0 1 B ( y ) d y
Using Binomial Expansion in Equation (13):
1 3 y 2 α β 2 y 3 α β s = i = 0 s ( 1 ) i s i 3 y 2 α β 2 y 3 α β i = i = 0 s ( 1 ) i s i 3 y 2 α β i 1 2 3 y 1 α β i = i = 0 s ( 1 ) i s i 3 y 2 α β i m = 0 i ( 1 ) m i m 2 3 y 1 α β m = i = 0 s ( 1 ) i s i 3 i y 2 i α β m = 0 i ( 1 ) m i m 2 3 m y m α β .
Now exchange integration with summation in Equations (13) & (14):
0 1 A ( y ) d y = 0 1 6 α β y 2 α β 1 3 y 2 α β 2 y 3 α β s d y = 0 1 6 α β y 2 α β i = 0 s ( 1 ) i s i 3 i y 2 i α β m = 0 i ( 1 ) m i m 2 3 m y m α β d y = 6 α β i = 0 s ( 1 ) i s i 3 i m = 0 i ( 1 ) m i m 2 3 m 0 1 y 2 α β + 2 i α β + m α β d y .
Where the integral is:
0 1 y 2 α β y 2 i α β y m α β d y = α β 2 + 2 i + m + α β
so:
0 1 A ( y ) d y = 0 1 6 α β y 2 α β 1 3 y 2 α β 2 y 3 α β s d y =
6 α β i = 0 s ( 1 ) i s i 3 i m = 0 i ( 1 ) m i m 2 3 m α β 2 + 2 i + m + α β
The same steps follow for:
0 1 B ( y ) d y = 0 1 6 α β y 3 α β 1 3 y 2 α β 2 y 3 α β s d y = 6 α β i = 0 s ( 1 ) i s i 3 i m = 0 i ( 1 ) m i m 2 3 m α β 3 + 2 i + m + α β
Substitute s=1 in Equations (17) and (18):
0 1 A ( y ) d y = 6 α β ( 1 ) 0 1 0 3 0 ( 1 ) 0 0 0 2 3 0 α β 2 + 2 ( 0 ) + 0 + α β + ( 1 ) 1 0 1 2 3 1 α β 2 + 2 ( 0 ) + 1 + α β + ( 1 ) 1 1 1 3 1 ( 1 ) 0 1 0 2 3 0 α β 2 + 2 ( 1 ) + 0 + α β + ( 1 ) 1 1 1 2 3 1 α β 2 + 2 ( 1 ) + 1 + α β 1 1 = 6 α β α β 2 + α β 3 α β 4 + α β + 2 α β 5 + α β = 6 1 2 + α β 3 4 + α β + 2 5 + α β
For
0 1 B ( y ) d y = 6 α β { ( 1 ) 0 1 0 3 0 [ ( 1 ) 0 0 0 2 3 0 α β 3 + 2 ( 0 ) + 0 + α β + ( 1 ) 1 0 1 2 3 1 α β 3 + 2 ( 0 ) + 1 + α β ] + ( 1 ) 1 1 1 3 1 [ ( 1 ) 0 1 0 2 3 0 α β 3 + 2 ( 1 ) + 0 + α β + ( 1 ) 1 1 1 2 3 1 α β 3 + 2 ( 1 ) + 1 + α β ] } = 6 α β α β 3 + α β 3 α β 5 + α β + 2 α β 6 + α β = 6 1 3 + α β 3 5 + β + 2 6 + α β
Substitute Equations (19) and (20) into Equation (14):
M 1 , 0 , 1 = 360 + 108 α β 1044 α β + 580 α 2 β + 155 α 3 β + 20 α 4 β + α 5 β + 720

2.2.2. Calculating M 1 , 1 , 0 Equations (21-26)

M 1 , 1 , 0 = E ( y p [ F ( y ) ] r ) = 0 1 y p [ F ( y ) ] r f ( y ) d y
0 1 y 3 y 2 α β 2 y 3 α β r 6 α β 1 y 1 α β y 2 α β 1 d y
= 0 1 6 α β y 2 α β 3 y 2 α β 2 y 3 α β r d y 0 1 6 α β y 3 α β 3 y 2 α β 2 y 3 α β r d y = 0 1 C ( y ) d y 0 1 D ( y ) d y
0 1 C ( y ) d y = 0 1 6 α β y 2 α β 3 y 2 α β 2 y 3 α β r d y = 6 α β 0 1 y 2 α β 3 y 2 α β r 1 2 3 y 1 α β r d y
Using binomial expansion: 1 2 3 y 1 α β r = i = 0 r ( 1 ) i r i 2 3 y 1 α β i
So
0 1 C ( y ) d y = 6 α β 0 1 y 2 α β 3 y 2 α β r i = 0 r ( 1 ) i r i 2 3 y 1 α β i d y
Exchange the integral and the sum 6 α β i = 0 r 3 r ( 1 ) i r i 2 3 i 0 1 y 2 α β y 2 r α β y i α β d y =
0 1 C ( y ) d y = 6 α β i = 0 r 3 r ( 1 ) i r i 2 3 i α β 2 + 2 r + i + α β
The same steps are followed and give:
0 1 D ( y ) d y = 6 α β i = 0 r 3 r ( 1 ) i r i 2 3 i α β 3 + 2 r + i + α β
Substitute Equation (24) and Equation (25) into Equation (22) and let r=1
M 1 , 1 , 0 = 6 α β i = 0 r 3 r ( 1 ) i r i 2 3 i α β 2 + 2 r + i + α β α β 3 + 2 r + i + α β = 6 α β 3 1 ( 1 ) 0 1 0 2 3 0 α β 2 + 2 ( 1 ) + 0 + α β α β 3 + 2 ( 1 ) + ( 0 ) + α β + 3 1 ( 1 ) 1 1 1 2 3 1 α β 2 + 2 ( 1 ) + ( 1 ) + α β α β 3 + 2 ( 1 ) + ( 1 ) + α β = 6 α β 3 α β 4 + α β 2 α β 5 + α β 3 α β 5 + α β + 2 α β 6 + α β = 18 4 + α β 30 5 + α β + 12 6 + α β M 1 , 1 , 0 = 6 ( 10 + α β ) ( 4 + α β ) ( 5 + α β ) ( 6 + α β ) = 60 + 6 α β 74 α β + 15 α 2 β + α 3 β + 120
To sum up, PWMs method for estimating the parameters using M 1 , 0 , 1 and M 1 , 1 , 0 :
Step 1: Calculate the population PWMs for the order p = 1
M 1 , 0 , 1 = 360 + 108 α β 1044 α β + 580 α 2 β + 155 α 3 β + 20 α 4 β + α 5 β + 720
M 1 , 1 , 0 = 60 + 6 α β 74 α β + 15 α 2 β + α 3 β + 120
Step 2: Calculate the estimated sample PWMs whether the unbiased or biased estimators and equate these estimators with the corresponding population PWMs.
α ^ s = 1 n i = 1 n 1 ( n i ) ( n 1 ) y i : n = M 1 , 0 , 1 , o r α ^ s = 1 n i = 1 n y i : n ( p i : n ) 0 ( 1 ( 1 p i : n ) s = M 1 , 0 , 1
β ^ r = 1 n i = 1 + 1 n ( i 1 ) ( n 1 ) y i : n = M 1 , 1 , 0 , o r β ^ r = 1 n i = 1 n y i : n ( p i : n ) 1 ( 1 p i : n ) 0 = M 1 , 1 , 0
Step 3: The above equations construct a system of equations to be solved numerically. In this paper, the author used the Levenberg-Marquardt (LM) algorithm. The objective functions to be minimized are Equations (2728). Differentiate the previous Equations (2728) with respect to alpha and beta.
The Jacobian matrix is
α M _ 101 β M _ 101 α M _ 110 β M _ 110
y f ( θ ( a ) ) = α ^ s 360 + 108 α β 1044 α β + 580 α 2 β + 155 α 3 β + 20 α 4 β + α 5 β + 720 β ^ r 60 + 6 α β 74 α β + 15 α 2 β + α 3 β + 120 , & θ ( a ) = α ( a ) β ( a )
Apply LM algorithm: θ ( a + 1 ) = θ ( a ) + J ( θ ( a ) ) J ( θ ( a ) ) + λ ( a ) I ( a ) 1 J ( θ ( a ) ) y f ( θ ( a ) )
Where the parameters used in the first iteration are the initial guess, then they are updated according to the sum of squares of errors. The LM algorithm is an iterative algorithm. J ( θ ( a ) ) is the Jacobian function which is the first derivative of the objective function evaluated at the initial guess θ ( a ) and λ ( a ) is a damping factor that adjusts the step size in each iteration direction. The starting value of this factor is usually 0.001 and according to the sum square of errors (SSE) in each iteration this damping factor is adjusted:
S S E ( a + 1 ) S S E ( a ) s o u p d a t e : λ u p d a t e d = 10 * λ o l d
S S E ( a + 1 ) < S S E ( a ) s o u p d a t e : λ u p d a t e d = 1 10 * λ o l d , f ( θ ( a ) ) is the objective function (population PWMs, M 1 , 0 , 1 & M 1 , 1 , 0 ) evaluated at the initial guess, and y is the sample estimates of population PWMs, the M 1 , 0 , 1 and M 1 , 1 , 0 .
Steps of the LM algorithm:
  • Start with the initial guess of parameters (alpha and beta).
  • Substitute these values in the objective function and the Jacobian.
  • Choose the damping factor, say lambda=0.001
  • Substitute in the equation (LM equation) to get the new parameters.
  • Calculate the SSE at these parameters and compare this SSE value with the previous one when using the initial parameters to adjust for the damping factor.
  • Update the damping factor accordingly as previously explained.
  • Start new iteration with the new parameters and the new updated damping factor, i.e., apply the previous steps many times till convergence is achieved or a pre-specified number of iterations is accomplished.
The value of this quantity: [ J ( θ ( a ) ) J ( θ ( a ) ) + λ ( a ) I ( a ) ] 1 can be considered a good approximation to the variance-covariance matrix of the estimated parameters. Standard errors for the estimated parameters are the square root of the diagonal of the elements in this matrix.

2.2.3. Calculating M 1 , 0 , 2 Equation (29)

Substitute s=2 in Equation (13) and solve to get Equation (29):
6 α β { ( 1 ) 0 2 0 3 0 [ ( 1 ) 0 0 0 2 3 0 α β 2 + 2 ( 0 ) + 0 + α β + ( 1 ) 1 0 1 2 3 1 α β 2 + 2 ( 0 ) + 1 + α β + ( 1 ) 2 0 2 2 3 2 α β 2 + 2 ( 0 ) + 2 + α β ] + ( 1 ) 1 2 1 3 1 [ ( 1 ) 0 1 0 2 3 0 α β 2 + 2 ( 1 ) + 0 + α β + ( 1 ) 1 1 1 2 3 1 α β 2 + 2 ( 1 ) + 1 + α β + ( 1 ) 2 1 2 2 3 2 α β 2 + 2 ( 1 ) + 2 + α β ] + ( 1 ) 2 2 2 3 2 [ ( 1 ) 0 2 0 2 3 0 α β 2 + 2 ( 2 ) + 0 + α β + ( 1 ) 1 2 1 2 3 1 α β 2 + 2 ( 2 ) + 1 + α β + ( 1 ) 2 2 2 2 3 2 α β 2 + 2 ( 2 ) + 2 + α β ] } 6 α β { ( 1 ) 0 2 0 3 0 [ ( 1 ) 0 0 0 2 3 0 α β 3 + 2 ( 0 ) + 0 + α β + ( 1 ) 1 0 1 2 3 1 α β 3 + 2 ( 0 ) + 1 + α β + ( 1 ) 2 0 2 2 3 2 α β 3 + 2 ( 0 ) + 2 + α β ] + ( 1 ) 1 2 1 3 1 [ ( 1 ) 0 1 0 2 3 0 α β 3 + 2 ( 1 ) + 0 + α β + ( 1 ) 1 1 1 2 3 1 α β 3 + 2 ( 1 ) + 1 + α β + ( 1 ) 2 1 2 2 3 2 α β 3 + 2 ( 1 ) + 2 + α β ] + ( 1 ) 2 2 2 3 2 [ ( 1 ) 0 2 0 2 3 0 α β 3 + 2 ( 2 ) + 0 + α β + ( 1 ) 1 2 1 2 3 1 α β 3 + 2 ( 2 ) + 1 + α β + ( 1 ) 2 2 2 2 3 2 α β 3 + 2 ( 2 ) + 2 + α β ] } M 1 , 0 , 2 = 6 α β α β 2 + α β 6 α β 4 + α β + 4 α β 5 + α β + 9 α β 6 + α β 12 α β 7 + α β + 4 α β 8 + α β 6 α β 1 α β 3 + α β 6 α β 5 + α β + 4 α β 6 + α β + 9 α β 7 + α β 12 α β 8 + α β + 4 α β 9 + α β M 1 , 0 , 2 = 6 1 2 + α β 6 4 + α β + 4 5 + α β + 9 6 + α β 12 7 + α β + 4 8 + α β 6 1 3 + α β 6 5 + α β + 4 6 + α β + 9 7 + α β 12 8 + α β + 4 9 + α β M 1 , 0 , 2 = K L where : K = 58320 α β + 6480 α 2 β + 120960 , L = 663696 α β + 509004 α 2 β + 214676 α 3 β + 54649 α 4 β + 8624 α 5 β + 826 α 6 β + 444 α 7 β + α 8 β + 362880 .

2.2.4. Calculating M _ 1 , 2 , 0 Equation (30)

Substitute r=2 in Equation (21) and solve to get Equation (30):
M 1 , 2 , 0 = 6 α β i = 0 2 3 r ( 1 ) i r i 2 3 i α β 2 + 2 r + i + α β α β 3 + 2 r + i + α β = 6 α β [ 3 2 ( 1 ) 0 2 0 2 3 0 α β 2 + 2 ( 2 ) + ( 0 ) + α β α β 3 + 2 ( 2 ) + ( 0 ) + α β + 3 2 ( 1 ) 1 2 1 2 3 1 α β 2 + 2 ( 2 ) + ( 1 ) + α β α β 3 + 2 ( 2 ) + ( 1 ) + α β + 3 2 ( 1 ) 2 2 2 2 3 2 α β 2 + 2 ( 2 ) + ( 2 ) + α β α β 3 + 2 ( 2 ) + ( 2 ) + α β ] M 1 , 2 , 0 = 6 α β 9 α β 6 + α β 12 α β 7 + α β + 4 α β 8 + α β 9 α β 7 + α β + 12 α β 8 + α β 4 α β 9 + α β M 1 , 2 , 0 = 6 9 6 + α β 21 7 + α β + 16 8 + α β 4 9 + α β M 1 , 2 , 0 = 150 α β + 6 α 2 β + 1008 1650 α β + 335 α 2 β + 30 α 3 β + α 4 β + 3024
In summary, PWMs method for estimating the parameters using M 1 , 0 , 2 and M 1 , 2 , 0 can be explained in the following steps: the first step is to calculate the population PWMs for the order p = 1 , using Equations (2930). The second step is to calculate the estimated sample PWMs, whether the unbiased or biased estimators, and equate these estimators with the corresponding population PWMs.
α ^ s = 2 = 1 n i = 1 n s ( n i ) ( n i 1 ) ( n 1 ) ( n 2 ) y i : n = M 1 , 0 , 2 or α ^ s = 1 n i = 1 n y i : n ( p i : n ) 0 ( 1 p i : n ) 2 = M 1 , 0 , 2 β ^ r = 2 = 1 n i = 1 + r n ( i 1 ) ( i 2 ) ( n 1 ) ( n 2 ) y i : n = M 1 , 2 , 0 or β ^ r = 1 n i = 1 n y i : n ( p i : n ) 2 ( 1 p i : n ) 0 = M 1 , 2 , 0 .
The last step is to construct a system of the above equations to solve it numerically. Using the Levenberg-Marquardt (LM) algorithm to minimize Equations (29) and (30). Differentiate the previous Equations (2930) with respect to alpha and beta.
The Jacobian matrix is
α M _ 102 β M _ 102 α M _ 120 β M _ 120
y f θ ( a ) = α ^ r K L β ^ s 150 α β + 6 α 2 β + 1008 1650 α β + 335 α 2 β + 30 α 3 β + α 4 β + 3024 , & θ ( a ) = α ( a ) β ( a )
Apply the LM algorithm: θ ( a + 1 ) = θ ( a ) + J ( θ ( a ) ) J ( θ ( a ) ) + λ ( a ) I ( a ) 1 J ( θ ( a ) ) y f ( θ ( a ) )
See Appendix A for the Jacobian matrices for both M 1 , 0 , 1 , M 1 , 1 , 0 and M 1 , 0 , 2 , M 1 , 2 , 0 .

2.3. Asymptotic Distribution of PWM Estimators

For optimal evaluations of estimators derived from various Probability Weighted Moments (PWMs) method, it is essential to leverage analytical expressions for their variances, as established by asymptotic theory. This approach not only streamlines the process but also eliminates the need for extensive and potentially flawed computer simulations. Unfortunately, in numerous cases, deriving straightforward expressions for the asymptotic variance of moments and parameter estimators proves to be a significant challenge. Moreover, as highlighted by Hosking (1986) and further explored by [4], the reliability of asymptotic variance expressions diminishes, particularly when working with small sample sizes.
In his 1986 study, Hosking offered critical insights into the asymptotic variances of generalized Pareto parameters estimated through classical PWMs. These asymptotic variances can be misleading, significantly misrepresenting true sampling variances when the sample size (n) is below 50. The asymptotic distribution of sample PWMs manifests as a linear combination of order statistics. [23] proved that the vector of b ^ = ( b 1 ^ , b 2 ^ ) has asymptotically a multivariate normal distribution with mean b ^ = ( b 1 ^ , b 2 ^ ) and covariance matrix n 1 V . Supporting this, [24] employed the delta method to show that the covariance matrix can be articulated in terms of the variance of the parameter being estimated; hence, using the delta method, the covariance matrix has the expression of n 1 G V G , where V is the variance of the parameter. The quantity J ( θ ( a ) ) J ( θ ( a ) ) + λ ( a ) I ( a ) 1 serves as a good approximation of the variance-covariance matrix for the estimated parameters, known as the V matrix. Moreover, the G matrix, integral to the LM algorithm, functions as the Jacobian matrix, reflecting the relationships between the parameters being estimated. Harnessing these methods can vastly improve the accuracy and reliability of estimators, ultimately leading to more credible and impactful results in statistical analysis.
Using Taylor series expansion of the g 1 ( θ ) = M 101 and g 2 ( θ ) = M 110 around the θ 0 gives: g ( θ ^ ) = g ( θ 0 ) + ( θ ^ θ 0 ) g ( θ 0 ) . Therefore, the delta method can be applied. Taking the variance on both sides yields v a r [ g ( θ ^ ) ] = [ g ( θ 0 ) ] v a r ( θ ^ ) [ g ( θ 0 ) ] T where g ( θ 0 ) = g ( θ 0 ) θ 0 | θ 0 = θ ^ which is the Jacobian matrix in LM algorithm. Hence, v a r ( θ ^ ) = [ [ g ( θ 0 ) ] T ( v a r [ g ( θ ^ ) ] ) 1 [ g ( θ 0 ) ] ] 1 and v a r [ g ( θ ^ ) ] = v a r ( M 101 c o v ( M 101 , M 110 ) c o v ( M 101 , M 110 ) v a r ( M 110 ) . This matrix, [ g ( θ 0 ) ] T ( v a r [ g ( θ ^ ) ] ) 1 [ g ( θ 0 ) ] may have a high condition number that inflates the v a r ( θ ^ ) .
To ameliorate this, an appropriate regularization factor ( ξ ) can be added to the matrix like this [ g ( θ 0 ) ] T ( v a r [ g ( θ ^ ) ] 1 [ g ( θ 0 ) ] + ξ I ] 1 , where ξ = λ m a x C * λ m i n C * 1 such that λ m a x is the maximum eigenvalue of the matrix [ g ( θ 0 ) ] T ( v a r [ g ( θ ^ ) ] 1 [ g ( θ 0 ) ] and λ m i n is the minimum eigenvalue for the same matrix. While the C * is the desired condition number. The desired value of this number is usually from 5 up to 10. Other types of variance is the value of this quantity [ J ( θ ( a ) ) J ( θ ( a ) ) + λ ( a ) I ( a ) ] 1 which can be considered a good approximation to the variance-covariance matrix of the estimated parameters θ ^ . The delta method is applied to derive variance of the function of the estimated parameters.
The same discussion is applied to M 1 , 0 , 2 , M 1 , 2 , 0 . See appendix B for the variance and covariance of the lower and higher order PWMs.

3. Monte Carlo Simulation Study

3.1. Simulation Study for the Parameters

A simulation was conducted by generating 1000 replicates (N) of the MBUW random variable y, each replicate with different sample sizes (n). The author chose suitable pairs of α and β parameters respectively as follows: ( 0.5 , 0.6 ) , ( 0.5 , 1.3 ) , ( 1.1 , 0.6 ) , and ( 1.1 , 1.6 ) . The sample sizes were 20, 50, 100, and 500. For each sample, the parameters were estimated using the PWMs method (the low and high orders), maximum likelihood (MLE) and the Method of Moments (MOM: the first and the second raw moments). For the MOM, m 1 ( y ) = μ 1 and m 2 ( y ) = μ 2 where m 1 ( y ) = 1 n i = 1 n y i = y ¯ and m 2 ( y ) = 1 n i = 1 n y i 2 , μ 1 = 6 ( 2 + α β ) ( 3 + α β ) and μ 2 = 6 ( 2 + 2 α β ) ( 3 + 2 α β ) . For each method, where θ ^ = α ^ β ^ , the statistical indices used for the comparisons are:
  • Average Absolute Bias (AAB) = 1 N j = 1 N θ ^ i θ
  • Mean Square Error (MSE) = 1 N j = 1 N θ ^ i θ 2
  • Mean Relative Error (MRE) = 1 N j = 1 N θ ^ θ θ
The empirical variance for each parameter = MSE − bias2.
The 2.5 th and the 97.5 th quantiles of the sampling distribution for each parameter are also recorded.
The number of the valid samples, which shows GoF, is recorded. These results are illuminated in Tables (Table 1, Table 2, Table 3 and Table 4). Each table shows the results of the comparisons between the methods for each sample size. The method yielding the estimator with the least bias is colored. Tables (Table 5, Table 6, Table 7 and Table 8) show the empirical variance-covariance matrices obtained from the sampling distribution of the parameters. These last four tables demonstrate the Pearson correlation coefficient between the parameters, the condition number and eigenvalues of the empirical covariance matrix. Table (Table 9) expounds the variance decay rate for each simulation validating the asymptotic consistency of the estimator. ( θ ^ n ) as n increases. In asymptotic theory, for a consistent estimator θ ^ n of a parameter θ , the variance typically behaves like var ( θ ^ n ) C n for large n, where C is some constant. This means that the variance decays inversely with sample size, it decays as 1 / n . Taking log of both sides of var ( θ ^ n ) C n gives log [ var ( θ ^ n ) ] log ( C ) log n so the slope in the log-log plot of the variance vs. sample size should be approximately ( 1 ) . For each simulation, the last two values of the variance of each parameter at n = 500 and n = 100 will be substituted in this log equation; in other words, the slope is calculated as log [ var ( θ ^ n = 500 ) ] log [ var ( θ ^ n = 100 ) ] log 500 log 100 1 . This is calculated for ( α ^ ) & ( β ^ ). Tables (Table 10, Table 11, Table 12 and Table 13) illustrate the computational time for each simulation and the relative efficiency of each method vs. other methods. The computational time is recorded in seconds per replication, and the total time of all replicates is also recorded. The relative efficiency of estimator A obtained from method A to estimator B gained from method B is defined as RE ( A vs B ) = M S E B ^ M S E A ^ . This means that estimator A is more efficient than estimator B if this RE is more than 1.

3.2. Simulation Study for the Function of the Parameters (Validation of the Delta Method)

Another simulation was conducted to validate the asymptotic normality of the delta method as follows: for sample size, n = 500 , α = 0.5 , β = 1.3 , and N = 1000
Step 1: for each replicate the g 1 ( θ ^ ) = M ^ 101 and g 2 ( θ ^ ) = M ^ 110 .
Step 2: g 1 ( θ 0 ) = M 101 and g 2 ( θ 0 ) = M 110 where θ 0 = [ α = 0.5 , β = 1.3 ] .
Step 3: e r r o r 101 = M ^ 101 M 101 and e r r o r 110 = M ^ 110 M 110
Step 4: calculate the variance-covariance between e r r o r 101 and e r r o r 110 , say it is ( cov e m p ). This is the empirical covariance matrix (2 by 2 matrix).
Step 5: obtain the theoretical covariance matrix ( cov t h e o ) using the delta method applied on each replicate then take the mean, let us call this mean _ cov t h e o
Step 6: apply the singular value decomposition (SVD) on the error matrix error = [ error 101 error 110 ] . The error matrix is N by 2 matrix. [ error 101 error 110 ] = U 1 * S 1 * V 1 T , where S 1 is the large non-zero singular value, and V 1 is the corresponding principal vector, λ e m p i r i c a l = S 1 2 / ( N 1 ) . Let’s call S e m p = [ error 101 error 110 ] * V 1 , is the projection of the error matrix in the direction of the principle vector. Let’s call Z e m p i r i c a l = S e m p / λ e m p i r i c a l , this standardized empirical Z should be distributed as standard normal. These Z s are the standardized errors after projection along the direction of maximum variability.
Step 7: get the trace of cov e m p and the trace of Mean _ cov t h e o . Let’s call scalar =
trace ( cov emp ) / trace ( mean _ cov theo ) . Let’s call cov theo_scaled = scalar * mean _ cov theo .
Step 8: apply the singular value decomposition (SVD) on mean _ cov t h e o so the mean _ cov t h e o = U 2 * S 2 * V 2 T where S 2 is the large non-zero singular value, and V 2 is the corresponding principal vector. Let us call S t h e o = [ error 101 error 110 ] * V 2 is the projection of the error matrix in the direction of the maximum variability of the theoretical covariance matrix.
Step 9: calculate λ theo_scaled = scalar * V 2 * mean_cov theo × V 2 T . Then calculate Z theo_scaled = S theo λ theo_scaled . This standardized theoretical Z should be distributed as standard normal.
The intuition behind this algorithm is that the theoretical covariance matrix (obtained from the delta method) should approximate the empirical covariance matrix from the replicates. The empirical and the theoretical standardized error can be defined, respectively, as
Z emp = Σ emp 1 / 2 g ( θ ^ ) g ( θ 0 ) and Z theo_scaled = Σ theo_scaled 1 / 2 g ( θ ^ ) g ( θ 0 ) and to hold for asymptotic normality these Z’s should be distributed as standard normal. A consistent observation across simulations (for all pairs of parameter and at different sample sizes) was that both covariance matrices (empirical and theoretical) shared nearly identical eigenvectors, as revealed by the singular value decomposition (SVD). This indicates that the empirical and theoretical covariance matrices capture a similar dependence structure between the two parameters. However, notable differences in their singular values were observed, particularly for large sample size. The theoretical matrix was a full rank for small sample size (n=20), but it is rank-one as n increases (n=500). This degeneration reflects the increasing linear dependency between the two parameter estimators as the sample size grows, implying that one parameter becomes an almost deterministic function of the other in large samples. To adjust for the overall discrepancy between the empirical and the theoretical covariance, a scalar rescaling factor was applied to the estimated theoretical matrix. This factor is trace ( cov emp ) / trace ( mean_cov theo ) , where the trace of covariance matrix represents the total marginal variance. Multiplying the theoretical covariance matrix by this scalar aligns the overall variance magnitude of the theoretical covariance with that of the empirical one while preserving its correlation structure. In other words, the rescaling equalizes the total variability but does not alter the orientation (the dependence structure, the eigenvectors) between the parameters. After this adjustment, the standardized vector, the theoretical Z, exhibited approximately zero mean and unit standard deviation, confirming that the scaling effectively normalized the parameter estimators (function of parameters). Hence, the asymptotic normality of the delta method was thus validated: as n increased, the distribution of the standardized estimators (PWMs are functions of the parameters) approached the standard normal, and the theoretical covariance structure aligned closely with the empirical results, despite the asymptotic rank deficiency inherent in the parameter dependence. Figure 3 illustrates the histogram, the fitted standard normal curves, and the QQ plot for the Z empirical and the Z theoretical (scaled).

4. Results and Discussion of the Simulation Study

Table 1, Table 2, Table 3 and Table 4 show that at various sample sizes where ( α = 0.5 & β = 0.6 ), the estimated ( α ^ ) obtained by MLE has the least Average Absolute Bias (AAB) and the least Mean Square Error (MSE) among the methods while the estimated ( β ^ ) obtained by M 1 , 0 , 2 , M 1 , 2 , 0 has the least AAB and MSE among the methods. This is also true at simulation runs using the pair ( α = 0.5 & β = 1.3 ) and the pair ( α = 1.1 & β = 1.6 ). However, at true values of the pair ( α = 1.1 & β = 0.6 ), the higher order PWMs ( M 1 , 0 , 2 , M 1 , 2 , 0 ) produce ( α ^ ) and ( β ^ ) with the least AAB and MSE among the methods. The tables record the number of valid samples that pass the GoF in each run of the simulation. Method of Moments (MOM) gives non-identifiable estimates of the parameters at ( α = 1.1 & β = 0.6 ) and ( α = 1.1 & β = 1.6 ) especially at small sample size; n = 20 and n = 50 . Moreover, at large sample sizes; n = 100 and n = 500 the number of valid samples gained by (MOM) are less than the number of valid samples obtained by other methods at the same values of the parameters pair. MOMs almost give the highest AAB among other methods at different sample sizes and at different pairs of parameters. MOMs almost yield the highest MSE among other methods across different sample sizes and across different pairs of the simulated parameters. The AAB of the higher order PWMs is approximately equal to that obtained from the lower order PWMs especially at ( α = 0.5 & β = 1.3 ) and at almost all sample sizes. The standard error of each of the estimated parameter is recorded beside the estimated value of the parameter. It is also obvious that the AAB and the MSE decrease as the sample size increases. This is almost true for all methods and for both estimated parameters. And this supports that the methods yield consistent estimators for the parameters.
Table 5 shows the variance-covariance matrix for each estimated parameter ( α ^ ) and ( β ^ ) obtained by the lower order PWMs ( M 1 , 0 , 1 , M 1 , 1 , 0 ). These estimated parameters are perfectly correlated as seen in all simulations across different sample sizes and across different pairs. These matrices show a high condition number and one eigenvalue that is nearly zero. This may lead to an identifiability problem and inflate the estimated variance which impairs the confidence interval construction. The lower-order PWMs are derived from the order statistics so correlation is inherent in the method. The question is does the distribution contribute to this high dependency? To answer this question, the correlation between the estimated parameters obtained from deploying other methods was investigated using the sampling distribution gained from the simulation. Table 6 depicts the sampling distribution obtained from applying the higher order PWMs ( M 1 , 0 , 2 , M 1 , 2 , 0 ) for parameters estimation. The same outcomes were observed. But this may be due to the PWMs inheriting the high dependency from the order statistics. But as the same results were also returned from MLE and MOMs as shown in Table 7Table 8, this is strong evidence that the distribution is the main root cause. Although the method can contribute, but when every reasonable method shows the same correlation patterns, so the distribution itself is likely to cause this coupling. The Weibull distribution and any derived distribution from the Weibull, like the one between our hands (MBUW), the parameters control related features of the sample data (mean, spread, and tail) by giving information about some combinations of the parameters more strongly than about each parameter separately. Whether the dependency relation is linear, nonlinear, monotonic, or non-monotonic, the method can influence this. Figure 1 shows the relationship between the estimated ( α ^ ) and estimated ( β ^ ) obtained from the simulation using the lower order PWMs at n=20. Table 6, Table 7 and Table 8 show the sampling distribution of the estimated parameters obtained by the simulation study. The empirical variance-covariance matrices and the Pearson correlation coefficient are recorded. The condition number and the eigenvalues of the covariance matrix for each simulation are also shown. Other figures for the different shapes of the dependency are seen in Appendix C.
The importance to recognize that the cause of the correlation is the distribution itself is the application of the practical remedies to lessen this correlation and hence to deflate the variance and to construct a valid confidence interval. Some of these remedies is to reparameterize the parameter and use the log of the parameter. These remedies can be a blueprint for future studies.
Table 9 expounds the pattern of the variance decay. The slopes in this table are obtained from the empirical variances for the estimated parameters recorded in Table 5, Table 6, Table 7 and Table 8. Using the log (variance) at n=500, log (variance) at n=100, log (n=500) and log (n=100), the slope can be calculated as previously explained in section 3, (Monte Carlo Simulation Study). Figure 2 also supports this asymptotic variance decay at a rate (1/n) as n enlarges, where the estimated parameter ( θ ^ n ) is asymptotically consistent at large sample size. Table 9 elucidates that the lower and the higher PWMs show better validity for this asymptotic consistency rather than the MLE and MOMs.
Table 10, Table 11, Table 12 and Table 13 show the relative efficiency of each method against other methods. The higher order PWMs ( M 1 , 0 , 2 , M 1 , 2 , 0 ) are slightly more efficient than the lower order PWMs ( M 1 , 0 , 1 , M 1 , 1 , 0 ). This can be confirmed by calculating the relative efficiency of the lower order to the higher order PWMs. Across all the sample sizes, this RE at the pairs ( α = 0.5 & β = 0.6 ) is in the range from 0.94 to 0.95 for both estimated parameters. While at the pairs ( α = 0.5 & β = 1.3 ), it is in the range from 0.998 to1. Furthermore, the RE at the pair ( α = 1.1 & β = 0.6 ) is around 0.8. Additionally, the ER at the pairs ( α = 1.1 & β = 1.6 ) is in the range from 0.94 to 0.95. So they are more or less close to each other if taking into account the computational time reported in seconds per replicate. At sample size 20, the time taken by the lower PWMs are slightly less than that taken by the higher order PWMs to accomplish the estimation. This is also true for sample size 50 except at the pair ( α = 0.5 & β = 1.3 ) where the time taken by the lower order PWMs is 0.1289 seconds per replicate while this time is 0.1264 seconds per replicate for the higher order PWMs to complete the task of estimation. At sample size 100, the lower order PWMs take less time than the higher order PWMs. At sample size 500, and at pairs ( α = 0.5 & β = 0.6 ) and ( α = 0.5 & β = 1.3 ), the lower order PWMs are faster than the higher order PWMs, in contrast to the pair ( α = 1.1 & β = 0.6 ) and the pair ( α = 1.1 & β = 1.6 ), where the higher order PWMs are faster than the lower order PWMs. The time taken by both the lower and higher PWMs is linearly increasing across different sample sizes as n increases. The time taken by the MLE is less than that taken by PWMs of both orders. The time taken by MOMs is the least among other methods. The time taken by MLE and MOMs does not vary markedly across the sample sizes as seen in PWMs.
Table 10, Table 11, Table 12 and Table 13 almost depict that the descending order of the efficient methods from the most efficient to the least efficient in estimating ( α ) parameter across different sample sizes and across different pairs is MLE, ( M 1 , 0 , 2 , M 1 , 2 , 0 ), ( M 1 , 0 , 1 , M 1 , 1 , 0 ) then MOM. The exception for this is the pair ( α = 1.1 & β = 0.6 ), which shows the descending order is ( M 1 , 0 , 2 , M 1 , 2 , 0 ), MLE, ( M 1 , 0 , 1 , M 1 , 1 , 0 ) and finally MOM across different sample sizes excluding n=500 & for this pair ( α = 1.1 & β = 0.6 ) the order is ( M 1 , 0 , 2 , M 1 , 2 , 0 ), ( M 1 , 0 , 1 , M 1 , 1 , 0 ), MLE then MOM. While the descending order of the efficient methods in estimating ( β ) across different sample sizes and different pairs is ( M 1 , 0 , 2 , M 1 , 2 , 0 ), ( M 1 , 0 , 1 , M 1 , 1 , 0 ), MLE then MOM. The exception for this is the pair ( α = 0.5 & β = 0.6 ), which shows the descending order is ( M 1 , 0 , 2 , M 1 , 2 , 0 ), ( M 1 , 0 , 1 , M 1 , 1 , 0 ), MOM and lastly MLE across different sample sizes. This last order is also true when n=500 and the pair is ( α = 0.5 & β = 1.3 ).
Figure 3 shows the validation of asymptotic normality of the delta method. The error obtained from the difference between g ( θ ^ ) and the g ( θ 0 ) where θ 0 are the pair of parameter ( α = 0.5 & β = 1.3 ), the sample size is 500 and the function is the lower order PWMs, is standardized by the empirical covariance matrix. The same error is also standardized by the theoretical covariance matrix obtained via applying the delta method. However, both matrices the empirical and the theoretical matrices are nearly singular whose one of their eigenvalues are nearly zero. So both matrices are subjected to singular value decomposition (SVD) so as to project the error along the principle vector with high variability (the vector corresponding to the large non-zero singular value). And to match the empirical with the theoretical matrix, the last one is multiplied by a small scalar that is the ratio between the trace of the empirical covariance and the trace of the theoretical covariance. Both standardized error should be a standard normal variable as shown in Figure 3. This procedure was repeated for the higher order PWMs with different sample sizes and different pairs of the parameters. The same results were obtained which confirms the asymptotic normality of the delta method. The regularity conditions essential to apply the delta method is that the function should be a continuously differentiable function which also holds for the PWMs of both orders.
Figure 1. shows the perfect linear relationship between estimated parameters from the simulation study at n=20 using lower order PWMs
Figure 1. shows the perfect linear relationship between estimated parameters from the simulation study at n=20 using lower order PWMs
Preprints 189293 g001
Figure 2. Shows the variance decay emphasizing asymptotic consistency. The slope in a log-log plot of the empirical variance and sample size is approximately 1 , as asymptotic consistency of ( θ ^ n ) entails
Figure 2. Shows the variance decay emphasizing asymptotic consistency. The slope in a log-log plot of the empirical variance and sample size is approximately 1 , as asymptotic consistency of ( θ ^ n ) entails
Preprints 189293 g002
Figure 3. shows the histogram, fitted standard normal curve and QQ plot for the error derived from the difference between the function of the parameter at the estimated parameters and the true parameters; g ( θ ^ ) g ( θ 0 ) . The error standardized by the empirical covariance matrix obtained from the empirical distribution is a standard normal variable (left upper and lower panels). The same error standardized by the covariance matrix obtained from rescaling the theoretical covariance matrix obtained via applying the delta method on each replicate is also a standard normal variable. (right upper and lower panels). This is the results of the simulation using n=500, N=1000, α = 0.5 and β = 1.3 and the lower order PWMs (M101,M110)
Figure 3. shows the histogram, fitted standard normal curve and QQ plot for the error derived from the difference between the function of the parameter at the estimated parameters and the true parameters; g ( θ ^ ) g ( θ 0 ) . The error standardized by the empirical covariance matrix obtained from the empirical distribution is a standard normal variable (left upper and lower panels). The same error standardized by the covariance matrix obtained from rescaling the theoretical covariance matrix obtained via applying the delta method on each replicate is also a standard normal variable. (right upper and lower panels). This is the results of the simulation using n=500, N=1000, α = 0.5 and β = 1.3 and the lower order PWMs (M101,M110)
Preprints 189293 g003

5. Real data analysis

The OECD platform is available at https://stats.oecd.org/index.aspx?DataSetCode=BLI
In OECD platform, many indicators are recorded. The author transformed these indicators into unit ratios and conducted a distributional fit for those indicators. In this paper, the author discussed three indicators; the water quality, the educational attainment and the self-reported health. Table 14 shows the descriptive analysis of the indicators. Figure 4 shows the boxplot of each indicator.
Water quality: is the percentage of people who report being satisfied with the quality of their water. This reflects the proportion of people who have access to a clean water supply. The values of this indicator are 0.92, 0.92, 0.79, 0.90, 0.62, 0.82, 0.87, 0.89, 0.93, 0.86, 0.97, 0.78, 0.91, 0.67, 0.81, 0.97, 0.8, 0.77, 0.77, 0.87, 0.82, 0.83, 0.83, 0.85, 0.75, 0.91, 0.85, 0.98, 0.82, 0.89, 0.81, 0.93, 0.76, 0.97, 0.96, 0.62, 0.82, 0.88, 0.7, 0.62, 0.72.
Educational attainment: is presented as the percentage of a given population who have completed a specific level of education, mainly the tertiary education. The values of this indicator are 0.84, 0.86, 0.8, 0.92, 0.67, 0.59, 0.43, 0.94, 0.82, 0.91, 0.91, 0.81, 0.86, 0.76, 0.86, 0.76, 0.85, 0.88, 0.63, 0.89, 0.89, 0.94, 0.74, 0.42, 0.81, 0.81, 0.82, 0.93, 0.55, 0.92, 0.90, 0.63, 0.84, 0.89, 0.42, 0.82, 0.92, 0.57, 0.95, 0.48.
Self-reported health is presented as the percentage of the population that rates their health as good or very good, and this reflects access to health services, healthy living conditions, and better lifestyle choices (diets and exercises). The values of this indicator are 0.85, 0.71, 0.74, 0.89, 0.60, 0.8, 0.73, 0.62, 0.7, 0.57, 0.68, 0.67, 0.66, 0.79, 0.58, 0.77, 0.84, 0.74, 0.73, 0.37, 0.34, 0.47, 0.46, 0.72, 0.66, 0.75, 0.86, 0.75, 0.6, 0.5, 0.65, 0.67, 0.75, 0.76, 0.81, 0.67, 0.73, 0.88, 0.43.
Table 14. Summary of descriptive statistics for each indicator.
Table 14. Summary of descriptive statistics for each indicator.
Preprints 189293 i009
Table 15. Estimators and validation indices for the water quality dataset
Table 15. Estimators and validation indices for the water quality dataset
Preprints 189293 i010
The variables exhibit left skewness with different degrees. Educational attainment shows significant left skewness then the self-reported health, and lastly the water quality indicator. Water quality shows mesokurtic shape, while the educational attainment and the self-reported health show slightly more than excess kurtosis (leptokurtic, the kurtosis coefficient is more than 3). So the data exhibit different shapes.
The different unit distributions tested was Beta, Kumaraswamy, unit Lindley, Topp Leone and the new MBUW distribution. MLE method was applied using the Nelder Mead optimizer in MATLAB. For each distribution, the estimated parameters, variance, Log-Likelihood, AIC, CAIC, BIC, and HQIC are recorded. Also the KS-test, AD, CVM test are reported. Figures for the fitted CDFs and fitted PDFs are shown.
Beta Distribution: f ( y ; α , β ) = Γ ( α + β ) Γ ( α ) Γ ( β ) y α 1 ( 1 y ) β 1 , 0 < y < 1 , α > 0 , β > 0
Kumaraswamy Distribution: f ( y ; α , β ) = α β y α 1 ( 1 y α ) β 1 , 0 < y < 1 , α > 0 , β > 0
Topp–Leone Distribution: f ( y ; θ ) = θ ( 2 2 y ) ( 2 y y 2 ) θ 1 , 0 < y < 1 , θ > 0
Unit–Lindley: f ( y ; θ ) = θ 2 1 + θ ( 1 y ) 3 exp θ y 1 y , 0 < y < 1 , θ > 0
Table 15 shows that all the unit distributions fit the water quality data well. The variance-covariance obtained after fitting the MBUW distribution is not identified, although the other statistical indices like AIC, CAIC, BIC, HQIC, LL are comparable to other distributions. Figure 5Figure 6 show that the fitted CDF and the fitted PDF of the MBUW align with the fitted CDFs and fitted PDFs of both the Beta distribution and the Kumaraswamy distribution. But the problem is mainly the variance obtained from employing the MLE method utilizing Nelder Mead optimizer in MATLAB. Using the values of the estimated parameter as initial guesses to substitute in the LM algorithm and PWMs (lower and higher orders) gives the results shown in Table 18 and in Figure 11Figure 12.
Table 16 shows that all the unit distributions fit the educational attainment data well except the Topp–Leone distribution. The same trouble of the variance-covariance obtained after fitting the MBUW distribution is shown, even though the other statistical indices like AIC, CAIC, BIC, HQIC, LL are more or less analogous to other fitting distributions. Figures (Figure 7Figure 8) show that the fitted CDF and the fitted PDF of the MBUW is side by side with the fitted CDFs and fitted PDFs of both the Beta distribution and the Kumaraswamy distribution. But then again; the trouble is essentially the variance gained from deploying the MLE method consuming Nelder Mead optimizer in MATLAB. Treating the values of the estimated parameters as initial guesses to replace in the LM algorithm and PWMs (lower and higher orders) produces the outcomes shown in (Table 19) and in Figures (Figure 13Figure 14).
Table 16. Estimators and validation indices for the educational attainment dataset
Table 16. Estimators and validation indices for the educational attainment dataset
Beta Kumaraswamy MBUW Topp-Leone Unit-Lindley
theta α = 6.0373 α = 5.4735 α = 0.4333 12.3585 0.2891
β = 1.7313 β = 1.9272 β = 1.3477
Var-cov 2.9116 0.9141 1.0268 0.3651 0.0000 0.0000 3.8183 0.0011
0.9141 0.3410 0.3651 0.2227 0.0000 0.0000
SE 1.7063 1.0133 1.9540 0.0332
0.5839 0.4719
AIC −48.1413 −48.8385 −48.0574 −40.9642 −57.3738
CAIC −47.8170 −48.5142 −47.7331 −40.8589 −57.2685
BIC −44.7636 −45.4607 −44.6796 −39.2753 −55.6849
HQIC −46.9200 −47.6172 −46.8361 −40.3535 −56.7631
LL 26.0707 26.4193 26.0287 21.4821 29.6869
K-S Value 0.1481 0.1434 0.1577 0.2598 0.0697
H0 Fail to reject Fail to reject Fail to reject Reject Fail to reject
P-value 0.1613 0.1841 0.1219 0.0023 0.9442
AD 1.1861 1.1575 1.4386 4.5709 0.2596
CVM 0.2068 0.1985 0.2546 0.8485 0.0344
Table 17. Estimators and validation indices for the self-reported health dataset
Table 17. Estimators and validation indices for the self-reported health dataset
Beta Kumaraswamy MBUW Topp-Leone Unit-Lindley
theta α = 8.0698 α = 5.7073 α = 0.5771 7.3278 0.5929
β = 3.8357 β = 5.2250 β = 1.2824
Var-cov 3.3978 1.6515 0.7936 1.2015 0.0000 0.0000 1.3768 0.0058
1.6515 0.9425 1.2015 2.5192 0.0000 0.0000
SE 1.8433 0.8908 1.1734 0.0693
0.9708 1.5872
AIC −46.2684 −47.3865 −38.5536 −44.4298 −40.09617
CAIC −45.9351 −47.0531 −38.2202 −44.3217 −39.9617
BIC −42.9413 −44.0594 −35.2264 −42.7662 −38.4062
HQIC −45.0746 −46.1927 −37.3598 −43.8329 −39.4729
LL 25.1342 25.6932 21.2768 23.2149 21.0349
K-S Value 0.0867 0.0684 0.1625 0.1234 0.1400
H0 Fail to reject Fail to reject Fail to reject Fail to reject Fail to reject
P-value 0.6674 0.8577 0.2284 0.5512 0.2102
AD 0.4674 0.3414 1.4935 0.8744 1.6862
CVM 0.0807 0.0547 0.2492 0.1397 0.2886
Table 18. Comparisons of the results of the PWMs method for parameter estimation of the water quality dataset using M 1 , 0 , 1 & M 1 , 1 , 0 and M 1 , 0 , 2 & M 1 , 2 , 0
Table 18. Comparisons of the results of the PWMs method for parameter estimation of the water quality dataset using M 1 , 0 , 1 & M 1 , 1 , 0 and M 1 , 0 , 2 & M 1 , 2 , 0
Using unbiasedSample estimator for M 1 , 0 , 1 and M 1 , 1 , 0 Using unbiasedSample estimator for M 1 , 0 , 2 and M 1 , 2 , 0
thetas α 0.3564 0.3529
β 1.4307 1.4316
Var-cov matrixof parameter 10.7370 22.9418 627.8954 2.4064e+3
22.9418 94.1037 2.4064e+3 9.3821e+3
Eigenvalues=4.8406, 100 Eigenvalue=10.0408, 10000
AD 0.3418 (p=0.8810) 0.3820 (p=0.865)
CVM 0.0531 (p=0.8420) 0.0637 (p=0.785)
KS & p-value 0.0980 (p=0.7901) 0.1061 (p=0.7051)
H0 Fail to reject Fail to reject
SSE 1.8169e-19 1.7110e-19
α ^ s , unbiased estimator 0.3891 0.2491
β ^ r , unbiased estimator 0.4441 0.3041
Sig. of α parameter 11.0082 ( p < 0.001 ) 11.0053 ( p < 0.001 )
Sig. of β parameter 17.9465 ( p < 0.001 ) 18.1268 ( p < 0.001 )
Variance of thefunction of theparameter, after thedelta method application.Determinant and traceof this matrix 0.0181 0.0096 0.0214 0.0080
0.0096 0.0051 0.0080 0.0030
Eigenvalues=0.0232, 8.6736e-19Determinant=0Trace=0.0232 Eigenvalues=0.0244, -4.3368e-19Determinant=0Trace=0.0244
Var-cov betweenM101&M110 0.1187e-3 0.0543e-3
0.0543e-3 0.0430e-3
Var-cov betweenM102 & M120 0.7094e-4 0.1719e-4
0.1719e-4 0.1547e-4
var ( θ ^ ) andassociated ξ 0.0010 0.0015 0.0010 0.0014
0.0015 0.0064 0.0014 0.0062
ξ = 148.5817 to achieved condition number=10 ξ = 151.4053 to achieved condition number=10
Jacobian matrix 0.3796 0.0977 0.2863 0.0735
1.4718 0.3786 0.1067 0.0274
Table 19. Comparisons of the results of the PWMs method for parameter estimation of the educational attainment dataset using M 1 , 0 , 1 & M 1 , 1 , 0 and M 1 , 0 , 2 & M 1 , 2 , 0
Table 19. Comparisons of the results of the PWMs method for parameter estimation of the educational attainment dataset using M 1 , 0 , 1 & M 1 , 1 , 0 and M 1 , 0 , 2 & M 1 , 2 , 0
Using unbiasedSample estimator for M 1 , 0 , 1 and M 1 , 1 , 0 Using unbiasedSample estimator for M 1 , 0 , 2 and M 1 , 2 , 0
thetas α 0.4310 0.4462
β 1.3483 1.3442
Var-cov matrixof parameter 679.8845 2.5075e+3 679.7206 2.4966e+3
2.5075e+3 9.3254e+3 2.4966e+3 9.3313e+3
Eigenvalues=5.2672, 10000 Eigenvalue=10.9762, 10000
AD 1.4027 (p=0.2090) 1.6993( p=0.1330)
CVM 0.2420 (p=0.1990) 0.3333(p=0.1110)
KS & p-value 0.1787 (p=0.1371) 0.2041 (p=0.0615)
H0 Fail to reject Fail to reject
SSE 2.2362e-19 2.4395e-19
α ^ s , unbiased estimator 0.3485 0.2139
β ^ r , unbiased estimator 0.4325 0.2979
Sig. of α parameter 10.1191 ( p < 0.001 ) 12.0627 ( p < 0.001 )
Sig. of β parameter 13.0970 ( p < 0.001 ) 18.8075 ( p < 0.001 )
Variance of thefunction of theparameter, after thedelta method application.Determinant and traceof this matrix 0.0190 0.0107 0.0213 0.0089
0.0107 0.0060 0.0089 0.0037
Eigenvalues=0.0250, 2.6021e-18Determinant=0Trace=0.0250 Eigenvalues=0.0250, 4.3368e-19Determinant=0Trace=0.0250
Var-cov betweenM101&M110 0.1848e-3 0.0924e-3
0.0924e-3 0.0785e-3
Var-cov betweenM102 & M120 0.1136e-3 0.0321e-3
0.0321e-3 0.0328e-3
var ( θ ^ ) andassociated ξ 0.0018 0.0025 0.0019 0.0027
0.0025 0.0106 0.0027 0.0114
ξ =88.6209 to achieve a condition number=10 ξ =82.3445 to achieved condition number=10
Jacobian matrix 0.3668 0.0987 0.2691 0.0721
0.2059 0.0554 0.1119 0.0300
Table 20. Comparisons of the results of the PWMs method for parameter estimation of the self-reported health dataset using M 1 , 0 , 1 & M 1 , 1 , 0 and M 1 , 0 , 2 & M 1 , 2 , 0
Table 20. Comparisons of the results of the PWMs method for parameter estimation of the self-reported health dataset using M 1 , 0 , 1 & M 1 , 1 , 0 and M 1 , 0 , 2 & M 1 , 2 , 0
Using unbiasedSample estimator for M 1 , 0 , 1 and M 1 , 1 , 0 Using unbiasedSample estimator for M 1 , 0 , 2 and M 1 , 2 , 0
thetas α 0.5843 0.5752
β 1.2806 1.2829
Var-cov matrixof parameter 11.610 21.6717 71.1767 230.3148
21.6717 94.6865 230.3148 942.8902
Eigenvalues=6.2964, 100 Eigenvalue=14.0669, 1000
AD 1.4443 (p=0.2010) 1.5100 (p=0.1780)
CVM 0.2382 (p=0.2100) 0.2529 (p=0.1900)
KS & p-value 0.1549 (p=0.2768) 0.1645 (p=0.2167)
H0 Fail to reject Fail to reject
SSE 5.9631e-19 0
α ^ s , unbiased estimator 0.3019 0.1865
β ^ r , unbiased estimator 0.3775 0.2621
Sig. of α parameter 9.7573 ( p < 0.001 ) 9.7503 ( p < 0.001 )
Sig. of β parameter 8.5322 ( p < 0.001 ) 8.7119 ( p < 0.001 )
Variance of thefunction of theparameter, after thedelta method application.Determinant and traceof this matrix 0.0172 0.0108 0.0205 0.0099
0.0108 0.0068 0.0099 0.0047
Eigenvalues=0.0240, 1.7347e-18Determinant=0Trace=0.0240 Eigenvalues=0.0253, 0Determinant=0Trace=0.0253
Var-cov betweenM101&M110 0.2856e-3 0.1636e-3
0.1636e-3 0.1541e-3
Var-cov betweenM102 & M120 0.1556e-3 0.0524e-3
0.0524e-3 0.0604e-3
var ( θ ^ ) andassociated ξ 0.0036 0.0049 0.0035 0.0048
0.0049 0.0225 0.0048 0.0217
ξ =42.1234 to achieve a condition number=10 ξ =43.7133 to achieve a condition number=10
Jacobian matrix 0.3062 0.0717 0.2317 0.0574
1.5296 0.3579 0.1112 0.0276
Figure 7. shows empirical CDF vs. fitted CDFs for different competitors for the educational attainment dataset
Figure 7. shows empirical CDF vs. fitted CDFs for different competitors for the educational attainment dataset
Preprints 189293 g007
Figure 8. shows the histogram and the fitted PDFs for different competitors for the educational attainment dataset
Figure 8. shows the histogram and the fitted PDFs for different competitors for the educational attainment dataset
Preprints 189293 g008
Figure 9. shows the empirical CDF vs the fitted CDFs for different competitors for self-reported health
Figure 9. shows the empirical CDF vs the fitted CDFs for different competitors for self-reported health
Preprints 189293 g009
Figure 10. shows the histogram vs the fitted PDFs for different competitors for self-reported health
Figure 10. shows the histogram vs the fitted PDFs for different competitors for self-reported health
Preprints 189293 g010
Figure 11. shows the eCDF vs. fitted CDF for MLE & low order PWMs, M101, M110 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for water quality dataset
Figure 11. shows the eCDF vs. fitted CDF for MLE & low order PWMs, M101, M110 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for water quality dataset
Preprints 189293 g011
Figure 12. shows the eCDF vs. fitted CDF for MLE & high order PWMs, M102, M120 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for water quality dataset
Figure 12. shows the eCDF vs. fitted CDF for MLE & high order PWMs, M102, M120 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for water quality dataset
Preprints 189293 g012
Figure 13. shows the eCDF vs. fitted CDF for MLE & low order PWMs, M101, M110 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for educational attainment dataset
Figure 13. shows the eCDF vs. fitted CDF for MLE & low order PWMs, M101, M110 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for educational attainment dataset
Preprints 189293 g013
Figure 14. shows the eCDF vs. fitted CDF for MLE & high order PWMs, M102, M120 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for educational attainment dataset
Figure 14. shows the eCDF vs. fitted CDF for MLE & high order PWMs, M102, M120 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for educational attainment dataset
Preprints 189293 g014
Figure 15. shows the eCDF vs. fitted CDF for MLE & low order PWMs, M101, M110 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for self-reported health dataset
Figure 15. shows the eCDF vs. fitted CDF for MLE & low order PWMs, M101, M110 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for self-reported health dataset
Preprints 189293 g015
Figure 16. shows the eCDF vs. fitted CDF for MLE & high order PWMs, M102, M120 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for self-reported health dataset
Figure 16. shows the eCDF vs. fitted CDF for MLE & high order PWMs, M102, M120 (upper left subplot), histogram & fitted PDFs ( upper right subplot), QQ plot (lower left) and PP plot (lower right) for self-reported health dataset
Preprints 189293 g016
Table 17 shows that all the unit distributions fit the self-reported health data well. The same dilemma of the variance-covariance obtained after fitting the MBUW distribution is displayed, albeit the other statistical indices like AIC, CAIC, BIC, HQIC, LL are slightly less than that of the Beta, Kumaraswamy, and the Topp–Leone distributions. Figure 9Figure 10 show that the fitted CDF and the fitted PDF of the MBUW is lined up with the Beta and Kumaraswamy at the lower tail and is aligned with the Topp–Leone and the Unit Lindley distribution at the upper tail. Nevertheless, the misfortune is fundamentally the variance returned from implementing the MLE method executing Nelder Mead optimizer in MATLAB. Employing the values of the estimated parameters as initial guesses to insert in the LM algorithm and PWMs (lower and higher orders) generates the sequels shown in Table 20 and in Figure 15Figure 16.
Table 18 shows the results of the water quality dataset. The values of the estimated parameters using the lower PWMs and higher PWMs are nearly equal. The approximated variance-covariance obtained from the LM algorithm using the lower order PWMs is less than that obtained from using the higher order PWMs. Applying the delta method on both lower and higher PWMs gives variances comparable to each other. The determinant for the matrix of the lower order (0.0232) is less than that of the higher order (0.0244). The Jacobian matrix is an ill-conditioned matrix for both PWMs. Using a regularization factor enhances the estimated variance of the estimated parameters. This regularized factor is chosen so the condition number of this matrix [ g ( θ 0 ) ] T ( v a r [ g ( θ ^ ) ] ) 1 [ g ( θ 0 ) ] is 10. This is because the estimated v a r ( θ ^ ) = [ g ( θ 0 ) ] T ( v a r [ g ( θ ^ ) ] ) 1 [ g ( θ 0 ) ] + ξ I 1 requires this factor ( ξ ). The standard errors of the estimated parameters are obtained by taking the square root of the diagonal of this last matrix. The estimated variance, v a r ( θ ^ ) , obtained from the lower PWMs is comparable to that obtained from the higher PWMs. The covariance between the samples PWMs is also reported: v a r [ g ( θ ^ ) ] = v a r ( M 101 ) c o v ( M 101 , M 110 ) c o v ( M 101 , M 110 ) v a r ( M 110 ) and it is smaller for the higher order PWMs than the covariance for the lower order PWMs. The AD, CVM, and KS tests are statistically significant for both order PWMs. The estimated ( α ^ ) & ( β ^ ) obtained from both PWMs are statistically significant. Figure 11Figure 12 show the fitted CDFs and the fitted PDFs for the lower and higher order PWMs respectively. The curves of the fitted CDF and the fitted PDF obtained from MLE are perfectly aligned with the curves obtained from both orders of PWMs. The QQ plot and PP plot are almost perfectly aligned with the diagonal.
Table 19 clarifies the same type of results for the educational attainment dataset. The values of the estimated parameters are nearly alike. The approximated variance yielded from the LM algorithm is nearly comparable between the two orders of the PWMs. After applying the delta method, the determinant of the covariance matrix obtained from both orders of the PWMs is identical (0.025). The variance of the estimated parameter, v a r ( θ ^ ) , is nearly comparable between the two orders of PWMs. The AD, CVM, KS tests are statistically significant. The estimated ( α ^ ) & ( β ^ ) obtained from both PWMs are statistically significant. Figure 13Figure 14 disclose the fitted CDFs and the fitted PDFs for the lower and higher order PWMs respectively. The fitted CDF and the fitted PDF exhibit curves gained from MLE that are seamlessly aligned with those curves attained from using both orders of PWMs. The QQ plot and PP plot display fair alignment with the diagonal at the upper tail and at the lower tail respectively.
Table 20 expresses similar type of the outcomes for the self-reported health dataset. The values of the estimated parameters are nearly indistinguishable. The approximated variance gained from the LM algorithm using the lower-order PWMs is smaller than that obtained with the higher-order PWMs. After applying the delta method, the determinant of the covariance matrix obtained from lower order PWMs (0.0240) is smaller than that produced from the higher order PWMs (0.0253). The variance of the estimated parameter, v a r ( θ ^ ) , is comparable between the two orders of the PWMs. The AD, CVM, KS tests are statistically significant. The estimated ( α ^ ) & ( β ^ ) obtained from both PWMs are statistically significant. Figure 15Figure 16 unveil the fitted CDFs and the fitted PDFs for the lower and higher order PWMs respectively. The curves of the fitted CDF and the fitted PDF obtained from MLE are perfectly aligned with the curves obtained from using both orders of PWMs. The QQ plot and PP plot illustrate reasonable alignment with the diagonal.

6. Conclusions

The classic (PWMs) method can be effectively utilized for the parameter estimation of the new Median Based Unit Weibull (MBUW) distribution. This method is robust against outliers and is more straightforward to obtain than the maximum likelihood estimator. In the context of the fitting datasets employed in this study, the unbiased estimators for both order of PWMs yielded comparable results. The PWMs method offers several advantages over alternative estimation techniques. It is both rapid and easy to compute, consistently producing feasible values for the estimated parameters and the estimated variances. Although the higher order PWMs take much longer time to execute at larger sample sizes than MLE, they are more efficient in estimating ( β ^ ). Furthermore, PWMs estimators exhibit asymptotic normal distributions. It was observed that higher-order moments like ( M 1 , 0 , 2 and M 1 , 2 , 0 ) did not significantly contribute more information than the commonly used moments ( M 1 , 0 , 1 and M 1 , 1 , 0 ). The relative efficiency of the lower order PWMs to the higher order PWMs is in the range from 0.8 to 1 depending on the values of the parameters. The Average Absolute Biases for both orders are more or less comparable and nearly equal so there is no big difference between them. The computational time is less for lower order PWMs than the time for the higher PWMs. The estimated parameters from both orders of the PWMs are asymptotically consistent. The delta method used to estimate the variance of the function of the parameters is asymptotically normal. Using regularization factor to estimate the variance of the estimated parameters, after accounting for the covariance between the sample PWMs, can be a solution to mitigate the dependency between the parameters. The main root cause of this dependency between the parameters is the distribution itself. The method may configure the shape of the dependency, being linear or non-linear and monotonic or non-monotonic.
In this paper, the formula employed for defining PWMs is primarily based on the binomial expansion of the cumulative distribution function and the survival function, thus integrating with respect to the variable (dy) rather than integrating with respect to the cumulative distribution function (dF). The quantile function of the MBUW is difficult to integrate in a sense to obtain a system of equations that can be solved numerically.

7. Future Work

PWMs serve as a powerful foundation for L-moments, paving the way for a wealth of analytical possibilities. In future research, we can leverage the estimation of various L-moments, including L-skewness and L-kurtosis, alongside the L-moment method for precise parameter estimation. Furthermore, by extending the ( β ) parameter of the Median Based unit Weibull (MBUW) distribution to accommodate negative values, we can broaden our analytical horizons. In such scenarios, the Generalized Probability-Weighted Moments (GPWMs) method stands out as an essential tool. Additionally, we can effectively apply Partial GPWMs to handle censored data, ensuring a comprehensive approach to our analyses. The Bayesian methods with different loss functions can be consumed in the future to estimate the parameters. The correlation between the two parameters is a proposal for future studies to alleviate this dependency and hence to enhance the estimation and inference procedures. The tail indices can be further assessed to see if they can benefit from higher order PWMs. These directions not only enhance our statistical methodologies but also significantly enrich our understanding of the underlying data distributions.

Appendix A

A code for the PWMs method using the LM algorithm is available at: https://doi.org/10.5281/zenodo.17867288
The Jacobian for the M101 & M110:
M 1 , 0 , 1 = 360 + 108 α β 1044 α β + 580 α 2 β + 155 α 3 β + 20 α 4 β + α 5 β + 720 = p q
Let us call numerator for M101 p and the denominator q
p α = 108 β α β 1 & p β = 108 α β ln α
q α = 1044 β α β 1 + 580 ( 2 β ) α 2 β 1 + 155 ( 3 β ) α 3 β 1 + 20 ( 4 β ) α 4 β 1 + ( 5 β ) α 5 β 1
q β = 1044 α β ln α + 580 ( 2 ) α 2 β ln α + 155 ( 3 ) α 3 β ln α + 20 ( 4 ) α 4 β ln α + ( 5 ) α 5 β ln α
M 1 , 0 , 1 α = p α ( q ) q α ( p ) q 2 & M 1 , 0 , 1 β = p β ( q ) q β ( p ) q 2
M 1 , 1 , 0 = 60 + 6 α β 74 α β + 15 α 2 β + α 3 β + 120 = R T
Let us call numerator for M110 R and the denominator T
R α = 6 β α β 1 & R β = 6 α β ln α
T α = 74 β α β 1 + 15 ( 2 β ) α 2 β 1 + ( 3 β ) α 3 β 1
T β = 74 α β ln α + 15 ( 2 ) α 2 β ln α + ( 3 ) α 3 β ln α
M 1 , 1 , 0 α = R α ( T ) T α ( R ) T 2 & M 1 , 1 , 0 β = R β ( T ) T β ( R ) T 2
The Jacobian for the M102 & M120:
M 1 , 0 , 2 = K L where K = 58320 α β + 6480 α 2 β + 120960
L = 663696 α β + 509004 α 2 β + 214676 α 3 β + 54649 α 4 β + 8624 α 5 β + 826 α 6 β + 44 α 7 β + α 8 β + 362880
K α = 58320 β α β 1 + 6480 ( 2 β ) α 2 β 1
K β = 58320 α β ln α + 6480 ( 2 ) α 2 β ln α
L α = 663696 β α β 1 + 509004 ( 2 β ) α 2 β 1 + 214676 ( 3 β ) α 3 β 1 + 54649 ( 4 β ) α 4 β 1 + 8624 ( 5 β ) α 5 β 1 + 826 ( 6 β ) α 6 β 1 + 44 ( 7 β ) α 7 β 1 + ( 8 β ) α 8 β 1
L β = 663696 α β ln α + 509004 ( 2 ) α 2 β ln α + 214676 ( 3 ) α 3 β ln α + 54649 ( 4 ) α 4 β ln α + 8624 ( 5 ) α 5 β ln α + 826 ( 6 ) α 6 β ln α + 44 ( 7 ) α 7 β ln α + ( 8 ) α 8 β ln α
M 1 , 0 , 2 α = K α ( L ) L α ( K ) L 2 & M 1 , 0 , 2 β = K β ( L ) L β ( K ) L 2
M 1 , 2 , 0 = 150 α β + 6 α 2 β + 1008 1650 α β + 335 α 2 β + 30 α 3 β + α 4 β + 3024 = H G
H α = 150 β α β 1 + 6 ( 2 β ) α 2 β 1 & H β = 150 α β ln α + 6 ( 2 ) α 2 β ln α
G α = 1650 β α β 1 + 335 ( 2 β ) α 2 β 1 + 30 ( 3 β ) α 3 β 1 + ( 4 β ) α 4 β 1
G β = 1650 α β ln α + 335 ( 2 ) α 2 β ln α + 30 ( 3 ) α 3 β ln α + ( 4 ) α 4 β ln α
M 1 , 2 , 0 α = H α ( G ) G α ( H ) G 2 & M 1 , 2 , 0 β = H β ( G ) G β ( H ) G 2

Maximum Likelihood Estimation

Let y 1 , y 2 , . . . , y n an observed random sample from MBUW distribution with parameters α , β . The likelihood function and the log-likelihood function are defined in
L ( α , β ; y ) = 6 α β n i = 1 n 1 y i 1 α β i = 1 n y i 2 α β 1 , 0 < y < 1 , α & β > 0
l ( α , β ; y ) = n ln ( 6 ) n β ln ( α ) + i = 1 n ln 1 y i 1 α β + 2 α β 1 i = 1 n ln ( y i )
The Maximum Likelihood (ML) equations are the first derivative of the log-likelihood function concerning each parameter as defined in
d l ( α , β ; y ) d α = β n α + i = 1 n y i 1 α β ( ln [ y i ] ) ( β α β 1 ) 1 y i 1 α β 2 β α β 1 i = 1 n ln ( y i )
d l ( α , β ; y ) d β = n ln α + α β i = 1 n y i 1 α β ln [ α ] ln [ y i ] 1 y i 1 α β 2 α β [ ln α ] i = 1 n ln ( y i )

Appendix B

The workflow for estimating the variance (after controlling for the variance between the PWMs)
Step 1: Parameter estimation. Using the LM algorithm to fit the parameters. The LM algorithm uses a small regularization or damping factor (0.001) for numerical stability in optimization as described in Section (2.2). This does not introduce bias in the estimated parameters because this ridge is extremely small.
Step 2: Bootstrap PMWs. Using the estimated parameters to generate new variables and to compute the sample unbiased estimators for the PWMs for each bootstrap sample. The author used 5000 bootstrap samples, each sample is equal to the sample size of the observations used in the real data analysis.
Step 3: Tayler series expansion to estimate the variance of the estimated parameters. Adding a large regularization factor to achieve a condition number 10 as described in Section (2.3) only controls the condition number and the matrix inversion. Hence the large ridge at this stage only stabilizes the variance composition computation and does not introduce bias to the parameter estimation itself.

Appendix C

Preprints 189293 i001 The first picture contains 8 figures with the following description: The first four figures are obtained from the replicates of sample size n=20, while the last four figures are from the replicates of sample size, n=50. The figures for each sample size is arranged as the pairs mentioned in the text, the first pair is ( α = 0.5 & β = 0.6 ), the second pair is ( α = 0.5 & β = 1.3 ), the third pair ( α = 1.1 & β = 0.6 ), and the fourth pair ( α = 1.1 & β = 1.6 ). These are the runs obtained from the MLE method.
Preprints 189293 i002 The second picture contains 8 figures with the following description: The first four figures are obtained from the replicates of sample size n=100, while the last four figures are from the replicates of sample size, n=500. The figures for each sample size is arranged as the pairs mentioned in the text, the first pair is ( α = 0.5 & β = 0.6 ), the second pair is ( α = 0.5 & β = 1.3 ), the third pair ( α = 1.1 & β = 0.6 ), and the fourth pair ( α = 1.1 & β = 1.6 ). These are the runs obtained from the MLE method.
Preprints 189293 i003 The third picture contains 5 figures with the following description: The first three figures are obtained from the replicates of sample size n=20, while the last two figures are from the replicates of sample size, n=50. The figures for each sample size is arranged as the pairs mentioned in the text, the first pair is ( α = 0.5 & β = 0.6 ), the second pair is ( α = 0.5 & β = 1.3 ), the third pair ( α = 1.1 & β = 0.6 ), and the fourth pair ( α = 1.1 & β = 1.6 ). These are the runs obtained from the MOM method.
Preprints 189293 i004 The fourth picture contains 8 figures with the following description: The first four figures are obtained from the replicates of sample size n=100, while the last four figures are from the replicates of sample size, n=500. The figures for each sample size is arranged as the pairs mentioned in the text, the first pair is ( α = 0.5 & β = 0.6 ), the second pair is ( α = 0.5 & β = 1.3 ), the third pair ( α = 1.1 & β = 0.6 ), and the fourth pair ( α = 1.1 & β = 1.6 ). These are the runs obtained from the MOM method.

References

  1. Attia, Iman M. Median-Based Unit Weibull (MBUW): A New Variants of Generalized Method of Moments and Percentile Estimators. Journal of Probability and Statistics 2025, vol. 2025(no. 1), 1503091. [Google Scholar] [CrossRef]
  2. Greenwood, J. A.; Landwehr, J. M.; Matalas, N. C.; Wallis, J. R. Probability weighted moments: Definition and relation to parameters of several distributions expressable in inverse form. Water Resources Research 1979, vol. 15(no. 5), 1049–1054. [Google Scholar] [CrossRef]
  3. Hosking, J. R. M.; Wallis, J. R.; Wood, E. F. Estimation of the Generalized Extreme-Value Distribution by the Method of Probability-Weighted Moments. Technometrics 1985, vol. 27(no. 3), 251–261. [Google Scholar] [CrossRef]
  4. Hosking, J. R. M.; Wallis, J. R. Parameter and Quantile Estimation for the Generalized Pareto Distribution. Technometrics 1987, vol. 29(no. 3), 251–261. [Google Scholar] [CrossRef]
  5. Hooda, Ekta; Hooda, B K; Tanwar, Nitin. Probability weighted moments (PWMs) and partial probability weighted moments (PPWMs) of type-II extreme value distribution. Conference: National Conference on Mathematics and Its applications in Science and technology, Department of Mathematics, GJUS&T, Hisar, Haryana, India, Oct. 2018. [Google Scholar]
  6. Ashkar, F.; Mahdi, S. Comparison of two fitting methods for the log-logistic distribution. Water Resources Research 2003, vol. 39(no. 8), 1–8. [Google Scholar] [CrossRef]
  7. Caeiro, F.; Mateus, A. A New Class of Generalized Probability-Weighted Moment Estimators for the Pareto Distribution. Mathematics 2023, vol. 11(no. 5), 1076. [Google Scholar] [CrossRef]
  8. Wang, Q. J. Estimation of the GEV distribution from censored samples by method of partial probability weighted moments. Journal of Hydrology 1990, vol. 120(no. 1–4), 103–114. [Google Scholar] [CrossRef]
  9. Caeiro, F.; Mateus, A. A Log Probability Weighted Moments Method for Pareto distribution. In Proceedings of the 17th Applied Stochastic Models and Data Analysis International Conference with 6th Demographics Workshop; London, UK, Skiadas, C.H., Ed.; 6-9 June 2017; pp. 211–218. [Google Scholar]
  10. F. Caeiro and D. Prata Gomes, A Log Probability Weighted Moment Estimator of Extreme Quantiles, in Theory and Practice of Risk Assessment, vol. 136, C. P. Kitsos, T. A. Oliveira, A. Rigas, and S. Gulati, Eds., Springer Proceedings in Mathematics & Statistics, vol. 136. Cham: Springer International Publishing, 2015, pp. 293–303. Springer International Publishing: Cham. [CrossRef]
  11. Caeiro, F.; Gomes, M. I.; Vandewalle, B. Semi-Parametric Probability-Weighted Moments Estimation Revisited. Methodology and Computing in Applied Probability 2014, vol. 16(no. 1), 1–29. [Google Scholar] [CrossRef]
  12. Caeiro, F.; Gomes, M. I. Semi-parametric tail inference through probability-weighted moments. Journal of Statistical Planning and Inference 2011, vol. 141(no. 2), 937–950. [Google Scholar] [CrossRef]
  13. Caeiro, F.; Gomes, M. I. A Class of Semi-parametric Probability Weighted Moment Estimators. In Recent Developments in Modeling and Applications in Statistics; Oliveira, P. E., Da Graça Temido, M., Henriques, C., Vichi, M., Eds.; Springer Berlin Heidelberg: Berlin, Heidelberg, 2013; pp. 139–147. [Google Scholar] [CrossRef]
  14. Munir, Rizwan; Saleem, Muhammad; Aslam, Muhammad; Ali, Sajid. Comparison of different methods of parameters estimation for Pareto Model. Caspian Journal of Applied Sciences Research 2013, vol. 2(no. 1), 45–56. [Google Scholar]
  15. Chen, H.; Cheng, W.; Zhao, J.; Zhao, X. Parameter estimation for generalized Pareto distribution by generalized probability weighted moment-equations. Communications in Statistics - Simulation and Computation 2017, vol. 46(no. 10), 7761–7776. [Google Scholar] [CrossRef]
  16. Vogel, R. M.; McMahon, T. A.; Chiew, F. H. S. Floodflow frequency model selection in Australia. Journal of Hydrology 1993, vol. 146, 421–449. [Google Scholar] [CrossRef]
  17. Hosking, J.R.M. L-moment: analysis and estimation of distributions using linear combinations of order statistics. Journal of the Royal Statistical Society 1990, vol. 52(no. 1), 105–124. [Google Scholar] [CrossRef]
  18. Hosking, J.R.M. The theory of probability weighted moments, Research Report, RC12210, IBM Thomas J. Watson Research Center, New York, 1986.
  19. Rasmussen, P. F. Generalized probability weighted moments: Application to the generalized Pareto Distribution. Water Resources Research 2001, vol. 37(no. 6), 1745–1751. [Google Scholar] [CrossRef]
  20. Jing, D.; Dedun, S.; Ronfu, Y.; Yu, H. Expressions relating probability weighted moments to parameters of several distributions inexpressible in inverse form. Journal of Hydrology 1989, vol. 110, 259–270. [Google Scholar] [CrossRef]
  21. Landwehr, J. M.; Matalas, N. C.; Wallis, J. R. Probability weighted moments compared with some traditional techniques in estimating Gumbel Parameters and quantiles. Water Resources Research 1979, vol. 15(no. 5), 1055–1064. [Google Scholar] [CrossRef]
  22. Landwehr, J. M.; Matalas, N. C.; Wallis, J. R. Estimation of parameters and quantiles of Wakeby Distributions: 1. Known lower bounds. Water Resources Research 1979, vol. 15(no. 6), 1361–1372. [Google Scholar] [CrossRef]
  23. Chernoff, H.; Gastwirth, J. L.; Johns, M. V. Asymptotic distribution of linear combinations of functions of order statistics with applications to estimation. Annals of Mathematical Statistics 1967, vol. 38, 52–72. [Google Scholar] [CrossRef]
  24. Rao, C.R. Linear Statistical Inference and Its Applications, 2nd ed.; John Wiley: New York, NY, 1973. [Google Scholar]
Figure 4. shows the box plot of the indicators. The variables are negatively skewed
Figure 4. shows the box plot of the indicators. The variables are negatively skewed
Preprints 189293 g004
Figure 5. shows the empirical CDF vs. fitted CDFs for the different competitors of the water quality dataset
Figure 5. shows the empirical CDF vs. fitted CDFs for the different competitors of the water quality dataset
Preprints 189293 g005
Figure 6. shows the histogram and the fitted PDFs for the different competitors of the water quality dataset
Figure 6. shows the histogram and the fitted PDFs for the different competitors of the water quality dataset
Preprints 189293 g006
Table 1. results of simulation study using sample size 20 with different values of α & β
Table 1. results of simulation study using sample size 20 with different values of α & β
Statistical indices MLE MOM PWM M101,M110 PWM M102,M120
n = 20 α = 0 . 5 β = 0 . 6 Mean( α ) 0.5055,(0.0842) 0.4943,(0.1113) 0.4975,(0.1007) 0.4978,(0.0985)
Mean( β ) 0.5963,(0.1077) 0.6298,(0.0812) 0.6014,(0.0581) 0.6013,(0.0569)
AAB( α ) 0.0647 0.0878 0.0798 0.0778
AAB( β ) 0.0794 0.0648 0.0461 0.0454
MSE( α ) 0.0071 0.0124 0.0101 0.0096
MSE( β ) 0.0116 0.0075 0.0034 0.0032
MRE( α ) 0.1293 0.1756 0.1596 0.1559
MRE( β ) 0.1324 0.1079 0.0768 0.0751
Quantile( α ) (0.3653,0.7004) (0.2838,0.7266) (0.3008,0.7042) (0.2961,0.6909)
Quantile( β ) (0.3366,0.7596) (0.4767,0.8019) (0.4821,0.7151) (0.4897,0.7177)
Number of valid samples 1000 of 1000 987 out of 1000 999 out of 1000 999 out of 1000
n = 20
α = 0 . 5
β = 1 . 3
Mean( α ) 0.5000 (0.0372) 0.4441 (0.0648) 0.4955 (0.0580) 0.4954 (0.0578)
Mean( β ) 1.3210,(0.0963) 1.1365 (0.0724) 1.3012 (0.0155) 1.3102 (0.0154)
AAB( α ) 0.02297 0.0706 0.0467 0.0468
AAB( β ) 0.0800 0.1637 0.0124 0.0125
MSE( α ) 0.0014 0.0073 0.0034 0.0034
MSE( β ) 0.0097 0.0320 2.4062e-4 2.3897e-4
MRE( α ) 0.0593 0.1412 0.0934 0.0934
MRE( β ) 0.0615 0.1259 0.0096 0.0096
Quantile( α ) (0.4281,0.5786) (0.2679,0.6069) (0.3815,0.6117) (0.3837,0.6118)
Quantile( β ) (1.1111,1.5046) (1.0330,1.3073) (1.2702,1.3316) (1.2702,1.3310)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000 1000 of 1000
n = 20
α = 1 . 1
β = 0 . 6
Mean( α ) 1.1056 (0.2651) 1.1772 (0.3668) 1.0920 (0.2883) 1.0942 (0.2586)
Mean( β ) 0.6509 (0.0407) 0.4717 (0.2795) 0.5986 (0.0504) 0.5990 (0.0542)
AAB( α ) 0.2153 0.2933 0.2287 0.2052
AAB( β ) 0.0518 0.2534 0.0400 0.0359
MSE( α ) 0.0702 0.1371 0.0831 0.0668
MSE( β ) 0.0043 0.0926 0.0025 0.0020
MRE( α ) 0.1958 0.2666 0.2079 0.1865
MRE( β ) 0.0864 0.4224 0.0666 0.0598
Quantile( α ) (0.6824,1.7008) (0.5143,1.8809) (0.5476,1.6682) (0.6037,1.5896)
Quantile( β ) (0.5945,0.7521) (0.0074,0.8971) (0.5035,0.6993) (0.5133,0.6856)
Number of valid samples 1000 of 1000 40 out of 1000 1000 of 1000 1000 of 1000
n = 20
α = 1 . 1
β = 1 . 6
Mean( α ) 1.0928 (0.1050) 1.0979 (0.0172) 1.0983 (0.1139)
Mean( β ) 1.6595 (0.0546) 1.5999 (0.0011) 1.5999 (0.0075)
AAB( α ) 0.0860 0.0959 0.0927
AAB( β ) 0.0646 0.0063 0.0061
MSE( α ) 0.0111 0.0138 0.0130
MSE( β ) 0.0065 5.9109e-5 5.5617e-5
MRE( α ) 0.0782 0.0128 0.0843
MRE( β ) 0.0403 5.7592e-4 0.0038
Quantile( α ) (0.8940,1.2883) (0.8661,1.13024) (0.8700,1.3045)
Quantile( β ) (1.5711,1.7822) (1.5847,1.6133) (1.5849,1.6134)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000
Table 2. results of simulation study using sample size of 50 with different values of α & β
Table 2. results of simulation study using sample size of 50 with different values of α & β
Statistical indices MLE MOM PWM M101,M110 PWM M102,M120
n = 50 α = 0 . 5 β = 0 . 6 Mean( α ) 0.5052 (0.0533) 0.4988 (0.074) 0.5002 (0.0666) 0.5001 (0.0645)
Mean( β ) 0.6012 (0.0652) 0.6135 (0.0417) 0.5999 (0.0385) 0.5999 (0.0373)
AAB( α ) 0.04427 0.0585 0.0540 0.0522
AAB( β ) 0.0520 0.0348 0.0312 0.0301
MSE( α ) 0.0029 0.0055 0.0044 0.0042
MSE( β ) 0.0042 0.0019 0.0015 0.0014
MRE( α ) 0.0854 0.1170 0.1080 0.1043
MRE( β ) 0.0866 0.0579 0.0520 0.0502
Quantile( α ) (0.4097,0.6114) (0.3495,0.6400) (0.3693,0.6242) (0.3718,0.6231)
Quantile( β ) (0.4535,0.7133) (0.5302,0.6930) (0.5283,0.6755) (0.5289,0.6740)
Number of valid samples 999 out of 1000 999 out of 1000 1000 of 1000 999 out of 1000
n = 50 α = 0 . 5 β = 1 . 3 Mean( α ) 0.5049 (0.0237) 0.4699 (0.041) 0.5015 (0.0374) 0.5016 (0.0375)
Mean( β ) 1.3136 (0.0652) 1.1947 (0.0546) 1.2996 (0.0100) 1.2996 (0.0100)
AAB( α ) 0.0192 0.0415 0.0301 0.0302
AAB( β ) 0.0537 0.1057 0.0080 0.0081
MSE( α ) 5.8625e-4 0.0026 0.0014 0.0014
MSE( β ) 0.0044 0.0141 9.9379e-5 1.0005e-4
MRE( α ) 0.0384 0.0830 0.0602 0.0605
MRE( β ) 0.0413 0.0813 0.0062 0.0062
Quantile( α ) (0.4540,0.5486) (0.3916,0.5492) (0.4313,0.5756) (0.4286,0.5769)
Quantile( β ) (1.1705,1.4177) (1.1100,1.3013) (1.2798,1.3183) (1.2795,1.3190)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000 1000 of 1000
n = 50 α = 1 . 1 β = 0 . 6 Mean( α ) 1.1094 (0.1773) 1.1040 (0.1804) 1.1034 (0.1610)
Mean( β ) 0.6336 (0.0282) 0.6007 (0.0315) 0.6006 (0.0281)
AAB( α ) 0.1439 0.1444 0.1292
AAB( β ) 0.0351 0.0252 0.0226
MSE( α ) 0.0315 0.0325 0.0259
MSE( β ) 0.0019 9.9281e-4 7.9139e-4
MRE( α ) 0.1308 0.1312 0.1175
MRE( β ) 0.0584 0.0420 0.0376
Quantile( α ) (0.8060,1.4720) (0.7622,1.4710) (0.7980,1.4223)
Quantile( β ) (0.5924,0.6997) (0.5410,0.6648) (0.5472,0.6563)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000
n = 50 α = 1 . 1 β = 1 . 6 Mean( α ) 1.0989 (0.0687) 1.1014 (0.0747) 1.1012 (0.0726)
Mean( β ) 1.6399 (0.0394) 1.6001 (0.0049) 1.6001 (0.0048)
AAB( α ) 0.0557 0.0597 0.0581
AAB( β ) 0.0458 0.0039 0.0038
MSE( α ) 0.0047 0.0056 0.0053
MSE( β ) 0.0031 2.3924e-5 2.2608e-5
MRE( α ) 0.0506 0.0543 0.0528
MRE( β ) 0.0286 0.0024 0.0024
Quantile( α ) (0.9697,1.2345) (0.9606,1.2525) (0.9610,1.2454)
Quantile( β ) (1.5658,1.7226) (1.5909,1.6100) (1.5909,1.6095)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000
Table 3. results of simulation study using sample size of 100 with different values of α & β
Table 3. results of simulation study using sample size of 100 with different values of α & β
Statistical indices MLE MOM PWM M101,M110 PWM M102,M120
n = 100 α = 0 . 5 β = 0 . 6 Mean( α ) 0.5043 (0.0353) 0.5027 (0.0484) 0.5005 (0.0449) 0.5001 (0.0434)
Mean( β ) 0.6042 (0.0433) 0.6106 (0.0274) 0.5997 (0.0259) 0.5999 (0.0251)
AAB( α ) 0.0282 0.0383 0.0359 0.0347
AAB( β ) 0.0349 0.0232 0.0207 0.0200
MSE( α ) 0.0013 0.0023 0.0020 0.0019
MSE( β ) 0.0019 8.6160e-4 6.7061e-4 6.2760e-4
MRE( α ) 0.0564 0.0767 0.0718 0.0694
MRE( β ) 0.0581 0.0387 0.0345 0.0334
Quantile( α ) (0.4332,0.5752) (0.4059,0.5999) (0.4118,0.5884) (0.4110,0.5851)
Quantile( β ) (0.5107,0.6776) (0.5577,0.6669) (0.5490,0.6509) (0.5509,0.6514)
Number of valid samples 1000 of 1000 1000 of 1000 1000 out of 1000 1000 of 1000
n = 100 α = 0 . 5 β = 1 . 3 Mean( α ) 0.5048 (0.0164) 0.4845 (0.0277) 0.5000 (0.0256) 0.5002 (0.0253)
Mean( β ) 1.3183 (0.0441) 1.247 (0.0330) 1.3000 (0.0068) 1.2999 (0.0067)
AAB( α ) 0.0137 0.0254 0.0207 0.0204
AAB( β ) 0.0379 0.0534 0.0055 0.0054
MSE( α ) 2.9281e-4 0.001 6.5674e-4 6.4032e-4
MSE( β ) 0.0023 0.0039 4.6676e-5 4.5509e-5
MRE( α ) 0.0273 0.0508 0.0414 0.0407
MRE( β ) 0.0292 0.0411 0.0042 0.0042
Quantile( α ) (0.4731,0.5354) (0.4322,0.5397) (0.4503,0.5508) (0.4507,0.5494)
Quantile( β ) (1.2219,1.3884) (1.1916,1.3021) (1.2865,1.3132) (1.2868,1.3131)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000 1000 of 1000
n = 100 α = 1 . 1 β = 0 . 6 Mean( α ) 1.1025 (0.122) 1.2234 (0.4310) 1.1006 (0.1254) 1.1014 (0.1114)
Mean( β ) 0.6232 (0.0216) 0.2910 (0.2066) 0.6001 (0.0219) 0.6002 (0.0195)
AAB( α ) 0.0974 0.3368 0.0999 0.0888
AAB( β ) 0.0248 0.3267 0.0175 0.0155
MSE( α ) 0.0149 0.2005 0.0157 0.0124
MSE( β ) 0.0010 0.1381 4.7959e-4 3.7833e-4
MRE( α ) 0.0885 0.3062 0.0908 0.0807
MRE( β ) 0.0413 0.5445 0.0291 0.0259
Quantile( α ) (0.8780,1.3561) (0.2606,2.1869) (0.8532,1.3501) (0.8829,1.3176)
Quantile( β ) (0.5907,0.6774) (0.0076,0.7279) (0.5569,0.6437) (0.5621,0.6380)
Number of valid samples 1000 of 1000 409 out of 1000 1000 of 1000 1000 of 1000
n = 100 α = 1 . 1 β = 1 . 6 Mean( α ) 1.0989 (0.0474) 1.3354 (0.1711) 1.1010 (0.0505) 1.1008 (0.0492)
Mean( β ) 1.6365 (0.0318) 0.6937 (0.3946) 1.6001 (0.0033) 1.6001 (0.0032)
AAB( α ) 0.0379 0.2446 0.0409 0.0398
AAB( β ) 0.0405 0.9064 0.0027 0.0026
MSE( α ) 0.0022 0.0847 0.0026 0.0024
MSE( β ) 0.0023 0.9770 1.0953e-5 1.0394e-5
MRE( α ) 0.0345 0.2224 0.0372 0.0362
MRE( β ) 0.0253 0.5665 0.0017 0.0016
Quantile( α ) (1.0089,1.1856) (1.0331,1.6236) (1.0036,1.1972) (1.0056,1.1941)
Quantile( β ) (1.5761,1.6994) (0.2126,1.5451) (1.5937,1.6064) (1.5938,1.6062)
Number of valid samples 1000 of 1000 997 out of 1000 1000 of 1000 1000 of 1000
Table 4. results of simulation study using sample size of 500 with different values of α & β
Table 4. results of simulation study using sample size of 500 with different values of α & β
Statistical indices MLE MOM PWM M101,M110 PWM M102,M120
n = 500 α = 0 . 5 β = 0 . 6 Mean( α ) 0.5044 (0.0162) 0.5011 (0.0214) 0.4988 (0.0202) 0.4989 (0.0197)
Mean( β ) 0.6097 (0.0201) 0.6066 (0.0123) 0.6007 (0.0117) 0.6006 (0.0114)
AAB( α ) 0.0131 0.0171 0.0161 0.0156
AAB( β ) 0.0177 0.0110 0.0093 0.0090
MSE( α ) 2.8252e-4 4.5718e-4 4.1113e-4 3.8840e-4
MSE( β ) 4.9920e-4 1.9447e-4 1.3717e-4 1.2959e-4
MRE( α ) 0.0261 0.0341 0.0322 0.0312
MRE( β ) 0.0294 0.0184 0.0155 0.0150
Quantile( α ) (0.4731,0.5360) (0.4590,0.5431) (0.4581,0.5388) (0.4578,0.5380)
Quantile( β ) (0.5659,0.6457) (0.5819,0.6303) (0.5776,0.6242) (0.5781,0.6244)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000 1000 of 1000
n = 500 α = 0 . 5 β = 1 . 3 Mean( α ) 0.5051 (0.0073) 0.4971 (0.0118) 0.5000 (0.0113) 0.4999 (0.0113)
Mean( β ) 1.3195 (0.0242) 1.2894 (0.0068) 1.3000 (0.003) 1.3000 (0.0030)
AAB( α ) 0.0068 0.0098 0.0092 0.0092
AAB( β ) 0.0237 0.0108 0.0024 0.0024
MSE( α ) 7.9362e-5 1.4693e-4 1.2707e-4 1.2682e-4
MSE( β ) 9.6674e-4 1.5855e-4 9.0313e-6 9.0134e-6
MRE( α ) 0.0137 0.0197 0.0183 0.0183
MRE( β ) 0.0182 0.0083 0.0019 0.0019
Quantile( α ) (0.4932,0.5216) (0.4753,0.5205) (0.4795,0.5223) (0.4793,0.5219)
Quantile( β ) (1.2824,1.3668) (1.2766,1.3018) (1.2940,1.3055) (1.2942,1.3055)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000 1000 of 1000
n = 500 α = 1 . 1 β = 0 . 6 Mean( α ) 1.0978 (0.0556) 1.2494 (0.0868) 1.1001 (0.0548) 1.1000 (0.0488)
Mean( β ) 0.6137 (0.0126) 0.2754 (0.1023) 0.6000 (0.0096) 0.6000 (0.0085)
AAB( α ) 0.0453 0.1532 0.0444 0.0396
AAB( β ) 0.0155 0.3247 0.0077 0.0069
MSE( α ) 0.0031 0.0299 0.0030 0.0024
MSE( β ) 3.4713e-4 0.1158 9.1734e-5 7.2737e-5
MRE( α ) 0.0412 0.1393 0.0403 0.0360
MRE( β ) 0.0258 0.5412 0.0129 0.0115
Quantile( α ) (0.9962,1.2096) (1.075,1.4236) (0.9952,1.2103) (1.0074,1.1972)
Quantile( β ) (0.5904,0.6388) (0.0953,0.5332) (0.5817,0.6193) (0.5838,0.6170)
Number of valid samples 1000 of 1000 675 out of 1000 1000 of 1000 1000 of 1000
n = 500 α = 1 . 1 β = 1 . 6 Mean( α ) 1.0976 (0.0213) 1.1300 (0.0454) 1.1001 (0.0227) 1.1000 (0.0220)
Mean( β ) 1.6352 (0.0279) 1.3005 (0.2355) 1.6000 (0.0015) 1.6000 (0.0014)
AAB( α ) 0.0171 0.0388 0.0183 0.0178
AAB( β ) 0.0384 0.2995 0.0012 0.0012
MSE( α ) 4.5729e-4 0.0030 5.1549e-4 4.8480e-4
MSE( β ) 0.0020 0.1451 2.2133e-6 2.0816e-6
MRE( α ) 0.0155 0.0353 0.0167 0.0162
MRE( β ) 0.0240 0.1872 7.5140e-4 7.3023e-4
Quantile( α ) (1.0584,1.1414) (1.0631,1.248) (1.0564,1.1454) (1.0584,1.1441)
Quantile( β ) (1.5789,1.6842) (0.6479,1.600) (1.5971,1.603) (1.5973,1.6029)
Number of valid samples 1000 of 1000 1000 of 1000 1000 of 1000 1000 of 1000
Table 5. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) using PWMs M 1 , 0 , 1 , M 1 , 1 , 0 .
Table 5. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) using PWMs M 1 , 0 , 1 , M 1 , 1 , 0 .
n
PWM: M 1 , 0 , 1 , M 1 , 1 , 0
α = 0.5 , β = 0.6

PWM: M 1 , 0 , 1 , M 1 , 1 , 0
α = 0.5 , β = 1.3

PWM: M 1 , 0 , 1 , M 1 , 1 , 0
α = 1.1 , β = 0.6

PWM: M 1 , 0 , 1 , M 1 , 1 , 0
α = 1.1 , β = 1.6
500 Var_cov 4.1004e-4 -2.368e-4 1.272e-4 -3.391e-5 0.003 5.255e-4 5.1600e-4 3.3811e-5
-2.368e-4 1.3681e-4 -3.391e-5 9.0404e-6 5.255e-4 9.1825e-5 3.3811e-5 2.2155e-6
Corr -1 -1 1 1
Cond num. Eigen_val Cond. num=1.395e+16
Eigen_val=-1.3553e-20, 5.4685e-4
Cond_num=9.0086e+16
Eigen_val=1.6941e-21, 1.3624e-4
Cond_num=2.2935e+17
Eigen_val=0, 0.0031
Cond_num=6.6920e+17
Eigen_val=0, 5.1822e-4
100 Var_cov 0.0020 -0.0012 6.574e-4 -1.7526e-4 0.0157 0.0027 0.0026 1.6725e-4
-0.0012 6.7119e-4 -1.7526e-4 4.6723e-5 0.0027 4.8006e-4 1.6725e-4 1.0959e-5
Corr -1 -1 1 1
Cond num. Eigen_val Cond. num=1.632e+16
Eigen_val=1.0842e-19, 0.0027
Cond. num=5.8349e+16
Eigen_val=0, 7.0412e-4
Cond. num=1.2587e+17
Eigen_val=0.0162, 2.1684e-19
Cond. num=2.0801e+18
Eigen_val=0.0026, 3.3881e-21
50 Var_cov 0.0044 -0.0026 0.0014 -3.7257e-4 0.0325 0.0057 0.0056 3.6534e-4
-0.0026 0.0015 -3.7257e-4 9.9325e-5 0.0057 9.9332e-4 3.6534e-4 2.3939e-5
Corr -1 -1 1 1
Cond num. Eigen_val Cond. num=2.548e+16
Eigen_val=0, 0.0059
Cond. num=5.654e+16
Eigen_val=0.0015, 1.3551e-20
Cond. num=5.2012e+16
Eigen_val=0.0335, 4.3368e-19
Cond. num=6.3331e+17
Eigen_val=0.0056, -6.7763e-21
20 Var_cov 0.0101 -0.0059 0.0034 -8.9798e-4 0.0831 0.0145 0.0138 9.0267e-4
-0.0059 0.0034 -8.9798e-4 2.3940e-4 0.0145 0.0025 9.0267e-4 5.9148e-5
Corr -1 -1 1 1
Cond num. Eigen_val Cond.num=2.7699e+16
Eigen.val=0.0135, 4.3368e-19
Cond.num=1.0764e+17
Eigen.val=0.0036, 5.4210e-20
Cond.num=9.2362e+16
Eigen.val=0.0856, 4.3368e-20
Cond.num=1.328e+18
Eigen.val=0.0138, -1.3553e-20
Table 6. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) using PWMs M 1 , 0 , 2 , M 1 , 2 , 0 .
Table 6. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) using PWMs M 1 , 0 , 2 , M 1 , 2 , 0 .
n
PWM: M 1 , 0 , 2 , M 1 , 2 , 0
α = 0.5 , β = 0.6

PWM: M 1 , 0 , 2 , M 1 , 2 , 0
α = 0.5 , β = 1.3

PWM: M 1 , 0 , 2 , M 1 , 2 , 0
α = 1.1 , β = 0.6

PWM: M 1 , 0 , 2 , M 1 , 2 , 0
α = 1.1 , β = 1.6
500 Var_cov 3.8765e-4 -2.2390e-4 1.2694e-4 -3.3840e-5 0.0024 4.1669e-4 4.8529e-4 3.1799e-5
-2.2390e-4 1.2934e-4 -3.3840e-5 9.0221e-6 4.1669e-4 7.2810e-5 3.1799e-5 2.0836e-6
Corr -1 -1 1 1
Cond num. Eigen_val Cond.num=4.1394e+17
Eigen_val=-4.0658e-20, 5.1700e-4
Cond.num=2.7053e+17
Eigen_val=3.3881e-21, 1.3596e-4
Cond.num=1.3574e+17
Eigen_val=1.3553e-20, 0.0025
Cond.num=7.5123e+17
Eigen_val=1.2705e-21, 4.8727e-4
100 Var_cov 0.0019 -0.0011 6.4091e-4 -1.7086e-4 0.0124 0.0022 0.0024 1.5875e-4
-0.0011 6.2822e-4 -1.7086e-4 4.5552e-5 0.0022 3.7865e-4 1.5875e-4 1.0402e-5
Corr -1 -1 1 1
Cond num. Eigen_val Cond.num=1.4424e+16
Eigen_val=1.0842e-19, 0.0025
Cond.num=1.0285e+17
Eigen_val=-1.3553e-20, 6.8647e-4
Cond.num=1.2979e+17
Eigen_val=-2.1684e-19, 0.0128
Cond.num=8.5287e+17
Eigen_val=-5.0822e-21, 0.0024
50 Var_cov 0.0042 -0.0024 0.0014 -3.7490e-4 0.0259 0.0045 0.0053 3.4528e-4
-0.0024 0.0014 -3.7490e-4 9.9962e-5 0.0045 7.9183e-4 3.4528e-4 2.2625e-5
Corr -1 -1 1 1
Cond num. Eigen_val Cond.num=1.8731e+16
Eigen_val=0, 0.0055
Cond.num=4.3350e+16
Eigen_val=-4.0658e-20, 0.0015
Cond.num=2.0248e+17
Eigen_val=1.0842e-19, 0.0267
Cond.num=4.8561e+17
Eigen_val=1.3553e-20, 0.0053
20 Var_cov 0.0096 -0.0056 0.0033 -8.9170e-4 0.0668 0.0117 0.0130 8.4944e-4
-0.0056 0.0032 -8.9170e-4 2.3772e-4 0.0117 0.0020 8.4944e-4 5.5660e-5
Corr -1 -1 1 1
Cond num. Eigen_val Cond.num=2.0154e+16
Eigen_val=8.6736e-19, 0.0128
Cond.num=3.6958e+17
Eigen_val=0.0036, 6.8647e-4
Cond.num=1.1993e+17
Eigen_val=8.6736e-19, 0.0689
Cond.num=6.3129e+17
Eigen_val=-2.7105e-20, 0.0130
Table 7. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) of MLE.
Table 7. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) of MLE.
n
MLE
α = 0.5 , β = 0.6

MLE
α = 0.5 , β = 1.3

MLE
α = 1.1 , β = 0.6

MLE
α = 1.1 , β = 1.6
500 Var_cov 2.6310e-4 -2.9000e-4 5.3680e-5 -9.9890e-5 0.0031 1.6830e-4 4.5220e-4 -7.0280e-5
-2.9000e-4 4.0480e-4 -9.9890e-5 5.8690e-4 1.6830e-4 1.5960e-4 -7.0280e-5 7.7790e-4
Corr -0.8887(p=0) -0.5628(p=1.26e-84) 0.2395(p=1.644e-14) -0.1185(p=1.729e-4)
Cond num. Eigen_val Cond.num=17.8648
Eigen_val=3.5402e-05, 6.3245e-4
Cond.num=17.0047
Eigen_val=3.5578e-05, 6.0499e-4
Cond.num=20.6931
Eigen_val=1.5003e-04, 0.0031
Cond.num=1.8107
Eigen_val=4.3764e-04, 7.9246e-4
100 Var_cov 0.0012 -0.0015 2.6990e-4 -6.4590e-4 0.0149 -6.4070e-4 0.0022 1.6973e-4
-0.0015 0.0019 -6.4590e-4 0.0019 -6.4070e-4 4.6760e-4 1.6973e-4 0.0010
Corr -0.9670(p=0) -0.8917(p=0) -0.2429(p=6.8e-15) 0.1128(p=3.7e-4)
Cond num. Eigen_val Cond.num=62.3364
Eigen_val=4.9263e-05, 0.0031
Cond.num=43.5652
Eigen_val=4.9688e-05, 0.0022
Cond.num=33.9381
Eigen_val=4.3920e-04, 0.0149
Cond.num=2.3023
Eigen_val=9.8549e-04, 0.0023
50 Var_cov 0.0028 -0.0034 5.6260e-4 -0.0014 0.0315 -0.0013 0.0047 -2.8410e-4
-0.0034 0.0042 -0.0014 0.0043 -0.0013 7.9280e-4 -2.8410e-4 0.0016
Corr -0.9823(p=0) -0.9326(p=0) -0.2514(p=7e-16) 0.1050(p=0.0009)
Cond num. Eigen_val Cond.num=116.3342
Eigen_val=6.0377e-05, 0.0070
Cond.num=72.3631
Eigen_val=6.5640e-05, 0.0047
Cond.num=42.4821
Eigen_val=7.4153e-04, 0.0315
Cond.num=3.1071
Eigen_val=0.0015, 0.0047
20 Var_cov 0.0071 -0.0089 0.0014 -0.0035 0.0703 -0.0044 0.011 -0.0013
-0.0089 0.0116 -0.0035 0.0093 -0.0044 0.0017 -0.0013 0.0030
Corr -0.9788(p=0) -0.9683(p=0) -0.4087(p=0) -0.2297(p=0)
Cond num. Eigen_val Cond.num=99.0764
Eigen_val=0.0185, 1.8665e-4
Cond.num=139.9554
Eigen_val=0.0106, 7.5630e-5
Cond.num=51.2278
Eigen_val=0.0014, 0.0705
Cond.num=4.0477
Eigen_val=0.0028, 0.0112
Table 8. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) of MOMs.
Table 8. Analysis of the sampling distribution of the estimated ( α ^ ) & ( β ^ ) of MOMs.
n
MOM
α = 0.5 , β = 0.6

MOM
α = 0.5 , β = 1.3

MOM
α = 1.1 , β = 0.6

MOM
α = 1.1 , β = 1.6
500 Var_cov 4.5650e-4 -2.5350e-4 1.3850e-5 -2.4080e-5 0.0075 0.0038 0.0021 -0.0071
-2.5350e-4 1.5150e-4 -2.4080e-5 4.6150e-5 0.0038 0.0105 -0.0071 0.0555
Corr -0.964(p=0) -0.3011(p=2.09e-22) 0.4323(p=4.0974e-32) -0.662(p=4.0381e-127)
Cond num. Eigen_val Cond.num=73.6573
Eigen_val=8.1441e-06, 5.9988e-4
Cond.num=3.5871
Eigen_val=4.0254e-5, 1.4440e-4
Cond.num=2.6805
Eigen_val=0.0049, 0.0131
Cond.num=49.5788
Eigen_val=0.0011, 0.0564
100 Var_cov 0.0023 -0.0013 7.6600e-4 4.6600e-5 0.1857 0.0101 0.0293 -0.0396
-0.0013 7.5020e-4 4.6600e-5 0.0011 0.0101 0.0427 -0.0396 0.1557
Corr -0.9846(p=0) 0.0511(p=0.1065) 0.1138(p=0.0214) -0.587(p=2.2243e-93)
Cond num. Eigen_val Cond.num=176.1722
Eigen_val=1.7454e-5, 0.0031
Cond.num=1.4435
Eigen_val=7.5939e-4, 0.0011
Cond.num=4.4403
Eigen_val=0.042, 0.1864
Cond.num=9.3439
Eigen_val=0.0179, 0.1671
50 Var_cov 0.0055 -0.0030 0.0017 3.2350e-4
-0.0030 0.0017 3.2350e-4 0.0030
Corr -0.9600(p=0) 0.1444(p=4.525e-6)
Cond num. Eigen_val Cond.num=67.5162
Eigen_val=1.0526e-4, 0.0071
Cond.num=1.8992
Eigen_val=0.0016, 0.0031
20 Var_cov 0.0124 -0.0041 0.0042 5.1310e-4 0.1345 0.0016
-0.0041 0.0066 5.1310e-4 0.0052 0.0016 0.0781
Corr -0.4506(p=1.619e-50) 0.1093(p=-5.333e-4) 0.0153(p=0.9255)
Cond num. Eigen_val Cond.num=3.2243
Eigen_val=0.0045, 0.0145
Cond.num=1.3664
Eigen_val=0.0040, 0.0055
Cond.num=1.724
Eigen_val=0.0781, 0.1246
Table 9. The variance decay of the estimated ( α ^ ) & ( β ^ ) in different simulations
Table 9. The variance decay of the estimated ( α ^ ) & ( β ^ ) in different simulations

PWM: M101,M110
α = 0.5 & β = 0.6

PWM: M101,M110
α = 0.5 & β = 1.3

PWM: M101,M110
α = 1.1 & β = 0.6

PWM: M101,M110
α = 1.1 & β = 1.6
Slope for α 0.9846 1.0206 1.0283 1.0048
Slope for β 0.9882 1.0206 1.0277 0.9933
PWM: M102,M120
α = 0 . 5 & β = 0 . 6
PWM: M102,M120
α = 0 . 5 & β = 1 . 3
PWM: M102,M120
α = 1 . 1 & β = 0 . 6
PWM: M102,M120
α = 1 . 1 & β = 1 . 6
Slope for α 0.9876 1.006 1.0204 0.9932
Slope for β 0.9820 1.0061 1.0244 0.9990
MLE
α = 0 . 5 & β = 0 . 6
MLE
α = 0 . 5 & β = 1 . 3
MLE
α = 1 . 1 & β = 0 . 6
MLE
α = 1 . 1 & β = 1 . 6
Slope for α 0.9429 1.0035 0.9755 0.9830
Slope for β 0.9607 0.7299 0.6679 0.1561
MOM
α = 0 . 5 & β = 0 . 6
MOM
α = 0 . 5 & β = 1 . 3
MOM
α = 1 . 1 & β = 0 . 6
MOM
α = 1 . 1 & β = 1 . 6
Slope for α 1.0047 1.0627 1.9940 1.6376
Slope for β 0.9940 1.9704 0.8719 0.6409
Table 10. The relative efficiency of different methods at sample size 20
Table 10. The relative efficiency of different methods at sample size 20
Preprints 189293 i005
Table 11. The relative efficiency of different methods at sample size 50
Table 11. The relative efficiency of different methods at sample size 50
Preprints 189293 i006
Table 12. The relative efficiency of different methods at sample size 100
Table 12. The relative efficiency of different methods at sample size 100
Preprints 189293 i007
Table 13. The relative efficiency of different methods at sample size 500
Table 13. The relative efficiency of different methods at sample size 500
Preprints 189293 i008
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated