Abstract: The aim of this study is finding classical and Bayesian estimators for the shape parameter of the Kumaraswamy distribution using un-grouped data and also considers relationship between them. We show how the classical estimators can be derived from various choices made within a Bayesian framework. We compare the classical estimators based on their Mean Squared Errors (MSE's). Then, we obtain Bayesian and non-Bayesian estimators of the shape parameter of this distribution under Grouped data. In Bayesian estimation, we consider three types of loss functions; the Squared error, Precautionary and General entropy loss functions which are symmetric and asymmetric, respectively. In all cases, we considered both point and interval estimations. These the point and interval estimations are compared empirically using Monte-Carlo simulation. Bayes approach under Precautionary loss function is best estimator for estimating the parameter of Kumaraswamy distribution and this is true for both un-grouped and grouped data.
INTRODUCTION
In life testing experiments, we observe the failure time of a component to
the nearest hour, day or month. Data for which true values are known only up
to subsets of the sample space are called grouped data (Alodat
and Al-Saleh, 2000; Surles and Padgett, 2001; Wu
and Perloff, 2005; Pipper and Ritz, 2007). In general
grouped data can be formulated as follows: Let X1, X2,...,
Xn be a random sample from the density f (x; θ), x ε χ,
θ ε
where, I X (θ) and I Y (θ) are the Fisher's information number obtained from X and Y, respectively and Eθ[I X\Y (θ\Y)] is the conditional score function. If we replace Y by the grouped sample n = (N1 N2,..., Nk+1), then I X (θ)≥In (θ) for all θ, which means that the information in the sample X about θ is reduced to In (θ) because of grouping. Kuldorff (1961) considered non-Bayesian estimation from grouped data when the data come from normal and exponential distributions. Alodat and Al-Saleh (2000) considered the Bayesian estimation from grouped data when the underlying distribution is exponential. Alodat et al. (2007) obtained Bayesian prediction intervals from grouped data when the underlying distribution is exponential. Aludaat et al. (2008) obtained the Bayesian and non-Bayesian estimation from grouped data when the underlying distribution is Burr type X.
The Kumaraswamy distributions were constructed by Kumaraswamy (1980). Jones (2009) said about its properties. The probability density function of a Kumaraswamy distributed random variable is given by:
(1) |
where, θ and λ are shape parameters, respectively. Here we assume that λ parameter is known. The distribution function is:
(2) |
The reliability and failure rate functions of Kumaraswamy distribution are given, respectively by:
(3) |
And
(4) |
Figure 1 shows the shape of for different values of θ and λ.
Bayesian estimators derived from an improper prior distribution can be use to derive the classical estimators. The technique of deriving the classical estimators from the Bayesian estimator is not new. Rossman et al. (1998) and Elfessi and Reineke (2001) presented some thought provoking insights on the relationship between Bayesian and classical estimation using the continuous uniform and exponential distributions, respectively. We will explore these relationships using the Kumaraswamy distribution from un-grouped data.
CLASSICAL AND BAYESIAN ESTIMATIONS BASED ON THE UN-GROUPED DATA
Here, first we obtain the classical estimators of and compare these estimators based on their Mean Squared Errors (MSE's) and then we get the Bayes estimators of under the symmetric and asymmetric loss function and show how the classical estimators can be derived from various choices made within a Bayesian framework. Also, we present the credible and confidence intervals for θ.
Classical point and interval estimations: Let X1, X2,..., Xn be a random sample from density (Eq. 2). The likelihood function is given by:
(5) |
Fig. 1: | Pdf of Kumaraswamy for different values of θ and λ |
Then the log-likelihood function is:
(6) |
Hence,
Thus the MLE of is:
(7) |
where,
The above estimator obtained by Gupta and Kundu (1999).
Here, we obtain the Uniformly Minimum Variance Unbiased Estimator (UMVUE) of θ. since family of density (2) belongs to an exponential family, therefore, statistic T is a complete sufficient statistic for θ. It is easy to show that statistic T is distributed as gamma distribution with parameters n and 1/θ, with the density:
thus:
Hence, the UMVUE of is:
(8) |
We can find the Minimum Mean Squared Error (Min MSE) estimator in the class of estimators of the form c/T. Therefore:
Where as:
thus:
And:
Then:
(9) |
The derivative of r (c) is:
That thereby c = n-2. Thus, the Min MSE estimator of θ is:
(10) |
From Eq. 7, the MSE of the classical estimators of are calculated as follow:
And
Easily can show that:
Now, we find a 100 (1-τ)% confidence interval for θ with obtain L and U, where. P(L<θ<U) = 1-τ. Let X1, X2,..., Xn be a random sample from Kumaraswamy. Since,T~ (n,1/θ), there by 2θT~χ2 (2n) , thus
or
Therefore a classical 100 (1-τ)% confidence interval for θ is given by:
Bayesian point and interval estimations: Here, we obtain the Bayes estimators
of θ under the improper prior distribution. Consider the improper prior
distribution
Notice that this prior distribution is the kernel of a Gamma distribution when
α>0. However, such a restriction on α is not necessary and decreases
the flexibility of the resulting estimator. Whereas
where:
The posterior distribution of θ is proper when, n+α>0, i.e.:
In this case, the Bayes estimator of θ under Squared error loss function is the posterior mean, i.e:
(11) |
Where,
The above estimator obtained by Kundu and Gupta (2008).
The classical estimators derived in the previous subsection can be obtained
from the above Bayes estimator. To do this, we put the different values of α
and β. For example, If α = 0 and β = 0 then
Now, we obtained the Bayes estimator of θ under an asymmetric loss function. Norstrom (1996) introduced an alternative asymmetric precautionary loss function and also presented a general class of precautionary loss function with quadratic loss function as a special case. These loss functions approach infinitely near the origin to prevent underestimation and thus giving conservative estimators, especially when low failure rates are being estimated. These estimators are very useful when underestimation may lead to serious consequences. A very useful and simple asymmetric precautionary loss function is:
(12) |
The Bayes estimator under this asymmetric loss function is denoted by
(13) |
This special case of the Precautionary loss function (Eq. 10) and the Entropy loss function are the same (Norstrom, 1996).
As said, under the gamma prior distribution, i.e:
(14) |
the posterior density of θ is gamma with the shape and scale parameters as α + n and 1/(β+t), respectively, therefore:
Hence, the Bayes estimator of θ with respect to the precautionary loss function under the gamma prior distribution is as follows:
(15) |
Where:
In many practical situations, it appears to be more realistic to express the
loss in terms of the ratio
(16) |
whose minimum occurs at
(17) |
Provided that Eθ {θ-q} exists and is finite, where, Eθ denoted to the expected value with respect to the posterior function of θ.
The Bayes estimator of θ with respect to the General Entropy loss function under the gamma prior distribution is as follows:
(18) |
where,
The Bayesian analog to the confidence interval is called a credibility interval. In general, the interval (L(x), U(x)) is a 100 (1-τ)% credibility interval for θ if:
Since,
Thereby, 2θ (β+T)∼χ2 (2(n+α)), thus:
Therefore, a 100 (1-τ)% Bayesian credibility interval for θ is (L(x), U(x)) where:
(19) |
And:
(20) |
The classical and Bayesian interval estimators are therefore the same when α = 0 and β = 0.
CLASSICAL AND BAYESIAN ESTIMATIONS BASED ON THE GROUPED DATA
Here, we obtain the MLE and Bayes estimators of θ and also the Fisher's information number when the data given in Groups. Also, we use the Fisher's information number to construct a Asymptotic Confidence interval for θ.
Likelihood function and MLE: Here, first we derive the likelihood density based on the grouped data. Let X1, X2,..., Xn be a random sample from Kumaraswamy. Assume that the sample space of f (x; 0) is partitioned into k+1 equally-spaced intervals as follows. Let Ij = [(j-1)δ, jδ), j = 1,..., k and IK+1 = [kδ,1), δ>0. If Nj denotes the number of Xj s that fall in Ij, j = 1,2,..., k+1, then n = N1 +...+ Nk+1. Let:
For j = 1,...k and
(21) |
where, C is a normalizing constant.
In continue, we find the MLE of θ based on the density (20). To do this, we maximize the log likelihood function:
The first derivative of the log-likelihood is:
(22) |
The MLE for θ is the solution of
(23) |
We use
(24) |
where,
And
Here, the initial solution θ0 should be selected from the MLE of θ based on the un-grouped data.
Fisher's information number and confidence interval: To find the fisher's information number contained in the grouped sample about θ, we find the expectation of the second derivative of the log-likelihood. So:
(25) |
where,
If IG(θ) denotes the Fisher's information number from grouped data, then:
And since, E[Nj] = nPj, thus:
(26) |
Using IG(θ), we can find a large sample (1-α) 100% confidence interval for θ as follows:
(27) |
Simple calculations can show that the Fisher's information number about θ in a random sample X1, X2,..., Xn from (1) is I(θ) = n/θ2.
Bayesian estimation: Here, we obtain the Bayes estimators of θ under the Gamma prior distribution, Eq. 13, with respect to the Squared Error and precautionary loss functions when the data given in groups.
Using the Binomial theorem, we rewrite the likelihood function of the grouped data, Eq. 17, as follows:
(28) |
where:
Combining the likelihood information with the prior information yields the posterior distribution θ of given n, i.e:
So we get:
(29) |
Where,
The Bayesian estimate of θ with respect to the squared error loss function
from the Grouped data, say
(30) |
The Bayesian estimate of θ with respect to the Precautionary loss function
from the Grouped data, say
(31) |
The Bayesian estimate of θ with respect to the General Entropy loss function
from the Grouped data, say
(32) |
Now, we compare all these estimators in terms of Biases and Mean Squared Errors (MSE's), using Monte-Carlo simulation.
SIMULATION STUDY
The estimators
Our main aim is to compare these estimators in terms of Biases and MSE's. As
mentioned earlier,
The simulation study was carried out with sample size N = 6,8,10,12,15 and 20. We put these samples into five intervals (k = 4) with δ = 1/5. Prior parameters are arbitrarily taken as α = 2 and β = 1. All the results are summarized in Table 1 and Fig. 2-4.
CONCLUSION
In this study, we obtained Bayesian and non-Bayesian estimators for the shape parameter of the Kumaraswamy distribution based on the grouped and un-grouped data. Meanwhile, we have shown the relationship of Bayesian estimators of the shape parameter of this distribution to three classical estimators, namely the MLE, UMVUE and Min MSE estimator and illustrate how Bayesian methods can yield classical estimators. We considered both point and interval estimators. We derived the Bayes estimators under symmetric and asymmetric loss functions. Our observations about the results are stated in the following points:
Table 1 shows that the Bayes estimates under Squared Error and precautionary loss functions have the smallest estimated MSE's as compared with the classical estimates, These are true for both un-grouped and grouped data. Also, the Bayes estimates under the precautionary loss function have the smallest estimated MSE's as compared with the Bayes estimates under squared error and general entropy loss functions. These are true for both un-grouped and grouped data, (Fig. 2 , 3). It is immediate to note that MSE's decrease as sample size increases. On the other hand the Bayes estimates and the MLE's are overestimation, this is true for the un-grouped data, but UMVUE's and the Min MSE estimates are underestimation. Also MLE's are underestimation but Bayes estimates are overestimation for grouped data. Meantime, the confidence intervals work quite well unless the sample size is very small and this is true for both un-grouped and grouped data.
Whereas, the performance of the Bayes estimates under Precautionary loss function are better than the rest, thus we suggest to use Bayes approach under Precautionary loss function for estimating the parameter of Kumaraswamy distribution and this is true for both un-grouped and grouped data (Fig. 4).
Fig. 2: | MSE's of the Classical and Bayesian point estimators based on the Ungrouped data, for different values of n |
Table 1: | Biases and Mean Squared Errors (MSE's) of the Point Estimates, and Lengths of the Interval Estimates from the un-grouped and grouped data, when k = 4, δ = 1, θ = 2, α = 2, β = 1, λ = 2 and τ = 0.05 (MSE in parenthesis) |
Fig. 3: | MSE's of the Classical and Bayesian point estimators based on the Grouped data, for different values of n |
Fig. 4: | MSE's of the Bayesian point estimators based on both the Grouped and Ungrouped data, for different values of n |
In general, the (Bayes) estimators yield of the Grouped data work very well. Therefore, we can use the estimators presented when the data given in Groups, for example in life testing experiments.