HOME JOURNALS CONTACT

Journal of Applied Sciences

Year: 2011 | Volume: 11 | Issue: 12 | Page No.: 2154-2162
DOI: 10.3923/jas.2011.2154.2162
Classical and Bayesian Estimations on the Kumaraswamy Distribution using Grouped and Un-grouped Data under Difference Loss Functions
Ramin Gholizadeh, Aliakbar Mastani Shirazi and Samira Mosalmanzadeh

Abstract: The aim of this study is finding classical and Bayesian estimators for the shape parameter of the Kumaraswamy distribution using un-grouped data and also considers relationship between them. We show how the classical estimators can be derived from various choices made within a Bayesian framework. We compare the classical estimators based on their Mean Squared Errors (MSE's). Then, we obtain Bayesian and non-Bayesian estimators of the shape parameter of this distribution under Grouped data. In Bayesian estimation, we consider three types of loss functions; the Squared error, Precautionary and General entropy loss functions which are symmetric and asymmetric, respectively. In all cases, we considered both point and interval estimations. These the point and interval estimations are compared empirically using Monte-Carlo simulation. Bayes approach under Precautionary loss function is best estimator for estimating the parameter of Kumaraswamy distribution and this is true for both un-grouped and grouped data.

Fulltext PDF Fulltext HTML

How to cite this article
Ramin Gholizadeh, Aliakbar Mastani Shirazi and Samira Mosalmanzadeh, 2011. Classical and Bayesian Estimations on the Kumaraswamy Distribution using Grouped and Un-grouped Data under Difference Loss Functions. Journal of Applied Sciences, 11: 2154-2162.

Keywords: prior distribution, loss function, reliability, Likelihood estimator, Kumaraswamy distribution and credibility interval

INTRODUCTION

In life testing experiments, we observe the failure time of a component to the nearest hour, day or month. Data for which true values are known only up to subsets of the sample space are called grouped data (Alodat and Al-Saleh, 2000; Surles and Padgett, 2001; Wu and Perloff, 2005; Pipper and Ritz, 2007). In general grouped data can be formulated as follows: Let X1, X2,..., Xn be a random sample from the density f (x; θ), x ε χ, θ ε and χ1, χ2,..., χk+1 be a partition of the sample space χ and Nj the number of X j’s that fall in for j = 1,2,..., k+1. The set of pairs {(χ1, N1),..., (χk+1, N + 1)} is called grouped data. The grouped data problem is to use these data to draw inferences about the parameter θ. Since we don’t have complete information about the sample, then there will be a loss in the information due to the grouping. Schervish (1995) showed the following:

where, I X (θ) and I Y (θ) are the Fisher's information number obtained from X and Y, respectively and Eθ[I X\Y (θ\Y)] is the conditional score function. If we replace Y by the grouped sample n = (N1 N2,..., Nk+1), then I X (θ)≥In (θ) for all θ, which means that the information in the sample X about θ is reduced to In (θ) because of grouping. Kuldorff (1961) considered non-Bayesian estimation from grouped data when the data come from normal and exponential distributions. Alodat and Al-Saleh (2000) considered the Bayesian estimation from grouped data when the underlying distribution is exponential. Alodat et al. (2007) obtained Bayesian prediction intervals from grouped data when the underlying distribution is exponential. Aludaat et al. (2008) obtained the Bayesian and non-Bayesian estimation from grouped data when the underlying distribution is Burr type X.

The Kumaraswamy distributions were constructed by Kumaraswamy (1980). Jones (2009) said about its properties. The probability density function of a Kumaraswamy distributed random variable is given by:

(1)

where, θ and λ are shape parameters, respectively. Here we assume that λ parameter is known. The distribution function is:

(2)

The reliability and failure rate functions of Kumaraswamy distribution are given, respectively by:

(3)

And

(4)

Figure 1 shows the shape of for different values of θ and λ.

Bayesian estimators derived from an improper prior distribution can be use to derive the classical estimators. The technique of deriving the classical estimators from the Bayesian estimator is not new. Rossman et al. (1998) and Elfessi and Reineke (2001) presented some thought provoking insights on the relationship between Bayesian and classical estimation using the continuous uniform and exponential distributions, respectively. We will explore these relationships using the Kumaraswamy distribution from un-grouped data.

CLASSICAL AND BAYESIAN ESTIMATIONS BASED ON THE UN-GROUPED DATA

Here, first we obtain the classical estimators of and compare these estimators based on their Mean Squared Errors (MSE's) and then we get the Bayes estimators of under the symmetric and asymmetric loss function and show how the classical estimators can be derived from various choices made within a Bayesian framework. Also, we present the credible and confidence intervals for θ.

Classical point and interval estimations: Let X1, X2,..., Xn be a random sample from density (Eq. 2). The likelihood function is given by:

(5)

Fig. 1: Pdf of Kumaraswamy for different values of θ and λ

Then the log-likelihood function is:

(6)

Hence,

Thus the MLE of is:

(7)

where,

The above estimator obtained by Gupta and Kundu (1999).

Here, we obtain the Uniformly Minimum Variance Unbiased Estimator (UMVUE) of θ. since family of density (2) belongs to an exponential family, therefore, statistic T is a complete sufficient statistic for θ. It is easy to show that statistic T is distributed as gamma distribution with parameters n and 1/θ, with the density:

thus:

Hence, the UMVUE of is:

(8)

We can find the Minimum Mean Squared Error (Min MSE) estimator in the class of estimators of the form c/T. Therefore:

Where as:

thus:

And:

Then:

(9)

The derivative of r (c) is:

That thereby c = n-2. Thus, the Min MSE estimator of θ is:

(10)

From Eq. 7, the MSE of the classical estimators of are calculated as follow:

And

Easily can show that:

Now, we find a 100 (1-τ)% confidence interval for θ with obtain L and U, where. P(L<θ<U) = 1-τ. Let X1, X2,..., Xn be a random sample from Kumaraswamy. Since,T~ (n,1/θ), there by 2θT~χ2 (2n) , thus

or

Therefore a classical 100 (1-τ)% confidence interval for θ is given by:

Bayesian point and interval estimations: Here, we obtain the Bayes estimators of θ under the improper prior distribution. Consider the improper prior distribution for θ of the form:

Notice that this prior distribution is the kernel of a Gamma distribution when α>0. However, such a restriction on α is not necessary and decreases the flexibility of the resulting estimator. Whereas , therefore the posterior distribution of θ is:

where:

The posterior distribution of θ is proper when, n+α>0, i.e.:

In this case, the Bayes estimator of θ under Squared error loss function is the posterior mean, i.e:

(11)

Where,

The above estimator obtained by Kundu and Gupta (2008). The classical estimators derived in the previous subsection can be obtained from the above Bayes estimator. To do this, we put the different values of α and β. For example, If α = 0 and β = 0 then and the prior distribution is the Jeffreys' prior, π (θ)∝1/θ (Lehmann and Casella, 1998). If α = 0 and β = 0 then and if α = 0 and β = 0 then . For α = 0 and β = 0 the prior distribution reduce to the flat improper prior distribution, i.e. π (θ)∝ 1. In this case the resulting estimator is (n+1)/T. We saw that a Bayesian analysis with a simple family of improper prior distributions provides a direct link among several classical estimators.

Now, we obtained the Bayes estimator of θ under an asymmetric loss function. Norstrom (1996) introduced an alternative asymmetric precautionary loss function and also presented a general class of precautionary loss function with quadratic loss function as a special case. These loss functions approach infinitely near the origin to prevent underestimation and thus giving conservative estimators, especially when low failure rates are being estimated. These estimators are very useful when underestimation may lead to serious consequences. A very useful and simple asymmetric precautionary loss function is:

(12)

The Bayes estimator under this asymmetric loss function is denoted by and may be obtained by solving the following Eq:

(13)

This special case of the Precautionary loss function (Eq. 10) and the Entropy loss function are the same (Norstrom, 1996).

As said, under the gamma prior distribution, i.e:

(14)

the posterior density of θ is gamma with the shape and scale parameters as α + n and 1/(β+t), respectively, therefore:

Hence, the Bayes estimator of θ with respect to the precautionary loss function under the gamma prior distribution is as follows:

(15)

Where:

In many practical situations, it appears to be more realistic to express the loss in terms of the ratio . In this case, a useful asymmetric loss function is the General Entropy (GE) loss proposed by Calabria and Pulcini (1996):

(16)

whose minimum occurs at . This loss function is a generalization of the Entropy loss used by several authors where the shape parameter q = 1, (Dey and Liu, 1992). When q>0, a positive error causes more serious consequences than a negative error. The Bayes estimate of θ under the General Entropy loss (Eq. 16) is:

(17)

Provided that Eθ-q} exists and is finite, where, Eθ denoted to the expected value with respect to the posterior function of θ.

The Bayes estimator of θ with respect to the General Entropy loss function under the gamma prior distribution is as follows:

(18)

where,

The Bayesian analog to the confidence interval is called a credibility interval. In general, the interval (L(x), U(x)) is a 100 (1-τ)% credibility interval for θ if:

Since,

Thereby, 2θ (β+T)∼χ2 (2(n+α)), thus:

Therefore, a 100 (1-τ)% Bayesian credibility interval for θ is (L(x), U(x)) where:

(19)

And:

(20)

The classical and Bayesian interval estimators are therefore the same when α = 0 and β = 0.

CLASSICAL AND BAYESIAN ESTIMATIONS BASED ON THE GROUPED DATA

Here, we obtain the MLE and Bayes estimators of θ and also the Fisher's information number when the data given in Groups. Also, we use the Fisher's information number to construct a Asymptotic Confidence interval for θ.

Likelihood function and MLE: Here, first we derive the likelihood density based on the grouped data. Let X1, X2,..., Xn be a random sample from Kumaraswamy. Assume that the sample space of f (x; 0) is partitioned into k+1 equally-spaced intervals as follows. Let Ij = [(j-1)δ, jδ), j = 1,..., k and IK+1 = [kδ,1), δ>0. If Nj denotes the number of Xj ’s that fall in Ij, j = 1,2,..., k+1, then n = N1 +...+ Nk+1. Let:

For j = 1,...k and . If we let, , then , for j = 1,..., k and . So the density of n = (N1, N2,..., Nk+1) is given by the multinomial distribution as follows:

(21)

where, C is a normalizing constant.

In continue, we find the MLE of θ based on the density (20). To do this, we maximize the log likelihood function:

The first derivative of the log-likelihood is:

(22)

The MLE for θ is the solution of . So the M.L.E is θ such that:

(23)

We use the notation to denote the M.L.E of θ obtained from the Grouped data. We can solve (21) by Newton-Raphson method. Hence, solution of the equation is:

(24)

where,

And

Here, the initial solution θ0 should be selected from the MLE of θ based on the un-grouped data.

Fisher's information number and confidence interval: To find the fisher's information number contained in the grouped sample about θ, we find the expectation of the second derivative of the log-likelihood. So:

(25)

where,

If IG(θ) denotes the Fisher's information number from grouped data, then:

And since, E[Nj] = nPj, thus:

(26)

Using IG(θ), we can find a large sample (1-α) 100% confidence interval for θ as follows:

(27)

Simple calculations can show that the Fisher's information number about θ in a random sample X1, X2,..., Xn from (1) is I(θ) = n/θ2.

Bayesian estimation: Here, we obtain the Bayes estimators of θ under the Gamma prior distribution, Eq. 13, with respect to the Squared Error and precautionary loss functions when the data given in groups.

Using the Binomial theorem, we rewrite the likelihood function of the grouped data, Eq. 17, as follows:

(28)

where:

Combining the likelihood information with the prior information yields the posterior distribution θ of given n, i.e:

So we get:

(29)

Where,

The Bayesian estimate of θ with respect to the squared error loss function from the Grouped data, say , is the posterior mean, i.e.,

(30)

The Bayesian estimate of θ with respect to the Precautionary loss function from the Grouped data, say , is obtained as follows:

(31)

The Bayesian estimate of θ with respect to the General Entropy loss function from the Grouped data, say , is obtained as follows:

(32)

Now, we compare all these estimators in terms of Biases and Mean Squared Errors (MSE's), using Monte-Carlo simulation.

SIMULATION STUDY

The estimators and are the classical and Bayesian estimations of the shape parameter of the Kumaraswamy distribution obtained from the un-grouped data. Meanwhile and are the MLE and Bayes estimators of θ under the Squared-error, Precautionary and General Entropy loss functions, respectively, based on the Grouped data. We also use the notations CL and BCL to denote the 95% Confidence and Bayesian Credibility Length, respectively for θ based on the un-grouped data and use notation CLG to denote the 95% Confidence Length for θ based on the Grouped data.

Our main aim is to compare these estimators in terms of Biases and MSE's. As mentioned earlier, and hence its MSE can not be put in a convenient closed form. Therefore, MSE's of the estimators are empirically evaluated based on a Monte-Carlo simulation study of 1000 samples. We generated these samples from Kumaraswamy distribution with θ = 2 by using MATLAB.

The simulation study was carried out with sample size N = 6,8,10,12,15 and 20. We put these samples into five intervals (k = 4) with δ = 1/5. Prior parameters are arbitrarily taken as α = 2 and β = 1. All the results are summarized in Table 1 and Fig. 2-4.

CONCLUSION

In this study, we obtained Bayesian and non-Bayesian estimators for the shape parameter of the Kumaraswamy distribution based on the grouped and un-grouped data. Meanwhile, we have shown the relationship of Bayesian estimators of the shape parameter of this distribution to three classical estimators, namely the MLE, UMVUE and Min MSE estimator and illustrate how Bayesian methods can yield classical estimators. We considered both point and interval estimators. We derived the Bayes estimators under symmetric and asymmetric loss functions. Our observations about the results are stated in the following points:

Table 1 shows that the Bayes estimates under Squared Error and precautionary loss functions have the smallest estimated MSE's as compared with the classical estimates, These are true for both un-grouped and grouped data. Also, the Bayes estimates under the precautionary loss function have the smallest estimated MSE's as compared with the Bayes estimates under squared error and general entropy loss functions. These are true for both un-grouped and grouped data, (Fig. 2 , 3). It is immediate to note that MSE's decrease as sample size increases. On the other hand the Bayes estimates and the MLE's are overestimation, this is true for the un-grouped data, but UMVUE's and the Min MSE estimates are underestimation. Also MLE's are underestimation but Bayes estimates are overestimation for grouped data. Meantime, the confidence intervals work quite well unless the sample size is very small and this is true for both un-grouped and grouped data.

Whereas, the performance of the Bayes estimates under Precautionary loss function are better than the rest, thus we suggest to use Bayes approach under Precautionary loss function for estimating the parameter of Kumaraswamy distribution and this is true for both un-grouped and grouped data (Fig. 4).

Fig. 2: MSE's of the Classical and Bayesian point estimators based on the Ungrouped data, for different values of n

Table 1: Biases and Mean Squared Errors (MSE's) of the Point Estimates, and Lengths of the Interval Estimates from the un-grouped and grouped data, when k = 4, δ = 1, θ = 2, α = 2, β = 1, λ = 2 and τ = 0.05 (MSE in parenthesis)

Fig. 3: MSE's of the Classical and Bayesian point estimators based on the Grouped data, for different values of n

Fig. 4: MSE's of the Bayesian point estimators based on both the Grouped and Ungrouped data, for different values of n

In general, the (Bayes) estimators yield of the Grouped data work very well. Therefore, we can use the estimators presented when the data given in Groups, for example in life testing experiments.

REFERENCES

  • Al Odat, M.T. and M.F. Al-Saleh, 2000. Bayesian estimation using grouped data with application to the exponential distribution. Soochow J. Math., 25: 343-357.
    Direct Link    


  • Alodat, M.T., K.M. Aludaat and T.T. Alodat, 2007. Bayesian prediction intervals from grouped data: Exponential distribution. Abhath Al-Yarmouk, 2007.


  • Aludaat, K.M., M.T. Alodat and T.T. Alodat, 2008. Parameter estimation of Burr type X distribution for grouped data. J. Appl. Math. Sci., 2: 415-423.
    Direct Link    


  • Calabria, R. and G. Pulcini, 1996. Point estimation under asymmetric loss functions for left-truncated exponential samples. Commun. Stat.: Theory Methods, 25: 585-600.
    CrossRef    Direct Link    


  • Dey, D.K. and P.S.L. Liu, 1992. On comparison of estimators in a generalized life model. Microelectron. Reliab., 32: 207-221.
    CrossRef    


  • Elfessi, A. and D.M. Reineke, 2001. A bayesian look at classical estimation: The exponential distribution. J. Stat. Educ., 9: 1-7.


  • Gupta, R.D. and D. Kundu, 1999. Generalized-exponential distribution. Aust. N. Z. J. Stat., 41: 173-188.
    CrossRef    


  • Jones, M.C., 2009. Kumaraswamys distribution: A β-type distribution with some tractability advantages. Stat. Methodol., 6: 70-81.
    CrossRef    


  • Kuldorff, G., 1961. Estimation from Grouped and Partially Grouped Samples. John Wiley and Sons, New York, USA., pp: 144


  • Kundu, D. and R.D. Gupta, 2008. Generalized exponential distribution: Bayesian inference Comput. Stat. Data Anal., 52: 1873-1883.
    Direct Link    


  • Kumaraswamy, P., 1980. A generalized probability density function for double-bounded random processes. J. Hydrol., 46: 79-88.
    CrossRef    Direct Link    


  • Lehmann, E.L. and G. Casella, 1998. Theory of Point Estimation. 2nd Edn., Springer-Verlag, New York, ISBN: 0-387-98502-6


  • Norstrom, J.G., 1996. The use of precautionary loss function in risk analysis. IEEE Trans. Reliab., 45: 400-403.
    CrossRef    


  • Pipper, C.B. and C. Ritz, 2007. Cheking the grouped data version of the cox model for interval-grouped survival data. Scand. J. Stat., 34: 405-418.
    CrossRef    Direct Link    


  • Rossman, A.J., T.H. Short and M.T. Parks, 1998. Bayes estimators for the continuous uniform distribution. J. Stat. Educ., 6: 1-7.


  • Schervish, M.J., 1995. Theory of Statistics. Springer-Verlag Inc., New York, USA., Pages: 702


  • Surles, J.G. and W.J. Padgett, 2001. Inference for reliability and stress-strength for a scaled Burr type X distribution. Lifetime Data Anal., 7: 187-202.
    CrossRef    


  • Wu, X. and J.M. Perloff, 2005. Chinas income distribution: 1985-2001. Rev. Econ. Stat., 87: 763-775.
    CrossRef    

  • © Science Alert. All Rights Reserved