Escolar Documentos
Profissional Documentos
Cultura Documentos
M. S. Youssef
Abstract
In this paper, Bayesian estimation for the Pareto distribution, this thesis aims
to investigate the possibility of using the quasi-likelihood function in the Bayesian
approach. We introduce a new method which we called "Quasi-Likelihood
Estimation". This method reduces to the usual Bayesian estimation if the distribution
is a member of the exponential family. We use the maximum quasi-likelihood
estimating the unknown parameters of the Pareto distribution.
1- Introduction
The quasi-likelihood function was introduced by Wedderburn (1974), to be
used for estimating the unknown parameters in generalized linear models. The idea of
quasi-likelihood weakens the assumption that we know exactly the distribution of the
random component in the model, and replace it by an assumption about how the
variance changes with the mean.
The quasi-likelihood function could be used for estimation in the same way as
the usual likelihood function. Wedderburn (19974) and McCullagh (1983) showed
that the maximum quasi-likelihood estimates have many properties similar to
particular, the maximum quasi-likelihood estimate of the vector β (the vector of
parameters in regression models) is a asymptotically normal with mean β , and
asymptotic covariance may be derived in the usual fashion from the second derivative
matrix of the quasi-likelihood function. Also, if the underlying distribution comes
from a natural exponential family the maximum quasi-likelihood estimate maximizes
the likelihood function and so it has full asymptotic efficiency; under more general
distributions there is some loss of efficiency, which had been investigated by Firth
(1987) and Hill & Tsai (1988).
510 M. S. Youssef
2- Quasi-Bayesian Estimation
In the present section, we present a new Bayesian estimation technique. We
can apply it without specifying the likelihood function of the sample observations, if
the relationship between the mean and the variance is know. That is, to construct a
posterior distribution the likelihood function could be replaced with the natural
exponential of the quasi-likelihood function. This method reduces to the usual
Bayesian estimation if the quasi-likelihood and the log-likelihood function are
identical.
Let x1 , x 2 , L , x n be an independent random sample, with mean μ = μ (θ ) ,
where θ is a vector of parameters, and the variance var ( x ) = φ V (μ ) , where V (.) is
some know variance function, and φ is a dispersion parameter which could be known
or unknown. The quasi-likelihood Q ( x; μ , φ ) can be derived as defined by the
relation;
∂ x − μi
Q ( xi , μ i ) = i , (A)
∂ μi V ( μi )
and the natural exponential of Q ( x; μ , φ ) will be used as a likelihood function. Using
⎛ φ⎞
a suitable prior density g (θ , φ ) the posterior distribution f * * ⎜θ , ⎟ can be
⎝ x⎠
constructed as
f * * (θ , φ x ) ∝ ⎛⎜ ∏ { exp [Q ( x; μ , φ )] }⎞⎟ g (θ , φ )
n
(2.1)
⎝ i =1 ⎠
where μ = μ (θ ) , θ = (θ 1 , θ 2 , L , θ n ) , θ i ∈ Θ , Θ is the parameter space,
i = 1, 2 , L , n , and φ > 0 . The joint mode of (2.1) can be used to obtain a Bayes
estimate for θ , and marginal densities may be obtained from (2.1) by integrating out
one or more of the unknown parameters, another set of Bayes estimates for the
unknown parameters can be obtained with respect to some loss functions, using these
marginal densities.
Now, let us apply the above method for a simple variance function. Suppose
that x1 , x 2 , L , x n is an independent random sample from an unknown distribution and
suppose that the variance and the mean are known to be proportional and the constant
of proportionality is φ = 1 , so that
Ε (x μ ) = μ , and Var (x μ ) = μ V (μ ) . (2.2)
Where V (μ ) denotes the variance function from (A), we have
Q (x; μ ) = log μ − μ and
Bayesian estimation for the Pareto parameters 511
e x p [ Q ( x; μ )] = μ x θ − μ . (2.3)
In order to construct a posterior distribution, a natural conjugate prior density of μ
is
g (μ ) = (μ υ )υ e − μ υ ,υ > 0 , μ > 0
1
(2.4)
μ Γυ
Using (2.3) and (2.4), the posterior probability distribution function (p. d. f) of μ is
⎛∏ n x −μ ⎞ υ −1 −μ υ
⎜ μ ie ⎟ μ e
⎝ i =1 ⎠
f * * (μ / x1 , x 2 , L , x n ) = ∞
⎛ n x −μ ⎞ μ −1 −μ υ
∫ ⎜∏ μ ie ⎟ μ e ∂μ
0 ⎝ i =1 ⎠
λ
= (λ μ )θo −1 e −λ μ , θ o , λ, μ > 0 (2.5)
Γ(θ o )
o
Where θ o = ∑ x i + υ and λ = n + υ .this, corresponding to a quadratic, loss function,
i =1
the quasi-Bayesian estimator is the posterior mean for (2.5) viz
μ ** = Ε(μ / x ) =
∞ (λμ )o e −λμ ∂μ = θ o
∫ (2.6)
0 Γ (θ o ) λ
the Bays risk of μ ** is
∞
(
var(μ / x ) = ∫ μ − μ ** )2
f ** (μ / x )∂μ =
θo
(2.7)
0 λ3
another Bays estimator is the mode of the posterior p. d. f (2.5), viz
θ −1
μ= o (2.8)
λ
It may be noted that the variance function V (μ ) = μ is associated with the Poisson
distribution which is a member of the exponential family and so the natural
exponential of the quasi-likelihood and the likelihood function are equivalent and this
make the quasi-Bayesian results identical to the usual Bayesian results.
αk
μ= (3.2)
(α − 1 )
and the variance is
α −1 α −1
var ( y ) = μ2 = V (μ ) (3.3)
α (α − 2 ) α (α − 2 )
where V (μ ) = μ 2 is the variance function. Thus for a sample of size n , the
quasi-likelihood function is given by
n
∑ yi − n μ
∂Q i =1
= (3.4)
∂μ μ2
which gives
n
− ∑ yi − n μ
Q ( y; μ ) = i =1
(3.5)
μ − n log μ
Substituting for μ from (3.2) in (3.5), the quasi-likelihood function (3.5) as a
function of α , k , becomes
) = − α − 1 ∑ y i − n log ⎜⎜ α − 1 ⎟⎟
⎛ ⎞
Q (y ; k , α
n
(3.6)
αk ⎝ αk ⎠
i =1
~ nα
We note that k is the maximum likelihood estimate of k , but α~ is not, , the
(n − 2 )
same estimates (3.8), and (3.10), could be derived from the extended quasi-likelihood
function associated with the variance function (3.3).
3.2.2 Quasi-Bayesian estimation for the Pareto distribution:
In this sub-section, the method of quasi-Bayesian estimation is applied to the
Pareto distribution. From (3.6), the natural exponential of the quasi-likelihood
function for a sample of size n observation is from the Pareto distribution is given by
⎛ α −1 ⎞
− D o ⎜⎜ ⎟⎟
exp (Q ( y1 ; α , k )) = α −n
(α − 1) n
k −n
e ⎝αk ⎠
α > 1 , 0 < k ≤ D1 (3.11)
where Do = ∑ y i , and D1 = min ( y i )
o
i =1
now, we derive three posterior distributions of α and k ; when α is known
and k is unknown, when k is known and α is unknown, and when both α and k
are unknown.
Case (1): α is known and k is unknown
Since α is known and k is unknown, a prior P.d.f of k is provided by (3.2.1), as
f 1 (k ) = a o bo o k ao − 1
−a
0 < k < bo (3.12)
where a o , bo are positive. From (3.11) and (3.12) the posterior P.d.f of k is
k D2 − D3
f 1* ( k α y ) = e k
0 < k ≤ D4 (3.13)
A1
α −1
where D 2 = a o n − 1 , D3 = Do , D 4 = min (bo − D1 ) and A1 is the
α
normalizing constant, given by
D4 − D3 k
A1 = ∫ k D2
e ∂k (3.14)
0
i.e. A1 is an incomplete gamma function, and could be calculated numerically.
The mode of (3.13), which is a Bays estimate of k and is
n
∑ yi
⎛ α − 1 ⎞ i =1
k* =⎜ ⎟ (3.15)
⎝ α ⎠ n − ao + 1
Also, the med ian of (3.13) is another Bays estimate of k and is given by
k **
k 2 − D3 k
D
∫ e ∂ k = 0.5 (3.16)
0 A1
Where k ** is the median of (3.13).
Still another Bays estimate of k is the mean of the posterior p.d.f (3.13), viz
1 D 2 +1 − D3 k
Ε (k α , y ) = k e ∂ k = 0.5 (3.17)
A1
The corresponding Bays risk is the variance of the posterior p.d.f (3.13), viz
514 M. S. Youssef
D3
2 D2 −
var (k α , y ) =
1 k
∫ [k − Ε (k α , y )] k e k
∂ k (3.18)
A1 0
The median, the mean and the variance given by (3.16), (3.13) and (3.18)
respectively.
The mode, the median, the mean and the variance given by (3.28), (3.23) and (3.24)
and (3.25) respectively.
We calculate the quasi-Bayesian estimates and the coefficient of variation for k when
α is known which is given by (3.17) and (3.18). also the corresponding results of
Bayesian estimation are obtained and theses results are listed in table (3.1).
Table.(3.2) summarizes the results of Bayesian and quasi-Bayesian estimation for α
when k is known, which is given by (3.24) and (3.25). tables, (3.3) and (3.4) gives
the efficiency of the quasi-Bayesian estimation of k when α is known, and of α
when k is known, respectively.
Bayesian estimation for the Pareto parameters 515
Table (3.1): Bayesian and Quasi-Likelihood Estimation for Pareto Data for k when
α is known (α = 2)
Sample size n = 10 n = 20 n = 30
Table (3.2): Bayesian and Quasi-Likelihood Estimation for Pareto Data for α when
k is known (k = 2)
Sample size n = 10 n = 20 n = 30
Prior Parameters n = 10 n = 20 n = 20
c0 b0
0.5 0.5 0.6136 0.6084 0.2280
1 0.75 0.6342 0.3900 0.2199
1 1 0.6990 0.2512 0.2161
1.5 1.5 0.8069 0.3047 0.2190
2.1 1 0.6152 0.1923 0.2101
2 1.1 0.6452 0.2100 0.2109
2.1 1.5 0.7123 0.2469 0.2010
2 2 0.8086 0.3590 0.2231
Prior Parameters n = 10 n = 20 n = 20
a0 b0
1 1 0.1158 0.0620 0.0381
2 2 0.1173 0.0363 0.0371
0.5 3 0.1126 0.0365 0.0425
0.42 3.1 0.1147 0.0367 0.0312
0.6 3.1 0.1147 0.0363 0.0372
1 3.2 0.1150 0.0360 0.0380
2 3 0.1172 0.0368 0.0399
3 3.5 0.1270 0.0375 0.0415
Tables. (3.1) and (3.2) shows that the difference between the coefficients of variation
of the Bayesian estimates and the quasi-Bayesian estimate is small, further the
difference is decreases as the sample size and the values of the prior parameters
increases. also, the coefficient of variation of the quasi-Bayesian estimates of α is
smaller then the coefficient of variation of the Bayesian estimate of α .
Table. (3.3) shows that the quasi-Bayesian estimate of the parameter k has a low
efficiency relative to the Bayesian estimate, also the efficiency decreases as the
sample size increases.
Table. (3.4) shows the quasi-Bayesian estimate of the parameter α has a high
efficiency even for small sample size and large values of prior parameters. But, the
efficiency decreases as the sample size increases.
Bayesian estimation for the Pareto parameters 517
References
1- Arnold, B.C. and Press, S. J. (1983). Bayesian inference for Pareto population.
Journal of Econometrices. 21, 287-306.
2- Arnold, B.C. and Press, S. J. (1986). Bayesian analysis of censored or grouped date
from Pareto population in Bayesian inference and decision techniques. Eds. P.
Goel and A. Zellner, Amsterdam: North Holland, pp. 157 – 173.
3- Arnold, B.C. and Press, S. J. (1989). Bayesian estimation and prediction for Pareto
data. JJASA, vol. 4. No. 408, 1079 – 1084.
4- Hill, J. R. and Tsai, C.L. (1988). Calculating the efficiency of maximum quasi-
likelihood estimation. Apple. Statist. 37, 2, 219 – 230.
5- McCullagh, P. and Nelder, J.A. (1983). Generalized linear models. London:
Chapman and Hall.
6- Wedderburn, R.W. M. (1974). Quasi-likelihood functions, generalized linear
models, and the Gauss-Newton method. Biometrika 61, 439-447.