, to form your coefficient covariance estimate. If you employ BHHH, the coefficient covariance will be estimated using the inverse of the outer product of the scores
, where
and
are the gradient (or score) and Hessian of the log likelihood evaluated at the ML estimates. ![]() | (31.62) |
, but as with all QML estimation, caution is advised.
belongs to the exponential family and that the conditional mean of
is a (smooth) nonlinear transformation of the linear part
:![]() | (31.63) |
, it does not possess any efficiency properties. An alternative consistent estimate of the covariance is obtained if we impose the GLM condition that the (true) variance of
is proportional to the variance of the distribution used to specify the log likelihood:![]() | (31.64) |
that is independent of
. The most empirically relevant case is
, which is known as overdispersion. If this proportional variance condition holds, a consistent estimate of the GLM covariance is given by:![]() | (31.65) |
![]() | (31.66) |
. When you select GLM standard errors, the estimated proportionality term
is reported as the variance factor estimate in EViews.
groups, and let
be the number of observations in group
. Define the number of
observations and the average of predicted values in group
as:![]() | (31.67) |
![]() | (31.68) |
distribution with
degrees of freedom. Note that these findings are based on a simulation where
is close to
.
groups. Since
is binary, there are
cells into which any observation can fall. Andrews (1988a, 1988b) compares the
vector of the actual number of observations in each cell to those predicted from the model, forms a quadratic form, and shows that the quadratic form has an asymptotic
distribution if the model is specified correctly.
be an
matrix with element
, where the indicator function
takes the value one if observation
belongs to group
with
, and zero otherwise (we drop the columns for the groups with
to avoid singularity). Let
be the
matrix of the contributions to the score
. The Andrews test statistic is
times the
from regressing a constant (one) on each column of
and
. Under the null hypothesis that the model is correctly specified,
is asymptotically distributed
with
degrees of freedom.