This paper concerns matrix computations within algorithms for variance
and covariance component estimation. Hemmerle and Hartley [Technometr
ics, 15 (1973), pp. 819-831] showed how to compute the objective funct
ion and its derivatives fbr maximum likelihood estimation of variance
components using matrices with dimensions of the order of the number o
f coefficients rather than that of the number of observations, Their a
pproach was extended by Corbeil and Searle [Technometrics, 18 (1976),
pp. 31-38] for restricted maximum likelihood estimation. A similar red
uction in dimension is possible using expectation-maximization (EM) al
gorithms. In most cases, variance components are assumed to be strictl
y positive. We advocate the use of a modification that is numerically
stable even if variance component estimates are small in magnitude. Fo
r problems in which the number of coefficients is large, Fellner [Proc
. Statistical Computing Section, American Statistical Association, 198
4, pp. 150-154], [Comm. Statist. Simulation Comput. B, 16 (1987), pp.
439-463] discusses the use of sparse matrix methods for positive defin
ite systems in EM algorithms. We show how to compute the likelihood fu
nctions and their derivatives via sparse matrix methods for symmetric-
indefinite systems, thus making solution of a much wider class of larg
e-scale problems realizable. Results are formulated for the more gener
al case of covariance components whenever possible.