Jas. Freeman et D. Saad, RADIAL BASIS FUNCTION NETWORKS - GENERALIZATION IN OVER-REALIZABLE AND UNREALIZABLE SCENARIOS, Neural networks, 9(9), 1996, pp. 1521-1529
Learning and generalization in a two-layer radial basis function netwo
rk, with fixed centres of the basis functions, is examined within a st
ochastic training paradigm. Employing a Bayesian approach, expressions
for generalization error are derived under the assumption that the ge
nerating mechanism (leacher) for the training data is also a radial ba
sis function network, but one for which the basis function centres and
widths need not correspond to those of the student network. The effec
ts of regularization, via a weight decay term, are examined. The cases
in which the student has greater representational power than the teac
her (over-realizable), and in which the teacher has greater power than
the student (unrealizable) are studied. Dependence on knowing the cen
tres of the teacher is eliminated by introducing a single degree-of-co
nfidence parameter. Finally, simulations are performed which validate
the analytic results. Copyright (C) 1996 Elsevier Science Ltd.