ON THE GEOMETRY OF FEEDFORWARD NEURAL-NETWORK ERROR SURFACES

Citation
Am. Chen et al., ON THE GEOMETRY OF FEEDFORWARD NEURAL-NETWORK ERROR SURFACES, Neural computation, 5(6), 1993, pp. 910-927
Citations number
22
Categorie Soggetti
Computer Sciences","Computer Applications & Cybernetics",Neurosciences
Journal title
ISSN journal
08997667
Volume
5
Issue
6
Year of publication
1993
Pages
910 - 927
Database
ISI
SICI code
0899-7667(1993)5:6<910:OTGOFN>2.0.ZU;2-U
Abstract
Many feedforward neural network architectures have the property that t heir overall input-output function is unchanged by certain weight perm utations and sign flips. In this paper, the geometric structure of the se equioutput weight space transformations is explored for the case of multilayer perceptron networks with tanh activation functions (simila r results hold for many other types of neural networks). It is shown t hat these transformations form an algebraic group isomorphic to a dire ct product of Weyl groups. Results concerning the root spaces of the L ie algebras associated with these Weyl groups are then used to derive sets of simple equations for minimal sufficient search sets in weight space. These sets, which take the geometric forms of a wedge and a con e, occupy only a minute fraction of the volume of weight space. A sepa rate analysis shows that large numbers of copies of a network performa nce function optimum weight vector are created by the action of the eq uioutput transformation group and that these copies all lie on the sam e sphere. Some implications of these results for learning are discusse d.