SUPERVISED LEARNING OF PERCEPTRON AND OUTPUT-FEEDBACK DYNAMIC NETWORKS - A FEEDBACK ANALYSIS VIA THE SMALL GAIN THEOREM

Authors
Citation
M. Rupp et Ah. Sayed, SUPERVISED LEARNING OF PERCEPTRON AND OUTPUT-FEEDBACK DYNAMIC NETWORKS - A FEEDBACK ANALYSIS VIA THE SMALL GAIN THEOREM, IEEE transactions on neural networks, 8(3), 1997, pp. 612-622
Citations number
21
Categorie Soggetti
Computer Application, Chemistry & Engineering","Engineering, Eletrical & Electronic","Computer Science Artificial Intelligence","Computer Science Hardware & Architecture","Computer Science Theory & Methods
ISSN journal
10459227
Volume
8
Issue
3
Year of publication
1997
Pages
612 - 622
Database
ISI
SICI code
1045-9227(1997)8:3<612:SLOPAO>2.0.ZU;2-2
Abstract
This paper provides a time-domain feedback analysis of the perceptron learning algorithm and of training schemes for dynamic networks with o utput feedback, It studies the robustness performance of the algorithm s in the presence of uncertainties that might be due to noisy perturba tions in the reference signals or to modeling mismatch, In particular, bounds are established on the step-size parameters in order to guaran tee that the resulting algorithms will behave as robust filters, The p aper also establishes that an intrinsic feedback structure can be asso ciated with the training schemes, The feedback configuration is motiva ted via energy arguments and is shown to consist of two major blocks: a time-variant lossless (i.e., energy preserving) feedforward path and a time-variant feedback path, The stability of the feedback structure is then analyzed via the small gain theorem and choices for the step- size parameter in order to guarantee faster convergence are deduced by appealing to the mean-value theorem, Simulation results are included to demonstrate the findings.