M. Rupp et Ah. Sayed, SUPERVISED LEARNING OF PERCEPTRON AND OUTPUT-FEEDBACK DYNAMIC NETWORKS - A FEEDBACK ANALYSIS VIA THE SMALL GAIN THEOREM, IEEE transactions on neural networks, 8(3), 1997, pp. 612-622
This paper provides a time-domain feedback analysis of the perceptron
learning algorithm and of training schemes for dynamic networks with o
utput feedback, It studies the robustness performance of the algorithm
s in the presence of uncertainties that might be due to noisy perturba
tions in the reference signals or to modeling mismatch, In particular,
bounds are established on the step-size parameters in order to guaran
tee that the resulting algorithms will behave as robust filters, The p
aper also establishes that an intrinsic feedback structure can be asso
ciated with the training schemes, The feedback configuration is motiva
ted via energy arguments and is shown to consist of two major blocks:
a time-variant lossless (i.e., energy preserving) feedforward path and
a time-variant feedback path, The stability of the feedback structure
is then analyzed via the small gain theorem and choices for the step-
size parameter in order to guarantee faster convergence are deduced by
appealing to the mean-value theorem, Simulation results are included
to demonstrate the findings.