ASYMPTOTIC STATISTICAL-THEORY OF OVERTRAINING AND CROSS-VALIDATION

Citation
S. Amari et al., ASYMPTOTIC STATISTICAL-THEORY OF OVERTRAINING AND CROSS-VALIDATION, IEEE transactions on neural networks, 8(5), 1997, pp. 985-996
Citations number
35
Categorie Soggetti
Computer Application, Chemistry & Engineering","Engineering, Eletrical & Electronic","Computer Science Artificial Intelligence","Computer Science Hardware & Architecture","Computer Science Theory & Methods
ISSN journal
10459227
Volume
8
Issue
5
Year of publication
1997
Pages
985 - 996
Database
ISI
SICI code
1045-9227(1997)8:5<985:ASOOAC>2.0.ZU;2-9
Abstract
A statistical theory for overtraining is proposed. The analysis treats general realizable stochastic neural networks, trained with Kullback- Leibler divergence in the asymptotic case of a large number of trainin g examples. It is shown that the asymptotic gain in the generalization error Is small if we perform early stopping, even if we have access t o the optimal stopping time. Considering cross-validation stopping we answer the question: In what ratio the examples should be divided into training and cross-validation sets in order to obtain the optimum per formance. Although cross-validated early stopping is useless in the as ymptotic region, it surely decreases the generalization error in the n onasymptotic region. Our large scale simulations done on a CM5 are in nice agreement with our analytical findings.