OVERTRAINING IN NEURAL NETWORKS THAT INTERPRET CLINICAL-DATA

Citation
Ml. Astion et al., OVERTRAINING IN NEURAL NETWORKS THAT INTERPRET CLINICAL-DATA, Clinical chemistry, 39(9), 1993, pp. 1998-2004
Citations number
34
Categorie Soggetti
Chemistry Medicinal
Journal title
ISSN journal
00099147
Volume
39
Issue
9
Year of publication
1993
Pages
1998 - 2004
Database
ISI
SICI code
0009-9147(1993)39:9<1998:OINNTI>2.0.ZU;2-P
Abstract
Backpropagation neural networks are a computer-based pattern-recogniti on method that has been applied to the interpretation of clinical data . Unlike rule-based pattern recognition, backpropagation networks lear n by being repetitively trained with examples of the patterns to be di fferentiated. We describe and analyze the phenomenon Of overtraining i n backpropagation networks. Overtraining refers to the reduction in ge neralization ability that can occur as networks are trained. The clini cal application we used was the differentiation of giant cell arteriti s (GCA) from other forms of vasculitis (OTH) based on results for 807 patients (593 OTH, 214 GCA) and eight clinical predictor variables. Th e 807 cases were randomly assigned to either a training set with 404 c ases or to a cross-validation set with the remaining 403 cases. The cr oss-validation set was used to monitor generalization during training. Results were obtained for eight networks, each derived from a differe nt random assignment of the 807 cases. Training error monotonically de creased during training. In contrast, the cross-validation error usual ly reached a minimum early in training while the training error was st ill decreasing. Training beyond the minimum cross-validation error was associated with an increased cross-validation error. The shape of the cross-validation error curve and the point during training correspond ing to the minimum cross-validation error varied with the composition of the data sets and the training conditions. The study indicates that training error is not a reliable indicator of a network's ability to generalize. To find the point during training when a network generaliz es best, one must monitor cross-validation error separately.