THE CASCADE-CORRELATION LEARNING - A PROJECTION PURSUIT LEARNING PERSPECTIVE

Citation
Jn. Hwang et al., THE CASCADE-CORRELATION LEARNING - A PROJECTION PURSUIT LEARNING PERSPECTIVE, IEEE transactions on neural networks, 7(2), 1996, pp. 278-289
Citations number
33
Categorie Soggetti
Computer Application, Chemistry & Engineering","Engineering, Eletrical & Electronic","Computer Science Artificial Intelligence","Computer Science Hardware & Architecture","Computer Science Theory & Methods
ISSN journal
10459227
Volume
7
Issue
2
Year of publication
1996
Pages
278 - 289
Database
ISI
SICI code
1045-9227(1996)7:2<278:TCL-AP>2.0.ZU;2-F
Abstract
Cascade-correlation (Cascor) is a popular supervised learning architec ture that dynamically grows layers of hidden neurons of fixed nonlinea r activations (e.g., sigmoids), so that the network topology (size, de pth) can be efficiently determined. Similar to a cascade-correlation l earning network (CCLN), a projection pursuit learning network (PPLN) a lso dynamically grows the hidden neurons. Unlike a CCLN where cascaded connections from the existing hidden units to the new candidate hidde n unit are required to establish high-order nonlinearity in approximat ing the residual error, a PPLN approximates the high-order nonlinearit y by using trainable parametric or semiparametric nonlinear smooth act ivations based on minimum mean squared error criterion. An analysis is provided to show that the maximum correlation training criterion used in a CCLN tends to produce hidden units that saturate and thus makes it more suitable for classification tasks instead of regression tasks as evidenced in the simulation results. It is also observed that this critical weakness in CCLN can also potentially carry over to classific ation tasks, such as the two-spiral benchmark used in the original CCL N paper.