ON THE COMPLEXITY OF TRAINING NEURAL NETWORKS WITH CONTINUOUS ACTIVATION FUNCTIONS

Citation
B. Dasgupta et al., ON THE COMPLEXITY OF TRAINING NEURAL NETWORKS WITH CONTINUOUS ACTIVATION FUNCTIONS, IEEE transactions on neural networks, 6(6), 1995, pp. 1490-1504
Citations number
27
Categorie Soggetti
Computer Application, Chemistry & Engineering","Engineering, Eletrical & Electronic","Computer Science Artificial Intelligence","Computer Science Hardware & Architecture","Computer Science Theory & Methods
ISSN journal
10459227
Volume
6
Issue
6
Year of publication
1995
Pages
1490 - 1504
Database
ISI
SICI code
1045-9227(1995)6:6<1490:OTCOTN>2.0.ZU;2-S
Abstract
We deal with computational issues of loading a fixed-architecture neur al network with a set of positive and negative examples. This is the f irst result on the hardness of loading a simple three-node architectur e which does not consist of the binary-threshold neurons, but rather u tilizes a particular continuous activation function, commonly used in the neural-network literature. We observe that the loading problem is polynomial-time if the input dimension is constant. Otherwise, however , any possible learning algorithm based on particular fixed architectu res faces severe computational barriers. Similar theorems have already been proved by Megiddo and by Blum and Rivest, to the case of binary- threshold networks only, Our theoretical results lend further suggesti on to the use of incremental (architecture-changing) techniques for tr aining networks rather than fixed architectures. Furthermore, they imp ly hardness of learnability in the probably approximately correct sens e as well.