Equilibrium statistical physics is applied to the off-line training of
layered neural networks with differentiable activation functions. A f
irst analysis of soft-committee machines with an arbitrary number (K)
of hidden units and continuous weights learning a perfectly matching r
ule is performed. Our results are exact in the limit of high training
temperatures (beta --> 0). For K = 2 we find a second-order phase tran
sition from unspecialized to specialized student configurations at a c
ritical size P of the training set, whereas for K greater than or equa
l to 3 the transition is first order. The limit K --> infinity can be
performed analytically, the transition occurs after presenting on the
order of NK/beta examples. However, an unspecialized metastable state
persists up tu P proportional to NK2/beta.