Limitations and Future Trends in Neural Computation

Front Cover
Sergey Ablameyko
IOS Press, 2003 - Computers - 245 pages
0 Reviews
This book reports critical analyses on complexity issues in the continuum setting and on generalization to new examples, which are two basic milestones in learning from examples in connectionist models. The problem of loading the weights of neural networks, which is often framed as continuous optimization, has been the target of many criticisms, since the potential solution of any learning problem is severely limited by the presence of local minimal in the error function. The maturity of the field requires to convert the quest for a general solution to all learning problems into the understanding of which learning problems are likely to be solved efficiently. Likewise, the notion of efficient solution needs to be formalized so as to provide useful comparisons with the traditional theory of computational complexity in the discrete setting. The book covers these topics focussing also on recent developments in computational mathematics, where interesting notions of computational complexity emerge in the continuum setting.

What people are saying - Write a review

We haven't found any reviews in the usual places.


The Complexity of Computing with Continuous Time Devices
EnergyBased Computation with Symmetric Hopfield Nets
Computational Complexity and the Elusiveness of Glohal Optima
Impact of Neural Networks on Signal Processing and Communications
Empirical Risk
Leaming Highdimensional Data
The Curse of Dimensionality and the Blessing of Multiply Hyhrid
TeraOPS Stored
Reliahility of ManSystem Interaction and Theory of Neural Networks
Author Index

Common terms and phrases

Bibliographic information