The Computational Complexity of Machine Learning
MIT Press, 1990 - Computers - 165 pages
The Computational Complexity of Machine Learning is a mathematical study of the possibilities for efficient learning by computers. It works within recently introduced models for machine inference that are based on the theory of computational complexity and that place an explicit emphasis on efficient and general algorithms for learning. Theorems are presented that help elucidate the boundary of what is efficiently learnable from examples. These results take the form of both algorithms with proofs of their performance, and hardness results demonstrating the intractability of learning in certain natural settings. In addition the book contains lower bounds on the resources required for learning, an extensive study of learning in the presence of errors in the sample data, and several theorems demonstrating reducibilities between learning problems. Michael J. Kearns is Postdoctoral Associate in the Laboratory for Computer Science at MIT. Contents: Definitions, Notations, and Motivation. Overview of Recent Research in Computational Learning Theory. Useful Tools for Distribution-Free Learning. Learning in the Presence of Errors. Lower Bounds on Sample Complexity. Cryptographic Limitations on Polynomial-Time Learning. Distribution-Specific Learning in Polynomial Time. Equivalence of Weak Learning and Group Learning.
What people are saying - Write a review
We haven't found any reviews in the usual places.
Recent Research in Computational Learning Theory
Tools for Distributionfree Learning
Learning in the Presence of Errors
Lower Bounds on Sample Complexity
Cryptographic Limitations on Polynomialtime Learning
Distributionspecific Learning in Polynomial Time
Equivalence of Weak Learning and Group Learning
Conclusions and Open Problems
accuracy allow appear apply approximation assume assumptions Blum Chapter choose circuits classification coloring combinatorial optimization computational concept consider consistent constant Corollary cost cover define definitions denote depend describe determine distribution-free model draw drawn equivalent expect Fact factor finding finite fixed follows functions give given hardness results high probability holds hypothesis input instance interesting known learning algorithm learning problem least length lower bound machine malicious error rate monomial monotone natural negative examples Note obtain optimal oracle output parameterized Pitt polynomial-time polynomial-time algorithm polynomial-time learning polynomially learnable positive examples possible presented problem Proof prove random Recently reducibility representation class respectively restricted sample sample complexity satisfying significant simply simulation Step target distributions target representation Theorem theory tolerated trapdoor uniform Valiant variables
Page 151 - IEEE Symp. on Foundations of Computer Science, 1988, pp. 100-109.  S. Judd. Learning in neural networks. Proc. of the 1988 Workshop on Computational Learning Theory, Morgan Kaufmann Publishers, 1988, pp 2-8.  M. Kearns, M. Li, L. Pitt, LG Valiant. On the learnability of Boolean formulae.