An introduction to computational learning theory

Emphasizing issues of computational efficiency, Michael Kearns and Umesh Vazirani introduce a number of central topics in computational learning theory for researchers and students in artificial intelligence, neural networks, theoretical computer science, and statistics.Computational learning theory...

Full description

Bibliographic Details
Main Author: Kearns, Michael J.
Other Authors: Vazirani, Umesh Virkumar
Format: eBook
Language:English
Published: Cambridge, Mass. MIT Press 1994
Subjects:
Online Access:
Collection: MIT Press eBook Archive - Collection details see MPG.ReNa
LEADER 02736nmm a2200337 u 4500
001 EB002070683
003 EBX01000000000000001210773
005 00000000000000.0
007 cr|||||||||||||||||||||
008 220922 ||| eng
020 |a 9780262276863 
020 |a 0262276860 
020 |a 9780585350530 
020 |a 0585350531 
050 4 |a Q325.5 
100 1 |a Kearns, Michael J. 
245 0 0 |a An introduction to computational learning theory  |h Elektronische Ressource  |c Michael J. Kearns, Umesh V. Vazirani 
260 |a Cambridge, Mass.  |b MIT Press  |c 1994 
300 |a xii, 207 pages  |b illustrations 
653 |a Machine learning 
653 |a COMPUTER SCIENCE/Machine Learning & Neural Networks 
653 |a Algorithms 
653 |a Neural networks (Computer science) 
653 |a Artificial intelligence 
700 1 |a Vazirani, Umesh Virkumar 
041 0 7 |a eng  |2 ISO 639-2 
989 |b MITArchiv  |a MIT Press eBook Archive 
028 5 0 |a 10.7551/mitpress/3897.001.0001 
856 4 0 |u https://doi.org/10.7551/mitpress/3897.001.0001?locatt=mode:legacy  |x Verlag  |3 Volltext 
082 0 |a 006.3 
520 |a Emphasizing issues of computational efficiency, Michael Kearns and Umesh Vazirani introduce a number of central topics in computational learning theory for researchers and students in artificial intelligence, neural networks, theoretical computer science, and statistics.Computational learning theory is a new and rapidly expanding area of research that examines formal models of induction with the goals of discovering the common methods underlying efficient learning algorithms and identifying the computational impediments to learning.Each topic in the book has been chosen to elucidate a general principle, which is explored in a precise formal setting. Intuition has been emphasized in the presentation to make the material accessible to the nontheoretician while still providing precise arguments for the specialist. This balance is the result of new proofs of established theorems, and new presentations of the standard proofs.The topics covered include the motivation, definitions, and fundamental results, both positive and negative, for the widely studied L. G. Valiant model of Probably Approximately Correct Learning; Occam's Razor, which formalizes a relationship between learning and data compression; the Vapnik-Chervonenkis dimension; the equivalence of weak and strong learning; efficient learning in the presence of noise by the method of statistical queries; relationships between learning and cryptography, and the resulting computational limitations on efficient learning; reducibility between learning problems; and algorithms for learning finite automata from active experimentation