This is the first comprehensive introduction to computational learning theory. The author's uniform presentation of fundamental results and their applications offers AI researchers a theoretical perspective on the problems they study. The book presents tools for the analysis of probabilistic models of learning, tools that crisply classify what is and is not efficiently learnable. After a general introduction to Valiant's PAC paradigm and the important notion of the Vapnik-Chervonenkis dimension, the author explores specific topics such as finite automata and neural networks. The presentation is intended for a broad audience--the author's ability to motivate and pace discussions for beginners has been praised by reviewers. Each chapter contains numerous examples and exercises, as well as a useful summary of important results. An excellent introduction to the area, suitable either for a first course, or as a component in general machine learning and advanced AI courses. Also an important reference for AI researchers.
To view this DRM protected ebook on your desktop or laptop you will need to have Adobe Digital Editions installed. It is a free software. We also strongly recommend that you sign up for an AdobeID at the Adobe website. For more details please see FAQ 1&2. To view this ebook on an iPhone, iPad or Android mobile device you will need the Adobe Digital Editions app, or BlueFire Reader or Txtr app. These are free, too. For more details see this article.
|Size: ||19.4 MB|
|Publisher: ||Morgan Kaufmann|
|Date published: || 2014|
|ISBN: ||9780080510538 (DRM-PDF)|
|Read Aloud: ||not allowed|