An Introduction to Computational Learning Theory

An Introduction to Computational Learning Theory
Author :
Publisher : MIT Press
Total Pages : 230
Release :
ISBN-10 : 0262111934
ISBN-13 : 9780262111935
Rating : 4/5 (34 Downloads)

Synopsis An Introduction to Computational Learning Theory by : Michael J. Kearns

Emphasizing issues of computational efficiency, Michael Kearns and Umesh Vazirani introduce a number of central topics in computational learning theory for researchers and students in artificial intelligence, neural networks, theoretical computer science, and statistics. Emphasizing issues of computational efficiency, Michael Kearns and Umesh Vazirani introduce a number of central topics in computational learning theory for researchers and students in artificial intelligence, neural networks, theoretical computer science, and statistics. Computational learning theory is a new and rapidly expanding area of research that examines formal models of induction with the goals of discovering the common methods underlying efficient learning algorithms and identifying the computational impediments to learning. Each topic in the book has been chosen to elucidate a general principle, which is explored in a precise formal setting. Intuition has been emphasized in the presentation to make the material accessible to the nontheoretician while still providing precise arguments for the specialist. This balance is the result of new proofs of established theorems, and new presentations of the standard proofs. The topics covered include the motivation, definitions, and fundamental results, both positive and negative, for the widely studied L. G. Valiant model of Probably Approximately Correct Learning; Occam's Razor, which formalizes a relationship between learning and data compression; the Vapnik-Chervonenkis dimension; the equivalence of weak and strong learning; efficient learning in the presence of noise by the method of statistical queries; relationships between learning and cryptography, and the resulting computational limitations on efficient learning; reducibility between learning problems; and algorithms for learning finite automata from active experimentation.

Boosting

Boosting
Author :
Publisher : MIT Press
Total Pages : 544
Release :
ISBN-10 : 9780262526036
ISBN-13 : 0262526034
Rating : 4/5 (36 Downloads)

Synopsis Boosting by : Robert E. Schapire

An accessible introduction and essential reference for an approach to machine learning that creates highly accurate prediction rules by combining many weak and inaccurate ones. Boosting is an approach to machine learning based on the idea of creating a highly accurate predictor by combining many weak and inaccurate “rules of thumb.” A remarkably rich theory has evolved around boosting, with connections to a range of topics, including statistics, game theory, convex optimization, and information geometry. Boosting algorithms have also enjoyed practical success in such fields as biology, vision, and speech processing. At various times in its history, boosting has been perceived as mysterious, controversial, even paradoxical. This book, written by the inventors of the method, brings together, organizes, simplifies, and substantially extends two decades of research on boosting, presenting both theory and applications in a way that is accessible to readers from diverse backgrounds while also providing an authoritative reference for advanced researchers. With its introductory treatment of all material and its inclusion of exercises in every chapter, the book is appropriate for course use as well. The book begins with a general introduction to machine learning algorithms and their analysis; then explores the core theory of boosting, especially its ability to generalize; examines some of the myriad other theoretical viewpoints that help to explain and understand boosting; provides practical extensions of boosting for more complex learning problems; and finally presents a number of advanced theoretical topics. Numerous applications and practical illustrations are offered throughout.

Understanding Machine Learning

Understanding Machine Learning
Author :
Publisher : Cambridge University Press
Total Pages : 415
Release :
ISBN-10 : 9781107057135
ISBN-13 : 1107057132
Rating : 4/5 (35 Downloads)

Synopsis Understanding Machine Learning by : Shai Shalev-Shwartz

Introduces machine learning and its algorithmic paradigms, explaining the principles behind automated learning approaches and the considerations underlying their usage.

Computational Learning Theory and Natural Learning Systems: Intersections between theory and experiment

Computational Learning Theory and Natural Learning Systems: Intersections between theory and experiment
Author :
Publisher : Mit Press
Total Pages : 449
Release :
ISBN-10 : 0262581337
ISBN-13 : 9780262581332
Rating : 4/5 (37 Downloads)

Synopsis Computational Learning Theory and Natural Learning Systems: Intersections between theory and experiment by : Stephen José Hanson

Annotation These original contributions converge on an exciting and fruitful intersection of three historically distinct areas of learning research: computational learning theory, neural networks, and symbolic machine learning. Bridging theory and practice, computer science and psychology, they consider general issues in learning systems that could provide constraints for theory and at the same time interpret theoretical results in the context of experiments with actual learning systems. In all, nineteen chapters address questions such as, What is a natural system? How should learning systems gain from prior knowledge? If prior knowledge is important, how can we quantify how important? What makes a learning problem hard? How are neural networks and symbolic machine learning approaches similar? Is there a fundamental difference in the kind of task a neural network can easily solve as opposed to those a symbolic algorithm can easily solve? Stephen J. Hanson heads the Learning Systems Department at Siemens Corporate Research and is a Visiting Member of the Research Staff and Research Collaborator at the Cognitive Science Laboratory at Princeton University. George A. Drastal is Senior Research Scientist at Siemens Corporate Research. Ronald J. Rivest is Professor of Computer Science and Associate Director of the Laboratory for Computer Science at the Massachusetts Institute of Technology.

Learning Theory

Learning Theory
Author :
Publisher : Cambridge University Press
Total Pages :
Release :
ISBN-10 : 9781139462860
ISBN-13 : 1139462865
Rating : 4/5 (60 Downloads)

Synopsis Learning Theory by : Felipe Cucker

The goal of learning theory is to approximate a function from sample values. To attain this goal learning theory draws on a variety of diverse subjects, specifically statistics, approximation theory, and algorithmics. Ideas from all these areas blended to form a subject whose many successful applications have triggered a rapid growth during the last two decades. This is the first book to give a general overview of the theoretical foundations of the subject emphasizing the approximation theory, while still giving a balanced overview. It is based on courses taught by the authors, and is reasonably self-contained so will appeal to a broad spectrum of researchers in learning theory and adjacent fields. It will also serve as an introduction for graduate students and others entering the field, who wish to see how the problems raised in learning theory relate to other disciplines.

The Principles of Deep Learning Theory

The Principles of Deep Learning Theory
Author :
Publisher : Cambridge University Press
Total Pages : 473
Release :
ISBN-10 : 9781316519332
ISBN-13 : 1316519333
Rating : 4/5 (32 Downloads)

Synopsis The Principles of Deep Learning Theory by : Daniel A. Roberts

This volume develops an effective theory approach to understanding deep neural networks of practical relevance.

Foundations of Machine Learning, second edition

Foundations of Machine Learning, second edition
Author :
Publisher : MIT Press
Total Pages : 505
Release :
ISBN-10 : 9780262351362
ISBN-13 : 0262351366
Rating : 4/5 (62 Downloads)

Synopsis Foundations of Machine Learning, second edition by : Mehryar Mohri

A new edition of a graduate-level machine learning textbook that focuses on the analysis and theory of algorithms. This book is a general introduction to machine learning that can serve as a textbook for graduate students and a reference for researchers. It covers fundamental modern topics in machine learning while providing the theoretical basis and conceptual tools needed for the discussion and justification of algorithms. It also describes several key aspects of the application of these algorithms. The authors aim to present novel theoretical tools and concepts while giving concise proofs even for relatively advanced topics. Foundations of Machine Learning is unique in its focus on the analysis and theory of algorithms. The first four chapters lay the theoretical foundation for what follows; subsequent chapters are mostly self-contained. Topics covered include the Probably Approximately Correct (PAC) learning framework; generalization bounds based on Rademacher complexity and VC-dimension; Support Vector Machines (SVMs); kernel methods; boosting; on-line learning; multi-class classification; ranking; regression; algorithmic stability; dimensionality reduction; learning automata and languages; and reinforcement learning. Each chapter ends with a set of exercises. Appendixes provide additional material including concise probability review. This second edition offers three new chapters, on model selection, maximum entropy models, and conditional entropy models. New material in the appendixes includes a major section on Fenchel duality, expanded coverage of concentration inequalities, and an entirely new entry on information theory. More than half of the exercises are new to this edition.

Algebraic Geometry and Statistical Learning Theory

Algebraic Geometry and Statistical Learning Theory
Author :
Publisher : Cambridge University Press
Total Pages : 295
Release :
ISBN-10 : 9780521864671
ISBN-13 : 0521864674
Rating : 4/5 (71 Downloads)

Synopsis Algebraic Geometry and Statistical Learning Theory by : Sumio Watanabe

Sure to be influential, Watanabe's book lays the foundations for the use of algebraic geometry in statistical learning theory. Many models/machines are singular: mixture models, neural networks, HMMs, Bayesian networks, stochastic context-free grammars are major examples. The theory achieved here underpins accurate estimation techniques in the presence of singularities.

A Probabilistic Theory of Pattern Recognition

A Probabilistic Theory of Pattern Recognition
Author :
Publisher : Springer Science & Business Media
Total Pages : 631
Release :
ISBN-10 : 9781461207115
ISBN-13 : 1461207118
Rating : 4/5 (15 Downloads)

Synopsis A Probabilistic Theory of Pattern Recognition by : Luc Devroye

A self-contained and coherent account of probabilistic techniques, covering: distance measures, kernel rules, nearest neighbour rules, Vapnik-Chervonenkis theory, parametric classification, and feature extraction. Each chapter concludes with problems and exercises to further the readers understanding. Both research workers and graduate students will benefit from this wide-ranging and up-to-date account of a fast- moving field.

Learning Theory and Kernel Machines

Learning Theory and Kernel Machines
Author :
Publisher : Springer
Total Pages : 761
Release :
ISBN-10 : 9783540451679
ISBN-13 : 3540451676
Rating : 4/5 (79 Downloads)

Synopsis Learning Theory and Kernel Machines by : Bernhard Schölkopf

This book constitutes the joint refereed proceedings of the 16th Annual Conference on Computational Learning Theory, COLT 2003, and the 7th Kernel Workshop, Kernel 2003, held in Washington, DC in August 2003. The 47 revised full papers presented together with 5 invited contributions and 8 open problem statements were carefully reviewed and selected from 92 submissions. The papers are organized in topical sections on kernel machines, statistical learning theory, online learning, other approaches, and inductive inference learning.