Amazon cover image
Image from Amazon.com
Image from Google Jackets

A probabilistic theory of pattern recognition / Luc Devroye, László Györfi, Gábor Lugosi.

By: Contributor(s): Material type: TextTextSeries: Applications of mathematics ; 31.Publication details: New Delhi : Springer, CBS Publishers, 2014, c1996.Description: xv, 636 p. : ill. ; 25 cmISBN:
  • 9788132214977
Subject(s): DDC classification:
  • 003.52015192 22 DEV
LOC classification:
  • Q327 .D5 1996
Contents:
1. Introduction -- 2. The Bayes Error -- 3. Inequalities and Alternate Distance Measures -- 4. Linear Discrimination -- 5. Nearest Neighbor Rules -- 6. Consistency -- 7. Slow Rates of Convergence -- 8. Error Estimation -- 9. The Regular Histogram Rule -- 10. Kernel Rules -- 11. Consistency of the k-Nearest Neighbor Rule -- 12. Vapnik-Chervonenkis Theory -- 13. Combinatorial Aspects of Vapnik-Chervonenkis Theory -- 14. Lower Bounds for Empirical Classifier Selection -- 15. The Maximum Likelihood Principle -- 16. Parametric Classification -- 17. Generalized Linear Discrimination -- 18. Complexity Regularization -- 19. Condensed and Edited Nearest Neighbor Rules -- 20. Tree Classifiers -- 21. Data-Dependent Partitioning -- 22. Splitting the Data -- 23. The Resubstitution Estimate -- 24. Deleted Estimates of the Error Probability -- 25. Automatic Kernel Rules -- 26. Automatic Nearest Neighbor Rules -- 27. Hypercubes and Discrete Spaces --
28. Epsilon Entropy and Totally Bounded Sets -- 29. Uniform Laws of Large Numbers -- 30. Neural Networks -- 31. Other Error Estimates -- 32. Feature Extraction.
Summary: Pattern recognition presents one of the most significant challenges for scientists and engineers, and many different approaches have been proposed. The aim of this book is to provide a self-contained account of probabilistic analysis of these approaches. The book includes a discussion of distance measures, nonparametric methods based on kernels or nearest neighbors, Vapnik-Chervonenkis theory, epsilon entropy, parametric classification, error estimation, tree classifiers, and neural networks.Summary: Wherever possible, distribution-free properties and inequalities are derived. A substantial portion of the results or the analysis is new. Over 430 problems and exercises complement the material.
Tags from this library: No tags from this library for this title. Log in to add tags.
Holdings
Item type Current library Collection Call number Status Date due Barcode Item holds
Books Books Learning Resource Centre Reserve Books 003.52015192 DEV (Browse shelf(Opens below)) Not for loan 9801
Books Books Learning Resource Centre 003.52015192 DEV (Browse shelf(Opens below)) Available 9800
Books Books Learning Resource Centre 003.52015192 DEV (Browse shelf(Opens below)) Available 7503
Total holds: 0

Includes bibliographical references (p. [593]-618) and indexes.

1. Introduction -- 2. The Bayes Error -- 3. Inequalities and Alternate Distance Measures -- 4. Linear Discrimination -- 5. Nearest Neighbor Rules -- 6. Consistency -- 7. Slow Rates of Convergence -- 8. Error Estimation -- 9. The Regular Histogram Rule -- 10. Kernel Rules -- 11. Consistency of the k-Nearest Neighbor Rule -- 12. Vapnik-Chervonenkis Theory -- 13. Combinatorial Aspects of Vapnik-Chervonenkis Theory -- 14. Lower Bounds for Empirical Classifier Selection -- 15. The Maximum Likelihood Principle -- 16. Parametric Classification -- 17. Generalized Linear Discrimination -- 18. Complexity Regularization -- 19. Condensed and Edited Nearest Neighbor Rules -- 20. Tree Classifiers -- 21. Data-Dependent Partitioning -- 22. Splitting the Data -- 23. The Resubstitution Estimate -- 24. Deleted Estimates of the Error Probability -- 25. Automatic Kernel Rules -- 26. Automatic Nearest Neighbor Rules -- 27. Hypercubes and Discrete Spaces --

28. Epsilon Entropy and Totally Bounded Sets -- 29. Uniform Laws of Large Numbers -- 30. Neural Networks -- 31. Other Error Estimates -- 32. Feature Extraction.

Pattern recognition presents one of the most significant challenges for scientists and engineers, and many different approaches have been proposed. The aim of this book is to provide a self-contained account of probabilistic analysis of these approaches. The book includes a discussion of distance measures, nonparametric methods based on kernels or nearest neighbors, Vapnik-Chervonenkis theory, epsilon entropy, parametric classification, error estimation, tree classifiers, and neural networks.

Wherever possible, distribution-free properties and inequalities are derived. A substantial portion of the results or the analysis is new. Over 430 problems and exercises complement the material.

There are no comments on this title.

to post a comment.
Powered by Koha & maintained by LRC, JK Lakshmipat University, Jaipur
Contact: [email protected]
Copyright © 2022 LRC, JK Lakshmipat University, Jaipur. All Rights Reserved.