Feature Extraction [electronic resource] : Foundations and Applications / edited by Isabelle Guyon, Masoud Nikravesh, Steve Gunn, Lotfi A. Zadeh.

By: Guyon, Isabelle [editor.]Contributor(s): Nikravesh, Masoud [editor.] | Gunn, Steve [editor.] | Zadeh, Lotfi A [editor.] | SpringerLink (Online service)Material type: TextTextLanguage: English Series: Studies in Fuzziness and Soft Computing: 207Publisher: Berlin, Heidelberg : Springer Berlin Heidelberg, 2006Description: XXIV, 778 p. With CD-ROM. Also available online. online resourceContent type: text Media type: computer Carrier type: online resourceISBN: 9783540354888Subject(s): Engineering | Artificial intelligence | Computer vision | Computer aided design | Mathematics | Engineering mathematics | Engineering | Appl.Mathematics/Computational Methods of Engineering | Artificial Intelligence (incl. Robotics) | Computer Imaging, Vision, Pattern Recognition and Graphics | Computer-Aided Engineering (CAD, CAE) and Design | Applications of Mathematics | Operations Research/Decision TheoryAdditional physical formats: Printed edition:: No titleDDC classification: 519 LOC classification: TA329-348TA640-643Online resources: Click here to access online
Contents:
An Introduction to Feature Extraction -- An Introduction to Feature Extraction -- Feature Extraction Fundamentals -- Learning Machines -- Assessment Methods -- Filter Methods -- Search Strategies -- Embedded Methods -- Information-Theoretic Methods -- Ensemble Learning -- Fuzzy Neural Networks -- Feature Selection Challenge -- Design and Analysis of the NIPS2003 Challenge -- High Dimensional Classification with Bayesian Neural Networks and Dirichlet Diffusion Trees -- Ensembles of Regularized Least Squares Classifiers for High-Dimensional Problems -- Combining SVMs with Various Feature Selection Strategies -- Feature Selection with Transductive Support Vector Machines -- Variable Selection using Correlation and Single Variable Classifier Methods: Applications -- Tree-Based Ensembles with Dynamic Soft Feature Selection -- Sparse, Flexible and Efficient Modeling using L 1 Regularization -- Margin Based Feature Selection and Infogain with Standard Classifiers -- Bayesian Support Vector Machines for Feature Ranking and Selection -- Nonlinear Feature Selection with the Potential Support Vector Machine -- Combining a Filter Method with SVMs -- Feature Selection via Sensitivity Analysis with Direct Kernel PLS -- Information Gain, Correlation and Support Vector Machines -- Mining for Complex Models Comprising Feature Selection and Classification -- Combining Information-Based Supervised and Unsupervised Feature Selection -- An Enhanced Selective Naïve Bayes Method with Optimal Discretization -- An Input Variable Importance Definition based on Empirical Data Probability Distribution -- New Perspectives in Feature Extraction -- Spectral Dimensionality Reduction -- Constructing Orthogonal Latent Features for Arbitrary Loss -- Large Margin Principles for Feature Selection -- Feature Extraction for Classification of Proteomic Mass Spectra: A Comparative Study -- Sequence Motifs: Highly Predictive Features of Protein Function.
In: Springer eBooksSummary: This book is both a reference for engineers and scientists and a teaching resource, featuring tutorial chapters and research papers on feature extraction. "This book compiles some very promising techniques, coming from an extremely smart collection of researchers, delivering their best ideas in a competitive environment." Trevor Hastie, Stanford University "Feature selection is a key technology for making sense of the high dimensional data. Isabelle Guyon et al. have done a splendid job in designing a challenging competition, and collecting the lessons learned." Bernhard Schoelkopf, Max Planck Institute "There has been until now insufficient consideration of feature selection algorithms, no unified presentation of leading methods, and no systematic comparisons. This volume is noteworthy for the breadth of methods covered, the clarity of presentations, the unity in notation and the helpful statistical appendices." David G. Stork, Ricoh Innovations "Feature extraction finds application in biotechnology, industrial inspection, the Internet, radar, sonar, and speech recognition. This book will make a difference to the literature on machine learning." Simon Haykin, Mc Master University "This book sets a high standard as the public record of an interesting and effective competition." Peter Norvig, Google Inc.
Tags from this library: No tags from this library for this title. Log in to add tags.
    Average rating: 0.0 (0 votes)
No physical items for this record

An Introduction to Feature Extraction -- An Introduction to Feature Extraction -- Feature Extraction Fundamentals -- Learning Machines -- Assessment Methods -- Filter Methods -- Search Strategies -- Embedded Methods -- Information-Theoretic Methods -- Ensemble Learning -- Fuzzy Neural Networks -- Feature Selection Challenge -- Design and Analysis of the NIPS2003 Challenge -- High Dimensional Classification with Bayesian Neural Networks and Dirichlet Diffusion Trees -- Ensembles of Regularized Least Squares Classifiers for High-Dimensional Problems -- Combining SVMs with Various Feature Selection Strategies -- Feature Selection with Transductive Support Vector Machines -- Variable Selection using Correlation and Single Variable Classifier Methods: Applications -- Tree-Based Ensembles with Dynamic Soft Feature Selection -- Sparse, Flexible and Efficient Modeling using L 1 Regularization -- Margin Based Feature Selection and Infogain with Standard Classifiers -- Bayesian Support Vector Machines for Feature Ranking and Selection -- Nonlinear Feature Selection with the Potential Support Vector Machine -- Combining a Filter Method with SVMs -- Feature Selection via Sensitivity Analysis with Direct Kernel PLS -- Information Gain, Correlation and Support Vector Machines -- Mining for Complex Models Comprising Feature Selection and Classification -- Combining Information-Based Supervised and Unsupervised Feature Selection -- An Enhanced Selective Naïve Bayes Method with Optimal Discretization -- An Input Variable Importance Definition based on Empirical Data Probability Distribution -- New Perspectives in Feature Extraction -- Spectral Dimensionality Reduction -- Constructing Orthogonal Latent Features for Arbitrary Loss -- Large Margin Principles for Feature Selection -- Feature Extraction for Classification of Proteomic Mass Spectra: A Comparative Study -- Sequence Motifs: Highly Predictive Features of Protein Function.

This book is both a reference for engineers and scientists and a teaching resource, featuring tutorial chapters and research papers on feature extraction. "This book compiles some very promising techniques, coming from an extremely smart collection of researchers, delivering their best ideas in a competitive environment." Trevor Hastie, Stanford University "Feature selection is a key technology for making sense of the high dimensional data. Isabelle Guyon et al. have done a splendid job in designing a challenging competition, and collecting the lessons learned." Bernhard Schoelkopf, Max Planck Institute "There has been until now insufficient consideration of feature selection algorithms, no unified presentation of leading methods, and no systematic comparisons. This volume is noteworthy for the breadth of methods covered, the clarity of presentations, the unity in notation and the helpful statistical appendices." David G. Stork, Ricoh Innovations "Feature extraction finds application in biotechnology, industrial inspection, the Internet, radar, sonar, and speech recognition. This book will make a difference to the literature on machine learning." Simon Haykin, Mc Master University "This book sets a high standard as the public record of an interesting and effective competition." Peter Norvig, Google Inc.

There are no comments on this title.

to post a comment.

Implemented and Maintained by Biju Patnaik Central Library.
For any Suggestions/Query Contact to library or Email: library@nitrkl.ac.in OR bpcl-cir@nitrkl.ac.in. Ph:91+6612462103
Website/OPAC best viewed in Mozilla Browser in 1366X768 Resolution.

Powered by Koha