Skip to main content

2023 | Buch

Improving Classifier Generalization

Real-Time Machine Learning based Applications

verfasst von: Rahul Kumar Sevakula, Nishchal K. Verma

Verlag: Springer Nature Singapore

Buchreihe : Studies in Computational Intelligence

insite
SUCHEN

Über dieses Buch

This book elaborately discusses techniques commonly used to improve generalization performance in classification approaches. The contents highlight methods to improve classification performance in numerous case studies: ranging from datasets of UCI repository to predictive maintenance problems and cancer classification problems. The book specifically provides a detailed tutorial on how to approach time-series classification problems and discusses two real time case studies on condition monitoring. In addition to describing the various aspects a data scientist must consider before finalizing their approach to a classification problem and reviewing the state of the art for improving classification generalization performance, it also discusses in detail the authors own contributions to the field, including MVPC - a classifier with very low VC dimension, a graphical indices based framework for reliable predictive maintenance and a novel general-purpose membership functions for Fuzzy Support Vector Machine which provides state of the art performance with noisy datasets, and a novel scheme to introduce deep learning in Fuzzy Rule based classifiers (FRCs). This volume will serve as a useful reference for researchers and students working on machine learning, health monitoring, predictive maintenance, time-series analysis, gene-expression data classification.

Inhaltsverzeichnis

Frontmatter
Chapter 1. Introduction
Abstract
With the advent of Cloud Computing, Big Data, and Internet of Things, the twenty-first century can be fairly seen as an age of data and unearthing patterns from them. Everywhere, the move is toward extracting information from data and embedding intelligence in devices/applications like smartphones, security systems, home appliances, etc. Such advances indeed help us to make our lives easier, more efficient, and resourceful.
Rahul Kumar Sevakula, Nishchal K. Verma
Chapter 2. Methods Used to Improve Generalization Performance
Abstract
Historically, the intuition behind developing classification algorithms was often to identify a hypothesis that minimizes training error. A major problem encountered with this approach is overfitting, which occurs when the hypothesis becomes too complex in comparison to the size of training data. In such cases, it is likely that an algorithm minimizing training error will find a hypothesis that fits the training data very well, but generalizes poorly to previously unseen data. Good generalization here refers to low generalization error which is defined as the difference between the training error and true error.
Rahul Kumar Sevakula, Nishchal K. Verma
Chapter 3. MVPC—A Classifier with Very Low VC Dimension
Abstract
In real-world problems, patterns found within the training data can be different from those found in the test data (dataset shift); obtaining reliable results with handcrafted (low-level) features becomes difficult even with linear kernel SVM. Such problems necessitate the creation of classifiers with low variance and high generalization as a driving function. In this chapter, we show the variance of a class of majority vote classifiers named Majority Vote Point (MVP) classifier, to be lower than that of linear classifiers, on account of lower VC dimension.
Rahul Kumar Sevakula, Nishchal K. Verma
Chapter 4. Framework for Reliable Fault Detection with Sensor Data
Abstract
This chapter presents a real-time health monitoring setup for condition-based maintenance of rotary machines to identify machine states using acoustic and vibrational signals. Diagnosis of faults is primarily done by analyzing changes in signals. The chapter specifically makes contributions in procedures pertaining to (1) organizing the data recordings, (2) data pre-processing, (3) feature selection, and (4) identification of sensitive position(s) for a pair of classes. The contributions are generic in nature and can be used for many other applications.
Rahul Kumar Sevakula, Nishchal K. Verma
Chapter 5. Membership Functions for Fuzzy Support Vector Machine in a Noisy Environment
Abstract
This chapter introduces the compounding of General Purpose Membership Functions (GPMFs) for overcoming class noise in a Fuzzy Support Vector Machine (FSVM). Traditional membership functions (MFs) characterize all samples of the class with a single MF.
Rahul Kumar Sevakula, Nishchal K. Verma
Chapter 6. Stacked Denoising Sparse Autoencoder-Based Fuzzy Rule Classifiers
Abstract
With time, machine learning experts are unanimously agreeing that finding good features is one of the most important problems in pattern classification [1]. In fact, if given features are good, even a linear classifier would suffice to give excellent classification results.
Rahul Kumar Sevakula, Nishchal K. Verma
Chapter 7. Epilogue
Abstract
In this chapter, we shall briefly discuss developments and use of autoencoders and transfer learning to obtain good generalization performance in tumor classification and time-series classification problems. Furthermore, we shall analyze the research work presented in the earlier chapters to discuss possible directions for future work. The chapter shall also conclude this monograph.
Rahul Kumar Sevakula, Nishchal K. Verma
Backmatter
Metadaten
Titel
Improving Classifier Generalization
verfasst von
Rahul Kumar Sevakula
Nishchal K. Verma
Copyright-Jahr
2023
Verlag
Springer Nature Singapore
Electronic ISBN
978-981-19-5073-5
Print ISBN
978-981-19-5072-8
DOI
https://doi.org/10.1007/978-981-19-5073-5

Premium Partner