Skip to main content
main-content
Top

Hint

Swipe to navigate through the articles of this issue

09-11-2019 | Methodologies and Application | Issue 13/2020 Open Access

Soft Computing 13/2020

An ensemble based on neural networks with random weights for online data stream regression

Journal:
Soft Computing > Issue 13/2020
Authors:
Ricardo de Almeida, Yee Mey Goh, Radmehr Monfared, Maria Teresinha Arns Steiner, Andrew West
Important notes
Communicated by V. Loia.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Abstract

Most information sources in the current technological world are generating data sequentially and rapidly, in the form of data streams. The evolving nature of processes may often cause changes in data distribution, also known as concept drift, which is difficult to detect and causes loss of accuracy in supervised learning algorithms. As a consequence, online machine learning algorithms that are able to update actively according to possible changes in the data distribution are required. Although many strategies have been developed to tackle this problem, most of them are designed for classification problems. Therefore, in the domain of regression problems, there is a need for the development of accurate algorithms with dynamic updating mechanisms that can operate in a computational time compatible with today’s demanding market. In this article, the authors propose a new bagging ensemble approach based on neural network with random weights for online data stream regression. The proposed method improves the data prediction accuracy as well as minimises the required computational time compared to a recent algorithm for online data stream regression from literature. The experiments are carried out using four synthetic datasets to evaluate the algorithm’s response to concept drift, along with four benchmark datasets from different industries. The results indicate improvement in data prediction accuracy, effectiveness in handling concept drift, and much faster updating times compared to the existing available approach. Additionally, the use of design of experiments as an effective tool for hyperparameter tuning is demonstrated.
Literature
About this article

Other articles of this Issue 13/2020

Soft Computing 13/2020 Go to the issue

Premium Partner

    Image Credits