A broad ensemble learning system for drifting stream classification

buir.contributor.authorBakhshi, Sepehr
buir.contributor.authorGhahramanian, Pouya
buir.contributor.authorCan, Fazlı
buir.contributor.orcidBakhshi, Sepehr|0000-0003-2292-6130
buir.contributor.orcidGhahramanian, Pouya|0000-0003-3479-8842
buir.contributor.orcidCan, Fazlı|0000-0003-0016-4278
dc.citation.epage89330en_US
dc.citation.spage89315
dc.citation.volumeNumber11
dc.contributor.authorBakhshi, Sepehr
dc.contributor.authorGhahramanian, Pouya
dc.contributor.authorBonab, H.
dc.contributor.authorCan, Fazlı
dc.date.accessioned2024-03-15T07:48:40Z
dc.date.available2024-03-15T07:48:40Z
dc.date.issued2023-08-21
dc.departmentDepartment of Computer Engineering
dc.description.abstractIn a data stream environment, classification models must effectively and efficiently handle concept drift. Ensemble methods are widely used for this purpose; however, the ones available in the literature either use a large data chunk to update the model or learn the data one by one. In the former, the model may miss the changes in the data distribution, while in the latter, the model may suffer from inefficiency and instability. To address these issues, we introduce a novel ensemble approach based on the Broad Learning System (BLS), where mini chunks are used at each update. BLS is an effective lightweight neural architecture recently developed for incremental learning. Although it is fast, it requires huge data chunks for effective updates and is unable to handle dynamic changes observed in data streams. Our proposed approach, named Broad Ensemble Learning System (BELS), uses a novel updating method that significantly improves best-in class model accuracy. It employs an ensemble of output layers to address the limitations of BLS and handle drifts. Our model tracks the changes in the accuracy of the ensemble components and reacts to these changes. We present our mathematical derivation of BELS, perform comprehensive experiments with 35 datasets that demonstrate the adaptability of our model to various drift types, and provide its hyperparameter, ablation, and imbalanced dataset performance analysis. The experimental results show that the proposed approach outperforms 10 state-of-the-art baselines, and supplies an overall improvement of 18.59% in terms of average prequential accuracy.
dc.identifier.doi10.1109/ACCESS.2023.3306957en_US
dc.identifier.eissn2169-3536en_US
dc.identifier.urihttps://hdl.handle.net/11693/114782en_US
dc.language.isoEnglishen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.relation.isversionofhttps://dx.doi.org/10.1109/ACCESS.2023.3306957
dc.source.titleIEEE Access
dc.subjectData stream mining
dc.subjectConcept drift
dc.subjectEnsemble learning
dc.subjectNeural networks
dc.subjectBig data
dc.titleA broad ensemble learning system for drifting stream classification
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
A_broad_ensemble_learning_system_for_drifting_stream_classification.pdf
Size:
2.76 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.01 KB
Format:
Item-specific license agreed upon to submission
Description: