Feature interval learning algorithms for classification

Date
2010
Authors
Dayanik, A.
Editor(s)
Advisor
Supervisor
Co-Advisor
Co-Supervisor
Instructor
Source Title
Knowledge-Based Systems
Print ISSN
0950-7051
Electronic ISSN
Publisher
Elsevier BV
Volume
23
Issue
5
Pages
402 - 417
Language
English
Journal Title
Journal ISSN
Volume Title
Series
Abstract

This paper presents Feature Interval Learning algorithms (FIL) which represent multi-concept descriptions in the form of disjoint feature intervals. The FIL algorithms are batch supervised inductive learning algorithms and use feature projections of the training instances to represent induced classification knowledge. The concept description is learned separately for each feature and is in the form of a set of disjoint intervals. The class of an unseen instance is determined by the weighted-majority voting of the feature predictions. The basic FIL algorithm is enhanced with adaptive interval and feature weight schemes in order to handle noisy and irrelevant features. The algorithms are empirically evaluated on twelve data sets from the UCI repository and are compared with k-NN, k-NNFP, and NBC classification algorithms. The experiments demonstrate that the FIL algorithms are robust to irrelevant features and missing feature values, achieve accuracy comparable to the best of the existing algorithms with significantly less average running times. © 2010 Elsevier B.V. All rights reserved.

Course
Other identifiers
Book Title
Citation
Published Version (Please cite this version)