Name: | Description: | Size: | Format: | |
---|---|---|---|---|
953.13 KB | Adobe PDF |
Advisor(s)
Abstract(s)
Discrete data representations are necessary, or at least convenient, in many machine learning problems. While feature selection (FS) techniques aim at finding relevant subsets of features, the goal of feature discretization (FD) is to find concise (quantized) data representations, adequate for the learning task at hand. In this paper, we propose two incremental methods for FD. The first method belongs to the filter family, in which the quality of the discretization is assessed by a (supervised or unsupervised) relevance criterion. The second method is a wrapper, where discretized features are assessed using a classifier. Both methods can be coupled with any static (unsupervised or supervised) discretization procedure and can be used to perform FS as pre-processing or post-processing stages. The proposed methods attain efficient representations suitable for binary and multi-class problems with different types of data, being competitive with existing methods. Moreover, using well-known FS methods with the features discretized by our techniques leads to better accuracy than with the features discretized by other methods or with the original features. (C) 2013 Elsevier B.V. All rights reserved.
Description
Keywords
Feature discretization Static discretization Incremental discretization Filter Wrapper Feature selection
Citation
FERREIRA, Artur Jorge; FIGUEIREDO, Mário Alexandre Teles de – Incremental filter and wrapper approaches for features discretization. In Neurocomputing. Amsterdam : Elsevier Science BV, 2014. ISSN: 0925-2312. Vol. 123, p. 60-74.
Publisher
Elsevier Science BV