Relief-Based Feature Selection: Introduction and Review

11/22/2017
by   Ryan J. Urbanowicz, et al.
0

Feature selection plays a critical role in data mining, driven by increasing feature dimensionality in target problems and growing interest in advanced but computationally expensive methodologies able to model complex associations. Specifically, there is a need for feature selection methods that are computationally efficient, yet sensitive to complex patterns of association, e.g. interactions, so that informative features are not mistakenly eliminated prior to downstream modeling. This paper focuses on Relief-based algorithms (RBAs), a unique family of filter-style feature selection algorithms that strike an effective balance between these objectives while flexibly adapting to various data characteristics, e.g. classification vs. regression. First, this work broadly examines types of feature selection and defines RBAs within that context. Next, we introduce the original Relief algorithm and associated concepts, emphasizing the intuition behind how it works, how feature weights generated by the algorithm can be interpreted, and why it is sensitive to feature interactions without evaluating combinations of features. Lastly, we include an expansive review of RBA methodological research beyond Relief and its popular descendant, ReliefF. In particular, we characterize branches of RBA research, and provide comparative summaries of RBA algorithms including contributions, strategies, functionality, time complexity, adaptation to key data characteristics, and software availability.

READ FULL TEXT

page 2

page 15

page 18

page 20

research
08/07/2020

Review of Swarm Intelligence-based Feature Selection Methods

In the past decades, the rapid growth of computer and database technolog...
research
06/28/2023

Feature Selection: A perspective on inter-attribute cooperation

High-dimensional datasets depict a challenge for learning tasks in data ...
research
11/30/2022

Universal Feature Selection Tool (UniFeat): An Open-Source Tool for Dimensionality Reduction

The Universal Feature Selection Tool (UniFeat) is an open-source tool de...
research
05/07/2015

Integrating K-means with Quadratic Programming Feature Selection

Several data mining problems are characterized by data in high dimension...
research
10/27/2021

Feature selection revisited in the single-cell era

Feature selection techniques are essential for high-dimensional data ana...
research
10/22/2019

Orthogonal variance decomposition based feature selection

Existing feature selection methods fail to properly account for interact...
research
08/22/2019

Applications of Nature-Inspired Algorithms for Dimension Reduction: Enabling Efficient Data Analytics

In [1], we have explored the theoretical aspects of feature selection an...

Please sign up or login with your details

Forgot password? Click here to reset