askvity

What is Correlation-Based Feature Selection (CFS)?

Published in Feature Selection 4 mins read

Correlation-Based Feature Selection (CFS) is a powerful filter-based feature selection technique that identifies and selects optimal subsets of features for machine learning models by assessing their relevance and redundancy.

Core Principle of Correlation-Based Feature Selection (CFS)

At its heart, CFS operates on the principle that good feature subsets are those that contain features highly correlated with the target variable but have low correlation with each other. This balance is crucial for building robust and accurate predictive models. It aims to find a subset of features that are strongly predictive of the outcome while minimizing redundancy among the selected features.

The "Good Feature Subset" Hypothesis

The fundamental hypothesis underpinning CFS is:

  • Relevance: Features should have a strong predictive relationship with the class (target variable).
  • Redundancy: Features should have a low inter-correlation with each other, meaning they provide unique information rather than duplicating insights.

Why Correlation-Based Feature Selection Matters

In the realm of machine learning, dealing with high-dimensional datasets is common. Too many features can lead to:

  • Overfitting: Models learn noise from irrelevant features, performing poorly on new data.
  • Increased Computational Cost: Training models becomes slower and more resource-intensive.
  • Reduced Model Interpretability: Understanding the model's decisions becomes challenging.

CFS addresses these issues by reducing the feature space, leading to:

  • Improved model accuracy.
  • Faster training times.
  • Enhanced model generalization.
  • Simplified model interpretation.

How Correlation-Based Feature Selection Works

CFS typically involves a search algorithm that explores different feature subsets and an evaluation function that assesses the "merit" of each subset based on its correlations.

  1. Correlation Measurement: It calculates the correlation between each feature and the target variable (feature-class correlation) and the correlation between all pairs of features (feature-feature correlation). Common correlation metrics include Pearson correlation (for linear relationships), Spearman correlation (for monotonic relationships), or mutual information (for non-linear dependencies).

  2. Subset Evaluation: The evaluation function quantifies the "goodness" of a feature subset using a heuristic often formulated as:

    $Merit = \frac{k \cdot \bar{r}{cf}}{\sqrt{k + k(k-1)\bar{r}{ff}}}$

    Where:

    • $k$ is the number of features in the subset.
    • $\bar{r}_{cf}$ is the average correlation between features in the subset and the target variable. (High desired)
    • $\bar{r}_{ff}$ is the average inter-correlation between features within the subset. (Low desired)

    A higher merit indicates a better subset, striking a balance between relevance and non-redundancy.

  3. Search Strategy: A search algorithm (e.g., best-first search, genetic algorithm, or greedy stepwise search) is used to navigate the vast space of possible feature subsets, aiming to find the one with the highest merit.

Practical Example

Consider a dataset for predicting house prices with features like size, number_of_bedrooms, number_of_bathrooms, and living_area.

Feature Correlation with Target (Price) Correlation with Other Features (Example) Desired Outcome (CFS)
size High Moderate with living_area Likely to be selected
number_of_bedrooms Medium Low with size Likely to be selected
number_of_bathrooms Medium Low with number_of_bedrooms Likely to be selected
living_area High High with size May be excluded if size is similar

In this scenario, CFS would likely select size, number_of_bedrooms, and number_of_bathrooms, potentially excluding living_area if it's highly correlated with size and doesn't provide significantly new information for predicting price.

Benefits of CFS

  • Effectiveness: Proven to select robust feature subsets that improve model performance.
  • Efficiency: As a filter method, it's generally faster than wrapper methods because it evaluates features independently of the learning algorithm.
  • Reduced Overfitting: By removing redundant and irrelevant features, it helps models generalize better to unseen data.
  • Interpretability: A smaller, non-redundant feature set makes the resulting model easier to understand.

Challenges and Considerations

  • Computation for Large Datasets: While efficient, searching through all possible feature subsets for very large datasets can still be computationally intensive. Heuristic search algorithms are employed to mitigate this.
  • Correlation Limitations: Standard correlation measures might not capture complex, non-linear relationships as effectively as some other feature selection methods.
  • Threshold Selection: The choice of correlation thresholds or the specific evaluation function can influence the selected features.

Related Articles