watch: AML | Feature Selection

Table of contents

Video 7 2021-10-04

[toc]

Dimensionality Reduction

  • Avoids the curse of Dimensionality

October 6, 2021

Curse of Dimensionality

  • When dimensionality increases, data becomes increasingly sparse
    1. Concepts become less meaningful: density and distance
    2. Subspace combinations grow very fast

Dimentionality Reduction

  • Eliminate irrelevant features and reduces noise

  • $X$ is a set of $N$ features: $X=\{X_1, X_2, \cdots X_N\}$,a reduced set $X'$ is a transformation of $X$ and consists of $d$ features so that $d $$ X' = T(X) = \{ X_1',\ X_2',\ \cdots,\ X_d'\} \\ T: \R^N \rightarrow \R^d,\ d

  • Avoids the curse of dimensionality. Reduces time and space required for computations.>

  • Two ways:

    1. Feature Extraction: transformation to a lower dimension
      • Wavelet transforms and PCA
    2. Feature Selection: transformation is limited to only selection from original features
      • Filters, Wrappers, Embedded.
  • 3 Features

    Feature Selection

    Assume a binary classification model, X → Model → Y ∈ {0, 1}, where X consists of N different features, e.g., age, weight, temperature, blood pressure, etc. X = {X1, X2, X3 , . . . , XN } N could be small, or relatively large value, e.g., an image of size of 300 × 300.

    Class Separation Criterion

    Feature Selection

    Search Strategies

    Objective Function

    Filters Approaches

    Wrapper Approches

    Best Individual Features

    General SFS and General SBS

    PLUS-L TAKE-R

    Sequential Floating Selection