Skip to content

Commit 76693d1

Browse files
committed
dimensionaliity reduction
1 parent d1ee040 commit 76693d1

File tree

6 files changed

+34
-1
lines changed

6 files changed

+34
-1
lines changed

Diff for: README.md

+1
Original file line numberDiff line numberDiff line change
@@ -87,6 +87,7 @@ Excerpts from the [Foreword](./docs/foreword_ro.pdf) and [Preface](./docs/prefac
8787
- [Does regularization in logistic regression always results in better fit and better generalization?](./faq/regularized-logistic-regression-performance.md)
8888
- [How was classification, as a learning machine, developed?](./faq/classifier-history.md)
8989
- [What are some good books/papers for learning deep learning?](./faq/deep-learning-resources.md)
90+
- [What are the different dimensionality reduction methods in machine learning?](./faq/dimensionality-reduction.md)
9091

9192
### Questions about the Book
9293

Diff for: faq/README.md

+2-1
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,8 @@ Sebastian
2828
- [What factors should I consider when choosing a predictive model technique?](./choosing-technique.md)
2929
- [Does regularization in logistic regression always results in better fit and better generalization?](./regularized-logistic-regression-performance.md)
3030
- [How was classification, as a learning machine, developed?](./classifier-history.md)
31-
- [What are some good books/papers for learning deep learning?](./deep-learning-resources.md)
31+
- [What are some good books/papers for learning deep learning?](./deep-learning-resources.md)
32+
- [What are the different dimensionality reduction methods in machine learning?](./dimensionality-reduction.md)
3233

3334
### Questions about the Book
3435

Diff for: faq/dimensionality-reduction.md

+31
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,31 @@
1+
# What are the different dimensionality reduction methods in machine learning?
2+
3+
Since there are so many different approaches, let's break it down to "feature selection" and "feature extraction."
4+
5+
Some examples of feature selection:
6+
7+
- L1 regularization (e.g., Logistic regression) and sparsity
8+
- variance thresholds
9+
- recursive feature elimination based on the weights of linear models
10+
- random forests / extra trees and feature importance (calculated as average information gain)
11+
- sequential forward/backward selection
12+
- genetic algorithms
13+
- exhaustive search
14+
15+
Some examples of feature extraction:
16+
17+
- Principal Component Analysis (PCA), unsupervised, returns axes of maximal variance given the constraint that those axes are orthogonal to each other
18+
- Linear Discriminant Analysis (LDA; not to be confused with Latent Dirichlett Allocation), supervised, returns axes that maximizes class separability (same constraint that axes are also orthogonal); and another article: Linear Discriminant Analysis bit by bit
19+
- kernel PCA: uses kernel trick to transform non-linear data to a feature space were samples may be linearly separable (in contrast, LDA and PCA are linear transformation techniques
20+
- supervised PCA
21+
- and many more non-linear transformation techniques, which you can find nicely summarized here: [Nonlinear dimensionality reduction](https://en.wikipedia.org/wiki/Nonlinear_dimensionality_reduction)
22+
23+
** So, which technique should we use? **
24+
25+
This also follows the "No Lunch Theorem" principle in some sense: there is no method that is always superior; it depends on your dataset. Intuitively, LDA would make more sense than PCA if you have a linear classification task, but empirical studies showed that it is not always the case. Although kernel PCA can separate concentric circles, it fails to unfold the Swiss Rroll, for example; here, locally linear embedding (LLE) would be more appropriate.
26+
27+
![](./dimensionality-reduction/swiss-roll.png)
28+
29+
![](./dimensionality-reduction/rbf-kpca.png)
30+
31+
![](./dimensionality-reduction/lle.png)

Diff for: faq/dimensionality-reduction/lle.png

82 KB
Loading

Diff for: faq/dimensionality-reduction/rbf-kpca.png

56 KB
Loading

Diff for: faq/dimensionality-reduction/swiss-roll.png

73.9 KB
Loading

0 commit comments

Comments
 (0)