Unsupervised 1 - traditional
Introduction
Supervised Learning = you have label.
Unsupervised Learning = you don't have label.
PCA / t-SNE = dimension reduction (faster, reduce noise)
We can use Unsupervised Learning to improve Supervised Learning.
PCA
Z = XQ, Q is a matrix, then the direction is changed.
Measure information -> variation
A deterministic variable = 0 variation. We want to transform Z to have most info in 1st column, 2nd most info in 2nd column, etc.
De-correlation
Z as latent variable
AV = LV
eignenvector(V) direction is not changed by matrix A.
eignenvalues(L) a scale factor.
t-SNE
t-SNE = t-distributed + stochastic neighbor embedding
nonlinear method
overcome PCA limitation
No train & test data, t-SNE modifies output directly in order to minimize cost function.
It didn't know labels, and just try to preserve distance between each input vector.
Huge RAM requirement
t-SNE will fail on Donuts or XOR problem.
Last updated
Was this helpful?