Unsupervised Learning Flashcards
unsupervised learning
learning algorithm works out the patterns data contains for the purpose of compact representation (dimensionality reduction) and categorisation & analysis (clustering)
dimensionality reduction
transform data into a latent space representation of that data
latent space
reduced dimensional space in which essential features of input data are encoded
latent space in autoencoder
compressed representation of input data
encoding
process of transforming/compressing data into latent representation
decoding
process of recovering/reconstructing data from latent space compressed
auto encoder
MLP (fully connected network) whose first hidden layer constitutes an encoder with the output of layer being the latent representation then the next layer after latent space is decoder
how does dimensionality reduction happen with autoencoders
encoding input data into a lower-dimensional space by setting #neurons to be smaller than # of neurons
how is network trained with autoencoder
network is trained end to end to minimise MSE (difference between input and reconstructed output)
usages of autoencoders
compressing file into a zip
noise reduction
auto encoder architecture picture
input and output number same
hourglas shape
more hidden layers allows you to find more features
reduced neuron count in encoder to find compressed latent data (important patterns) then expand neuron count to extract
PCA stands for
principal component analysis
what is PCA and what is it used for
used for dimensionality reduction while preserving as much variance as possible in the data
It transforms the original dataset into a new coordinate system, where the axes (called principal components) are ordered by the amount of info it conveys about data
how to do PCA to compute components
given Nx D matrix
compute mean value for each attribute
compute covariance matrix C
get sorted eigen values
first K rows of E gives K most important components of X
applications of PCA
data visualisation (visualise high-dimensional data in 2D or 3D)
noise reduction (retains only the most significant principal components)
feature reduction