Quant- Machine Learning Flashcards
Neural networks
Include highly flexible ML algorithms that have been successfully applied to a variety of tastes characterized by non linearities and interactions amount features.
The foundation for deep learning and reinforcement learning.
K-fold cross validation
Technique for mitigating the holdout sample problem (excessive dedication of the training set size)
Validation technique in which the data are shuffled randomly and then are divided into k equal sub samples with k-1 samples used as training samples and one sample the k-th used as a validation sample.
Repeat multiple times to minimize bias and variance
Penalized regression
Includes a constraint such that the regression coefficients are chosen to minimize the sum of squared residuals plus a penalty term that increases in size with the number of included features. Therefore a feature must make a sufficient contribution to model fit to offset the penalty from including it.
LASSO
Type of penalized regression / regularization technique
Stands for least absolute shrinkage and selection operator
LASSO also minimizes the sum of the absolute values of the regression coefficients and eliminates the least important features from the model
Equation includes Lambda which is hyper-parameter whose value is set by research
Support Vector Machine (SVM)
Supervised algorithm used for classification, regression, and outlier detection
Determines the hyper plane that optimally separates the observations into two sets of data points
K-nearest neighbour
To classify a new observation by finding similarities between the new observation and the existing data
Classification and regression tree (CART)
Applied to predict either a categorical target variable, producing a classification tree or a continuous target variable, producing a regression tree
What are the two categories of supervised learning?
- Regression- if the target variable is predicted to be continuous
- Classification- if the target variable is categorical or ordinal
Describe the difference between bias error and variance error
Bias error- the degree to which a model fits the training data
Variance error- how much a models results change in response to new data from validation and test samples
Out of sample error = bias + variance + base error (randomness in the data)
Random forest classifier
A collection of many different decision trees generated by a bagging method or by randomly reducing the number of features available during training
Principals components analysis (PCA)
An unsurprised ML algorithm that reduces highly correlated features into fewer uncorrelated composite variables by transforming the feature covariance matrix
K-means
Unsupervised ML algorithm that partitions observations into a fixed number of non overlapping clusters. Each cluster is characterized by its centroid and each observation belongs to the cluster with the centroid to which that observation is closest
Neural networks
Consist of nodes connected by links
They have three types of layers: an input layer, hidden layers, and an output layer. Learning takes place in the hidden layer nodes, each of which consists of a summation operator and an activation function.
There are usually at least 3 hidden layers (also called deep learning nets) which are the backbone of artificial intelligence
Machine learning
Machine learning aims at extracting knowledge from large amounts of data by learning from known examples to determine an underlying structure in the data. The emphasis is on generating structure or predictions without human intervention. An elementary way to think of ML algorithms is to “find the pattern, apply the pattern.”
Unsupervised Learning
-algorithms are trained with no labeled data, so
they must infer relations between features, summarize them, or present an interesting underlying structure in their distributions that has not been explicitly provided
-two important types of problems well suited to unsupervised ML are dimension reduction and clustering.