AI+ML+DS Flashcards
- Deep Learning, Goodfellow - Artificial Intelligence, Norvig - Introduction to Statistical Learning in Python, Hastie
Knowledge Based Approach to AI
Hard-coding knowledge about the world in formal languages for the computer system to make logical inference rules.
Machine Learning
Subset of artificial intelligence that enables computers to learn from data and improve their performance on a specific task without being explicitly programmed. Instead of being hand-coded with specific rules, machine learning algorithms can identify patterns and make predictions based on the data they are trained on.
Artificial Intelligence
A broad field of computer science that aims to create intelligent agents, which are systems that can reason, learn, and act autonomously.
In simpler terms, AI involves developing machines that can think and behave like humans.
Deep Learning
A subset of machine learning that uses artificial neural networks with multiple layers to learn from data. These neural networks are inspired by the structure and function of the human brain, and they can learn complex patterns and relationships in data that are difficult for traditional machine learning algorithms to capture.
Solves a central problem in representation learning by introducing representations that are expressed in terms of other, simpler representations
Representation of Data
Refers to the way data is structured and encoded so that it can be processed by machine learning algorithms. The choice of data representation can significantly impact the performance and efficiency of a machine learning model.
Feature
Each piece of information included in the representation of an observation
Representation Learning
Use of machine learning to discover not only the mapping from representation to output but also the representation itself
Autoencoder
Quintessential example of representation learning
Combination of an encoder function, which converts the input data into a different representation, and a decoder function, which converts the new representation back into the original format
Trained to preserve as much information as possible when an input is run through the encoder and then the decoder, but also trained to make the new representation have various nice properties
Encoder
Converts the input data into a different representation
Decoder
Converts the new representation (encoded) back into the original format
Factors of Variation
Concepts of abstractions that help us make sense of the rich variability in the data
Multilayer Perceptron (MLP)
A type of artificial neural network that consists of multiple layers of interconnected neurons. Each neuron takes a weighted sum of its inputs, applies an activation function, and passes the result to the next layer.
Input/Visible Layer
First layer of neural network that contains the variables we are able to observe
Hidden Layer
Layers of a neural network that are not the first (input) or last (output) layers of the network. Extracts increasingly abstract features from the data. Their values are not given in the data; instead, the model must determine which concepts are useful for explaining the relationships in the observed data.
Adaptive Linear Element (ADALINE)
A type of single-layer artificial neural network used for linear regression and classification tasks. It is similar to the perceptron but uses a least mean squares (LMS) algorithm for training, which allows it to learn more efficiently.
Rectified Linear Unit (ReLu)
A popular activation function used in artificial neural networks. It introduces non-linearity into the model, allowing it to learn complex patterns.
How does it work?
- If the input (x) is positive, the output is the input itself.
- If the input is negative, the output is zero.
Linear Algebra
Branch of mathematics that deals with the study of vectors, matrices, and linear transformations. It provides a framework for solving systems of linear equations and analyzing the properties of linear relationships between variables.
Scalar
A single number
Written in italics with lowercase variable names
Can be thought of as a matrix with a single entry
Vector
An array of ordered scalars = 1-D
(each number has a specific location in the array)
Written in lowercase names with bold typeface
Can be thought of as matrices that contain only one column
Matrix
2-D array of scalars
Written in uppercase names with bold typeface
Tensor
N-D array of scalars
Written in uppercase names with bold-tensor typeface
bold-tensor typeface is slightly different than our traditional bold typeface
Matrix Operation: Transpose
Taking the mirror image of a matrix across the main diagonal
Main Diagonal
Diagonal line on a matrix running down to the right, starting from its upper left corner.
Broadcasting
The implicit copying of a scalar to many locations when performing a matrix operation