ML Flashcards

1
Q

Artificial Intelligence

A

Program that can sense, reason, act & adapt

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Machine Learning

A

Algorithms where performance improve as they re exposed to more data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Deep Learning

A

Multilayered neural networks learn from lots of data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Types of analytics in order of complexity

A

Descriptive
Predictive
Prescriptive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Ability to process large data because of ________. Using:
1.
2.
3.

A

Infrastructure

  1. CLoud Services
  2. GPU: Handles graphic rendering tasks
  3. TPU: Gives high performance & pure efficiency when running tensor flow
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Types of ML (3 types)

A

Supervised - training data
Unsupervised - no training data
Reinforcement - interaction, +/-ve feedback

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Summary of what to use:
Continuous = …
Discrete = …

A

…Accuracy
…Confusion Matrix

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Support Vector Machines use

A

Discriminative classifier
Clear margin of separation between categories
Used in small clean data sets
Don’t suffer from over fitting as much as other methods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

When should you not use Support Vector Machines

A

large data sets because required training time is higher in noisy data sets

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Decision trees

A

Used for regression lasts - continous variable decision trees
Used for classification - Categorical variable decision trees
Why use them?
→ involves stratifying or segmenting the predictor space spare into # of regions
→ good for non-linear data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Random Forests

A

Belong in the general category of Ensemble Methods
Increases predictive accuracy but sometimes at the expense of explainability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Ways to increase accuracy

A

Bagging:
Learns from each other independently in parallel & combines them for determining model average

Boosting:
Learns sequentially & adaptively to increase model predictions of learning algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Pruning

A

Decreases size of decision trees by removing sections of the tree that are non-critical & redundant to classify instances.

Decreases of the final classifier .’. increase accuracy by decrease of overfitting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Why use Unsupervised Learning

A

Easier to obtain unlabelled data
Takes place in real time
Decrease complexity in comparison to supervised
finds all kinds of unknown patterns in data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

K-means cluster analysis

A

Help with data-driven insights
Deep domain knowledge in required
No right or wrong matter for interpretation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Isolation Forest Algorithm

A

Basic idea is to overfit decision tree models

  • Grows in random decisions see model until exact instance is in its own leaf
  • Model is forced to keep splitting
17
Q

Dimensionality reduction

A

Decrease features to easily understandable dimensions

Helps project data in dimensions the brain can understand

18
Q

Deep Learning

A

Can be supervised or unsupervised
Uses multilayered neural networks called DNN to simulate complex decision making power of human brain

19
Q

Convolutional neural nets

A
20
Q

What’s the difference between loss function and optimisation?

A

Loss function: Measures how far a model’s predictions are from the actual values. Lower loss indicates a better fit.

Optimization: The process of adjusting a model to minimize the loss function and improve its performance.

Analogy: Imagine training a dog to fetch. The loss function is like the distance between the ball and where the dog drops it. Optimization is the process of training the dog to minimize that distance (bring the ball back closer each time).

21
Q

Backpropagation

A

A method in deep learning to adjust the internal parameters of a neural network. It calculates the error between predictions and actual values, then propagates it backwards through the network to update weights and biases, making the model learn from its mistakes.

Analogy: Imagine a team game where players need to improve their coordination. Backpropagation is like a coach reviewing plays, identifying where things went wrong (high error), and guiding each player (weights and biases) on how to adjust for the next round.

22
Q

Transfer Learning

A

Reusing a pre-trained neural network model on a new task. The pre-trained model (usually trained on a massive dataset) acts as a starting point, with its weights fine-tuned for the specific needs of the new task. This saves time and data compared to training a new model from scratch.

Analogy: Learning to ride a bike. Once you know the basics (balance, pedaling), you can transfer that knowledge to ride a motorcycle (similar concept, new controls). Transfer learning leverages pre-existing knowledge (pre-trained model) for a new task (riding a motorcycle).

23
Q

Pros and Cons of Deep Learning

A

Pros:

High Accuracy: Can achieve state-of-the-art performance on complex tasks like image recognition, natural language processing, and speech recognition.
Feature Learning: Discovers patterns and features from raw data automatically, eliminating feature engineering.
Scalability: Handles large and complex datasets effectively.
Cons:

High Computational Cost: Requires powerful GPUs and large amounts of data for training, making it resource-intensive.
Black Box Problem: Deep models can be complex and difficult to interpret, making it challenging to understand their decision-making process.
Overfitting: Prone to overfitting if not trained properly, leading to poor performance on unseen data.

24
Q

Stemming

A

Popular way to decrease size of vocab in Neural language tasks by conflating words with related meanings

Aims to convert words with same stem/root to single word types

25
Q

Tokenisation

A

Splitting texts into individual words or sequences of words [N-grams]

26
Q

K-means clustering

A

An unsupervised learning algorithm that groups similar data points together into clusters. It iteratively assigns data points to the closest cluster center (mean) and recalculates the center until convergence (minimal changes).

Analogy: Sorting socks by color. K-means separates data points (socks) into piles (clusters) based on their features (color) by minimizing the distance between each sock and the average color of its pile.

27
Q

Supervised vs Unsupervised Learning

A

Supervised learning: Learns from labeled data (data with known outcomes) to make predictions on unseen data. It’s like learning with a teacher who provides examples and corrects mistakes.

Unsupervised learning: Discovers patterns and structures in unlabeled data (data without predefined categories). Imagine exploring a new city with no map - you learn by looking for landmarks and patterns.

Key Difference: Labeled data. Supervised learning uses labeled data for training, while unsupervised learning doesn’t.