Lecture 11 - Neural Networks Flashcards

1
Q

What are the purposes for language modelling and natural language processing?

A
  • Search
  • Query and answering
  • Translation
  • Sentence generator
  • Speech recognition
  • Topics/Summaries
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the XOR problem?

A
Simple unit (Perceptron) has a binary output.
Cannot compute XOR (Complex interactions)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is one solution to the XOR problem?

A

Neural Networks due to their hidden layer and non-linear activation function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is meant by activation function?

A

Neural networks take an activation function, which the values and weights modify -> Basically a graph that gets modified

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is meant by a feed-forward network?

A

A network that feeds information forward - I.e only goes one way (not recurrent)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

If you have multiple neurons, are the activation function applied to each neuron separately?

A

Yes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the three layers of neural networks?

A

Input Layer
Hidden Layer (Could have multiple)
Output Layer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How do we adjust weights?

A

Backpropagation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why would we need to adjust weights?

A

To fix ‘error’ (i.e computed output != correct output)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

True or False: Backpropagation uses Gradient Descent to fix error

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

True or False: Backpropagation is a heavy process

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

True or False: Error is computed over all output nodes

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

In a hidden layer, we do not have a target output value - So how do we compute the error?

A

We compute how much each node contributed to downstream error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How do neural networks initialize weights?

A

Randomly -> Uniformly from interval

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are the advantages of Neural Networks?

A
  • Neural Models don’t need any smoothing
  • Provide higher accuracy than N-gram models
  • XOR solution: Complex Interactions
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What are the disadvantages of Neural Networks?

A

Training Neural Models is strikingly slow when compared with traditional models

17
Q

True or False: BERT models need to be trained by you to use them

A

False - A BERT model is pretrained, you just need to fine-tune it