Machine Learning Flashcards

1
Q

What is the definition of a hidden Markov model?

A

A HHM consists of:
- a set of states S
- a transition state probability matrix A
- a state PDF matrix defined for the set of possible observations
The current observation is dependent only on the current state, there is no dependence on previous states or observations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the HMM assumptions?

A

Temporal Independence
Piecewise Stationarity
Random Variability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a word level HMM?

A

Each word is modelled by a dedicated HMM
+ good performance
+ good for small well defined vocabularies
- many examples of each word required for training
- fails to exploit regularities in spoken language

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a sub-word level HMM?

A

HMMs are built for a complete set of sub-word building blocks
Word-level HMMs are constructed by concatenating sub-word level HMMs
+ can exploit regularities in speech data
+ efficient use of training data
+ flexible (models can be built for words not included in the training data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is a discrete HMM?

A

The state output PDF is defined by a list of probabilities

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the advantages and disadvantages of a discrete HMM?

A

+ computational advantages

  • vector quantisation may introduce non-recoverable errors
  • outperformed by continuous HMMs
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a continuous HMM?

A

The state output PDF must be defined for any value in the continuous observation set. Parametric continuous state output PDFs are used.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is Viterbi Decoding?

A

An algorithm used to find the sequence of HMM states which are most likely to have produced a given observation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the Viterbi decoding formula?

A

P(Y,X)=b_(x_1 ) (y_t)∏{a_(x_(t-1) x_t ) b_(x_t ) (y_t )}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the steps in error back propagation for a neural network?

A
  1. Choose initial weights
  2. Propagate each training sample through the network to obtain an output and calculate the error
  3. Calculate ∂E/∂w for each weight by propagating the error back up the network
  4. When all training samples have been seen change the weights by an amount proportional to ∂E/∂w
  5. Repeat until error falls below the threshold
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the formula for the change in weights in a neural network?

A

∆W= -ηδo

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the formula for delta in a neural network?

A

(∑wδ_k)o(1-o) for a non-output neuron

(o-t)o(1-o) for an output neuron

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the Baum-Welch algorithm?

A

A soft version of the Viterbi estimation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the formula for the Baum-Welch algorithm?

A

μ(i)=(∑γ_t (i) y_t )/(∑γ_t (i) )

γ_t (i)=P(x_t=i|Y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly