3. Generative models for discrete data Flashcards

1
Q

<b>The beta-binomial</b>
Likelihood
- what’s the sufficient statistics?
- likelihood in beta-binomial model

A

p. 75

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

<b>The beta-binomial</b>
Prior
- what’s the conjugate prior?
- what’s the conjugate prior of the Bernoulli (or Binomial) distribution?
- how are the parameters of the prior called?
- exercise 3.15
- exercise 3.16
- hyperparameters in uniform prior the the beta-binomial model

A

p. 76

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q
<b>The beta-binomial</b>
Posterior
- posterior in beta-binomial
- what pseudo counts are?
- what's equivalent sample size?
- posterior MAP
- posterior MLE
- when MAP = MLE?
- posterior mean
- posterior variance
A

p. 77

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q
<b>The beta-binomial</b>
Posterior predictive distribution
- p(x|D)
- add-one smoothing
- beta-binomial distribution (def, mean, var)
A

p. 79

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

<b>The Dirichlet-multinomial model</b>

Likelihood and prior

A

p. 81

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

<b>The Dirichlet-multinomial model</b>

Prior

A

p. 81

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

<b>The Dirichlet-multinomial model</b>
Posterior
- MAP and MLE

A

p. 81

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

<b>The Dirichlet-multinomial model</b>

Posterior predictive

A

p. 83

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

<b>Naive Bayes</b>

  • NBC definition
  • binary, categorical, and real-valued features
A

p. 84

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q
<b>Naive Bayes</b>
Model fitting
- log p(D|theta)
- MLE
- BNBC
A

p. 85

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

<b>Naive Bayes</b>
Using the model for prediction
- p(y=c|x,D)
- special case if the posterior is Dirichlet
- what if the posterior is approximated by a single point?

A

p. 87

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

<b>Naive Bayes</b>

The log-sum-exp trick

A

p. 88

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

<b>Naive Bayes</b>

Feature selection using mutual information

A

p. 89

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q
<b>Naive Bayes</b>
Classifying documents using bag of words
- Bernoulli product model (binary independence model)
- x_ij and theta_jc interpretation
- adapt the model to use the number of occurrences of each word
- burstiness phenomenon
- Dirichlet Compound Multinomial (DCM)
- What's Polya urn?
A

p. 90

How well did you know this?
1
Not at all
2
3
4
5
Perfectly