Graphical models Flashcards

1
Q

Describe the “Probabilistic Pipeline”

A

Build model -> Discovere patterns -> Predict and Explore -> Critisise Model -> Build model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is probabilistic inference?

A

Given latent variables and observed variables, propabilistic inference is learning about the unknown latent variables trough the posterior distrubution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What can we do if the marginal likelihood is intractable?

A

We can approximate using e.g. sampling and markov chain monte carlo.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What kind of graphs are used for:

  1. Bayesiean networks
  2. Markov random fields
  3. Factor graphs
A
  1. Directed
  2. Undirected
  3. Bipartite
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why do we use graphical models?

A
  1. Simple way to vizualise
  2. Gives insights into properties of the model
  3. Can be used to design/ motivate new models
  4. Can be used to express complex computations
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What type of nodes do we have in Bayesian networks?

A

Shaded nodes: Observed random variables

Unshaded nodes: Latent random variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What kind of depenedens do we have in a Bayesian network if and arrow is drawn from node a to b

A

p(b|a)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

For Bayesian graphs, split into three sets, A,B,C, when can we say that A is conditional independent of B given C?

A

If all paths from A to B are blocked. A path is blocked if:

1 The arrows meet tail to tail or head to tail at a node and the node is in C
2. The arrows meet head to head at a node and neither the node nore it’s decendands are in C

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

In a Markov Random field, how can we se if two variables are conditional independent given the rest of the variables?

A

They arn’t connected directly in the graph

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is a clique in Markov random fields?

A

A fully connected subgraph

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the formula for factorizing the joint distrubution in Markov Random fields?

A

p(x) = (1/Z) prod_C phi_C(x_C)

Z is the normalizing constant
C are the maximum liques
phi is the clique potential

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the formula for the normalizing constant Z in markov random fields, and what is the complexity of calculating it?

A

sum_x prod_C phi_C(x_C) = Z

If the model has M nodes and each node has K states the complexity is M^K

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

When do the clique potentials have a probabilistic interpretation?

A

If we convert a directed grap to a MRF.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the problem factorizing using MRF for continues models?

A

We have to solve an integral for the normalizing constant Z, which is often intractable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How can we see which variables are conditional independent in MRF?

A

For sets A, B, C, A and B are conditional independent if all paths from A to B go trough C.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How is the clique potential related to energy functions?

A

phi_C(x_C) = exp(-E(x_c)) for some energy function E.

17
Q

What is the total energy regarding clique potentials?

A

total_energy = -log(p(x)) = sum -log (phi_C(x_C))

18
Q

Describe the MRF energy minimization model for image denoising

A

Let y_i be the observed variables and y_i be the latent variables.

E(yi, xi) = - axiyi
E(xi, xj) = -bxixj for neighbour pixels
E(xi) = cxi Bias term
E(x) = -a
sum(xiyi) - bsum(xixj) + csum(x_i)

Initialize xi to yi, calculate energy set xi to {-1, +1} minimizing the energy

19
Q

What does a link between a variable node and a factor node indicate in a factor graph?

A

That the factor depends on the variable

20
Q

How do we convert a Directed graph to a MRF?

A
  1. Add undirected links between all pairs of parents for all nodes and consider the rest of the links undirected
  2. Identify maximum cliques
  3. Initialize clique potential to 1
  4. Multiply each conditional distribution factor in the directed graph into one of the clique potentials
21
Q

Can a directed graph have ore than one representation as a factor graph?

A

Yes

22
Q

How can we create a factor graph from a MRF

A

Add a factor node for each clique

23
Q

What is one advantage of factor graphs over MRF graphs?

A

Can remove cycles

24
Q

What is the goal of the message passing algorithm?

A

Finding the (marginal) posterior distrubution

25
Q

How do we pass messages from variable to factor nodes?

A

Send the product of all incoming messages

26
Q

How do we send a message from factor nodes to variable nodes

A

marginalize out all incoming variables from

f(x1, x2, …, xM) prod incoming messages

27
Q

How do we initialize leaf nodes in factor graphs?

A

If the node is a variable, initialize to 1, if it is a factor initialize to f(x)

28
Q

How can we use observed results in the message passing algorithm to say something about the posterior?

A

If v_hat are the observerd variables, h are the unobserved variables and x are all variables, we have:

p(x)p(v=v_hat) = p(h,v=v_hat) prop. to p(h|v=v_hat)

29
Q

Name some applications of inferenec in graphical models

A

Ranking, Computer Vision, Coding Theory, Linear Algebra, Signal Processing

30
Q

If xi, xj are conditionally independent given everything else in a MRF (They don’t have a link) can they appear in a common factor in the factorization?

A

No