markov models Flashcards

1
Q

when we do not have control over the state transitions but the states are completely observable then we have a:

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

how do you solve a POMDP?

A

the model has to find, for any action/observation history, the action that maximizes the expected discounted reward. the model has to find, for any action/observation history, the action that maximizes the expected discounted reward:

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

a POMDP model contains: (3)

A
  • A state transition (probability) function: P(s_{t+1}|s_t, a_t) (probability of the next state, given the current state and the action)
  • An observation function: P(o_t|s_t,a_t) (probability of the observation, given the current state and the action)
  • A reward function: E(r_t|s_t,a_t) (expected reward, given the current state and the action)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Partially Observable Markov Decision Processes are

A

MDPs that have partially observable states instead of the agent knowing exactly what the current state is.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is the difference between POMDPs and MPDs?

A

in MDPs, an agent knows exactly what the current state is.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what are 2 methods to find the optimal policy of an MDP

A

value iteration

policy iteration

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is the utility in a MPD?

A

the utility of a state is the expected sum of discounted rewards if the agent executes the policy pi. The true utility of a state corresponds to the optimal policy pi*.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is a markov decision process?

A

MPD’s are functions to choose actions, given the state of the world which are generally called policies. They associate an optimal decision with every state that the agent might reach in uncertain environments.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

why / when are HMM’s used?

A

because an experimenter can not always observe the states, only measure them by the observable outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

how is a hidden markov model different form a markov model?

A

hidden markov models do not have observations given but rather represent the states by the probability distribution of several possible observations that can be expected to occur in this state.: the probability that we observe k given we’re in state i.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

how do markov chains work?

A

1) starts in a state decided by starting distribution probilities.
2) visit states based on the probability to go from one state to the next
3) is a sttochastic process

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

when we have control over the state transitions but the states aren’t completely observable then we have a:

A

partially observable markov decision process: POMDP

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

when we do not have control over the state transitions and the states are not completely observable then we have a:

A

Hidden markov model (HMM)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

when we do have control over the state transitions but the states are completely observable then we have a:

A

markov decision process (MDP)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

when we do not have control over the state transitions but the states are completely observable then we have a:

A

markov chain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

when the observations are given, the Markov model is also called a:

A

markov chain model

17
Q

what is the markov property?

A

assuming that future states depend only on the current state, not on the events that occurred before it.

18
Q

what is a markov model?

A

A Markov model has observable states of the world, and an observed sequence and it assumes the Markov property.

19
Q

where are markov models applied?

A

in areas where there is recognition of temporal patterns with variable duration, such as speech and handwriting recognition. This has the problem that the patterns can be very short- or very long- lasting.