8 - Value Function Approximation Flashcards

1
Q

List a couple of advantages of Value Function Approximation methods?

A

One advantage is that we don’t have to explicitly store or learn for every single state a reward model, value, state-action value, or policy. It allows us a more compact representation that can generalize across states and actions, thus reducing memory, computation, and experience needs. (We assume this is reasonable if we have a “smooth structure”).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Neural Networks / Deep Learning is the most popular differentiable function approximator. What is the second most popular differentiable function approximator?

A

Linear Feature Representations (ie - linear combinations of features)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Gradient descent is guaranteed to find a _________ {local, global} optima.

A

Gradient descent is guaranteed to find a local optima.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does the linear part of Linear Value Function Approximation mean?

A

This means that we represent a value function (or state-action value function) for a particular policy using a weighted linear combination of features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly