A.I. Flashcards
which classifier is most easilt understandable to humand
decision tree
You have about 1 million training examples in a 6-dimensional space.You only
expect to be asked to classify 100 test examples.
Nearest Neighbours is a good choice. The dimensionality is low and so
appropriate for NN. For NN, training is very slow and since there are few classications, the fact that this will be slow does not matter
You are trying to predict the average rainfall in Bristol as a function of the
measured currents and tides in the Atlantic ocean in the previous six months.
This is a regression problem; least squares regression, regression trees or
locally weighted nearest neighbours are all appropriate choices
MDP or reinforcement learning?
mental model of how the world works
MDP
MDP or reinforcement learning?
Perform actions in the world to nd out and collect rewards.
RL
MDP or reinforcement learning?
The goal is to find policy to collect maximum rewards.
MDP
MDP or reinforcement learning?
its an offline process
MDP
Value iteration vs Policy iteration
VALUE:
- Cheaper
- good for large state spaces and a few actions
Policy Iteration
-good for small state spaces
When to use Policy iteration or Value iteration?
For small state spaces: policy evaluation using exact solution methods is often the most efficient approach, typically very fast and converges quickly.
For large state spaces, when time might be prohibitive. Value iteration is preferred.
Reinforcement learning
Reinforcement learning is on-line methodology that we use when the model of world is unknown and/or rewards are delayed.
Describe the main ideas behind the EM algorithm
Expectation Maximization (EM), tries to maximize the marginal likelihood
The EM algorithm works analogously
EM consists of alternating between
two steps, the E-step and the M-step.
In the E-step, we dont know what
the hidden variables are, so we compute the posterior distribution over them
given our current parameters
In the M-step, we take in our set of full
assignments with weights, and we just do maximum likelihood estimation
If we repeat the E-step and the M-step
over and over again, we are guaranteed to converge to a local optima
Define what is meant by an admissible heuristic
estimate of the cost of reaching the goal state
in an informed search algorithm.
In order for a heuristic to be admissible to
the search problem, the estimated cost must always be lower than or equal
to the actual cost of reaching the goal state
What is the main idea of Principal Component Analysis (PCA)?
Maximize the variance of projection along each component or minimize the
reconstruction error
reconstruction error
the squared distance between the original data and
its estimate
Why is PCA useful for visualization tasks
Dimensionality reduction. the first and second component
can be plotted against each other to obtain a two-dimensional representation
of the data that captures most of the variance, useful to analyze and interpret
the structure of a dataset
Indicate how the method (K-NN) could overfit training data
Every point in dataset (including noise) denfies its
own decision boundary. The distance function can be chosen to do well on training set but less well on new data.
2. How can you reduce overtting? Use larger K for K-NN. Use cross-validation to choose K and the metric.