Predictive Analytics Problem Definition (10-20%) Flashcards
Define descriptive analytics.
Focuses on insights from the past and answers the question “what happened?”
Define predictive analytics.
Focuses on future insights and addresses “what will happen next?”
Define prescriptive analytics.
Suggests decision options ie. “what would happen if i do this?” or “what is the best course of action?”
What are the characteristics of predictive modeling problems?
- Identify and define business issues that needs to be addressed
- clarify the business issue
- develop hypotheses, a testable hypothesis to guide the project
- an ability to assess the outcome, gain clarity, and measurable key
performance indicators (KPIs)
- Address the issue with a few well-defined questions - what you need, what’s the target outcome and how well it is representing the problem, how will the model performance be evaluated
- Having good/useful data that can answer these questions - high quality, permission to use data. amd identify info gaps
- Knowing the predictions will drive actions or increase understanding - is the approach practical, understandable, adn implementable
- Knowing the approach is better than the existing one - easier approach
- Monitor and update the model as new data comes in
What is a Field Test?
When the model is implemented in the exact way it will be used, but it is not yet used for decision making, only observed on its potential to help users make good decisions. Particularly when the problem, data, or type of model is new
Explain the Bia-Variance Trade-off.
- Bias: the expected loss arising from the model not being complex/flexible enough to capture the underlying signal
- Variance: the expected loss arising from the model being too complex and overfitting to a specific instance of the data
- High bias means that the model won’t be accurate because it doesn’t have the capacity to capture the signal in the data, whereas high variance means that the model won’t be accurate because it overfit to the data it was trained on and thus won’t generalize well to new, unseen data
For y=f(x) + Ɛ, what are the assumptions for the error term (Ɛ)?
- The errors are independent - with observed data, we know the form and parameters of distribution of Ɛ, but for a new observation, the past tells us nothing about the value of Ɛ that will go with it.
- The errors have a mean of zero - if f(x) is known, on average, predictions will be correct.
- The errors have a constant variance - variance doesn’t depend on x.
- The errors have a Gaussian (normal) distribution.
Define Loss Function.
A function of two variables - the prediction and a single observed new value that measures error ie. squared error loss function.
Note: the loss is a random variable since many items in the expression are random variables. Because the loss function is a random variable, it may not be useful as a measure of model quality so we typically look at its expected value (average loss over predictions we might make and over all training sets we might use).
Define bias-variance decomposition.
A useful theoretical window into how model complexity affects the quality of models and helps demonstrate the important insight that more complex models aren’t always better models.