Interpretability and Feature Importance Flashcards
Define intrinsic vs post-hoc interpretability
Intrinsic means the model itself is interpretable (e.g. linear models) and post-hoc means some method is needed to understand feature importance.
What is the regularisation term in Total Variation? and in Sparse Total Variation?
J(w) = |∇w| i.e. the discrete gradient of w in 3D
Sparse: J(w) = |∇w| + |w|
Write down the regularisation term of a Laplacian model and Sparse Total Laplacian
Check notes
Define Gini importance
total reduction of the impurity criterion brought by that feature
Define permutation feature importance
Randomly shuffle values in a given feature and measure drop in performance
Define LIME
Local Interpretable Model-agnostic Explanations (LIME) perturbs samples in a given region and trains an interpretable model on the o.g. model’s predictions to check what is driving decisions
What is SHAP?
Game theory approach to feature importance. Changes one feature at a time and samples to see how much that feature contributed to prediction change.
Explain the difference between filter, wrapper, and embedded methods for feature selection
Filter methods: use a proxy measure to rank the features with respect to their relationship with the labels/targets/outcomes (e.g. correlation, statistical test, mutual information).
Wrapper methods: use a predictive model to score/rank the features according to their predictive power (e.g. Recursive Feature Elimination, aka RFE).
Embedded methods: perform feature selection as part of the model construction process (e.g. LASSO and Elastic-net regularization).
How does stability selection work?
By training various sparse models on perturbed data to see which features are kept most often.