B10 Improving Performance Flashcards
Exam Prep
The process of adjusting a model’s parameters to identify the best fit is called _____________.
Parameter tuning
Automated parameter tuning requires you to consider:
- What type of machine learning model (and specific
implementation) should be trained on the data? - Which model parameters can be adjusted, and how
extensively should they be tuned to find the optimal
settings? - What criteria should be used to evaluate the models to find the best candidate?
The technique of combining and
managing the predictions of multiple
models is known as ____________.
meta-learning
_______ and ________ are statistics that evaluate the performance of Classification models, while _______ or ________ are used for numeric models.
Accuracy; Kappa
R-squared;RMSE
Cost-sensitive measures such as _______, _______, and ___________ can also be used to evaluate performance.
sensitivity, specificity, AUC
The meta-learning approach that
utilizes the principle of creating a varied
team of experts is known as an
_______.
ensemble
The _________ dictates how much of the training data
each model receives
allocation function
The __________ governs how disagreements among the predictions are reconciled
combination function
Some ensembles utilize another model to learn a combination function from various combinations of predictions. This is known as _______.
stacking
The two main families of ensemble methods are:
1.
2.
Averaging methods
Boosting methods
Ensemble methods provide a number of performance
advantages over single models:
-___________ to future problems.
- Improved performance on _____ or _______
datasets.
- The ability to synthesize data from distinct
domains.
- A more nuanced understanding of difficult learning
tasks.
Better generalizability
massive or miniscule
Independently built models with their predictions averaged or combined by a voting scheme. They attempt to reduce the _______ of a ________. Examples include _________ and _________.
variance
single base estimator
Bagging methods
Random Forest
___________ or __________ is a
technique that generates a number of
training datasets by __________
sampling the original training data.
Bootstrap Aggregating
Bagging
Bootstrap
In the Bagging process: 1. The training datasets are used to generate a set of models using a \_\_\_\_\_\_\_\_\_\_. 2. The models' predictions are combined using \_\_\_\_\_\_\_ (for classification) or \_\_\_\_\_\_\_ (for numeric prediction).
single learner
voting
averaging
The Random Forest (or Decision Tree Forest) learner focuses only on ensembles of decision trees. It combines the base principles of \_\_\_\_\_\_\_ with \_\_\_\_\_\_\_\_ to add additional diversity to decision tree models.
bagging
random feature selection