Sagemaker Flashcards
Complete with the name of each Sagemaker functionality:
-_________ is useful for more complex inference workflows
-_________ can help deployment on edge devices
-_________ can accelerate inference for Deep Learning models
-_________ evaluates new models against currently deployed models to catch errors
-Inference Pipeline
-Sagemaker Neo
-Elastic Inference
-Shadow Testing
When training models on AWS, what is de the difference between File mode and Pipe mode?
File mode copies the training files to the EBS volume of the instance, while Pipe mode streams the data from S3, reducing the volume space needed
True or False: RecordIO-Protobuf’s advantages over CSV are that is faster, more eficient and can be used in Pipe mode
False, CSV can also be sed in pipe mode
True or False: Sagemaker Linear Learner can perform Regression, Binary classification and multi-class classification
True
What training data types and training types does Linear Learner support?
Data type: CSV and RecordIO-Protobuf (float32)
Training type: Pipe and File
True or False: When training Linear Learner, it is recommended to launch multiple training jobs at the same time, since the model is very sensitive to hyperparameters
False, multiple models are trained by default and the best one is selected by default
True or False: Linear Learner supports both L1 and L2 regularization
True
Identify the following Linear Learner Hyperparameters by their description:
- ______ equalizes the importance given to each class in a multi-class classification model
-_______ defines the speed with which the SGDC algorithm converges
-_______ governs L1 regularization
-_______ governs L2 regularization weigh decay
-_______ keeps the precision at the specified value then maximizes recall
-_______ keeps the recall at the specified value then maximizes precision
-Balance_multiclass_weights
-Learning_rate
-L1
-Wd
-target_precision
-target_recall
True or False: Linear Learner can benefit from single and multi-machine CPUs, bot not GPUs
False, it can benefit from single machine GPUs, but not multi-machine ones
What is XGBoost?
It is an ensemble model that trains multiple decision trees based on the errors of previous trees
True or False: XGBoost accepts only CSV and RecordIO-Protobuf training data
False, also accepts Parquet and libsvm
Identify the following XGBoost Hyperparameters by their description:
- ______ prevents overfitting
-_______ step size shrinkage, helps with overfitting
-_______ minimum loss reduction to create a partition
-_______ governs L1 regularization
-_______ governs L2 regularization
-_______ metric to use on the model evaluation process
-_______ adjust balance between positive and negative weights (helps with unbalanced data)
-_______ max depth of the tree, lower values help with overfitting
-Subsample
-Eta
-Gamma
-Alpha
-Lambda
-eval_metric
-scale_pos_weight
-max_depth
True or False: XGBoost is compute limited, so the best type of training instance for it is a compute focused one, such as C
False, it is memory bound, so something like an M5 instance
True or False: XBoost nowadays accepts both single and distributed instance GPU training
True, as long as you configure the hyperparameters adequately
What is Seq2seq’s use cases?
Any case where a sequence is received as input and passed as output (text to text, audio to text, etc)
What model types are used to implement Seq2seq?
RNNs and CNNs with attention
What types of input data does Seq2seq accept?
Only RecordIO-Protobuf with integer tokens
What types of metrics can Seq2seq be optimized on?
Accuracy, BLEU score, Perplexity (Cross-entropy)
Which of the following are Seq2Seq hyperparameters?
-Batch_size
-Optimizer_type (adam, sgd, rmsprop)
-Learning_rate
-Alpha
-Lambda
-Num_layers_encoder
-Num_layers_decoder
-Top_k
-Top_n
-Batch_size
-Optimizer_type (adam, sgd, rmsprop)
-Learning_rate
-Num_layers_encoder
-Num_layers_decoder
True or False: Seq2seq can run on both CPU and GPU instances
False, only single or multi GPU instances
What is Amazon DeepAR useful for?
Performing 1D timeseries inference
What XGBoost Hyperparameters do you have to configure to enable single-GPU and multi-GPU training?
-Single GPU: tree_method = gpu_hist
-Multi GPU: use_dask_gpu_training = true and TrainingInput distribution = fully_replicated
True or False: XGBoost distributed GPU training works only for RecordIO-Protobuf
False, CSV and Parquet
What Seq2seq metrics are good for measuring machine translation problems
-BLEU score and perplexity