ML Implementations and Operations Flashcards

1
Q

What is the name of the library used for making containers compatible with Sagemaker?

A

sagemaker-containers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the file structure of a training container on Sagemaker?

A

/opt/ml
├── input
│ ├── config
│ │ ├── hyperparameters.json
│ │ └── resourceConfig.json
│ └── data
│ └── <channel_name>
│ └── <input></input>
├── model

├── code
│ └──



└── output
└── failure
</channel_name>
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the file structure of a docker image?

A

WORKDIR
* nginx.conf
* predictor.py
* serve/
* train/
* wsgi.py

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are production variants of a model?

A

They variants of a model that you can test on live traffic through the usage of weighted traffic redistribution. Allows you to perform A/B testing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

True or False: Shadow Testing is an example of a production variant of a model

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Sagemaker Neo?

A

It is a service that optimizes machine learning model execution for specific devices, allowing them to run on edge devices

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

True or False: Sagemaker Neo is composed of a singular compiler that optimizes your code

A

False, it is composed of a compiler and a runtime

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What other AWS Service would you use with Sagemaker Neo to deploy a model to the Edge?

A

IoT Greengrass

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

True or False: Sagemaker Neo optimized models can be deployed to an HTTP endpoint

A

False, only HTTPS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

True or False: There are no restirction to which instance a Sagemaker Neo optimized model can be deployed to

A

False, it must be deployed to the same type of instance as the one where it was optimized

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

True or False: IoT Greengrass uses Lambda inference applications

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

True or False: You can use KMS integrations to encrypt all the data on Sagemaker Notebooks and jobs

A

False, all notebook data can be encrypted, but on jobs only data /opt/ml/ and /tmp can be encrypted

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

True or False: When running a training job, inter-node training communication may be optionally encrypted with no impacts on performance (AKA inter-container traffic encryption)

A

False, it can result in increased training cost and time for deep learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

True or False: You can run Sagemaker jobs inside VPCs, but Notebooks are internet enabled by default. If this is disabled, your notebook needs an interface endpoint (PrivateLink) or NAT
Gateway, and allow outbound connections, for training and hosting to
work. The same is true for inference and training images.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What on Sagemaker can Cloudwatch log and monitor?

A

-Ground Truth (active workers, how much they are doing)
-Instance health metrics
-Invocations and latency of endpoints

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What are some frameworks you can use to perform distributed training on Sagemaker?

A

Horovod and Parameter Servers

17
Q

True or False: Sagemaker does not support Spot Training

18
Q

What is Sagemaker Elastic Inference?

A

It is a feature that accelerates inference at a fraction of the cost of using a GPU

19
Q

True or False: Sagemaker Elastic Inference works by leeting you select instance types optimized for inference when you choose your instance

A

False, it allows you to use an accelerator alongside a CPU instance

20
Q

True or False: Elastic Inference cannot be applied to Sagemaker notebooks

A

False, it can

21
Q

What built-in Sagemaker algorithms work with Elastic Inference?

A

-Image Classification and Object Detection

22
Q

True or False: Elastic Inference works with both pre-built and custom MXNet, Tensorflow and PyTorch containers

23
Q

Which use cases are good for Serverless Inference?

A

Use cases where traffic is unpredictable

24
Q

What is Sagemaker Inference Recommender?

A

It is a Sagemaker feature that recommends the best instance type and configuration for your model deployments

25
True or False: Sagemaker Inference Recommender automates load testing model tuning
True
26
How does Sagemaker Inference Recommender work?
It benchmarks your model registered on Model Registry for various instace types and endpoint configurations and generates metrics for all of them
27
True or False: Sagemaker Inference Recommender has 2 types of recommendation, Instance Recommendation and Endpoint Recommendation, and each of the need their own benchmarks
True
28
How do Sagemaker Deployments interact with AZs?
Sagemaker automatically tries to distribute instances across AZs, but this only works if you have more than 1 instance and subnets configured in different AZs
29
What is Amazon Sagemaker Operators for Kubernetes?
It is a framework that integrates Sagemaker with Kubernetes to facilitate MLOps
30
True or False: Amazon Sagemaker Operators for Kubernetes enables Hybrid ML workflows (on-prem + cloud)
True
31
What are Sagemaker Projects?
They are Sagemaker Studio's native MLOps solution with CI/CD.
32
True or False: Sagemaker Projects uses code repositories for building and deploying solutions using Step-Functions
False, it uses Sagemaker Pipelines
33
What is a Sagemaker inference pipeline?
It is linear sequence of 2-15 containers where the output of one container is piped to the next in a way as to make the whole inference work as a pipeline. Useful for dealing with pre-processing, post-processing, etc
34
True or False: Sagemaker Inference Pipelines are compatible with Scikit-learn containers and SparkML, which can be run with Glue or EMR
True