Renewal2 - Work with Compute Targets Flashcards
The 5 computer types/targets you can use
- Compute Instances
- Computer Clusters
- Kubernetes Clusters
- Attached Compute
- SERVERLESS Compute (new one!) - Spark or Azure Machine Learning Serverless Computes (Managed Compute!)
Use these compute targets during experimentation and development, and when you prefer working in a Jupyter notebook. A notebook experience benefits most from a compute that is continuously running.
Compute Instances, Spark Serverless Computer
Use these compute targets when moving to production and you want the compute target to be ready to handle large volumes of data i.e. scaling needs
Compute Clusters, Azure Machine Learning Serverless Compute
Use these compute targets when generating batch predictions, particularly in a pipeline job in Azure Machine Learning, since you’ll need them on-demand and scalable…
Computer Clusters and AML Serverless Compute
You want it light weight and cost efficient…
Use this compute target when generating real-time predictions, where you’d want the compute to run continuously, where you manage the compute
Kubernetes Clusters!
Not really a Compute Target…you wanted to do this in our own ML
Use this compute target when generating real-time predictions, where you’d want the compute to run continuously, where you let Azure manage the compute
Managed Online Endpoints with Azure ML managing the Containers
Compute Instance requirements
- Unique name across REGIONS
- To work on an instance, it needs to be assigned to your user account
- Can only be assigned to one user at a time (can’t handle Parallel Workloads)
Compute Cluster creation and parameters to consider
- size: VM type and compute option (CPU or GPU)
- max_instances: Max nodes per cluster
- tier: low priority or dedicated
- idle_time_before_scale_down: The length of time your cluster’s nodes stay alive before they scale down. Keep this high if you want to reduce startup times and scaling costs
Tier - setting Low Priority can lower costs but no guarantee of availability
Three scenarios you would use a Compute Cluster and why
- Running a pipeline job you built in the Designer.
- Running an Automated Machine Learning job.
- Running a script as a job.
Because each scenario allows you to scale up or down as necessary. Clusters also allow you to train multiple models in parallel.