6. Buidling Secure ML Pipelines Flashcards
What are the three types of encryption used in GCP?
Encryption at rest (Cloud Storage and BigQuery tables)
Encryption in transit (Transport Layer Security)
Encryption in use
What encryption is used to encrypt individual table values in BigQuery?
Authenticated Encryption with Associated Data encryption
What are the differences between server-side and client-side encryption for encryption at rest?
Server-side encryption:
Encryption that occurs after the cloud storage receives your data, but before the data is written to disk and stored. The keys are managed by Key Management Service.
Client-side encryption:
Encryption that occurs before data is sent to Cloud Storage and BigQuery but it will still be encrypted on the server-side. Client-side keys are managed by the user.
How does encryption in use work?
Confidential Computing protects your data in memory from compromise by encrypting it while it is being processed. You can encrypt your data in use with Confidential VMs and Confidential GKE Nodes.
What are the two levels of roles in Identity and Access Management in GCP?
Project‐level roles: Assign roles to a principal (user, group, or service account).
Resource‐level roles: Grant access to a specific resource (individual users). The resource must support resource‐level policies.
What are the three types of IAM roles available in Vertex AI?
Predefined roles: Administrator and User
Basic roles: Owner, Editor, and Viewer at the project level.
Custom roles: allow you to choose a specific set of permissions, create your own role with those permissions
What are two types of Vertex AI notebooks with Vertex AI Workbench?
User‐managed notebook instances are highly customizable
Managed notebook is less customizable.
Advantages are its integration with Cloud Storage and BigQuery in JupyterLab and automatic shutdown
What are two ways to set up user access modes (permission) for both user‐managed and managed notebooks?
Single User Only access mode grants access only to the user that you specify.
Service Account access mode grants access to a service account. You can grant access to one or more users through this service account.
How can you run Vertex AI APIs in Google Colab?
Create a service account key with access to the Vertex AI administrator and Cloud Storage owner permission.
Then you can provide the location of the JSON key file to the GOOGLE_APPLICATION_CREDENTIALS environment variable to authenticate your Google Colab project
What are Google Cloud shared responsibility and shared fate models?
Shared responsibility model: The cloud provider must monitor and respond to security threats related to the cloud itself and its underlying infrastructure. Meanwhile, end users are responsible for protecting data and other assets they store in any cloud environment.
Shared fate model: Focuses on how all parties can better interact to continuously improve security.
1. Security recommendations enabled by default
2. Risk protection program
3. Assured workloads and governance
What is the best practice to secure your workbench?
Use a private IP address
Connect your instance to a VPC network in the same project
Shared VPC network. You can use VPC Service Controls to allow or deny access to specific services
What are the artifacts protected when you use VPC Service Controls?
Training data for an AutoML model or custom model
Models that you created
Requests for online predictions
Results from a batch prediction request
How to secure Vertex AI endpoints?
Public endpoint is publicly accessible to the Internet.
For private endpoints, use private connection to talk to your endpoint without your data ever traversing the public Internet,
How do you secure your Vertex AI training jobs?
Using private IP addresses to connect to your training jobs provides more network security and lower network latency than using public IP addresses.
What is Federated Learning?
Federated learning enables mobile phones to collaboratively learn a shared prediction model while keeping all the training data on the device.
The updated model is then sent for consolidation.
Lower latency, and less power consumption, all while ensuring privacy.