Whizlabs, Practice Questions Flashcards

Google Cloud Certified Professional Cloud Architect

1
Q

You are working for a Startup company as a Solutions Architect. Recently an application was deployed to production. There is a requirement to monitor the key performance indicators like CPU, memory, and Disk IOPS for the application, and also a dashboard needs to be set up where metrics are visible to the entire team. Which service will you use?
A. Use Cloud monitoring to monitor key performance indicators and create Dashboards with key indicators that can be used by the team
B. Use Cloud Logging to monitor key performance indicators and create Dashboards with key indicators that can be used by the team
C. Use Third-party service from marketplace to monitor key performance indicators and create Dashboards with key indicators that can be used by the team
D. Use Cloud Trace to monitor key performance indicators and create Dashboards with key indicators that can be used by the team

A

Option A is correct Cloud monitoring provides detailed visibility into the application by monitoring several key performance indicators like CPU, memory and disk IOPS, etc. You can create dashboards to visualize the performance and also can share with the team to provide detailed visibility into the application performance.
Option B is incorrect because Cloud logging is a fully managed service which allows you to store, search and analyze logs
Option C is incorrect because there is no need to use a third-party service you can use Cloud monitoring for such requirements
Option D is incorrect because Cloud trace is used to detect the latency issues in your application

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

You are working as a Solutions Architect for a large enterprise. They are using the GKE cluster for their production workload. In the upcoming weeks, they are expecting a huge traffic increase and thus want to enable autoscaling on the GKE cluster. What is the command to enable autoscaling on the existing GKE cluster?
A. gcloud container clusters update cluster-name –enable-autoscaling –min-nodes 1 –max-nodes 10 –zone compute-zone –node-pool demo
B. gcloud container clusters create cluster-name –enable-autoscaling –min-nodes 1 –max-nodes 10 –zone compute-zone –node-pool demo
C. You cannot enable autoscaling on existing GKE cluster
D. gcloud container clusters update cluster-name –no-enable-autoscaling –node-pool pool-name [–zone compute-zone –project project-id

A

Option A is correct It is the right command to enable autoscaling on existing GKE cluster gcloud container clusters update cluster-name –enable-autoscaling –min-nodes 1 –max-nodes 10 –zone compute-zone –node-pool demo
Option B is incorrect because it is used to create a new GKE cluster with auto-scaling enabled
Option C is incorrect because you can enable autoscaling on an existing GKE cluster
Option D is incorrect because the command will disable autoscaling on a GKE cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

There is a requirement to make some files from the Google Cloud Storage bucket publicly available to the customers. Which of the below commands you will use to make some objects publicly available?
A. gsutil acl ch -u allUsers:R gs://new-project-bucket/example.png
B. gsutil signurl -d 10m keyfile.json gs://new-project-bucket/example.png
C. gsutil acl ch -g my-domain.org:R gs://gcs.my-domain.org
D. gsutil requesterpays get gs://new-project-bucket

A

Option A is correct This is the right command to make specific files publicly available from Google Cloud storage bucket https://cloud.google.com/storage/docs/gsutil/commands/acl
Option B is incorrect because this command is used to generate a Signed URL which is mostly used to share private content securely for a limited period of time
Option C is incorrect and is used when you have to share the objects with a particular G-suite domain
Option D is incorrect because this enables the requester pay feature on the bucket

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

You are working as a Solutions Architect for a Startup company that is planning to migrate an on-premise application to Google Cloud. They want to transfer a large number of files to Google Cloud Storage using the gsutil command line. How can you speed up the transfer process?
A. Use -m option with gsutil command
B. Use -o option with gsutil command
C. Use du option with gsutil command
D. Use mb option with gsutil command

A

Option A is correct When you have to transfer a large number of files from on-premise to Cloud storage using gsutil command then -m is the best option as it enables parallel multithreading copying ttps://cloud.google.com/storage/docs/gsutil/commands/cp
Option B is incorrect because it is used when you have to copy a file which is large in size
Option C is incorrect because it is used to get object size usage
Option D is incorrect because it is used to create a bucket

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

You are working with a large enterprise as a Solutions architect which is planning to migrate its application from AWS cloud to GCP cloud. There is a requirement to copy data from the AWS S3 bucket to Google Cloud Storage using a command-line utility. How will you fulfill this requirement?
A. Add AWS credentials in the boto configuration file and use the gsutil command to copy data
B. Configure the AWS credentials in gcloud configuration and use the gsutil command to copy files
C. First, download the S3 data using the AWS command-line utility and then copy files to Google cloud storage using gsutil commands
D. Use –s3 flag with gsutil commands to supply AWS credentials while copying files to Google cloud storage

A

Option A is correct You can directly use the AWS S3 bucket as the source or destination while using the gsutil command-line utility. Just you have to put the AWS credentials in the credentials section of the .boto configuration file. https://cloud.google.com/storage/docs/interoperability Options B & D are incorrect because there are no such commands
Option C can be a possible answer but adding AWS credentials in the .boto file is the preferred and easy way.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

A Financial Organization has been growing at a rapid rate and dealing with massive data sets has become an issue. The management has decided to move from on premise to Google Cloud to meet the scaling demands. The data analysts are looking at services which can analyze massive amount of data and can run SQL queries -does data manipulation and visualization in Python. What Google Cloud services can fulfill the requirements?
A. Use Bigquery to run the SQL queries and use Cloud Datalab for detailed data manipulation and visualization in Python.
B. Use Bigtable to run SQL queries and use use Cloud Datalab for detailed data manipulation and visualization in Python.
C. Use Datastore to analyze massive data and use Dataprep for data manipulation and visualization in Python.
D. Use Cloud Spanner to analyze massive data and use Data Studio for data manipulation and visualization in python.

A

Option A is correct. Big Query can analyze large amounts of data also you can run SQL queries , Cloud Datalab does detailed data manipulation and visualization in Python.
Option B is incorrect. Cloud Bigtable is Google’s NoSQL Big Data database service , and it doesn’t support SQL queries ,use it when you need low latency for high writes and high reads.
Option C is incorrect. Cloud Datastore is a NoSQL document database built for automatic scaling, high performance, and ease of application development not suitable for the current scenario and Dataprep is data service for visually exploring, cleaning, and preparing structured and unstructured datasets of any size with the ease of clicks(UI), not code.
Option D is incorrect. The workload is analytics and Bigquery is the right choiceand Data Studio is a decision report generator service. Reference(s): https://cloud.google.com/solutions/time-series/analyzing-financial-time-series-using-bigquery-and-cloud-datalab https://cloud.google.com/datalab/docs/ https://cloud.google.com/bigquery/

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Your organization deals with a huge amount of data and lately, it has become time-consuming and complicated to handle the ever-increasing data volume that needs to be protected and classified based on data sensitivity. The management has set the objective to automate data quarantine and classification system using Google Cloud Platform services. Please select the services that would achieve the objective.
A. Cloud Storage, Cloud Function, Cloud Pub/Sub, DLP API
B. Cloud Storage, Cloud Function, VPC Service control, Cloud Pub/sub
C. Cloud Storage, Cloud Function, Cloud Armour, DLP API
D. Cloud Storage, Cloud Pub/Sub, Cloud Classifier, Cloud Function

A

Option A is the Correct choice because, the data is uploaded to Cloud Storage and later we create buckets example classification_bucket_1 ( for sensitive information) and classification_bucket_2 (for non-sensitive information), use Cloud Function to invoke the DLP API when files are uploaded to cloud storage, use Cloud Pub/Sub topic and subscription to notify when file processing is completed, use Cloud DLP to understand and manage sensitive data(classification).
Option B is Incorrect because VPC service doesn’t help in data classification better choice would be to use Cloud DLP API. VPC Service Controls allow users to define a security perimeter around Google Cloud Platform resources such as Cloud Storage buckets, Bigtable instances, and BigQuery datasets to constrain data within a VPC and help mitigate data exfiltration.
Option C is Incorrect because Google Cloud Armor delivers defense at scale against infrastructure and application Distributed Denial of Service (DDoS) attacks using Google’s global infrastructure and security systems which don’t fulfill the objective set by the management.
Option D is Incorrect because Cloud Classifier is a fictitious service. Using Cloud DLP API serves the purpose of classifying data. Cloud DLP helps you better understand and manage sensitive(protected ) data. The numbers in this pipeline correspond to these steps: You upload files to Cloud Storage. You invoke a Cloud Function. The DLP API inspects and classifies the data. The file is moved to the appropriate bucket. Read more about it here: https://cloud.google.com/solutions/automating-classification-of-data-uploaded-to-cloud-storage https://cloud.google.com/dlp/

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

You are working as a Solutions Architect for a large media company that is planning to migrate its on-premise data warehouse to Google Cloud BigQuery. As a part of the migration, you want to write some migration scripts to interact with BigQuery. Which Command Line utility will you use?
A. gsutil
B. bq
C. gcloud
D. kubectl

A

Option B is correct Bq is a command-line tool for BigQuery which can be used to perform any operations on BigQuery
Option A is incorrect because it is used to interact with Google Cloud storage
Option C is incorrect because Bigquery is having its own command-line utility
Option D is incorrect because kubectl is used to manage Kubernetes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

You are working as a Solutions Architect for a startup company that has recently started using Google cloud for their development environment. The developers want to know if they can persist data on Cloud shell, so they can use Cloud shell for their day to day tasks. What will you suggest to them?
A. Cloud shell can persist up to 10GB data
B. Cloud Shell can persist up to 5GB data
C. Cloud shell data is ephemeral
D. You can attach an additional persistent disk to the Cloud shell

A

Option B is correct Cloud shell comes with 5GB of persistent disk space which is mounted to your $HOME directory where you can keep your data. This persistent disk persists between your sessions.
Option A is incorrect because Cloud shell comes with 5GB of persistent disk
Option C is incorrect because you can persist data on the Cloud shell
Option D is incorrect because you cannot attach an additional persistent disk to the cloud shell session

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

You are working with a startup company as Solutions Architect which is planning to use Google Cloud Storage as a backup location for its on-prem application data. There is a requirement to sync a directory from an on-premise server to Google Cloud bucket. Which gsutil command you will use to sync the data on a daily basis?
A. Use lsync option with gsutil
B. Use rsync option with gsutil
C. Use -m option with gsutil
D. Use mb option with gsutil

A

Option B is correct rsync option is used to sync data between buckets/directories. By using the rsync option only the changed data from the source is copied to the destination bucket https://cloud.google.com/storage/docs/gsutil/commands/rsync
Option A is incorrect because there is no option like lsync
Option C is incorrect because it is used for parallel multithreading copying
Option D is incorrect because mb option is used to create a bucket

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

You are working as a DevOps engineer for an enterprise. Recently one of the microservices was facing intermittent database connectivity issues. This issue was rarely seen and whenever this problem occurs it triggers a few lines in the log file. There is a requirement to set up alerting for such a scenario. What will you do?
A. Use Cloud trace and setup alerting policies
B. Use Cloud logging to set up log-based metrics and set up alerting policies.
C. Manually monitor the log file
D. Use Cloud profiler to set up log-based metrics and set up alerting policies.

A

Option B is correct You can set up a log-based metric that is based on the entries in the log files. For example, you can count the number of occurrences of a specific line entry in the log file and create a metric based on the count. You can also set up alerting policies on the metric if the count goes beyond any threshold value. https://cloud.google.com/logging/docs/logs-based-metrics
Option A is incorrect because Cloud trace is used to detect the latency issues in your application
Option C is incorrect because you need to automate this procedure and also setup required alerting
Option D is incorrect because Cloud Profiler helps you to analyze the CPU and memory usage of your functions in the application

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Your company is migrating the application from AWS to Google Cloud. There is a requirement to copy the data from the AWS S3 bucket to the Google Cloud Storage bucket. Which transfer service would you use to migrate the data to Google Cloud in the easiest way?
A. Storage Transfer Appliance
B. gsutil utility
C. Storage Transfer Service
D. S3cmd

A

Option C is correct Storage Transfer Service is used to quickly transfer data from any other cloud provider to Google cloud storage bucket using Console
Option B can also be used but they have not mentioned any specific command line requirement
Option A is incorrect because it is used to transfer data from on-premise
Option D is incorrect because it used for AWS S3 service

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

You are running a web application on a Compute Engine VM that is using the LAMP stack. There is a requirement to monitor the HTTP response latency of the application, diagnose, and get notified whenever the response latency reaches a defined threshold. Which GCP service will you use?
A. Use Cloud monitoring and setup alerting policies
B. Use Cloud monitoring and setup uptime checks
C. Use Cloud Trace and setup alerting policies
D. Use Cloud Logging and setup uptime checks

A

Option C is correct You can use cloud trace to setup and track latency based metric which will monitor the HTTP response latency and setup alerting policy on this metric which will send an alert when a certain threshold is reached https://cloud.google.com/trace
Option B is incorrect because the uptime check is used to check the system availability Options A & D are incorrect because Cloud trace is used to detect the latency issues in your application

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

You are using gcloud command-line utility to interact with Google Cloud resources. There is a requirement to create multiple gcloud configurations for managing resources. What is the command to create a gcloud configuration?
A. gcloud config create example-config
B. gcloud config configurations activate example-config
C. gcloud configurations create example_config
D. gcloud config configurations create example-config

A

Option D is correct gcloud config configurations create is the right command to create a new configuration gcloud command Options A & C are incorrect because the commands are not right
Option B is incorrect because is used to activate an existing gcloud configuration Ref URL: https://cloud.google.com/sdk/gcloud/reference/topic/configurations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

You are using Cloud shell for accessing Google cloud resources and for your day to day tasks. There is a requirement to install some packages when the Cloud Shell boots. How will you fulfill this requirement?
A. Schedule a cronjob on restart
B. Add the script in /$HOME/.bashrc file
C. Add the script in /$HOME/.profile file
D. Add the script in /$HOME/.customize_environment file

A

Option D is correct To install any packages or run bash script while the cloud shell boots up you must write the script in /$HOME/.customize_environment file. This will install the required things and you view the execution logs in /var/log/customize_environment https://cloud.google.com/shell/docs/configuring-cloud-shell#environment_customization All other options are invalid with respect to cloud shell

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

You are working for a company that is using Google Cloud for its production workload. As per their new security policy, all the Admin activity logs must be retained for at least 5 and will be accessed once a year for auditing purposes. How will you ensure that all IAM Admin Activity logs are stored for at least 5 years keeping cost low?
A. Create a sink to Cloud Storage bucket with Coldline as a storage class
B. Create a sink to BigQuery
C. Create a sink to Pub/Sub
D. Store it in Cloud logging itself

A

A) Option is correct All the admin activity logs are enabled by default and stored in cloud logging. The default retention period for Admin activity logs is 400 Days. If you want to store logs for a longer period, you must create a sink. In our case since logs will be accessed once a year for auditing purposes then Cloud storage sink is the most suitable option.
Option B is incorrect because BigQuery is not a cost-effective solution
Option C is incorrect because Pub Sub is not used for long term storage
Option D is incorrect because Cloud Logging default retention period is 400 days

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Your company recently performed an audit on your production GCP project. The audit revealed that recently an SSH port was opened to the world on a compute engine VM. The management has requested entire details of the API call made. How will you provide detailed information?
A. Navigate to the Logs viewer section from the console, select VM Instance as a resource and search for the required entry
B. Navigate to the Stackdriver trace section from the console, select GCE Network as a resource and search for the required entry
C. Connect to the compute engine VM and check system logs for API call information
D. Navigate to the Stackdriver monitoring section from the console, select GCE Network as a resource and search for the required entry

A

A) Option is correct All the IAM admin related activity logs are stored in the logs viewer section of Cloud Logging. You can see the entire details for an API call made in the logs viewer section of that resource. You can see what network tags were added to the particular VM in this section.
Option B is incorrect because Stackdriver trace is used to collect latency details from applications
Option C is incorrect because system logs will contain all logs related to the operating system only, not the Google cloud resources
Option D is incorrect because Stackdriver monitoring is used to monitor CPU, memory, disk, or any other custom metrics.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

You are working as a Solutions Architect for a large Media Company. They are using BigQuery for their data warehouse purpose with multiple datasets in it. There is a requirement that a data scientist wants full access to a particular dataset only on which he can run queries against the data. How will you assign appropriate IAM permissions keeping the least privilege principle in mind?
A. Grant bigquery.dataEditor at the required dataset level and bigquery.user at the project level
B. Grant bigquery.dataEditor and bigquery.user at the project level
C. Grant bigquery.dataEditor at the project level and bigquery.user at the required dataset level
D. Grant bigquery.admin at required dataset level and bigquery.user at the project level

A

A) Option is correct bigquery.dataEditor on the required dataset will grant write access to the particular Dataset only and bigquery.user at the project level will grant him access to run query jobs in project https://cloud.google.com/bigquery/docs/access-control All other options are incorrect because they are too broad access roles as per our requirement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

You are working for a large enterprise as a DevSecOps engineer. They are running several applications on compute engine VM. The database credentials required by an application are stored in the Cloud Secret Manager service. As per the best practices, what is the recommended approach for the application to authenticate with Google Secret manager service in order to obtain the credentials?
A. Ensure that the service account used by the VM’s have appropriate Cloud Secret Manager IAM roles and VM’s have proper access scopes
B. Ensure that the VM’s have full access scope to all Cloud APIs and do not have access to Cloud Secret Manager service in IAM roles
C. Generate OAuth token with appropriate IAM permissions and use it in your application
D. Create a service account and access key with appropriate IAM roles attached to access secrets and use that access key in your application

A

A) Option is correct In order to access Cloud services for an application running on compute engine VM, you should use a service account attached to the VM. If you are using the default service account you need to set access scope for API’s and also need to attach appropriate IAM roles to the service account https://googleapis.dev/python/google-api-core/latest/auth.html https://cloud.google.com/compute/docs/access/service-accounts
Option B is incorrect because you also need to attach IAM roles to the service account with required Cloud API’s access scope
Option C is incorrect because as per Google’s recommended best practices you should use service account attached with the service
Option D is incorrect because as per Google’s recommended best practices you should use service account attached with the service

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

You have been hired by a large enterprise as a Solutions Architect which has several departments like HR, development, and finance. There is a requirement that they want to control IAM policies for each department separately but centrally. Which hierarchy should you use?
A. A single organization with separate folders for each department
B. A separate organization for each department
C. A single organization with a separate project for each department
D. A separate organization with multiple folders

A

A) Option is correct As per Google recommended best practices you should have multiple folders within an organization for each department. Each department can have multiple teams and projects. By using folders, you can group resources for each department that shares common IAM policies. For example, you have multiple projects for the HR department and want to assign a Compute Instance Admin role to a user for each project in the HR department. You can assign a Compute Instance Admin role to the user at the HR folder level which will grant him access to each project within the HR folder. https://cloud.google.com/resource-manager/docs/creating-managing-folders
Option B is incorrect because you cannot manage IAM Policies centrally if you create separate Organization for each department
Option C is incorrect because each department can have multiple teams and multiple projects under it. So it will become difficult to manage IAM policy centrally for each project within the department
Option D is incorrect because you cannot manage IAM Policies centrally if you create separate Organizations for each department.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

You are working for a Company as a Solutions architect. They want to develop a new application that will have two environments development and Production. The initial requirement is that all the resources deployed in development and Production must be able to communicate with each other using the same RFC-1918 Address space. How will you fulfill the requirement considering the least privilege principle?
A. Create a separate project for each environment and Use shared VPC
B. Create a single GCP project and single VPC for both environments
C. Create a separate project for each environment and create individual VPC in each project with VPC peering
D. Create a separate project and use direct peering

A

Ansawer : A Shared VPC allows you to share a single VPC in one project with another project within an organization called service project. By using shared VPC, the resources in service project can be deployed in shared VPC and they will use the same IP range from shared VPC The main advantage of Shared VPC is that we can delegate administrative responsibilities, such as creating and managing resources which will use one common VPC that allows each team to manage their own resources individually with proper access control In our case, we will create a VPC in production project which will be called a host project, and share it with the development project which will be called a service project. https://cloud.google.com/vpc/docs/shared-vpc
Option B is incorrect because if we use Single project and VPC for both environments we cannot segregate the access control for example if you want to give someone access to create resources only for staging, not production. Such kind of access control is not possible if we are using Single Project and the same VPC
Option C is incorrect because we want the Same RFC-1918 address space. VPC peering is used to connect two different VPC
Option D is incorrect because direct peering is a connection between the on-prem network and Google’s edge network

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

You are working with a large finance company as a Consultant which is planning to migrate petabytes of data from the on-premise data centre to Google Cloud storage. They are having 1gbps network connectivity from on-premise to Google Cloud. Which option will you recommend to transfer data?
A. Storage Transfer Service
B. Transfer Appliance
C. gsutil command-line tool
D. Transfer Service for On-premise

A

Answer - B Since they are having petabytes of data to transfer, a transfer Appliance is the best option. Transfer appliance is an offline data transfer service in which data is transferred via the transfer appliance which comes in two sizes 100TB version and 480TB version
Option A is incorrect because this service scales to available bandwidth and can deliver seamless transfers in just minutes and the available bandwidth is 1gbps, which is too low to transfer petabytes of data.
Option C is incorrect because they are having petabytes of data to transfer and using gsutil command-line utility will take a long time even if the bandwidth is good
Option D is incorrect because it is used when we have data in TB’s. Reference: https://cloud.google.com/storage-transfer/docs/on-prem-overview https://cloud.google.com/transfer-appliance/docs/4.0/overview

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

You are working for a large enterprise as a Solutions architect. They are running several applications on the Compute Engine in Development, Staging, and Production environments. The CTO has informed you that Development and Staging environments are not used on weekends and must be shut down on weekends for cost savings. How will you automate this procedure?
A. Apply appropriate tags on development and staging environments. Write a Cloud function that will shut down compute engine VM’s as per applied the tags. Write a Cron Job in Cloud Scheduler which will invoke cloud functions endpoint on weekends only.
B. Apply appropriate tags on development and staging environments. Write a Cloud function that will shut down compute engine VM’s as per the tags. Write a Cron Job in Cloud Tasks which will invoke cloud functions endpoint on weekends only.
C. Apply appropriate tags on development and staging environments. Write a Cloud function that will shut down compute engine VM’s as per the applied tags. Write a Cron Job in Cloud build which will invoke cloud functions endpoint on weekends only.
D. Apply appropriate tags on development and staging environments. Write a Cloud function that will shut down compute engine VM’s as per the applied tags. Write a Cron Job in Cloud Run which will invoke cloud functions endpoint on weekends only.

A

Answer: A
Apply tags to the development and staging Compute Engine VM’s. Write a Cloud Functions using any preferred language which will filter the VM’s based on applied tags and will shut down them. Select the trigger type as HTTP while configuring cloud function and write a cronjob in Cloud Scheduler which will trigger the HTTP endpoint only on a weekly basis. https://www.google.com/search?client=firefox-b-d&q=gcp+cloud+scheduler
Option B is incorrect because Cloud Task is used for management of a large number of distributed tasks
Option C is incorrect because Cloud Build is used to create CICD pipelines
Option D is incorrect because Cloud Run is used to run Containers where the entire infrastructure management is fully handled by GCP

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

For this question, refer to the Dress4Win case study: “https://cloud.google.com/certification/guides/cloud-architect/casestudy-dress4win-rev2 In the initial phase of migration how will you isolate development and test environments?
A. Create a separate project for testing and separate project for development
B. Create a Single VPC for all environments, separate by subnets
C. Create a VPC network for development and separate VPC network for testing
D. You cannot isolate access between different environments in Google cloud

A

Answer: A
As per the IAM best practices, you should create a separate project for each environment to isolate each environment. https://cloud.google.com/blog/products/gcp/iam-best-practice-guides-available-now
Option B is incorrect because as per IAM best practice you should create a separate project for each team
Option C is incorrect because you cannot isolate each env by creating 2 VPC in the same project. If anyone has permission to start/stop VM he can stop both environments VM’s if they are in the project
Option D is incorrect because you can isolate env’s by creating a separate project for each

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

You are working for a company which develops online games. Recently one of their online games started becoming more popular which is deployed on a compute engine. As the traffic is increasing they are struggling to provision additional instances globally for any time of the day. How will you design the architecture which will meet the demand of growing users and maintain the performance globally?
A. Use Global Load balancer and Managed Instance Group
B. Use Global Load balancer and Unmanaged Instance Group
C. Use Regional Load balancer and Managed Instance Group
D. Use Regional Load balancer and Unmanaged Instance Group

A

Answer: A
As the game is becoming more popular globally they should use Global load balancer and Managed instance groups deployed in several regions in multiple zones. Using global load balancer will distribute the traffic to the managed instance group which is closer to the user automatically. Enable autoscaling on Managed instance groups to dynamically scale up and down as the traffic increases.
Option B is incorrect because unmanaged instance group does not support autoscaling
Option C is incorrect because Regional load balancer cannot load balance managed instance group deployed in multiple regions
Option D is incorrect because unmanaged instance group does not support autoscaling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

You are working as a Solutions Architect for a large financial firm. The data scientists team wants to run batch jobs on a nightly basis which will perform data analytics. These jobs can be disrupted or restarted and will use Spark and Hadoop clusters. Which GCP managed services will you use to keep analytics processing fast, easy, and more secure and cost-effective?
A. Use Cloud Dataproc with preemptible compute engine option.
B. Run Spark and Hadoop clusters on a preemptible compute engine.
C. Run Spark and Hadoop clusters on a standard compute engine.
D. Use Cloud Dataproc with standard compute engine option.

A

Answer: A
As they want to run data analytics jobs which will be using Hadoop and spark clusters. Dataproc is a good option because it is a managed service based on Hadoop and spark which is used for ETL workload and data analysis. https://cloud.google.com/dataproc/docs Dataproc clusters can use preemptible VM instances, which will result in huge cost saving https://cloud.google.com/dataproc/docs/concepts/compute/preemptible-vms
Option B is incorrect because they want a managed service
Option C is incorrect because they want a managed service
Option D is incorrect because they want a cost-effective solution so standard compute engine is not a good choice

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

You are working for a large enterprise where the management of network and security resources such as firewalls are typically managed by a dedicated Security team for the entire organization. The development teams only want flexibility to launch instances and carry out other actions related to instances in the dev project only. How will you grant respective IAM permission to the development team and security team keeping the least privilege principle in mind?
A. Compute Network Admin role for Security team at Organization level and Compute Instance Admin role for development team for dev project only
B. Compute Network Admin role for Security team at Organization level and Compute Instance Admin role for development team at organization level
C. Compute Network Admin role for Security team at Organization level and Compute Network Admin role for development team for dev project only
D. Compute Instance Admin role for Security team at Organization level and Compute Network Admin role for development team at Organization level

A

Answer: A
Assign Compute Network Admin role to the Security team at the Organization level. This will grant them Permissions to administer networking resources. The network admin role does not allow the security team to control the compute engine resources. By assigning this role at Organization level the Security team will have access to every project within an organization Assign Compute Instance Admin role to the Development team at a specific project-level i.e. dev project. This will grant the dev-team full access to compute engine resources only and read-only access to networking resources. By assigning this role at a specific project level will grant them access to resources in that particular project only. https://cloud.google.com/compute/docs/access/iam https://cloud.google.com/iam/docs/resource-hierarchy-access-control
Option B is incorrect because it will grant dev-team access to all projects under the organization.
Option C is incorrect because the Network admin role will not allow dev-team to control compute resources
Option D is incorrect because Compute Instance admin role will not allow the Security team to administrate networking resources and Network Admin role will not allow the dev team to administrate compute resources

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

You have been hired as a DevSecOps Engineer by an enterprise who are planning to migrate their application on Google Cloud Platform but as per the compliance requirement they want to use their existing Active Directory domain to manage user identities. What you should suggest in this scenario?
A. Use Google Cloud Directory Sync to sync Active Directory username with Cloud Identity
B. Use Identity-Aware Proxy configured with your Active Directory Domain
C. There is no option for using Active Directory Domain. Use G-Suite for user management
D. Create an Active Directory domain controller on Compute Engine that is a replica of on-premise AD and use Google Cloud Directory Sync

A

Answer: A
By using Google Cloud Directory Sync you can sync Active directory username with Cloud Identity. In order to sync users and groups, you need to install GCDS agent in you AD servers https://support.google.com/a/answer/106368?hl=en
Option B is incorrect because Identity aware proxy lets you manage access to the applications which are running on App Engine, Kubernetes engine and VM’s
Option C is incorrect because you can sync Active directory users using GCDS
Option D is incorrect because there is no need to move AD servers to compute engine, you can directly install GCDS agent on AD servers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

You have been hired as Solutions Architect by a large enterprise who has recently migrated to GCP. The database warehouse team came to you as they want to know which managed service they can use for cleaning, preparing structured and unstructured data for analysis, reporting, and machine learning?
A. Cloud Dataprep
B. Cloud Dataproc
C. Cloud Dataflow
D. Cloud Datalab

A

Answer: A
Cloud Dataprep is a serverless service that can be used for large dataset cleaning and preparing the data for analysis and reporting. It provides a GUI for cleaning and preparing the data. https://www.youtube.com/watch?v=Q5GuTIgmt98
Option B is incorrect because it is used to run Apache spark and Hadoop clusters
Option C is incorrect because dataflow is used for real-time and batch processing of data
Option D is incorrect because Datalab is used to visualize data and build machine learning models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

You are working for a large Finance company as a Solutions architect. They have multiple applications running in production. All the applications log data is stored in GCS bucket for future analysis to improve the application performance. What is the recommended approach to De-identify personally identifiable information or payment card information stored in logs?
A. Use Cloud DLP
B. Use thread detection
C. Use Web Security Scanner
D. Use Cloud Armor

A

Answer: A
Cloud DLP is a fully managed service used to de-identify sensitive data like credit card numbers, Phone numbers, and any other PII information stored in text files within cloud storage and Bigquery. After detecting sensitive data the DPL API provides various options like mask the data or delete the data https://cloud.google.com/dlp/docs/deidentify-sensitive-data
Option B is incorrect because it is used to detect threats like Burt force attack from logs and reports to Security command center
Option C is incorrect because it is used to find any vulnerable library used in your application code
Option D is incorrect because it is used to mitigate DDoS attack and provide WAF

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

You are working for a company that is planning to migrate its entire application to GCP. During the initial phase of migration, there is a requirement to set up a site-to-site VPN connection between on-prem and GCP which provides 99.99 availability on the GCP side connection. Which service will you use?
A. Cloud HA VPN
B. Cloud Classic VPN
C. Direct Peering
D. Configure Openswan on two compute engine instances and create two VPN tunnels

A

Answer: A
Cloud HA VPN provides an SLA of 99.99% service availability. https://cloud.google.com/network-connectivity/docs/vpn/how-to/creating-ha-vpn2?hl=nl
Option B is incorrect because Cloud classic VPN provides 99.9 availability.
Option C is used to connect on-premise location to Google’s Point of presence location(PoP)
Option D can also work but there will be lots of management work so it is not preferable. GCP has its fully managed service i.e. Cloud HA VPN

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

You are working as Solutions Architect for Company which is running the entire application on-premise. There is a new requirement to migrate the SQL server enterprise edition to GCP which runs in the availability group for High availability in the datacenter. Which option you will choose from below which will provide less management work ahead and can also provide data redundancy?
A. Create a Cloud SQL server instance with high availability option enabled
B. Create a Compute instance in the different zone within a region and install SQL server with always-on availability groups for data redundancy
C. Create a Cloud SQL instance, by defaults it comes with high availability
D. Create a Compute instance in a single zone with always-on availability groups

A

Answer: A
Cloud SQL is a fully managed service where Google manages all the heavy lifting work like patching, failover, backups and replication. Cloud SQL server instance is the best choice with a high availability option enabled on it. When you enable High Availability(regional) option, if there is an outage, your instance fails over to another zone in the region where your instance is located There are also several licensing options available for Cloud SQL. https://cloud.google.com/sql/docs/sqlserver/high-availability
Option B is incorrect because to reduce the management work ahead, we will be using managed service i.e Cloud SQL
Option C is incorrect because we need to enable the high availability option while creating Cloud SQL
Option D is incorrect because it will not provide high availability and also will not reduce management work

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

You are running Apache Kafka on a compute engine for real-time data processing pipeline. The machine size is n1-standard-4 with 1TB of SSD persistent disk and as per the monitoring, you are not getting the desired disk throughput required for the job to do. What configuration will you change to increase disk performance?
A. Increase the machine to n1-standard-8
B. Increase the disk size to 2TB
C. Increase the machine memory
D. change the storage type to standard persistent disk

A

Answer: A
Disk performance depends on its size, instance vCPU count and I/o block size In our case, we are already having a large disk size i.e. 1TB which can support 480 Mbps read/write throughput. But as per our machine size i.e. n1-standard-4 (4vcpu), the disk is only limited to 240mbps read/write throughput. We need to increase the CPU count to 8 or above to support desired disk performance. For example, consider a 1,000 GB SSD persistent disk attached to an instance with an N2 machine type and 4 vCPUs. The read limit based solely on the size of the disk is 30,000 IOPS. However, because the instance has 4 vCPUs, the read limit is restricted to 15,000 IOPS.” https://cloud.google.com/compute/docs/disks/performance#size_price_performance https://cloud.google.com/compute/docs/disks/performance#machine-type-disk-limits
Option B is incorrect because we already have large disk size, the bottleneck was CPU
Option C is incorrect because RAM does not limit the disk performance
Option D will more degrade the performance. see above URL for comparison

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

You are working for a large enterprise as a Solutions Architect. The development team is building a new application that will be deployed on Compute Engine. How will you set compute engine VM configuration in such a way that there is no downtime when GCP performs periodic infrastructure maintenance on the compute engine?
A. Set the on-host maintenance option to Migrate VM instance
B. Set the Automatic restart option to ON
C. You need to restart VM when there is such kind of maintenance activity from GCP
D. Set the on-host maintenance option to Terminate VM instance

A

Answer: A
GCP performs maintenance activity on compute engine infrastructure which includes. Host kernel upgrades, hardware repair, or upgrade. This activity occurs once every two weeks. You can configure compute engine VM to perform live migration to another host in case of such maintenance activity without downtime. You just need to set instance On host maintenance property to Migrate VM instance and the entire process is handled by GCP on your behalf. You can see compute. instances. migrateOnHostMaintenance operation type performed in Operations Suite (formerly Stackdriver) logging when such activity is carried out. https://cloud.google.com/compute/docs/instances/live-migration https://cloud.google.com/compute/docs/instances/setting-instance-scheduling-options
Option B is incorrect because it is used when the host machine crashes which holds your VM. If this property is enabled, whenever there is a host machine failure. Your compute engine will be automatically restarted
Option C is incorrect because there is no need to perform any kind of operation from your side
Option D is incorrect because if the property is set to Terminate VM instance, GCP will terminate your VM when there is a maintenance event.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

You are working for a Media company as a Solutions Architect. They are having a mobile application which is used by journalists to capture and upload images on a daily basis to the GCS bucket from a different location for any Breaking News. There is a requirement to process these file images in real-time to detect any offensive content and if there is any offensive content it should be made blur and re-uploaded to the bucket. Which services will you include in your Architecture?
A. Cloud Functions, Cloud Vision API
B. Cloud functions, Cloud ML Engine
C. App Engine, Cloud Vision API
D. Cloud Tasks, Cloud Vision API

A

Answer: A
Google Cloud’s Vision API is an AI service provided by GCP to detect objects in an image, detect any explicit content in images, and also can extract text from images. As soon as the image is uploaded to the GCS bucket Cloud Function is invoked which will call Vision API and perform Offensive Image Detection operation. If any offensive image is detected another Cloud function will be called which will make the offensive content Blur using python pillow library and upload it to the same bucket.
Option B is incorrect because Cloud ML engine is used to train machine learning models
Option C is incorrect because we will need event-based service for such kind of requirement
Option D is incorrect because Cloud Tasks is a fully managed service used to manage distributed tasks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

You have been hired by a large U.S based healthcare firm as a Consultant which is planning to migrate entire application and on-premise data to Google Cloud. As the data includes medical records of different hospitals located in the U.S. What regulations would you look to for more guidance on complying with relevant regulations?
A. HIPAA
B. PCI-DS
C. GDPR
D. SOX

A

Answer: A
HIPAA (Health Insurance Portability and Accountability Act) is regulatory compliance in the U.S which is used to protect the healthCare data collected by websites and application for business purpose in the U.S https://cloud.google.com/security/compliance/hipaa-compliance
Option B is incorrect because it is a Payment Card Industry Data Security Standard to protect credit card information collected for business
Option C is incorrect because GDPR(General Data Protection Regulation) is regulatory compliance in Europe which is used to protect any personally identifiable information collected for business purpose within the Europe region
Option D is incorrect because
Option SOX compliance is used for financial auditing purpose

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

There is a new requirement to Deploy a web application on Google Kubernetes Engine which will be accessed by multiple users around the world. How will you enable autoscaling on the application which will scale automatically based on the CPU Utilization?
A. Create a HorizontalPodAutoscaler with CPU as target and enable autoscaling on your GKE cluster
B. Create a HorizontalPodAutoscaler with CPU as target and enable autoscaling on your managed instance group
C. Create a Deployment with the max unavailable and max surge properties and enable autoscaling on your GKE cluster
D. Create a Deployment with the max unavailable and max surge properties and enable autoscaling on your managed instance group

A

Answer: A
Horizontal Pod Autoscaler is used to automatically scale the pods in a deployment based on the CPU utilization or memory utilization Kubectl autoscale command is used to create HorizontalPodAutoscaler kubectl autoscale deployment example-app –max 5–min 2–cpu-percent 60 You can also enable autoscaling on your GKE cluster which can add or remove nodes from node pool based on the demands of your workloads You can use gcloud command to enable autoscaling on your GKE cluster gcloud container clusters update example-cluster –enable-autoscaling \ –min-nodes 2–max-nodes 6–zone compute-zone –node-pool default-pool
Option B is incorrect because you need to enable autoscaling on GKE cluster not managed instance group Options C & D is incorrect because deployment is a Kubernetes object which is used to run multiple replicas of your pod and will automatically replace any failed or unresponsive pod

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

You are working for a company that has several applications running on a compute engine. Daily files are uploaded to the GCS bucket from these instances. These files are accessed once a month by developers for analysis. After 1 year all the files are accessed only once a year but must be retained for 5 years as per compliance. How will you configure data storage in a cost-effective way?
A. Set the default storage class of the bucket to the near line and create a lifecycle rule to move objects older than 1 year to Coldline storage class
B. Set the default storage class of the bucket to standard and create a lifecycle rule to move objects older than 1 year to nearline storage class
C. Set the default storage class of the bucket to standard and create a lifecycle rule to move objects older than 1 year to Coldline storage class
D. Set the default storage class of the bucket to Coldline storage.

A

Answer: A
Set the default class to Nearline Nearline Storage is the best choice when you want to access objects stored in the bucket once a month After one year as the files stored in the bucket will be accessed only once a year then you should create lifecycle rule to migrate nearline objects to cold line storage https://cloud.google.com/storage/docs/lifecycle
Option B &
Option C are incorrect because standard storage is used for objects which are accessed very frequently
Option D is incorrect because Coldline storage is used for objects which are accessed once a year Note - GCP has launched a new storage class called Archival storage which was Generally available on January 08, 2020. This may reflect in exam https://cloud.google.com/storage/docs/storage-classes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

You are working for a company that is using GCP for their production workload. One of the applications is using Cloud CDN for static content caching in front of the https load balancer. As per the cloud logging, you see lower than expected cache hit ratios. How will you increase the cache-hit ratio?
A. Use custom cache keys
B. Increase the cache expiration time
C. Use cache invalidation frequently
D. Decrease cache expiration time

A

Answer: A
To improve the cache hit ratio you should reduce the cache key by removing host and protocol information. This final URL is called a custom cache key For e.g. https://demo.com/test/cloud.jpg and https://demo2.com/test/cloud.jpg have the same image i.e. cloud.jpg but URL is different you can remove protocol and host information from the cache key https://cloud.google.com/cdn/docs/best-practices
Option B is incorrect it used to define the time that how long content is cached at PoP location
Option C is incorrect because it is used to clear cache entry manually
Option D is used to when you have an application where content is frequently updated So you can keep low cache expiration time on cache contents

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

You are working for a large enterprise as a GCP Cloud Architect. As per the new compliance requirement, you should regularly save your all admin activity and VM system logs within your project centrally for third party auditing which will happen once every month. How will you achieve this requirement keeping the cost low?
A. All admin and VM system logs are automatically collected by Stackdriver, just create sink for selected logs to GCS nearline bucket
B. Stackdriver automatically collects admin activity logs for most services. Only the Stackdriver Logging agent must be installed on each instance to collect system logs and create sink for selected logs to GCS nearline bucket
C. Stackdriver automatically collects admin activity logs for most services. Only the Stackdriver Logging agent must be installed on each instance to collect system logs and create sink for selected logs to GCS cold storage bucket
D. All admin and VM system logs are automatically collected by Stackdriver, just create sink for selected logs to GCS cold storage bucket

A

Answer: B
Admin activity logs are automatically collected for most of the services in GCP. For the VM system logs, you need to install a Logging agent in each VM whose logs you want to export to stackdriver logging. As per the compliance requirement, you must retain logs for auditing for that you should create a sink to GCS nearline bucket. These logs will be accessed once a month that’s why the nearline bucket is the best storage option. https://cloud.google.com/logging/docs/agent https://cloud.google.com/logging/docs/audit
Option A is incorrect because VM system logs are not automatically collected. You need to install a stackdriver agent to get VM system logs.
Option C is incorrect because the audit will happen once a month and for that Coldline storage is not a good option
Option D is incorrect because VM system logs are not automatically collected you need to install stackdriver agent to get VM system logs and also coldline is not a right storage option

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

You have been hired as a DevOps Engineer by a large finance company. As per their regulatory compliance the CTO has informed you that any resources which will be created in Google Cloud must be created in the U.S region only and all other regions are restricted by default. How can you restrict the resources creation limited to the U.S region only?
A. Create a custom IAM policy at Organization level
B. Create an Organization Policy at Organization level
C. Create an Organization policy at individual project level
D. You cannot apply such kind of restriction in Google cloud

A

Answer: B
An organization policy is a configuration of restrictions. You can create Organization policy at Organization level which will inherit to all resource under it and with Constraint for Google Cloud Platform - Resource Location Restriction set to the U.S only https://cloud.google.com/resource-manager/docs/organization-policy/overview https://cloud.google.com/resource-manager/docs/organization-policy/org-policy-constraints
Option A is incorrect because IAM policy is attached to resources which are used to define access control
Option C is incorrect because we want to apply the restriction for all the projects under the organization, not a specific project
Option D is incorrect because we can have such kind of restriction using Organization Policy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

You have been hired as a DevSecOps Engineer by a large finance company. As per their regulatory compliance the CTO has informed you that by default all VM instances which are created in the entire organization are not allowed to use external IP addresses. How can you fulfill this requirement?
A. Create a custom IAM policy at Organization level
B. Create an Organization Policy at Organization level
C. Create an Organization policy at individual project level
D. You cannot apply such kind of restriction in Google cloud

A

Answer: B
An organization policy is a configuration of restrictions. You can create Organization policy at Organization level which will inherit to all resource under it and with Constraints for Compute Engine service which include Define allowed external IPs for VM instances set to Deny All https://cloud.google.com/resource-manager/docs/organization-policy/overview https://cloud.google.com/resource-manager/docs/organization-policy/org-policy-constraints
Option A is incorrect because IAM policy is attached to resources which are used to define access control
Option C is incorrect because we want to apply the restriction for all the projects under the organization, not a specific project
Option D is incorrect because we can have such kind of restriction using Organization Policy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

You have been hired as a solutions architect by Large Finance firm. The development team is developing an application which will be hosted on Google cloud and will access the Oracle Database in its own datacenter. The Network engineers have determined that a link between the on-premises network and GCP will require an 8 Gbps connection and low latency to meet the business requirements with an SLA. Which Option you will select?
A. Dedicated Interconnect
B. Partner Interconnect
C. Cloud VPN
D. Hybrid Interconnect

A

Answer: B
B Is right because Partner Interconnect is good up to 10Gbps and provides SLA also https://cloud.google.com/network-connectivity/docs/interconnect/concepts/partner-overview
Option A is incorrect because Dedicated interconnect is suitable and cost-effective above 10Gbps
Option C is incorrect option because it is not suitable for High-speed connections where latency is a key requirement
Option D is incorrect because there is no such service

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

You have been hired as a Solutions Architect by a large enterprise. They are planning the migration of their on-premise application to GCP. The application ingests time-series data at low latency collected from sensors from chemical plants located across different locations. They are using Cassandra clusters as database storage and RabbitMQ as a messaging service. One of the business requirements is to maximize the use of managed services while moving to GCP. Please select services as per the business requirements
A. Use Cloud Datastore and Pub/Sub
B. Use Cloud Bigtable and Pub/Sub
C. Use Cloud Bigquery and Pub/Sub
D. Use Dataproc and Pub/Sub

A

Answer: B
Cloud Bigtable is the best choice when you want to ingest time series data from sensors at low latency. It is a fully managed service used for large NoSQL analytical workloads. https://cloud.google.com/bigtable As they are using RabbitMQ as a messaging service on-premise and want to move to managed service while migration then Pub/Sub is a good choice. Pub-Sub is a fully managed service which provides asynchronous service to service communication mostly used in event-driven architectures https://cloud.google.com/pubsub/docs/overview
Option A is incorrect because Datastore is not ideal for where low latency is a key requirement.
Option C is incorrect because Bigquery is used for SQL data
Option D is incorrect because it is used to run Apache Hadoop and Spark clusters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

You are working for Media Company as a Solutions Architect. There is a new requirement that the visual effects artists team requires a file share system that can be easily mounted on several Compute Engine instances for media workflow processing like video editing and video rendering which usually require common file share. Which storage solution will you use for this kind of scenario?
A. Cloud Storage
B. Cloud Filestore
C. Relational database
D. Cloud datastore

A

Answer: B
Cloud filestore is a fully managed network-attached storage which uses NFS protocol where multiple Linux instances can mount a common file share over a network. https://cloud.google.com/filestore
Option A is incorrect because Cloud storage is object storage and cannot be mounted on compute engines
Option C is incorrect because the Relational database is used to store SQL data
Option D is incorrect because Cloud datastore is a NoSQL database

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

You are working as a consultant for a company which has thousands of IoT devices installed in several chemical plants for monitoring humidity, temperature and electrochemical gas. There is a requirement to capture the data from this sensor in real-time, ingest it, run through a data processing pipeline and store it for analysis. SQL queries will be run against data for analysis and also there is a requirement for a data visualization tool that can analyze the data interactively. Which architecture you will suggest for the above requirements?
A. Cloud IoT core, Pub/Sub, Dataproc, Bigtable, Data Lab
B. Cloud IoT core, Pub/Sub, Dataflow, Bigquery, Data studio
C. Cloud IoT core, Pub/Sub, Dataprep, Biquery, Data Lab
D. Cloud IoT core, Pub/Sub, Dataflow, Bigtable, Data studio

A

Answer: B
Cloud IoT core, Pub/Sub, Dataflow, Bigquery, Data studio is the correct option. Cloud IoT Core is a fully managed service which will accept data from sensors and will manage the connection with sensors. After the data arrives at IoT core it is sent to Pub/Sub which will act as an asynchronous message bus and further this real-time data is processed by data flow and stored in Bigquery for analysis as they want to run SQL queries so Bigurey is the best choice. You can use Data studio which will use Bigquery as a source and create dashboards and reports for visualization as per requirement.
Option A is incorrect because Dataproc is used to run Hadoop and Spark clusters
Option C is incorrect because data prep is used to cleanse and prepare data for analysis and machine learning
Option D is incorrect because we want to run SQL queries against the data so BigTable is not the right choice as it is NoSQL database

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

You have been hired by a large enterprise as a Solutions Architect. The development team came to you with a requirement that they want a global load balancing solution that can support Non-HTTPS traffic and SSL termination at the load-balancing level. Which load balancer will you recommend?
A. HTTPS
B. SSL Proxy
C. TCP Proxy
D. Internal TCP/UDP

A

Answer: B
SSL Proxy load balancer is the best choice for non-https traffic and can also handler SSL termination. It is a Global load-balancing Solution Provided by GCP https://cloud.google.com/load-balancing/docs/choosing-load-balancer
Option A is incorrect because HTTPS load balancer is used for HTTP traffic
Option C is incorrect because the requirement is to terminate the SSL at the load balancing level. TCP proxy does not support SSL termination
Option D is incorrect because Internal TCP/UDP load balancer is used to load balancer internal traffic inside a VPC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

You are working for a company as a Solutions architect. The Development team is developing a new stateful application that will be deployed on the Google Kubernetes Engine. What type of Kubernetes resource will you create for stateful application?
A. Pods
B. StatefulSets
C. Deployments
D. DaemonSets

A

Answer: B
StatefulSets are used for stateful applications where you want to persist application data. When you create StatefulSet, replica pods are created in order and each replica pod have its unique id, own PVC and state https://cloud.google.com/kubernetes-engine/docs/concepts/statefulset
Option A is incorrect because the pod is the smallest unit of Kubernetes and mostly managed by Kubernetes objects like deployment, replica set, and StatefulSet
Option C is incorrect deployments are mostly used for stateless application
Option D is incorrect because DaemonSets are used when you want a run a copy of each pod on each node in Kubernetes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

You are working for a large enterprise as a Solutions Architect. One of your application is running on-premise. There is a requirement that an application running on google cloud needs to access a few APIs of the on-premise applications without exposing them to the internet. Which type of topology will you implement to fulfill the requirements?
A. Meshed topology
B. Gated egress topology
C. Gated egress and ingress topology
D. Gated Ingress topology

A

Answer: B
This type of topology is useful when you want to expose on-premise application API’s to the workload running on Google Cloud without exposing them to internet Please refer to below link for different hybrid and multi-cloud network topologies https://cloud.google.com/solutions/hybrid-and-multi-cloud-network-topologies
Option A is incorrect because Meshed topology is used to establish flat network connectivity where every system can communicate with each other
Option C is incorrect because Gated egress and ingress topology is used when you have to expose a few API’s from on-premise to cloud and from cloud to on-premise in a secure way
Option D is incorrect because it used when you want to expose a few API’s from an application which is running on Google Cloud to On-premise in a secure way

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

You are working for a large enterprise as a Solutions Architect. As per their compliance requirement all the data which is stored in Cloud SQL, Compute Engine and Cloud storage must be encrypted by customer-managed encryption keys with rotation schedule for symmetric keys to automatically generate a new key, please suggest the right choice for encryption?
A. Use default encryption which is provided by Google Cloud
B. Use CMEK using Cloud KMS
C. Use CSEK
D. Use third party service from Marketplace for customer-managed-encryption

A

Answer: B
Use CMEK using Cloud KMS Customer-managed encryption keys (CMEK) using Cloud KMS lets you create your own encryption keys in Cloud KMS where you can create, rotate, automatically rotate and destroy symmetric encryption keys https://cloud.google.com/storage/docs/encryption/customer-managed-keys
Option A is incorrect because default encryption is fully managed by GCP from creating keys to encrypting the data and storing the keys and rotating them
Option C is incorrect because CSEK is used when there is a requirement to store the encryption keys on-premise and only supports two services i.e cloud storage compute engine
Option D is incorrect because you cannot use a third-party solution for encrypting GCP services

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

You are working for a large finance company as a Solutions Architect. As per the FINRA compliance regulation, the data stored in GCS buckets must be retained for 5 years. How can you ensure that the current objects or any objects uploaded to the buckets are not deleted for at least 5 years?
A. Apply Lifecycle rules to buckets
B. Apply retention policy to buckets
C. Apply IAM policy with appropriate roles
D. Enable versioning on buckets

A

Answer: B
when you set a retention policy on the bucket you cannot delete any objects in that bucket for the specified period of time mentioned in a retention policy https://cloud.google.com/storage/docs/using-bucket-lock
Option A is incorrect because lifecycle rules are used to move objects between different storage classes
Option C is incorrect because it is used to control access management
Option D is incorrect because versioning is used to create multiple versions of a single object

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

You are working for a Company as a Consultant which has recently acquired a Software Company which has their entire application on Google Cloud Platform. There is a new requirement that the application in your GCP VPC requires RFC 1918 connectivity to VPC in the acquired GCP account. How will you create connectivity?
A. Shared VPC
B. Cloud VPN
C. VPC Peering
D. Direct Peering

A

Answer: C

Option A is incorrect because it is used to share the VPC from the host project to service projects within an organization
Option C is correct because VPC peering is always preferred when you want to connect two VPC’s within GCP cloud because the traffic stay’s inside Google’s private network
Option D is incorrect because direct peering is a connection between the on-prem network and Google’s edge network

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

You have been hired as Consultant for a company which is planning migration of their enterprise application to GCP. As the company holds sensitive data and has the requirement to generate own encryption keys and manage it on-premises as per their regulatory compliance. The CTO has asked you to list the Google Cloud Products Which Supports Customer-supplied Keys(CSEK) before they perform migration. Please select the services which support CSEK.
A. All Google Cloud Products support CSEK
B. Compute Engine, Cloud Storage and Cloud SQL
C. Compute Engine and Cloud Storage
D. BigQuery, Cloud SQL and Datastore

A

Answer: C

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

You have been hired as DevSecOps Engineer by a finance firm. They are developing a new application that will be used for financial transactions thus needs to be PCI compliant and will be deployed on a compute engine. As per the security team, the infrastructure on which the application will run must be hardened by security controls to protect against rootkits and bootkit. Which compute engine option you will use?
A. Enable encryption on the Boot disk
B. Use Sole-Tenant VM
C. Use Shielded VM
D. Use Preemptible VM

A

CSEK is a feature in Google Cloud Storage and Google Compute Engine services https://cloud.google.com/security/encryption-at-rest/customer-supplied-encryption-keys Options A, B & D are incorrect because only Cloud Storage and Compute Engine supports CSEK.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

You have been hired as Consultant by an enterprise. The company is running their production workload on Google Cloud. One of your clients requested a penetration testing report for your application and your CTO has decided to hire a Security specialist to perform penetration testing on your application, what is the procedure to conduct penetration testing on Google Cloud?
A. You need to raise a support ticket with Google cloud for permission to perform Penetration testing
B. Google Cloud does not allow to perform any kind of penetration testing
C. You do not have to notify Google when conducting a penetration test on your application
D. Raise a support ticket with Google to perform penetration testing on your behalf

A

Answer: C
Shielded VM is an option in a compute engine instance that comes with a set of security controls which helps to protect against rootkits and bootkits. For an application which required hardened OS, Shielded VM is a good option https://cloud.google.com/shielded-vm
Option A is incorrect because by enabling encryption on the boot disk will only encrypt the data. It will not protect against rootkits and bootkit
Option B is incorrect because this option provides us a dedicated physical server, which is allotted to us only for running compute engine instances
Option D is incorrect because Preemptible instances are short-lived instance which can run for max 24 hour and provide huge cost saving as compared to standard instances

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

For this question, refer to the TerramEarth case study: https://cloud.google.com/certification/guides/cloud-architect/casestudy-terramearth-rev2 Initially, TerramEarth will be testing BigQuery service as the preferred replacement of their On-Premise data warehouse system. During the testing phase, they only want access to the most recent data on BigQuery. Any data older than 15 days must be deleted to optimize storage use. How will you fulfill this requirement?
A. Set the default table expiration to 15 days
B. Create a script using bq that removes records older than 15 days
C. Take advantage of BigQuery long-term storage
D. Make the tables Date-partitioned, and configure the partition expiration at 15 days

A

Answer: C
You can perform penetration testing on your application without informing Google Cloud but you must satisfy all the terms and conditions of Google Cloud https://support.google.com/cloud/answer/6262505?hl=en.
Option A is incorrect because there is no need to raise a support ticket to conduct penetration testing on your application
Option B is incorrect because you can perform penetration testing on GCP
Option D is incorrect because google does not perform penetration testing on your behalf. You can perform yourself without notifying google cloud

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

You have been working as a Solutions Architect for a company who has recently developed an online mobile game which will be mostly used for children ages 10 to 14 and will be deployed on Google Cloud in the us-west1 region. The online game will collect the personal information of the player such as name, address, age and hobbies. With which regulation would you advise them to comply with?
A. HIPAA
B. PCI-DS
C. GDPR
D. COPPA

A

Answer: D
As TerramEarth will be testing BigQuery initially, they don’t want data older than 15 days You can partition the table based on date and set the default table expiration to 15 days which will automatically delete data older than 15 days providing you the most recent data. https://cloud.google.com/bigquery/docs/best-practices-storage
Option B is incorrect because there is no read to write a script, it can be done by the default table expiration feature on BigQuery
Option C is incorrect because this is used when you have a table that is not edited for the last 90 Days. After 90 days the storage price drops by 50% which is similar to Nearline storage pricing
Option A is incorrect because it will directly set the default table expiration time to 15 days which will delete data older than 15 days. Please refer to https://cloud.google.com/bigquery/docs/managing-tables for more information

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

You have been hired as a Solutions Architect by a large enterprise. They are planning to migrate an application that is running in the AWS cloud to the GCP cloud. During the initial phase of migration, there is a requirement to create RFC-1918 connectivity with a minimum of 5Gbps bandwidth between AWS VPC and GCP VPC for secure migration. What service you will use at the GCP side with the least management work ahead?
A. Use a Cloud HA VPN
B. Use an OpenSwan VPN solution on the Compute engine with more CPU
C. Use VPC Peering
D. Use Cloud Partner Interconnect

A

Answer: D
COPPA is regulatory compliance in the U.S which is related to protecting the privacy of children below 13 age in the U.S https://www.ftc.gov/tips-advice/business-center/guidance/complying-coppa-frequently-asked-questions-0
Option A is incorrect because HIPAA is related to protecting the privacy of healthcare data in the U.S
Option B is incorrect because it is a Payment Card Industry Data Security Standard to protect credit card information collected for business
Option C is incorrect because GDPR(General Data Protection Regulation) is regulatory compliance in Europe which is used to protect any personally identifiable information collected for business purpose within the Europe region

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

You have been hired as a Cloud Consultant for a company that is planning the migration of their entire Application and data from AWS cloud to Google Cloud Platform. During the initial phase of migration, there is a requirement to migrate data from AWS S3 buckets to GCS buckets. One of the key requirements is that any new data which gets added to S3 bucket should be copied to GCS bucket on a daily basis until the migration is completed. How will you accomplish this task?
A. Use Transfer Appliance
B. Create a Linux Compute VM on GCP and schedule a cron job which will copy data on a daily basis with proper authentication
C. Use gsutil cp cmd and run on a daily basis
D. Use GCP Storage Transfer Service

A

Answer: D
For multicloud diagram, you can use Cloud Interconnect by resource Partner Interconnect with solutions such as megaport https://www.megaport.com/services/google-cloud-partner-interconnect/ or Equinix ECX https://cloud.google.com/architecture/connection-google-cloud-vpcs-to-aws-equinix-network-edge
Option A is incorrect since HA VPN cannot provide 6 Gbps bandwidth. VPN tunnel can support bandwidth up to 3gbps.
Option B also can be used to connect two networks, but we want a managed service.
Option C is incorrect because it is used to connect VPCs within Google Cloud.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

You are working for a Company which is planning to develop a new application Which will be deployed in the Frankfurt region in Europe. The company offers an online vehicle insurance service that collects user data like name, address and vehicle-related details. Which regulation must your company comply with?
A. SOX
B. HIPAA
C. COPPA
D. GDPR

A

Answer: D
GCP Storage Transfer Service offers Quick transfer of data from online sources like AWS S3 and Azure Blob Storage to Cloud Storage in one simple process. You can also create a schedule in transfer service to sync data on a daily basis https://cloud.google.com/storage-transfer/docs/create-manage-transfer-console#amazon-s3
Option A is incorrect because it is used to Transfer data from on-premise to Google cloud
Option B is incorrect because managed GCP service will do most of the work for you.
Option C is incorrect because you can use gsutil cmd but as per GCP, Transfer service is the best option that will do all the work in a single process.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

For this question, refer to the MountKirk Games case study: https://cloud.google.com/certification/guides/cloud-architect/casestudy-mountkirkgames-rev2 As per the Technical requirements of MountKirk Games which Compute Option is best suitable for them?
A. A Single Compute instance with sustained discounts and instance property as Preemptible
B. A Single Compute instance with sustained discounts and instance property as non-Preemptible
C. A Managed Instance group with sustained discounts and instance property as Preemptible
D. GKE

A

Answer: D
GDPR(General Data Protection Regulation) is regulatory compliance in Europe which is used to protect any personally identifiable information collected for business purpose within the Europe region
Option A is incorrect because SOX compliance is used for financial auditing purpose
Option B is incorrect because HIPAA is related to protecting the privacy of healthcare data in U.S
Option C is incorrect because COPPA is related to protecting the privacy of children below 13 age in the U.S

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

Your team has developed an application that will be deployed on the Google Kubernetes Engine. There a requirement to persist the application data on the Kubernetes pods. How will you persist the data beyond the lifetime of the pods?
A. Ingress
B. Deployments
C. ReplicaSets
D. PersistentVolumes

A

Answer: D
Kubernetes Engine
Option A is incorrect because MountKirk wants a scalable environment so using Single compute engine instance will not fulfill the requirement
Option B is incorrect because MountKirk wants a scalable environment so using Single compute engine instance will not fulfill the requirement
Option C is incorrect because preemptible VM is not recommended for Production workload

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

You are working as a Solutions Architect for an enterprise. Your company recently developed a Web-App which will be deployed on App Engine. Following the IAM best practices which roles you will grant to the members where Team Lead is responsible for auditing App Engine code in production only requires read-only access to deployed source code and where developers can release code into production?
A. roles/appengine.appAdmin, roles/appengine.appViewer
B. roles/appengine.appAdmin, roles/appengine.codeViewer
C. roles/appengine.serviceAdmin, roles/appengine.deployer
D. roles/appengine.codeViewer, roles/appengine.deployer

A

Answer: D
PersistentVolumes(PV) is cluster-wide storage which is used to store data. Persistent Volume has a lifecycle independent of any pod that uses the persistent Volume. When we create a persistent Volume in GKE a compute engine persistent disk is created https://kubernetes.io/docs/concepts/storage/persistent-volumes/
Option A is incorrect because it used to expose a Kubernetes service to the public internet
Option B is incorrect because deployment is a Kubernetes object which is used to run multiple replicas of your pod and will automatically replace any failed or unresponsive pod
Option C incorrect because it used to manage the number pods running in a deployment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

You have been hired as a Cloud consultant for a company which is already using Google Cloud for their production and staging workload in separate GCP projects within an organization. Recently they came across a situation where an application running on a compute engine in a staging project requires a read access to a private GCS bucket which is in a production project. According to IAM best practices how will you grant access?
A. Create a service account in production with access keys, grant Storage object viewer role and configure the application in to use access keys
B. Create a service account in a staging project and attach the service account to the compute engine where the application is running. In production project grant staging projects service account Storage object viewer role in GCS bucket permission section.
C. Create a service account in a staging project and attach the service account to the compute engine where the application is running. In production project grant staging projects service account Storage object viewer role in project IAM section.
D. Add allUsers as a member in the permission section of GCS bucket in production project and grant Storage object viewer role.

A

Answer: D
The team lead is responsible for auditing App engine code in production so he will need only roles/appengine.codeViewer to perform his duties. This role grants read-only access to deployed source code and application configurations. Developers can be granted roles/appengine.deployer which grants them read-only access to all application configuration, settings and allow them to create a new version of the application https://cloud.google.com/appengine/docs/admin-api/access-control#roles
Option A is incorrect because it grants Read/Write/Modify permission to team lead and developers will not have permission to create a new version of an application
Option B is incorrect because it will grant Read/Write/Modify permission to team lead
Option C is incorrect because it will not allow the team lead to read the deployed source code

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

You have been hired as a Security consultant for a Financial company. The company holds sensitive data like customer account numbers, credit card information in the GCS bucket. The CTO wants additional security for mitigating exfiltration of data from a discontinued employee or attacker who has stolen identities. How will you mitigate this security risk by providing access to only authorized projects?
A. Cloud Armor
B. Threat Detection
C. VPC service controls
D. DLP

A

Answer B As per IAM best practices, you should add staging project’s service account in GCS bucket permission section and grant Storage object viewer role to provide cross account access https://cloud.google.com/dataprep/docs/concepts/gcs-buckets
Option A is a possible option but directly using access keys in the compute engine is not a good security practice.
Option C is incorrect because assigning the role in IAM section of the project will give access to all buckets in that project, not a particular bucket, this will grant access permissions
Option D is incorrect because adding allUsers will make the bucket public and anyone can access it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

One of your clients is using customer –managed encryption, which of the following statements are true when you are applying customer-managed encryption key to an object.[Select any 3]
A. the encryption key is used to encrypt the object’s data
B. the encryption key is used to encrypt the object’s CRC32C checksum
C. the encryption key is used to encrypt the object’s name
D. the encryption key is used to encrypt the object’s MD5 hash

A

Answer C VPC service controls allow you to lock down GCP resources. In VPC service control you can define which projects can call on your GCP APIs allowing you to whitelist the project which you want to grant access to. This can protect sensitive data from attackers or stolen identity The most common use cases for VPC service controls are Mitigate threats such as data exfiltration Isolate parts of the environment by trust level Secure access to multi-tenant services https://cloud.google.com/vpc-service-controls
Option A is incorrect because it is used to mitigate DDoS attack and provides WAF
Option B is incorrect because it is used to detect threats like Burt force attack from logs and reports to Security command center
Option D is incorrect because Cloud DLP is used to detect and de-identify any sensitive information like credit card number or any PII data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

You have a long-running job that one of your employees has permissions to start. You don’t want that job to be terminated when the employee who last started that job leaves the company. What would be the best way to address the concern in this scenario?
A. Create many IAMusers and give them the permission.
B. Create a service account. Grant the Service Account User permission to the employees who needs to start the job. Also, provide “Compute Instance Admin” permission to that service account.
C. Give full permissions to the Service Account and give permission to the employee to access this service account.
D. Use Google-managed service accounts in this scenario.

A

Answer:
Option A, B, D are the CORRECT choice because, When you apply a customer-managed encryption key to an object, the encryption key is used to encrypt the object, its CRC32C checksum, and its MD5 hash. The remaining metadata for the object, including the object’s name, is encrypted using standard server-side keys. This allows you to always read and update metadata, as well as list and delete objects, provided you have permission to do so. https://cloud.google.com/storage/docs/encryption/customer-managed-keys

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

A Global Media company is configuring a Global load balancer for non-http(s) traffic. They are looking for a service with SSL offloading and as a Cloud Architect what would be your load balancing choice?
A. HTTPS load balancing
B. SSL proxy Load balancing.
C. TCP proxy Load balancing for all non-http(s) traffic
D. Network TCP/UDP load balancing

A

Answer:
Option B is the CORRECT because, creating service accounts for each service with only the permissions required for that service is the best practice, even if the employee leaves the organization other employees can use the service account .
Option A is INCORRECT because Service Account is used to give permission to Application or VMs. A service account is a special type of Google account that belongs to your application or a virtual machine (VM), instead of to an individual end user. Your application assumes the identity of the service account to call Google APIs so that the users aren’t directly involved. With Admin access, the employees will be able to create Compute Engine instances which runs the service account, connect to them, and use the service account to start the job. So in nutshell,admin empowers to effectively run code as the service accounts used to run these instances, and indirectly gain access to all the resources for which the service accounts has access.
Option C is INCORRECT because Granting the service account only the minimum set of permissions required to achieve their goal is the best practice.
Option D is INCORRECT because Google Managed service accounts are created and owned by Google. These accounts represent different Google services and each account is automatically granted IAM roles to access your GCP project. This service account is designed specifically to run internal Google processes on your behalf and is not listed in the Service Accounts section of GCP Console. More reading at https://cloud.google.com/iam/docs/understanding-service-accounts

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

Which of the following are the best practices recommended by Google Cloud when dealing with service Accounts. Select 3 relevant options
A. Grant the service account full set of permissions
B. Do not delete service accounts that are in use by running instances on Google App Engine or Google Compute Engine
C. Grant serviceAccountUser role to all the users in the organization.
D. Use the display name of a service account to keep track of the service accounts. When you create a service account, populate its display name with the purpose of the service account.
E. Create service accounts for each service with only the permissions required for that service.

A

Answer:
Option B is the CORRECT choice because SSL proxy Loadbalancing supports SSL offloading and it is availability is Global and it handles non-http(s) traffic.
Option A is INCORRECT because the traffic is non-http(s).
Option C is INCORRECT because TCP proxy can handle non-http(s) traffic but it doesn’t come with SSL offloading feature.
Option D is INCORRECT because Network TCP/UDP load balancing is Regional and it doesn’t handle SSL offloading. Google Cloud SSL Proxy Load Balancing terminates user SSL (TLS) connections at the load balancing layer, then balances the connections across your instances using the SSL or TCP protocols. Cloud SSL proxy is intended for non-HTTP(S) traffic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

One of the large data Analysis company uses Big Query, Big Table, Data Proc and Cloud Storage services. They use a Hybrid Architecture involving on premise and Google Cloud, Cloud VPN is used to connect to Google Cloud Platform. One of the main challenges for the Organization is mitigating Data exfiltration risks stemming from stolen identities, IAM policy misconfigurations, malicious insiders and compromised virtual machines. What Google Cloud Service can they use to address the challenge?
A. Shared VPC
B. Cloud Armour
C. VPC Service Controls
D. Resource Manager

A

Answer:
Option B, D & E are the CORRECT choices.
Option A is INCORRECT because always grant the service account only the minimum set of permissions required to achieve their goal.
Option C is INCORRECT because always restrict who can act as service accounts. Users who are Service Account Users for a service account can indirectly access all the resources the service account has access to. Therefore, be cautious when granting the serviceAccountUser role to a user.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
71
Q

A power generation company is looking to use the Google cloud platform to monitor a power station. They have installed several IoT sensors in the power station like temperature sensors, smoke detectors, motion detectors, etc. Sensor data will be continuously streamed to the cloud. Those data need to be handled by different components for real-time monitoring and alerts, analysis, and performance improvement. What Google Cloud Architecture would serve their purpose?
A. Cloud IoT Core receives data from IoT devices and redirects the requests to aCloud Pub/Sub Topic. AfterPub/Sub, data is retrieved by a streaming job running in Cloud Dataflow that transforms the data and sends it to BigQuery for analysis.
B. Send IoT devices data to Cloud Storage, load data from cloud storage to Big Query.
C. Cloud IoT core receives data from IoT sensors, then sends the data to cloud storage, transform the data using Cloud Dataflow and send the data to BigQuery for Analysis.
D. Cloud IoT core receives the data from IoT devices, Cloud IoT core transforms and redirects the request to Pub/Sub, use data proc to transform the data and send it to BigQuery for Analysis.

A

Answer:
Option C is CORRECT because , VPC Service Controls create a security perimeter around data stored in API-based GCP services such as Google Cloud Storage, BigQuery and Bigtable. This helps mitigate data exfiltration risks stemming from stolen identities, IAM policy misconfigurations, malicious insiders and compromised virtual machines.
Option A is INCORRECT because , Shared VPC allows an organization to connect resources from multiple projects to a common VPC network, so that they can communicate with each other securely and efficiently using internal IPs from that network. When you use Shared VPC, you designate a project as a host project and attach one or more other service projects to it. The VPC networks in the host project are called Shared VPC networks. Eligible resources from service projects can use subnets in the Shared VPC network .Here the challenge is to mitigate Data exfiltration and VPC Service Controls is the right choice.
Option B is INCORRECT because, Cloud Armor is used for delivering defense at scale against infrastructure and application Distributed Denial of Service (DDoS) attacks using Google’s global infrastructure and security systems.
Option D is INCORRECT because , . Resource Manager enables you to programmatically manage these resource containers. Google Cloud Platform provides Resource containers such as Organizations, Folders, and Projects, that allow you to group and hierarchically organize other Cloud Platform resources. This hierarchical organization lets you easily manage common aspects of your resources such as access control and configuration settings. Security benefits of VPC Service Controls VPC Service Controls helps mitigate the following security risks without sacrificing the performance advantages of direct private access to GCP resources: Access from unauthorized networks using stolen credentials: By allowing private access only from authorized VPC networks, VPC Service Controls protects against theft of OAuth credentials or service account credentials. Data exfiltration by malicious insiders or compromised code: VPC Service Controls complements network egress controls by preventing clients within those networks from accessing the resources of Google-managed services outside the perimeter. VPC Service Controls also prevents reading data from or copying data to a resource outside the perimeter using service operations such as copying to a public Cloud Storage bucket using the gsutil cp command or to a permanent external BigQuery table using the bq mk command. The restricted VIPs feature can be used to prevent access from a trusted network to storage services that are not integrated with VPC Service Controls. Public exposure of private data caused by misconfigured Cloud IAM policies: VPC Service Controls provides an additional layer of security by denying access from unauthorized networks, even if the data is exposed by misconfigured Cloud IAM policies. By assigning the Access Context Manager Policy Admin role for Cloud IAM, VPC Service Controls can be configured by a user who is not the Cloud IAM policy administrator. VPC Service Controls is configured for your GCP organization to create a broad, uniform policy that applies consistently to all protected resources within the perimeter. You retain the flexibility to process, transform, and copy data within the perimeter. The security controls automatically apply to all new resources created within a perimeter. Read more about VPC Service Control here : https://cloud.google.com/vpc-service-controls/docs/overview A service perimeter creates a security boundary around GCP resources. You can configure a service perimeter to control communications from virtual machines (VMs) to a GCP service (API), and between GCP services. A service perimeter allows free communication within the perimeter but, by default, blocks all communication across the perimeter. For example: A VM within a Virtual Private Cloud (VPC) network that is part of a service perimeter can read from or write to a Cloud Storage bucket in the same perimeter. However, any attempt to access the bucket from VPC networks that are not inside the perimeter is denied. A copy operation between two Cloud Storage buckets will succeed if both buckets are in the same service perimeter, but will fail if one of the buckets is outside the perimeter. A VM within a VPC network that is part of a service perimeter can privately access any Cloud Storage buckets in the same perimeter. However, the VM will be denied access to Cloud Storage buckets that are outside the perimeter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
72
Q

Your company just finished a rapid lift and shift to Google Compute Engine for your compute needs. You have another 9 months to design and deploy a more cloud-native solution. The business team is looking for services with lesser responsibility and easy manageability. Please select the order of services with lesser responsibility to more responsibility
A. GKE >Google App Engine Standard Environment >Cloud Functions >Compute Engine with containers >Compute Engine
B. Cloud Functions >Google App Engine Standard Environment>GKE >Compute Engine with containers >Compute Engine
C. Cloud Functions >GKE >Google App Engine Standard Environment >Compute Engine >Compute Engine with containers
D. Google App Engine Standard Environment >Cloud Functions>Compute Engine with containers>GKE>Compute Engine

A

Answer: A
Option A is CORRECT because Cloud IoT Core can accept data from IoT devices and Cloud Pub/Sub acts as a connector service and sends the data to Cloud Data Flow for transformation. Data Flow transforms the data and sends it to Big Query for analysis.
Option B is INCORRECT because Cloud Storage isn’t the right choice for streaming data, using Cloud Pub/Sub is the best choice.
Option C is INCORRECT because Cloud IoT Core can stream the data directly to Cloud Pub/Sub .(use Cloud Storage for Batch Upload)
Option D is INCORRECT because Dataproc is a fully managed cloud service for running Apache Spark and Apache Hadoop clusters. Sources: https://cloud.google.com/community/tutorials/cloud-iot-rtdp

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
73
Q

One of the customers want to redact the sensitive data like credit card numbers , social security numbers that are generated by the application logs .Please select the suitable service that fulfils the necessary requirement .
A. Cloud Data Loss Prevention
B. Cloud Secure
C. VPC Service control
D. Cloud Armour

A

Answer: B is the CORRECT choice, Cloud Functions is the least no-Ops, then App Engine, then followed by GKE and then Compute Engine with containers and at last Compute Engine.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
74
Q

Your organization is developing an event-driven application in which cloud functions will access Cloud SQL for managing data. As per the security best practices, you want to store the Cloud SQL credentials securely. Where will you store the Cloud SQL credentials?
A. In the Cloud function code
B. In the Cloud function environment variable
C. In Cloud Secret Manager
D. In Cloud KMS

A

Answer:A
Option A is the Correct choice because , Cloud DLP helps you better understand and manage sensitive data. It provides fast, scalable classification and redaction for sensitive data elements like credit card numbers, names, social security numbers, US and selected international identifier numbers, phone numbers, and GCP credentials .
Option B is Incorrect because Cloud Secure Service doesn’t exist in GCP.
Option C is incorrect because , VPC Service Controls allow users to define a security perimeter around Google Cloud Platform resources such as Cloud Storage buckets, Bigtable instances, and BigQuery datasets to constrain data within a VPC and help mitigate data exfiltration risk but it doesn’t help in data redaction .
Option D is Incorrect because ,Cloud Armour Google Cloud Armor delivers defence at scale against infrastructure and application Distributed Denial of Service (DDoS) attacks using Google’s global infrastructure and security systems but it doesn’t help in data redaction . Read more about it here : https://cloud.google.com/dlp/

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
75
Q

You have been hired as a DevSecOps engineer by a finance company. They want to upload files from an on-premise server to Google Cloud Storage. But as per there security policy, the files must be encrypted using customer-supplied encryption keys on Google Cloud storage. How will you fulfill this requirement?
A. Use –encryption_key flag with gsutil command to supply encryption key while uploading files
B. Supply the encryption key in Cloud KMS and use that key for encryption
C. Add the encryption_key option in the boto configuration file and use gsutil command to upload files
D. Configure the encryption key in gcloud configuration and use gsutil to upload files

A

C)
Option is correct You should store Cloud SQL credentials in Cloud Secret Manager where you can rotate, create versions and can manage access to credentials https://cloud.google.com/secret-manager/docs/creating-and-accessing-secrets
Option A is incorrect because storing the credentials in the code itself will make it accessible to anyone having access to cloud functions and it will also become difficult to rotate credentials
Option B is incorrect because storing the credentials in the environment variable will make it accessible to anyone having access to cloud functions
Option D is incorrect because Cloud KMS is used to managing encryption and decryption

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
76
Q

To set up a virtual private network between your office network and Google Cloud Platform and have the routes automatically updated when the network topology changes, what is the minimal number of each type of component you need to implement?
A. 2 Cloud VPN Gateways and 1 Peer Gateway
B. 1 Cloud VPN Gateway, 1 Peer Gateway, and 1 Cloud Router
C. 2 Peer Gateways and 1 Cloud Router
D. 2 Cloud VPN Gateways and 1 Cloud Router

A

C) Option is correct To use customer-supplied encryption keys with Google Cloud Storage while uploading files, you must add the encryption_key option in [GSUtil] section of the boto configuration file. Boto configuration file is the file where you can configure all configurations related to gsutil command line https://cloud.google.com/storage/docs/gsutil/addlhelp/UsingEncryptionKeys
Option A is incorrect because you need to add encryption_key option in GSUtil section of boto configuration file there is no such flag while using gsutil commands
Option B is incorrect because as per security policy they don’t want to store Keys in Google Cloud. So CMEK is not an option.
Option D is incorrect because encryption_key must be added in boto file, not the gcloud configuration

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
77
Q

A Digital Media company has recently moved its infrastructure from On-premise to Google Cloud. They have deployed several instances under a Global HTTPS load balancer. A few days ago an Application and Infrastructure were subjected to DDOS attacks. Hence they are looking for a service that would provide a defence mechanism against the DDOS attacks. Please select the relevant service from below.
A. Cloud Armor
B. Cloud-Identity Aware Proxy
C. GCP Firewalls
D. IAM policies

A

Concert answer B The question describes a topology for Dynamic routing The minimal number of each type of component you need to implement Dynamic routing: 1 Cloud VPN Gateway (Show as VPN in GCP network on left), 1 Peer Gateway (Show as VPN Gateway with BGP in peer network on right), and 1 Cloud Router, displayed in the diagram

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
78
Q

Your infrastructure includes two 100-TB enterprise file servers. You need to perform a one-way, one-time migration of this data to the Google Cloud securely. Only users in Germany will access this data. You want to create the most cost-effective solution. What should you do?
A. Use Transfer Appliance to transfer the offsite backup files to a Cloud Storage - Regionbucket as a final destination.
B. Use Transfer Appliance to transfer the offsite backup files to a Cloud Storage - Multi-Regionbucket as a final destination.
C. Use Storage Transfer Service to transfer the offsite backup files to a Cloud Storage - Region bucket as a final destination.
D. Use Storage Transfer Service to transfer the offsite backup files to a Cloud Storage - Multi-Region bucket as a final destination.

A

Correct Answer - A
Option A is CORRECT because Cloud Armor delivers defence at scale against infrastructure and application Distributed Denial of Service (DDoS) attacks using Google’s global infrastructure and security systems.
Option B is INCORRECT because, Cloud-Identity Aware Proxy lets you establish a central authorization layer for applications accessed by HTTPS, so you can use an application-level access control model instead of relying on network-level firewalls.
Option C is INCORRECT because GCP firewalls rules don’t apply for HTTP(S) Load Balancers, while Cloud Armor is delivered at the edge of Google’s network, helping to block attacks close to their source.
Option D is INCORRECT IAM policies don’t help in mitigating DDOS attacks. Read more about Cloud Armor: https://cloud.google.com/blog/products/gcp/getting-to-know-cloud-armor-defense-at-scale-for-internet-facing-services

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
79
Q

You are working as a Google Cloud Architect for a large enterprise. They are using the GKE cluster for their production workload. There is a requirement to expose an existing deployment to the public internet using a service type of load balancer. Which command will you use to create a service type of load balancer?
A. kubectl expose deployment demo –port=80 –target-port=80 --name=example-service –type=LoadBalancer
B. kubectl expose deployment demo –type=LoadBalancer –expose 80
C. kubectl expose service demo –port=443 –target-port=80 –name=new-application
D. kubectl expose deployment demo –type=NodePort –name=example-service

A

Correct Answer - A
Option A is correct because you are performing a one-time (rather than an ongoing series) data transfer from on-premises to Google Cloud Platform for users in a single region (Germany). Using a Region storage bucket will reduce cost and also conform to regulatory requirements Options B, C, and D are incorrect because you should not use a Multi-Region storage bucket for users in a single region (B, D). Also, Storage Transfer Service does not work for data stored on-premises file servers (C, D). Reference GCS Region storage for single location access: https://cloud.google.com/storage/docs/storage-classes Google Cloud transfer appliance: https://cloud.google.com/transfer-appliance/

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
80
Q

Your company plans to migrate a multi-petabyte data set to the cloud. The data set must be available 24hrs a day. Your business analysts have experience only with using an SQL interface. How should you store the data to optimize it for ease of analysis?
A. Load data into Google BigQuery.
B. Insert data into Google Cloud SQL.
C. Put flat files into Google Cloud Storage.
D. Stream data into Google Cloud Datastore.

A

Correct Answer - A
Option A is correct With the help of this command, we can create a service type of LoadBalancer and expose the port on which our application is hosted. https://kubernetes.io/docs/tasks/access-application-cluster/create-external-load-balancer/
Option B is incorrect because the flag –name is not optional and is missing in
Option B.
Option C is incorrect because it will create a service type of ClusterIP.
Option D is incorrect because we need a service type of load balancer that we can expose to the public internet.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
81
Q

One of your clients are storing highly sensitive data on Google Cloud Storage, they strictly adhere to their compliance, hence they do not want their keys to be stored in a cloud, please suggest them the right choice of encryption.
A. Google recommends the usage of Cloud External Key Manager (Cloud EKM)
B. All objects on Google Storage are encrypted by default hence additional encryption isn’t required
C. Give your Cloud Storage service account access to an encryption key, that service account encrypts
D. Google recommends the usage of cloud KMS for storing CMEK.

A

Correct Answer - A
Option A is correct. BigQuery is the only of these Google products that support an SQL interface and can handle petabyte data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
82
Q

For this question, refer to the TerramEarth case study. TerramEarth’s 20 million vehicles are scattered around the world. Based on the vehicle’s location its telemetry data is stored in a Google Cloud Storage (GCS) regional bucket (US. Europe, or Asia). The CTO has asked you to run a report on the raw telemetry data to determine why vehicles are breaking down after 100 K miles. You want to run this job on all the data. What is the most cost-effective way to run this job?
A. Launch a cluster in each region to preprocess and compress the raw data, then move the data into a regional bucket and use a Cloud Dataproc cluster to finish the job.
B. Move all the data into 1 region, then launch a Google Cloud Dataproc cluster to run the job.
C. Launch a cluster in each region to preprocess and compress the raw data, then move the data into a multi-region bucket and use a Dataproc cluster to finish the job.
D. Move all the data into 1 zone, then launch a Cloud Dataproc cluster to run the job.

A

Correct Answer - A
Option A is the correct choice because the client doesn’t want to store the encryption keys on Google Cloud. With Cloud EKM, you can use keys that you manage within a supported external key management partner to protect data within Google Cloud.
Option B is incorrect because, even though All objects on Google Storage are encrypted by default, the client is storing sensitive data and hence default encryption isn’t the best option. https://cloud.google.com/security/encryption-at-rest/
Option C is incorrect because giving your Cloud Storage service account access to an encryption key, that the service account encrypts comes under Customer-Managed Encryption Keys, these keys are stored in Google cloud, hence not the correct choice here.
Option D is incorrect because, in a customer-managed encryption key, your encryption keys are stored within Cloud KMS. The client doesn’t want to store keys on the Cloud. Reference: https://cloud.google.com/kms/docs/ekm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
83
Q

You have an application server running on Compute Engine in the europe-west1-d zone. You need to ensure high availability and replicate the server to the europe-west2-c zone using the fewest steps possible. What should you do to achieve the requirement?
A. Create a snapshot from the disk. Create a disk from the snapshot in the europe-west2-c zone. Create a new VM with that disk.
B. Create a snapshot from the disk. Create a disk from the snapshot in the europe-west1-d zone and then move the disk to europe-west2-c. Create a new VM with that disk.
C. Use “gcloud” to copy the disk to the europe-west2-c zone. Create a new VM with that disk.
D. Use “gcloud compute instances move” with parameter “–destination-zone europe-west2-c” to move the instance to the new zone.

A

Correct Answer - A A (Correct answer) - Launch a cluster in each region to preprocess and compress the raw data, then move the data into a regional bucket and use a Cloud Dataproc cluster to finish the job. Since the raw data are saved based on the vehicle’s location all over the world, most likely they’ll scatter in many different regions, and eventually they need to move to a centralized location for final processing. Preprocessing raw data and compressing them from each location to reduce the size so to save the between-region data egress cost. Dataproc is a region-specific resource and since you want to run this job on all data and you or your group probably are the only consumers for the data, moving the data into a regional bucket same or closest to the DataProc cluster’s region for final analysis is most cost-effective. Use a regional location to help optimize latency, availability, and network bandwidth for data consumers grouped in the same region. Use a multi-regional location when you want to serve content to data consumers that are outside of the Google network and distributed across large geographic areas. Store frequently accessed data, or data that needs to be geo-redundant as Multi-Regional Storage.
B - Move all the data into 1 region, then launch a Google Cloud Dataproc cluster to run the job. Since the raw data are save based on the vehicles’ location all over the world, moving them to a centralized region without preprocessing and compressing would incur additional between-region data egress cost
C - Launch a cluster in each region to preprocess and compress the raw data, then move the data into a multi-region bucket and use a Dataproc cluster to finish the job. Dataproc is Region-specific resource and since you want to run this job on all data and data consumption occurs in a centralized location, then moving the data into a multi-region bucket for Dataproc cluster jobs is not most cost-effective. Use a multi-regional location when you want to serve content to data consumers that are outside of the Google network and distributed across large geographic areas. · Store frequently accessed data or data that needs to be geo-redundant as Multi-Regional Storage.
D - Move all the data into 1 zone, then launch a Cloud Dataproc cluster to run the job. GCS is either Regional or Multi-Regional not Zonal Resources

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
84
Q

You need to reduce the number of unplanned rollbacks of erroneous production deployments in your company’s web hosting platform. Improvement to the QA and Test processes accomplished an 80% reduction. Which additional two approaches can you take to further reduce the rollbacks? (Choose two)
A. Introduce a blue-green deployment model.
B. Fragment the monolithic platform into microservices.
C. Remove the QA environment. Start executing canary releases.
D. Remove the platform’s dependency on relational database systems.
E. Replace the platform’s relational database systems with a NoSQL database.

A

Correct Answer - A A is correct because this makes sure the VM gets replicated in the new zone. B is incorrect because this takes more steps than A. C is incorrect because this will generate an error because gcloud cannot copy disks. D is incorrect because the original VM will be moved, not replicated. References: https://cloud.google.com/compute/docs/instances/create-start-instance#createsnapshot

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
85
Q

You are now working for an international company that has many Kubernetes projects on various Cloud platforms. These projects involve mainly microservices web applications and are executed either in GCP or other cloud providers. They have many inter-relationships and there is the involvement of many teams related to development, staging, and production environments. Your new task is to find the best way to organize these systems. You need a solution for gaining control on application organization and networking: monitor functionalities, performances, and security in a complex environment. Which of the following services may help you?
A. Traffic Director
B. Istio on GKE
C. Apigee
D. App Engine Flexible Edition

A

Correct Answer - A and B A (Correct Answer) - The blue-green model allows for extensive testing of the application in the green environment before sending traffic to it. Typically, the two environments are identical otherwise which gives the highest level of testing assurance. B (Correct Answer) - Microservices allows for smaller, more incremental rollouts of updates (each microservice can be updated individually) which will reduce the likelihood of an error in each rollout. C is incorrect - Would remove a well proven step from the general release strategy, a canary release platform is not a replacement for QA, it should be additive. D is incorrect - Doesn’t really help the rollout strategy, there is no inherent property of a relational database that makes it more subject to failed releases than any other type of data storage. E is incorrect - Doesn’t really help either since NoSQL databases do not offer anything over relational databases that would help with release quality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
86
Q

Your company plans to host a large donation website on the Google Cloud Platform. You anticipate a large and undetermined amount of traffic that will create many databases writes. Which managed service hosted on GCP would you suggest to ensure no drop for any write traffic to a database?
A. Cloud SQL with Bigger (More CPU, Memory, and Disk Size) machine type with throughput capacity matching to the anticipated peak write throughput.
B. Cloud Pub/Sub for capturing the writes and draining the queue to write to the database.
C. Memcached to store the writes until the writes are committed to the database.
D. Install your MySQL database on Compute instance and enable autoscaling.

A

Correct Answer - A What you need is Service Management with capabilities of real-time monitoring, security, and telemetry data collection in a multi-cloud microservices environment. They are called Service Mesh. The most popular product in this category is ISTIO, which collects traffic flows and telemetry data between microservices, enforcing security, with the help of proxies that operate without changes to application code. Traffic Director can help in a global service mesh because it is a fully managed Service Management control plane. With Traffic Director, you can manage on-premise and multi-cloud destinations, too. B is incorrect because Istio on Google Kubernetes Engine is a tool for GKE that offers automated installation and management of Istio Service Mesh. So, only inside GCP. C is incorrect because Apigee is a powerful tool for API Management suitable also for on-premise and multi-cloud environments. But API Management is for managing application APIs and Service Mesh is for managing Service to Service communication, security, Service Levels, and control. Similar services with different scopes. D is incorrect because App Engine Flexible Edition is a PaaS for microservices applications within Google Cloud. For any further detail: https://cloud.google.com/traffic-director/docs/overview Choosing between service management and API management

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
87
Q

Your customer needs a dedicated System with MongoDB and 2 replicas. He also wants maximum availability and protection against failures and interruptions for maintenance/ updates to the instances. The Database operates only in one US region and is actively queried and updated 24/7. So, you cannot select a comfortable maintenance Window. What do you advise?
A. Use an internal load balancing Service with a Managed Instance Group and Regional persistent disks
B. Use a 3rd party MongoDB Managed Service like MongoDBAtlas
C. Implement Live Migration and use persistent regional SSDs
D. Use internal TCP/UDP Load Balancing with local SSD disks

A

Correct Answer - B
A - you anticipate a “large and undetermined amount of traffic”, so regardless of any provisioned IOPS there is always a risk it will not be enough and potentially high none necessary cost B (Correct answer) - Cloud Pub/Sub for capturing the writes and draining the queue to write to the database. Cloud Pub/Sub brings the scalability, flexibility, and reliability of enterprise message-oriented middleware to the cloud. By providing many-to-many, asynchronous messaging that decouples senders and receivers, it allows for secure and highly available communication between independently written applications. Cloud Pub/Sub delivers low-latency, durable messaging that helps developers quickly integrate systems hosted on the Google Cloud Platform and externally.
C - Memcached is for reading not for write
D - Install your MySQL database on Compute instance and enable autoscaling. If you roll out your own MySql instance, then you don’t have the advantage from manage Google Cloud SQL. Furthermore, it’ll be complicated and costly to implement a horizontal autoscaling feature even if you can try some sharding and master/slave. So, Answer B is the clear winner.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
88
Q

You are designing a relational data repository on Google Cloud to grow as needed. The data will be transactional consistent and added from any location in the world. You want to monitor and adjust node count for input traffic, which can spike unpredictably. What should you do?
A. Use Cloud Spanner for storage. Monitor storage usage and increase node count if more than 70% utilized.
B. Use Cloud Spanner for storage. Monitor CPU utilization and increase node count if more than 70% utilized for your time span.
C. Use Cloud Bigtable for storage. Monitor data stored and increase node count if more than 70% utilized.
D. Use Cloud Bigtable for storage. Monitor CPU utilization and increase node count if more than 70% utilized for your time span.

A

Correct Answer - B
Option A is incorrect. A load balancer with MIG will provide scalability with an inbuilt feature of Live Migration if needed. But it is not as optimized response as compared to a managed service
Option B is correct. The requirement is to have a dedicated System. MongoDB Atlas provides the same. MongoDB Atlas provides customers a fully managed service on Google’s globally scalable and reliable infrastructure. Atlas allows you to manage your databases easily with just a few clicks in the UI or an API call, is easy to migrate to, and offers advanced features such as global clusters for low-latency read and write access anywhere in the world.
Option C is incorrect. Live migration is not an option but an inbuilt feature provided by Google.
Option D is incorrect. A DB Instance, even if NoSQL, cannot scale in a simple way. In case of failover, it is likely to have inconsistencies and loss of services. In addition, local SSD disks are really fast but they persist only until the instance is stopped or deleted. Definitely not according to the requirements. For any further detail, please check the following URLs: https://cloud.google.com/mongodb https://www.mongodb.com/cloud/atlas/

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
89
Q

How are subnetworks (VPC Networks) different than the legacy networks?
A. They’re the same, only the branding is different.
B. Each subnetwork controls the IP address range used for instances that are allocated to that subnetwork.
C. With subnetworks IP address allocation occurs at the global network level.
D. Legacy networks are the preferred way to create networks.

A

Correct Answer - B
Option B is correct because of the requirement to globally scalable transactions—use Cloud Spanner. CPU utilization is the recommended metric for scaling, per Google best practices, see linked below. A is incorrect because you should not use storage utilization as a scaling metric. C, D are incorrect because you should not use Cloud Bigtable for this scenario: The data will be transactional consistent and added from any location in the world. References: Cloud Spanner Monitoring Using Operations Suite (formerly Stackdriver) https://cloud.google.com/spanner/docs/monitoring Best Practices: https://cloud.google.com/spanner/docs/best-practice-list

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
90
Q

You have been hired as a DevSecOps Engineer by a large enterprise. They recently migrated their on-premise servers to GCP. There is a requirement that the instances running in the VPC should only send traffic to Active Directory Servers in the same VPC and all other outgoing traffic should be blocked. How will you create the firewall rules for this scenario?
A. Create firewall rules which deny all egress traffic and assign a priority of 100. Also, create a firewall rule which allows egress traffic to Active Directory Servers and assign a priority of 1000 and apply to all instances.
B. Create firewall rules which deny all egress traffic and assign a priority of 1000. Also, create a firewall rule which allows egress traffic to Active Directory Servers and assign a priority of 100 and apply both rules to all instances.
C. Create firewall rules which deny all ingress traffic and assign a priority of 100. Also, create a firewall rule which allows egress traffic to Active Directory Servers and assign a priority of 1000 and apply to all instances.
D. Create firewall rules which deny all ingress traffic and assign a priority of 1000. Also, create a firewall rule which allows egress traffic to Active Directory Servers and assign a priority of 100 and apply both rules to all instances.

A

Correct Answer - B Google Cloud Platform (GCP) legacy networking vs. VPC subnet: Legacy networking Legacy networks have a single RFC 1918 range, which you specify when you create the network. The network is global in scope and spans all cloud regions. In a legacy network, instance IP addresses are not grouped by region or zone. One IP address can appear in one region, and the following IP address can be in a different region. Any given range of IPs can be spread across all regions, and the IP addresses of instances created within a region are not necessarily contiguous. It is not possible to create regional subnets with a legacy network. Legacy networking Example: Subnets and IP ranges Each VPC network consists of one or more useful IP range partitions called subnetworks or subnets. Each subnet is associated with a region. Networks can contain one or more subnets in any given region. Subnets are regional resources. Each subnet must have a primary address range, which is a valid RFC 1918 CIDR block. Subnets in the same network must use unique IP ranges. Subnets in different networks, even in the same project, can re-use the same IP address ranges. VPC network example: subnet3 is defined as 10.2.0.0/16, in the us-east1 region. One VM instance in the us-east1-a zone and a second instance in the us-east1-b zone, each receiving an IP addresses from its available range. Note: Legacy networks are not recommended. Many newer GCP features are not supported in legacy networks. It is still possible to create legacy networks through the gcloud command-line tool and the REST API. It is not possible to create legacy networks using the Google Cloud Platform Console. Reference resources Virtual Private Cloud (VPC) Network Overview https://cloud.google.com/vpc/docs/vpc Google Cloud Platform (GCP) legacy networking vs. VPC subnet https://cloud.google.com/vpc/docs/legacy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
91
Q

You are working as a DevOps engineer for a large enterprise. Recently there was an update deployed for an application running on the Compute engine server which caused a memory leak issue. Due to this, the compute engine memory was full leading to an outage. How will you avoid this issue in the future by setting up a proper alerting solution for memory metric, so the SRE teams get notified in time?
A. Install Cloud Logging agent in VM to monitor memory usage and setup alerting policies to notify the SRE team
B. Install monitoring agent on VM for memory usage monitoring and set up alerting policies in Cloud Operations to notify the SRE team using Cloud Operations.
C. Install Cloud monitoring agent in VM to monitor memory usage and setup uptime checks policies to notify the SRE team
D. Memory metrics are by default available for a VM, just setup alerting policies to notify the SRE team

A

Correct Answer - B Since we need to allow egress traffic to Active Directory servers only, we will create an egress rule which has a destination IP range of Active Directory servers and Assign a Low priority number because the lower the number, the higher the priority. The second rule to deny all egress traffic with Higher Priority Number i.e. 1000 https://cloud.google.com/vpc/docs/using-firewalls
Option A is incorrect because creating a deny rule for all egress traffic with priority 100 will block all traffic including Active Directory. The lower number takes the first precedence Options C & D are incorrect because there is a requirement to all block outgoing traffic except traffic to the active directory server, so configuring ingress rules will not work.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
92
Q

You are working as a DevOps engineer for a startup company. Recently they deployed a Python based application on the Google Kubernetes engine which is running slow and using more infrastructure resources than expected as per monitoring alerts. Which GCP service can you use to troubleshoot such an issue?
A. Cloud Monitoring
B. Cloud Trace
C. Cloud Profiler
D. Cloud Logging

A

Correct Answer - B We can monitor the GCP services using the Operations suite from Google. Though it doesn’t provide memory metrics out of the box. We have to install the agent on VM for additional metrics. After configuring the Cloud monitoring you can set up alerting policies in Cloud monitoring to notify the SRE team.
Option A is incorrect because Cloud logging is used for logging application logs or any other logs to Cloud logging.
Option C is incorrect because the uptime check is used to check the system availability.
Option D is incorrect because by default memory metrics are not available on Cloud monitoring.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
93
Q

You have to deploy an update to your scalable app, operating with managed instance groups but you cannot undergo any service disruption during the migration. You have already tested the new configuration and you need to deploy it in the fastest and safest way. Which is the best solution?
A. Use a new Template and everything will be automatic.
B. Use a new Template, then start new instances and stop the old ones.
C. Use a new Template and ask for a Rolling update.
D. Use a new Template and ask for a Canary update.

A

Correct Answer - C
Option A is incorrect because it is used to monitor resource utilization or any custom metric.
Option B is incorrect because Cloud trace is used to detect the latency issues in your application.
Option C is correct because Cloud profiler is a Google cloud service that helps you to analyze the CPU and memory usage of your functions in the application https://codelabs.developers.google.com/codelabs/cloud-stackdriver-profiler/#0
Option D is incorrect because Cloud logging is a fully managed service which allows you to store, search and analyze logs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
94
Q

Rules must be set to allow data traffic to database servers only from application servers, in 3 different projects: A, B, and C. The resources of the 3 projects must be isolated from each other. You want to organize operations in order to create simple and intuitive standards to use, which can be repeated for other projects. In your organization, it is not necessary to provide different security for various projects. Which of the following strategies will you choose?
A. Create 2 Firewall Rules, one in ingress and one in egress, between each Database Server and App Server using the ephemeral external IP address
B. Create 1 Firewall Rule, in ingress, between each Database Server and App Server using private IP addresses
C. Configure your Servers with appropriate Network Tags (AppVM and DBVM, for example) and create 1 Firewall Rule, in ingress, between each Database Server and App Server using these Tags
D. Configure your Servers with appropriate Network Tags (AppVM and DBVM, for example) and create 2 Firewall Rules, in ingress and egress, between each Database Server and App Server using these Tags
E. Create and assign appropriate Service Accounts and rights to the VMs and create a Firewall Rule between each Database Server and App Server using source-service-accounts and target-service-accounts

A

Correct Answer - C A is incorrect. Instance Template are immutable so you have to create a new Instance Template and update the Managed Group Definition B is incorrect. It is not advisable to do such a manual operation. It’s cumbersome and prone to errors. C is correct. With managed instance group updater, you may roll out an update automatically based on your specifications: maxSurge is the number of Instances beyond the targetSize of the group maxUnavailable set the number of instances unavailable at any time during the update Minimal action: if the updater has to REPLACE or RESTART the Instances D is incorrect. A canary update is a partial update to a few numbers of instances in the instance group. The requirement was to deploy it in all the VMs in the fastest and safest way For more details, please refer to the URLs below: https://cloud.google.com/compute/docs/instance-groups/ https://cloud.google.com/compute/docs/instance-groups/rolling-out-updates-to-managed-instance-groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
95
Q

You can SSH into an instance from another instance in the same VPC by its internal IP address, but not its external IP address. What is the possible reason for this scenario?
A. The outgoing instance does not have correct permission granted to its service account.
B. The internal IP address is disabled.
C. The SSH firewall rule is restricted only to the internal VPC.
D. The receiving instance has an ephemeral address instead of a reserved address.

A

Correct Answer - C GCP firewall rules are stateful. When a connection is allowed through the firewall in either direction, return traffic matching this connection is also allowed. You cannot configure a firewall rule to deny associated response traffic. Return traffic must match the 5-tuple (source IP, destination IP, source port, destination port, protocol) of the accepted request traffic, but with the source and destination addresses and ports reversed. Options A and D are incorrect. A service account represents an identity associated with an instance. Only one service account can be associated with an instance. So it is the best option in case of strict security constraints. Be careful because you cannot mix and match service accounts and network tags in any firewall rules.
Option E is incorrect because it is not necessary to provide different security to various projects. So service accounts are not required for this requirement. For any further detail, please refer to the URLs below: https://cloud.google.com/vpc/docs/using-firewalls https://cloud.google.com/vpc/docs/firewalls#service-accounts-vs-tags https://cloud.google.com/vpc/docs/firewalls#specifications

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
96
Q

You are creating a solution to remove backup files older than 90 days from your backup Cloud Storage bucket. You want to optimize ongoing Cloud Storage spending. What should you do?
A. Write a lifecycle management rule in XML and push it to the bucket with gsutil.
B. Schedule a cron script using gsutil ls -lr gs://backups/** to find and remove items older than 90 days.
C. Schedule a cron script using gsutil ls -1 gs://backups/** to find and remove items older than 90 days and schedule it with cron.
D. Write a lifecycle management rule in JSON and push it to the bucket with gsutil.

A

Correct Answer - C The firewall rule to allow SSH is restricted to the internal VPC Instances can have both Internal and External IP addresses. When connecting to another instance by its external address, you’re going out of your internal network to the external Internet and coming back to access the instance by its external address. If traffic is restricted to the local VPC, it will reject this attempt as it is coming from an external source. Reference: https://cloud.google.com/vpc/docs/firewalls#firewall_rules_in

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
97
Q

If external auditors need to be able to access your admin activity logs once a year for compliance, what is the best method of preserving and sharing that log data? (Choose two)
A. If they need access to multiple logs in a single bucket, and they have a GCP account, export logs to a Cloud Storage bucket for long-term retention and grant auditor accounts the Storage Object Viewer role to the bucket.
B. Create GCP accounts for the auditors and grant the Project Viewer role to view logs in Operations Suite (formerly Stackdriver) Logging.
C. If they do not need a GCP account and need to view a single date’s object, export the logs to a Cloud Storage bucket for long-term retention and generate a signed URL for temporary object-level access.
D. Export logs to Cloud Storage bucket and email a list of the logs once per year.

A

Correct Answer - D Opion
A - Write a lifecycle management rule in XML and push it to the bucket with gsutil: you can set lifecycle configuration for an existing bucket with a PUT API call request (NOT the “gsutil lifecycle” command!). You must include an XML document in the request body that contains the lifecycle configuration. https://cloud.google.com/storage/docs/xml-api/put-bucket-lifecycle#request_body_elements B and C can be eliminated. They do the similar thing slightly different: write script listing object and get their timestamps gsutil ls -[l or lr] gs://[BUCKET_NAME]/** If an object’s age is older than 90 days, do deleting, then schedule a cron job for the recurring process. However, gsutil ls -l/-lr does not list versioned objects. To list versioned object, need gsutil ls -a. Using this approach, versioned archives won’t be deleted. There is a better, easier, and more consistent way to do this in Answer D D (Correct answer) - Write a lifecycle management rule in JSON and push it to the bucket with gsutil. To enable lifecycle management for a bucket: https://cloud.google.com/storage/docs/managing-lifecycles · Create a .json file with the lifecycle configuration rules you would like to apply (see examples below). · Use the lifecycle set command to apply the configuration gsutil lifecycle set [LIFECYCLE_JSON-CONFIG_FILE] gs://[BUCKET_NAME] The following lifecycle configuration JSON document specifies that all objects in this bucket that are more than 90 days old will be deleted automatically: { “rule”: [ { “action”: {“type”: “Delete”}, “condition”: {“age”: 90} } ] }

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
98
Q

When creating firewall rules, what forms of segmentation can narrow which resources the rule is applied to? (Choose all that apply)
A. Network range in source filters
B. Zone
C. Region
D. Network tags

A

Correct Answer A and C Explanation For long-term logs preserving and retention: There are 3 type of sink destinations you can export Logs to: Cloud Storage, Cloud Pub/Sub, BigQuery. Export logs to Cloud Storage via an export sink. Cloud Storage is perfect solution for long-term logs retention. For Sharing: The choice to use IAM or signed URL’s depends on if the auditors need a GCP account or need access to a single object or all logs in a bucket. You could either create a GCP account for auditor ACL object access or signed URL’s depending on if they need to have a GCP account or not. Answer A is correct. If Auditors have GCP accounts, you can grant them “roles/storage.objectViewer” which can view objects and their metadata. Note the different between “storage.objectViewer” and “Project Viewer” https://cloud.google.com/storage/docs/access-control/iam-roles Cloud Storage IAM Roles Answer C is correct: “A signed URL is associated with a bucket or object and gives time-limited read or write access to that specific resource. Anyone in possession of the URL has the access granted by the URL, regardless of whether they have a Google account.” https://cloud.google.com/storage/docs/access-control/create-signed-urls-program Answer B is incorrect: Project Viewer role is not enough to view Admin Activity logs in Operations Suite (formerly Stackdriver) Logging. “To view the logs, you must have the IAM roles Logging/Private Logs Viewer or Project/Owner”. https://cloud.google.com/logging/docs/audit/#admin-activity Note: the Operations Suite (formerly Stackdriver) Admin activity log retention period is 400 days which meets and exceeds the required once-a-year access. Answer D is incorrect due to this part: “email a list of the logs once per year”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
99
Q

You are transferring a very large number of small files to Google Cloud Storage from an on-premises location. You need to speed up the transfer of your files. Assuming a fast network connection, what two actions can you do to help speed up the process? Choose the 2 correct answers:
A. Compress and combine files before transferring.
B. Use the -r option for large transfers.
C. Copy the files in bigger pieces at a time.
D. Use the -m option for multi-threading on transfers.

A

Correct Answer A and D Explanation You can restrict network access on the firewall by network tags and network ranges/subnets. Here is the console screenshot showing the options when you create firewall rules - network tags and network ranges/subnets are highlighted

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
100
Q

Your company wants to try out the cloud with low risk. They want to archive approximately 100 TB of their log data to the cloud and test the analytics features available to them there, while also retaining that data as a long-term disaster recovery backup. Which two steps should they take? Choose 2 answers
A. Load logs into Google BigQuery.
B. Import logs into Google Operations Suite (formerly Stackdriver)
C. Insert logs into Google Cloud Bigtable.
D. Load logs into Google Cloud SQL.
E. Upload log files into Google Cloud Storage.

A

Correct Answer A and D
B - Use the -r option for large transfers. The -R and -r options are synonymous. Causes directories, buckets, and bucket subdirectories to be copied recursively.
C - Copy the files in bigger pieces at a time. No applicable to the question requirements D (Correct answer) - Use the -m option for multi-threading on transfers. If you have a large number of files to transfer you might want to use the gsutil -m option, to perform a parallel (multi-threaded/multi-processing) copy: gsutil -m cp -r dir gs://my-bucket A (Correct answer) - Compress and combine files before transferring. Compressing and combining smaller files info fewer larger files is also a best practice for speeding up transfer speeds because it saves network bandwidth and space in Google Cloud Storage gsutil cp -z html -a public-read cattypes.html tabby.jpeg gs://mycats Reference cp - Copy files and objects https://cloud.google.com/storage/docs/gsutil/commands/cp

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
101
Q

You need to deploy an update to an application in Google App Engine. The update is risky, but it can only be tested in a live environment. What is the best way to introduce the update to minimize risk?
A. Deploy a new version of the application but use traffic splitting to only direct a small number of users to the new version.
B. Deploy the application temporarily and be prepared to pull it back if needed.
C. Warn users that a new app version may have issues and provide a way to contact you if there are problems.
D. Create a new project with the new app version, then redirect users to the new version.

A

Correct Answer A and E The order should be Upload log files into Google Cloud Storage and then Load logs into Google BigQuery. E (Correct answer) - Upload log files into Google Cloud Storage Cloud Storage is best solution for Long-term disaster recovery backup. You can do SQL query direct against data in Cloud Storage. It also meets the low risk requirement to prevent potential accidental data loss and modification. A (Correct answer) - Load logs into Google BigQuery - BigQuery is most suitable solution for doing analytics against large amount of data; You can do SQL query direct against data in Cloud Storage.
B - Import logs into Google Operations Suite (formerly Stackdriver) - Operations Suite (formerly Stackdriver) is not a suitable solution for Long-term disaster recovery backup
C - Insert logs into Google Cloud Bigtable: BigTable is not a suitable solution for Long-term disaster recovery backup
D - Load logs into Google Cloud SQL - Cloud SQL is relation database designed for transactional CRUD OLTP processing suitable for data less than 10 TB. Note: Our requirement is “analytics feature” ————- BigQuery is Google’s Cloud-based data warehousing solution. It targets data in big picture and can query huge volume of data in a short time. As the data is stored in columnar data format, it is much faster in scanning large amounts of data compared with BigTable. BigQuery allows to scale to petabyte and is great enterprise data warehouse for analytics. BigQuery is serverless BigTable is designed in NoSQL architecture, but can still use row-based data format. With data read/write under 10 milliseconds, it is good for applications that have frequent data ingestion. It can be scaleable to hundreds of petabytes and handle millions of operations per second. ———————–

102
Q

What is the best practice for separating responsibilities and access for production and development environments?
A. Separate project for each environment, each team only has access to their project.
B. Separate project for each environment, both teams have access to both projects.
C. Both environments use the same project, but different VPC’s.
D. Both environments use the same project, just note which resources are in use by which group.

A

Correct Answer A Explanation A (Correct Answer) - Deploying a new version without assigning it as the default version will not create downtime for the application. Using traffic splitting allows for easily redirecting a small amount of traffic to the new version and can also be quickly reverted without application downtime
B - Deploy the application temporarily and be prepared to pull it back if needed. Deploying the application new version as default requires moving all traffic to the new version. This could impact all users and disable the service during the new version’s live time.
C - Warn users that a new app version may have issues and provide a way to contact you if there are problems. We won’t recommend this practice.
D - Create a new project with the new app version, then redirect users to the new version. Deploying a second project requires data synchronization and having an external traffic splitting solution to direct traffic to the new application. While this is possible, with Google App Engine, these manual steps are not required.

103
Q

You are creating a single preemptible VM instance named “preempt” to be used as scratch space for a single workload. If your VM is preempted, you need to ensure that disk contents can be re-used. Which gcloud command would you use to create this instance?
A. gcloud compute instances create “preempt” –preemptible gcloud compute instances set-disk-auto-delete preempt –no-auto-delete
B. gcloud compute instances create “preempt” –preemptible –boot-disk-auto-delete=no
C. gcloud compute instances create “preempt” –preemptible
D. gcloud compute instances create “preempt” –no-auto-delete

A

Correct Answer A Explanation A (Correct answer) - Separate project for each environment, each team only has access to their project. For least privilege and separation of duties, the best practice is to separate both environments into different projects, development or production team gets their own accounts, and each team is assigned to only their projects. The best practices: · You should not use same account for both Development and production environments regardless how do you create projects inside that account for different environments. You should use different account for each environment which associated with different group of users. You should use project to isolate user access to resource not to manage users. · Using a shared VPC allows each team to individually manage their own application resources, while enabling each application to communicate between each other securely over RFC1918 address space. So VPC’s isolate resources but not user/service accounts. B, C, and D are incorrect Answer B is the scenario that use same account for both development and production environments attempting to isolate user access with different projects Answer C is the scenario that use same account for both development and production environments with same project attempting to isolate user access with network separation. Answer D is the scenario that use same account for both development and production environments with same project attempting to isolate user access with user group at resource level. You may grant roles to group of users to set policies at organization level, project level, or (in some cases) the resource (e.g., existing Cloud Storage and BigQuery ACL systems as well as and Pub/Sub topics) level. The best practice: Set policies at the Organization level and at the Project level rather than at the resource level. This is because as new resources get added, you may want them to automatically inherit policies from their parent resource. For example, as new Virtual Machines gets added to the project through auto scaling, they automatically inherit the policy on the project. https://cloud.google.com/iam/docs/resource-hierarchy-access-control#best_practices Additional Resources: To recap: IAM lets you control who (users) has what access (roles) to which resources by setting IAM policies. IAM policies grant specific role(s) to a user giving the user certain permissions. https://cloud.google.com/resource-manager/docs/access-control-org Using Resource Hierarchy for Access Control https://cloud.google.com/iam/docs/resource-hierarchy-access-control#background

104
Q

A production database virtual machine on Google Compute Engine has an ext4-formatted persistent disk for data files. The database is about to run out of storage space. How can you remediate the problem with the least amount of downtime?
A. In the Cloud Platform Console, increase the size of the persistent disk and use the resize2fs command in Linux.
B. Shut down the virtual machine, use the Cloud Platform Console to increase the persistent disk size, then restart the virtual machine.
C. In the Cloud Platform Console, increase the size of the persistent disk and verify the new space is ready to use with the fdisk command in Linux.
D. In the Cloud Platform Console, create a new persistent disk attached to the virtual machine, format and mount it, and configure the database service to move the files to the new disk.
E. In the Cloud Platform Console, create a snapshot of the persistent disk, restore the snapshot to a new larger disk, unmount the old disk, mount the new disk, and restart the database service.

A

Correct Answer A Explanation A (Correct answer) - Specifying ‘–no–auto-delete’ preserves the disk. This flag is not enabled by default so if not specify, it causes the disk to be auto-deleted. https://cloud.google.com/sdk/gcloud/reference/compute/instances/set-disk-auto-delete
B - The default is boot disk automatically delete and no flag needed, also the syntax is incorrect for this type of flags
C - if you don’t specify ‘–no-boot-disk-auto-delete’. The default would be boot disk automatically delete Here is the corresponding console setting displaying the default option
D - when instance created without this flag: –preemptible, it’ll be standard instance Here is the corresponding console setting in “Availability Policy” when you create instance with –preemptible flag

105
Q

Over time, you’ve created 5 snapshots of a single instance. To save space, you delete snapshots number 3 and 4. What has happened to the fifth snapshot?
A. The data from both snapshots 3 and 4 necessary for continuance are transferred to snapshot 5.
B. It is no longer useable and cannot restore data.
C. All later snapshots, including 5, are automatically deleted as well.
D. The data from snapshot 4 necessary for continuance was transferred to snapshot 5, however snapshot 3’s contents were transferred to snapshot 2.

A

Correct Answer A Explanation Answers B (Shut down VM), D (Move files to new attached disk), and E (Use snapshot to restore … restart the database service) all have some sorts of downtime, so they can be ruled out. A (Correct answer) - In the Cloud Platform Console, increase the size of the persistent disk and use the resize2fs command in Linux. You can resize persistent disks when your instances require more storage space and attach multiple secondary disks only when you need to separate your data into unique partitions. You can resize disks at any time, regardless of whether the disk is attached to a running instance. You can use console or command line to resize the disk: gcloud compute disks resize [DISK_NAME] –size [DISK_SIZE] After you resize your persistent disk, you must configure the file system on the disk to use the additional disk space. If the disk has a partition table, such as a boot disk, you must grow the partition and resize the file system on that partition. If your persistent disk has only a file system and no partition table, you can just resize the file system. Extend the file system on the disk or the partition to use the added space. If you grew a partition on your disk, specify the partition. If your disk does not have a partition table, specify only the disk ID. The resize2fs is Linux program to resize ext2, ext3, or ext4 file systems. sudo resize2fs /dev/[DEVICE_ID][PARTITION_NUMBER]
C - In the Cloud Platform Console, increase the size of the persistent disk and verify the new space is ready to use with the fdisk command in Linux. This answer is incomplete: after indicating size increase in console, to make the new size effective, you have two options: restart the VM or configure (Grow partition if needed and expand partition/file system) in the VM’s operating systems, windows or linux Reference Resources https://cloud.google.com/compute/docs/disks/add-persistent-disk Adding or Resizing Persistent Disks Additional Resource Update: Now you have the option to enable “Automatic storage increase” Instance Settings https://cloud.google.com/sql/docs/mysql/instance-settings#automatic-storage-increase-2ndgen Automatic storage increase If this setting is enabled, your available storage is checked every 30 seconds. If available storage falls below a threshold size, additional storage capacity is automatically added to your instance.

106
Q

You need to have a backup/rollback plan in place for your application that is distributed across a large managed instance group. What is the preferred method for doing so?
A. Use the Rolling Update feature to deploy/roll back versions with different managed instance group templates.
B. Use the managed instance group snapshot function that is included in Compute Engine.
C. Have each instance write critical application data to a Cloud Storage bucket.
D. Schedule a cron job to take snapshots of each instance in the group.

A

Correct Answer A Explanation Deleting a snapshot: https://cloud.google.com/compute/docs/disks/restore-and-delete-snapshots When you delete a snapshot, Compute Engine immediately marks the snapshot as DELETED in the system. If the snapshot has no dependent snapshots, it is deleted outright. However, if the snapshot does have dependent snapshots: 1) Any data that is required for restoring other snapshots is moved into the next snapshot, increasing its size. 2) Any data that is not required for restoring other snapshots is deleted. This lowers the total size of all your snapshots. 3) The next snapshot no longer references the snapshot marked for deletion, and instead references the snapshot before it. Because subsequent snapshots might require information stored in a previous snapshot, keep in mind that deleting a snapshot does not necessarily delete all the data on the snapshot. As mentioned in the first bullet above, if any data on a snapshot that is marked for deletion is needed for restoring subsequent snapshots, that data is moved into the next corresponding snapshot. To definitively delete data from your snapshots, you should delete all snapshots. The diagram below illustrates the process described above:

107
Q

For this question, refer to the Mountkirk Games case study. Mountkirk Games has deployed their new backend on Google Cloud Platform (GCP). You want to create a thorough testing process for new versions of the backend before they are released to the public. You want the testing environment to scale in an economical way. How should you design the process?
A. Create a scalable environment in GCP for simulating production load.
B. Build stress tests into each component of your application using resources internal to GCP to simulate load.
C. Use the existing infrastructure to test the GCP-based backend at scale.
D. Create a set of static environments in GCP to test different levels of load – for example, high, medium, and low.

A

Correct Answer A Explanation Individual VM’s in a managed instance group should be treated as disposable entities and should not be individually backed up. Using the rolling update feature in your managed instance group allows you to use alternate managed instance templates for managing different versions of your application. https://cloud.google.com/compute/docs/instance-groups/updating-managed-instance-groups

108
Q

You are designing a large distributed application with 30 microservices. Each of your distributed microservices needs to connect to a database back-end. You want to store the credentials securely. Where should you store the credentials?
A. In a secret management system
B. In the source code
C. In an environment variable
D. In a config file that has restricted access through ACLs

A

Correct Answer A A (Correct answer) - Create a scalable environment in GCP for simulating production load. With this disposable and repeatable testing resources, you can do load test whenever needed. Shutdown or stop the services or simplify delete and recreate it based on the test plans, to keep the cost low. It meets the requirements “create a thorough testing process for new versions of the backend before they are released to the public” and” testing environment to scale in an economical way”. Doing thorough testing on production infrastructure is risky to other running application, not feasible, not scale in economical way.
B - Build stress tests into each component of your application using resources internal to GCP to simulate load. This is not scale nor economical and too complicated to implement.
C - Use the existing infrastructure to test the GCP-based backend at scale. At first glance, reuse exiting environments so it’ll be scalable, economical, and in the real situation. If we read the case study again, we know that Mountkirk games is a popular game platform targeting to global users with very high traffic and heavy load. Doing a load test on the production is no longer an option, nor is it necessarily a scale in an economical way if you mix the production and testing load. Comparing to the solution creating disposable and reputable testing environment simulating production load and execute test plans on demanding, Answer A is the winner.
D - Create a set of static environments in GCP to test different levels of load - for example, high, medium, and low. This is nor scale nor economical

109
Q

Your company collects and stores security camera footage in Google Cloud Storage. Within the first 30 days, footage is processed regularly for threat detection, object detection, trend analysis, and suspicious behavior detection. You want to minimize the cost of storing all the data. How should you store the videos?
A. Use Standard Google CloudStorage for the first 30 days, and then move to Coldline Storage.
B. Use Google Cloud Nearline Storage for the first 30 days, and then move to Coldline Storage.
C. Use Google Cloud Regional Storage for the first 30 days, and then move to Nearline Storage.
D. Use Google Cloud Regional Storage for the first 30 days, and then move to Google Persistent Disk.

A

Correct Answer A A (Correct answer) - In a secret management system Applications often require access to small pieces of sensitive data at build or run time. These pieces of data are often referred to as secrets. Secrets are similar in concept to configuration files, but are generally more sensitive, as they may grant access to additional data, such as user data. https://cloud.google.com/kms/docs/secret-management
B - In the source code: This is exactly again the best practice “Do not embed secrets related to authentication in source code, such as API keys, OAuth tokens, and service account credentials.” (see below the best practice #1)
C - In an environment variable - you use environment variable to point to the location where the secrets (credentials) are stored other than store the secrete directly (see below the best practice #1
D - In a configuration file that has restricted access through ACLs - Secrets are similar to but generally more sensitive than configuration and also, ACLs may not enough for the secrete management. Here is example for Storing secrets https://cloud.google.com/kms/docs/store-secrets Additional Resource https://cloud.google.com/docs/authentication/production#providing_credentials_to_your_application Best practices for managing credentials Credentials provide access to sensitive data. The following practices help protect access to these resources: 1) Do not embed secrets related to authentication in source code, such as API keys, OAuth tokens, and service account credentials. You can use an environment variable pointing to credentials outside of the application’s source code, such as Cloud Key Management Service. 2) Do use different credentials in different contexts, such as in testing and production environments. 3) Do transfer credentials only over HTTPS to prevent a third party from intercepting your credentials. Never transfer in clear text or as part of the URL. 4) Never embed long-lived credentials into your client-side app. For example, do not embed service account credentials into a mobile app. Client-side apps can be examined, and credentials can easily be found and used by a third party. Do revoke a token if you no longer need it.

110
Q

You have created a Kubernetes engine cluster named ‘mycluster’. You’ve realized that you need to change the machine type for the cluster from n1-standard-1 to n1-standard-4. What is the command to make this change?
A. You must create a new node pool in the same cluster and migrate the workload to the new pool.
B. gcloud container clusters update mycluster –machine-type n1-standard-4
C. This action is not possible.
D. gcloud container clusters resize mycluster –machine-type n1-standard-4

A

Correct Answer A A (Correct Answer) - Since the data is accessed frequently within the first 30 days, using Standard Google Cloud Storage will enable the most cost-effective solution for storing and accessing the data. For videos older than 30 days, Google Cloud Coldline Storage offers the most cost-effective solution since it won’t be accessed.
B - While Google Cloud Coldline storage is cost-effective for long-term video storage, Google Cloud Nearline Storage would not be an effective solution for the first 30 days as the data is expected to be accessed frequently.
C - While Google Cloud Regional Storage is the most cost-effective solution for the first 30 days, Google Cloud Nearline Storage is not cost effective for long-term storage.
D - While Google Cloud Regional Storage is the most cost-effective solution for the first 30 days, storing the data on Google Cloud Persistent Disk would not be cost-effective for long term storage.

111
Q

You have a GKE cluster with 1 node-pool. The cluster receives a lot of traffic and needs to grow. You decide to add a node. What should you do?
A. Use “gcloud container clusters resize” with the desired number of nodes.
B. Use “kubectl container clusters resize” with the desired number of nodes.
C. Edit the managed instance group of the cluster and increase the number of VMs by 1.
D. Edit the managed instance group of the cluster and enable autoscaling.

A

Correct answer A A (Correct answer) - You must create a new node pool in the same cluster and migrate the workload to the new pool. You cannot change the machine type for an individual node pool after creation. You need to create a new pool and migrate your workload over. Here are the steps for “Migrating workloads to different machine types” https://cloud.google.com/kubernetes-engine/docs/tutorials/migrating-node-pool
B - gcloud container clusters update mycluster –machine-type n1-standard-4 updates cluster settings for an existing container cluster. You can use this command to specify –max-nodes –min-nodes for autoscaling purpose not for changing machines type https://cloud.google.com/sdk/gcloud/reference/container/clusters/update
C - This action is not possible. It’s possible to migrate workloads running on a Kubernetes Engine cluster to a new set of nodes without incurring downtime for your application. See “Migrate the workloads” https://cloud.google.com/kubernetes-engine/docs/tutorials/migrating-node-pool#step_4_migrate_the_workloads
D - gcloud container clusters resize mycluster –machine-type n1-standard-4. Resizes an existing cluster for running containers not for changing machine type https://cloud.google.com/sdk/gcloud/reference/container/clusters/resize

112
Q

The application reliability team at your company has added a debug feature to their backend service to send all server events to Google Cloud Storage for eventual analysis The event records are at least 50 KB and at most 15 MB and are expected to peak at 3,000 events per second. You want to minimize data loss. Which process should you implement?
A. Append metadata to file body. Compress individual files. Name files with a random prefix pattern. Save files to one bucket
B. Batch every 10,000 events with a single manifest file for metadata. Compress event files and manifest file into a single archive file. Name files using serverName-EventSequence. Create a new bucket if bucket is older than 1 day and save the single archive file to the new bucket. Otherwise, save the single archive file to existing bucket.
C. Compress individual files. Name files with serverName-EventSequence. Save files to one bucket Set custom metadata headers for each object after saving.
D. Append metadata to file body. Compress individual files. Name files with serverName-Timestamp. Create a new bucket if bucket is older than 1 hour and save individual files to the new bucket. Otherwise, save files to existing bucket

A

Correct answer A A is correct because this resizes the cluster to the desired number of nodes. B is not correct because you need to use gcloud, not kubectl. C is not correct because you should not manually manage the MIG behind a cluster. D is not correct because you should not manually manage the MIG behind a cluster.

113
Q

You have a mission-critical database running on an instance on Google Compute Engine. You need to automate a database backup once per day to another disk. The database must remain fully operational and functional and can have no downtime. How can you best perform an automated backup of the database with minimal downtime and minimal costs?
A. Use a cron job to schedule your application to backup the database to another persistent disk.
B. Use a cron job to schedule a disk snapshot once per day.
C. Write the database to two different disk locations simultaneously, then schedule a snapshot of the secondary disk, which will allow the primary disk to continue running.
D. Use the automated snapshot service on Compute Engine to schedule a snapshot.

A

Correct Answer A Avoid using sequential filenames such as timestamp-based filenames if you are uploading many files in parallel. Because files with sequential names are stored consecutively, they are likely to hit the same backend server, meaning that throughput will be constrained. In order to achieve optimal throughput, you can add the hash of the sequence number as part of the filename to make it non-sequential https://cloud.google.com/storage/docs/best-practices Answer A (Correct) - since it uses “Name files with a random prefix pattern.” Answer C , B, and D are incorrect since they use either “Name files with serverName-EventSequence” Or “Name files with serverName-Timestamp” which will cause the files unevenly distributed in the backend. For example, a specific server may generate much more events than other, or at certain time period the system may generate much more events than other period…

114
Q

Your company is planning on deploying a web application to Google Cloud hosted on a custom Linux distribution. Your website will be accessible globally and needs to scale to meet demand. Choose all of the components that will be necessary to achieve this goal. Choose the 2 correct answers:
A. App Engine Standard environment
B. HTTP Load Balancer
C. Managed Instance Group on Compute Engine
D. Network Load Balancer

A

Correct answer A To both minimize costs (don’t want extra disks) and minimize downtime (cannot freeze database). Backing up just the database to another disk using a cron job is the preferred answer. It is also possible to backup the database to a Cloud Storage bucket instead of a disk, which would be cheaper for the same amount of storage. B and D all have some sort of Database downtime due to the snapshot. Answer C would be hard to implement and use doubled resources. You’ll also lost the data consistency if you don’t freeze the primary database when you take snapshot on secondary database. Overall, it’s not worthwhile for your efforts for this task when you have better solution like answer A.

115
Q

Your application is hosted across multiple regions and consists of both relational database data and static images. Your database has over 10 TB of data. You want to use a single storage repository for each data type across all regions. Which two products would you choose for this task? (Choose two)
A. Cloud Bigtable
B. Cloud Spanner
C. Cloud SQL
D. Cloud Storage

A

Correct Answer B and C Explanation B and C (Correct answers) - A managed instance group can auto scale and use a custom Linux distribution. An HTTP load balancer serves web traffic and is a global load balancer (single load balancer for all GCP regions worldwide).
A - App Engine would support autoscaling but does not use custom Linux distributions.
D - A Network Load Balancer might be used in conjunction with an HTTP Load Balancer for backend functions but is not the required component in this scenario.

116
Q

To ensure that your application will handle the load even if an entire zone fails, what should you do? Select all correct options.
A. Don’t select the “Multizone” option when creating your managed instance group.
B. Spread your managed instance group over two zones and overprovision by 100%. (for Two Zone)
C. Create a regional unmanaged instance group and spread your instances across multiple zones.
D. Overprovision your regional managed instance group by at least 50%. (for Three Zones)

A

Correct Answer B and D Cloud Spanner acts as an SQL database that is horizontally scalable for cross-region support and can host large datasets. Cloud Storage supports multi-regional buckets for high performance from different regions.

117
Q

You are asked to design the next generation of smart helmet for accident detection and reporting system. Each helmet will push 10kb of biometric data In JSON format every 1 second to a collection platform that will process and use a trained machine learning model to predict and detect if an accident happens and send a notification. Management has tasked you to architect the platform ensuring the following requirements are met: · Provide the ability for real-time analytics of the inbound biometric data · Ensure the processing of the biometric data is highly durable. Elastic and parallel · The results of the analytic processing should be persisted for data mining to improve the accident detection ML model in the future. Which architecture outlined below meet the initial requirements for the platform?
A. Utilize Cloud Storage to collect the inbound sensor data, analyze data with Dataproc and save the results to BigQuery.
B. Utilize Cloud Pub/Sub to collect the inbound sensor data, process the data with DataFlow and save the results to BigQuery for further analysis.
C. Utilize Cloud Pub/Sub to collect the inbound sensor data, analyze the data with DataFlow and save the results to Cloud SQL.
D. Utilize Cloud Pub/Sub to collect the inbound sensor data, analyze the data with DataFlow and save the results to BigTable.

A

Correct answer B and D B is correct if one zone fails you still have 100% desired capacity in another zone C is incorrect because it won’t be able to handle the full load since, it’s unmanaged group and won’t auto scale accordingly. D is correct since you have at least total 150% desired capacity spread over 3 zones, each zone has 50% capacity. You’ll have 100% desired capacity in two zones if any single zone failed at given time. Reference Resources https://cloud.google.com/compute/docs/instance-groups/distributing-instances-with-regional-instance-groups If you are creating a regional managed instance group in a region with at least three zones, Google recommends overprovisioning your instance group by at least 50%.

118
Q

A small number of API requests to your microservices-based application take a very long time. You know that each request to the API can traverse many services. You want to know which service takes the longest in those cases. What should you do?
A. Set timeouts on your application so that you can fail requests faster.
B. Instrument your application with Operations Suite (formerly Stackdriver) Trace to break down the request latencies at each microservice.
C. Send custom metrics for each of your requests to Operations Suite (formerly Stackdriver) Monitoring.
D. Use Operations Suite (formerly Stackdriver) Monitoring to look for insights that show when your API latencies are high.

A

Correct Answer B Answer B meet all of the 3 requirements: Cloud Pub/Sub is a simple, reliable, scalable foundation for stream analytics and event-driven computing systems. As part of Google Cloud’s stream analytics solution, the service ingests event streams and delivers them to Cloud Dataflow for processing and BigQuery for analysis as a data warehousing solution. Relying on the Cloud Pub/Sub service for delivery of event data frees you to focus on transforming your business and data systems with applications such as: · check Real-time personalization in gaming · check Fast reporting, targeting and optimization in advertising and media · check Processing device data for healthcare, manufacturing, oil and gas, and logistics · check Syndicating market-related data streams for financial services Also, Use Cloud Dataflow as a convenient integration point to bring predictive analytics to fraud detection, real-time personalization and similar use cases by adding TensorFlow-based Cloud Machine Learning models and APIs to your data processing pipelines. https://cloud.google.com/ml-engine/ BigQuery provides a flexible, powerful foundation for Machine Learning and Artificial Intelligence. BigQuery provides integration with CloudML Engine and TensorFlow to train powerful models on structured data. Moreover, BigQuery’s ability to transform and analyze data helps you get your data in shape for Machine Learning. https://cloud.google.com/bigquery/ Other solutions may work one way or other but only the combination of theses 3 components integrate well in data ingestion, collection, and real-time analysis, and data mining in a highly durable, elastic, and parallel manner.
A - Cloud storage is not suitable for this kind of real-time streaming data collection; Dataproc is GCP’s BigData Hadoop/Spark that can do ETL and analysis, but DataFlow provides a simple unified programming model for ETL and analysis in both real-time and batch.
C - Cloud SQL is mainly for OLTP (Transactional, CRUD) not for OLAP (On-line Analytical Processing, Data Warehouse). It does not have the scalability, elasticity, and parallel to absorb this amount of Data in real-time. Instead, BigQuery integrates well with DataFlow and can absorb both streaming and batch data from it.
D - Bigtable is one of the possible Data sink for DataFlow and have the capability to absorb this amount of real time data but it lacks the Data mining features like BigQuery. Further Explanation Pub/Sub is a kind of ‘shock absorber’, allowing asynchronous messaging between large numbers of devices. Cloud Dataflow acts as your data processing pipeline for ETL functions on both streaming and batch data. BigQuery is a data warehouse, able to run analysis on petabytes of data using SQL queries. Below is a reference architect Google recommending for similar scenario in Real-time streaming data collection and analysis https://cloud.google.com/solutions/mobile/mobile-gaming-analysis-telemetry Real-time processing of events from game clients and game servers Data Transformation with Cloud Dataflow - Dataflow acts as your data processing pipeline for ETL functions on both streaming and batch data.

119
Q

Your solution is producing performance bugs in production that you did not see in staging and test environments. You want to adjust your test and deployment procedures to avoid this problem in the future. What should you do? Select one.
A. Deploy changes to a small subset of users before rolling out to production.
B. Increase the load on your test and staging environments and Deploy smaller changes to production.
C. Deploy smaller changes to production.
D. Deploy fewer changes to production.

A

Correct Answer B Explanation
A - Set timeouts on your application so that you can fail requests faster - This won’t be able to tell you directly where the bottleneck is. B (Correct Answer) - Instrument your application with Operations Suite (formerly Stackdriver) Trace to break down the request latencies at each microservice. This is exactly Operations Suite (formerly Stackdriver) Trace comes to play.
C - Send custom metrics for each of your requests to Operations Suite (formerly Stackdriver) Monitoring - without knowing where the bottleneck is beforehand, it’s not easy, if not impossible, to setup custom metrics to capture the latency causes. Besides, the question itself is about to find where the latency/bottleneck exists.
D - Use Operations Suite (formerly Stackdriver) Monitoring to look for insights that show when your API latencies are high - this could tell you when the API call latency reaching to certain threshold/criteria but can hardly tell where the root causes is without additional setup and analysis. Reference Resources Operations Suite (formerly Stackdriver) Trace can help you answer the following questions: https://cloud.google.com/trace/docs/overview · How long does it take my application to handle a given request? · Why is it taking my application so long to handle a request? · Why do some of my requests take longer than others? · What is the overall latency of requests to my application? · Has latency for my application increased or decreased over time? · What can I do to reduce application latency? “As micro-services become more popular, the cross-application tracing provided by Operations Suite (formerly Stackdriver) Trace becomes essential in pinpointing the root cause of latency issues.”

120
Q

For this question refer to the TerramEarth case study. Which of TerramEarth’s legacy enterprise processes in their existing data centers would experience significant change as a result of increased Google Cloud Platform adoption.
A. Opex (operational expenditures) and capex (capital expenditures) allocation, LAN changes, capacity planning.
B. Capacity planning, TCO calculations, Opex and Capex allocation.
C. Capacity planning, utilization measurement, data center expansion.
D. Data Center expansion, TCO calculations, utilization measurement.

A

Correct Answer B
A - Deploy changes to a small subset of users before rolling out to production. This is the practice in Canary deployment. The bug slip into production may be caused by the discrepancy between test/staging and production environments or testing data. With Canary deployment or Canary test, you have the ability to test code with live data at any time, you increase the chance discovering the bug earlier and reduced the risk bring the bug into production with minimums impact and down time by rolling back quickly. But the canary deployment will not able to test the performance bugs in the environment. B (Correct Answer) - Increase the load on your test and staging environments. Increase the load in your test and staging environment will help to discover the bugs revolving around the performance issue. C and
D - Deploy smaller or fewer changes to production. Although those are generally good agile practices for cloud native microservice, they don’t address the issues to adjust your test and deployment procedures to discover the bugs before in production. The Bug can still slip into production no matter how small how often you test the changes in same environment and same set of test data in same procedures.

121
Q

Your company has reserved a monthly budget for your project. You want to be informed automatically of your project spend so that you can take action when you approach the limit. What should you do?
A. Link a credit card with a monthly limit equal to your budget.
B. Create a budget alert for desired percentages such as 50%, 90%, and 100% of your total monthly budget.
C. In App Engine Settings, set a daily budget at the rate of 1/30 of your monthly budget.
D. In the GCP Console, configure billing export to BigQuery. Create a saved view that queries your total spend.

A

Correct Answer B
A - Opex and capex allocation is part of answers; GCP adoption would not cause significant LAN changes. B (Correct Answer) - Capacity planning, TCO calculations, Opex and Capex allocation - those are all in the scopes concerned. From the case study, it can conclude that Management (CXO) all concern rapid provision of resources (infrastructure) for growing as well as cost management, such as Cost optimization in Infrastructure, trade up front capital expenditures (Capex) for ongoing operating expenditures (Opex), and Total cost of ownership (TCO)
C - Capacity planning, utilization measurement, data center expansion - their data center would be shrinking instead of expanding if increasing Google Cloud Platform adoption.
D - Data Center expansion, TCO calculations, utilization measurement - “Data Center expansion” is wrong choice; “utilization measurement” is not necessary a significant change caused by GCP adoption; Also, this answer is not as complete as Answer B Additional Resource Please read TerramEarth case study carefully to draw and extract your conclusions applicable to this questions and answers.

122
Q

Your customer is moving their corporate applications to Google Cloud Platform. The security team wants detailed visibility of all projects in the organization. You provision the Google Cloud Resource Manager and set up yourself as the org admin. Which Google Cloud Identity and Access Management (Cloud IAM) roles should you give to the security team?
A. Org viewer, project owner
B. Org viewer, project viewer
C. Org admin, project browser
D. Project owner, network admin

A

Correct answer B A is not correct because this will just give you the spend but will not alert you when you approach the limit. B Is correct because a budget alert will warn you when you reach the limits set. C Is not correct because those budgets are only on App Engine, not other GCP resources. Furthermore, this makes subsequent requests fail, rather than alert you in time so you can mitigate appropriately. D is not correct because if you exceed the budget, you will still be billed for it. Furthermore, there is no alerting when you hit that limit by GCP. Reference AppEngine Spending Limit https://cloud.google.com/appengine/pricing#spending_limit Set Budgets and Alerts https://cloud.google.com/billing/docs/how-to/budgets

123
Q

Your company is developing a next generation pet collar that collects biometric information to assist potential millions of families with promoting healthy lifestyles for their pets. Each collar will push 30kb of biometric data In JSON format every 2 seconds to a collection platform that will process and analyze the data providing health trending information back to the pet owners and veterinarians via a web portal. Management has tasked you to architect the collection platform ensuring the following requirements are met. Provide the ability for real-time analytics of the inbound biometric data Ensure processing of the biometric data is highly durable, elastic and parallel The results of the analytic processing should be persisted for data mining Which architecture outlined below win meet the initial requirements for the platform?
A. Utilize Cloud Storage to collect the inbound sensor data, analyze data with Dataproc and save the results to BigQuery.
B. Utilize Cloud Pub/Sub to collect the inbound sensor data, analyze the data with DataFlow and save the results to BigQuery.
C. Utilize Cloud Pub/Sub to collect the inbound sensor data, analyze the data with DataFlow and save the results to Cloud SQL.
D. Utilize Cloud Pub/Sub to collect the inbound sensor data, analyze the data with DataFlow and save the results to BigTable.

A

Correct Answer B Answer B gives the security team read only access to everything your company produces, anything else gives them the ability to, accidentally or otherwise, change things, a violation to the principle of least privilege.

124
Q

You are working on a project with two compliance requirements. The first requirement states that your developers should be able to see the Google Cloud Platform billing charges for only their own projects. The second requirement states that your finance team members can set budgets and view the current charges for all projects in the organization. The finance team should not be able to view the project contents. You want to set permissions. What should you do?
A. Add the finance team members to the default IAM Owner role. Add the developers to a custom role that allows them to see their own spend only.
B. Add the finance team members to the Billing Administrator role for each of the billing accounts that they need to manage. Add the developers to the Viewer role for the Project.
C. Add the developers and finance managers to the Viewer role for the Project.
D. Add the finance team to the Viewer role for the Project. Add the developers to the Security Reviewer role for each of the billing accounts.

A

Correct Answer B Answer B meet all of the 3 requirements: Cloud Pub/Sub is a simple, reliable, scalable foundation for stream analytics and event-driven computing systems. As part of Google Cloud’s stream analytics solution, the service ingests event streams and delivers them to Cloud Dataflow for processing and BigQuery for analysis as a data warehousing solution. Relying on the Cloud Pub/Sub service for delivery of event data frees you to focus on transforming your business and data systems with applications such as: · check Real-time personalization in gaming · check Fast reporting, targeting and optimization in advertising and media · check Processing device data for healthcare, manufacturing, oil and gas, and logistics · check Syndicating market-related data streams for financial services Other solutions may work one way or other but only the combination of theses 3 components integrate well in data ingestion, collection, and real-time analysis, and data mining in a highly durable, elastic, and parallel manner.
A - Cloud storage is not suitable for this kind of real-time streaming data collection; Dataproc is GCP’s BigData Hadoop/Spark that can do ETL and analysis, but DataFlow provide simple unified programming model for ETL and analysis in bot Realtime and batch and integrate well with PubSub.
C - Cloud SQL is mainly for OLTP (Transactional, CRUD) not for OLAP (On-line Analytical Processing, Data Warehouse). It does not have the scalability, elasticity, and parallel to absorb this amount of Data in real time. Instead BigQuery integrate well with DataFlow and can absorb both steaming and batch data from it.
D - Bigtable is one of the possible Data sink for DataFlow and have the capability to absorb this amount of real time data but it lacks the Data mining features like BigQuery. Further Explanation Pub/Sub is kind of ‘shock absorber’, allowing asynchronous messaging between large numbers of devices. Cloud Dataflow acts as your data processing pipeline for ETL functions on both streaming and batch data. BigQuery is a data warehouse, able to run analysis on petabytes of data using SQL queries. Below is a reference architect Google recommending for similar scenario in Real-time streaming data collection and analysis https://cloud.google.com/solutions/mobile/mobile-gaming-analysis-telemetry Real-time processing of events from game clients and game servers Data Transformation with Cloud Dataflow - Dataflow acts as your data processing pipeline for ETL functions on both streaming and batch data.

125
Q

Your company is building a large-scale web application. Each team is responsible for its own service component of the application and wants to manage its own individual projects. You want each service to communicate with the others over RFC1918 address space. What should you do?
A. Deploy each service into a single project within the same VPC.
B. Configure Shared VPC and add each project as a service of the Shared VPC project.
C. Configure each service to communicate with the others over HTTPS protocol.
D. Configure a global load balancer for each project and communicate between each service using the global load balancer IP addresses.

A

Correct answer B B (Correct Answer) - B is correct because it uses the principle of least privilege for IAM roles; use the Billing Administrator IAM role for that job function. A, C, and D are not correct because is it a Google best practice to use pre-defined IAM roles when they exist and match your business scenario; see the link below. Reference IAM for Billing: https://cloud.google.com/iam/docs/job-functions/billing

126
Q

Your company’s user-feedback portal comprises a standard LAMP stack replicated across two zones. It is deployed in the us-central1 region and uses autoscaled managed instance groups on all layers, except the database. Currently, only a small group of select customers have access to the portal. The portal meets a 99.99% availability SLA under these conditions. However next quarter, your company will be making the portal available to all users, including unauthenticated users. You need to develop a resiliency testing strategy to ensure the system maintains the SLA once they introduce additional user load. What should you do?
A. Expose the new system to a larger group of users and increase group’s size each day until autoscale logic is triggered on all layers. At the same time, terminate random resources on both zones.
B. Create synthetic random user input, replay synthetic load until autoscale logic is triggered on at least one layer, and introduce “chaos” to the system by terminating random resources on both zones.
C. Capture existing users input and replay captured user load until autoscale is triggered on all layers. At the same time, terminate all resources in one of the zones.
D. Capture existing users input and replay captured user load until resource utilization crosses 80%. Also, derive estimated number of users based on existing users’ usage of the app, and deploy enough resources to handle 200% of expected load.

A

Correct Answer B B (Correct Answer) - Configure Shared VPC and add each project as a service of the Shared VPC project. Using a shared VPC allows each team to individually manage their own application resources, while enabling each application to communicate between each other securely over RFC1918 address space. The following example illustrates a simple Shared VPC scenario: Instead of a VPC being within a single project, Shared VPC allows the VPC to exist in multiple projects: - Instance/Applications in Service Project A can communicate with Instance/Applications in Service Project B. - Neither Instance/Applications in Project A or Project B can communicate with the Standalone Project (Bottom) - Of course, within the Standalone Project, Instance 1 can commutate with Instance 2 normally.
A - Deploy each service into a single project within the same VPC. Deploying services into a single project results in every team accessing and managing the same project resources. This is difficult to manage and control as the number of teams involved increases.
C - Configure each service to communicate with the others over HTTPS protocol. HTTPS is a valid option; however, this answer does not address the need to ensure management of individual projects.
D - Configure a global load balancer for each project and communicate between each service using the global load balancer IP addresses The global load balancer uses a public IP address, and therefore it does not conform to the requirement of communication over RFC1918 address space.

127
Q

You’ve created a Kubernetes engine cluster named “mycluster”, which has a cluster pool named ‘primary-node-pool’. You’ve realized that you need more total nodes within your cluster pool to meet capacity demands from 10 to 20. What is the command to change the number of nodes in your pool?
A. gcloud container clusters resize mycluster –node-pool ‘primary-node-pool’ –max-nodes= 20
B. gcloud container clusters resize mycluster –node-pool ‘primary-node-pool’ –num-nodes = 20
C. You must create a new node pool and transfer the workload to the new pool.
D. gcloud container clusters update mycluster –node-pool ‘primary-node-pool’ –num-nodes = 20

A

Correct Answer B B (Correct answer) - Create synthetic random user input, replay synthetic load until autoscale logic is triggered on at least one layer, and introduce “chaos” to the system by terminating random resources on both zones. Ideally, the test environment should always resemble production in term of infrastructure and workload but it’s not always realistic. For example, testing data with full user population is not available, or test environment at production level is not ready. In the question scenario, you can test in pre-production with a synthetic workload from existing users which simulates a production workload and introduce “chaos” to simulate one zone failure to test the availability for the SLA. This enables you to decouple deployment from release, get real user feedback, test for bugs, and assess infrastructure performance. Answers A, C, and D either incorrect or incomplete in workload simulation for testing users’ generation, or zone (s) failure simulation for availability SLA. Additional Resources Provisioning a regional managed instance group in two zones: https://cloud.google.com/compute/docs/instance-groups/distributing-instances-with-regional-instance-groups#provisioning_a_regional_managed_instance_group_in_two_zones

128
Q

Your developer currently maintains a J2EE application. What two considerations should he consider for moving his application to the cloud to meet demand and minimize overhead? (Choose two)
A. Host resources in Cloud Bigtable.
B. Configure application for Cloud Dataproc.
C. Re-tooling the application to run on App Engine.
D. Re-tool to run on managed instance group with appropriate storage backend.

A

Correct Answer B: B (Correct Answer). The command to resize an existing GKE node pool is: gcloud container clusters resize NAME (–num-nodes=NUM_NODES | –size=NUM_NODES) [–async] [–node-pool=NODE_POOL] [–region=REGION | –zone=ZONE
, -z ZONE] [GCLOUD_WIDE_FLAG …]
Option B is correct as you have to use –num-nodes flag.
Option A uses a wrong flag “–max-nodes”
Option C is describing a situation for changing machine type
Option D “gcloud container clusters update”. This updates cluster settings for an existing container cluster. You can use this command to specify –max-nodes –min-nodes for autoscaling purpose. Also “–num-nodes” is a wrong flag option for this command. https://cloud.google.com/sdk/gcloud/reference/container/clusters/update Reference gcloud container clusters resize - resizes an existing cluster for running containers https://cloud.google.com/sdk/gcloud/reference/container/clusters/resize

129
Q

Your customer is receiving reports that their recently updated Google App Engine application is taking approximately 30 seconds to load for some of their users. This behavior was not reported before the update. What strategy should you take?
A. Work with your ISP to diagnose the problem.
B. Open a support ticket to ask for network capture and flow data to diagnose the problem, then roll back your application.
C. Roll back to an earlier known good release initially, then use Operations Suite (formerly Stackdriver) Trace and logging to diagnose the problem in a development/test/staging environment.
D. Roll back to an earlier known good release, then push the release again at a quieter period to investigate. Then use Operations Suite (formerly Stackdriver) Trace and logging to diagnose the problem.

A

Correct Answer C and D Explanation J2EE is Java, which can run on App Engine. He can also configure his application to run on a managed instance group for scaling, as long as he configures a data storage backend for the group as well.

130
Q

Your organization requires that metrics from all applications be retained for 5 years for future analysis in possible legal proceedings. Which approach should you use?
A. Configure Operations Suite (formerly Stackdriver) Monitoring for all Projects, and export to BigQuery.
B. Configure Operations Suite (formerly Stackdriver) Monitoring for all Projects with the default retention policies.
C. Configure Operations Suite (formerly Stackdriver) Monitoring for all Projects, and export to Google Cloud Storage.
D. Grant the security team access to the logs in each Project.

A

Correct Answer C Explanation
A - You ISP normally won’t help in this level. Also, the problem most likely is caused by recent update. The good approach is to rollback first and then investigate later. Similarly, this also apply to answer B. To investigate this kind of issue, use Operations Suite (formerly Stackdriver) Trace and logging to diagnose the bottleneck C and D have something in common for both “use Operations Suite (formerly Stackdriver) Trace and logging”, either in test/dev or in production environment and “Roll back to an earlier known good release”. At this moment, only the “earlier known good release” version starts receiving traffic. The difference lines between C’s “to diagnose the problem in a development/test/staging environment.” and D’s “then push the release again at a quieter period to investigate”. If you want to debug in production environments, “then push the release again at a quieter period to investigate” is not necessary - you can simply switch “default” version or split the traffic between the “earlier known good release” version and the new problem version. Essentially D’s “then push the release again at a quieter period to investigate” disqualifies itself as good answer - the default would be the new pushed version (the one with problem) starts receiving traffic “at a quieter period”, and the slow loading users may not present. But with answer C in development/test/staging environment, you can arbitrarily load those suffering users if you know them or simulate production load to reveal the problem users and then do further investigation. So, C is the correct answer: First, rollback to “the earlier known good release” and then use the test/dev/staging envs to investigate. Additional Resource https://cloud.google.com/appengine/docs/flexible/python/testing-and-deploying-your-app Testing and Deploying your Application

131
Q

You need to take streaming data from thousands of Internet of Things (IoT) devices, ingest it, run it through a processing pipeline, and store it for analysis. You want to run SQL queries against your data for analysis. What services in which order should you use for this task?
A. Cloud Dataflow, Cloud Pub/Sub, BigQuery
B. Cloud Pub/Sub, Cloud Dataflow, Cloud Dataproc
C. Cloud Pub/Sub, Cloud Dataflow, BigQuery
D. App Engine, Cloud Dataflow, BigQuery

A

Correct Answer C Explanation B and D can be quickly ruled out because none of them is solution for the requirements “retained for 5 years” Between A and C, the different is where to store, BigQuery or Cloud Storage. Since the main concern is extended storing period, C (Correct Answer) is better choice, and the “retained for 5 years for future analysis” further qualifies it, for example, using Archive storage class. With regards of BigQuery, while it is also a low-cost storage, but the main purpose is for analysis. Also, logs stored in Cloud Storage is easy to transport to BigQuery or do query directly against the files saved in Cloud Storage if and whenever needed. Additional Resource Overview of storage classes, price, and use cases https://cloud.google.com/storage/docs/storage-classes. Why export logs? https://cloud.google.com/logging/docs/export/ Operations Suite (formerly Stackdriver) Quotas and Limits for Monitoring https://cloud.google.com/monitoring/quotas The BigQuery pricing. https://cloud.google.com/bigquery/pricing

132
Q

Your company currently hosts an AWS S3 bucket. You need to keep the contents of this bucket in sync with a new Google Cloud Storage bucket to support a backup storage destination. What is the best method to achieve this?
A. Once per week, use a gsutil cp command to copy over newly modified files.
B. Use gsutil rsync commands to keep both locations in sync.
C. Use Storage Transfer Service to keep both the source and destination in sync.
D. Use gsutil -m cp to keep both locations in sync.

A

Correct Answer C Explanation C (Correct answer) - Cloud Pub/Sub, Cloud Dataflow, BigQuery Cloud Pub/Sub is a simple, reliable, scalable foundation for stream analytics and event-driven computing systems. As part of Google Cloud’s stream analytics solution, the service ingests event streams and delivers them to Cloud Dataflow for processing and BigQuery for analysis as a data warehousing solution. Relying on the Cloud Pub/Sub service for delivery of event data frees you to focus on transforming your business and data systems with applications such as: · check Real-time personalization in gaming · check Fast reporting, targeting and optimization in advertising and media · check Processing device data for healthcare, manufacturing, oil and gas, and logistics · check Syndicating market-related data streams for financial services Also, Use Cloud Dataflow as a convenient integration point to bring predictive analytics to fraud detection, real-time personalization and similar use cases by adding TensorFlow-based Cloud Machine Learning models and APIs to your data processing pipelines. https://cloud.google.com/ml-engine/ BigQuery provides a flexible, powerful foundation for Machine Learning and Artificial Intelligence. BigQuery provides integration with CloudML Engine and TensorFlow to train powerful models on structured data. Moreover, BigQuery’s ability to transform and analyze data helps you get your data in shape for Machine Learning. https://cloud.google.com/bigquery/ Other solutions may work one way or other but only the combination of theses 3 components integrate well in data ingestion, collection, and real-time analysis, and data mining in a highly durable, elastic, and parallel manner.
A - Wrong order. You don’t normally ingest IoT data directly to DataFlow
C - DataProc is GCP version of Apache Hadoop/Spark. Although it has the SQL-like Hive, it does not provide SQL interface as sophisticated as BigQuery does.
D - App Engine is compute resources. It is not designed to ingest IoT data like PubSub. Also. It’s rare use case App Engine ingests data to DataFlow directly. Below two pictures illustrate the typical toles played by DataFlow and PubSub Dataflow PubSub

133
Q

You want to optimize the performance of an accurate, real-time, weather-charting application. The data comes from 50,000 sensors sending 10 readings a second, in the format of a timestamp and sensor reading. Where should you store the data?
A. Google BigQuery
B. Google Cloud Storage
C. Google Cloud Bigtable
D. Google Cloud SQL

A

Correct Answer C Explanation You can use gsutil rsync to keep two locations in sync. Here are the rules of thumb when deciding whether to use gsutil or Storage Transfer Service: · When transferring data one time or have to do it rarely, use gsutil. · When transferring data continuously for long periods of time, use Storage Transfer Service. Otherwise, evaluate both tools with respect to your specific scenario. Since, here the requirement is to keep both the bucket in sync all the time it’s preferable to use Storage Transfer Service.

134
Q

Your company’s test suite is a custom C++ application that runs tests throughout each day on Linux virtual machines. The full test suite takes several hours to complete, running on a limited number of on premises servers reserved for testing. Your company wants to move the testing infrastructure to the cloud, to reduce the amount of time it takes to fully test a change to the system, while changing the tests as little as possible. Which cloud infrastructure should you recommend?
A. Google Cloud Dataproc to run Apache Hadoop jobs to process each test
B. Google App Engine with Google Operations Suite (formerly Stackdriver) for logging
C. Google Compute Engine managed instance groups with auto-scaling
D. Google Compute Engine unmanaged instance groups and Network Load Balancer

A

Correct Answer C
A - Google BigQuery: A scalable, fully-managed Enterprise Data Warehouse (EDW) with SQL and fast response times. It is for analytics and OLAP workload, though it also provides storage capacity and price similar to GCS and it cannot stand for this amount (50000 X 10 per second) of data streaming ingestion in real-time.
B - Google Cloud Storage: A scalable, fully-managed, highly reliable, and cost-efficient object / blob store. It cannot stand for this amount of data streaming ingestion rate in real-time C (Correct answer) - Google Cloud Bigtable: A scalable, fully-managed NoSQL wide-column database that is suitable for both real-time access and analytics workloads. https://cloud.google.com/storage-options/ Bigtable is Good for · Low-latency read/write access · High-throughput analytics · Native time series support For the following common work load IoT, finance, adtech Personalization, recommendations Monitoring Geospatial datasets Graphs
D - Google Cloud SQL: A fully-managed MySQL and PostgreSQL relational database service for Structured data and OLTP workloads. It also won’t stand for this type of high ingesting rate in real time

135
Q

You have been asked to select the storage system for the click-data of your company’s large portfolio of websites. This data is streamed in from a custom website analytics package at a typical rate of 6,000 clicks per minute, with bursts of up to 8,500 clicks per second. It must be stored for future analysis by your data science and user experience teams. Which storage infrastructure should you choose?
A. Google cloud Datastore
B. Google Cloud SQL
C. Google Cloud Bigtable
D. Google Cloud Storage

A

Correct Answer C
A - Google Cloud Dataproc to run Apache Hadoop jobs to process each test Apache Hadoop runs Java not C++; If the questions meant to use Hadoop to manage and process the test, it’s overkill and also needs significant changes to the testing infrastructure to integrate with Dataproc.
B - Google App Engine with Google Operations Suite (formerly Stackdriver) for logging App Engine did not natively support C++, also it’s probably hard to port their “runs tests throughout each day on Linux virtual machines” to App Engine “while changing the tests as little as possible”; Operations Suite (formerly Stackdriver) logging won’t help porting the test to GCP, either. Between C and D, the main difference is managed or unmanaged instance group Unmanaged instance groups are groups of dissimilar instances that you can arbitrarily add and remove from the group. Unmanaged instance groups do not offer autoscaling, rolling update support, or the use of instance templates so Google recommends creating managed instance groups whenever possible. Use unmanaged instance groups only if you need to apply load balancing to your pre-existing configurations or to groups of dissimilar instances. https://cloud.google.com/compute/docs/instance-groups/ From the question, there is no such requirement for unmanaged instance group and not mention that dissimilar Linux machine types are required. In addition, judging from what they suffered “The full test suite takes several hours to complete, running on a limited number of on-premises servers”, it seems they simply need more computation power - bigger and/or more instances for the testing. So the managed instance group with autoscaling is preferred.

136
Q

You work in a small company where everyone should be able to view all resources of a specific project. You want to grant them access following Google’s recommended practices. What should you do?
A. Create a script that uses “gcloud projects add-iam-policy-binding” for all users’ email addresses and the Project Viewer role.
B. A. Create a script that uses “gcloud iam roles create” for all users’ email addresses and the Project Viewer role.
C. Create a new Google Group and add all users to the group. Use “gcloud projects add-iam-policy-binding” with the Project Viewer role and Group email address.
D. Create a new Google Group and add all members to the group. Use “gcloud iam roles create” with the Project Viewer role and Group email address.

A

Correct Answer C
A - Google cloud Datastore. Doesn’t meet this requirement “It must be stored for future analysis by your data science and user experience teams.” Google Cloud Datastore is a NoSQL document database built for automatic scaling, high performance, and ease of application development and integrating well with App Engine. Datastore: A scalable, fully-managed NoSQL document database for your web and mobile applications. Good for: Semi-structured application data Hierarchical data Durable key-value data Workload: User profiles Product catalogs Game state
B - Google Cloud SQL. Cloud SQL is mainly for OLTP (Transactional, CRUD) not for taking and storing streaming data. It does not have the scalability and elasticity to absorb this amount of data in real time. C (Correct Answer) - Google Cloud Bigtable. The reason is that data is in IoT nature and it will be used for analytics. Bigtable: A scalable, fully-managed NoSQL wide-column database that is suitable for both real-time access and analytics workloads. Bigtable is ideal for very large NoSQL datasets and is useful for high-speed transactions and analysis. It integrates well with ML. Dataproc, and analytics Good for Low-latency read/write access High-throughput analytics Native time series support Work load IoT, finance, adtech Personalization, recommendations Monitoring Geospatial datasets Graphs Although both Datastore and Bigtable are NoSQL databases, Bigtable is able to support over a petabyte of data and is useful for high speed analytics as well, whereas Datastore is not.
D - Google Cloud Storage. GCS is ideally for Object storage purpose although it has pretty good scalability. It’s not suitable for IoT kind of spiky streaming data. Its buckets initially support roughly 1000 writes per second and then scale as needed. As the request rate for a given bucket grows, Cloud Storage automatically increases the IO capacity for that bucket by distributing the request load across multiple servers. Especially considering the click stream rate of 6,000 clicks per minute, with bursts of up to 8,500 clicks per second, the way GCS handle and absorb this kind high and low data stream by scale up and down make it not suitable for this task.

137
Q

You have a definition for an instance template that contains a web application. You are asked to deploy the application so that it can scale based on the HTTP traffic it receives. What should you do?
A. Create a VM from the instance template. Create a custom image from the VM’s disk. Export the image to Cloud Storage. Create an HTTP load balancer and add the Cloud Storage bucket as its backend service.
B. Create a VM from the instance template. Create an App Engine application in Automatic Scaling mode that forwards all traffic to the VM.
C. Create a managed instance group based on the instance template. Configure autoscaling based on HTTP traffic and configure the instance group as the backend service of an HTTP load balancer.
D. Create the necessary amount of instances required for peak user traffic based on the instance template. Create an unmanaged instance group and add the instances to that instance group. Configure the instance group as the Backend Service of an HTTP load balancer.

A

Correct answer C A is not correct because groups are recommended over individual assignments. B is not correct because this command is to create roles, not to assign them. C is correct because Google recommends to use groups where possible. D is not correct because this command is to create roles, not to assign them. Reference gcloud iam https://cloud.google.com/sdk/gcloud/reference/iam/

138
Q

Your infrastructure runs on another cloud and includes a set of multi-TB enterprise databases that are backed up nightly both on premises and also to that cloud. You need to create a redundant backup to Google Cloud. You are responsible for performing scheduled monthly disaster recovery drills. You want to create a cost-effective solution. What should you do?
A. Use Transfer Appliance to transfer the offsite backup files to a Cloud Storage Nearline storage bucket as a final destination.
B. Use Transfer Appliance to transfer the offsite backup files to a Cloud Storage Coldline bucket as a final destination.
C. Use Storage Transfer Service to transfer the offsite backup files to a Cloud Storage Nearline storage bucket as a final destination.
D. Use Storage Transfer Service to transfer the offsite backup files to a Cloud Storage Coldline storage bucket as a final destination.

A

Correct answer C A Is not correct because the Load Balancer will just load balance access to the uploaded image itself, and not create or autoscale VMs based on that image. B Is not correct because while the App Engine can scale as a proxy, all requests will still end up on the same Compute Engine instance, which needs to scale itself. C is correct because a managed instance group can use an instance template to scale based on HTTP traffic. D is not correct because unmanaged instance groups do not offer autoscaling. Reference Managed instance groups and autoscaling https://cloud.google.com/compute/docs/instance-groups/#managed_instance_groups_and_autoscaling Exporting an Image https://cloud.google.com/compute/docs/images/export-image Adding a Cloud Storage Bucket to Content-based Load Balancing https://cloud.google.com/compute/docs/load-balancing/http/adding-a-backend-bucket-to-content-based-load-balancing

139
Q

One of your primary business objectives is being able to trust the data stored in your application. You want to log all changes to the application data. How can you design your logging system to verify the authenticity of your logs?
A. Create a JSON dump of each log entry and store it in Google Cloud Storage.
B. Write the log concurrently in the cloud and on premises.
C. Digitally sign each timestamp and log entry and store the signature.
D. Use an SQL database and limit who can modify the log table.

A

Correct answer C C (Correct Answer) - C is correct because you will need to access your backup data monthly to test your disaster recovery process, so you should use a Nearline bucket; also, because you will be performing ongoing, regular data transfers, so you should use the storage transfer service. A, B, and D are not correct because you should not use Coldline if you want to access the files monthly (B, D) and you should not use Transfer Appliance for repeated data transfers (A, B). Reference GCS Nearline for once-per-month access https://cloud.google.com/storage/docs/storage-classes#nearline Storage Transfer Service Documentation https://cloud.google.com/storage-transfer/docs/

140
Q

You set up an autoscaling instance group to serve web traffic for an upcoming launch. After configuring the instance group as a backend service to an HTTP(S) load balancer, you notice that virtual machine (VM) instances are being terminated and re-launched every minute. The instances do not have a public IP address. You have verified the appropriate web response is coming from each instance using the curl command. You want to ensure the backend is configured correctly. What should you do?
A. Ensure that a firewall rule exists to allow source traffic on HTTP/HTTPS to reach the load balancer.
B. Create a tag on each instance with the name of the load balancer. Configure a firewall rule with the name of the load balancer as the source and the instance tag as the destination.
C. Ensure that a firewall rule exists to allow load balancer health checks to reach the instances in the instance group.
D. Assign a public IP to each instance and configure a firewall rule to allow the load balancer to reach the instance public IP.

A

Correct Answer C C (Correct answer) - Digitally sign each timestamp and log entry and store the signature. Answer A, B, and D don’t have any added value to verify the authenticity of your logs. Besides, Logs are mostly suitable for exporting to Cloud storage, BigQuery, and PubSub. SQL database is not the best way to be exported to nor store log data. Simplified Explanation To verify the authenticity of your logs if they are tampered with or forged, you can use a certain algorithm to generate digest by hashing each timestamp or log entry and then digitally sign the digest with a private key to generate a signature. Anybody with your public key can verify that signature to confirm that it was made with your private key and they can tell if the timestamp or log entry was modified. You can put the signature files into a folder separate from the log files. This separation enables you to enforce granular security policies. Ref URL: https://cloud.google.com/logging/docs/reference/tools/gcloud-logging

141
Q

Your company wants to reduce cost on infrequently accessed data by moving it to the cloud. The data will still be accessed approximately once a month to refresh historical charts. In addition, data older than 5 years is no longer needed. How should you store and manage the data?
A. In Google Cloud Storage and stored in a Multi-Regional bucket. Set an Object Lifecycle Management policy to delete data older than 5 years.
B. In Google Cloud Storage and stored in a Multi-Regional bucket. Set an Object Lifecycle Management policy to change the storage class to Coldline for data older than 5 years.
C. In Google Cloud Storage and stored in a Nearline bucket. Set an Object Lifecycle Management policy to delete data older than 5 years.
D. In Google Cloud Storage and stored in a Nearline bucket. Set an Object Lifecycle Management policy to change the storage class to Coldline for data older than 5 years.

A

Correct Answer C C (correct answer) - Ensure that a firewall rule exists to allow load balancer health checks to reach the instances in the instance group. HTTP health check probes are sent from the IP ranges depending on LB types used. These are IP address ranges that the load balancer uses to connect to backend instances. You must create firewall rules that allows traffic from those ranges to reach your instances For Network load balancing When a health check is used with Network load balancing, the health check probes come from addresses in the ranges 209.85.152.0/22, 209.85.204.0/22, and 35.191.0.0/16. For HTTP(S). SSL proxy. TCP proxy, and Internal load balancing When a health check is used with HTTP(S), SSL proxy, TCP proxy, or Internal load balancing, the health check probes come from addresses in the ranges 130.211.0.0/22 and 35.191.0.0/16.
A - Ensure that a firewall rule exists to allow source traffic on HTTP/HTTPS to reach the load balancer. Firewall controls access at instance level, not load balancer. Must allow load balancer traffic to connect backend instance allowing health check
B - Create a tag on each instance with the name of the load balancer. Configure a firewall rule with the name of the load balancer as the source and the instance tag as the destination. At this moment it is not possible to set firewall rules over the GCE Load Balancers. You need to create firewall rules that at subnet or instances level allowing specific health check IP ranges (See Answer A above), not the LB tags, to connect to all your load balanced instances.
D - Assign a public IP to each instance and configure a firewall rule to allow the load balancer to reach the instance public IP. This is not mandatory since your LB could be Internal load balancing so instances’ external IPs may be removed

142
Q

A lead software engineer tells you that his new application design uses websockets and HTTP sessions that are not distributed across the web servers. You want to help him ensure his application will run properly on Google Cloud Platform. What should you do?
A. Help the engineer to convert his websocket code to use HTTP streaming.
B. Review the encryption requirements for websocket connections with the security team.
C. Meet with the cloud operations team and the engineer to discuss load balancer options.
D. Help the engineer redesign the application to use a distributed user session service that does not rely on websockets and HTTP sessions.

A

Correct Answer C C (Correct Answer) - The access pattern fits Nearline storage class requirements and Nearline is a more cost-effective storage approach than Multi-Regional. The object lifecycle management policy to delete data is correct versus changing the storage class to Coldline. A and
B - For the requirement: “accessed approximately once a month” A and B can be quickly eliminated due to the incorrect Multi-Regional storage class vs Nearline storage class
D - Need deleting them, not changing the storage class to Coldline.

143
Q

What is the command for creating a storage bucket that has once per month access and is named ‘archive_bucket’?
A. gsutil rm -coldline gs://archive_bucket
B. gsutil mb -c coldline gs://archive_bucket
C. gsutil mb -c nearline gs://archive_bucket
D. gsutil mb gs://archive_bucket

A

Correct Answer C C (Correct Answer) - The HTTP(S) load balancer in GCP handles websocket traffic natively. Backends that use WebSocket to communicate with clients can use the HTTP(S) load balancer as a front end, for scale and availability.
A - There is no compelling reason to move away from websockets as part of a move to GCP.
B - This may be a good exercise anyway, but it doesn’t really have any bearing on the GCP migration.
D - There is no compelling reason to move away from websockets as part of a move to GCP.

144
Q

The security team has disabled external SSH access into production virtual machines in GCP. The operations team needs to remotely manage the VMs and other resources. What can they do?
A. Develop a new access request process that grants temporary SSH access to cloud VMs when an operations engineer needs to perform a task.
B. Grant the operations team access to use Google Cloud Shell.
C. Have the development team build an API service that allows the operations team to execute specific remote procedure calls to accomplish their tasks.
D. Configure a VPN connection to GCP to allow SSH access to the cloud VMs.

A

Correct answer C mb is to make the bucket. Nearline buckets are for once per month access. Coldline buckets require only accessing once per 90 days and would incur additional charges for greater access. Further Explanation Synopsis gsutil mb [-c class] [-l location] [-p proj_id] url… If you don’t specify a -c option, the bucket is created with the default storage class Standard Storage, which is equivalent to Multi-Regional Storage or Regional Storage, depending on whether the bucket was created in a multi-regional location or regional location, respectively. If you don’t specify a -l option, the bucket is created in the default location (US). -l option can be any multi-regional or regional location. Reference mb - Make buckets: https://cloud.google.com/storage/docs/gsutil/commands/mb

145
Q

For this question, refer to the TerramEarth case study. TerramEarth has equipped unconnected trucks with servers and sensors to collect telemetry data. Next year they want to use the data to train machine learning models. They want to store this data in the cloud while reducing costs. What should they do?
A. Have the vehicle’ computer compress the data in hourly snapshots and store it in a Google Cloud storage (GCS) Nearline bucket.
B. Push the telemetry data in Real-time to a streaming dataflow job that compresses the data and store it in Google BigQuery.
C. Push the telemetry data in real-time to a streaming dataflow job that compresses the data and store it in Cloud Bigtable.
D. Have the vehicle’s computer compress the data in hourly snapshots, and store it in a GCS Coldline bucket.

A

Correct Answer D
Option
D - Configure a VPN connection to GCP to allow SSH access to the cloud VMs. The questions tell that the “blocking” happens on GCP, especially in production environments. That means that there are firewall rules preventing access from public IPs on port 22. Therefore, using a VPN and configuring a firewall that allows TCP connections from RFC1918 on port 22, would work best. In this case, answer D is better.
Option
B - SSH access will not allow access if Port 22 is blocked. Options A and C are possible options that might require more setup than worthwhile for the needs.

146
Q

You’re writing a Python application and want your application to run in a sandboxed managed environment with the ability to scale up in seconds to account for huge spikes in demand. Which service should you host your application on?
A. Compute Engine
B. App Engine Flexible Environment
C. Kubernetes Engine
D. App Engine Standard Environment

A

Correct Answer D D (Correct answer) - Have the vehicle’s computer compress the data in hourly snapshots, a Store it in a GCS Coldline bucket. This is the Lowest cost for storage for infrequent access that meets the requirement (“next year …”). There is no good reason to use nearline instead of low-cost storage Coldline for one-year-after access.
A - Have the vehicle’s computer compresses the data in hourly snapshots and store it in a Google Cloud Storage (GCS) Nearline bucket. Nearline does not fit the usage pattern described in the question. Nearline fits this usage pattern: For example, if you want to continuously add files to cloud storage, and plan to access those files once a month for analysis, nearline storage is a great choice B and C can be eliminated for this reason “Push the telemetry data in real-time to a streaming dataflow job …” since vehicles are unconnected Reference Resource Comparison of storage classes https://cloud.google.com/storage/docs/storage-classes

147
Q

For this question, refer to the TerramEarth case study: Based on TerramEarth’s current data flow environment, what are the direct GCP services needed to replicate the same structure for batch uploads?
A. Cloud Spanner - Cloud SQL - BigQuery
B. Cloud Dataflow - Cloud Bigtable - Cloud Dataproc
C. Cloud Dataproc - Cloud Storage - BigQuery
D. Cloud Storage - Cloud Dataflow - BigQuery

A

Correct Answer D Explanation App Engine Standard environment intend to · Language: Python, Java, Node.js, PHP, and Go · Experiences sudden and extreme spikes of traffic which require immediate scaling. · Instance startup time in seconds Reference Choosing an App Engine Environment https://cloud.google.com/appengine/docs/the-appengine-environments

148
Q

For this question, refer to the Mountkirk Games case study Mountkirk Games needs to create a repeatable and configurable mechanism for deploying isolated application environments. Developers and testers can access each other’s environments and resources, but they cannot access staging or production resources. The staging environment needs access to some services from production. What should you do to isolate development environments from staging and production?
A. Create a project for development and test and another for staging and production.
B. Create a network for development and test and another for staging and production.
C. Create one subnetwork for development and another for staging and production.
D. Create one project for development and test, a second for staging and a third for production.

A

Correct answer D Explanation Based on their current batch upload model, the direct equivalent would be to use Cloud Storage for storing files, Dataflow for their ETL processing, and BigQuery for their data warehouse needs. Below illustrates the solution concept. TerramEarth’s Existing Technical Environment One Possible GCP solution for batch upload flow

149
Q

For this question, refer to the MountKirk Games case study: MountKirk Games needs to build out their streaming data analytics pipeline to feed from their game backend application. What GCP services in which order will achieve this?
A. Cloud Storage - Cloud Dataflow - BigQuery
B. Cloud Dataproc - Cloud Storage - BigQuery
C. Cloud Pub/Sub - Cloud Dataflow - Cloud Bigtable
D. Cloud Pub/Sub - Cloud Dataflow - BigQuery

A

Correct Answer D Explanation D (Correct answer) - among the available answers, D is the closest solution to meet the isolate and inter-access requirements. In this example, you’ll create one Host project for Developer and Tester and another Host project for staging, and the third one for production. Staging and Production environments can access resources as per the cross accessible Service Accounts created for just the required needs.
B - This is incomplete and not the best solution. Network isolation is for separation of resources communication, the project is for IAM resource access control. If the question meant putting resources in different networks but in the same project, it’s not enough to separate developer from access Stage/Product unless access policy is set at each specific resource level which is not only against the best practice but also hard to manage especially if you consider Mountkirk Games is not a small shop. Answer C is incorrect for the same or similar reasons. Answer A indeed enables the isolation but sharing staging and Production in the same project might have some cross access of the resources by human error. On the other hand, if the quest meant Developer and tester are in the same group called Development (based on “What should you do to isolate development environments from staging and production?”), D could be an answer since it isolates the development from staging and production, though no inter-project access issues addressed. Overall, judging from Mountkirk Games application, environments, and company size, most likely they’ll have separated Development and Testing while they do share access to some resources such as access testing data as well as computing resources. So, Answer option D is closer to the requirements

150
Q

Your company has decided to make a major revision of their API in order to create better experiences for their developers. They need to keep the old version of the API available and deployable, while allowing new customers and testers to try out the new API. They want to keep the same SSL and DNS records in place to serve both APIs. What should they do?
A. Configure a new load balancer for the new version of the API.
B. Reconfigure old clients to use a new endpoint for the new API.
C. Have the old API forward traffic to the new API based on the path.
D. Use separate backend pools for each API path behind the load balancer.

A

Correct answer D Explanation Pub/Sub is kind of ‘shock absorber’, allowing asynchronous messaging between large numbers of devices. Cloud Dataflow acts as your data processing pipeline for ETL functions on both streaming and batch data. BigQuery is a data warehouse, able to run analysis on petabytes of data using SQL queries. Below is a reference architect Google recommending for similar scenario in Real-time streaming data collection and analysis https://cloud.google.com/solutions/mobile/mobile-gaming-analysis-telemetry Real-time processing of events from game clients and game servers Data Transformation with Cloud Dataflow - Dataflow acts as your data processing pipeline for ETL functions on both streaming and batch data.

151
Q

You created an update for your application on App Engine. You want to deploy the update without impacting your users. You want to be able to roll back as quickly as possible if it fails. What should you do?
A. Delete the current version of your application. Deploy the update using the same version identifier as the deleted version.
B. Notify your users of an upcoming maintenance window. Deploy the update in that maintenance window.
C. Deploy the update as the same version that is currently running.
D. Deploy the update as a new version. Migrate traffic from the current version to the new version.

A

Correct Answer D
A - Configure a new load balancer for the new version of the API. Then you’ll have two load balancers: new and old one. This would break the API contract and also need to configure SSL and DNS for new load balancer
B - Reconfigure old clients to use a new endpoint for the new API. This would break the contract between client and API and also can no longer reach to old API.
C - Have the old API forward traffic to the new API based on the path. This is against the design principle and best practice for webservices API. API should not play this role - API should not know how to route client requests. In fact, API even shouldn’t have knowledge about client at all. D (Correct answer) - Use separate backend pools for each API path behind the load balancer. This solution meets both requirements: continue servicing old and new API and keep the same SSL and DNS records Further Explanation HTTP(S) Load Balancing configuration and it backend service: A configured backend service contains one or more backends. Here is demonstrating configuration for the concept similar to the question scenarios: you just need to configure URL mapping pointing to the new and old APIs which are hosted in corresponding backends - this implementation meets all the requirement: No contract breaking; clients continue to access new and old API without reconfigure SSL and DNS References Setting Up HTTP(S) Load Balancing https://cloud.google.com/compute/docs/load-balancing/http/ Backend service components: https://cloud.google.com/compute/docs/load-balancing/http/backend-service Creating Content-Based Load Balancing https://cloud.google.com/load-balancing/docs/https/content-based-example

152
Q

You are running an application in Google App Engine that is serving production traffic. You want to deploy a risky but necessary change to the application. It could take down your service if not properly coded. During development of the application, you realized that it can only be properly tested by live user traffic. How should you test the feature?
A. Deploy the new application version temporarily, and then roll it back.
B. Create a second project with the new app in isolation, and onboard users.
C. Set up a second Google App Engine service, and then update a subset of clients to hit the new service.
D. Deploy a new version of the application and use traffic splitting to send a small percentage of traffic to it.

A

Correct answer D A and B are not correct because this will make the application temporarily unavailable to users. C is not correct because to roll back, you’ll need to redeploy the previous deployment because the app was overwritten with the same version number. Therefore this takes longer than a rollback using method D. D is correct because this makes sure there is no downtime and you can roll back the fastest. Reference Migrating and Splitting Traffic https://cloud.google.com/appengine/docs/admin-api/migrating-splitting-traffic

153
Q

One of the microservices in your application has an intermittent performance problem. You have not observed the problem when it occurs but when it does, it triggers a particular burst of log lines. You want to debug a machine while the problem is occurring. What should you do?
A. Log into one of the machines running the microservice and wait for the log storm.
B. In the Operations Suite (formerly Stackdriver) Error Reporting dashboard, look for a pattern in the times the problem occurs.
C. Configure your microservice to send traces to Operations Suite (formerly Stackdriver) Trace so you can find what is taking so long.
D. Set up a log metric in Operations Suite (formerly Stackdriver) Logging, and then set up an alert to notify you when the number of log lines increases past a threshold.

A

Correct Answer D D (Correct Answer) - Deploying a new version without assigning it as the default version will not create downtime for the application. Using traffic splitting allows for easily redirecting a small amount of traffic to the new version and can also be quickly reverted without application downtime.
A - Deploying the application version as default requires moving all traffic to the new version. This could impact all users and disable the service.
B - Deploying a second project requires data synchronization and having an external traffic splitting solution to direct traffic to the new application. While this is possible, with Google App Engine, these manual steps are not required.
C - App Engine services are intended for hosting different service logic. Using different services would require manual configuration of the consumers of services to be aware of the deployment process and manage from the consumer side who is accessing which service.

154
Q

Your marketing department wants to send out a promotional email campaign. The development team wants to minimize direct operation management. They project a wide range of possible customer responses, from 100 to 500,000 click-throughs per day. The link leads to a simple website that explains the promotion and collects user information and preferences. Which infrastructure should you recommend?
A. Use a single compute Engine virtual machine (VM) to host a web server, backed by Google Cloud SQL.
B. Use a Google Container Engine cluster to serve the website and store data to persistent disk.
C. Use a managed instance group to serve the website and Google Cloud Bigtable to store user data.
D. Use Google App Engine to serve the website and Google Cloud Datastore to store user data.

A

Correct Answer D D (Correct Answer) - Since you know that there is a burst of log lines you can set up a metric that identifies those lines. Operations Suite (formerly Stackdriver) will also allow you to set up a text, email or messaging alert that can notify promptly when the error is detected so you can hop onto the system to debug.
A - Logging into an individual machine may not see the specific performance problem as multiple machines may be in the configuration and reducing the chances of interacting with an intermittent performance problem.
B - Error reporting won’t necessarily catch the log lines unless they are stack traces in the proper format. Additionally, just because there is a pattern doesn’t mean you will know exactly when and where to log in to debug.
C - Trace may tell you where time is being spent but won’t let you know in on the exact host that the problem is occurring on because you generally only send samples of traces. There is also no alerting on traces to notify exactly when the problem is happening. Additional Resource https://cloud.google.com/logging/docs/logs-based-metrics/

155
Q

Your customer is moving an existing corporate application from an on-premises data center to Google Cloud Platform . The business owner requires minimal user disruption. There are strict security team requirements for storing passwords. What authentication strategy should they use?
A. Use G Suite Password Sync to replicate passwords into Google.
B. Ask users to set their Google password to match their corporate password.
C. Provision users in Google using the Google Cloud Directory Sync tool.
D. Federate authentication via SAML 2.0 to the existing Identity Provider.

A

Correct Answer D D (Correct answer) - Use Google App Engine to serve the website and Google Cloud Datastore to store user data - This solution meets all the requirements and has all the element of App Engine features App Engine Standard environment intend to · Multiple Languages supported; · Experiences sudden and extreme spikes of traffic which require immediate scaling; · Instance startup time in seconds Google Cloud Datastore is a NoSQL document database built for automatic scaling, high performance, and ease of application development and integrating well with AppEngine Datastore: A scalable, fully-managed NoSQL document database for your web and mobile applications. Good for: Semi-structured application data Hierarchical data Durable key-value data Workload: User profiles Product catalogs Game state The combination of App Engine and DataStore make answer D a clear choice.
A - Use a single compute Engine virtual machine (VM) to host a web server, backed by Google Cloud SQL - Not as good as App Engine solution for “minimize direct operation management” and single VM may not scalable enough to handle the traffic spikes from 100 to 500,000 click throughs per day.
B - Use a Google Container Engine cluster to serve the website and store data to persistent disk - Slightly better than A but still not as good as App Engine solution for “minimize direct operation management”. Also store huge amount of user information and preferences data to disk is improper in term of programming model, efficiency, cost, and scalability
C - Use a managed instance group to serve the website and Google Cloud Bigtable to store user data - this would overkill for this kind of simple application in term of programming model, efficiency, and cost. It deviates further from the requirements “to minimize direct operation management.” Additional Resource Choosing an App Engine Environment https://cloud.google.com/appengine/docs/the-appengine-environments Build highly scalable applications on a fully managed serverless platform https://cloud.google.com/appengine/ Cloud Datastore Overview, What it’s good for: https://cloud.google.com/appengine/docs/standard/python/datastore/

156
Q

You have a managed instance group comprised of preemptible VM’s. All of the VM’s keep deleting and recreating themselves every minute. What is a possible cause of this behavior?
A. Your zonal capacity is limited, causing all preemptible VM’s to be shutdown to recover capacity. Try deploying your group to another zone.
B. You have hit your instance quota for the region.
C. Your managed instance group’s VM’s are toggled to only last 1 minute in preemptible settings.
D. Your managed instance group’s health check is repeatedly failing, either to a misconfigured health check or misconfigured firewall rules not allowing the health check to access the instances.

A

Correct Answer D D is Correct answer - Federate authentication via SAML 2.0 to the existing Identity Provider. This meets both “minimal user disruption” and “strict security team requirements for storing passwords” User’s passwords are stored on-premise, authentication happens on premise, there is no user disruption, on successful authentication, access token is shared to access application or GCP services.
Option
A - Use G Suite Password Sync to replicate passwords into Google - This is a violation against “strict security team requirements for storing passwords” https://support.google.com/a/answer/2611859?hl=en
Option
B - Ask users to set their Google password to match their corporate password - this violate “minimal user disruption” and “strict security team requirements for storing passwords”
Option
C - Provision users in Google using the Google Cloud Directory Sync tool. With google cloud directory sync, only the SHA-1 and MD5 unsalted passwords gets synced from source. Plus this may break the strict password requirement. Your credential details are now stored at 2 places.

157
Q

Your company wants to track whether someone is present in a meeting room reserved for a scheduled meeting. There are 1000 meeting rooms across 5 offices on 3 continents. Each room is equipped with a motion sensor that reports its status every second. The data from the motion detector includes only a sensor ID and several different discrete items of information. Analysts will use this data, together with information about account owners and office locations. Which database type should you use?
A. Relational database
B. Blobstore
C. Flat file
D. NoSQLdatabase

A

Correct Answer D In this case, the preemptible VM’s are not part of the problem. Likely, the cause is that the health check for the instance group is not receiving a success reply from the VM’s, causing it to recreate the VM’s over and over. This is due to either the health check not being correctly configured or the firewall for the instance group not allowing traffic from the load balancer/health check.

158
Q

The database administration team has asked you to help them improve the performance of their new database server running on Google Compute Engine. The database is for importing and normalizing their performance statistics and is built with MySQL running on Debian Linux. They have an n1-standard-8 virtual machine with 80 GB of SSD persistent disk. What should they change to get better performance from this system?
A. Create a new virtual machine running PostgreSQL.
B. Migrate their performance metrics warehouse to BigQuery.
C. Increase the virtual machine’s memory to 64 GB.
D. Modify all of their batch jobs to use bulk inserts into the database.
E. Dynamically resize the SSD persistent disk to 500 GB.

A

Correct Answer D The 3 characters: 1000 rooms; a sensor that reports its status every second; the data includes only a sensor ID and several different discrete items are a clear indication for a BigTable (a NoSQL database) usage pattern. Google Cloud Bigtable: A scalable, fully-managed NoSQL wide-column database that is suitable for both real-time access and analytics workloads. https://cloud.google.com/storage-options/ Bigtable is Good for: · Low-latency read/write access · High-throughput analytics · Native time series support For the following common workload: IoT, finance, adtech Personalization, recommendations Monitoring Geospatial datasets Graphs Bigtable single value in each row is indexed; Cloud Bigtable tables are sparse; if a cell does not contain any data, it does not take up any space, which satisfies for storing this type of data: “the data includes only a sensor ID and several different discrete items” Reference: Overview of Cloud Bigtable https://cloud.google.com/bigtable/docs/overview

159
Q

If you have object versioning enabled on a multi-regional bucket, what will the following lifecycle config file do?
A. Archive objects older than 30 days (the second rule doesn’t do anything)
B. Delete objects older than 30 days (the second rule doesn’t do anything)
C. Archive objects older than 30 days and move objects to Coldline Storage after 365 days
D. Delete all the versions that are not live and 30 days old. Move theremaining current versionsto Coldline after 365 days.

A

Correct Answer E All of other answers are either not applicable or not specified by the question scenarios. The following console screenshot show the effect to IO performance by changing memory or disk - it’s self-explained (please enlarge to see the details). Answer E is a clear straightforward winner over answer C as well as others. Increase disk size to 500GB take significant effect on IO performance than original configuration and higher memory configuration. Taking IOPS for read as example, all instances are configure with 8vCPU: 80G DISK/30G MEM = 2400 IOPS; 500G DISK/30G MEM = 15000 IOPS; 80G Disk/52G MEM=2400 IOPS And here is how to dynamically resize disk: https://cloud.google.com/compute/docs/disks/add-persistent-disk Adding or Resizing Persistent Disks

160
Q

You are using DataFlow to ingest a large amount of data and later you send the data to Bigquery for Analysis, but you realize the data is dirty and cannot be cleaned using predefined models, what would be the best choice to be used to clean the data in the stream with serverless approach?
A. Fetchdata from Bigquery and clean data from DataPrep and send it back to Bigquery.
B. Fetchdata from Bigquery and create one more pipeline, clean data using DataFlow and send it back to Bigquery.
C. Fetch the data from Bigquery andclean data from DataProc and send it back to Bigquery.
D. Send thedata to Data Storage and use BigTable to clean the data.

A

Correct Answer is D With formatted json, it’ll be easy to understand without lots of knowledge about Storage lifecycle syntax. It says (note the line number in [ ]): [1] Create lifecycle [2] rule to take [5] action of [6] delete [8] if the object is [9] 30 days old and [10] the version isLive: false (non-current version) [12] And [14] take action of [15] SetStorageClass to [16] COLDLINE [18] if the object is [19] 365 days old and [20] if its StorageClass matches [21] MULTI-REGIONAL The Correct Answer is D: Delete all the versions that are not current and 30 days old. Move the remaining current versions to Coldline after 365 days. Note: Here there are 2 rules 1st rule has 2 condition a. isLive: false => this means object should be declared as non-current verison b. age 30: 30 days old Then action will run “action”:{ “type”: “Delete”}, In another way: In this question there are 2 rules Rule 1 is with 2 condition a. 30 days b. isLive: false if these 2 condition meets: the object will be have action. i.e. delete Here IsLive means object need to be tagged as non-current which means versioning is enabled for that bucket. Rule 2 also have 2 condition a. 365 days b. MatchesStorageClass: Multi-Regional This means if Object is in Multi-Regional for over 365 days then the action will trigger Move the object to StorageClass:ColdLine Hence our suggested answer option D is correct. For a deep dive, please find time to study at https://cloud.google.com/solutions/data-lifecycle-cloud-platform

161
Q

You are working with a custom VPC and network. You need to allow traffic from specific virtual machines in ‘subnet-a’ network access to machines in ‘subnet-b’ without giving the entirety of subnet-a access. How can you accomplish this?
A. Create a firewall rule to allow traffic from resources with specific network tags, then assign the specific machines in subnet-a the same tags.
B. Relocate the subnet-a machines to a different subnet and give the new subnet the needed access.
C. Create a rule to deny all traffic to the entire subnet, then create a second rule with higher priority giving access to tagged VM’s in subnet-a.
D. You can only grant firewall access to an entire subnet and not individual VM’s inside.

A

Correct Answer:
Option B
Option B is CORRECT Dataflow is a good choice if data has no implementation with Spark or Hadoop and does not run on clusters, instead, it is based on parallel data processing. As such data is split processed on multiple microprocessors to reduce processing time.
Option A is INCORRECT. Dataprep uses Dataflow/ BigQuery under the hood, enabling to process structured or unstructured datasets of any size with the ease of clicks. Dataprep optimizes the job execution leveraging Trifacta’s in-memory processing engine for small datasets, BigQuery SQL pushdown (ELT) when the data is already in BigQuery/Dataflow by parallel data processing framework for large distributed datasets. Dataprep uses predefined models to clean the data, which is not valid in this scenario.
Option C is INCORRECT because Dataproc is designed to run on clusters which makes it compatible with Apache Hadoop, hive, and spark. It is significantly faster at creating clusters and can autoscale clusters without interruption of the running jobs.
Option D is INCORRECT because Cloud Bigtable is Google’s NoSQL Big Data database service, it is not used to clean data.

162
Q

You need to create a new Kubernetes Cluster on Google Cloud Platform that can autoscale the number of worker nodes. What should you do?
A. Create a cluster on Kubernetes Engine and enable autoscaling during cluster creation.
B. Create a cluster on Kubernetes Engine and enable autoscaling on the instance group of the cluster.
C. Configure a Compute Engine instance as a worker and add it to an unmanaged instance group. Add a load balancer to the instance group and rely on the load balancer to create additional Compute Engine instances when needed.
D. Create Compute Engine instances for the workers and the master and install Kubernetes. Rely on Kubernetes to create additional Compute Engine instances when needed.

A

Correct Answer: A
A - Create a firewall rule to allow traffic from resources with specific network tags, then assign the specific machines in subnet-a the same tags. Network tags allow more granular access based on individually tagged instances - Instances by target tags: The firewall rule is applicable only to VMs if they have a matching network tag. It allows specific VMs in the subnet-a to reach the VMs in subnet-b.
B - Relocate the subnet-a machines to a different subnet and give the new subnet the needed access. This would give the entire subnet access which is against the requirements: allow traffic from specific virtual machines in ‘subnet-a’ network access to machines in ‘subnet-b’ without giving the entirety of subnet-a access.
C - Create a rule to deny all traffic to the entire subnet, then create a second rule with higher priority giving access to tagged VM’s in subnet-a. Every custom VPC by default has a firewall rule that denies network traffic between subnet.
D - You can only grant firewall access to an entire subnet and not individual VM’s inside.

163
Q

Once a month Terram Earth’s vehicles are serviced and the data is downloaded from the maintenance port. The data analysts would want to query the large amount of data collected from these vehicles and analyze the overall condition of the vehicles. Terram Earth’s management is looking at a solution that is cost-effective and would scale for future requirements. Please select the right choice based on the requirement.
A. Load the data from Cloud Storage to Bigquery and Run queries based on date using an appropriate filter on DATE for the data stored in Bigquery based on the date partitioned table.
B. Store the data in Bigtable and run queries on it.
C. Load the data from Cloud Storage to Bigquery and run queries on Bigquery.
D. Run queries against the data stored in Cloud Spanner.

A

Correct answer: A
Option A is correct because this is the way to set up an autoscaling Kubernetes cluster.
Option B is incorrect because you should not manage the scaling of Kubernetes through the MIG.
Option C is incorrect because a UMIG cannot scale based on a load balancer and this is not the correct way to set up Kubernetes.
Option D is incorrect because Kubernetes will not create additional instances when deployed on Compute Engine. Reference: Kubernetes Engine’s cluster Autoscaler https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-autoscaler

164
Q

For this question, refer to the Mountkirk Games case study. Mountkirk Games wants to set up a real-time analytics platform for their new game. The new platform must meet its technical requirements. Which combination of Google technologies will meet all of their requirements?
A. Cloud Dataflow, Cloud Storage, Cloud Pub/Sub, and BigQuery
B. Cloud SQL, Cloud Storage, Cloud Pub/Sub, and Cloud Dataflow
C. Container Engine, Cloud Pub/Sub, and Cloud SQL
D. Cloud Pub/Sub, Compute Engine, Cloud Storage, and Cloud Dataproc
E. Cloud Dataproc, Cloud Pub/Sub, Cloud SQL, and Cloud Dataflow

A

Correct Answer: A
Option A is correct. Running queries based on month-for-date partitioned tables is an efficient and cost-optimized solution.
Option B is incorrect. While Bigtable can provide low latency for a high volume of reads and writes but it isn’t a requirement here.
Option C is incorrect. One of the requirements is the solution to be cost-effective, and loading the data from Cloud Storage to Bigquery and running queries on Bigquery is a cost-optimized but not a viable and performance-optimized solution.
Option D is incorrect. Cloud Spanner is a transactional database, the requirement suggests a data warehouse service. Reference : Introduction to partitioned tables | BigQuery | Google Cloud

165
Q

Case Study TerramEarth 2 In order to speed up the transmission, TerramEarth deployed 5g devices in their vehicles with the goal of achieving an unplanned vehicle downtime to a minimum. But a set of older vehicles will be still using the old technology for a while. So, on these vehicles, data is stored locally and can be accessed for analysis only when a vehicle is serviced. In this case, data is downloaded via a maintenance port. You need to integrate this old procedure with the new one, building a workflow, in the simplest way. Which of the following tools would you choose?
A. Cloud Composer
B. Cloud Interconnect
C. Appengine
D. Cloud Build

A

Correct Answer: A (see Mountkirk Games case study for details or below for briefing summary) For requirements: Process incoming data on the fly, directly from the game servers - Cloud Dataflow (for both stream and batch), hence we can eliminate C and D since they don’t have DataFlow
Option C is incorrect - Container Engine, Cloud Pub/Sub, and Cloud SQL
Option D is incorrect - Cloud Pub/Sub, Compute Engine, Cloud Storage, and Cloud Dataproc For requirements: Allow SQL queries to access at least 10 TB of historical data - BigQuery, hence we can eliminate B and E since they don’t have BigQuery
Option B is incorrect - Cloud SQL, Cloud Storage, Cloud Pub/Sub, and Cloud Dataflow
Option E is incorrect - Cloud Dataproc, Cloud Pub/Sub, Cloud SQL, and Cloud Dataflow The only correct answer left is A, which meets all of their requirements
Option A is correct - Cloud Dataflow, Cloud Storage, Cloud Pub/Sub, and BigQuery Below is a reference architect Google recommends for a similar scenario in data collection and analysis https://cloud.google.com/solutions/mobile/mobile-gaming-analysis-telemetry Building a Mobile Gaming Analytics Platform - a Reference Architecture Mountkirk Games real-time analytics platform Solution Concept: Mountkirk Games is building a new game, which they expect to be very popular. They plan to deploy the game’s backend on Google Compute Engine so they can capture streaming metrics, run intensive analytics, take advantage of its autoscaling server environment and integrate with a managed NoSQL database.

166
Q

Case Study Mountkirk Games 2 You’re working for a gaming company that collects a huge amount of player telemetry data and game events. The main task is to capture, transform and archive in real time millions of messages on system performance, for analysis on player behavior and for gaming purposes. Data must be available very quickly. Therefore it was thought to use, also in order not to create overload, the information that the system already produces and that are in Json format. The archival time series database service selected must allow queries to access at least 10 TB of historical data How do you process the collected information in order to be used real-time and batch and which storage solution would you choose?
A. Cloud Dataflow and Cloud Bigtable
B. Cloud Dataproc and Cloud Storage
C. Cloud Functions and Cloud Spanner
D. Data Catalog and Cloud Datalab
E. Cloud Dataprep and Cloud Data Fusion

A

Correct Answer: A A is correct. Cloud Composer is a fully managed workflow service that can author, schedule, and monitor pipelines that span across clouds and on-premises data centers. B is wrong. Cloud Interconnect gives fast (10/100Gb) connections to your Google VPC. It is too expensive to connect the fields’ offices in this way. C is wrong. Appengine is a PaaS, so you have to prepare a program for that. It is not simple at all. D is wrong. Cloud Build is a service that builds tour code on GCP for deploy; any kind of code. A Cloud Composer task, when started with automated commands, uses Cloud Identity-Aware Proxy for security, controls processing, and manage storage with Cloud Storage bucket. In this way, it is possible in a simple, standard, and safe way to automate all the processes. Once the files are correctly stored, a triggered procedure can start the new and integrated procedures. For more details, please refer to the URLs below: https://cloud.google.com/composer/ https://cloud.google.com/composer/docs/concepts/cloud-storage

167
Q

Your company has decided to move to Google Cloud Platform. You are asked to explore the GCP environment. You have created a new project called “test-project-ch3” .The technical team would want to know which services are enabled when you create a project in GCP. Please select the right command to list the enabled services also select the services that are enabled when you create a project. Select any two.
A. ” gcloud services list “ is the command, the services that are enabled when you create a project are BigQuery API, Google Cloud APIs, Operations SuiteLogging API, Operations Suite Monitoring API, Datastore API, Service Management API, Service Usage API, Cloud SQL API, Cloud Storage JSON API & Cloud Storage API.
B. “gcloud services list –enabled “ is the command, the services that are enabled when you create a project are BigQuery API, Google Cloud APIs, Operations Suite Logging API, Operations Suite Monitoring API, Datastore API, Service Management API, Service Usage API, Cloud SQL API, Cloud Storage JSON API & Cloud Storage API.
C. ”gcloud services list –available “ is the command, the services that are enabled when you create a project are BigQuery API, Compute Engine API, Operations Suite (formerly Stackdriver) Logging API, Operations Suite (formerly Stackdriver) Monitoring API, Datastore API, Service Management API, Service Usage API, Cloud SQL API, Cloud Storage JSON API & Cloud Storage API.
D. ”gcloud services list –upservices “ is the command , the services that are enabled when you create a project are BigQuery API, Compute Engine API, Operations Suite (formerly Stackdriver) Logging API, Operations Suite (formerly Stackdriver) Monitoring API, Datastore API, Service Management API, Service Usage API, Cloud SQL API ,Cloud Storage JSON API & Cloud Storage API.

A

Correct Answer: A A is correct. Cloud Dataflow is the only product that can process streaming and batch data at the same time. In addition, there are ready to use Templates, useful for streaming data transformation. Look at this link for any further detail: https://github.com/GoogleCloudPlatform/DataflowTemplates Cloud Bigtable provides a massively scalable NoSQL database suitable for low-latency and high-throughput workloads. B is wrong. Cloud Dataproc is a service for running Apache Spark and Apache Hadoop clusters. It is not completely serveless and has limited support for streaming data. Cloud Storage is an object storage solution that allows worldwide storage and retrieval of any amount of data at any time. It is not a Database. C is wrong. Cloud Functions is a serverless execution environment for building and connecting cloud services. You write simple, single-purpose functions that are attached to events released from your cloud infrastructure and services. Also, in this case, there is no direct support for streaming data. In any case, you have to write all your code. Cloud Spanner is a powerful, mission-critical, scalable relational database service, built to support transactions, strong consistency, and high availability across regions and continents. It doesn’t have the required speed of Bigtable. D is wrong. Data Catalog is a fully managed and scalable metadata management service that helps to discover, manage, and understand all their data in Google Cloud. So, it performs a completely different task. The same goes for Cloud Datalab which is a visual tool created to explore, analyze, transform, and visualize data already stored somewhere and build machine-learning models on Google Cloud Platform. E is wrong. Cloud Dataprep is a tool for cleaning and preparing structured and unstructured data for analysis. Cloud Data Fusion is a fully managed, cloud-native data integration service that helps users efficiently build and manage ETL/ELT data pipelines. With a graphical interface and a broad open-source library of preconfigured connectors and transformations, Data Fusion shifts an organization’s focus away from code and integration to insights and action. The problem, even in this case, is that there is no direct support for streaming data. For any further detail, please visit the following URLs: https://cloud.google.com/dataflow/docs/ https://github.com/GoogleCloudPlatform/DataflowTemplates https://cloud.google.com/bigtable/docs/

168
Q

Using the principle of least privilege and allowing for maximum automation, what steps can you take to store audit logs for the long term and to allow access for external auditors to view? (Choose two)
A. Generate a signed URL of the GCS bucketfor auditors to access.
B. Create an account for auditors to have view access to Operations Suite (formerly Stackdriver) Logging.
C. Export audit logs to Cloud Storage via an export sink.
D. Export audit logs to BigQuery via an export sink.

A

Correct Answer: A and B To list the services the current project has enabled for consumption, run: gcloud services list -enabled also –enabled is the default when you don’t use any flag. The services that are enabled when you create a project are BigQuery API, Google Cloud APIs, Operations Suite (formerly Stackdriver) Logging API, Operations Suite (formerly Stackdriver) Monitoring API, Datastore API, Service Management API, Service Usage API, Cloud SQL API, Cloud Storage JSON API & Cloud Storage API.
Option C is incorrect because, it lists the services the current project can enable for consumption, run: .”gcloud services list –available “. Also Compute Engine API isn’t enabled when you create a project, once you click on Compute Engine in the console, it gets enabled.
Option D is incorrect because –upservices is a fictitious flag and it doesn’t exist. References: https://cloud.google.com/sdk/gcloud/reference/services/list

169
Q

Your company places a high value on being responsive and meeting customer needs quickly. Their primary business objectives are release speed and agility. You want to reduce the chance of security errors being accidentally introduced. Which two actions can you take? (Select TWO)
A. Use source code security analyzers as part of the CI/CD pipeline.
B. Ensure you have stubs to unit test all interfaces between components.
C. Ensure every code check-in is peer reviewed by a security SME.
D. Run a vulnerability security scanner as part of your continuous-integration /continuous- delivery (CI/CD) pipeline.
E. Enable code signing and a trusted binary repository integrated with your CI/CD pipeline.

A

Correct Answer: A and C Explanation: C (Correct answer) - Export audit logs to Cloud Storage via an export sink. Cloud Storage is the perfect solution for long-term logs storage. There is 3 type of sink destinations you can export Operations Suite (formerly Stackdriver) Logs to Cloud Storage, Cloud Pub/Sub, BigQuery. While you could export to BigQuery for low-cost storage, BigQuery is mainly and best for analysis not for long-term storage. Besides, whenever you need to do analysis with BigQuery, you can always easily export the logs from GCS to BigQuery or do queries directly against data in the GCS bucket. A (Correct answer) - You could either create a GCP account for auditor ACL object access or a signed URL depending on if they need to have a GCP account or not. Since the requirement is to “allow access for external auditors to view”, hence signed URL is the right choice
B - Does not meet the “for long-term access” requirement.
D - It works but for the “for long-term access” storage consideration, Cloud Storage is a better choice over BigQuery. References: https://cloud.google.com/logging/docs/export/ Generate a signed URL to the Stackdriver export destination for auditors to access.

170
Q

For this question, refer to the TerramEarth case study. The TerramEarth development team wants to create an API to meet the company’s business requirements. You want the development team to focus their development effort on business value versus creating a custom framework. Which method should they use?
A. Use Google App Engine with Google Cloud Endpoints. Focus on an API for dealers and partners.
B. Use Google App Engine with a JAX-RS Jersey Java-based framework. Focus on an API for the public.
C. Use Google App Engine with the Swagger (Open API Specification) framework. Focus on an API for the public.
D. Use Google Container Engine with a Django Python container. Focus on an API for the public.
E. Use Google Container Engine with a Tomcat container with the Swagger (Open API Specification) framework. Focus on an API for dealers and partners.

A

Correct Answer: A and D A (Correct answer) - it’s generally considered as a good practice to leverage source code security analyzers integrated with your CI/CD pipeline. D (Correct Answer) - Run a vulnerability security scanner as part of your continuous-integration /continuous-delivery (CI/CD) pipeline - it’s generally considered as a good practice to do Security scanning of the application and infrastructure as part of the CI/CD pipeline.
B - Ensure you have stubs to unit test all interfaces between components - this is just one of the specific approaches to unit testing your code, not for security error detection. C and
E - The process is not required for an agile practice and it would slow down not speed up the release. Also, those do not specifically have added value for security error detection.

171
Q

You are working for an e-commerce company as a Solutions architect which is planning to migrate the entire application and data to Google Cloud. As a part of the migration, you will be migrating a MySQL database of Size 200GB to Cloud. During the migration, you can have a downtime of a few hours in non-production hrs. Which cost-effective method would you consider first to migrate this database to the cloud?
A. Use the MySQL Dump utility, and copy the exported file to the GCS using gsutil
B. Set up a replica of the database in the cloud, synchronize the data, and then switch traffic to the instance in the cloud.
C. Contact Database Migration Partners on Marketplace
D. Use the MySQL Dump utility, and copy the export file to the cloud using Google Transfer Appliance

A

Correct Answer: A Answer B, C, and D can be simply ruled out since none of the business requirements are public-facing (see TerramEarth case study for details or below for briefing summary). Also, the frameworks mentioned are too specific. A (Correct answer) - Use Google App Engine with Google Cloud Endpoints. Focus on an API for dealers and partners. Google Cloud Endpoints is a distributed API Management system comprising services, runtimes, and tools. Cloud Endpoints is configured using the OpenAPI Specification (formerly Swagger Specification), which provides management, monitoring, and authentication, to help you create, share, maintain, and secure your APIs.
E - Use Google Container Engine with a Tomcat container with the Swagger (Open API Specification) framework. Focus on an API for dealers and partners. Google Cloud Endpoints is OpenAPI Specification (formerly Swagger Specification) based API management to help you create, share, maintain, and secure your APIs. Cloud Endpoints is designed to allow developers to easily choose the language and framework they want for their backend. Tomcat container is just one of the supported containers which primarily provide Java Runtime. TerramEarth Business Requirements Predict and detect vehicle malfunction and rapidly ship parts to dealerships for just-in-time repair where possible. Decrease cloud operational costs and adapt to seasonality. Increase speed and reliability of development workflow. Allow remote developers to be productive without compromising code or data security. Create a flexible and scalable platform for developers to create custom API services for dealers and partners. Cloud Endpoints Architecture

172
Q

For this question, refer to the EHR Healthcare case study. EHR Healthcare is migrating its workloads to the Cloud with the aim of achieving greater security, availability, innovation capabilities, all supported by stronger control and organization. Top Management needs to understand what are the trends in the healthcare world, also by comparing their data stored in Cloud SQL and MongoDB with other information from Open Data, Insurance and Institutions such as WHO. It also wants to carry out forecast analyzes on trends and time series. They do not have an in-house Data Scientist team and therefore want GCP services to facilitate this process. Which GCP service is best suited to these requirements?
A. BigQuery
B. Dataproc
C. Dataflow
D. BigTable

A

Correct Answer: A Because of the small data volume i.e. 200GB and as we can have a few hrs. downtime while migration MySQL Dump utility is the best option for performing the migration.
Option B is incorrect because we can consider downtime while migration. This option is valid when you have strict SLAs that the database should not be down for hrs.
Option C is incorrect because there is no such option.
Option D is incorrect because of the small data volume.

173
Q

For this question, refer to the EHR Healthcare case study. EHR Healthcare wants to create a single, globally accessible, high-performance SQL transactional database that provides EHRs to all customers with minimal latency and allows their management. Insurance providers will provide and access data globally. Which service do you think is the best?
A. Cloud Spanner
B. Cloud SQL with MySQL and global Read Replicas
C. Cloud SQL with SQL Server and global Read Replicas
D. Firestore replicated in multiple regions

A

Correct answer: A BigQuery allows to manage data with the standard SQL language, well known in the company, but not only: it allows access to data on text files and other databases, both in GCP, on-premises and in other Clouds. Furthermore, with BigQuery ML, it is possible to create Machine Learning models for the required forecasting services without a team of Data Scientists. B is wrong because Dataproc performs data analysis with the tools of the Hadoop ecosystem, but you need to know it well and it has neither the ease of use, nor the ability to integrate external data, nor the versatility of BigQuery. C is wrong because Dataflow is aimed to manage parallel Data Pipelines, that is, data transformations. In Batch and Streaming mode. It is not an analytics tool. D is wrong because BigTable is a noSQL database, powerful and very fast, used by Google for Gmail and Maps. It is not so suitable for data integration and prediction. For any further detail: https://cloud.google.com/bigquery-ml/docs/introduction https://cloud.google.com/bigquery/docs/introduction

174
Q

You have been hired as DevSecOps Engineer by a large finance company. They have recently developed a sensitive application that runs on Google Compute Engine that does not have external IP. Your Instance needs to upload Application log files to Google Cloud storage for Logs Archival daily. Which method would you choose for enabling secure access to Google Cloud storage without using external IP with Instance.
A. Enable Private Google Access for Subnet
B. Create a Cloud NAT Gateway
C. Enable Private Google Access for Entire VPC
D. Enable Private Services Access

A

Correct answer: A Cloud Spanner is a fully managed, globally distributed, ACID-compliant relational database with read-write unlimited scale, strong consistency, and up to 99.999% availability. It handles replicas, sharding, and transaction processing. B is wrong because the DB must be globally in read-write mode. In addition Cloud SQL may be a too small solution for a such growing business C is wrong also because SQL Server do not support global Read Replicas D is wrong because Firestore is not a SQL transactional database For any further detail: https://cloud.google.com/spanner/docs/replication

175
Q

For this question, refer to the EHR Healthcare case study. The case study explains that: EHR hosts several legacy file-based and API integrations with on-site insurance providers, which are expected to be replaced in the coming years. Hence, there is no plan to upgrade or move these systems now. But EHR wants to use these APIs from its applications in Google Cloud so that they remain on-premise and private, securely exposing them. In other words, EHR wants to protect these APIs and the data they process, connect them only to its VPC environment in Google Cloud, with its systems in a protected DMZ that is not accessible from the Internet. Providers will be able to access integrations only through applications and with all possible precautions. Which technique allows you to fulfill these requirements?
A. Gated Egress and VPC Service Controls
B. Cloud Endpoint
C. Cloud VPN
D. Cloud Composer

A

Correct Answer: A Enable Private Google Access for Subnet When you enable Private Google Access for a subnet, instances in that subnet that don’t have a public IP address can access Google APIs and services like Cloud Storage, etc. https://cloud.google.com/vpc/docs/configure-private-google-access
Option B is incorrect because Cloud NAT Gateway is used by instances that require internet access. By using a NAT gateway, instances in a private subnet can download or upload files to Cloud storage
Option C is incorrect because Private Google Access is a subnet-level setting.
Option D is incorrect because Private Services Access is a not available for GCE and GCS. https://cloud.google.com/vpc/docs/private-services-access

176
Q

You have been hired as a Solutions Architect by an e-commerce company which is planning full migration to GCP cloud. They have several Hadoop and Spark Clusters running on-premise which are managed by DevOps and an IT team which do interruptible data analysis work. After the data analysis is finished, the data is stored on NAS located in the same Datacenter for future use. The CTO has tasked you to reduce the expenses while planning the migration. How will you accomplish this task?
A. Use managed service like Dataproc with preemptible VM for data processing and GCS for storage
B. Use Compute VM for Hadoop & Spark clusters and GCS for storage
C. Use managed service like Dataproc with standard VM for data processing and GCS for storage
D. Use Compute VM for Hadoop & Spark clusters and create NAS cluster on Compute VM for storage

A

Correct Answer: A Gated egress topology lets APIs in on-premise environments be available only to processes inside Google Cloud without direct public internet access. Applications in Google Cloud communicate with APIs in on-premise environments only with private IP addresses and are eventually exposed to the public via an Application Load Balancer and using VPC Service Controls. VPC Service Controls create additional security for Cloud applications: Isolate services and data Monitor against data theft and accidental data loss Restrict access to authorized IPs, client context, and device parameters B is wrong because Cloud Endpoint is an API Gateway that could create an application facade as required. But Cloud Endpoint does not support on-premises endpoints. C is wrong because Cloud VPN is just a way to connect the local network to a VPC. D is wrong because Cloud Composer is a workflow management service. For any further detail: https://cloud.google.com/architecture/hybrid-and-multi-cloud-network-topologies#gated-egress https://cloud.google.com/vpc-service-controls#all-features

177
Q

You work as an architect in a company that develops statistical studies on big data and produces reports for its customers. Analysts often allocate VMs to process data with ad hoc development procedures. You have been called by the administrative department because they have been billed for a very large number of Compute Engine instances, which you also consider excessive in relation to operational needs. How can you check, without controlling them one by one, which of these systems can be accidentally left active by junior technicians?
A. Use the Recommender CLI Command
B. Use Cloud Billing Reports
C. Use Idle Systems Report in GCP Console
D. Use Security Command Center Reports

A

Correct Answer: A The use of Managed services reduces the workload on systems administrators and DevOps engineers because this will eliminate some of the work required when managing your own implementation of a platform. The use of preemptible machines will cost significantly less than standard VMs.
Option B is incorrect because using Compute VM for Hadoop & Spark clusters will not eliminate the management and operational work
Option C is incorrect because using Standard VM will not reduce operational cost
Option D is incorrect because using Compute VM for Hadoop & Spark clusters will not eliminate the management and operational work.

178
Q

Your international Company called CompA has acquired the firm CompB that also uses GCP and has its own VPC named VPC-B. CompA administers the Cloud Services through a global network that incorporates its VPC (named VPC-A) with the on-premise technical environment through a redundant VPN with BGP dynamic routing. It is in addition required to extend this organization to VPC-B in order to manage the necessary integrations and migrations of data and applications in an easier and smoother way. All of this, of course, should be done in the fastest, most manageable and convenient way. Which of the following solutions would you recommend?
A. VPC peering
B. Dedicated Interconnect between the 2 VPCs
C. Another VPN between on-premise and VPC-B
D. Partner Interconnect between the 2 VPCs
E. Shared VPC

A

Correct Answer: A This command : gcloud recommender recommendations list –recommender = google.compute.instance.IdleResourceRecommender gives to all the idle VMs based on Cloud Monitoring metrics of the previous 14 days. There is no equivalent in the Console. B is wrong because Cloud Billing Reports don’t give details about activities. C is wrong because there is no Idle Systems Report in the GCP Console. D is wrong because the Security Command Center is used for Security threats, not for ordinary technical operations. For any further detail: https://cloud.google.com/compute/docs/instances/viewing-and-applying-idle-vm-recommendations

179
Q

You have a few media files over 5GB each that you need to migrate to Google Cloud Storage. The files are in your on-premises data center. What migration method can you use to help speed up the transfer process?
A. Use multi-threaded uploads using the -m option.
B. Use parallel uploads to break the file into smaller chunks then transfer it simultaneously.
C. Use the Cloud Transfer Service to transfer.
D. Start a recursive upload.

A

Correct Answer: A VPC peering allows any integration of VPC between different Organizations, so it meets all requirements. B, C, and D are wrong because the need is to connect the 2 VPCs, without realizing other data link on-premises. E is wrong because shared VPC is only within an organization. Check the following URLs: https://cloud.google.com/blog/products/gcp/getting-started-with-shared-vpc https://cloud.google.com/blog/products/gcp/getting-started-with-shared-vpc For any further detail, please visit the URLs below: https://cloud.google.com/vpn/docs/concepts/choosing-networks-routing https://cloud.google.com/blog/products/networking/google-cloud-networking-in-depth-simplify-routing-between-your-vpcs-with-vpc-peering

180
Q

What of the following activity should be carried out to reverse the process of compression, deduplication, and encryption which has occurred while transferring the data to a Transfer Appliance?
A. Link Aggregation
B. Data Rehydration
C. Data Capture
D. Data Recapture

A

Correct Answer: B
Option A is incorrect- Use multi-threaded uploads using the -m option. If you have a large number of files to transfer you might want to use the gsutil -m option, to perform a parallel (multi-threaded/multi-processing) copy: gsutil -m cp -r dir gs://my-bucket
Option B (Correct answer) - Parallel uploads are for breaking up larger files into pieces for faster uploads. gsutil can automatically use object composition to perform uploads in parallel for large, local files being uploaded to Google Cloud Storage. If enabled (see below), a large file will be split into component pieces that are uploaded in parallel and then composed in the cloud (and the temporary components finally deleted). gsutil -o GSUtil:parallel_composite_upload_threshold=150M cp bigfile gs://your-bucket
Option C is incorrect. Use the Cloud Transfer Service to transfer. Storage Transfer Service is limited to AWS S3, Google Cloud Storage, On-premise and HTTP/HTTPS locations. Our requirement is “help speed up the transfer process” Using
Option C, we can migrate but will not help to increase speed.
Option D is incorrect Start a recursive upload: The -R and -r options are synonymous. Causes directories, buckets, and bucket subdirectories to be copied recursively. Reference: cp - Copy files and objects https://cloud.google.com/storage/docs/gsutil/commands/cp

181
Q

You have been delegated Access to XYZ Organization, you want to create Shared VPC, even with the delegated access you aren’t able to create the Shared VPC, what solution would resolve the issue?
A. With Delegated Access, you don’t need any other extra permission.
B. Give yourself Compute Shared VPC Admin role
C. Give yourself Compute Admin Access
D. Add your member and give them a Shared Network Admin Role.

A

Correct Answer: B
Option B is Correct because Data rehydration is the process by which you fully reconstitute the files so you can access and use the transferred data. To rehydrate data, the data is first copied from the Transfer Appliance to your Cloud Storage staging bucket. The data uploaded to your staging bucket is still compressed, deduplicated, and encrypted. Data rehydration reverses this process and restores your data to a usable state.
Option A is Incorrect because Link aggregation is the bundling of multiple network interfaces (NIC) into a single link, which allows maximum data throughput and redundancy in case of a network connection failure.
Option C is Incorrect because Data capture jobs are used to identify data on your network and stream it to Google Transfer Appliance.
Option D is Incorrect because Data Recapture activity doesn’t exist. Reference: https://cloud.google.com/transfer-appliance/docs/2.0/data-rehydration

182
Q

You are working for a major Bank that operates worldwide and has recently acquired a small Hong Kong financial institution that offers truly innovative services. The IT Management is planning to migrate these apps to make them global and with a much larger number of loyal users. These systems use a mySQL Database which, while preserving data integrity and consistency, is absolutely inadequate for the new scope. Which of the following solutions would you recommend?
A. Cloud SQL with Postgres e Memory and MemoryStore
B. Cloud Spanner
C. Cloud Firestore
D. Cloud SQL with MS SQL Server
E. Cloud Datastore

A

Correct Answer: B
Option B is the CORRECT because Compute shared VPC Admin role gives you the permission to set up Shared VPC.
Option A is INCORRECT because even with delegated access you need to give yourself Compute Shared Admin role.
Option C is INCORRECT because giving the Compute Admin Access role doesn’t give you permission to set up Shared VPC.
Option D is INCORRECT because there is NO Shared Network Admin Role in GCP Read More at https://cloud.google.com/vpc/docs/provisioning-shared-vpc

183
Q

For this question, refer to the TerramEarth case study. TerramEarth receives daily data in the Cloud using network interconnects with private on-premises data centers. A subset of the data is transmitted and processed in real time and the rest daily, when the vehicles return to home base. You have been asked to prepare a complete solution for the ingestion and management of this data, which must be both fully stored and aggregated for analytics with Bigquery. Which of the following actions do you think is the best solution (pick 2)?
A. Real-time data is streamed to BigQuery, and each day a job creates all the required aggregate processing
B. Real-time data is sent via Pub / Sub and processed by Dataflow that stores data in Cloud Storage and computes the aggregates for BigQuery.
C. The Daily Sensor data is uploaded to Cloud Storage using pub/sub with parallel composite uploads and at the end with a Cloud Storage Trigger a Dataflow procedure is activated
D. Daily Sensor data is loaded quickly with BigQuery Data Transfer Service and processed on demand via job

A

Correct Answer: B A is wrong: Cloud SQL is not global B is correct: it meets all the requirements (see the main features below) C is wrong: Cloud Firestore is not global and it is a NoSQL DB D is wrong: Cloud SQL is not global E is wrong: Cloud Datastore is not global and it is a NoSQL DB A small recap of the targeted technologies: Cloud Spanner is a mission-critical, scalable relational database service, built to support transactions, strong consistency, and high availability across regions and continents. Main features: scalable, enterprise-grade, globally-distributed, combine the benefits of relational database structure with a non-relational horizontal scale industry-leading 99.999% availability SLA no planned downtime enterprise-grade security. For more details, please visit the link: https://cloud.google.com/spanner/ Cloud SQL is a fully-managed database service that offers to set up, maintain, manage, and administer your relational PostgreSQL, MySQL, and SQL Server databases in the cloud. Main features: high performance and good scalability (not linear). database infrastructure for applications running anywhere. Regional with multi-regional Backups. Live migration High availability with continuous health-checking and automatic failover For more details, please visit the link: https://cloud.google.com/sql/ Cloud Firestore is a fast, fully managed, serverless, cloud-native NoSQL document database that simplifies storing, syncing, and querying data for your mobile, web, and IoT apps at a global scale. Cloud Firestore is the next generation of Cloud Datastore. Main features: ACID transactions Datastore mode Regional (multi-zone) with multi-region replication, if needed Automatic horizontal scaling in and out Realtime Database For more details, please visit the link: https://cloud.google.com/firestore/ Cloud Datastore is a highly-scalable NoSQL database for your applications. Cloud Datastore automatically handles sharding and replication Main features: ACID transactions Regional with multi-region replication Regional (multi-zone) with multi-region replication, if needed It will be replaced by Firestore For more details, please visit the link: https://cloud.google.com/datastore/

184
Q

Helicopter Racing League (HRL) offers premium contents and, among their business requirements, has: To increase the number of concurrent viewers and To create a merchandising revenue stream So, they want to offer service subscriptions for their and partner services and manage monetization, pay-as-use management, flat-use control, and rate-limiting. All the functionalities that can assure a managed revenue stream in the simplest way. Which is the best GCP Service to achieve that?
A. Cloud Endpoints
B. Apigee
C. Cloud Tasks
D. Cloud Billing
E. API Gateway

A

Correct Answer: B and C Pub/Sub is the solution recommended by Google because it provides flexibility and security. Flexibility because, being loosely coupled with a publish / subscribe mechanism, it allows you to modify or add functionality without altering the application code. Security because it guarantees reliable, many-to-many, asynchronous messaging with at-least-once delivery. Uploading to both Cloud Storage and Bigquery is important because you want to store the data both in its entirety and in aggregate form. Parallel composite uploads are recommended because the daily files are of considerable size (200 to 500 megabytes). Using Dataflow allows you to manage processing in real-time and to use the same procedures for daily batches. A is incorrect because it stores data only in BigQuery and does not provide real-time processing when the requirements are to have both global and aggregated data. D is incorrect because, also here, data is stored only in BigQuery and because BigQuery Data Transfer Service involves passing through cloud sources, not from on-premise archives. It also doesn’t talk about how the data is decompressed and processed. For any further detail: https://cloud.google.com/bigquery-transfer/docs/introduction https://cloud.google.com/storage/docs/uploads-downloads#parallel-composite-uploads https://www.youtube.com/playlist?list=PLIivdWyY5sqJcBvDh5dfPoblLGhG1R1-O https://cloud.google.com/pubsub

185
Q

A tour company deals with sensitive data and must meet high compliance requirements. All the interactions regarding Customer Care Offices are stored with BigQuery and Cloud Storage. You have to give advice for protecting private data with a quick and managed solution. Which of the following is the best solution?
A. Develop a Cloud Function internally that detects sensitive data and transfers data and records in a special vault
B. Use DLP
C. Use Versioning to classify data
D. Use PCI Data Security Standard compliance

A

Correct Answer: B Apigee is the GCP top product for API management. It offers all the functionalities requested: monetization, traffic control, throttling, security and hybrid (third -parties) integration. GCP offers 3 different products for API management: Apigee, Cloud Endpoints (only GCP) and API Gateway (for Serverless workloads). A is wrong because Cloud Endpoints is an API product, too but doesn’t support monetization and hybrid C. is wrong because Cloud Tasks is a dev tool for thread management D is wrong because Cloud Billing is for GCP services accounting, billing and reporting, not for end-user services E is wrong because API Gateway is an API product, too but doesn’t support monetization and hybrid For any further detail: https://cloud.google.com/apigee/docs/api-platform/get-started/what-apigee

186
Q

Case Study Mountkirk Games 1 Mountkirk Games collects a huge amount of player telemetry data and game events. The main task is to capture, transform, and archive millions of messages on system performance in real time, to analyze player behavior and for gaming purposes. It has to process incoming data on the fly directly from the game servers, so data must be available very quickly. As the system scales, it is important to ensure that data is not lost due to processing backlogs. Therefore, it was thought to use and create overload, the information that the system already produces and that is in JSON format. How would you acquire and collect information?
A. Cloud Composer
B. Pub/Sub
C. Cloud Task
D. Cloud Functions
E. Cloud SQL

A

Correct Answer: B Cloud DLP can identify where sensitive data is stored, then use tools to redact those sensitive identifiers. Cloud DLP uses more than 90 predefined detectors to identify patterns, formats, and checksums, and de-identification techniques like masking, secure hashing, and tokenization to redact sensitive data, all without replicating customer data. A is wrong. This solution is not at all quick and managed B is correct. Cloud DLP is the perfect solution C is wrong. Versioning doesn’t protect sensitive data but preserves an old version of the Objects. D is wrong. PCI Data Security Standard compliance, that is Payment Card Industry Data Security Standard (PCI DSS) compliance is only related to electronic payments. It sets the requirements for organizations and sellers to accept, store, process, and transmit cardholder data safely and securely during a credit card transaction. For more details, please check the URLs below: https://cloud.google.com/blog/products/storage-data-transfer/scan-your-cloud-storage-buckets-for-sensitive-data-using-cloud-dlp

187
Q

For this question, refer to the EHR Healthcare case study. EHR Healthcare wants, in an initial project phase, to transfer its workloads to the Cloud in an efficient way and minimizing the impact for applications. This also applies to its on-premises databases, which are mySQL, SQL Server and MongoDB. At a later stage, the data management will be reorganized together with the management applications and procedures. There is a need to bring MongoDB to the Cloud. MongoDB is used for non-transactional, scalable systems and you want to continue using a noSQL system of this type, in order to have full control of its management but the final goal is to have a service that is fully managed in the Google Cloud Platform. Given these requirements, which solution would you recommend for the first migration and subsequent integration?
A. Install MongoDB in Cloud on a VM and then switch to Datastore
B. Deploy MongoDB in Cloud as a StatefulSet and then switch to Firestore
C. Directly migrate MongoDB to Firestore
D. Directly migrate MongoDB to DocumentDB and then switch to Datastore

A

Correct Answer: B Cloud Pub/Sub is the perfect product for this project: you can send and receive messages between independent applications and transmit data across projects and applications running on cloud, on-premise, or hybrid environments. Cloud Pub/Sub is perfect to decouple systems and components hosted on GCP or elsewhere on the internet. It provides “at least once” delivery at low latency with on-demand scaling to tens of millions of messages per second. A is wrong. Cloud Composer is a managed workflow orchestration. So, different processing are often related to each other B is correct. C is wrong. Cloud Tasks manage queues, but within an Application, so its scalability cannot meet requirements. It is used with App Engine. D is wrong. Cloud Functions are the small pieces of code executed in an event driven mode. Natively, it doesn’t handle messages. E is wrong. It a managed Database Service. Its scalability cannot meet requirements, definitely. For any further detail, please refer to the URL below: https://cloud.google.com/pubsub/

188
Q

For this question, refer to the MountKirk Games case study: Based on MountKirk Games’ technical requirements, what GCP services/infrastructure will they use to host their game backend platform?
A. App Engine Flexible Environment
B. Google Kubernetes Engine
C. App Engine Standard Environment
D. Managed Instance Group on Compute Engine

A

Correct answer: B MongoDB is a horizontally scalable noSQL Document Database that is not offered in GCP as a managed service; in order to obtain horizontal scalability and data redundancy, however, you need to install it on a cluster of VMs. Given the volumes of EHR, the best solution is certainly to use Kubernetes and its StatefulSets and Volumes. Kubernetes is known for offering a perfect environment for Microservices, with its Stateless Deployments , but it is not the only possible solution. In fact, there are also StatefulSets, Demons and Jobs. It should be noted that Google uses its own internal Kubernetes for any type of workload. The closest MongoDB-like product in the GCP is Firebase, so this is the best strategy. A is wrong because a single VM deployment doesn’t offer any scalability and failover. C and D are wrong because a migration is out of the requirements scopes. Moreover, DocumentDB is an AWS Service, not a GCP one. For any further detail: https://kubernetes.io/blog/2017/01/running-mongodb-on-kubernetes-with-statefulsets/ https://cloud.google.com/kubernetes-engine/docs/concepts/statefulset?hl=it

189
Q

Helicopter Racing League (HRL) offers a paid service to stream the races all over the world with live telemetry and predictions throughout each race. Management wants to increase telemetry and create additional insights; they want to start as soon as possible, before the migration in GCP, for all kinds of services, and they want to keep on improving this process. Which of the following is the best solution?
A. Develop a sw module to be deployed inside each Pod in Kubernetes
B. Use open source tools like Prometheus and gradually integrate with Operation Suite
C. Use Operations Suite (formerly Stackdriver) metrics and data from all the services
D. Use Istio telemetry functions
E. Use Cloudwatch metrics and data from all the services

A

Correct Answer: B Since the case study clearly mentions that “They plan to deploy the game’s backend on Google Kubernetes Engine so they can scale rapidly”, hence Google Kubernetes Engine can be used. Case Study GKE Documentation

190
Q

You have a Cloud Function that sometimes fails because of an error that is still not well identified. The error happens randomly, sometimes it occurs and sometimes it doesn’t. Is there a method to minimize the effect while the developers are looking for the solution?
A. Change the function with an exception routine that logs the error
B. Use the Retry on failure option
C. Change the function with an exception routine that repeats the operation
D. Use Pub/Sub to decouple the operation and store the results
E. Use Operations Suite (formerly Stackdriver) Error Reporting

A

Correct Answer: B Telemetry includes all control data and metrics from cloud systems. Telemetry needs to be extracted, organized and transmitted from multiple locations to a central point of analysis. We go beyond the scalability and availability of services, because the purpose of this Data is to optimize processes and costs; for example, Telemetry is used for the security and integrity of applications, to improve the user experience, to maximize performance and, more generally, the quality of systems. Prometheus is a widely used open-source (Cloud Native Computing Foundation graduated) for the acquisition, integration, query and analysis of telemetry data. Its main features are: All data is collected into an easy to use multidimensional data model Uses a flexible query language called PromQL Flexible System Management and, security and no overhead graphing and dashboarding support, even if often used with Kabana, another open tool GCP has new features for Service and Telemetry reporting that may be integrated after migration A is wrong because this is not an easy and fast solution, even if it is the Google internal solution; Pods in Google are often made by the app container plus the control container that deals with telemetry. C is wrong because Operations Suite (formerly Stackdriver) don’t cover all the needs for a wide-range telemetry D is wrong because Istio is a service-mesh for Kubernetes; so, it is aimed at microservices architectures. It is open-source and offers important telemetry functions, but it doesn’t cover all of our requirements. E is wrong because Cloudwatch is the Operations Suite (formerly Stackdriver) counterpart product in AWS For any further detail: https://prometheus.io/docs/introduction/overview/ https://cloud.google.com/service-infrastructure/docs/telemetry-reporting https://istio.io/latest/about/service-mesh/

191
Q

For this question, refer to the EHR Healthcare case study. EHR Healthcare aims to migrate applications from on-premises to Cloud, from monolithic to microservices and to improve all levels of quality. At the same time EHR Healthcare wants to set up a gradual process, with modern DevOps techniques, which allows it to take as few risks as possible and to keep the services for customers unchanged. So it wants to manage agile processes, with the deployment of new features while preserving the user interface, optimizing traffic and latency, and finally increasing the level of security. Which of the following techniques would you use?
A. Canary deployment
B. Api Management with Quota and Rate Limiting
C. Istio with Kubernetes
D. App Engine Standard Edition
E. App Engine Flexible Edition

A

Correct Answer: B The ‘retry on failure’ option is aimed to automatically retry a background function until it completes successfully. And this is exactly what is needed. You have to use it carefully and you should handle permanent problems inside the function. The other proposed solutions modify the function which is not advisable because developers are already looking for the solution or just the debugging methods. For any further detail, please follow the URL below: https://cloud.google.com/blog/products/serverless/cloud-functions-pro-tips-using-retries-to-build-reliable-serverless-systems

192
Q

Your team is redacting a new application that is about to go into production. During testing, it emerges that a developer code allows user input to be used to modify the application and execute commands. This event has thrown everyone into despair and has generated the fear that there are other problems of this type in the system. Which of the following services may help you?
A. Cloud Armor
B. Web Security Scanner
C. Security Command Center
D. Shielded GKE nodes

A

Correct answer: B The best and simplest way to achieve that is with an API Gateway, possibly Apigee. Apigee is the most complete service in GCP for API management and supports any environment. The official documentation says: Apigee on the other hand is a comprehensive API management platform built for Enterprises, with deployment options on cloud, on-premises, or hybrid. The feature set includes a full-fledged API gateway, customizable portal for on-boarding partners and developers, rate and Traffic limiting and management, monetization, and deep analytics around your APIs. A is wrong because Canary deployment doesn’t create an API facade service as required. C is wrong because ISTIO is a service mesh management and even if it can help for app evolving, it doesn’t create an API facade service. D and E are PaaS; they manage app versions and advanced deployments but not hybrid configurations as required. For any further detail: https://cloud.google.com/api-gateway/docs/deployment-model#multiregion https://docs.apigee.com/api-platform/develop/rate-limiting

193
Q

Company rules require that the VMs within the private subnets never have a public address. As an administrator, you need to perform software updates for all your Instances using Internet connections. You need a fast, easy and reliable automated solution. Which of the following would be the best solution?
A. Use Cloud CDN
B. Use a NAT Instance or a Cloud NAT gateway
C. Setup a VPN Connection
D. Use Cloud Interconnect
E. Use a Bastion Host

A

Correct Answer: B What you need is a service that examines your code and finds out if something is vulnerable or insecure. Web Security Scanner does exactly this: it performs managed and custom web vulnerability scanning. It performs scans for OWASP, CIS GCP Foundation, PCI-DSS (and more) published findings. A is wrong because Cloud Armor is a Network Security Service, with WAF rules, DDoS and application attacks defenses. C is wrong because the Security Command Center suite contains Web Security Scanner and many other services. D is wrong because Shielded GKE nodes are special and secured VMs. For any further detail: https://cloud.google.com/security-command-center/docs/concepts-web-security-scanner-overview

194
Q

Your customer is moving their storage product to Google Cloud Storage (GCS). The data contains personally identifiable information (PII) and sensitive customer information. What security strategy should you use for GCS?
A. Use signed URLs to generate time bound access to objects.
B. Grant IAM read-only access to users and use default ACLs on the bucket.
C. Do not grant any Access Management (Cloud IAM) roles to users belonging to Cloud Identity and use granular ACLs on the bucket.
D. Create randomized bucket and object names. Enable public access, but only provide specific file URLs to people who do not have Google accounts and need access.

A

Correct Answer: B You could have solved the problem with a NAT system or a Bastion Host too, but the first option is far simpler and meets requirements. Network Address Translation (NAT) performs Private to Public Address translations, so VMs in private subnets can access the internet—for updates, patching, config management, and more—in a controlled and efficient manner, but outside resources cannot directly access those instances. A is totally out of scope. C and D could solve the problem but are too expensive and complex for this case. For any further detail, please refer to the URLs below: https://cloud.google.com/nat/docs/overview https://cloud.google.com/blog/products/networking/cloud-nat-deep-dive-into-our-new-network-address-translation-service

195
Q

In your company it is now necessary for auditors and control managers to be able to access and query the IT systems, in a simple way, for both the administrative functions and for data access. IT management has the problem of providing simple access, organized by historical period. At the same time, they need to restrict costs. Which solution would you recommend?
A. Let auditors and control managers use Operations Suite (formerly Stackdriver) Logging
B. Export selected logs to Cloud Storage
C. Export selected logs to Cloud BigQuery
D. Develop an interface application for accessing log data

A

Correct Answer: C
Option C is correct - This grants the least privilege required to access the data and minimizes the risk of accidentally granting access to the wrong people.
Option A is incorrect. Signed URLs could potentially be leaked.
Option B is incorrect. This is needlessly permissive, users only require one permission in order to get access.
Option D is incorrect - This is security through obscurity, also known as no security at all.

196
Q

You have been asked to migrate to the Cloud, a legacy system, listening to specific TCP ports, in a “lift and shift way”, that needs different types of instances but is accessed by a single IP address. Which configuration among the following would you evaluate as the best one?
A. Layer 7 Load Balancer with an Unmanaged Instance Group
B. Layer 4Load Balancer with a Managed Instance Group
C. Layer 4Load Balancer with an Unmanaged Instance Group
D. Layer 7 Load Balancer with a Managed Instance Group

A

Correct Answer: C A is wrong. It is not easy and simple. Auditors would not be able to search and analyze data the way they would need it. B is wrong. Cloud Storage is not searchable at the granular level required. C is correct. It will be easy to query and organize the Log information by historical period D is wrong. It is neither easy nor simple and economical You may configure log export and deliver them to a dataset in BigQuery, granting permissions to limit access. Date-partitioned tables and expiration dates could help limit query costs by reducing the amount of data scanned as a part of queries and might keep auditing the logging data for the required period and then delete it. For any further detail, please refer to the URLs below: https://cloud.google.com/logging/ https://cloud.google.com/solutions/exporting-stackdriver-logging-for-security-and-access-analytics

197
Q

Your company has collected all its historical data in BigQuery and Documents in Cloud Storage. Most BigQuery tables are structured on historical sales data and have a large number of fields, many of which are rarely examined. Almost all the queries target only current and previous years and they have to select data by product families and sale regions. Older data is stored in tables with only statistical aggregates. They asked you to improve this process, with the following guidelines: Economicity Ease of use Security The system is going to be used less experienced users, so a special attention have to be cared for. Which one of the following strategies would you consider appropriate?
A. Leave the Organization untouched, It is just perfect
B. Use BigTables with Views
C. Use BigQuery Clustered tables withauthorized views
D. Cloud Spanner withauthorized views
E. Use Cloud Firebase

A

Correct Answer: C A is wrong. The application does not use HTTP(s), so an HTTP(S) Load Balancer is out of scope B is wrong. Managed Instance Groups are for scalable and identical Instances C is correct. It is the only feasible way D is wrong. The application does not use Http, and Managed Instance Groups is for scalable and identical Instances Flowchart for choosing a load balancer: For more details, please refer to the URLs below: https://cloud.google.com/load-balancing/docs/choosing-load-balancer https://cloud.google.com/compute/docs/instance-groups/

198
Q

Your business is connected to a VPN connection to GCP. On GCP, you are hosting an internal website using an HTTP load balancer in the same region, which serves traffic to a managed instance group. Your users are unable to access the internal website, though all systems are up. What is a possible cause for this problem?
A. You’ve enabled static IP address for the managed instance group servers.
B. Your load balancer is not set for the correct region.
C. Firewall rules are not correctly configured.
D. You need to enable Operations Suite (formerly Stackdriver) debugger to collect more information.

A

Correct Answer: C A is wrong. The key requirement that you need to meet is to select data by product families and sale regions B is wrong. BigTable is a NoSQL database, not an Analytics System C is correct. Clustering Partitioned Tables with Product Type and Sale Region you will gain in speed and spare money. D is wrong. Cloud Spanner is a global SQL database and not an Analytics System E is wrong. Cloud Firestore is a NoSQL database and not an Analytics System BigQuery is a fully managed, low cost, serverless, columnar and ANSI SQL data warehouse that can analyze terabytes to petabytes of data in blazing-fast speeds. Additional features: Analyze geospatial data using familiar SQL with BigQuery GIS ML models on large-scale structured or semi-structured data: BigQuery ML Real-time interactive dashboarding with sub-second query latency using BigQuery BI Engine. Transferring services and flexible data ingestion Pay-for-what-you-use pricing It doesn’t use indexes, but only full-scan searches. It is important because you will be charged for the amount of data processed for a query. So, the use of Partitioned Tables, Clustering Tables, and View are uttermost important. Clustering Tables improve the performance of queries when filter clauses based on the contents of clustered columns are used. When data is written to a clustered table by a query job or a load job, BigQuery sorts the data using the values in the clustering columns. These values are used to organize the data into multiple blocks in BigQuery storage. When you submit a query containing a clause that filters data based on the clustering columns, BigQuery uses the sorted blocks to eliminate scans of unnecessary data. Similarly, when you submit a query that aggregates data based on the values in the clustering columns, performance is improved because the sorted blocks find out rows with similar values. For any further detail, please check the URLs below: https://cloud.google.com/bigquery/docs/clustered-tables https://cloud.google.com/bigquery/docs/authorized-views

199
Q

An e-commerce System is operating in an App Engine flex with Node.js and has to perform many operations while registering orders. You have been asked to find a way to decouple the service with a procedure that will send an email to the customer with an order confirmation, at the end. You must find the fastest, serverless, simple, economic, and secure solution to perform this procedure. What would you recommend?
A. Make an asynchronous call with Node.js
B. Use Pub / Sub to call a Cloud function
C. Use Cloud Task and define an appropriate worker server
D. Record orders in a database and manage them with a cron job within App Engine

A

Correct Answer: C Check to make sure you have proper firewall rules allowing the correct subnets access. Also, make sure another rule with higher priority is not conflicting with it.

200
Q

Mountkirk Games wants to set up a continuous delivery pipeline. Their architecture includes many small services that they want to be able to update and roll back quickly. Mountkirk Games has the following requirements: * Services are deployed redundantly across multiple regions in the US and Europe. * Only frontend services are exposed on the public internet. * They can provide a single frontend IP for their fleet of services. * Deployment artifacts are immutable. Which set of products should they use?
A. Google Cloud Functions, Google Cloud Pub/Sub, Google Cloud Deployment Manager
B. Google Cloud Storage, Google App Engine, Google Network Load Balancer
C. Google Container Registry, Google Kubernetes Engine, Google HTTP(s) Load Balancer
D. Google Cloud Storage, Google Cloud Dataflow, Google Compute Engine

A

Correct Answer: C Cloud Tasks is an asynchronous task execution service that encodes and executes Tasks using Queues. It enables serverless execution for Systems operating in App Engine standard or flexible environments. With this Service, it is possible to offload long-running and background activities, decouple services from one another and make your applications much more resilient to failures. The other answers depict feasible solutions, that are not Serverless (A,D) or incomplete (B) For more details, please refer to the URLs below: https://cloud.google.com/tasks/ https://cloud.google.com/appengine/docs/flexible/dotnet/scheduling-jobs-with-cron-yaml

201
Q

You work in an international company and manage many GCP Compute Engine instances using SSH and RDP protocols. The management, for security reasons, asks you that VMs cannot have public IP addresses. So you are actually no longer able to manage these VMs. How is it possible to manage in a simple and secure way, respecting the company rules, access and operations with these systems?
A. Bastion Hosts
B. Nat Instances
C. IAP’s TCP forwarding
D. Security Command Center

A

Correct Answer: C Google Container Registry, Google Kubernetes Engine, Google HTTP(s) Load Balancer. As per the requirements, Google Container Registry, and Google Kubernetes Engine meets the below requirements: “Their architecture includes many small services that they want to be able to update and roll back quickly”; And following specific requirements * Services are deployed redundantly across multiple regions in the US and Europe. * Deployment artifacts are immutable. And Google HTTP(s) Load Balancer meets the below requirements: * Only frontend services are exposed on the public internet. * They can provide a single frontend IP for their fleet of services. All other answers provide an incomplete or incorrect solution and don’t meet the requirements.

202
Q

You have been asked to design an application that allows you to manage XML Customer Orders documents as soon as they are written in a specific Cloud Storage Bucket. The document must be processed with an existing routine written in Node.js. The solution must be as simple as possible, fully managed, asynchronous, secure and economical. How could this be accomplished?
A. Develop a new App to be deployed with App Engine Standard that periodically reads the bucket, processes the files, and transfers them to a new bucket
B. Create a GSuite module that performs the required operations
C. Create a Cloud Function with the existing code triggered by change notifications from Google Cloud Storage
D. Develop an app, package it into a Container and deploy with App Engine Flexible Edition that periodically reads the bucket, processes the files, and transfers them to a new bucket

A

Correct Answer: C IAP- Identity-Aware Proxy is a service that lets you use SSH and RDP on your GCP VMs from the public internet, wrapping traffic in HTTPS and validating user access with IAM. Inside GCP there is a Proxy server with a listener that translates the communication and lets you operate in a safe way without the public exposure of your GCP resources. A is wrong because a Bastion Host needs a Public IP, so it is not feasible. B is wrong because a Nat Instance needs a Public IP, too. In addition, it is aimed at outgoing connectivity to the internet, blocking inbound traffic, thus preventing exactly what we need. D is wrong because Security Command Center is a reporting service for security that offers monitoring against vulnerabilities and threats. For any further detail: https://cloud.google.com/iap/docs/tcp-forwarding-overview https://cloud.google.com/security-command-center

203
Q

You are working as a Solutions Architect for a Startup company. They want to migrate an existing on-premise application to Google Cloud with minor or no modifications. Which type of migration would you choose for this requirement?
A. Improve and Move
B. Rip and Replace
C. Lift and Shift
D. Shift to SaaS

A

Correct Answer: C It is a simple and quick solution that meets all the requirements. The other solutions are much more complex and not so easy to manage and maintain. The GSuite module (solution B) requires to partially rewrite the routine. For any further detail, please refer to the URLs below: https://cloud.google.com/functions/docs/calling/storage https://cloud.google.com/functions/docs/tutorials/storage For more complex scenarios, it is also possible to use Pub/Sub Notifications for Cloud Storage. These solutions are suitable when there are multiple activities that arise after the change of an Object. For more details on Pub/Sub notifications, check the following URLs: https://cloud.google.com/functions/docs/calling/pubsub https://cloud.google.com/storage/docs/pubsub-notifications

204
Q

For this question, refer to the Mountkirk Games case study. Mountkirk Games is building a new multiplayer game that they expect to be very popular. They want to be able to improve every aspect of the game and the infrastructure. To do this, they plan to create a system for telemetry analysis. They want to minimize effort, maximize flexibility, and ease of maintenance. They also want to be able to perform real-time analyses. Which of the following services may help to fulfill these requirements?
A. Pub/Sub and Big Table
B. Kubeflow
C. Pub/Sub, Dataflow and BigQuery
D. Pub/Sub and Cloud Spanner

A

Correct Answer: C Lift and Shift is the best approach because the company wants to migrate the existing application to google cloud with minor modifications and refactoring. In lift and shift migration you migrate the application as it is from on-premise to Google Cloud with very fewer modifications and it is the fastest way of migrating an application to Google Cloud https://cloud.google.com/solutions/migration-to-gcp-getting-started
Option A is incorrect because this approach is used when you need to modernize the application while migrating to Google Cloud
Option B is incorrect because it used when you have to totally redesign the application while migrating
Option D is incorrect because there is no such type of migration as per Google’s Documentation

205
Q

For this question, refer to the Mountkirk Games case study. MountKirk Games uses Kubernetes and Google Kubernetes Engine and wants to Use GPU processing to render graphics server-side. But GPUs are expensive so they are planning to use preemptible GPU nodes. What do they have to perform for the above solution?
A. Create a node pool with preemptible VMs
B. Create a node pool with non-preemptible VMs
C. Remove the taint nvidia.com/gpu=present:NoSchedule
D. Let System Pods run with non-preemptibleVMs

A

Correct Answer: C Pub/Sub ingests and stores these messages, both from the user devices or the Game Server. Dataflow can transform data in schema-based and process it in real-time BigQuery will perform analytics. A is wrong because Big Table is not the service for real-time analytics B is wrong because Kubeflow is used for Machine Learning pipelines. D is wrong because Cloud Spanner is a global SQL Database and not an analytics tool. For any further detail: https://cloud.google.com/solutions/mobile/mobile-gaming-analysis-telemetry

206
Q

You work in a multinational company that is migrating to Google Cloud. The head office has the largest data center and manages a connection network to offices in various countries around the world. Each country has its own projects to manage the specific procedures of each location, but the management wants to create an integrated organization while maintaining the independence of the projects for the various branches. How do you plan to organize Networking?
A. Peered VPC
B. Cloud Interconnect
C. Shared VPC
D. Cloud VPN and Cloud Router

A

Correct Answer: C System Pods have to run with non-preemptible VMs, otherwise, it would be dangerous when the node gets removed. So you have to avoid having only node pools with GPU preemptible VMs. In this case, the taint nvidia.com/gpu=present:NoSchedule should be removed. And it is OK to have at least a node pool with non-preemptible VMs, in addition to GPU preemptible VMs.
Option A is wrong because it is OK to have a node pool with preemptible VMs when you have at least a node pool with non-preemptible VMs.
Option B is wrong because it is OK to use GPU preemptible VMs.
Option D is wrong because it would be dangerous when System Pods will get removed. For any further detail: https://cloud.google.com/kubernetes-engine/docs/how-to/preemptible-vms#gpu_preemptible_node_taints

207
Q

For this question, refer to the Mountkirk Games case study. MountKirk Games uses Kubernetes and Google Kubernetes Engine. For the management, it is important to use an open platform, cloud-native, and without vendor lock-ins. But they also need to use advanced APIs of GCP services and want to do it securely using standard methodologies, following Google-recommended practices but above all efficiently with maximum security. Which of the following solutions would you recommend?
A. API keys
B. Service Accounts
C. Workload identity
D. Workload identity federation

A

Correct Answer: C The headquarters office manages the global network so the networking specialists mainly work over there. Shared VPC lets create a single, global VPC organized by a central project (host project). All the other projects (service projects) maintain their independence but they don’t have the burden of network management. So we can have a balance between control policies at the network level and freedom to manage application projects A is wrong because with VPC peering there is no organization hierarchy. B is wrong because Cloud Interconnect is for on-premises networking. D is wrong because Cloud VPN and Cloud Router are used for Cloud and on-premises telecommunications. For any further detail: https://cloud.google.com/vpc/docs/shared-vpc https://cloud.google.com/architecture/best-practices-vpc-design#shared-vpc

208
Q

Helicopter Racing League (HRL) wants to migrate their existing cloud service to a new platform with solutions that allow them to use and analyze video of the races both in real-time and recorded for broadcasting, on-demand archive, forecasts, and deeper insights. There is a need to transform media into different streaming formats so as to optimize them for the various platforms. Which of the following strategies do you think is the best?
A. Use Dataflow and Pub/Sub
B. Use Elemental Medialive
C. Use Transcoder API
D. Use Dataproc

A

Correct Answer: C The preferred way to access services in a secured and authorized way is with Kubernetes service accounts, which are not the same as GCP service accounts. With Workload Identity, you can configure a Kubernetes service account so that workloads will automatically authenticate as the corresponding Google service account when accessing GCP APIs. Moreover, Workload Identity is the recommended way for applications in GKE to securely access GCP APIs because it lets you manage identities and authorization in a standard, secure and easy way. A is wrong because API keys offer minimal security and no authorization, just identification. B is wrong because GCP Service Accounts are GCP proprietary. Kubernetes is open and works with Kubernetes service accounts. D is wrong because Workload identity federation is useful when you have an external identity provider such as Amazon Web Services (AWS), Azure Active Directory (AD), or an OIDC-compatible provider. For any further detail: https://cloud.google.com/kubernetes-engine/docs/how-to/workload-identity https://cloud.google.com/kubernetes-engine/docs/how-to/workload-identity https://cloud.google.com/docs/authentication

209
Q

Your company’s architecture is shown in the diagram. You want to automatically and simultaneously deploy new code to each Google Container Engine cluster. Which method should you use?
A. Use an automation tool, such as Jenkins.
B. Change the clusters to activate federated mode.
C. Use Parallel SSH with Google Cloud Shell and kubectl.
D. Use Google Cloud Build to publish the new images

A

Correct Answer: C The Transcoder API can: create consumer streaming formats, like MPEG-4 (MP4), Dynamic Adaptive Streaming over HTTP (DASH, also known as MPEG-DASH), and HTTP Live Streaming (HLS). encode or decode a digital data stream with multiple techniques (codecs) Partionize video files for fast encoding decoding A is wrong because Dataflow could transform a stream of data in another format but you had to provide all the coding to realize it . B is wrong because Elemental medialive is a broadcasting service from Amazon D is wrong because Dataproc you should use Hadoop streaming and develop all the code for the solution. For any further detail: https://cloud.google.com/transcoder/docs/concepts/overview

210
Q

You’re reviewing an application that sometimes executes some SQL queries with unacceptable response times. You need to find a way to solve the problem and identify the causes. Which of the following methods would you suggest?
A. Extract the queries from the application code and try to identify the slowest ones
B. Use Cloud Debugger
C. Use Cloud Trace
D. Use Operations Suite (formerly Stackdriver) Logs and set up a metric
E. Look at the Operations Suite (formerly Stackdriver) Error Reporting dashboard

A

Correct Answer: D
Option A is incorrect - Since we have a managed service and a native solution in the option, it is preferred to pick that option.
Option B is incorrect. Federated mode allows for deployment in a federated way but does not do anything beyond that, you still have to have a tool like Jenkins to enable the “automated “ part of the question, and with Jenkins you can accomplish the goal without necessarily needing federation to be enabled.
Option C is incorrect. This may work in very simple examples, but as complexity grows this will become unmanageable.
Option D is correct. You can automate the deployment of your application to GKE by creating a trigger in Cloud Build. You can configure triggers to build and deploy images whenever you push changes to your code. https://cloud.google.com/build/docs/deploying-builds/deploy-gke

211
Q

Your company is using Bigquery for data analysis, many users have access to this service and the data set, you would want to know which user has run what query, what would be the best way to get the required information?
A. Go to job history, it has information about which user has run what query.
B. Query the Operations Suite (formerly Stackdriver) logs
C. Check the Audit logs for the user ID.
D. Go to Query history it has information about which user has run what query.

A

Correct Answer: D
Option A is incorrect. It is a manual operation, long and heavy, and the problem may be related to other elements and factors related to the application
Option B is incorrect. Cloud Debugger checks the code, and you already know the queries with problems
Option C is incorrect. Cloud Trace collects latency data from your applications and displays it in the Google Cloud Platform Console. You already know that this kind of problem exists.
Option D is correct. You can set a metric that accurately identifies the log lines related to queries. You can also create an alert that can promptly alert you when the problem is displayed, so you can review all the related logs and information at the right time.
Option E is incorrect. There are no errors, so this is a useless operation. For more details, please refer to the following URLs: https://cloud.google.com/logging/ https://cloud.google.com/logging/docs/logs-based-metrics/

212
Q

You have to quickly deploy an update to your scalable app operating with managed instance groups but you definitely know that there may be some errors. You want to test the new app in production before fully committing to the update. Which is the best solution?
A. Change Template and everything will be automatic
B. Change Template, then start new instances and stop the old ones
C. Change Template and ask for a Rolling update
D. Change Template and ask for a Canary update

A

Correct Answer: D
Option D is correct because Query history has the required information.
Option A is incorrect because it has information about export, load, copy etc. Options B and C are incorrect because they provide logs for user-specific actions on a high-level basis. They don’t provide query level logs specific to a user.

213
Q

You have the task of deploying a stateless Node.js already containerized application with a scalable configuration so that the start of a new instance is fast and easy to update. Additional requirements for the instances are: Ubuntu-1804-lts Nodejs 10 Free from boot or kernel-level malware or rootkits. Which is the best solution?
A. Instance Json Template with startup script
B. Instance custom image Template
C. Instance public image Template with startup script
D. Instance shielded public image Template with startup script

A

Correct Answer: D A is wrong. Instance Template are immutable, so you have to create a new Instance Template and update the Managed Group Definition B is wrong. It is not advisable to do such a manual operation. It is cumbersome and prone to errors C is wrong. With managed instance group updater, you may roll out all an update automatically. The question required a partial update. D is correct. A canary update is a partial update to a few numbers of instances in the instance group. You may view with console or gcloud for the currentAction being performed on each instance in your group, as well as the status of each instance. You may roll back quickly to the previous version and control the speed of an update with the parameters: minReadySeconds (wait for the next replacement), Enable health checking (wait until healthy), tune maxUnavailable (number of Instances beyond the targetSize of the group), and maxSurge (number of instances unavailable at any time during the update) A canary update is an update that is applied to a partial number of instances in the instance group. Canary updates let you test new features or upgrades on a subset of instances, instead of rolling out a potentially disruptive update to all your instances. If an update is not going well, you only need to roll back a small number of instances, minimizing the disruption for your users. From the perspective of the server, a canary update is the same as a standard rolling update, except that the number of instances that should be updated is less than the total size of the instance group. Like a standard rolling update, a canary update is disruptive to the instances affected; that is, the affected instances are deleted and replaced by new VM instances during the update. For more details, please visit the following URLs: https://cloud.google.com/compute/docs/instance-groups/ https://cloud.google.com/compute/docs/instance-groups/rolling-out-updates-to-managed-instance-groups

214
Q

A large financial company wants to collect and analyze (both in real-time and in batches) a huge amount of stock quotes and historical movements. This data must be processed as follows: It is necessary to calculate a complete set of statistical parameters, even in streaming mode A series of forecasting models, currently under development, is going to be set up and backtested The response times have to be, for every type of operation, in milliseconds and with linear scalability. Which of the following product groups would you recommend?
A. Pub/sub and Cloud Spanner
B. Cloud Dataproc
C. Composer, Cloud Dataprep, and Cloud Dataview
D. Pub/Sub, Cloud Dataflow, and BigTable
E. Cloud Task and Cloud Datastore

A

Correct Answer: D A is wrong. It is not the fastest way (images are quicker) and don’t malware protection B is wrong. It is the fastest but it don’t have malware protection C is wrong. It doesn’t have malware protection D is correct The key requirement is that is needed for the Shielded VM support, so you are compelled to use a Public Image. See https://cloud.google.com/compute/docs/images Instance template may use a public image and a startup script to prepare the instance after it starts running. Custom images are more deterministic and start more quickly than instances with startup scripts. However, startup scripts are more flexible and allow you to update the applications and settings in your instances more easily. For more details, please refer to the URLs below: https://cloud.google.com/compute/docs/instance-templates/create-instance-templates https://cloud.google.com/compute/docs/instances/create-vm-from-instance-template https://cloud.google.com/security/shielded-cloud/shielded-vm

215
Q

Your company wants to migrate to the GCP Cloud an on-premise J2EE application that uses a Postgres SQL Database. The solution is required to be: Economic Flexible Fully managed Highly scalable Platform independent Which of the following solution do you think is the most suitable?
A. GCE with Lift and Shift
B. Application Layer and Database Layer with Docker Containers
C. GKE for Layer and Cloud SQL for PostgreSQL
D. App Engine Flexible Edition and Cloud SQL for PostgreSQL

A

Correct Answer: D A is wrong. Pub/Sub is correct but Cloud Spanner is a global SQL Database with outstanding integrity and consistency, but don’t have milliseconds performances. We don’t need (and want to pay) all these features. B is wrong. Dataproc is the Hadoop solution in GCP. It doesn’t really solve the real-time requirement that hasn’t milliseconds performances. C is wrong. Cloud Dataprep is a tool for Data correction and refining (not required in the question). D is correct. It is the only solution that meets all requirements. E is wrong. Cloud Datastore is a performant NoSQL Database, inexpensive but not suitable for Big Data and realtime processing. Cloud Pub/Sub is a serverless product for stream analytics and event-driven computing. You can send and receive messages between independent applications and transmit data across projects and applications running on the cloud, on-premise, or hybrid environments. Cloud Pub/Sub is perfect to decouple systems and components hosted on GCP or elsewhere on the internet. It provides “at least once” delivery at low latency with on-demand scaling to tens of millions of messages per second. With Cloud Pub/Sub, data engineers can: Scale without provisioning, partitioning, or load isolation Expand applications and pipelines to new regions Enrich, deduplicate, order, aggregate, and land events using Cloud Dataflow Mix real-time and batch processing via Cloud Pub/Sub’s durable storage Cloud Dataflow is a fully managed service for transforming and enriching data in real-time and batch stream. Cloud Dataflow has a serverless approach that saves money because you only pay for what you use. Plus, Cloud Dataflow not only works with Google’s ingestion, data warehousing, and machine learning products but also third-party tools like Apache Spark and Apache Beam. https://cloud.google.com/dataflow/ Cloud Bigtable is a NoSQL database service for use cases where low latency reads and high throughput writes, scalability, and reliability are critical. Main features: Now it is global (used to be regional) It offers consistent sub-10ms latency It is ideal for Ad Tech, Fintech, and IoT It offers a storage engine for machine learning applications It provides easy integration with open-source big data tools For any further detail, please visit the following URLs: https://cloud.google.com/pubsub/ https://cloud.google.com/dataflow/ https://cloud.google.com/bigtable/

216
Q

You work for a multinational company and are migrating an Oracle database to a multi-region Spanner cluster. You have to plan the migration activities and the DBAs have told you that the migration will be almost immediate because no non-standard ISO / IEC features or stored procedures are used. But you know that there is an element that will necessarily require some maintenance work. Which is this element?
A. You need to drop the secondary indexes
B. You have to change most of the primary keys
C. You need to manage table partitions
D. You have to change the schema design of many tables

A

Correct Answer: D App Engine Flexible Edition is a PaaS solution without any constraints regarding the technologies adopted, as long as you can package the app into a Container. So it is the only solution that meets all the requirements. A is wrong because it is not fully managed, scalable and flexible B is wrong especially because Containers are not advisable with Database Systems. SQL Databases cannot scale simply starting a new Instance. C is wrong because it is not fully managed. For any further detail, please refer to the URLs below: https://cloud.google.com/appengine/docs/flexible/java/ https://cloud.google.com/sql/docs/postgres/connect-app-engine https://vsupalov.com/database-in-docker/

217
Q

You want to make a copy of a production Linux virtual machine in the US-Central region. You want to manage and replace the copy easily if there are changes in the production virtual machine. You will deploy the copy as a new instance in a different project in the US-East region. What steps must you take?
A. Use the Linux dd and netcat commands to copy and stream the root disk contents to a new virtual machine instance in the US-East region.
B. Create a snapshot of the root disk and select the snapshot as the root disk when you create a new virtual machine instance in the US-East region.
C. Create an image file from the root disk with Linux dd command, create a new disk from the image file, and use it to create a new virtual machine instance in the US-East region.
D. Create a snapshot of the root disk, create an image file in Google Cloud Storage from the snapshot, and create a new virtual machine instance in the US-East region using the image file for the root disk.

A

Correct Answer: D As mentioned in the official documentation, the schema data types vary from Oracle to Spanner. For instance, the usage of VARCHAR column should be converted into Spanner’s STRING equivalent. The same applies for INT, INTEGER, etc (to be converted into INT64); the same for DATETIM
E -> TIMESTAMP. This basically means we need to convert the table schemas in any case. Therefore, answer D is valid.
Option A is incorrect because Spanner handles secondary indexes.
Option C is incorrect because Spanner automatically manages the distribution of data in the clusters.
Option B is incorrect because there is no insight about the type of primary keys being used in Oracle schema: nowhere it is written the DB is using sequences. The source Oracle DB could be using GUIDs which will work well. https://cloud.google.com/spanner/docs/migrating-oracle-to-cloud-spanner#supported-data For any further detail: https://cloud.google.com/spanner/docs/schema-design#primary-key-prevent-hotspots

218
Q

You are a Cloud Architect in a medium-sized company. You have been summoned by the General Manager who has asked you to encrypt all data stored in Google Cloud and to put the encryption key in a safe place in the most convenient and cost effective way. You started laughing and you said it’s already being taken care of and you explained how. Which of these is the correct explanation?
A. Google doesn’t manage data encryption and let the customers use Customer-managed encryption keys (CMEK)
B. Google automatically encrypts the files with a data encryption key (DEK) using SHA-256
C. Google automatically encrypts data with DEK and KEK using RSA-256
D. Google automatically encrypts data with DEK and KEK using KMS and AES256/AES128

A

Correct Answer: D Explanation: D (Correct Answer) - This approach meets all of the requirements, it is easy to do and works cross project and cross region.
A - This approach affects the performance of the existing machine and incurs significate network costs.
B - We can share the snapshots of boot dist across the project and region but cannot create an instance using directly the snapshots.
C - dd will not work correctly on a mounted disk. Reference Resources https://cloud.google.com/compute/docs/machine-images#when-to-use

219
Q

You have to create several preemptible Linux virtual machine instances using Google Compute Engine. You want to properly shut down your application before the virtual machines are preempted. What should you do?
A. Create a shutdown script registered as a xinetd service in Linux and configure an Operations Suite (formerly Stackdriver) endpoint check to call the service.
B. Create a shutdown script, registered as a xinetd service in Linux, and use the gcloud compute instances add-metadata command to specify the service URL as the value for a new metadata entry with the key shutdown-script-url
C. Create a shutdown script named k99.shutdown in the /etc/rc.6.d/ directory.
D. Create a shutdown script and use it as the value for a new metadata entry with the key shutdown-script in the Cloud Platform Console when you create the new virtual machine instance.

A

Correct Answer: D Google, by default, encrypts each chunk of data with a data encryption key DEK using AES256/AES128, symmetric cryptography. DEKs are sent to KMS (the service for Key Management) encrypted with a key-encryption key KEK, and the wrapped DEKs are stored with the data chunks. KEKs are kept in KMS and are not exportable; so, all encryption and decryption must be done within KMS. KEKs are rotated periodically and automatically. A is wrong because in GCP all the storage is encrypted. B is wrong because SHA-256 is a hash algorithm, one way, no feasible. C is wrong because RSA-256 is asymmetric cryptography, not used for data encryption because it would be too slow. For any further detail: https://cloud.google.com/security/encryption/default-encryption/

220
Q

Your company has developed IT Systems deployed on Kubernetes and GKE. The company has also recently acquired a new company and wants to integrate its applications and services. Fortunately, the new company also uses Kubernetes but on Microsoft Azure. The requirement is to quickly integrate the two Cloud platforms and the on-premises organization, having in this way, the capability to use any type of service. The solution must be seamless, quick, safe, standard, and fast, without heavy migrations or network updates. You are asked to identify the best solution among those proposed.
A. It is not possible to meet all the requirements
B. Develop an interface between the various systems that work on-premise
C. Use Cloud Endpoints to create a set of services that encapsulate external ones and use JWT
D. Use Service Mesh
E. Adopt a proprietary solution from Microsoft

A

Correct Answer: D Running Shutdown Scripts: Create and run shutdown scripts that execute commands right before an instance is terminated or restarted, on a best-effort basis. This is useful if you rely on automated scripts to start up and shut down instances, allowing instances time to clean up or perform tasks, such as exporting logs or syncing with other systems. https://cloud.google.com/compute/docs/shutdownscript To setup Shutdown Scripts, go to the GCP console and follow the steps: Compute Engine -> VM instance -> Create Instance -> (Expand) Management, disks, networking, SSH keys Enter the key “shutdown-script” and the proper value.

221
Q

Rules must be set to allow data traffic to database servers only from application servers, in 3 different projects: A, B, and C. The resources of the 3 projects must be isolated from each other. You want to organize operations in order to create simple and intuitive standards to use, which can be repeated for other projects. In your organization, it is mandatory to provide different security to the various projects. Which of the following strategies is the best?
A. Create 2 Firewall Rules, one in ingress and one in egress, between each Database Server and App Server using the ephemeral external IP address
B. Create 1 Firewall Rule in ingress, between each Database Server and App Server using private IP addresses
C. Configure your Servers with appropriate Network Tags (AppVM and DBVM, for example) and create 1 Firewall Rule, in ingress, between each Database Server and App Server using these Tags
D. Configure your Servers with appropriate Network Tags (AppVM and DBVM, for example) and create 2 Firewall Rules, in ingress and egress, between each Database Server and App Server using these Tags
E. Create and assign appropriate Service Accounts and rights to the VMs and create a Firewall Rule between each Database Server and App Server using source-service-accounts and target-service-accounts

A

Correct Answer: D The perfect solution is Service Mesh. Anthos Service Mesh is Google’s implementation of the powerful Istio open-source project, allowing you to manage, observe, and secure your services without having to change your application code. Options A and B are incorrect. It is not possible to develop quickly an internal solution that meets these requirements.
Option C is incorrect because Endpoint creates APIs and doesn’t integrate all kinds of services. https://cloud.google.com/endpoints/
Option E is incorrect because it is not standard and cannot integrate all kinds of services. For any further detail, please follow the links below: https://cloud.google.com/service-mesh/docs/unified-install/multi-cloud-hybrid-mesh

222
Q

For this question, refer to the EHR Healthcare case study. EHR Healthcare requested you, to create a strategy for the progressive and secure migration of most of its workloads to the Cloud. As a Cloud Architect, you have decided to rely on Google Cloud and want to use the services managed by GCP as much as possible to maximize reliability and security while maintaining full control of the processes. In particular, he wants his systems to be implemented with automatic procedures both in the Cloud and on-premise with automatic procedures and without the development team having to manage the code differently depending on the systems in which they will be installed. In this way, a consistent way to manage applications will be provided. Which of the following strategies do you think is the best?
A. Design CI/CD pipelines for container-based workloads using Spinnaker for VMs on-premises and on Compute Engine
B. Design CI/CD pipelines for container-based workloads using Cloud Build and Cloud Run
C. Design CI/CD pipelines for container-based workloads using Cloud Build for VMs on-premises and on Compute Engine
D. Design CI/CD pipelines for container-based workloads using Spinnaker for VMs on-premises and on AppEngine
E. Design CI/CD pipelines for container-based workloads using Spinnaker for GKE and on-premises Kubernetes clusters

A

Correct Answer: E GCP firewall rules are stateful. If a connection is allowed between a source and a target, all subsequent traffic in either direction will be allowed as long as the connection is active. In other words, firewall rules allow bidirectional communication once a session is established. The connection is considered active if at least one packet is sent every 10 minutes. Firewall rules cannot allow traffic in one direction while denying the associated return traffic. So, A and D are wrong. A service account represents an identity associated with an instance. Only one service account can be associated with an instance. So it is the best option in case of strict security constraints. Be careful because you cannot mix and match service accounts and network tags in any firewall rules. C is wrong because it is necessary to provide different security to various projects. So network tags are the arbitrary attributes and are not enough for this requirement. For any further detail, please refer to the URLs below: https://cloud.google.com/vpc/docs/using-firewalls https://cloud.google.com/vpc/docs/firewalls#service-accounts-vs-tags

223
Q

Your company has collected all its historical data in BigQuery and Documents in Cloud Storage. It has just been completed, the organization of structured Data with your help; now it is time to cope with documents. Many documents originated from papers, so they are image files. Management wants to find a way to quickly classify docs. The actual organization will be perfectioned with use and practice. You are a Cloud Architect with a basic knowledge of Machine Learning. There are no AI experts in the Company. According to the requirements, which tool is the best suited for detecting text from image files?
A. Develop a complete solution with Tensorflow
B. AI Platform Data Labeling Service
C. Vision API
D. BigQuery ML
E. DocumentAI

A

Correct Answer: E The best way is to adopt Kubernetes on-prem and in Cloud. It could be used Anthos on-prem, too, in order to ease the migration. Kubernetes provides the best solution for orchestrating workloads of any kind in any Cloud and on-prem. Cloud Build cannot deploy local workloads so a tool like Spinnaker is the best way to achieve the desired consistency. A is wrong because VMs and Compute Engine are not containers optimized. B is wrong because Cloud Build and Cloud Run are optimal services only in Cloud. They are not suitable for on-premises deployments C is wrong Cloud Build cannot deploy local workloads D is wrong VMs and AppEngine are not containers optimized. For any further detail: https://cloud.google.com/containers https://cloud.google.com/build/docs/overview https://cloud.google.com/kubernetes-engine/docs/concepts/kubernetes-engine-overview https://cloud.google.com/anthos/docs/setup/on-premises

224
Q

You are hired as a consultant for a company and the management wants to migrate their systems to the cloud. The customer is concerned about cost control and asks some questions with respect to cost-effectiveness. Which of the below possibilities are valid? (Choose 2)
A. Is it possible to control your Google Cloud Platform charges?
B. Is it possible to create separate budgets for projects and resources?
C. Is it possible to have notifications?
D. Are the alerts sent, only when the limit is reached?
E. Is there a way to have a programmatic interface?

A

Correct Answer: E The DocumentAI can detect and extract text from images. Document AI uses machine learning and Google Cloud to help you create a scalable, cloud-based document understanding solution. Using Document AI, you can: Convert images to text Classify documents Analyze and extract entities For more details, please visit the following link: https://cloud.google.com/document-ai/docs/overview

225
Q

Your team has created a set of applications that will run in GCP. IT Management wants to activate and standardize a simple but effective security system. You have prepared a list of possibilities and features that you can use. You realize that some choices must be discarded because they are not safe enough or even wrong. What solutions would you, recommend at the end? (Choose 3)
A. Google groups for each category of users with associated roles
B. Service Accounts related to your applications
C. G Suite or Cloud Identity domains with associated roles
D. Service Accounts related to your VMs
E. Service Accounts related to your K8s Clusters

A

Correct Answers - A and B Options A and B are correct as they are valid questions with respect to cost effectiveness. Other options are incorrect, as they are general questions asked based upon curiousity. They are not related to cost or billing. For any further detail, please refer to the URLs below: https://cloud.google.com/billing/docs/how-to/budgets https://cloud.google.com/billing/docs/ https://cloud.google.com/billing/v1/getting-started

226
Q

Which of the following are characteristics of GCP VPC subnets? Choose 2 answers.
A. Each subnet can span over multiple availability zones within a region to provide a high availability environment.
B. Each subnet maps to a single Availability Zone.
C. CIDR block mask of /25 is the smallest range supported.
D. By default, all subnets can route between each other, whether they are private or public.

A

Correct Answers - A, B, D A is correct. As per google’s best practices, this is the best way to manage the access and users. B is correct. You can authorize users via Cloud Identity-Aware Proxy (IAP). They do not require direct access to the underlying GCP resources—just to the web app that utilizes the GCP resources. C is wrong. Feasible, but not advisable. Burdensome and difficult to maintain. D is correct. You create the basic Service Accounts and link them to the VMs. So, users that have access to the VM, inherit all the authorizations needed E is incorrect. Service Accounts are associated with your pods and not clusters. Service Accounts related to your applications Service Accounts related to your VMs For more details, please visit the following links: https://cloud.google.com/iam/docs/overview https://cloud.google.com/kubernetes-engine/docs/tutorials/authenticating-to-cloud-platform https://cloud.google.com/blog/products/identity-security/understanding-gcp-service-accounts-three-common-use-cases

227
Q

You need to regularly create disk-level backups of the root disk of a critical instance. These backups need to be able to be converted into new instances that can be used in different projects. How should you do this? Select the two possible ways to accomplish this.
A. Create snapshots, turn the snapshot into a custom image, and send the image across projects.
B. Use the VM migration tools in Compute Engine to copy a VM to a different project.
C. Create snapshots and send them to other projects.
D. Stream your VM’s data into Cloud Storage and share the exported data in the storage bucket with another project.

A

Correct Answers A and D A (Correct) - Each subnet can span over multiple Availability Zones to provide a high-availability environment. Each VPC network consists of one or more useful IP range partitions called subnetworks or subnets. Each subnet is associated with a region. Networks can contain one or more subnets in any given region. Subnets are regional resources. VPC network example: subnet3 is defined as 10.2.0.0/16, in the us-east1 region. One VM instance in the us-east1-a zone and a second instance in the us-east1-b zone, each receiving an IP addresses from its available range. D (Correct Answer) - By default, all subnets can route between each other, whether they are private or public. Because subnets are regional resources, instances can have their network interfaces associated with any subnet in the same region that contains their zones. Resources within a VPC network can communicate with one another using internal (private) IPv4 addresses, subject to applicable network firewall rules. The default network includes a “default-allow-internal” rule, which permits instance-to-instance communication within the network. C is incorrect: Each subnet defines a range of IP addresses. The minimum CIDR size for a subnet is /29. Reference: https://cloud.google.com/vpc/docs/vpc

228
Q

For this question, refer to the EHR Healthcare case study. EHR Healthcare wants to on-board new insurance providers as quickly as possible. To do this, it must be able to acquire records both in an atomic way, record by record, and through the transmission of files in various formats. So both streaming and batch. He wants to create standard data transformation procedures for loading into his data structures but to make the process as easy and fast as possible. What are the GCP services that can meet this requirement (pick 2) ?
A. Manage data acquisition in Streaming with Pub / Sub and DataFlow
B. Manage the acquisition of streaming data with web applications with a standard http interface
C. Manage the Acquisition of files with signed URL in Cloud Storage and use DataFlow for transformation
D. Sharing of public buckets with a naming convention for the acquisition of files

A

Correct Answers: A and C
B - is for migration not for “regularly creating disk-level backups of the root disk of a critical instance”. There are tools allowing copying (importing) on-premises virtual disk to Compute engine but you cannot copy GCP VM. C(Correct Answer) - Sharing storage resources across projects and organizations You can share access to images, disks, and snapshots using the following IAM roles or permissions: Images: The roles/compute.imageUser role or the compute.images.useReadOnly permission. Snapshots: The roles/compute.storageAdmin role or the compute.snapshots.useReadOnly permission. Disks: The roles/compute.storageAdmin role or the compute.disks.useReadOnly permission. These roles and permissions allow you to share each resource type independently with other team members outside of your projects. For example, your company might have a specific project with qualified images that the rest of your company can use. You can assign a group to the project that is explicitly responsible for creating and maintaining images. Then, you can grant the roles/compute.imageUser role to other team members so that team members can use these images in their own projects. Note: These roles allow users to use your storage resources in any project, including projects outside of your organization. To restrict image use to specific organizations or specific projects, set the ‘constraints/compute.storageResourceUseRestrictions’ constraint as an organization policy. https://cloud.google.com/compute/docs/images/sharing-images-across-projects A (Correct answer) - The proper method is to create a custom image either from an existing, stopped instance, or snapshots of a boot disk, which can then be shared across projects and used to create additional instances. https://cloud.google.com/compute/docs/instances/create-start-instance D- doesn’t meet the requirement “regularly create disk-level backups of the root disk of a critical instance” nor is it easy to convert into the new instance.

229
Q

Your company’s development teams use, as required by internal rules, service accounts. They just forget to delete the service accounts that are no longer used. A coordinator noticed the problem and ordered them to clean up. Now your team is faced with a huge, boring, and potentially dangerous job and has asked you for help. What advice can you give him?
A. Service account insights
B. Cloud Audit Logs
C. Activity Analyzer
D. Flow logs

A

Correct answers: A and C Dataflow can process both batch- and streaming-data parallel pipelines with the same code in a serverless way. It is based on Apache Beam so the procedures may be created as reusable templates in any environment. The dev team cab build programs that define the pipeline. Then, one of Apache Beam’s supported distributed processing backends, such as Dataflow, executes the pipeline. The data processing job supports parallel processing. B is wrong because it requires more developing work and may not sustain heavy traffic. D is wrong because the solution is not secure For any further detail: https://cloud.google.com/dataflow/docs/concepts/beam-programming-model https://cloud.google.com/storage/docs/access-control/signed-urls

230
Q

Case Study: EHR Healthcare EHR Healthcare manages identities with Microsoft Active Directory, which is also integrated into many applications. When deploying the migrated workload, they need the privileges granted by Google Cloud IAM and the identities and permissions from on-premise environments in a seamless way. They must create a simple and workable solution to minimize the transformation load. Which of the following techniques would you follow (pick 2)?
A. Configure Cloud Identity to use Active Directory as LDAP and authoritative source through federation with Google Cloud Directory Sync and Active Directory Federation Services (AD FS)
B. Configure Cloud Identity to use Active Directory as LDAPand authoritative source through federation with Google Cloud with Azure Active Directory
C. Use Cloud Identity and replicate changes to Active Directory with SSO
D. Use Cloud Identity and replicate changes to an LDAPServer compatible with Azure Active Directory

A

Correct Answers: A and C The best way to find out service accounts usage are: Service account insights, that lists service accounts not used in the past 90 days and Activity Analyzer, which reports about service account’s last usages. So they let you control the opposite aspects. B is wrong because Cloud Audit Logs contain audit trials, that is user activity and services modifications in GCP. D is wrong because Flow logs contain only network information to and from VM instances. For any further detail: https://cloud.google.com/iam/docs/manage-service-account-insights https://cloud.google.com/iam/docs/service-account-recent-usage

231
Q

For this question, refer to the TerramEarth case study. TerramEarth needs to migrate legacy monolithic applications into containerized RESTful microservices. The development team is experimenting with the use of packaged procedures with containers in a completely serverless environment, using Cloud Run. Before migrating the existing code into production it was decided to perform a lift and shift of the monolithic application and to develop the new features that are required with serverless microservices. So, they want to carry out a gradual migration, activating the new microservice functionalities while maintaining the monolithic application for all the other activities. The problem now is how to integrate the legacy monolithic application with the new microservices to have a consistent interface and simple management. Which of the following techniques can be used (pick 3)?
A. Use an HTTP(S) Load Balancer
B. Develop a proxy inside the monolithic application for integration
C. Use Cloud Endpoints/Apigee
D. Use Serverless NEGs for integration
E. Use App Engine flexible edition

A

Correct answers: A and C With AD SSO federation relevant users and groups are synchronized to Cloud Identity, but changes in Active Directory are replicated to Google Cloud but not vice versa. SO Active Directory remains the only system that manages these credentials. EHR Healthcare uses Active Directory on-premises but not Active Directory in Azure Cloud (Azure Active Directory). B and D are wrong because EHR Healthcare doesn’t use Azure Active Directory. For any further detail: https://cloud.google.com/architecture/identity/federating-gcp-with-active-directory-introduction

232
Q

Your Customer has been testing Kubernetes for a while and needs a punctual analysis comparing the deployment with instance template and Docker vs Kubernetes. You performed a list of advantages and disadvantages. Which of the following is correct? (Select THREE)
A. Money-Saving because of less computing power required
B. More manual intervention is needed with K8s
C. Allows continuous integration with Cloud Build
D. Difficult to change Cloud Provider, because K8s is a Google product
E. K8s Container-native load balancing is more performant than Instance Group Load Balancing

A

Correct answers: A, C and D The first solution (A+D) uses HTTP(S) Load Balancing and NEGs. Network endpoint groups (NEG) let you design serverless backend endpoints for external HTTP(S) Load Balancing. Serverless NEGs became target proxies and the forwarding is performed with the use of URL maps. In this way, you may integrate seamlessly with the legacy application. An alternative solution is API Management, which creates a facade and integrates different applications. GCP has 3 API Management solutions: Cloud Endpoints, Apigee, and API Gateway. API Gateway is only for serverless back ends. B is wrong because developing a proxy inside the monolithic application for integration means, keep on updating the old app with possible service interruptions and useless toil. E is wrong because App Engine’s flexible edition manages containers but cannot integrate the legacy monolithic application with the new functions. For any further detail: https://cloud.google.com/load-balancing/docs/negs/serverless-neg-concepts https://cloud.google.com/endpoints

233
Q

Your team has created a set of applications that will run in GKE Clusters. IT Management wants to activate and standardize a simple but effective security system. You have prepared a list of possibilities and features that you can use. You realize that some choices must be discarded because they are not safe enough or even wrong. Which solutions would you recommend? (Select 3)
A. In the cluster, the nodes will be assigned on internal RFC 1918 IP addresses only
B. Use Service Accounts and store keys in Configuration Files
C. In the cluster, you have to assign public IP addresses to the Master nodes
D. Use Service Accounts and store the keys as a Kubernetes secret
E. Use Workload Identity

A

Correct Answers: A, C, and E
Option A is correct. Kubernetes can fit more containers/pods in the same VM
Option B is incorrect. Kubernetes can automate more operations and hide the complexity of networking and load balancing
Option C is correct. You may create automation with Cloud Build such as when you commit your source, your production or staging environment will be seamlessly updated.
Option D is incorrect. Kubernetes has been adopted by all the major cloud platforms as the leading Containers Orchestrator
Option E is correct. As you can easily see from the following picture, the container-native load balancer communicates directly with the Pods, connections have fewer network hops, so both latency and throughput are improved. For more details, please visit the following URLs: https://cloud.google.com/kubernetes-engine/ https://cloud.google.com/kubernetes-engine/docs/how-to/container-native-load-balancing

234
Q

Using the principle of least privilege and allowing for maximum automation, what steps can you take to store audit logs for long-term access and to allow access for external auditors to view? (Select TWO)
A. Create account for auditors to have view access to Operations Suite (formerly Stackdriver) Logging.
B. Export audit logs to Cloud Storage via an export sink.
C. Export audit logs to BigQuery via an export sink.
D. Create account for auditors to have view access to export storage bucket with the Storage Object Viewer role.

A

Correct Answers: A, D, E A is correct. It is the most advisable way to build private clusters that can use an HTTP(S), an internal or a network load balancer to accept incoming traffic B is wrong. You have to use Service Accounts, but you cannot expose keys in clear text inside Configuration Files C is wrong. It is always possible to use private clusters, that can use an HTTP(S), an internal or a network load balancer to accept incoming traffic D is correct. It is the basic way to go, but now there is a better way: Workload Identity E is correct. It is the new security method: once you configure the relationship between a Kubernetes service account and a Google service account, any workload running as the Kubernetes service account automatically authenticates as the Google service account while accessing Google Cloud APIs. Workload Identity, is the new way for GKE applications to authenticate and consume other Google Cloud services. It works by creating a relationship between Kubernetes service accounts and Cloud IAM service accounts, so you can use Kubernetes-native concepts to define which workloads run as which identities, and permit your workloads to automatically access other Google Cloud services, all without having to manage Kubernetes secrets or IAM service account keys. For any further detail, please refer to the URLs below: https://cloud.google.com/kubernetes-engine/docs/how-to/workload-identity https://cloud.google.com/kubernetes-engine/docs/concepts/security-overview https://cloud.google.com/kubernetes-engine/docs/tutorials/authenticating-to-cloud-platform

235
Q

Helicopter Racing League (HRL) wants to migrate their existing cloud service to the GCP platform with solutions that allow them to use and analyze video of the races both in real-time and recorded for broadcasting, on-demand archive, forecasts, and deeper insights. During a race filming, how can you manage both live playbacks of the video and live annotations so that they are immediately accessible to users without coding (pick 2)?
A. Use HTTP protocol
B. Use Video Intelligence API Streaming API
C. Use DataFlow
D. Use HLS protocol
E. Use Pub/Sub

A

Correct Answers: B and D
A - Create an account for auditors to have view access to Operations Suite (formerly Stackdriver) Logging - continue storing log information in Operations Suite (formerly Stackdriver) is not a proper solution for “Long term access”. Between B and C, the difference is where to store the logs, BigQuery or Cloud Storage. Since the main concern is extended storing period, B (Correct Answer) is a better answer, and the “Long term access” further qualifies it, for example, using the Coldline storage class. With regards to BigQuery, while it is also a low-cost storage, but the main purpose is for analysis. Also, logs stored in Cloud Storage is easy to transport to BigQuery or do query directly against the files saved in Cloud Storage if and whenever needed. D (Correct Answer) - Create an account for auditors to have view access to the export storage bucket with the Storage Object Viewer role - this completes the answers by providing a view-only role to the auditors

236
Q

For this question, refer to the EHR Healthcare case study. EHR Healthcare needs to set up a general DR policy for all its relation Databases, distributed in all its Data Centers on-premises. The activity is preparatory to the migration to the Cloud, which includes managed services. EHR Healthcare wants to migrate data into managed services in the future without major impact on applications, which will all need to be containerized. Later it will adopt a global DB solution. The DR will be the first step towards migration The requirements are RPO and RTO in less than 1 hour. Which of the following solutions do you think are the best (Select TWO)?
A. Create daily snapshots of the Database and transfer them to Cloud Storage
B. Create MySQL external replica promotion migration into Cloud SQL
C. Create a SQL Server external replica promotion migration into Cloud SQL
D. Save Backup folders of the SQL Server Databases to Cloud Storage with gsutil rsync with hourly update procs to Cloud SQL
E. Save Backup folders of the MySQL Databases to Cloud Storage with gsutil rsync with daily update procs to Cloud SQL

A

Correct Answers: B and D D is correct because HTTP Live Streaming is a technology from Apple for sending live and on‐demand audio and video to a broad range of devices. It supports both live broadcasts and prerecorded content, from storage and CDN. B is correct because Video Intelligence API Streaming API is capable of analyzing and getting important metadata from live media, using the AIStreamer ingestion library. A is wrong because HTTP protocol alone cannot manage live streaming video. C is wrong because Dataflow manages streaming data pipelines but cannot derive metadata from binary data, unless you use customized code. E is wrong because Pub/Sub could ingest metadata, but not analyze and getting labels and other info from videos For any further detail: https://cloud.google.com/video-intelligence/docs/streaming/live-streaming-overview https://cloud.google.com/blog/products/data-analytics/streaming-video-using-cloud-data-platform https://developer.apple.com/streaming/

237
Q

Telemetry Helicopter Racing League (HRL) offers a paid service to stream the races all over the world with live telemetry and predictions throughout each race. Management wants to increase telemetry and create additional insights; they want to start as soon as possible, before the migration in GCP, for all kinds of services, and they want to keep on improving this process. Which of the following is the best solution?
A. Develop a sw module to be deployed inside each Pod in Kubernetes
B. Use open-source tools like Prometheus and gradually integrate with Operations Suite (formerly Stackdriver)
C. Use Operations Suite (formerly Stackdriver) metrics and data from all the services
D. Use Istio telemetry functions
E. Use Cloudwatch metrics and data from all the services

A

Correct Answers: B and D For MySQL Database it is possible to create a Cloud SQL read replica of a local DB. The Cloud SQL read replica is asynchronously synchronized and may be promoted to master DB. An easy and elegant solution. With SQL Server it is not possible; Cloud SQL read replicas for SQL Server are not supported so the traditional way (incremental backups and transaction logs) have to be followed
Option A is incorrect because don’t ensure RPO and RTO requirements
Option C is incorrect because Cloud SQL read replicas for SQL Server are not supported
Option E is incorrect because Cloud SQL read replicas for MySQL are supported For any further detail: https://cloud.google.com/architecture/dr-scenarios-for-data https://cloud.google.com/architecture/disaster-recovery-for-microsoft-sql-server https://cloud.google.com/architecture/migration-to-google-cloud-transferring-your-large-datasets https://cloud.google.com/architecture/migrating-mysql-to-cloudsql-concept#external_replica_promotion_migration

238
Q

Long Term Forecasts Helicopter Racing League (HRL) collects data and wants to update the predictions on the results of the championships. HRL wants to create long-term Forecasts with the data from video collected (on BigQuery) both while taking (first processing) and during streaming for users. On the advice of the Cloud Architects, they decided to use 2 different lines of action: A simple and immediate way to create forecast models with the data already collected The ability and culture to develop highly customized models that are continuously improved with the data that it gradually collects. Which of the following strategies do you think is the best (Select 2)?
A. Use pre-trained and Auto-ML models for the easy way
B. TensorFlow Enterprise and KubeFlow for the customized models
C. Use BigQuery ML for the easy way
D. Kubernetes and Tensorflow

A

Correct answers: B Telemetry includes all control data and metrics from cloud systems. Telemetry needs to be extracted, organized and transmitted from multiple locations to a central point of analysis. We go beyond the scalability and availability of services, because the purpose of this Data is to optimize processes and costs; for example, Telemetry is used for the security and integrity of applications, to improve the user experience, to maximize performance and, more generally, the quality of systems. Prometheus is a widely used open-source (Cloud Native Computing Foundation graduated) for the acquisition, integration, query and analysis of telemetry data. Its main features are: All data is collected into an easy to use the multidimensional data model Uses a flexible query language called PromQL Flexible System Management and, security and no overhead graphing and dashboarding support, even if often used with Kabana, another open tool GCP has new features for Service and Telemetry reporting that may be integrated after migration A is wrong because this is not an easy and fast solution, even if it is the Google internal solution; Pods in Google are often made by the app container plus the control container that deals with telemetry. C is wrong because Operation Suite (formerly Stackdriver) doesn’t cover all the needs for a wide-range telemetry D is wrong because Istio is a service mesh for Kubernetes; so, it is aimed at microservices architectures. It is open-source and offers important telemetry functions, but it doesn’t cover all of our requirements. E is wrong because Cloudwatch is the Operation Suite (formerly Stackdriver) counterpart product in AWS For any further detail: https://prometheus.io/docs/introduction/overview/ https://cloud.google.com/service-infrastructure/docs/telemetry-reporting https://istio.io/latest/about/service-mesh/

239
Q

You are attending a tech AWS meetup and the speaker is talking about Cloud Storage features. You, as an experienced GCP Cloud Architect, notice that there are some inaccuracies. Please choose the correct statements (Select 3).
A. Cloud Storage Object can be updated or appended
B. It has Global consistency
C. There isan archive and cheap Solution like AWS Glacier
D. Buckets may have a Retention Policy and objects can be versioned

A

Correct answers: B, C All the answers are correct, but the best solutions are: BigQuery ML for the easy way, because data is already in BigQuery and it can host many kinds of models, even custom TensorFlow and Auto ML TensorFlow Enterprise and KubeFlow are tailored by design for the creation of an MLOps environment in Google Cloud for developing, training, and continuously improve ML Models A is incorrect because BigQuery has all the capabilities without having to move or transform data D is incorrect because Kubernetes and Tensorflow are the foundation of the optimal solution, without the need to configure and create pipelines and automatic methods for process optimization. Kubeflow and TensorFlow Enterprise are already optimized for this purpose. For any further detail: https://cloud.google.com/bigquery-ml/docs https://www.kubeflow.org/ https://cloud.google.com/tensorflow-enterprise https://www.tensorflow.org/tfx

240
Q

You have been asked to set up a Disaster Recovery solution for a non-critical Database Server with multiple disks. The application can be stopped for hours without creating major issues. The data must be recovered from the beginning of the previous day. The solution must be simple and inexpensive. What would you advise?
A. Create a Custom Image of the Instance root disk and use it for re-creating the VM in another region
B. Set the autodelete flag to false for the persistent disks and make regular snapshots
C. Save the daily snapshots to Cloud Storage
D. Use Zonal SSD persistent disks
E. Regional SSD persistent disks, and daily snapshots stored to Cloud Storage

A

Correct Answers: B, C, D A is wrong: Cloud Storage Object cannot be updated or appended. Objects are immutable, you cannot make incremental changes to objects, such as append operations or truncate operations. You may overwrite objects, so incremental updates can be achieved by rewriting an object with the desired updates. B is correct: It has Global consistency. It is different from AWS because it has full consistency for all these operations: Read-after-write, Read-after-metadata-update, Read-after-delete, Bucket listing, Object listing and Granting access to resources C is correct: There is an archive and cheap Solution like AWS Glacier. There is the Archival Storage, that is available within milliseconds. It is the low-cost, highly durable storage service for data archiving, online backup, and disaster recovery. D is correct: Objects may have a Retention Policy and can be versioned. A retention period can be placed on a bucket. An object in the bucket cannot be deleted or overwritten until it reaches the specified age. Object Versioning can be enabled on a bucket in order to retain older versions of objects when they are deleted or overwritten. Object Versioning increases storage costs, but this can be partially mitigated by configuring Object Lifecycle Management to delete older object versions. For any further detail, please visit the following URLs: https://cloud.google.com/storage/docs/concepts https://cloud.google.com/storage/docs/best-practices https://cloud.google.com/blog/products/storage-data-transfer/hdfs-vs-cloud-storage-pros-cons-and-migration-tips

241
Q

Helicopter Racing League (HRL) wants that video footage will be stored in optimal formats for any device/encoding and accessible in a fast and fluid way. To achieve this, access to videos must have minimal latency and be as close as possible to users anywhere in the world. Also, the latest and most requested videos have to be cached. For the first set of requirements, HRL adopted the Transcoder API. Which are the additional services to be set up (pick 2)?
A. Apigee Edge
B. Vertex
C. Cloud global Load Balancing
D. Cloud Content Delivery Network (CDN)
E. Cloud Endpoints

A

Correct Answers: C All answers are a partial solution with the exception of option D and E, which is incorrect because Zonal SSD persistent disks are useless in the event of a disaster. Furthermore, regional disks may not be available, but in this case, all the snapshots in Cloud Storage are preserved. Remember that snapshots are incremental and compressed, so they are fast and inexpensive. Check the following link to get the complete procedure of this solution: https://cloud.google.com/solutions/dr-scenarios-for-data

242
Q

Reduce viewer latency and set content closer to the users, even in emerging regions Helicopter Racing League (HRL) wants that video footage will be both stored in optimal formats for any device/encoding and accessible in a fast and fluid way. To achieve this, access to videos must have minimal latency and be as close as possible to users anywhere in the world. Also, the latest and most requested videos have to be cached. For the first set of requirements, HRL adopted the Transcoder API. Which are the additional services to be set up (Select TWO)?
A. Apigee Edge
B. Vertex
C. Cloud global Load Balancing
D. Cloud Content Delivery Network (CDN)
E. Cloud Endpoints

A

Correct Answers: C and D Google Cloud global Load Balancing gives a single anycast IP for the users anywhere in the world. This IP address will be declined in the best network path to the Google Edge Network around the world, closest to the users. This is an important feature of the powerful Google Network. The Edge Network is the link between ISP (Internet Providers) and the Google Network and hosts also the CDN services that cache and accelerate static contents. Google Cloud global Load Balancing and CDN are designed to work together for these goals. A is wrong because Apigee Edge is a powerful and enterprise API (Application Programming Interface). B is wrong because Vertex is a complete cloud solution for AI and ML E is wrong becauseCloud Endpoints is also an API (Application Programming Interface), and so it is designed for completely different scopes. For any further detail: https://cloud.google.com/cdn/docs/overview https://cloud.google.com/load-balancing https://peering.google.com/

243
Q

For this question, refer to the EHR Healthcare case study. EHR is carrying out automatic DevOps methodologies for the migration, maintenance and development of its apps. Considering that it uses a hybrid environment and the applications are maintained by both internal and external international teams, its concern is to ensure that these processes take place in the safest way. In particular, they are worried to preserve the integrity of systems and avoid weaknesses that could allow attacks or data leaks. Development teams that use CI/CD methods have privileges according with the Least privilege and operate according to the principles of Zero Trust Security. Which of these tools are the best for these requirements (choose 2)?
A. Cloud DLP
B. Forseti
C. Vulnerability scanning
D. Event Threat Detection
E. Binary Authorization

A

Correct Answers: C and D Google Cloud Global Load Balancing gives a single anycast IP for users anywhere in the world. This IP address will be in the best network path to the Google Edge Network around the world, closest to the users. This is an important feature of the powerful Google Network. The Edge Network is the link between ISP (Internet Providers) and the Google Network and hosts also the CDN services that cache and accelerate static contents. Google Cloud Global Load Balancing and CDN are designed to work together for these goals. A is wrong because Apigee Edge is a powerful enterprise API (Application Programming Interface). B is wrong because Vertex is a complete cloud solution for AI and ML E is wrong because cloud Endpoints is also an API (Application Programming Interface), and so it is designed for completely different scopes. For any further detail: https://cloud.google.com/cdn/docs/overview https://cloud.google.com/load-balancing https://peering.google.com/

244
Q

You are working as a Consultant for a large IoT company which is planning to use Bigtable for their Production workload. The QA is deploying a tool for testing the scalability of the application which depends on Bigtable. What will you suggest to the QA team while testing scalability and performance against Bigtable? Select any three.
A. Ensure the load test are performed with enough data
B. Ensure heavy pre-test are performed for several minutes
C. There is no need to perform any testing against Bigtable as it a fully managed service
D. Ensure tests are running for at least 10 mins
E. Ensure that all third-party services used by the application are tested

A

Correct answers: C and E All the solutions indicated are useful for preserving and monitoring the security level of your environment. The question was about CI / CD processes, so the relevant techniques are: Vulnerability Scanning works in container images submitted to Artifact Registry repositories. As soon as a reference, for example, is classified as unsafe, an alarm is raised. Binary authorization is a control for container images deployed on Google Kubernetes Engine (GKE); they must be signed by trusted authorities to ensure their sources. A It is wrong because DLP is and is related to data content and privacy and is not connected to software development and maintenance B It’s wrong because Forseti Security is a community-driven, open source security set of tools, unrelated to CI / CD D It is wrong because Event Threat Detection is part of the Security Command Center that reads the logs and, with ML techniques, detects possible threats and dangerous situations For any further detail: https://cloud.google.com/binary-authorization/docs/overview https://cloud.google.com/security-command-center/docs/concepts-event-threat-detection-overview https://cloud.google.com/artifact-registry

245
Q

For this question, refer to the Dress4Win case study: https://cloud.google.com/certification/guides/cloud-architect/casestudy-dress4win-rev2 Dress4Win is planning to expand its business in the European region. Which regulation would you advise them to comply with?
A. HIPAA
B. PCI-DSS
C. COPPA
D. GDPR

A

Correct options are A, B & D As per the Google Best practices for testing the performance of Bigtable, you should include these test scenarios in your performance and scalability testing Ensure the load tests are performed with enough data - For example, If your production Bigtable consists of 150GB data, you should test the performance with the same amount of data. Ensure heavy pre-test are performed for several minutes - By performing pre-test for several minutes this gives Cloud Bigtable time to balance data across several nodes based on the access pattern Ensure tests are running for at least 10 mins - This ensures that the test read operations are performed from disk as well as cached memory https://cloud.google.com/bigtable/docs/performance#testing
Option C is incorrect because as per best practices you should include Bigtable while performance testing
Option E is incorrect because the question is asking suggestions related to Bigtable

246
Q

One of your application recently was a victim of a large scale DDOS attack and web application attacks which is running in the Managed instance group behind the HTTPS load balancer. The CTO has tasked you to look for the services which can mitigate the DDOS attack and also provide a web application firewall. Which GCP service will you use?
A. Threat Protection
B. GCP Firewalls
C. Web Security Scanner
D. Cloud Armor

A

D) Option correct GDPR (General Data Protection Regulation) is regulatory compliance in Europe which is used to protect any personally identifiable information collected for business purpose within the European region
Option A is incorrect because HIPAA is related to protecting the privacy of healthcare data in the U.S
Option B is incorrect because it is a Payment Card Industry Data Security Standard to protect credit card information collected for business
Option C is incorrect because COPPA is regulatory compliance in the U.S which is related to protecting the privacy of children below 13 age in the U.S

247
Q

You are working for a large finance company as a Solutions Architect. There is a requirement that the Security team wants read-only access to all the projects under the organization for internal auditing purposes. How will you grant access to the security team as per Google’s IAM best practices?
A. Organization viewer, Project editor
B. Organization viewer, Security Center Admin
C. Organization viewer, Project owner
D. Organization viewer, Project viewer

A

D) Option is correct Cloud Armor is a fully managed service that protects your application against DDOS attack and also provides a Web access firewall which can further provide protection against attacks like XSS (cross-site-scripting), SQL injection, etc. You can also have geo-based access control to your application using Cloud armor. https://cloud.google.com/armor
Option A is incorrect because it is used to detect threats like Burt force attack from logs and reports to Security command center
Option B is incorrect because it is used to control incoming and outgoing traffic to and from your compute engine VM’s
Option C is incorrect because it is used to find any vulnerable library used in your application code

248
Q

You are working with a large finance company that is planning to use API keys in their application for accessing some Google Cloud API. According to Google’s best practices, what are the correct methods for securing API keys? Please select 2 options.
A. Do not place API keys directly into the application code
B. Use the same API key for different applications
C. Store the API keys into the code
D. Periodically Renew API keys

A

D) Option is correct Organization viewer role will provide the ability to view the entire organization and Project viewer role will grant read-only access to all the projects and resources under it within the organization
Option A is incorrect because the Project editor is a too broad role and it will grant read-write access to the resources within the project
Option B is incorrect because the Security Center Admin will not grant read-only access to the project
Option C is incorrect because the Project editor is a too broad role and it will grant admin access to all the resources within the project

249
Q

A financial company has recently moved from on-premise to Google Cloud Platform, they have started to use Bigquery for data analysis, while the performance of Bigquery has been good, but they are concerned about controlling the cost for Bigquery usage. Select the relevant Bigquery best practices for controlling costs from the options given below. (Select 3)
A. Avoid SELECT * Query only the columns that you need.
B. Use the –dry_run flag in the CLI Before running queries, preview them to estimate costs
C. Use a LIMIT clause as a method of cost control
D. If possible, partition your BigQuery tables by date
E. Always keep large result sets in BigQuery storage

A

Options A & D are correct As per Google’s API key best practices you should never place the API key directly into application code & renew API keys periodically You should save the API keys as environment variables or in a secret management system. https://developers.google.com/maps/api-key-best-practices
Option B is incorrect because you should have a limited scope for each API key.
Option C is incorrect because as per Google’s best practices you should never store API keys directly into code.

250
Q

Your company has decided to build a backup replica of their on- premises user authentication PostgreSQL database on Google Cloud Platform. The database is 4 TB, and large updates are frequent. Replication requires RFC1918 private address space. Which networking approach would be the best choice?
A. Create two VPN tunnels within the same Cloud VPN gateway to the same destination VPN gateway.
B. Direct Peering.
C. Google Cloud Dedicated Interconnect or Google Cloud Partner Interconnect
D. Google Cloud VPN connected to the data center network.

A

Options A, B, D are the right choices. Using SELECT * is the most expensive way to query data. When you use SELECT *, BigQuery does a full scan of every column in the table. Queries are billed according to the number of bytes read. To estimate costs before running a query you could use The –dry_run flag in the CLI. If possible, partition your BigQuery tables by date. Partitioning your tables allows you to query relevant subsets of data which improves performance and reduces costs.
Option C is an incorrect choice because applying a LIMIT clause to a query does not affect the amount of data that is read. It merely limits the results set to output. You are billed for reading all bytes in the entire table as indicated by the query.
Option E is an incorrect choice because, Keeping large result sets in BigQuery storage has a cost. If you don’t need permanent access to the results, use the default table expiration to automatically delete the data for you. Reference(s) : https://cloud.google.com/bigquery/docs/best-practices-costs

251
Q

A development manager is building a new application. He asks you to review his requirements and identify what cloud technologies he can use to meet them. The application must 1. Be based on open-source technology for cloud portability 2. Dynamically scale compute capacity based on demand 3. Support continuous software delivery 4. Run multiple segregated copies of the same application stack 5. Deploy application bundles using dynamic templates 6. Route network traffic to specific services based on URL Which combination of technologies will meet all of his requirements?
A. Google Kubernetes Engine and Cloud Load Balancing
B. Google Compute Engine, Jenkins, and Cloud Load Balancing
C. Google Kubernetes Engine and Cloud Deployment Manager
D. Google Kubernetes Engine, Jenkins, and Helm

A

The correct answer is Option C.
Option C is correct - Google Cloud Dedicated Interconnect or Google Cloud partner Interconnect Both VPN and Dedicated Interconnect/Partner Interconnect provide private address space communication. “The database is 4 TB, and large updates are frequent” makes the Dedicated Interconnect/Partner Interconnect a suitable solution due to its bandwidth capability and SLA A single interconnect can be a single 10G link or a link bundle, connected to a single Google router
Option D is incorrect because Google Cloud VPN connected to the data center network
Option A is incorrect because you cannot create two VPN tunnels within the same Cloud VPN gateway to the same destination VPN gateway.
Option B is incorrect because Direct Peering exists outside of the Google Cloud Platform. https://cloud.google.com/hybrid-connectivity/ Dedicated Interconnect Overview: https://cloud.google.com/interconnect/docs/concepts/dedicated- overview

252
Q

You want to enable your running Google Kubernates Engine cluster to scale as demand for your application changes. What should you do? Select one
A. Add additional nodes to your Kubernates Engine cluster using the following command: gcloud container clusters resize CLUSTER_NAME –size 10
B. Add a tag to the instances in the cluster with the following command: gcloud compute instances add-tags INSTANCE –tags enable-autoscaling max-nodes-10
C. Update the existing Kubernates Engine cluster with the following command: gcloud container clusters update CLUSTER_NAME –enable-autoscaling –min-nodes=1–max-nodes=10
D. Create a new Kubernates Engine cluster with the following command: gcloud container clusters create CLUSTER_NAME –enable-autoscaling –min-nodes=1–max-nodes=10 and redeploy your application

A

The correct answer is Option D. The best approach is by elimination: start from any requirements, for example, you may start elimination by the requirement not supported by repeatedly appeared components (e.g., GCE and GKE) in the questions · Req 1: Be based on open-source technology for cloud portability · Req 3 Support continuous software delivery · Req 5 Deploy application bundles using dynamic templates · Req 6 Route network traffic to specific services based on URL If we start from “Be based on open-source technology for cloud portability”, we know that Container Engine (new name is K8s Engine, GKE for short) one of the unique features is “open-source and cloud portability”. Now we have followings left: · Google Kubernetes Engine and Cloud Load Balancing · Google Kubernetes Engine, Jenkins, and Helm At this point, if you have the experience or knowledge, you probably are able to make the right decision. If not then following the same approach, we can choose either requirement of LB or CICD. For example, if we chose CICD, then the only answer is: Answer D. Google Kubernetes Engine, Jenkins, and Helm At first glance it appears answer D does not meet “all of his requirements” since it seems misses the “Route network traffic to specific services based on URL”, an obvious feature for Cloud Load Balancing. If looking further, we know, unlike Compute Engine, the Kubernetes Engine offers integrated support for two types of cloud load balancing for a publicly accessible application. One of them is HTTP(S) load balancers are designed to terminate HTTP(S) requests and can make better context-aware load balancing decisions. https://cloud.google.com/kubernetes-engine/docs/tutorials/http-balancer For your information: Helm is a package manager for Kubernetes templates. It allows for defining the Kubernetes templates required to run an application and then replace the application options dynamically. It bundles all the templates in tgz packages called charts. https://helm.sh/ Note: The first requirement in the question is “Open source technology for cloud portability “. The Google Kubernetes Engine (GKE) is the most preferred choice for this requirement for this. Google document mentions the following about GKE: Requirement 3 in the question is Continues delivery. Hence the correct choice will be Google Container Engine, Jenkins, and Helm. Also the one more requirement here is “Route network traffic to specific services based on URL” which is the requirement make you think to select Cloud Load Balancing. Let’s summarise how load balancing is achieved here. The Google Document mentions the following: Kubernetes Engine offers integrated support for two types of cloud load balancing for a publicly accessible application. Reference link: https://cloud.google.com/kubernetes-engine/docs/tutorials/http-balancer