DP-203 Flashcards
Which roles does the user account used to sign into Azure need to be a member of to create a Data Factory instance?
Contributor and owner OR administrator
To create and manage Data Factory objects including datasets, linked services, pipelines, triggers, and integration runtimes, which requirements must be met?
To create and manage child resources in the Azure portal, you must belong to the Data Factory Contributor role at the resource group level or above.
To create and manage resources with PowerShell or the SDK, the contributor role at the resource level or above is sufficient.
What are the 3 types of Integration runtime?
Azure
Self-hosted
Azure-SSIS
Which Azure Data Factory component orchestrates a transformation job or runs a data movement command?
Activities
You are moving data from an Azure Data Lake Gen2 store to Azure Synapse Analytics. Which Azure Data Factory integration runtime would be used in a data copy activity?
Azure
What is the Azure IR used for?
When moving data between Azure data platform technologies, the Azure Integration runtime is used when copying data between two Azure data platform.
What is the self-hosted IR used for?
Self-hosted IR is used when working with data movement from private networks to the cloud and vice versa.
What is the Azure-SSIS IR used for?
Azure-SSIS IR is used when you lift and shift existing SSIS workload.
What is an integration runtime?
An integration runtime provides the bridge between the activity and linked services.
In Azure Data Factory authoring tool, where would you find the Copy data activity?
The Move & Transform section contains activities that are specific to Azure Data Factory copying data and defining data flows.
You want to ingest data from a SQL Server database hosted on an on-premises Windows Server. What integration runtime is required for Azure Data Factory to ingest data from the on-premises server?
A self-hosted integration runtime can run copy activities between a cloud data store and a data store in a private network. It also can dispatch transform activities against compute resources in an on-premises network or an Azure virtual network.
By default, how long are the Azure Data Factory diagnostic logs retained for?
45 days
Which transformation in the Mapping Data Flow is used to routes data rows to different streams based on matching conditions?
A Conditional Split transformation routes data rows to different streams based on matching conditions. The conditional split transformation is similar to a CASE decision structure in a programming language.
Which transformation is used to load data into a data store or compute resource?
A Sink transformation allows you to choose a dataset definition for the destination output data. You can have as many sink transformations as your data flow requires.
How long does Databricks retain cluster configuration information?
Up to 70 all-purpose clusters terminated in the last 30 days and up to 30 job clusters recently terminated by the job scheduler.
How to keep an all-purpose cluster configuration in Databricks for more than 30 days after it has been terminated?
An administrator can pin a cluster to the cluster list
What are the 2 cluster types?
All-purpose cluster or a job cluster
What is an All-purpose cluster?
All-purpose clusters can be shared by multiple users and are best for performing ad-hoc analysis, data exploration, or development. Once you’ve completed implementing your processing and are ready to operationalize your code, switch to running it on a job cluster.
What is a job cluster?
Job clusters terminate when your job ends, reducing resource usage and cost.
What are the 3 cluster modes?
Standard, High Concurrency, and Single Node. Most regular users use Standard or Single Node clusters.
What is a standard cluster used for?
Standard clusters are ideal for processing large amounts of data with Apache Spark.
What is a single node cluster used for?
Single Node clusters are intended for jobs that use small amounts of data or non-distributed workloads such as single-node machine learning libraries.