3. DP-203 Data Integration with Data Factory Flashcards

1
Q

Much of the functionality of Azure Data Factory appears in Azure Synapse Analytics as a feature called pipelines. You can use it to integrate data pipelines between which of the following?
Select all options that apply.

-Spark Pools
-Apache Hive
-SQL Pools
-SQL Serverless

A

-Spark Pools
-SQL Pools
-SQL Serverless

Pipelines enable you to integrate data pipelines between SQL Pools, Spark Pools, and SQL Serverless.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Which of the following provides a cloud-based data integration service that orchestrates the movement and transformation of data between various data stores and compute resources?

A

Azure Data Factory

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

ADF has native functionality to ingest and transform data, sometimes it will instruct another service, such as Databricks, to perform the actual work required on its behalf. Which of the following terms best describes this process?

A

Orchestration

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Which of the following terms describes analyzing past data patterns and trends by looking at historical data and customer insights?

-Descriptive Analytics
-Prescriptive Analytics
-Predictive Analytics

A

-Descriptive Analytics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Microsoft Azure provides a variety of data platform services that enables you to perform different types of analytics. Predictive analytics can be implemented through which of the following features?
Select all options that apply

-HDInsight
-Azure Data Lake Storage Gen2
-Machine Learning Services
-Azure Databricks

A

-HDInsight
-Machine Learning Services
-Azure Databricks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Data integration includes extraction, transformation, and loading of data. It is commonly referred to as Extract-Transform-Load or ETL.

At which stage in the ETL process is the splitting, combining, deriving, adding, and removing data carried out?

A

Transform

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

You are creating a new Azure Data Factory instance. The instance name must be unique within which of the following?

A

Globally within Azure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How would you define an Azure factory dataset?

A

A dataset is a named view that points to, or references, the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How would you define an Azure Data Factory pipeline?

A

A pipeline is a logical grouping of activities that together perform a task.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How would you define an Azure Data Factory ‘activity’?

A

Activities typically contain the transformation logic or the analysis commands of the Azure Data Factory’s work.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the three categories of activities within Azure Data Factory that define the actions to be performed on the data?
Select all options that apply.

-Data movement
-Linked Service
-Data transformation
-Control

A

-Data transformation
Data transformation activities can be performed natively within the authoring tool of Azure Data Factory using the Mapping Data Flow. Alternatively, you can call a compute resource to change or enhance data through transformation or perform analysis of the data.

-Control
You can use the control flow to orchestrate pipeline activities, including chaining activities in a sequence, branching, defining parameters at the pipeline level, and passing arguments while invoking the pipeline on-demand or from a trigger.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

When graphically authoring ADF solutions, you can use the control flow within the design to orchestrate which of the following pipeline activities?
Select all options that apply.

-Execute Pipeline Activity
-WebActivity
-ForEach Activity
-Parameters Activity

A

-Execute Pipeline Activity
-WebActivity
-ForEach Activity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Which of the following processes will allow data to be extracted and loaded in its native format?
Select all options that apply.

-ELTL
-ETL
-ELT
-ETLL

A

-ELTL
-ELT

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Select all options that apply.
Pipelines in Azure Data Factory typically perform the four distinct steps. Identify these steps?

-Connect and Collect
-Publish
-Transform and Enrich
-Monitor
-Data Analysis

A

-Connect and Collect
-Publish
-Transform and Enrich
-Monitor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

To create and manage child resources including publishing in the Azure portal, which Data Factory role must you belong to at the resource group level or above?

-Data Factory User
-Data Factory Contributor
-Data Factory Writer
-Data Factory Reader

A

-Data Factory Contributor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

When setting up Azure Data factory through the Azure Portal. Which of the following pieces of Information are required?
Select all options that apply.

-Subscription
-Region
-Instance Name
-Resource Group name
-Management Group Name

A

-Subscription
-Region
-Instance Name
-Resource Group name

17
Q

Azure Data Factory integration runtime enables bridging between activities and linked Services objects. Which of the following are types of Integration Runtime in Azure data factory?
Select all options that apply.

-Azure-SSIS
-Azure-SQL
-Azure
-Self Hosted
-Azure-Spark

A

-Azure-SSIS
-Azure
-Self Hosted

18
Q

An integration runtime provides which of the following?

-The action to be performed.
-The infrastructure for the activity and linked services.
-A target data store or a compute service.

A

-The infrastructure for the activity and linked services.

19
Q

Data Factory offers three types of Integration Runtime.

-Azure
-Self-hosted
-Azure-SSIS

Which of these provides support for both Public Networks and Private Networks?

A

Self Hosted and Azure-SSIS

20
Q

You are moving data from an Azure Data Lake Gen2 store to Azure Synapse Analytics. Which Azure Data Factory integration runtime would be used in a data copy activity?

A

Azure

21
Q

Which Azure Data Factory component orchestrates a data movement command?

-Datasets
-Linked Services
-Activities

A

Activities

22
Q

Which Azure Data Factory process involves using compute services to produce data to feed production environments with cleansed data?

-Publish
-Transform and enrich
-Monitor
-Connect and collect

A

-Transform and enrich