Vcehome > Microsoft > Microsoft Certified: Azure Data Engineer Associate > DP-203 > DP-203 Online Practice Questions and Answers

DP-203 Online Practice Questions and Answers

Questions 4

DRAG DROP

You have the following table named Employees.

You need to calculate the employee_type value based on the hire_date value.

How should you complete the Transact-SQL statement? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to

view content.

NOTE: Each correct selection is worth one point.

Select and Place:

Browse 380 Q&As
Questions 5

DRAG DROP

You have an Azure Synapse Analytics SQL pool named Pool1 on a logical Microsoft SQL server named Server1.

You need to implement Transparent Data Encryption (TDE) on Pool1 by using a custom key named key1.

Which five actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

Select and Place:

Browse 380 Q&As
Questions 6

HOTSPOT

You need to output files from Azure Data Factory.

Which file format should you use for each type of output? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Hot Area:

Browse 380 Q&As
Questions 7

You are creating an Azure Data Factory data flow that will ingest data from a CSV file, cast columns to specified types of data, and insert the data into a table in an Azure Synapse Analytic dedicated SQL pool. The CSV file contains three columns named username, comment, and date.

The data flow already contains the following:

1.

A source transformation.

2.

A Derived Column transformation to set the appropriate types of data.

3.

A sink transformation to land the data in the pool.

You need to ensure that the data flow meets the following requirements:

1.

All valid rows must be written to the destination table.

2.

Truncation errors in the comment column must be avoided proactively.

3.

Any rows containing comment values that will cause truncation errors upon insert must be written to a file in blob storage.

Which two actions should you perform? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

A. To the data flow, add a sink transformation to write the rows to a file in blob storage.

B. To the data flow, add a Conditional Split transformation to separate the rows that will cause truncation errors.

C. To the data flow, add a filter transformation to filter out rows that will cause truncation errors.

D. Add a select transformation to select only the rows that will cause truncation errors.

Browse 380 Q&As
Questions 8

You have an Azure Synapse Analytics dedicated SQL pool.

You need to ensure that data in the pool is encrypted at rest. The solution must NOT require modifying applications that query the data.

What should you do?

A. Enable encryption at rest for the Azure Data Lake Storage Gen2 account.

B. Enable Transparent Data Encryption (TDE) for the pool.

C. Use a customer-managed key to enable double encryption for the Azure Synapse workspace.

D. Create an Azure key vault in the Azure subscription grant access to the pool.

Browse 380 Q&As
Questions 9

You are designing a fact table named FactPurchase in an Azure Synapse Analytics dedicated SQL pool. The table contains purchases from suppliers for a retail store. FactPurchase will contain the following columns.

FactPurchase will have 1 million rows of data added daily and will contain three years of data. Transact-SQL queries similar to the following query will be executed daily. SELECT SupplierKey, StockItemKey, COUNT(*) FROM FactPurchase WHERE DateKey >= 20210101 AND DateKey <= 20210131 GROUP By SupplierKey, StockItemKey Which table distribution will minimize query times?

A. round-robin

B. replicated

C. hash-distributed on DateKey

D. hash-distributed on PurchaseKey

Browse 380 Q&As
Questions 10

You plan to use an Apache Spark pool in Azure Synapse Analytics to load data to an Azure Data Lake Storage Gen2 account.

You need to recommend which file format to use to store the data in the Data Lake Storage account. The solution must meet the following requirements:

Column names and data types must be defined within the files loaded to the Data Lake Storage account. Data must be accessible by using queries from an Azure Synapse Analytics serverless SQL pool. Partition elimination must be

supported without having to specify a specific partition.

What should you recommend?

A. Delta Lake

B. JSON

C. CSV

D. ORC

Browse 380 Q&As
Questions 11

You have an Azure data factory named ADF1.

You currently publish all pipeline authoring changes directly to ADF1.

You need to implement version control for the changes made to pipeline artifacts. The solution must ensure that you can apply version control to the resources currently defined in the Azure Data Factory Studio for ADF1.

Which two actions should you perform? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

A. From the Azure Data Factory Studio, run Publish All.

B. Create an Azure Data Factory trigger.

C. Create a Git repository.

D. Create a GitHub action.

E. From the Azure Data Factory Studio, select Set up code repository.

F. From the Azure Data Factory Studio, select Publish.

Browse 380 Q&As
Questions 12

What should you do to improve high availability of the real-time data processing solution?

A. Deploy a High Concurrency Databricks cluster.

B. Deploy an Azure Stream Analytics job and use an Azure Automation runbook to check the status of the job and to start the job if it stops.

C. Set Data Lake Storage to use geo-redundant storage (GRS).

D. Deploy identical Azure Stream Analytics jobs to paired regions in Azure.

Browse 380 Q&As
Questions 13

What should you recommend using to secure sensitive customer contact information?

A. Transparent Data Encryption (TDE)

B. row-level security

C. column-level security

D. data sensitivity labels

Browse 380 Q&As
Exam Code: DP-203
Exam Name: Data Engineering on Microsoft Azure
Last Update: May 11, 2024
Questions: 380 Q&As

PDF

$49.99

VCE

$59.99

PDF + VCE

$67.99