Weekend Sale Special - Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: top65certs

Microsoft Certified: Azure Data Engineer Associate DP-203 Exam Questions and Answers PDF

Page: 6 / 8
Total 316 questions

Data Engineering on Microsoft Azure Questions and Answers

Question 21

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You plan to create an Azure Databricks workspace that has a tiered structure. The workspace will contain the following three workloads:

  • A workload for data engineers who will use Python and SQL.
  • A workload for jobs that will run notebooks that use Python, Scala, and SOL.
  • A workload that data scientists will use to perform ad hoc analysis in Scala and R.

The enterprise architecture team at your company identifies the following standards for Databricks environments:

  • The data engineers must share a cluster.
  • The job cluster will be managed by using a request process whereby data scientists and data engineers provide packaged notebooks for deployment to the cluster.
  • All the data scientists must be assigned their own cluster that terminates automatically after 120 minutes of inactivity. Currently, there are three data scientists.

You need to create the Databricks clusters for the workloads.

Solution: You create a High Concurrency cluster for each data scientist, a High Concurrency cluster for the data engineers, and a Standard cluster for the jobs.

Does this meet the goal?

Options:

A.

Yes

B.

No

Question 22

You haw an Azure data factory named ADF1.

You currently publish all pipeline authoring changes directly to ADF1.

You need to implement version control for the changes made to pipeline artifacts. The solution must ensure that you can apply version control to the resources currently defined m the UX Authoring canvas for ADF1.

Which two actions should you perform? Each correct answer presents part of the solution

NOTE: Each correct selection is worth one point.

Options:

A.

Create an Azure Data Factory trigger

B.

From the UX Authoring canvas, select Set up code repository

C.

Create a GitHub action

D.

From the Azure Data Factor Studio, run Publish All.

E.

Create a Git repository

F.

From the UX Authoring canvas, select Publish

Question 23

You have an Azure data factor/ connected to a Git repository that contains the following branches:

• mam: Collaboration branch

• abc: Feature branch

• xyz: Feature branch

You save charges to a pipeline in the xyz branch.

You need to publish the changes to the live service

What should you do first?

Options:

A.

Push the code to a remote origin.

B.

Publish the data factory.

C.

Create a pull request to merge the changes into the abc branch.

D.

Create a pull request to merge the changes into the main branch.

Question 24

You are designing a data mart for the human resources (MR) department at your company. The data mart will contain information and employee transactions. From a source system you have a flat extract that has the following fields:

• EmployeeID

• FirstName

• LastName

• Recipient

• GrossArnount

• TransactionID

• GovernmentID

• NetAmountPaid

• TransactionDate

You need to design a start schema data model in an Azure Synapse analytics dedicated SQL pool for the data mart.

Which two tables should you create? Each Correct answer present part of the solution.

Options:

A.

a dimension table for employee

B.

a fabric for Employee

C.

a dimension table far EmployeeTransaction

D.

a dimension table for Transaction

E.

a fact table for Transaction

Page: 6 / 8
Total 316 questions