Pre-Summer Sale Special - Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmaspas7

Easiest Solution 2 Pass Your Certification Exams

DP-600 Microsoft Implementing Analytics Solutions Using Microsoft Fabric Free Practice Exam Questions (2026 Updated)

Prepare effectively for your Microsoft DP-600 Implementing Analytics Solutions Using Microsoft Fabric certification with our extensive collection of free, high-quality practice questions. Each question is designed to mirror the actual exam format and objectives, complete with comprehensive answers and detailed explanations. Our materials are regularly updated for 2026, ensuring you have the most current resources to build confidence and succeed on your first attempt.

Page: 1 / 2
Total 171 questions

You have a Fabric workspace named Workspace1.

You need to create a semantic model named Model1 and publish Model1 to Workspace1. The solution must meet the following requirements:

Can revert to previous versions of Model1 as required.

Identifies differences between saved versions of Model1.

Uses Microsoft Power BI Desktop to publish to Workspace1.

Can edit item definition files by using Microsoft Visual Studio Code.

Which two actions should you perform? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

A.

Enable Git integration for Workspace1.

B.

Save Model1 in Power BI Desktop as a PBIX file.

C.

Save Model1 in Power BI Desktop as a PBIP file.

D.

Enable users to edit data models in the Power BI service.

UESTION NO: 83 HOTSPOT

You have an Azure Data Lake Storage Gen2 account named storage! that contains a Parquet file named sales.parquet.

You have a Fabric tenant that contains a workspace named Workspace1.

Using a notebook in Workspace1, you need to load t he content of the file to the default lakehouse. The solution must ensure that the content will display automatically as a table named Sales in Lakehouse explorer.

How should you complete the code? To answer, select the appropriate options in the answer ar ea.

NOTE: Each correct selection is worth one point.

You have a Fabric workspace named Workspace1 that is assigned to a newly created Fabric capacity named Capacity1.

You create a semantic model named Model1 and deploy Model1 to Workspace1.

You need to publish changes to Model1 directly from Tabular Editor.

What should you do?

A.

For Workspace1, enable Git integration.

B.

For Workspace1, create a managed private endpoint.

C.

For Model1, enable external sharing.

D.

For Capacity1, set XMLA Endpoint to Read Write.

What should you use to implement calculation groups for the Research division semantic models?

A.

DAX Studio

B.

Microsoft Power Bl Desktop

C.

the Power Bl se rvice

D.

Tabular Editor

Which workspace rote assignments should you recommend for ResearchReviewersGroupl and ResearchReviewersGroupZ? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

You need to ensure that Contoso can use version control to meet the data analytics requirements and the general requirements. What should you do?

A.

Store all the semantic models and reports in Data Lake Gen2 storage.

B.

Modify the settings of the Research workspaces to use a GitHub repository.

C.

Store all the semantic models and reports in Microsoft OneDrive.

D.

Modify the settings of the Research division workspaces to use an Azure Repos repository.

You have source data in a CSV file that has the following fields:

• SalesTra nsactionl D

• SaleDate

• CustomerCode

• CustomerName

• CustomerAddress

• ProductCode

• ProductName

• Quantity

• UnitPrice

You plan to implement a star schema for the tables in WH1. Thedimension tables in WH1 will implement Type 2 slowly changing dimension (SCD) logic.

You need to design the tables that will be used for sales transaction analysis and load the source data.

Which type of target table should you specify for the CustomerName, CustomerCode, and SaleDate fields? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

CC

Which syntax should you use in a notebook to ac cess the Research division data for Productlinel?

A)

B)

C)

D)

A.

Option A

B.

Option B

C.

Option C

D.

Option D

Which type of data store should you recommend in the AnalyticsPOC workspace?

A.

a data lake

B.

a warehouse

C.

a lakehouse

D.

an external Hive metaStore

You need to resolve the issue with the pricing group classification.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

You need to design a semantic model for the customer satisfaction report.

Which data source authentication method and mode should you use? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

You need to implement the date dimension in the data store. The solution must meet the technical requirements.

What are two ways to achieve the goal? Each correct answer presents a complete solution.

NOTE: Each correct selection is worth one point.

A.

Populate the date dimension table by using a dataflow.

B.

Populate the date dimension table by using a Stored procedure activity in a pipeline.

C.

Populate the date dimension view by using T-SQL.

D.

Populate the date dimension table by using a Copy activity in a pipeline.

What should you recommend using to ingest the customer data into the data store in the AnatyticsPOC workspace?

A.

a stored procedure

B.

a pipeline that contains a KQL activity

C.

a Spark notebook

D.

a dataflow

You need to create a DAX measure to calculate the average overall satisfaction score.

How should you complete the DAX code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

You need to ensure the data loading activities in the AnalyticsPOC workspace are executed in the appropriate sequence. The solution must meet the technical requirements.

What should you do?

A.

Create a pipeline that has depen dencies between activities and schedule the pipeline.

B.

Create and schedule a Spark job definition.

C.

Create a dataflow that has multiple steps and schedule the dataflow.

D.

Create and schedule a Spark notebook.

You ne ed to recommend a solution to prepare the tenant for the PoC.

Which two actions should you recommend performing from the Fabric Admin portal? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

A.

Enable the Use rs can try Microsoft Fabric paid features option for specific security groups.

B.

Enable the Allow Azure Active Directory guest users to access Microsoft Fabric option for specific security groups.

C.

Enable the Users can create Fabric items option and exclude specific security groups.

D.

Enable the Users can try Microsoft Fabric paid features option for the entire organization.

E.

Enable the Users can create Fabric items option for specific security groups.

You have a Fabric warehouse named Warehousel that contains a table named Table! Tablel contains customer data.

You need to implement row-level security (RLS) for Tablel. The solution must ensure that users can see only their respective data.

Which two objects should you create? Each correct answer presents part of the solution.

NOTE: Each co rrect selection is worth one point.

A.

SECURITY POLICY

B.

STORED PROCEDURE

C.

DATABASE ROLE

D.

FUNCTION

E.

CONSTRAINT

You have the following KQL query.

For each of the following statements, select Yes if the statement is true. Otherwise, select No. NOTE: Each correct selection is worth one point.

You need to create a data loading pattern for a Type 1 slowly changing dimension (SCD).

Which two actions should you include in the process? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

A.

Update rows when the non-key attributes have changed.

B.

Insert new rows when the natural key exists in the dimension table, and the non-key attribute values have changed.

C.

Update the effective end date of rows when the non-key attribute values have changed.

D.

Insert new records when the natural key is a new value in the table.

You have a Fabric tenant that contains a lakehouse named lakehouse1. Lakehouse1 contains a table named Table1.

You are creating a new data pipeline.

You plan to copy external data to Table1. The schema of the external data changes regularly.

You need the copy operation to meet the following requirements:

• Replace Table1 with the schema of the external data.

• Replace all the data in Table1 with the rows in the external data.

You add a Copy data activity to the pipeline. What should you do for the Copy data activity?

A.

From the Source tab, add additional columns.

B.

From the Destination tab, set Table action to Overwrite.

C.

From the Settings tab, select Enable staging

D.

From the Source tab, select Enable partition discovery

E.

From the Source tab, select Recursively

Page: 1 / 2
Total 171 questions
Copyright © 2014-2026 Solution2Pass. All Rights Reserved