DP-200
Free trial
Verified
Question 1
You are a data engineer implementing a lambda architecture on Microsoft Azure. You use an open-source big data solution to collect, process, and maintain data.
The analytical data store performs poorly.
You must implement a solution that meets the following requirements:
✑ Provide data warehousing
✑ Reduce ongoing management activities
✑ Deliver SQL query responses in less than one second
You need to create an HDInsight cluster to meet the requirements.
Which type of cluster should you create?
- A: Interactive Query
- B: Apache Hadoop
- C: Apache HBase
- D: Apache Spark
Question 2
The data engineering team manages Azure HDInsight clusters. The team spends a large amount of time creating and destroying clusters daily because most of the data pipeline process runs in minutes.
You need to implement a solution that deploys multiple HDInsight clusters with minimal effort.
What should you implement?
- A: Azure Databricks
- B: Azure Traffic Manager
- C: Azure Resource Manager templates
- D: Ambari web user interface
Question 3
You are the data engineer for your company. An application uses a NoSQL database to store data. The database uses the key-value and wide-column NoSQL database type.
Developers need to access data in the database using an API.
You need to determine which API to use for the database model and type.
Which two APIs should you use? Each correct answer presents a complete solution.
NOTE: Each correct selection is worth one point.
- A: Table API
- B: MongoDB API
- C: Gremlin API
- D: SQL API
- E: Cassandra API
Question 4
A company is designing a hybrid solution to synchronize data and on-premises Microsoft SQL Server database to Azure SQL Database.
You must perform an assessment of databases to determine whether data will move without compatibility issues. You need to perform the assessment.
Which tool should you use?
- A: SQL Server Migration Assistant (SSMA)
- B: Microsoft Assessment and Planning Toolkit
- C: SQL Vulnerability Assessment (VA)
- D: Azure SQL Data Sync
- E: Data Migration Assistant (DMA)
Question 5
DRAG DROP -
You manage a financial computation data analysis process. Microsoft Azure virtual machines (VMs) run the process in daily jobs, and store the results in virtual hard drives (VHDs.)
The VMs product results using data from the previous day and store the results in a snapshot of the VHD. When a new month begins, a process creates a new
VHD.
You must implement the following data retention requirements:
✑ Daily results must be kept for 90 days
✑ Data for the current year must be available for weekly reports
✑ Data from the previous 10 years must be stored for auditing purposes
✑ Data required for an audit must be produced within 10 days of a request.
You need to enforce the data retention requirements while minimizing cost.
How should you configure the lifecycle policy? To answer, drag the appropriate JSON segments to the correct locations. Each JSON segment may be used once, more than once, or not at all. You may need to drag the split bat between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
Select and Place:
Question 6
A company plans to use Azure SQL Database to support a mission-critical application.
The application must be highly available without performance degradation during maintenance windows.
You need to implement the solution.
Which three technologies should you implement? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
- A: Premium service tier
- B: Virtual machine Scale Sets
- C: Basic service tier
- D: SQL Data Sync
- E: Always On availability groups
- F: Zone-redundant configuration
Question 7
A company plans to use Azure Storage for file storage purposes. Compliance rules require:
✑ A single storage account to store all operations including reads, writes and deletes
✑ Retention of an on-premises copy of historical operations
You need to configure the storage account.
Which two actions should you perform? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
- A: Configure the storage account to log read, write and delete operations for service type Blob
- B: Use the AzCopy tool to download log data from $logs/blob
- C: Configure the storage account to log read, write and delete operations for service-type table
- D: Use the storage client to download log data from $logs/table
- E: Configure the storage account to log read, write and delete operations for service type queue
Question 8
DRAG DROP -
You are developing a solution to visualize multiple terabytes of geospatial data.
The solution has the following requirements:
✑ Data must be encrypted.
✑ Data must be accessible by multiple resources on Microsoft Azure.
You need to provision storage for the solution.
Which four actions should you perform in sequence? To answer, move the appropriate action from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
Question 9
You are developing a data engineering solution for a company. The solution will store a large set of key-value pair data by using Microsoft Azure Cosmos DB.
The solution has the following requirements:
✑ Data must be partitioned into multiple containers.
✑ Data containers must be configured separately.
✑ Data must be accessible from applications hosted around the world.
✑ The solution must minimize latency.
You need to provision Azure Cosmos DB.
Which three actions should you perform? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
- A: Configure account-level throughput.
- B: Provision an Azure Cosmos DB account with the Azure Table API. Enable geo-redundancy.
- C: Configure table-level throughput.
- D: Replicate the data globally by manually adding regions to the Azure Cosmos DB account.
- E: Provision an Azure Cosmos DB account with the Azure Table API. Enable multi-region writes.
Question 10
A company has a SaaS solution that uses Azure SQL Database with elastic pools. The solution will have a dedicated database for each customer organization.
Customer organizations have peak usage at different periods during the year.
Which two factors affect your costs when sizing the Azure SQL Database elastic pools? Each correct answer presents a complete solution.
NOTE: Each correct selection is worth one point.
- A: maximum data size
- B: number of databases
- C: eDTUs consumption
- D: number of read operations
- E: number of transactions
Question 11
HOTSPOT -
You are developing a solution using a Lambda architecture on Microsoft Azure.
The data at rest layer must meet the following requirements:
Data storage:
✑ Serve as a repository for high volumes of large files in various formats.
✑ Implement optimized storage for big data analytics workloads.
✑ Ensure that data can be organized using a hierarchical structure.
Batch processing:
✑ Use a managed solution for in-memory computation processing.
✑ Natively support Scala, Python, and R programming languages.
✑ Provide the ability to resize and terminate the cluster automatically.
Analytical data store:
✑ Support parallel processing.
✑ Use columnar storage.
✑ Support SQL-based languages.
You need to identify the correct technologies to build the Lambda architecture.
Which technologies should you use? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
Question 12
DRAG DROP -
You develop data engineering solutions for a company. You must migrate data from Microsoft Azure Blob storage to an Azure SQL Data Warehouse for further transformation. You need to implement the solution.
Which four actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
Question 13
DRAG DROP -
Your company has an on-premises Microsoft SQL Server instance.
The data engineering team plans to implement a process that copies data from the SQL Server instance to Azure Blob storage once a day. The process must orchestrate and manage the data lifecycle.
You need to create Azure Data Factory to connect to the SQL Server instance.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
Question 14
A company runs Microsoft SQL Server in an on-premises virtual machine (VM).
You must migrate the database to Azure SQL Database. You synchronize users from Active Directory to Azure Active Directory (Azure AD).
You need to configure Azure SQL Database to use an Azure AD user as administrator.
What should you configure?
- A: For each Azure SQL Database, set the Access Control to administrator.
- B: For each Azure SQL Database server, set the Active Directory to administrator.
- C: For each Azure SQL Database, set the Active Directory administrator role.
- D: For each Azure SQL Database server, set the Access Control to administrator.
Question 15
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You have an Azure SQL database named DB1 that contains a table named Table1. Table1 has a field named Customer_ID that is varchar(22).
You need to implement masking for the Customer_ID field to meet the following requirements:
✑ The first two prefix characters must be exposed.
✑ The last four suffix characters must be exposed.
✑ All other characters must be masked.
Solution: You implement data masking and use a credit card function mask.
Does this meet the goal?
- A: Yes
- B: No
Question 16
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You have an Azure SQL database named DB1 that contains a table named Table1. Table1 has a field named Customer_ID that is varchar(22).
You need to implement masking for the Customer_ID field to meet the following requirements:
✑ The first two prefix characters must be exposed.
The last four suffix characters must be exposed.
✑ All other characters must be masked.
Solution: You implement data masking and use an email function mask.
Does this meet the goal?
- A: Yes
- B: No
Question 17
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You have an Azure SQL database named DB1 that contains a table named Table1. Table1 has a field named Customer_ID that is varchar(22).
You need to implement masking for the Customer_ID field to meet the following requirements:
✑ The first two prefix characters must be exposed.
✑ The last four suffix characters must be exposed.
✑ All other characters must be masked.
Solution: You implement data masking and use a random number function mask.
Does this meet the goal?
- A: Yes
- B: No
Question 18
DRAG DROP -
You are responsible for providing access to an Azure Data Lake Storage Gen2 account.
Your user account has contributor access to the storage account, and you have the application ID and access key.
You plan to use PolyBase to load data into an enterprise data warehouse in Azure Synapse Analytics.
You need to configure PolyBase to connect the data warehouse to the storage account.
Which three components should you create in sequence? To answer, move the appropriate components from the list of components to the answer area and arrange them in the correct order.
Select and Place:
Question 19
You plan to create a dimension table in Azure Synapse Analytics that will be less than 1 GB.
You need to create the table to meet the following requirements:
✑ Provide the fastest query time.
✑ Minimize data movement during queries.
Which type of table should you use?
- A: hash distributed
- B: heap
- C: replicated
- D: round-robin
Question 20
You have an enterprise data warehouse in Azure Synapse Analytics.
Using PolyBase, you create an external table named [Ext].[Items] to query Parquet files stored in Azure Data Lake Storage Gen2 without importing the data to the data warehouse.
The external table has three columns.
You discover that the Parquet files have a fourth column named ItemID.
Which command should you run to add the ItemID column to the external table?
- A: Option A
- B: Option B
- C: Option C
- D: Option D
Question 21
DRAG DROP -
You have a table named SalesFact in an enterprise data warehouse in Azure Synapse Analytics. SalesFact contains sales data from the past 36 months and has the following characteristics:
✑ Is partitioned by month
✑ Contains one billion rows
✑ Has clustered columnstore indexes
At the beginning of each month, you need to remove data from SalesFact that is older than 36 months as quickly as possible.
Which three actions should you perform in sequence in a stored procedure? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
Question 22
You plan to implement an Azure Cosmos DB database that will write 100,000,000 JSON records every 24 hours. The database will be replicated to three regions.
Only one region will be writable.
You need to select a consistency level for the database to meet the following requirements:
✑ Guarantee monotonic reads and writes within a session.
✑ Provide the fastest throughput.
✑ Provide the lowest latency.
Which consistency level should you select?
- A: Strong
- B: Bounded Staleness
- C: Eventual
- D: Session
- E: Consistent Prefix
Question 23
You develop data engineering solutions for a company. The company has on-premises Microsoft SQL Server databases at multiple locations.
The company must integrate data with Microsoft Power BI and Microsoft Azure Logic Apps. The solution must avoid single points of failure during connection and transfer to the cloud. The solution must also minimize latency.
You need to secure the transfer of data between on-premises databases and Microsoft Azure.
What should you do?
- A: Install a standalone on-premises Azure data gateway at each location
- B: Install an on-premises data gateway in personal mode at each location
- C: Install an Azure on-premises data gateway at the primary location
- D: Install an Azure on-premises data gateway as a cluster at each location
Question 24
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You have an Azure SQL database named DB1 that contains a table named Table1. Table1 has a field named Customer_ID that is varchar(22).
You need to implement masking for the Customer_ID field to meet the following requirements:
✑ The first two prefix characters must be exposed.
✑ The last four suffix characters must be exposed.
✑ All other characters must be masked.
Solution: You implement data masking and use a custom text mask.
Does this meet the goal?
- A: Yes
- B: No
Question 25
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You have an Azure Storage account that contains 100 GB of files. The files contain text and numerical values. 75% of the rows contain description data that has an average length of 1.1 MB.
You plan to copy the data from the storage account to an enterprise data warehouse in Azure Synapse Analytics.
You need to prepare the files to ensure that the data copies quickly.
Solution: You modify the files to ensure that each row is less than 1 MB.
Does this meet the goal?
- A: Yes
- B: No
Free preview mode
Enjoy the free questions and consider upgrading to gain full access!