Download Implementing an Azure Data Solution.DP-200.TestKing.2020-01-16.83q.vcex

Vendor: Microsoft
Exam Code: DP-200
Exam Name: Implementing an Azure Data Solution
Date: Jan 16, 2020
File Size: 4 MB

How to open VCEX files?

Files with VCEX extension can be opened by ProfExam Simulator.

Demo Questions

Question 1
You are a data engineer implementing a lambda architecture on Microsoft Azure. You use an open-source big data solution to collect, process, and maintain data. The analytical data store performs poorly. 
You must implement a solution that meets the following requirements:
  • Provide data warehousing 
  • Reduce ongoing management activities 
  • Deliver SQL query responses in less than one second 
You need to create an HDInsight cluster to meet the requirements. 
Which type of cluster should you create?
  1. Interactive Query
  2. Apache Hadoop
  3. Apache HBase
  4. Apache Spark
Correct answer: D
Explanation:
Lambda Architecture with Azure:Azure offers you a combination of following technologies to accelerate real-time big data analytics:Azure Cosmos DB, a globally distributed and multi-model database service. Apache Spark for Azure HDInsight, a processing framework that runs large-scale data analytics applications. Azure Cosmos DB change feed, which streams new data to the batch layer for HDInsight to process. The Spark to Azure Cosmos DB Connector       Note: Lambda architecture is a data-processing architecture designed to handle massive quantities of data by taking advantage of both batch processing and stream processing methods, and minimizing the latency involved in querying big data.References:https://sqlwithmanoj.com/2018/02/16/what-is-lambda-architecture-and-what-azure-offers-with-its-new-cosmos-db/
Lambda Architecture with Azure:
Azure offers you a combination of following technologies to accelerate real-time big data analytics:
  1. Azure Cosmos DB, a globally distributed and multi-model database service. 
  2. Apache Spark for Azure HDInsight, a processing framework that runs large-scale data analytics applications. 
  3. Azure Cosmos DB change feed, which streams new data to the batch layer for HDInsight to process. 
  4. The Spark to Azure Cosmos DB Connector 
   
 
  
Note: Lambda architecture is a data-processing architecture designed to handle massive quantities of data by taking advantage of both batch processing and stream processing methods, and minimizing the latency involved in querying big data.
References:
https://sqlwithmanoj.com/2018/02/16/what-is-lambda-architecture-and-what-azure-offers-with-its-new-cosmos-db/
Question 2
You develop data engineering solutions for a company. The company has on-premises Microsoft SQL Server databases at multiple locations. 
The company must integrate data with Microsoft Power BI and Microsoft Azure Logic Apps. The solution must avoid single points of failure during connection and transfer to the cloud. The solution must also minimize latency. 
You need to secure the transfer of data between on-premises databases and Microsoft Azure. 
What should you do?
  1. Install a standalone on-premises Azure data gateway at each location
  2. Install an on-premises data gateway in personal mode at each location
  3. Install an Azure on-premises data gateway at the primary location
  4. Install an Azure on-premises data gateway as a cluster at each location
Correct answer: D
Explanation:
You can create high availability clusters of On-premises data gateway installations, to ensure your organization can access on-premises data resources used in Power BI reports and dashboards. Such clusters allow gateway administrators to group gateways to avoid single points of failure in accessing on-premises data resources. The Power BI service always uses the primary gateway in the cluster, unless it’s not available. In that case, the service switches to the next gateway in the cluster, and so on. References:https://docs.microsoft.com/en-us/power-bi/service-gateway-high-availability-clusters
You can create high availability clusters of On-premises data gateway installations, to ensure your organization can access on-premises data resources used in Power BI reports and dashboards. Such clusters allow gateway administrators to group gateways to avoid single points of failure in accessing on-premises data resources. The Power BI service always uses the primary gateway in the cluster, unless it’s not available. In that case, the service switches to the next gateway in the cluster, and so on. 
References:
https://docs.microsoft.com/en-us/power-bi/service-gateway-high-availability-clusters
Question 3
You are a data architect. The data engineering team needs to configure a synchronization of data between an on-premises Microsoft SQL Server database to Azure SQL Database. 
Ad-hoc and reporting queries are being overutilized the on-premises production instance. The synchronization process must:
  • Perform an initial data synchronization to Azure SQL Database with minimal downtime 
  • Perform bi-directional data synchronization after initial synchronization 
You need to implement this synchronization solution. 
Which synchronization method should you use?
  1. transactional replication
  2. Data Migration Assistant (DMA)
  3. backup and restore
  4. SQL Server Agent job
  5. Azure SQL Data Sync
Correct answer: E
Explanation:
SQL Data Sync is a service built on Azure SQL Database that lets you synchronize the data you select bi-directionally across multiple SQL databases and SQL Server instances. With Data Sync, you can keep data synchronized between your on-premises databases and Azure SQL databases to enable hybrid applications. Compare Data Sync with Transactional Replication        References:https://docs.microsoft.com/en-us/azure/sql-database/sql-database-sync-data
SQL Data Sync is a service built on Azure SQL Database that lets you synchronize the data you select bi-directionally across multiple SQL databases and SQL Server instances. 
With Data Sync, you can keep data synchronized between your on-premises databases and Azure SQL databases to enable hybrid applications. 
Compare Data Sync with Transactional Replication 
   
 
   
References:
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-sync-data
Question 4
An application will use Microsoft Azure Cosmos DB as its data solution. The application will use the Cassandra API to support a column-based database type that uses containers to store items. 
You need to provision Azure Cosmos DB. Which container name and item name should you use?  Each correct answer presents part of the solutions. 
NOTE: Each correct answer selection is worth one point.
  1. collection
  2. rows
  3. graph
  4. entities
  5. table
Correct answer: BE
Explanation:
B: Depending on the choice of the API, an Azure Cosmos item can represent either a document in a collection, a row in a table or a node/edge in a graph. The following table shows the mapping between API-specific entities to an Azure Cosmos item:      E: An Azure Cosmos container is specialized into API-specific entities as follows:       References:https://docs.microsoft.com/en-us/azure/cosmos-db/databases-containers-items
B: Depending on the choice of the API, an Azure Cosmos item can represent either a document in a collection, a row in a table or a node/edge in a graph. The following table shows the mapping between API-specific entities to an Azure Cosmos item:
   
 
  
E: An Azure Cosmos container is specialized into API-specific entities as follows:
   
 
   
References:
https://docs.microsoft.com/en-us/azure/cosmos-db/databases-containers-items
Question 5
A company has a SaaS solution that uses Azure SQL Database with elastic pools. The solution contains a dedicated database for each customer organization. Customer organizations have peak usage at different periods during the year. 
You need to implement the Azure SQL Database elastic pool to minimize cost. 
Which option or options should you configure?
  1. Number of transactions only
  2. eDTUs per database only
  3. Number of databases only
  4. CPU usage only
  5. eDTUs and max data size
Correct answer: E
Explanation:
The best size for a pool depends on the aggregate resources needed for all databases in the pool. This involves determining the following:Maximum resources utilized by all databases in the pool (either maximum DTUs or maximum vCores depending on your choice of resourcing model). Maximum storage bytes utilized by all databases in the pool. Note: Elastic pools enable the developer to purchase resources for a pool shared by multiple databases to accommodate unpredictable periods of usage by individual databases. You can configure resources for the pool based either on the DTU-based purchasing model or the vCore-based purchasing model.References:https://docs.microsoft.com/en-us/azure/sql-database/sql-database-elastic-pool
The best size for a pool depends on the aggregate resources needed for all databases in the pool. This involves determining the following:
  • Maximum resources utilized by all databases in the pool (either maximum DTUs or maximum vCores depending on your choice of resourcing model). 
  • Maximum storage bytes utilized by all databases in the pool. 
Note: Elastic pools enable the developer to purchase resources for a pool shared by multiple databases to accommodate unpredictable periods of usage by individual databases. You can configure resources for the pool based either on the DTU-based purchasing model or the vCore-based purchasing model.
References:
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-elastic-pool
Question 6
A company manages several on-premises Microsoft SQL Server databases. 
You need to migrate the databases to Microsoft Azure by using a backup process of Microsoft SQL Server. 
Which data technology should you use?
  1. Azure SQL Database single database
  2. Azure SQL Data Warehouse
  3. Azure Cosmos DB
  4. Azure SQL Database Managed Instance
Correct answer: D
Explanation:
Managed instance is a new deployment option of Azure SQL Database, providing near 100% compatibility with the latest SQL Server on-premises (Enterprise Edition) Database Engine, providing a native virtual network (VNet) implementation that addresses common security concerns, and a business model favorable for on-premises SQL Server customers. The managed instance deployment model allows existing SQL Server customers to lift and shift their on-premises applications to the cloud with minimal application and database changes. References:https://docs.microsoft.com/en-us/azure/sql-database/sql-database-managed-instance
Managed instance is a new deployment option of Azure SQL Database, providing near 100% compatibility with the latest SQL Server on-premises (Enterprise Edition) Database Engine, providing a native virtual network (VNet) implementation that addresses common security concerns, and a business model favorable for on-premises SQL Server customers. The managed instance deployment model allows existing SQL Server customers to lift and shift their on-premises applications to the cloud with minimal application and database changes. 
References:
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-managed-instance
Question 7
The data engineering team manages Azure HDInsight clusters. The team spends a large amount of time creating and destroying clusters daily because most of the data pipeline process runs in minutes. 
You need to implement a solution that deploys multiple HDInsight clusters with minimal effort. 
What should you implement?
  1. Azure Databricks
  2. Azure Traffic Manager
  3. Azure Resource Manager templates
  4. Ambari web user interface
Correct answer: C
Explanation:
A Resource Manager template makes it easy to create the following resources for your application in a single, coordinated operation:HDInsight clusters and their dependent resources (such as the default storage account). Other resources (such as Azure SQL Database to use Apache Sqoop). In the template, you define the resources that are needed for the application. You also specify deployment parameters to input values for different environments. The template consists of JSON and expressions that you use to construct values for your deployment. References:https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-create-linux-clusters-arm-templates
A Resource Manager template makes it easy to create the following resources for your application in a single, coordinated operation:
  • HDInsight clusters and their dependent resources (such as the default storage account). 
  • Other resources (such as Azure SQL Database to use Apache Sqoop). 
In the template, you define the resources that are needed for the application. You also specify deployment parameters to input values for different environments. The template consists of JSON and expressions that you use to construct values for your deployment. 
References:
https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-create-linux-clusters-arm-templates
Question 8
You are the data engineer for your company. An application uses a NoSQL database to store data. The database uses the key-value and wide-column NoSQL database type. 
Developers need to access data in the database using an API. 
You need to determine which API to use for the database model and type. 
Which two APIs should you use? Each correct answer presents a complete solution. 
NOTE: Each correct selection is worth one point.
  1. Table API
  2. MongoDB API
  3. Gremlin API
  4. SQL API
  5. Cassandra API
Correct answer: BE
Explanation:
B: Azure Cosmos DB is the globally distributed, multimodel database service from Microsoft for mission-critical applications. It is a multimodel database and supports document, key-value, graph, and columnar data models.E: Wide-column stores store data together as columns instead of rows and are optimized for queries over large datasets. The most popular are Cassandra and HBase.References:https://docs.microsoft.com/en-us/azure/cosmos-db/graph-introductionhttps://www.mongodb.com/scale/types-of-nosql-databases
B: Azure Cosmos DB is the globally distributed, multimodel database service from Microsoft for mission-critical applications. It is a multimodel database and supports document, key-value, graph, and columnar data models.
E: Wide-column stores store data together as columns instead of rows and are optimized for queries over large datasets. The most popular are Cassandra and HBase.
References:
https://docs.microsoft.com/en-us/azure/cosmos-db/graph-introduction
https://www.mongodb.com/scale/types-of-nosql-databases
Question 9
A company is designing a hybrid solution to synchronize data and on-premises Microsoft SQL Server database to Azure SQL Database. 
You must perform an assessment of databases to determine whether data will move without compatibility issues. You need to perform the assessment. 
Which tool should you use?
  1. SQL Server Migration Assistant (SSMA)
  2. Microsoft Assessment and Planning Toolkit
  3. SQL Vulnerability Assessment (VA)
  4. Azure SQL Data Sync
  5. Data Migration Assistant (DMA)
Correct answer: E
Explanation:
The Data Migration Assistant (DMA) helps you upgrade to a modern data platform by detecting compatibility issues that can impact database functionality in your new version of SQL Server or Azure SQL Database. DMA recommends performance and reliability improvements for your target environment and allows you to move your schema, data, and uncontained objects from your source server to your target server. References:https://docs.microsoft.com/en-us/sql/dma/dma-overview
The Data Migration Assistant (DMA) helps you upgrade to a modern data platform by detecting compatibility issues that can impact database functionality in your new version of SQL Server or Azure SQL Database. DMA recommends performance and reliability improvements for your target environment and allows you to move your schema, data, and uncontained objects from your source server to your target server. 
References:
https://docs.microsoft.com/en-us/sql/dma/dma-overview
Question 10
A company plans to use Azure SQL Database to support a mission-critical application. 
The application must be highly available without performance degradation during maintenance windows. 
You need to implement the solution. 
Which three technologies should you implement? Each correct answer presents part of the solution. 
NOTE: Each correct selection is worth one point.
  1. Premium service tier
  2. Virtual machine Scale Sets
  3. Basic service tier
  4. SQL Data Sync
  5. Always On availability groups
  6. Zone-redundant configuration
Correct answer: AEF
Explanation:
A: Premium/business critical service tier model that is based on a cluster of database engine processes. This architectural model relies on a fact that there is always a quorum of available database engine nodes and has minimal performance impact on your workload even during maintenance activities.E: In the premium model, Azure SQL database integrates compute and storage on the single node. High availability in this architectural model is achieved by replication of compute (SQL Server Database Engine process) and storage (locally attached SSD) deployed in 4-node cluster, using technology similar to SQL Server Always On Availability Groups.       F: Zone redundant configurationBy default, the quorum-set replicas for the local storage configurations are created in the same datacenter. With the introduction of Azure Availability Zones, you have the ability to place the different replicas in the quorum-sets to different availability zones in the same region. To eliminate a single point of failure, the control ring is also duplicated across multiple zones as three gateway rings (GW). References:https://docs.microsoft.com/en-us/azure/sql-database/sql-database-high-availability
A: Premium/business critical service tier model that is based on a cluster of database engine processes. This architectural model relies on a fact that there is always a quorum of available database engine nodes and has minimal performance impact on your workload even during maintenance activities.
E: In the premium model, Azure SQL database integrates compute and storage on the single node. High availability in this architectural model is achieved by replication of compute (SQL Server Database Engine process) and storage (locally attached SSD) deployed in 4-node cluster, using technology similar to SQL Server Always On Availability Groups.
   
 
   
F: Zone redundant configuration
By default, the quorum-set replicas for the local storage configurations are created in the same datacenter. With the introduction of Azure Availability Zones, you have the ability to place the different replicas in the quorum-sets to different availability zones in the same region. To eliminate a single point of failure, the control ring is also duplicated across multiple zones as three gateway rings (GW). 
References:
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-high-availability
HOW TO OPEN VCE FILES

Use VCE Exam Simulator to open VCE files
Avanaset

HOW TO OPEN VCEX AND EXAM FILES

Use ProfExam Simulator to open VCEX and EXAM files
ProfExam Screen

ProfExam
ProfExam at a 20% markdown

You have the opportunity to purchase ProfExam at a 20% reduced price

Get Now!