Download Data Engineering on Microsoft Azure (beta).DP-203.NetworkArmy.2021-07-14.61q.vcex

Vendor: Microsoft
Exam Code: DP-203
Exam Name: Data Engineering on Microsoft Azure (beta)
Date: Jul 14, 2021
File Size: 5 MB

How to open VCEX files?

Files with VCEX extension can be opened by ProfExam Simulator.

Demo Questions

Question 1
You need to implement the surrogate key for the retail store table. The solution must meet the sales transaction dataset requirements. 
What should you create? 
  1. a table that has an IDENTITY property
  2. a system-versioned temporal table
  3. a user-defined SEQUENCE object
  4. a table that has a FOREIGN KEY constraint
Correct answer: A
Explanation:
Scenario: Implement a surrogate key to account for changes to the retail store addresses. A surrogate key on a table is a column with a unique identifier for each row. The key is not generated from the table data. Data modelers like to create surrogate keys on their tables when they design data warehouse models. You can use the IDENTITY property to achieve this goal simply and effectively without affecting load performance. Reference:https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-tables-identity 
Scenario: Implement a surrogate key to account for changes to the retail store addresses. 
A surrogate key on a table is a column with a unique identifier for each row. The key is not generated from the table data. Data modelers like to create surrogate keys on their tables when they design data warehouse models. You can use the IDENTITY property to achieve this goal simply and effectively without affecting load performance. 
Reference:
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-tables-identity 
Question 2
You need to ensure that the Twitter feed data can be analyzed in the dedicated SQL pool. The solution must meet the customer sentiment analytic requirements.  
Which three Transact-SQL DDL commands should you run in sequence? 
To answer, move the appropriate commands from the list of commands to the answer area and arrange them in the correct order.  
NOTE: More than one order of answer choices is correct. You will receive credit for any of the correct orders you select. 
 
Correct answer: To work with this question, an Exam Simulator is required.
Explanation:
Scenario: Allow Contoso users to use PolyBase in an Azure Synapse Analytics dedicated SQL pool to query the content of the data records that host the Twitter feeds. Data must be protected by using row-level security (RLS). The users must be authenticated by using their own Azure AD credentials.  Box 1: CREATE EXTERNAL DATA SOURCE External data sources are used to connect to storage accounts.  Box 2: CREATE EXTERNAL FILE FORMAT CREATE EXTERNAL FILE FORMAT creates an external file format object that defines external data stored in Azure Blob Storage or Azure Data Lake Storage. Creating an external file format is a prerequisite for creating an external table.  Box 3: CREATE EXTERNAL TABLE AS SELECT When used in conjunction with the CREATE TABLE AS SELECT statement, selecting from an external table imports data into a table within the SQL pool. In addition to the COPY statement, external tables are useful for loading data.  Incorrect Answers: CREATE EXTERNAL TABLE  The CREATE EXTERNAL TABLE command creates an external table for Synapse SQL to access data stored in Azure Blob Storage or Azure Data Lake Storage.  Reference: https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/develop-tables-external-tables
Scenario: Allow Contoso users to use PolyBase in an Azure Synapse Analytics dedicated SQL pool to query the content of the data records that host the Twitter feeds. 
Data must be protected by using row-level security (RLS). The users must be authenticated by using their own Azure AD credentials.  
Box 1: CREATE EXTERNAL DATA SOURCE External data sources are used to connect to storage accounts.  
Box 2: CREATE EXTERNAL FILE FORMAT CREATE EXTERNAL FILE FORMAT creates an external file format object that defines external data stored in Azure Blob Storage or Azure Data Lake Storage. Creating an external file format is a prerequisite for creating an external table.  
Box 3: CREATE EXTERNAL TABLE AS SELECT 
When used in conjunction with the CREATE TABLE AS SELECT statement, selecting from an external table imports data into a table within the SQL pool. In addition to the COPY statement, external tables are useful for loading data.  
Incorrect Answers: 
CREATE EXTERNAL TABLE  
The CREATE EXTERNAL TABLE command creates an external table for Synapse SQL to access data stored in Azure Blob Storage or Azure Data Lake Storage.  
Reference: 
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/develop-tables-external-tables
Question 3
You need to design the partitions for the product sales transactions. The solution must meet the sales transaction dataset requirements.  
What should you include in the solution? To answer, select the appropriate options in the answer area.  
NOTE: Each correct selection is worth one point. 
Correct answer: To work with this question, an Exam Simulator is required.
Explanation:
Box 1: Sales date Scenario: Contoso requirements for data integration include: Partition data that contains sales transaction records. Partitions must be designed to provide efficient loads by month. Boundary values must belong to the partition on the right.  Box 2: An Azure Synapse Analytics Dedicated SQL pool Scenario: Contoso requirements for data integration include: Ensure that data storage costs and performance are predictable.  The size of a dedicated SQL pool (formerly SQL DW) is determined by Data Warehousing Units (DWU).  Dedicated SQL pool (formerly SQL DW) stores data in relational tables with columnar storage. This format significantly reduces the data storage costs, and improves query performance.  Synapse analytics dedicated sql pool  Reference: https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-overview-what-is
Box 1: Sales date 
Scenario: Contoso requirements for data integration include: 
  • Partition data that contains sales transaction records. Partitions must be designed to provide efficient loads by month. Boundary values must belong to the partition on the right.  
Box 2: An Azure Synapse Analytics Dedicated SQL pool Scenario: Contoso requirements for data integration include: 
  • Ensure that data storage costs and performance are predictable.  
The size of a dedicated SQL pool (formerly SQL DW) is determined by Data Warehousing Units (DWU).  
Dedicated SQL pool (formerly SQL DW) stores data in relational tables with columnar storage. 
This format significantly reduces the data storage costs, and improves query performance.  
Synapse analytics dedicated sql pool  
Reference: 
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-overview-what-is
Question 4
You are designing a fact table named FactPurchase in an Azure Synapse Analytics dedicated SQL pool. 
The table contains purchases from suppliers for a retail store. FactPurchase will contain the following columns.  
   
FactPurchase will have 1 million rows of data added daily and will contain three years of data.  
Transact-SQL queries similar to the following query will be executed daily.  
SELECT  
  SupplierKey, StockItemKey, COUNT(*)  
FROM FactPurchase  
WHERE DateKey >= 20210101  
  AND DateKey <= 20210131  
GROUP By SupplierKey, StockItemKey 
Which table distribution will minimize query times?
  1. replicated
  2. hash-distributed on PurchaseKey
  3. round-robin
  4. hash-distributed on DateKey
Correct answer: B
Explanation:
Hash-distributed tables improve query performance on large fact tables, and are the focus of this article. Round-robin tables are useful for improving loading speed. Incorrect:Not D: Do not use a date column. . All data for the same date lands in the same distribution. If several users are all filtering on the same date, then only 1 of the 60 distributions do all the processing work. Reference:https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-tables-distribute
Hash-distributed tables improve query performance on large fact tables, and are the focus of this article. 
Round-robin tables are useful for improving loading speed. 
Incorrect:
Not D: Do not use a date column. . All data for the same date lands in the same distribution. If several users are all filtering on the same date, then only 1 of the 60 distributions do all the processing work. 
Reference:
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-tables-distribute
Question 5
You have a table in an Azure Synapse Analytics dedicated SQL pool. The table was created by using the following Transact-SQL statement. 
   
You need to alter the table to meet the following requirements:
  • Ensure that users can identify the current manager of employees.  
  • Support creating an employee reporting hierarchy for your entire company.  
  • Provide fast lookup of the managers’ attributes such as name and job title.  
Which column should you add to the table?  
  1. [ManagerEmployeeID] [int] NULL
  2. [ManagerEmployeeID] [smallint] NULL
  3. [ManagerEmployeeKey] [int] NULL
  4. [ManagerName] [varchar](200) NULL
Correct answer: A
Explanation:
Use the same definition as the EmployeeID column. Reference:https://docs.microsoft.com/en-us/analysis-services/tabular-models/hierarchies-ssas-tabular
Use the same definition as the EmployeeID column. 
Reference:
https://docs.microsoft.com/en-us/analysis-services/tabular-models/hierarchies-ssas-tabular
Question 6
You have an Azure Synapse workspace named MyWorkspace that contains an Apache Spark database named mytestdb.  
You run the following command in an Azure Synapse Analytics Spark pool in MyWorkspace.  
CREATE TABLE mytestdb.myParquetTable( 
  EmployeeID int,  
  EmployeeName string,  
  EmployeeStartDate date)  
USING Parquet  
You then use Spark to insert a row into mytestdb.myParquetTable. The row contains the following data.  
   
One minute later, you execute the following query from a serverless SQL pool in MyWorkspace.  
SELECT EmployeeID  
FROM mytestdb.dbo.myParquetTable  
WHERE name = 'Alice'; 
What will be returned by the query?
  1. 24
  2. an error
  3. a null value
Correct answer: A
Explanation:
Once a database has been created by a Spark job, you can create tables in it with Spark that use Parquet as the storage format. Table names will be converted to lower case and need to be queried using the lower case name. These tables will immediately become available for querying by any of the Azure Synapse workspace Spark pools. They can also be used from any of the Spark jobs subject to permissions. Note: For external tables, since they are synchronized to serverless SQL pool asynchronously, there will be a delay until they appear. Reference:https://docs.microsoft.com/en-us/azure/synapse-analytics/metadata/table
Once a database has been created by a Spark job, you can create tables in it with Spark that use Parquet as the storage format. Table names will be converted to lower case and need to be queried using the lower case name. These tables will immediately become available for querying by any of the Azure Synapse workspace Spark pools. They can also be used from any of the Spark jobs subject to permissions. 
Note: For external tables, since they are synchronized to serverless SQL pool asynchronously, there will be a delay until they appear. 
Reference:
https://docs.microsoft.com/en-us/azure/synapse-analytics/metadata/table
Question 7
You have files and folders in Azure Data Lake Storage Gen2 for an Azure Synapse workspace as shown in the following exhibit.  
   
You create an external table named ExtTable that has LOCATION='/topfolder/'. 
When you query ExtTable by using an Azure Synapse Analytics serverless SQL pool, which files are returned?
  1. File2.csv and File3.csv only
  2. File1.csv and File4.csv only
  3. File1.csv, File2.csv, File3.csv, and File4.csv
  4. File1.csv only
Correct answer: C
Explanation:
To run a T-SQL query over a set of files within a folder or set of folders while treating them as a single entity or rowset, provide a path to a folder or a pattern (using wildcards) over a set of files or folders. Reference:https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/query-data-storage#query-multiple-files-or-folders
To run a T-SQL query over a set of files within a folder or set of folders while treating them as a single entity or rowset, provide a path to a folder or a pattern (using wildcards) over a set of files or folders. 
Reference:
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/query-data-storage#query-multiple-files-or-folders
Question 8
You are designing the folder structure for an Azure Data Lake Storage Gen2 container.  
Users will query data by using a variety of services including Azure Databricks and Azure Synapse Analytics serverless SQL pools. The data will be secured by subject area. Most queries will include data from the current year or current month.  
Which folder structure should you recommend to support fast queries and simplified folder security?  
  1. /{SubjectArea}/{DataSource}/{DD}/{MM}/{YYYY}/{FileData}_{YYYY}_{MM}_{DD}.csv
  2. /{DD}/{MM}/{YYYY}/{SubjectArea}/{DataSource}/{FileData}_{YYYY}_{MM}_{DD}.csv
  3. /{YYYY}/{MM}/{DD}/{SubjectArea}/{DataSource}/{FileData}_{YYYY}_{MM}_{DD}.csv
  4. /{SubjectArea}/{DataSource}/{YYYY}/{MM}/{DD}/{FileData}_{YYYY}_{MM}_{DD}.csv
Correct answer: D
Explanation:
There's an important reason to put the date at the end of the directory structure. If you want to lock down certain regions or subject matters to users/groups, then you can easily do so with the POSIX permissions. Otherwise, if there was a need to restrict a certain security group to viewing just the UK data or certain planes, with the date structure in front a separate permission would be required for numerous directories under every hour directory. Additionally, having the date structure in front would exponentially increase the number of directories as time went on. Note: In IoT workloads, there can be a great deal of data being landed in the data store that spans across numerous products, devices, organizations, and customers. It’s important to pre-plan the directory layout for organization, security, and efficient processing of the data for down-stream consumers. A general template to consider might be the following layout:{Region}/{SubjectMatter(s)}/{yyyy}/{mm}/{dd}/{hh}/
There's an important reason to put the date at the end of the directory structure. If you want to lock down certain regions or subject matters to users/groups, then you can easily do so with the POSIX permissions. 
Otherwise, if there was a need to restrict a certain security group to viewing just the UK data or certain planes, with the date structure in front a separate permission would be required for numerous directories under every hour directory. Additionally, having the date structure in front would exponentially increase the number of directories as time went on. 
Note: In IoT workloads, there can be a great deal of data being landed in the data store that spans across numerous products, devices, organizations, and customers. It’s important to pre-plan the directory layout for organization, security, and efficient processing of the data for down-stream consumers. A general template to consider might be the following layout:
{Region}/{SubjectMatter(s)}/{yyyy}/{mm}/{dd}/{hh}/
Question 9
You need to design an Azure Synapse Analytics dedicated SQL pool that meets the following requirements: 
  • Can return an employee record from a given point in time.  
  • Maintains the latest employee information.  
  • Minimizes query complexity.  
How should you model the employee data?
  1. as a temporal table
  2. as a SQL graph table
  3. as a degenerate dimension table
  4. as a Type 2 slowly changing dimension (SCD) table
Correct answer: D
Explanation:
A Type 2 SCD supports versioning of dimension members. Often the source system doesn't store versions, so the data warehouse load process detects and manages changes in a dimension table. In this case, the dimension table must use a surrogate key to provide a unique reference to a version of the dimension member. It also includes columns that define the date range validity of the version (for example, StartDate and EndDate) and possibly a flag column (for example, IsCurrent) to easily filter by current dimension members. Reference:https://docs.microsoft.com/en-us/learn/modules/populate-slowly-changing-dimensions-azure-synapse-analytics-pipelines/3-choose-between-dimension-types
A Type 2 SCD supports versioning of dimension members. Often the source system doesn't store versions, so the data warehouse load process detects and manages changes in a dimension table. In this case, the dimension table must use a surrogate key to provide a unique reference to a version of the dimension member. It also includes columns that define the date range validity of the version (for example, StartDate and EndDate) and possibly a flag column (for example, IsCurrent) to easily filter by current dimension members. 
Reference:
https://docs.microsoft.com/en-us/learn/modules/populate-slowly-changing-dimensions-azure-synapse-analytics-pipelines/3-choose-between-dimension-types
Question 10
You have an enterprise-wide Azure Data Lake Storage Gen2 account. The data lake is accessible only through an Azure virtual network named VNET1. 
You are building a SQL pool in Azure Synapse that will use data from the data lake.  
Your company has a sales team. All the members of the sales team are in an Azure Active Directory group named Sales. POSIX controls are used to assign the Sales group access to the files in the data lake.  
You plan to load data to the SQL pool every hour.  
You need to ensure that the SQL pool can load the sales data from the data lake. 
Which three actions should you perform? Each correct answer presents part of the solution.  
NOTE: Each area selection is worth one point.
  1. Add the managed identity to the Sales group.
  2. Use the managed identity as the credentials for the data load process.
  3. Create a shared access signature (SAS).
  4. Add your Azure Active Directory (Azure AD) account to the Sales group.
  5. Use the snared access signature (SAS) as the credentials for the data load process.
  6. Create a managed identity.
Correct answer: ADF
Explanation:
The managed identity grants permissions to the dedicated SQL pools in the workspace. Note: Managed identity for Azure resources is a feature of Azure Active Directory. The feature provides Azure services with an automatically managed identity in Azure AD Reference:https://docs.microsoft.com/en-us/azure/synapse-analytics/security/synapse-workspace-managed-identity
The managed identity grants permissions to the dedicated SQL pools in the workspace. 
Note: Managed identity for Azure resources is a feature of Azure Active Directory. The feature provides Azure services with an automatically managed identity in Azure AD 
Reference:
https://docs.microsoft.com/en-us/azure/synapse-analytics/security/synapse-workspace-managed-identity
HOW TO OPEN VCE FILES

Use VCE Exam Simulator to open VCE files
Avanaset

HOW TO OPEN VCEX AND EXAM FILES

Use ProfExam Simulator to open VCEX and EXAM files
ProfExam Screen

ProfExam
ProfExam at a 20% markdown

You have the opportunity to purchase ProfExam at a 20% reduced price

Get Now!