DP200 - PracticeTests 1 AnswersAndExplanation
DP200 - PracticeTests 1 AnswersAndExplanation
Question 1
]A.
Azure Advanced Threat Protection
]B.
Auditing
]C.
Transparent Data Encryption
]D.
Azure Monitor Audit Logs
Explanation:
Answer – B
You can use Azure SQL Database auditing to get insights into how the data in your database is
being accessed.
Option C is incorrect since this is used to encrypt data at rest in SQL databases.
Option D is incorrect since this is used to look at the control plane activities carried out on
Azure resources.
For more information on Azure SQL Database auditing, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-auditing
Question 2
]A.
Resource Group
]B.
SQL Server
]C.
SQL Data Warehouse
]D.
Subscription
Explanation:
Answer – C
Since we need to create an alert based on the consumption of the Data Warehouse itself, we
should create an alert on the Warehouse itself.
The other options are incorrect since we need to ensure the monitoring is enabled on the Data
Warehouse itself.
For more information on querying Azure SQL Data Warehouse, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-
concept-resource-utilization-query-activity
Question 3
]A.
CPU used
]B.
DWU limit
]C.
DWU used
]D.
Data IO Percentage
Explanation:
Answer – C
You would monitor the number of DWU (Data Warehousing Units) being consumed.
Since this is a measurement for performance for Azure SQL Data Warehouses, all other options
are incorrect.
For more information on Azure SQL Data Warehouse DWU’s, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/what-is-a-data-warehouse-unit-dwu-
cdwu
Question 4
]A.Yes
]B.No
Explanation:
Answer – A
For more information on Automatic Tuning, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-automatic-tuning-
enable
Question 5
]A.Yes
]B.No
Explanation:
Answer – B
For more information, please refer to the below links:
https://docs.microsoft.com/en-us/azure/azure-sql/database/automatic-tuning-enable#enable-
automatic-tuning-on-server
https://docs.microsoft.com/en-us/azure/azure-sql/database/automatic-tuning-overview
Question 6
]A.Yes
]B.No
Explanation:
Answer – B
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-automatic-tuning-
enable
Question 7
]A.Yes
]B.No
Explanation:
Answer – B
The right option is to have a lookup collection and use the ProductName as the partition key,
but you also need to use OrderId as the value.
For more information on Cosmos DB partitioning, please visit the following URL-
https://azure.microsoft.com/en-us/blog/azure-cosmos-db-partitioning-design-patterns-
part-1/
Question 8
]A.Yes
]B.No
Explanation:
Answer – A
The right option is to have a lookup collection and use the ProductName as the partition key and
use OrderId as the value.
For more information on Cosmos DB partitioning, please visit the following URL-
https://azure.microsoft.com/en-us/blog/azure-cosmos-db-partitioning-design-patterns-
part-1/
Question 9
]A.Yes
]B.No
Explanation:
Answer – B
The Partition Key can only be based on a single attribute. The right option is to have a lookup
collection and use the ProductName as the partition key and use OrderId as the value.
For more information on Cosmos DB partitioning, please visit the following URL-
https://azure.microsoft.com/en-us/blog/azure-cosmos-db-partitioning-design-patterns-
part-1/
Question 10
]B.
Low Concurrency
]C.
Premium
]D.
Standard
Explanation:
Answer – A
The documentation for Azure Data bricks mentions that you should ideally set the Cluster Mode
to High Concurrency.
Since this is clearly mentioned in the documentation, all other options are incorrect.
For more information on using Azure AD for Azure Data bricks, please visit the following
URL-
https://docs.azuredatabricks.net/data/data-sources/azure/adls-passthrough.html
Question 11
]A.
Blob access control
]B.
Table access control
]C.
Credential Passthrough
]D.
Single Sign-On
Explanation:
Answer – C
The documentation for Azure Data bricks mentions that you should set Credential Passthrough
in the Advanced option.
Since this is clearly mentioned in the documentation, all other options are incorrect.
For more information on using Azure AD for Azure Data bricks, please visit the following
URL-
https://docs.azuredatabricks.net/data/data-sources/azure/adls-passthrough.html
Question 12
]A.
Azure
]B.
Self-Hosted
]C.
Primary
]D.
Azure-SSIS
Explanation:
Answer – A
You can use the Azure Integration runtime type to copy data between Azure-based resources.
For more information on the Integration runtime, please visit the following URL-
https://docs.microsoft.com/en-us/azure/data-factory/concepts-integration-runtime
Question 13
]B.
Session
]C.
Sliding
]D.
Tumbling
Explanation:
Answer – D
You need to use the Tumbling windowing function for this requirement.
For more information on the windowing functions, please visit the following URL-
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-window-
functions
Question 14
]A.
Hopping
]B.
Session
]C.
Sliding
]D.
Tumbling
Explanation:
Answer – A
You need to use the Hopping windowing function for this requirement.
For more information on the windowing functions, please visit the following URL-
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-window-
functions
Question 15
]B.
Session
]C.
Sliding
]D.
Tumbling
Explanation:
Answer – C
You need to use the Sliding windowing function for this requirement.
For more information on the windowing functions, please visit the following URL-
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-window-
functions
Question 16
A.
Write the results onto Azure Data Lake Storage.
B.
Drop the data frame.
C.
Perform transformations on the data frame.
D.
Mount the Data Lake Storage onto DBFS.
E.
Perform transformations on the file.
F.
Read the file into a data frame.
G.
Specify a temporary folder to stage the data.
H.
Write the results to a table in SQL Data Warehouse.
Explanation:
Answer - C, D, F, G and H
To connect to Azure Data Lake Gen 2 storage, we first need to mount the file system. This is
also given in the Azure data bricks documentation.
Next, we need to load the data from the Azure Data Lake Storage Gen2 account. An example is
also given in the Microsoft documentation.
Next, we need to transform the data in the data frame to combine the FirstName and LastName
values.
And finally, you copy the data onto the Azure SQL data warehouse.
Option A is incorrect since you don’t need to write the results onto Azure Data Lake storage.
The results are written to the Azure SQL Data warehouse table.
For more information on loading data from an Azure SQL Data warehouse, please visit the
following URL-
https://docs.microsoft.com/en-us/azure/azure-databricks/databricks-extract-load-sql-
data-warehouse#load-data-into-azure-sql-data-warehouse
Question 17
]A.
%
]B.
#
]C.
@
]D.
&
Explanation:
Answer – A
This can be done with the use of the % command. This is also mentioned in the databricks
documentation.
Since this is clearly given in the documentation, all other options are incorrect.
For more information on developing notebooks, please visit the following URL-
https://docs.databricks.com/notebooks/notebooks-use.html#develop-notebooks
Question 18
A.
Create an external data source and ensure to use the abfs location.
B.
Create an external data source and ensure to use the Hadoop location.
C.
Create an external file format and set the First_row option.
D.
Create a database scoped credential that uses OAuth2 token and a key.
E.
Use the CREATE EXTERNAL TABLE AS SELECT and create a view that removes the
empty row.
Explanation:
Answer – A, C and D
The Microsoft documentation highlights the steps required to load data from Azure Data Lake
Gen 2 to an Azure SQL Data warehouse.
Another step is to create the external data source using ‘abfs’ as the file location.
And you can use the FIRST_ROW parameter to skip the first row of the file.
For more information on an example of loading data from Azure Data Lake Store and for the
format of the external file format, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-load-
from-azure-data-lake-store
https://docs.microsoft.com/en-us/sql/t-sql/statements/create-external-file-format-
transact-sql?view=sql-server-ver15
Question 19
]A.
Strong
]B.
Bounded Staleness
]C.
Eventual
]D.
Session
]E.
Consistent Prefix
Explanation:
Answer – D
If you look at the Microsoft documentation, the Session consistency level matches the
requirement.
Options A and B are incorrect since these would lead to more latency and less throughput
Option C and E are incorrect since these would not provide session-based consistency.
For more information on Cosmos DB consistency levels, please visit the following URL-
https://docs.microsoft.com/en-us/azure/cosmos-db/consistency-levels
Question 20
A.
Create a new empty table named comp_salesfact_new that has the same schema as
comp_salesfact.
B.
Drop the comp_salesfact_new table.
C.
Copy the data to the new table by using CREATE TABLE AS SELECT (CTAS).
D.
Truncate the partition containing the stale data.
E.
Switch the partition containing the stale data from comp_salesfact to comp_salesfact_new.
F.
Execute the DELETE statement where the value in the Date column is greater than 12
months.
Explanation:
Answer – B, C and E
An example of this is given in a blog post. To achieve this, we first need to copy the data to a
new table using the “CREATE TABLE AS SELECT” command. Then we switch the partition
and then delete the staging table.
Option A is incorrect because we also need to copy the data onto the new table.
Option F is incorrect because issuing the DELETE statement would take time.
For more information on the blog article, please visit the following URL-
https://blogs.msdn.microsoft.com/apsblog/2018/06/18/azure-sql-dw-performance-
ctaspartition-switching-vs-updatedelete/
Question 21
]A.
ALTER EXTERNAL TABLE [Ext].[compitems] ADD [ItemID] int;
[ItemDescription] nvarchar(250))
WITH
LOCATION=’/Items/’,
DATA_SOURCE=AzureDataLakeStore,
FILE_FORMAT=PARQUET,
REJECT_TYPE=VALUE,
REJECT_VALUE=0
);
Explanation:
Answer – D
You can’t alter the external table to add a new column. You would need to drop the external
table and recreate the table again.
For more information to create the external table, please visit the following URL-
https://docs.microsoft.com/en-us/sql/t-sql/statements/create-external-file-format-
transact-sql?view=sql-server-ver15
Question 22
]A.
Hash distributed
]B.
Heap
]C.
Replicated
]D.
Round-Robin
Explanation:
Answer – C
The Microsoft documentation mentions that the ‘Replicated’ table type would be ideal for tables
less than 2 GB of size.
Since this is clearly given in the documentation, all other options are incorrect.
For more information on a cheat sheet for designing tables, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/cheat-sheet
Question 23
]A.Yes
]B.No
Explanation:
Answer – B
The Credit card masking function is specifically required for those columns that store credit
card information and cannot be used for this requirement.
For more information on dynamic data masking, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-dynamic-data-
masking-get-started
Question 24
]A.Yes
]B.No
Explanation:
Answer - B
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-dynamic-data-
masking-get-started
Question 25
]A.Yes
]B.No
Explanation:
Answer - B
The email masking function is specifically required for those columns that store email
information and cannot be used for this requirement.
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-dynamic-data-
masking-get-started
Question 26
A.
A database encryption key
B.
An Asymmetric key
C.
An external data source
D.
An external file format
E.
A database scoped credential
Explanation:
Answer – C, D and E
The Microsoft documentation mentions the steps required to load data into Azure SQL Data
warehouse from Azure Data Lake storage accounts.
First, you have to create a database master key and a database scoped credential.
Then you have to create an external data source.
For more information on loading data from Azure Data Lake storage onto SQL Data warehouse,
please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-load-
from-azure-data-lake-store
Question 27
]A.
columnstore index
]B.
non-durable table
]C.
materialized view
]D.
memory-optimized table
Explanation:
Answer – D
You can use Memory-optimized technologies to speed up read operations on transactional data.
Option B is incorrect since this is used for storing temporary data which is used for caching.
Option C is incorrect since this is used for creating a view from multiple tables.
For more information on using In-Memory technologies, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-in-memory
Question 28
]A.
columnstore index
]B.
non-durable table
]C.
materialized view
]D.
memory-optimized table
Explanation:
Answer – C
Here it is better to create a materialized view which can be used for faster comparison of code
snippets.
Option A is incorrect since this is ideal when you want to perform analytical queries on the
table.
Option B is incorrect since this is used for storing temporary data which is used for caching.
Option D is incorrect since this should be used as the storage design for the underlying tables.
For more information on Materialized views, please visit the following URL-
https://docs.microsoft.com/en-us/azure/architecture/patterns/materialized-view
Question 29
]A.
‘Java exception raised on call to HdfsBridge_Connect:Error
[com.microsoft.polybase.client.KerberosSecureLogin] occurred while accessing external
files’
]B.
Cannot execute the query “Remote Query” against OLE DB provider “SQLNCLI11”: for
linked server “(null)”, Query aborted-the maximum reject threshold (0 rows) was reached
while reading from external source
]C.
‘Java exception raised on call to HdfsBridge_Connect:Error[Unable to Instantiate
LoginClass] occurred while accessing external files.’
]D.
‘Java exception raised on call to HdfsBridge_Connect:Error[No FileSystem for
schema:wabs]
Explanation:
Answer – B
The Microsoft tech community mentions the error to look for when there is an invalid schema.
Since this is clearly mentioned, all other options are invalid.
For more information on the article, please visit the following URL-
https://techcommunity.microsoft.com/t5/DataCAT/PolyBase-Setup-Errors-and-
Possible-Solutions/ba-p/305297
Question 30
A.
In Azure Monitor, go ahead and create a new data source.
B.
Configure the current Azure Databricks cluster to use the monitoring library.
C.
Deploy an instance of Grafana to a new Azure virtual machine.
D.
Build the spark-listeners-loganalytics-1.0-SNAPSHOT.jar file.
E.
Create the required Dropwizard counters in the application code.
Explanation:
Answer – B, D and E
The steps for sending application logs and metrics to Azure Monitor are given in the Microsoft
documentation.
For sending Apache Spark level events and Spark Structured Streaming metrics, you can use the
monitoring library.
For more information on Azure Databricks monitoring, please refer to the following link-
https://docs.microsoft.com/en-us/azure/architecture/databricks-monitoring/application-
logs
Question 31
A.
Create a database encryption key using a certificate.
B.
Create a certificate protected by the master key.
C.
Set the context to the master database.
D.
Create a master key using a password.
E.
Set the context to the company database.
F.
Enable Encryption.
Explanation:
Answer – A, B, D and F
For more information on Transparent Data Encryption, please visit the following URL-
https://docs.microsoft.com/en-us/sql/relational-databases/security/encryption/
transparent-data-encryption?view=sql-server-ver15
Question 32
]A.
Azure Cosmos DB
]B.
Azure SQL Database
]C.
Azure Storage Account – Blob
]D.
Azure SQL Datawarehouse
Explanation:
Answer – A
This code is used to connect to an Azure Cosmos DB database. An example of the code is given
in the Microsoft documentation.
Since this is clearly mentioned in the Microsoft documentation, all other options are incorrect.
For more information on working with Cosmos DB from .Net, please visit the following URL-
https://docs.microsoft.com/en-us/azure/cosmos-db/sql-api-get-started
Question 33
]A.
Resource token
]B.
Master Key
]C.
Certificate
]D.
Password
Explanation:
Answer – B
Here the master keys of the account are being used for authentication.
Since this is clearly mentioned in the Microsoft documentation, all other options are incorrect.
For more information on securing access to data, please visit the following URL-
https://docs.microsoft.com/en-us/azure/cosmos-db/secure-access-to-data
Question 34
Domain :Implement data storage solutions
A company is planning to set up an Azure SQL database to store sensitive data. The company
wants to monitor data usage and data copied from the system to prevent data leakage. The
company also wants to configure the Azure SQL database to email a specific user when the data
leakage occurs. Which of the following activities would you need to perform? Choose 3 answers
from the options given below.
A.
In Auditing, enable the auditing feature.
B.
Configure the service to create alerts for threat detections of the type “Data Exfiltration”.
C.
In the Firewalls and virtual networks section, enable “Allow access to Azure services”.
D.
Enable Advanced threat protection.
E.
Configure the service to send email alerts to the IT security administrator.
Explanation:
Answer – B, D and E
For more information on Advanced threat protection, please visit the following URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-threat-detection
Question 35
]A.
CPU Percentage
]B.
DWU used
]C.
DWU percentage
]D.
Data IO percentage
Explanation:
Answer – B
The resources allocated to the data warehouse are based on a metric called DWU or Data
Warehouse Units. This determines the resources allocated to the data warehouse.
Option A and D are incorrect since DWU’s are used for allocated resources. But please note that
you can still monitor these aspects as well for your data warehouse.
Option C is incorrect because the right DWU to measure is the amount that has been used up.
For more information on Data Warehouse Units, please refer to the following link-
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/what-is-
a-data-warehouse-unit-dwu-cdwu
Question 36
]A.
Azure SQL Database
]B.
Azure Blob storage
]C.
Azure Cosmos DB
]D.
Azure Data Lake Storage
Explanation:
Answer – D
Azure Data Lake Storage fulfills all of the right aspects as being built for Big Data Analytics. It
can also scale in terms of storage. Here you can store different types of files, different file sizes
as well.
For more information on Azure Data Lake Storage, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-introduction
Question 37
]B.
HDInsight Hadoop
]C.
Azure Databricks
]D.
HDInsight Interactive Query
Explanation:
Answer – C
With Azure Databricks, you can set up clusters that can be terminated automatically. You can
configure these clusters to perform computational processing. You can make use of Notebooks
based on different programming languages.
For more information on Azure Databricks, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/azure-databricks/what-is-azure-databricks
Question 38
]A.
HDInsight Base
]B.
Azure SQL Data warehouse
]C.
Azure Analysis services
]D.
Azure Cosmos DB
Explanation:
Answer – B
For columnar storage, you can make use of Azure SQL data warehouse.
For more information on Azure SQL data warehouse, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-
overview-what-is
Question 39
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
Which of the following should be used as the API for the Cosmos DB account?
]A.
Cassandra
]B.
Gremlin
]C.
MongoDB
]D.
SQL
]E.
Table
Explanation:
Answer – C
Since the on-premise data store is a MongoDB database, it makes logical sense to migrate it to
the MongoDB API in the Cosmos DB account.
Since this is the most logical solution, all other options are incorrect.
For more information on Azure Cosmos DB MongoDB API, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/cosmos-db/mongodb-introduction
Question 40
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
Which of the following would you use for the consistency level for the database?
]A.
Eventual
]B.
Session
]C.
Strong
]D.
Consistent Prefix
Explanation:
Answer – A
There is a requirement for data to be written to the closest data center for Cosmos DB. We need
to ensure a multi-master setup for Cosmos DB wherein data can be written from multiple
regions. For such accounts, we can’t set the consistency level to Strong.
Because of the proposed logic to the consistency level, all other options are incorrect.
For more information on Azure Cosmos DB consistency levels and trade-offs, one can visit the
below URL-
https://docs.microsoft.com/en-us/azure/cosmos-db/consistency-levels-tradeoffs
https://docs.microsoft.com/en-us/azure/cosmos-db/consistency-levels
Question 41
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
You need to build the Azure SQL Data warehouse data store. Which of the following would you
use as the underlying table type?
]A.
Hash distributed
]B.
Replicated
]C.
Round-Robin
]D.
Primary
Explanation:
Answer – A
Since the data size is large and data is aggregated using a column, we can use the Hash
distributed table type to get the best performance.
For more information on Azure SQL Data warehouse table types, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-tables-
distribute
Question 42
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
You need to build the Azure SQL Data warehouse data store. Which of the following would you
use as the underlying index type?
]A.
Clustered
]B.
Clustered column store
]C.
Heap
]D.
Nonclustered
Explanation:
Answer – B
If you have large tables, you should consider a Clustered column store index.
Since this is a clear candidate for the index type, all other options are incorrect.
For more information on Azure SQL Data warehouse indexes, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-tables-
index
Question 43
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
Which of the following masking functions should be used for the “carID” column?
]A.
Credit Card
]B.
Default
]C.
Email
]D.
Random number
Explanation:
Answer – A
Only the last four digits of the values in the column carID must be shown.
Hence for this, we can use the Credit Card masking type.
Since this is clear from the Microsoft documentation, all other options are incorrect.
For more information on dynamic masking, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-dynamic-data-
masking-get-started
Question 44
Which of the following masking functions should be used for the “carWeight” column?
]A.
Credit Card
]B.
Default
]C.
Email
]D.
Random number
Explanation:
Answer – B
A zero value must be shown for all values in the column carWeight.
Hence for this, we can use the Default masking type.
Since this is clear from the Microsoft documentation, all other options are incorrect.
For more information on dynamic masking, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-dynamic-data-
masking-get-started
Question 45
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
]A.
A copy activity that needs to use a stored procedure as the source
]B.
A copy activity that needs to use schema mappings
]C.
A delete activity that needs to have logging enabled
]D.
A filter activity that needs to have a condition
Explanation:
Answer – B
In the case study, it mentions that for on-premise the attributes used to change when data was
transferred from MongoDB to SQL Server. The same needs to be done for Azure CosmosDB to
Azure SQL Server.
Since this is the logical step as per the case study, all other options are incorrect.
For more information on schema mapping in Azure Data Factory, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/data-factory/copy-activity-schema-and-type-
mapping
Question 46
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
The team is monitoring the Data Factory pipeline. They can see that the Cosmos DB to SQL
database run time is taking 45 minutes. Which of the following can be carried out to improve
the performance of the job?
]A.
Increase in the number of data integration units.
]B.
Ensure that the copy activity uses a staged copy.
]C.
Ensure that the copy activity performs compression.
]D.
Ensure to decrease the amount pf parallelism in the copy activities.
Explanation:
Answer – A
One of the ways to increase performance is to use more Data Integration Units.
Option C is incorrect since applying compression might slow down the process.
Option D is incorrect since having parallel copies can actually increase throughput.
For more information on increasing performance for Azure Data Factory copy activities, one
can visit the below URL-
https://docs.microsoft.com/en-us/azure/data-factory/copy-activity-performance
Question 47
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
Which of the following can be used to satisfy the case study requirement?
“The query performance for data in the Azure SQL database must be stable without the
need for administrative overhead.”
]A.
sp_update stored procedure
]B.
Using the Query store
]C.
Using the dbcc checkdb command
]D.
Using automatic tuning
Explanation:
Answer – D
You can use automatic tunning which can ensure optimum query performance without the need
for any sort of administrative overhead.
All of the other methods all need administrative intervention.
For more information on Azure SQL Database automatic tuning, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-automatic-tuning
Question 48
Overview
Comps is an online training provider. They also provide a yearly gaming competition for their students.
The competition is held every month in different locations.
Current Environment
The company currently has the following environment in place
The racing cars for the competition send their telemetry data to a MongoDB database.
The telemetry data has around 100 attributes.
A custom application is then used to transfer the data from the MongoDB database to a
SQL Server 2017 database. The attribute names are changed when they are sent to the
SQL Server database.
Another application named “Comp workflow” is then used to perform analytics on the
telemetry data to look for improvements on the racing cars.
The SQL Server 2017 database has a table named “cardata” which has around 1 TB of
data. “Comp workflow” performs the required analytics on the data in this table. Large
aggregations are performed on a column of the table.
Proposed Environment
The company now wants to move the environment to Azure. Below are the key requirements
The racing car data will now be moved to Azure Cosmos DB and Azure SQL database.
The data must be written to the closest Azure data center and must converge in the
least amount of time.
The query performance for data in the Azure SQL database must be stable without the
need of administrative overhead
The data for analytics will be moved to an Azure SQL Data warehouse
Transparent data encryption must be enabled for all data stores wherever possible
An Azure Data Factory pipeline will be used to move data from the Cosmos DB
database to the Azure SQL database. If there is a delay of more than 15 minutes for the
data transfer, then configuration changes need to be made to the pipeline workflow.
The telemetry data must be monitored for any sort of performance issues.
The Request Units for Cosmos DB must be adjusted to maintain the demand while also
minimizing costs.
The data in the Azure SQL Server database must be protected via the following
requirements
o Only the last four digits of the values in the column carID must be
shown
o A zero value must be shown for all values in the column carWeight
You need to monitor the telemetry data being sent to Cosmos DB so that you can decide on the
amount of Request Units to provision for Cosmos DB. Which of the following metrics must you
monitor? Choose 2 answers from the options given below.
A.
The number of requests
B.
The Session consistency
C.
The data and index storage consumed
D.
The Average Throughput/second
Explanation:
Answer – A and C
The Microsoft documentation gives the different elements to consider for the Request Units.
Here we need to consider the number of requests and the amount of storage being consumed.
Since this is clear from the Microsoft documentation, all other options are incorrect.
For more information on Cosmos DB request units, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/cosmos-db/request-units
Question 49
]A.Yes
]B.No
Explanation:
Answer – B
Here we are using the UNION clause which is different from the JOIN clause.
https://docs.microsoft.com/en-us/stream-analytics-query/union-azure-stream-analytics
Question 50
]A.Yes
]B.No
Explanation:
Answer – B
We need to match the partition key in the input and output scheme, but not necessarily need to
match the count.
The Microsoft documentation mentions the following when it comes to the steps for performing
parallel processing.
For more information on parallel processing in Azure Stream Analytics, please refer to the
following link-
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-
parallelization
Question 51
]A.Yes
]B.No
Explanation:
Answer - B
You can scale up to 6 streaming units for each step in a job. If you have partitions, you need to
multiply the number of partitions by 6.
Now in the query, we have 2 select queries in the input streams. And each has a partition count
of 10. That means we can scale the job to the following number of streaming units.
= 2*10*6 = 120
In the query, we have one SELECT statement in the output with no partition count. Hence the
calculation for the maximum number of streaming units is
Hence the total number of streaming units that can be assigned to the job is 126.
Hence this could be the ideal value for the optimization of the job.
For more information on parallel processing in Azure Stream Analytics, please refer to the
following link-
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-
parallelization
Question 52
]A.
Standard
]B.
Basic
]C.
Job
]D.
High concurrency
Explanation:
Answer – A
For more information on cluster configurations, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/databricks/clusters/configure
Question 53
]A.
Standard
]B.
Basic
]C.
Job
]D.
High concurrency
Explanation:
Answer – D
For more information on cluster configurations, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/databricks/clusters/configure
Question 54
]A.
Standard
]B.
Basic
]C.
Job
]D.
High concurrency
Explanation:
Answer – D
Since this is clearly mentioned in the documentation, all other options are incorrect.
For more information on cluster configurations, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/databricks/clusters/configure
Question 55
A.
Azure Event Hubs
B.
Azure Storage
C.
Azure Cosmos DB
D.
Azure SQL Data warehouse
Explanation:
Answer – A and B
The Microsoft documentation provides the ways you can stream the diagnostics log data.
Since this is clearly mentioned in the documentation, all other options are incorrect.
For more information on Azure SQL database diagnostic logging, one can visit the below URL-
https://docs.microsoft.com/en-us/azure/sql-database/sql-database-metrics-diag-logging