DP 203 Questions 2
DP 203 Questions 2
Your data engineering team has an Azure Stream Analytics job in place. Currently the
job is configured to take in events from an Azure Event Hub. It then outputs data to an
Azure Dedicated SQL pool within Azure Synapse Analytics. The engineers have been
reviewing the metrics. They are seeing a high number of Backlogged input events.
Which of the following can be done to ensure the Backlogged input events are kept in
check?
● Add another output to the Stream Analytics job
● Change the partition key of the incoming stream
● Add another input to the Stream Analytics job
● Increase the number of streaming units assigned to the job
● (Correct)
Explanation
One reason for a high value of Backlogged Inputs events could be because the job is not
able to keep up with the incoming stream. By adding more streaming units , it can help
to add more resources to ensure the job can keep up with the incoming streams.
For more information on monitoring for a Stream Analytics job, one can visit the
following URL
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-monitoring
Question 2: Correct
Your data engineering team is planning on setting up a dedicated SQL pool in an Azure
Synapse Analytics workspace. A separate set of users will be responsible for loading
data into the SQL pool. And another set of users will be responsible for querying of data
from the SQL pool. You have to ensure that the loading process has enough resources
assigned to it. Which of the following can be implemented for this requirement?
● Assign more resources via workload classification
● (Correct)
● Make sure to use the COPY statement while loading the data
● Make use of materialized views
Explanation
You need to make use of Workload Classifiers to ensure that more resources are
allocated to the users who will be performing the load process.
For more information on Workload Classifiers, one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-dat
a-warehouse-workload-classification
Question 3: Correct
Your team is designing the tables for a data warehouse. The data warehouse is going to
be hosted in a Dedicated SQL pool in Azure Synapse Analytics. The following tables are
going to be hosted initially in the pool
You have to choose the right distribution for each table. You have to ensure data
movement is minimized across tables
Which of the following distribution type would you choose for the Sales table?
● Hash
● (Correct)
● Round Robin
● Replicated
Explanation
Here since this is a large fact table, you should use Hash distributed tables
For more information on table distribution, one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-dat
a-warehouse-tables-distribute
Question 4: Correct
Your team is designing the tables for a data warehouse. The data warehouse is going to
be hosted in a Dedicated SQL pool in Azure Synapse Analytics. The following tables are
going to be hosted initially in the pool
You have to choose the right distribution for each table. You have to ensure data
movement is minimized across tables
Which of the following distribution type would you choose for the Customer table?
● Hash
● Round Robin
● Replicated
● (Correct)
Explanation
Here since this is a dimension table and you need to ensure data movement is
minimized , you should choose replicated tables so that the data is available across all
nodes in the SQL pool.
For more information on replicated table design, one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/design
-guidance-for-replicated-tables
Question 5: Correct
Your team is designing the tables for a data warehouse. The data warehouse is going to
be hosted in a Dedicated SQL pool in Azure Synapse Analytics. The following tables are
going to be hosted initially in the pool
You have to choose the right distribution for each table. You have to ensure data
movement is minimized across tables
Which of the following distribution type would you choose for the Date table?
● Hash
● Round Robin
● Replicated
● (Correct)
Explanation
Here since this is a dimension table and you need to ensure data movement is
minimized , you should choose replicated tables so that the data is available across all
nodes in the SQL pool.
For more information on replicated table design, one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/design
-guidance-for-replicated-tables
Question 6: Correct
Your team has several Azure Stream Analytics jobs in place. They need to make use of
several windowing functions based on the needed requirement. Which of the following
windowing function can be used for the below requirement?
“Ensure that the data stream is segmented into distinct time segments and ensure that
events don’t overlap.”
● Sliding window
● Session window
● Tumbling window
● (Correct)
● Hopping window
Explanation
Here we need to use the Tumbling window.
For more information on Azure Stream Analytics windowing functions, one can visit the
following URL
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-window-fun
ctions
Question 7: Correct
Your team has several Azure Stream Analytics jobs in place. They need to make use of
several windowing functions based on the needed requirement. Which of the following
windowing function can be used for the below requirement?
“Ensure to output events only for points in time when the content of the window
actually changes”
● Sliding window
● (Correct)
● Session window
● Tumbling window
● Hopping window
Explanation
Here we need to use the Sliding window.
For more information on Azure Stream Analytics windowing functions, one can visit the
following URL
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-window-fun
ctions
Question 8: Correct
Your team has several Azure Stream Analytics jobs in place. They need to make use of
several windowing functions based on the needed requirement. Which of the following
windowing function can be used for the below requirement?
For more information on Azure Stream Analytics windowing functions, one can visit the
following URL
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-window-fun
ctions
Question 9: Correct
You have to design a Fact table . The table will be used to store Orders-based data. The
size of the table will be around 10 GB on disk. The table needs to be partitioned based
on date values.
For more information on table distribution, one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-dat
a-warehouse-tables-distribute
Question 10: Correct
You have to design a Fact table . The table will be used to store Orders-based data. The
size of the table will be around 10 GB on disk. The table needs to be partitioned based
on date values.
For more information on table partitioning , one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-dat
a-warehouse-tables-partition
Question 11: Correct
You have to design a Fact table . The table will be used to store Orders-based data. The
size of the table will be around 10 GB on disk. The table needs to be partitioned based
on date values.
For more information on table partitioning , one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-dat
a-warehouse-tables-partition
Question 12: Correct
Your team is planning on using External tables in Azure Synapse Analytics. Which of the
following can be queried via the use of External tables?
● Files in Azure File shares
● Documents in Azure Cosmos DB
● Objects in Azure Data Lake Gen2
● (Correct)
● Tables in Azure SQL Databases
Explanation
With external tables you can query for data in Azure Blob Storage or Azure Data Lake
Gen2
For more information on External tables , one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/develop-tables-external-
tables?tabs=hadoop
Question 13: Incorrect
Your team is planning on using External tables in Azure Synapse Analytics. The team
will be using a set of Parquet-based files hosted in an Azure Data Lake Gen2 Storage
account.
Which of the following statement is used to reference the Azure Data Lake Gen2
Storage account and the associated credentials to access the account?
● CREATE EXTERNAL FILE FORMAT
● CREATE EXTERNAL TABLE
● (Incorrect)
● CREATE EXTERNAL DATA SOURCE
● (Correct)
Explanation
With the CREATE EXTERNAL DATA SOURCE statement, you will mention the source of
data which is in Azure Data Lake Gen2 Storage account. You will also specify the
credentials which will be used to access the storage account.
For more information on External tables , one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/develop-tables-external-
tables?tabs=hadoop
Question 14: Correct
Your team is planning on using External tables in Azure Synapse Analytics. The team
will be using a set of Parquet-based files hosted in an Azure Data Lake Gen2 Storage
account.
Which of the following statement is used to describe the format of the files?
● CREATE EXTERNAL FILE FORMAT
● (Correct)
● CREATE EXTERNAL TABLE
● CREATE EXTERNAL DATA SOURCE
Explanation
The CREATE EXTERNAL FILE FORMAT statement is used to specify the format of the
files.
For more information on External tables , one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/develop-tables-external-
tables?tabs=hadoop
Question 15: Correct
Your data engineering team currently has the following resources defined in Azure
1) An Azure Event Hub – This is used to stream events from external data sources onto
Azure.
2) An Azure Data Lake Gen2 Storage account – This is used to store the events
streamed via Azure Event Hubs
3) An Azure Data Factory instance – This is used to build various ETL pipelines
4) An Azure Synapse Analytics workspace – This is used to host a dedicated SQL pool.
You have to build a pipeline in Azure Data Factory to copy data at regular time intervals
from the Azure Data Lake Gen2 Storage account onto tables in the dedicated SQL pool.
You have to ensure that only data within a specified time window is copied onto tables
in the dedicated SQL pool.
Which of the following would you choose the Integration runtime type for the pipeline?
● Azure Integration runtime
● (Correct)
● Azure-SSIS Integration runtime
● Self-hosted Integration runtime
Explanation
Here since the source and destination of the data are Azure-based resources ,we can
make use of the Azure Integration runtime itself.
For more information on the Azure Integration runtime , one can visit the following URL
https://docs.microsoft.com/en-us/azure/data-factory/create-azure-integration-runtime?
tabs=data-factory
Question 16: Incorrect
Your data engineering team currently has the following resources defined in Azure
1) An Azure Event Hub – This is used to stream events from external data sources onto
Azure.
2) An Azure Data Lake Gen2 Storage account – This is used to store the events
streamed via Azure Event Hubs
3) An Azure Data Factory instance – This is used to build various ETL pipelines
4) An Azure Synapse Analytics workspace – This is used to host a dedicated SQL pool.
You have to build a pipeline in Azure Data Factory to copy data at regular time intervals
from the Azure Data Lake Gen2 Storage account onto tables in the dedicated SQL pool.
You have to ensure that only data within a specified time window is copied onto tables
in the dedicated SQL pool.
For more information on the tumbling window trigger , one can visit the following URL
https://docs.microsoft.com/en-us/azure/data-factory/how-to-create-tumbling-window-t
rigger?tabs=data-factory
Question 17: Correct
Your data engineering team has a table that has the following structure in a dedicated
SQL pool in an Azure Synapse Analytics workspace
Which of the following statement can be used to implement row level security in the
table?
● CREATE DYNAMIC MASK
● CREATE SECURITY POLICY
● (Correct)
● GRANT
● UPDATE
Explanation
You can implement row-level security with the use of the statement CREATE SECURITY
POLICY
For more information on row-level security , one can visit the following URL
https://docs.microsoft.com/en-us/sql/relational-databases/security/row-level-security?
view=sql-server-ver15
Question 18: Correct
Your data engineering team has a table that has the following structure in a dedicated
SQL pool in an Azure Synapse Analytics workspace
Which of the following statement can be used to implement column level security in the
table?
● CREATE DYNAMIC MASK
● CREATE SECURITY POLICY
● GRANT
● (Correct)
● UPDATE
Explanation
You can implement column-level security with the use of the GRANT statement
For more information on column-level security , one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/colum
n-level-security
Question 19: Correct
Your team has an Azure Databricks workspace. They need to create two clusters. Below
are the requirements for the clusters
Which of the following would you choose as the cluster mode for Cluster 1?
● Single Node
● Standard
● High Concurrency
● (Correct)
Explanation
All of these requirements are met with the use of the High Concurrency cluster
For more information on configuring clusters in Azure Databricks , one can visit the
following URL
https://docs.microsoft.com/en-us/azure/databricks/clusters/configure
Question 20: Correct
Your team has an Azure Databricks workspace. They need to create two clusters. Below
are the requirements for the clusters
Which of the following would you choose as the cluster mode for Cluster 2?
● Single Node
● Standard
● (Correct)
● High Concurrency
Explanation
We have to choose the Standard cluster because the High Concurrency cluster does not
support Scala. Also, the Single Node cluster will not be effective for a set of users.
For more information on configuring clusters in Azure Databricks , one can visit the
following URL
https://docs.microsoft.com/en-us/azure/databricks/clusters/configure
Question 21: Correct
Your team is going to make use of Azure Data Lake Gen 2 storage accounts for storage
of data. Data will be uploaded to Azure Data Lake Gen 2 storage account via a pipeline
in Azure Data Factory. The pipeline will run once every day.
You have to design the storage access for the storage account based on the following
requirements
1) During the first 2 weeks, the data in the storage account will be accessed frequently
2) After 2 weeks, the data will be accessed less frequently. But the data needs to be
accessed immediately whenever required.
3) After 3 months the data will be rarely accessed. Whenever an object is required , an
SLA of one day is in place to have the object in place.
For more information on configuring access tiers for Azure Blob storage , one can visit
the following URL
https://docs.microsoft.com/en-us/azure/storage/blobs/access-tiers-overview
Question 22: Correct
Your team is going to make use of Azure Data Lake Gen 2 storage accounts for storage
of data. Data will be uploaded to Azure Data Lake Gen 2 storage account via a pipeline
in Azure Data Factory. The pipeline will run once every day.
You have to design the storage access for the storage account based on the following
requirements
1) During the first 2 weeks, the data in the storage account will be accessed frequently
2) After 2 weeks, the data will be accessed less frequently. But the data needs to be
accessed immediately whenever required.
3) After 3 months the data will be rarely accessed. Whenever an object is required , an
SLA of one day is in place to have the object in place.
Which of the following access tier would you use for the objects after the first 2 weeks
and before 3 months?
● Archive
● Cool
● (Correct)
● Hot
Explanation
Here since the objects are not accessed that frequently, we can choose the Cool Access
tier. Here we will not choose the Archive Access tier because the objects need to be
accessed immediately whenever required.
For more information on configuring access tiers for Azure Blob storage , one can visit
the following URL
https://docs.microsoft.com/en-us/azure/storage/blobs/access-tiers-overview
Question 23: Correct
Your team is going to make use of Azure Data Lake Gen 2 storage accounts for storage
of data. Data will be uploaded to Azure Data Lake Gen 2 storage account via a pipeline
in Azure Data Factory. The pipeline will run once every day.
You have to design the storage access for the storage account based on the following
requirements
1) During the first 2 weeks, the data in the storage account will be accessed frequently
2) After 2 weeks, the data will be accessed less frequently. But the data needs to be
accessed immediately whenever required.
3) After 3 months the data will be rarely accessed. Whenever an object is required , an
SLA of one day is in place to have the object in place.
Which of the following access tier would you use for the objects after 3 months?
● Archive
● (Correct)
● Cool
● Hot
Explanation
Here we can opt for the Archive Access tier since the objects are rarely accessed. Also,
with the SLA of one day, that can be taken to rehydrate the object whenever required.
For more information on configuring access tiers for Azure Blob storage , one can visit
the following URL
https://docs.microsoft.com/en-us/azure/storage/blobs/access-tiers-overview
Question 24: Correct
Your data engineering team is developing a data analytics solution. Part of the solution
is to develop a data warehousing environment. Initially the below table design has been
proposed.
For more information on understanding the star schema , one can visit the following
URL
https://docs.microsoft.com/en-us/power-bi/guidance/star-schema
Question 25: Correct
Your data engineering team is developing a data analytics solution. Part of the solution
is to develop a data warehousing environment. Initially the below table design has been
proposed.
What type of table is the Orders table going to be?
● Dimension
● Fact
● (Correct)
Explanation
As per the Star schema design, the Orders table is going to be a Fact table
For more information on understanding the star schema , one can visit the following
URL
https://docs.microsoft.com/en-us/power-bi/guidance/star-schema
Question 26: Correct
Your data engineering team is developing a data analytics solution. Part of the solution
is to develop a data warehousing environment. Initially the below table design has been
proposed.
What type of table is the Customers table going to be?
● Dimension
● (Correct)
● Fact
Explanation
As per the Star schema design, the Customers table is going to be a Dimension table
For more information on understanding the star schema , one can visit the following
URL
https://docs.microsoft.com/en-us/power-bi/guidance/star-schema
Question 27: Incorrect
Your data engineering team is developing a data analytics solution. Part of the solution
is to develop a data warehousing environment. Initially the below table design has been
proposed.
What type of dimension is the Product Table designed to be?
● Type 0
● Type 1
● (Incorrect)
● Type 2
● (Correct)
Explanation
The Product table is designed to be a Type 2 Slowly changing dimension. Here the table
has the additional columns of StartDate, EndDate and IsCurrent. This design suggests
that it is going to be a Type 2 Slowly Changing Dimension.
For more information on understanding the star schema , one can visit the following
URL
https://docs.microsoft.com/en-us/power-bi/guidance/star-schema
Question 28: Correct
Your team needs to deploy an Azure Data Lake Gen2 storage account. You have to
ensure that the Storage account remains available even if there is a region-level failure.
Costs need to be minimized wherever possible.
Which of the following do you need to enable when deploying an Azure General Purpose
V2 Storage account to ensure that is behaves as a Data Lake Gen2 Storage account?
● Enable storage account key access
● Enable hierarchical namespace
● (Correct)
● Access tier set to the Hot Access tier
● Enable large file shares
Explanation
When creating an Azure Data Lake Gen 2 storage account, for a normal General Purpose
V2 storage account, you need to enable the hierarchical namespace.
Which of the following would you choose as a redundancy option for the storage
account?
● Locally redundant storage
● Zone-redundant storage
● (Incorrect)
● Geo-redundant storage
● (Correct)
● Read Access Geo-redundant storage
Explanation
When you set the data redundancy option to Geo-redundant storage, you can be ensured
that the data in the storage account will become available in a secondary location if the
primary location fails.
Using the redundancy option of Read Access Geo-redundant storage would increase the
costs. And we need to minimize the costs wherever possible.
For more information on Azure Storage account redundancy , one can visit the following
URL
https://docs.microsoft.com/en-us/azure/storage/common/storage-redundancy
Question 30: Incorrect
Your team currently has the following resources defined on Azure
A Notebook is being developed in Scala in Azure Databricks. The Notebook will take
data from the Azure Data Lake Gen2 storage account as batch updates and save the
data onto a delta table.
For more information on working with batch workloads in delta tables , one can visit the
following URL
https://docs.microsoft.com/en-us/azure/databricks/delta/delta-batch
Question 31: Correct
Your team currently has the following resources defined on Azure
1) An Azure Data Lake Gen2 Storage account
A Notebook is being developed in Scala in Azure Databricks. The Notebook will take
data from the Azure Data Lake Gen2 storage account as batch updates and save the
data onto a delta table.
For more information on working with batch workloads in delta tables , one can visit the
following URL
https://docs.microsoft.com/en-us/azure/databricks/delta/delta-batch
Question 32: Incorrect
Your team currently has the following resources defined on Azure
A Notebook is being developed in Scala in Azure Databricks. The Notebook will take
data from the Azure Data Lake Gen2 storage account as batch updates and save the
data onto a delta table.
For more information on working with batch workloads in delta tables , one can visit the
following URL
https://docs.microsoft.com/en-us/azure/databricks/delta/delta-batch
Question 33: Correct
You have to design a Data Analytics solution for your company. You need to decide on
the services that are going to be used based on the below requirements
For more information on Azure Data Lake Gen2 Storage accounts , one can visit the
following URL
https://docs.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-introduction
Question 34: Correct
You have to design a Data Analytics solution for your company. You need to decide on
the services that are going to be used based on the below requirements
For more information on Azure Databricks , one can visit the following URL
https://docs.microsoft.com/en-us/azure/databricks/scenarios/what-is-azure-databrick
s
Question 35: Correct
You have to design a Data Analytics solution for your company. You need to decide on
the services that are going to be used based on the below requirements
For more information on Azure Synapse , one can visit the following URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/overview-what-is
Question 36: Correct
You have a table named ProductDetails hosted in a Dedicated SQL Pool in an Azure
Synapse Analytics workspace. You have to segregate the status of each product in the
table. Below is the SQL statement that needs to be completed for this requirement
Which of the following would come in Area 1?
● UPDATE
● SELECT
● CASE
● (Correct)
● ELSE
Explanation
Here we can compare the different product status with the use of the CASE statement
For more information on the CASE statement , one can visit the following URL
https://docs.microsoft.com/en-us/sql/t-sql/language-elements/case-transact-sql
Question 37: Correct
You have a table named ProductDetails hosted in a Dedicated SQL Pool in an Azure
Synapse Analytics workspace. You have to segregate the status of each product in the
table. Below is the SQL statement that needs to be completed for this requirement
For more information on the CASE statement , one can visit the following URL
https://docs.microsoft.com/en-us/sql/t-sql/language-elements/case-transact-sql
Question 38: Incorrect
You have to develop the SQL statement for an Azure Stream Analytics Job. The Job will
take inputs from two separate Azure Event Hubs. And then write the data to a table in a
Dedicated SQL pool in an Azure Synapse Analytics workspace.
For more information on repartitioning data , one can visit the following URL
https://docs.microsoft.com/en-us/azure/stream-analytics/repartition
Question 39: Incorrect
You have to develop the SQL statement for an Azure Stream Analytics Job. The Job will
take inputs from two separate Azure Event Hubs. And then write the data to a table in a
Dedicated SQL pool in an Azure Synapse Analytics workspace.
For more information on repartitioning data , one can visit the following URL
https://docs.microsoft.com/en-us/azure/stream-analytics/repartition
Question 40: Correct
You have an Azure Databricks cluster. You want to keep the configuration of the cluster
even after it is terminated. Which of the following can you do for this requirement?
● Create a notebook in the cluster with the cluster configuration
● Pin the cluster
● (Correct)
● Configure the cluster init scripts
Explanation
If you want to maintain the configuration of the cluster, you just need to Pin the cluster.
For more information on managing clusters , one can visit the following URL
https://docs.microsoft.com/en-us/azure/databricks/clusters/clusters-manage
Question 41: Correct
Your team has an Azure Data Lake Gen2 storage account. Continuous Time
series-based data is going to be streamed into the Data Lake Gen2 storage account.
Which of the following is the right design pattern to follow when it comes to the folder
structure and file naming convention for the streaming data?
● \YYYY\MM\DD\DataSet\datafile_YYYY_MM_DD.csv
● \DataSet\YYYY\MM\DD\datafile_YYYY_MM_DD.csv
● (Correct)
● \DataSet\datafile_YYYY_MM_DD.csv
Explanation
The recommendation is to ensure to have a parent folder that could specify the data
source or the data set for the data. Then the format of the child folders needs to be in
the form of the year , then the month and then the day. And then finally you have the file.
For more information on the best practices for Azure Data Lake Storage , one can visit
the following URL
https://docs.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-best-practice
s
Question 42: Correct
Your team currently has an Azure Stream Analytics job in place. The job is used to take
in data being streamed via Azure Event Hubs. Here log-based metrics from an
application is being streamed from Azure Event Hubs onto the Stream Analytics job.
You have to find the difference in time between the First and the Final Event in the
stream over a 2-hour duration.
● LIMIT
● LAST
● COLLATE
● DATEDIFF
● (Correct)
Explanation
Here we need to use the DATEDIFF function to find the time difference
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-stream-ana
lytics-query-patterns
Question 43: Correct
Your team currently has an Azure Stream Analytics job in place. The job is used to take
in data being streamed via Azure Event Hubs. Here log-based metrics from an
application is being streamed from Azure Event Hubs onto the Stream Analytics job.
You have to find the difference in time between the First and the Final Event in the
stream over a 2-hour duration.
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-stream-ana
lytics-query-patterns
Question 44: Incorrect
Your team currently has an Azure Stream Analytics job in place. The job is used to take
in data being streamed via Azure Event Hubs. Here log-based metrics from an
application is being streamed from Azure Event Hubs onto the Stream Analytics job.
You have to find the difference in time between the First and the Final Event in the
stream over a 2-hour duration.
https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-stream-ana
lytics-query-patterns
Question 45: Incorrect
Your team needs to create an external table in an Azure Synapse Serverless SQL pool.
The table will be used to query parquet-based files in an Azure Data Lake Gen2 storage
account. Currently the storage account container is configured as shown below
You have to ensure the Serverless SQL Pool has the right authorization to query the data
in the storage account. Which of the following would you create for this requirement?
● An encryption key
● (Incorrect)
● An Azure Databricks scoped secret
● A database scoped credential
● (Correct)
Explanation
Here we need to create a database scoped credential that would have the right
authorization such as Shared Access Signatures. This would allow the external table to
query the data in the Azure Data Lake Gen2 storage account.
For more information on working with external tables , one can visit the below URL
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql/develop-tables-external-
tables