100% found this document useful (1 vote)
732 views25 pages

Databricks Certified Data Engineer Associate Practice Exams - 1

The document contains practice exam questions for the Databricks Certified Data Engineer Associate certification, covering topics such as data modeling, Delta Lake features, and SQL operations. It includes multiple-choice questions about calculated columns, job stability, and the differences between AUTO LOADER and COPY INTO, among others. The exam aims to test knowledge and skills relevant to data engineering within the Databricks environment.

Uploaded by

Lucas Siilva
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
732 views25 pages

Databricks Certified Data Engineer Associate Practice Exams - 1

The document contains practice exam questions for the Databricks Certified Data Engineer Associate certification, covering topics such as data modeling, Delta Lake features, and SQL operations. It includes multiple-choice questions about calculated columns, job stability, and the differences between AUTO LOADER and COPY INTO, among others. The exam aims to test knowledge and skills relevant to data engineering within the Databricks environment.

Uploaded by

Lucas Siilva
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 25

04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

Databricks Certified Data


Engineer Associate Practice Exams - 1

Simulado 1

You were asked to create a table that can store the below data, orderTime is
a timestamp but the finance team when they query this data normally prefer
the orderTime in date format, you would like to create a calculated column
that can convert the orderTime column timestamp datatype to date and
store it, fill in the blank to complete the DDL.

(1 Ponto)

AS DEFAULT (CAST(orderTime as DATE))

GENERATED ALWAYS AS (CAST(orderTime as DATE))

GENERATED DEFAULT AS (CAST(orderTime as DATE))

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 1/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

AS (CAST(orderTime as DATE))

Delta lake does not support calculated columns, value should be inserted into the table as
part of the ingestion process

The data engineering team noticed that one of the job fails randomly as a
result of using spot instances, what feature in Jobs/Tasks can be used to
address this issue so the job is more stable when using spot instances?
(1 Ponto)

Use Databrick REST API to monitor and restart the job

Use Jobs runs, active runs UI section to monitor and restart the job

Add second task and add a check condition to rerun the first task if it fails

Restart the job cluster, job automatically restarts

Add a retry policy to the task

What is the main difference between AUTO LOADER and COPY INTO?
(1 Ponto)

COPY INTO supports schema evolution.

AUTO LOADER supports schema evolution.

COPY INTO supports file notification when performing incremental loads.

AUTO LOADER supports reading data from Apache Kafka

AUTO LOADER Supports file notification when performing incremental loads.

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 2/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

Why does AUTO LOADER require schema location? (1 Ponto)

Schema location is used to identify the schema of target table

Schema location is used to store user provided schema

AUTO LOADER does not require schema location, because its supports Schema evolution

Schema location is used to store schema inferred by AUTO LOADER

Schema location is used to identify the schema of target table and source table

Which of the following statements are incorrect about the lakehouse


(1 Ponto)

Support end-to-end streaming and batch workloads

Supports ACID

Support for diverse data types that can store both structured and unstructured

Supports BI and Machine learning

Storage is coupled with Compute

You are designing a data model that works for both machine learning using
images and Batch ETL/ELT workloads. Which of the following features of data
lakehouse can help you meet the needs of both workloads? (1 Ponto)

Data lakehouse requires very little data modeling.

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 3/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

Data lakehouse combines compute and storage for simple governance.

Data lakehouse provides autoscaling for compute clusters.

Data lakehouse can store unstructured data and support ACID transactions.

Data lakehouse fully exists in the cloud.

Which of the following locations in Databricks product architecture hosts


jobs/pipelines and queries? (1 Ponto)

Data plane

Control plane

Databricks Filesystem

JDBC data source

Databricks web application

10

You are currently working on a notebook that will populate a reporting table
for downstream process consumption, this process needs to run on a
schedule every hour. what type of cluster are you going to use to set up this
job? (1 Ponto)

Since it’s just a single job and we need to run every hour, we can use an all-purpose
cluster

The job cluster is best suited for this purpose.

Use Azure VM to read and write delta tables in Python

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 4/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

Use delta live table pipeline to run in continuous mode

11

Which of the following developer operations in CI/CD flow can be


implemented in Databricks Repos? (1 Ponto)

Merge when code is committed

Pull request and review process

Trigger Databricks Repos API to pull the latest version of code into production folder

Resolve merge conflicts

Delete a branch

12

You are currently working with the second team and both teams are looking
to modify the same notebook, you noticed that the second member is
copying the notebooks to the personal folder to edit and replace the
collaboration notebook, which notebook feature do you recommend to
make the process easier to collaborate. (1 Ponto)

Databricks notebooks should be copied to a local machine and setup source control
locally to version the notebooks

Databricks notebooks support automatic change tracking and versioning

Databricks Notebooks support real-time coauthoring on a single notebook

Databricks notebooks can be exported into dbc archive files and stored in data lake

Databricks notebook can be exported as HTML and imported at a later time

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 5/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

13

You are currently working on a project that requires the use of SQL and
Python in a given notebook, what would be your approach (1 Ponto)

Create two separate notebooks, one for SQL and the second for Python

A single notebook can support multiple languages, use the magic command to switch
between the two.

Use an All-purpose cluster for python, SQL endpoint for SQL

Use job cluster to run python and SQL Endpoint for SQL

14

Which of the following statements are correct on how Delta Lake implements
a lake house? (1 Ponto)

Delta lake uses a proprietary format to write data, optimized for cloud storage

Using Apache Hadoop on cloud object storage

Delta lake always stores meta data in memory vs storage

Delta lake uses open source, open format, optimized cloud storage and scalable meta
data

Delta lake stores data and meta data in computes memory

15

You were asked to create or overwrite an existing delta table to store the
below transaction data.

(1 Ponto)

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 6/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

a) b)

c) d)

16

if you run the command VACUUM transactions retain 0 hours? What is the
outcome of this command? (1 Ponto)

Command will be successful, but no data is removed

Command will fail if you have an active transaction running

Command will fail, you cannot run the command with retentionDurationcheck enabled

Command will be successful, but historical data will be removed

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 7/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

Command runs successful and compacts all of the data in the table

17

You noticed a colleague is manually copying the data to the backup folder
prior to running an update command, incase if the update command did not
provide the expected outcome so he can use the backup copy to replace
table, which Delta Lake feature would you recommend simplifying the
process? (1 Ponto)

Use time travel feature to refer old data instead of manually copying

Use DEEP CLONE to clone the table prior to update to make a backup copy

Use SHADOW copy of the table as preferred backup choice

Cloud object storage retains previous version of the file

Cloud object storage automatically backups the data

18

Which one of the following is not a Databricks lakehouse object? (1 Ponto)

Tables

Views

Database/Schemas

Catalog

Functions

Stored Procedures

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 8/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

19

What type of table is created when you create delta table with below
command? (1 Ponto)

Managed delta table

External table

Managed table

Temp table

Delta Lake table

20

Which of the following command can be used to drop a managed delta table
and the underlying files in the storage? (1 Ponto)

DROP TABLE table_name CASCADE

DROP TABLE table_name

Use DROP TABLE table_name command and manually delete files using command
dbutils.fs.rm("/path",True)

DROP TABLE table_name INCLUDE_FILES

DROP TABLE table and run VACUUM command

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlNU… 9/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

21

Which of the following is the correct statement for a session scoped


temporary view? (1 Ponto)

Temporary views are lost once the notebook is detached and re-attached

Temporary views stored in memory

Temporary views can be still accessed even if the notebook is detached and attached

Temporary views can be still accessed even if cluster is restarted

Temporary views are created in local_temp database

22

Which of the following is correct for the global temporary view? (1 Ponto)

global temporary views cannot be accessed once the notebook is detached and attached

global temporary views can be accessed across many clusters

global temporary views can be still accessed even if the notebook is detached and
attached

global temporary views can be still accessed even if the cluster is restarted

global temporary views are created in a database called temp database

23

You are currently working on reloading customer_sales tables using the


below query

1. INSERT OVERWRITE customer_sales


2. SELECT * FROM customers c
https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 10/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

3. INNER JOIN sales_monthly s on s.customer_id = c.customer_id

After you ran the above command, the Marketing team quickly wanted to
review the old data that was in the table. How does
INSERT OVERWRITE impact the data in the customer_sales table if you want
to see the previous version of the data prior to running the above statement?
(1 Ponto)

Overwrites the data in the table, all historical versions of the data, you can not time travel
to previous versions

Overwrites the data in the table but preserves all historical versions of the data, you can
time travel to previous versions

Overwrites the current version of the data but clears all historical versions of the data, so
you can not time travel to previous versions.

Appends the data to the current version, you can time travel to previous versions

By default, overwrites the data and schema, you cannot perform time travel

24

Which of the following SQL statement can be used to query a table by


eliminating duplicate rows from the query results? (1 Ponto)

SELECT DISTINCT * FROM table_name

SELECT DISTINCT * FROM table_name HAVING COUNT(*) > 1

SELECT DISTINCT_ROWS (*) FROM table_name

SELECT * FROM table_name GROUP BY * HAVING COUNT(*) < 1

SELECT * FROM table_name GROUP BY * HAVING COUNT(*) > 1

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 11/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

25

Which of the below SQL Statements can be used to create a SQL UDF to
convert Celsius to Fahrenheit and vice versa, you need to pass two
parameters to this function one, actual temperature, and the second that
identifies if its needs to be converted to Fahrenheit or Celcius with a one-
word letter F or C?

(1 Ponto)

a) b)

c) d)

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 12/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

26

You are trying to calculate total sales made by all the employees by parsing a
complex struct data type that stores employee and sales data, how would
you approach this in SQL

Table definition,

batchId INT, performance ARRAY<STRUCT<employeeId: BIGINT, sales:


INT>>, insertDate TIMESTAMP

Sample data of performance column

1. [
2. { "employeeId":1234
3. "sales" : 10000},
4.
5. { "employeeId":3232
6. "sales" : 30000}
7. ]

Calculate total sales made by all the employees?

Sample data with create table syntax for the data:

create or replace table sales as


select 1 as batchId ,
from_json('[{ "employeeId":1234,"sales" : 10000 },{ "employeeId":3232,"sales" :
30000 }]',
'ARRAY<STRUCT<employeeId: BIGINT, sales: INT>>') as performance,
current_timestamp() as insertDate
union all
select 2 as batchId ,
from_json('[{ "employeeId":1235,"sales" : 10500 },{
"employeeId":3233,"sales" : 32000 }]',
'ARRAY<STRUCT<employeeId: BIGINT, sales: INT>>') as
performance,
current_timestamp() as insertDate

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 13/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

(1 Ponto)

a) b)

c) d)

e)

27

Which of the following statements can be used to test the functionality of


code to test number of rows in the table equal to 10 in python?

(1 Ponto)

assert (row_count = 10, "Row count did not match")

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 14/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

assert if (row_count = 10, "Row count did not match")

assert row_count == 10, "Row count did not match"

assert if row_count == 10, "Row count did not match"

assert row_count = 10, "Row count did not match"

28

How do you handle failures gracefully when writing code in Pyspark, fill in
the blanks to complete the below statement

(1 Ponto)

try: failure:

try: catch:

try: except:

try: fail:

try: error:

29

You are working on a process to query the table based on batch date, and
batch date is an input parameter and expected to change every time the
program runs, what is the best way to we can parameterize the query to run
without manually changing the batch date? (1 Ponto)

Create a notebook parameter for batch date and assign the value to a python variable
and use a spark data frame to filter the data based on the python variable

Create a dynamic view that can calculate the batch date automatically and use the view to
query the data

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 15/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

There is no way we can combine python variable and spark code

Manually edit code every time to change the batch date

Store the batch date in the spark configuration and use a spark data frame to filter the
data based on the spark configuration.

30

Which of the following commands results in the successful creation of a view


on top of the delta stream(stream on delta table)? (1 Ponto)

a) b)

c) d)

e)

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 16/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

31

Which of the following techniques structured streaming uses to create an


end-to-end fault tolerance? (1 Ponto)

Checkpointing and Water marking

Write ahead logging and water marking

Checkpointing and idempotent sinks

Write ahead logging and idempotent sinks

Stream will failover to available nodes in the cluste

32

Which of the following two options are supported in identifying the arrival of
new files, and incremental data from Cloud object storage using Auto
Loader? (1 Ponto)

Directory listing, File notification

Checking pointing, watermarking

Writing ahead logging, read head logging

File hashing, Dynamic file lookup

Checkpointing and Write ahead logging

33

Which of the following data workloads will utilize a Bronze table as its
destination? (1 Ponto)

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 17/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

A job that aggregates cleaned data to create standard summary statistics

A job that queries aggregated data to publish key insights into a dashboard

A job that ingests raw data from a streaming source into the Lakehouse

A job that develops a feature set for a machine learning application

A job that enriches data by parsing its timestamps into a human-readable format

34

Which of the following data workloads will utilize a silver table as its source?
(1 Ponto)

A job that enriches data by parsing its timestamps into a human-readable format

A job that queries aggregated data that already feeds into a dashboard

A job that ingests raw data from a streaming source into the Lakehouse

A job that aggregates cleaned data to create standard summary statistics

A job that cleans data by removing malformatted records

35

Which of the following data workloads will utilize a gold table as its source?
(1 Ponto)

A job that enriches data by parsing its timestamps into a human-readable format

A job that queries aggregated data that already feeds into a dashboard

A job that ingests raw data from a streaming source into the Lakehouse

A job that aggregates cleaned data to create standard summary statistics

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 18/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

A job that cleans data by removing malformatted records

36

You are currently asked to work on building a data pipeline, you have noticed
that you are currently working with a data source that has a lot of data
quality issues and you need to monitor data quality and enforce it as part of
the data ingestion process, which of the following tools can be used to
address this problem? (1 Ponto)

AUTO LOADER

DELTA LIVE TABLES

JOBS and TASKS

UNITY Catalog and Data Governance

STRUCTURED STREAMING with MULTI HOP

37

When building a DLT s pipeline you have two options to create a live tables,
what is the main difference between CREATE STREAMING LIVE
TABLE vs CREATE LIVE TABLE? (1 Ponto)

CREATE STREAMING LIVE table is used in MULTI HOP Architecture

CREATE LIVE TABLE is used when working with Streaming data sources and Incremental
data

CREATE STREAMING LIVE TABLE is used when working with Streaming data sources and
Incremental data

There is no difference both are the same, CREATE STRAMING LIVE will be deprecated
soon

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 19/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

CREATE LIVE TABLE is used in DELTA LIVE TABLES, CREATE STREAMING LIVE can only
used in Structured Streaming applications

38

A particular job seems to be performing slower and slower over time, the
team thinks this started to happen when a recent production change was
implemented, you were asked to take look at the job history and see if we
can identify trends and root cause, where in the workspace UI can you
perform this analysis? (1 Ponto)

Under jobs UI select the job you are interested, under runs we can see current active runs
and last 60 days historical run

Under jobs UI select the job cluster, under spark UI select the application job logs, then
you can access last 60 day historical runs

Under Workspace logs, select job logs and select the job you want to monitor to view the
last 60 day historical runs

Under Compute UI, select Job cluster and select the job cluster to see last 60 day historical
runs

Historical job runs can only be accessed by REST API

39

What are the different ways you can schedule a job in Databricks workspace?
(1 Ponto)

Continuous, Incremental

On-Demand runs, File notification from Cloud object storage

Cron, On Demand runs

Cron, File notification from Cloud object storage

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 20/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

Once, Continuous

40

You have noticed that Databricks SQL queries are running slow, you are
asked to look reason why queries are running slow and identify steps to
improve the performance, when you looked at the issue you noticed all the
queries are running in parallel and using a SQL endpoint(SQL Warehouse)
with a single cluster. Which of the following steps can be taken to improve
the performance/response times of the queries?

*Please note Databricks recently renamed SQL endpoint to SQL warehouse.


(1 Ponto)

They can turn on the Serverless feature for the SQL endpoint(SQL warehouse).

They can increase the maximum bound of the SQL endpoint(SQL warehouse)’s scaling
range

They can increase the warehouse size from 2X-Smal to 4XLarge of the SQL endpoint(SQL
warehouse).

They can turn on the Auto Stop feature for the SQL endpoint(SQL warehouse).

They can turn on the Serverless feature for the SQL endpoint(SQL warehouse) and change
the Spot Instance Policy to “Reliability Optimized.”

41

You currently working with the marketing team to setup a dashboard for ad
campaign analysis, since the team is not sure how often the dashboard
should be refreshed they have decided to do a manual refresh on an as
needed basis. Which of the following steps can be taken to reduce the
overall cost of the compute when the team is not using the compute?

*Please note that Databricks recently change the name of SQL Endpoint to
SQL Warehouses. (1 Ponto)

They can turn on the Serverless feature for the SQL endpoint(SQL Warehouse).

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 21/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

They can decrease the maximum bound of the SQL endpoint(SQL Warehouse) scaling
range.

They can decrease the cluster size of the SQL endpoint(SQL Warehouse).

They can turn on the Auto Stop feature for the SQL endpoint(SQL Warehouse).

They can turn on the Serverless feature for the SQL endpoint(SQL Warehouse) and change
the Spot Instance Policy from “Reliability Optimized” to “Cost optimized”

42

You had worked with the Data analysts team to set up a SQL
Endpoint(SQL warehouse) point so they can easily query and analyze data in
the gold layer, but once they started consuming the SQL
Endpoint(SQL warehouse) you noticed that during the peak hours as the
number of users increase you are seeing queries taking longer to finish,
which of the following steps can be taken to resolve the issue?

*Please note Databricks recently renamed SQL endpoint to SQL warehouse.


(1 Ponto)

They can turn on the Serverless feature for the SQL endpoint(SQL warehouse).

They can increase the maximum bound of the SQL endpoint(SQL warehouse) ’s scaling
range.

They can increase the cluster size from 2X-Small to 4X-Large of the SQL endpoint(SQL wa‐
rehouse) .

They can turn on the Auto Stop feature for the SQL endpoint(SQL warehouse) .

They can turn on the Serverless feature for the SQL endpoint(SQL warehouse) and change
the Spot Instance Policy from “Cost optimized” to “Reliability Optimized.”

43

The research team has put together a funnel analysis query to monitor the
customer traffic on the e-commerce platform, the query takes about 30 mins

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 22/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

to run on a small SQL endpoint cluster with max scaling set to 1 cluster.
What steps can be taken to improve the performance of the query? (1 Ponto)

They can turn on the Serverless feature for the SQL endpoint.

They can increase the maximum bound of the SQL endpoint’s scaling range anywhere
from between 1 to 100 to review the performance and select the size that meets the re‐
quired SLA.

They can increase the cluster size anywhere from X small to 3XL to review the perfor‐
mance and select the size that meets the required SLA.

They can turn off the Auto Stop feature for the SQL endpoint to more than 30 mins.

They can turn on the Serverless feature for the SQL endpoint and change the Spot Ins‐
tance Policy from “Cost optimized” to “Reliability Optimized.”

44

Unity catalog simplifies managing multiple workspaces, by storing and


managing permissions and ACL at _______ level (1 Ponto)

Workspace

Account

Storage

Data pane

Control pane

45

Which of the following section in the UI can be used to manage permissions


and grants to tables? (1 Ponto)

User Settings

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 23/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

Admin UI

Workspace admin settings

User access control lists

Data Explorer

46

Which of the following is not a privilege in the Unity catalog? (1 Ponto)

SELECT

MODIFY

DELETE

CREATE TABLE

EXECUTE

47

A team member is leaving the team and he/she is currently the owner of the
few tables, instead of transfering the ownership to a user you have decided
to transfer the ownership to a group so in the future anyone in the group
can manage the permissions rather than a single individual, which of the
following commands help you accomplish this? (1 Ponto)

ALTER TABLE table_name OWNER to 'group'

TRANSFER OWNER table_name to 'group'

GRANT OWNER table_name to 'group'

ALTER OWNER ON table_name to 'group'

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 24/25
04/04/2025, 11:38 Databricks Certified Data Engineer Associate Practice Exams - 1

GRANT OWNER On table_name to 'group'

Nunca forneça sua senha. Relatar abuso

Este conteúdo foi criado pelo proprietário do formulário. Os dados que você enviar serão enviados ao proprietário
do formulário. A Microsoft não é responsável pela privacidade ou práticas de segurança de seus clientes, incluindo
aqueles do proprietário deste formulário. Nunca forneça sua senha.
Microsoft Forms | Pesquisas, questionários e votações com tecnologia de IA Criar meu próprio formulário
O proprietário deste formulário não forneceu uma política de privacidade sobre como usará seus dados de resposta.
Não forneça informações pessoais ou confidenciais. | Condições de uso

https://forms.office.com/pages/responsepage.aspx?id=jTiTfGXygEyBuKHhQcVUb3pCG-Yl6ppJgtYu7zPu6fBUNEhUSkZaWkdINUU3QzlCRFlN… 25/25

You might also like