0% found this document useful (0 votes)
10 views1 page

Tech Mahindra

The document outlines key interview questions for a Data Engineering role at Tech Mahindra Group in 2025, focusing on topics such as Spark, ADF triggers, data governance, and SQL queries. It includes technical distinctions, optimization techniques, and practical coding tasks relevant to the role. Additionally, it provides an example of transforming input data into a specified output format.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as TXT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
10 views1 page

Tech Mahindra

The document outlines key interview questions for a Data Engineering role at Tech Mahindra Group in 2025, focusing on topics such as Spark, ADF triggers, data governance, and SQL queries. It includes technical distinctions, optimization techniques, and practical coding tasks relevant to the role. Additionally, it provides an example of transforming input data into a specified output format.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as TXT, PDF, TXT or read online on Scribd
You are on page 1/ 1

Are you preparing for a Data Engineering role in 2025?

Here are some key interview


questions based on recent experiences at Tech Mahindra Group:
Round : L1

- Explain your current project and your specific role in it.


- What is Trigger In ADF and Types of Triggers?
- What is Integration Runtime and its Types?
- Differentiate between repartition and coalesce in Spark.
- Explain the variance between caching and persisting in Spark.
- Understanding Unity Catalog in Databricks and its role in data governance.
- Familiarize yourself with the reading and writing modes available in Spark.
- Cluster in Spark and its types.
- Spark optimization techniques, including broadcasting.
- Wide versus narrow transformations in Spark.
- Autoscaling and Auto-Termination.
- RDD versus Data frame distinctions.
- Code implementation for string reversal without built-in functions.
- Parquet file format overview and its benefits.
- Delta Lake and a detailed explanation of versioning.
- Comparison between union and unionbyname.
- Differentiation of map and flatmap functions.
- Write a query to create sparksession.
- SQL query to find the second highest salary from the employee table.
- Query to transform the input table into the specified output format.

input:
ID Item_Name Value
1 Apple 100
1 Apple 50
1 Orange 200
1 Orange 300
1 Banana 50

output:
id Apple Orange Banana
1 75 250 50

You might also like