Ajay Kadiyala Resume 2023 PDF
Ajay Kadiyala Resume 2023 PDF
Ajay Kadiyala Resume 2023 PDF
📱: +91-9542380365, 📧: [email protected],
: https://www.linkedin.com/in/ajay026/,
: https://github.com/Ajay026.
Profile Summary:
• Over 5+ Years of overall IT experience in Application Development.
• Working experience in Hadoop ecosystem (Gen-1 and Gen-2) and its various components such as
HDFS, Job Tracker, Task Tracker, Name Node, Data Node, Resource Manager (YARN), Application master
Node manager.
• Experience with components such as Cloudera distribution encompassing components like
MapReduce, Spark, SQL, Hive, HBase, Sqoop, Pyspark.
• Good skills on NoSQL Database- Cassandra.
• Proficient in developing Hive scripts for various business requirements.
• Knowledge in Data Warehousing Concepts in OLTP/OLAP System Analysis and developing Database
Schemas like Star Schema and Snowflake Schema for Relational and Dimensional Modelling.
• Good hands on in creating custom UDF’s in Hive.
• Load and transform large sets of structured, semi-structured and unstructured data from Relational
Database Systems to HDFS and vice-versa using Sqoop tool.
• Working knowledge on Hive UDF and various joins.
• Good Experience on architecture and components of Spark, and efficient in working with Spark Core,
Data Frames/Data Sets/RDD API/Spark SQL, Spark streaming and expertise in building PySpark and
Spark-Scala applications for interactive analysis, batch processing and stream processing.
• Hands-on experience in Spark, Scala, SparkSQL, Hive Context for Data Processing.
• Working knowledge on GCP tools like Cloud Function, Dataproc, Big Query.
• Experience on Azure cloud i.e., ADF, ADLS, Blob Storage, Databricks, Synapse etc.
• Extensive working experience in an Agile development Methodology & Working knowledge on Linux.
• Expertise in working with big data distributions like Cloudera and Hortonworks.
• Automated data pipelines using streams & tasks Involved in loading the structured and semi
structured data into spark clusters using Spark SQL and Data Frames Application programming interface
(API).
• Experience in working with Hive data warehouse tool-creating tables, distributing data by doing static
partitioning and dynamic partitioning, bucketing, and using Hive optimization techniques.
• Experience in tuning and debugging Spark application and using Spark optimization techniques.
• Knowledge on architecture and components of Spark and demonstrated efficiency in optimizing and
tuning compute and memory for performance and price optimization.• Expertise in developing batch
data processing applications using Spark, Hive and Sqoop.
• Experience in working with CSV, JSON, XML, ORC, Avro and Parquet file formats.
• Good experience in creating and designing data ingest pipelines using technologies such as Apache
Kafka.
• Worked on most of the popular AWS stack like S3, EC2, EMR, Athena.
• Good knowledge in working with ETL methods for data extraction, transformation and loading in
corporate-wide ETL Solutions and Data Warehouse tools for reporting and data analysis.
• Basic Experience in implementing Snowflake Data Warehouse.
• Experience in working with version control systems like Git, GitHub, CI/CD pipelines.
TECHNICAL SKILLS:
• Big Data Technologies: Hadoop, Spark, Hive, Sqoop, Kafka, PySpark, HBase, Scala Spark &
Snowflake Basic,
• Cloud Technologies: Azure(Azure Storage, Azure Synapse, ADF, Azure Data Bricks), GCP(Bigquery,
Dataproc), AWS basics
• Languages: Scala, Python, SQL
• Databases and Tools: Oracle, MySQL, SQL, NoSQL.
• Platforms: Windows, Linux.
• IDEs: Eclipse, Cloudera, Hortonworks.
• Scheduling: Airflow.
• Project Management Tools: Jira, GitHub.
Certifications:
• Completed Microsoft Azure-Fundamentals (AZ-900).
• Completed Microsoft Azure Data Fundamentals (DP-900)
• Completed Microsoft Azure Data Scientist Associate (D100).
• Completed Microsoft Azure Power BI Data Analyst Associate (PL-300).
• Completed Microsoft Azure Data Engineer Associate (DP-203).
• GCP Associate Cloud Engineer.
PROFESSIONAL EXPERIENCE:
• Tech stack: Scala, Hadoop, Spark, Spark SQL, Spark Streaming, Hive, Cassandra, MySQL,
HDFS,Apache Kafka.
Client : Vodafone(VFQ)
Remote
Oct 2021 – June 2022
Project Description:
Provides a 360 degree view of the customer so that a Salesperson is well aware of all the
facts when talking to customer. This gives a much better chance to close the deal.
This involves building a data lake. Data sources use Hadoop tools to transfer data to and from
HDFS and some of the sources, were imported using sqoop, Then storing the raw data into
HIVE tables in ORC format in order to facilitate the data scientists to perform analytics using
HIVE. New use cases were developed and dumped into a NOSQL database (Hbase) for
further analytics.
Responsibilities:
• Developed SQOOP scripts to import the source data from Oracle database into HDFS for
further processing.
• Developed HIVE Script to store raw data in ORC format.
• Involved in gathering requirements, designing, development and testing.
• Generated reports using Hive for business requirements received on ADHOC basis.
Project Description:
The Project is about to handle Risk Management Team, where the Bank wanted to store,
process manage the huge amount of data in a day-to-day operation, collected from various
sources. The system Majority checks the credibility of the customer looks for the credit risks.
Responsibilities:
• Ingested data from multiple sources like MySQL.
• Created and worked on Sqoop jobs with incremental load.
• Design both managed & External tables in Hive.
• Developed Spark Code in Scala using Spark SQL & Data frames for optimization.
• Creating HBase layer for faster reporting.
• Tech Stack: HDFS · Apache Sqoop · HBase · Apache Spark · Hive · Hadoop
Freelancing
Project 1: Migration
Remote
April 2022 – June 2022
Role: Data Engineer
Responsibilities:
• Working with Structured data that is being ingested into Azure File storage explorer.
• Create ETL pipeline in Snap logic tool to bring the data into azure Databricks workspace.
• Applied transformation logic including, Spark sql, pyspark operations on data.
• Applied optimizations logics i.e., Partitioning, broadcast joins etc.
• Create ETL pipeline on Databricks transformed data to dump the target directory called
snowflake.
• Analyze the resultant data with data bricks tool.
Responsibilities:
• Loaded and transformed large sets of structured and semi structured data.
• Imported data using Sqoop into Hive and Hbase from existing SQL Server.
• Support code/design analysis, strategy development and project planning.
• Create reports for the BI team using Sqoop to export data into HDFS and Hive.
• Involve in Requirement Analysis, Design, and Development.
• Export and Import data into HDFS and Hive using Sqoop.
• Storing the data to HBase tables according to business requirements.
• Creating Hive Tables using advances concepts in Hive like bucketing, partitioning, UDF’s.
• Tech Stack: Hadoop Framework, HDFS, Spark, Spark SQL, Hive, Sqoop.
Side Projects:
Education:
• B Tech (Electronics & Communication) Siddhartha Institute of Engineering and Technology,
Puttur (A.P) India, 2017 with First Division marks 60%.
• Diploma in Electronics and Communication Engineering in Govt Polytechnic college,
Chandragiri (A.P) India, 2014 with 70%