0% found this document useful (0 votes)
20 views3 pages

Prithvi Analysis of Chauhan

Prithvi Nag Kolla is a Data Engineer with 6 years of experience in data application development and management, holding a Master's in Data Science and a Bachelor's in Computer Science. He has expertise in AWS services, data processing using PySpark, and CI/CD pipelines, along with experience in quality assurance and automation. His professional background includes roles at LPL Financial and Amazon, where he contributed to data quality, automation, and operational excellence.

Uploaded by

tovikram
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views3 pages

Prithvi Analysis of Chauhan

Prithvi Nag Kolla is a Data Engineer with 6 years of experience in data application development and management, holding a Master's in Data Science and a Bachelor's in Computer Science. He has expertise in AWS services, data processing using PySpark, and CI/CD pipelines, along with experience in quality assurance and automation. His professional background includes roles at LPL Financial and Amazon, where he contributed to data quality, automation, and operational excellence.

Uploaded by

tovikram
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

Prithvi Nag Kolla

Data Engineer
Email: [email protected]
Phone: +91 – 6361-663-590

Credly Skill Badges:

SUMMARY

• With 6 years of experience in developing, deploying, and managing data applications, as well as
conducting data validation, coupled with 6 years of academic experience including a Master's in
Data Science from the USA and a Bachelor's in Computer Science from India.
• Hands-on experience with AWS S3, AWS Lambda, AWS CloudWatch
• Created Data processing pipelines using PySpark, SQL for DDL, DML functionalities
• Worked effectively in a fast paced and demanding complex environmental systems
• Experienced in serverless services in AWS like Lambda and Athena
• Experienced in AWS logging service like CloudWatch
• Experienced in processing large datasets of different forms including structured and unstructured
data
• Expertise in designing, developing and integrating test cases into Data Test Automation Pipelines
for Data Quality Assurance
• Using PySpark to develop processes for extracting, cleansing, transforming, validating,
automating and creating reports
• Created KPI dashboards using Tableau
• Experienced in handling various file formats like text, CSV, JSON, Parquet
• Hands-on Experience in GitHub Version control, maintaining code base using git bash commands
• Experience in Building, storing and deploying docker images on GitHub Actions
• Experience creating CI CD pipelines with GitHub Actions
• Developed SQL Queries for validating business transformations and analyzing data.
• Experienced in writing queries using SQL & PostgreSQL
EDUCATION

• MS in Data science 2019 - 2021


University of North Texas, USA

• BTECH in Computer science 2014 - 2018


GITAM University, India

TECHNICAL SKILLS

Programming Languages: Python, PySpark, Git


Cloud Computing: AWS S3, AWS Lambda, Event Bridge, IAM Roles and Permission
Management
CI/CD Tools Terraform, GitHub Actions Integration and Deployment
Reporting Tools: Tableau, Excel
Bigdata Ecosystems: Pandas, boto3
Version Control Tools: GitHub, Git Bash
Programming IDE’s: Jupyter Notebook, PyCharm, VSCode
Databases: MySQL, PostgreSQL
Operating Systems: Windows and Mac

PROFESSIONAL EXPERIENCE

LPL Financial – Phoenix, AZ AUG 2021 – APRIL 2024


Role: Data Engineer
- Spearheaded black box testing initiatives on the AWS data processing framework, ensuring
robust data quality and reliability.
- Identified errors and meticulously documented bug issues in JIRA for resolution, streamlining the
debugging process.
- Closed bug issues after thorough retesting to ensure alignment with business requirements and
data integrity.
- Championed test automation by devising solutions in PySpark, automating redundant test cases
within regression testing on vast datasets, significantly improving data quality and testing
efficiency.
- Significantly reduced manual testing efforts and enhanced performance by expediting test report
delivery using PySpark automation techniques.
- Leveraged GitHub workflow actions, scripted YAML Workflow, and automated regression tests
with PySpark, mitigating issues and substantially improving performance and deployment
processes.
- Meticulously created test scenarios, test cases, and test data for development features, ensuring
comprehensive test coverage and data accuracy.
- Documented and audited test reports during daily sprints and on release days, maintaining
transparency and accountability in testing processes.
- Managed offshore team members by assigning tasks through JIRA and providing pre-sprint
presentations, ensuring seamless implementation of new processes and effective collaboration.
- Donned multiple hats – from Developer and Tester to DevSecOps, leveraging tools such as JIRA,
GitHub, Python, and PySpark to drive efficiency and innovation in data engineering projects.
- Achieved operational excellence and drove test automation solutions that aligned with company
standards, enhancing overall data processing and quality assurance frameworks using PySpark

Amazon – Hyderabad, India AUG 2018 – JULY 2019


Role: Data Analyst
- Ensured quality assurance of daily operations at North American Fulfillment Centers, utilizing MS
Excel for data analysis and reporting to maintain high operational standards.
- Instrumentally provided timely and relevant responses as part of an automation initiative,
leveraging Python scripting to automate repetitive tasks and improve response times.
- Contributed significantly to the enhancement of tools and features through active participation in
BETA testing programs, using data analysis with Pandas to identify areas of improvement.
- Actively utilized responses generated during video monitoring activities as crucial training data
for AI models, employing Python for data preprocessing and annotation.
- Showcased versatility and adaptability by commencing as a Training Associate, contributing
significantly to the development and delivery of comprehensive training programs using MS Excel
for curriculum tracking and progress analysis.
- Observed and managed agile processes to streamline project development and foster cross-
functional collaboration in the pursuit of quality excellence.
- Showcased versatility and adaptability by commencing as a Training Associate, contributing
significantly to the development and delivery of comprehensive training programs
- Provided valuable insights and feedback that directly impacted the enhancement of tools and
features, highlighting my dedication to maintaining the highest standards of quality and innovation
through data-driven approaches using Python and MS Excel.

PROGECT EXPERIENCE

Weather Data Analytics JAN 2017 – JAN 2018


Role: Project Lead
- - Designed and implemented AWS data pipelines for processing weather data and micro-batching
based on the Location feature.
- - Utilized AWS services such as S3, Lambda, and IAM to automate data processing tasks efficiently.
- - Developed and maintained Python scripts using Pandas and boto3 for data manipulation and
analysis.
- - Created manual data ingestion processes into Raw S3 buckets and design event triggers to initiate
Lambda functions.
- - Collaborated with team members to understand business requirements and translate them into
technical solutions.
- - Ensured data quality and integrity by implementing data validation and error handling
mechanisms.
- - Optimized data processing workflows for performance and scalability in a cloud environment.
- - Monitored and troubleshooted data pipeline issues, ensuring smooth operation and timely
delivery of results.
- - Documented system architecture, processes, and code to facilitate knowledge sharing and future
maintenance.

You might also like