Weekend Python Learning Plan for Aspiring Data
Engineers
Goal
Build a solid foundation in Python for data engineering by the end of the weekend. This plan assumes prior
programming knowledge (e.g., Java).
Day 1: Core Python & Scripting Essentials
Morning: Python Basics
• Variables and data types
• Lists, tuples, sets, and dictionaries
• Loops: for, while
• Conditionals: if, elif, else
• Functions, args and *kwargs
Practice: - Write a function to parse a CSV file and return rows as dictionaries.
Afternoon: File Handling, Strings, Exceptions
• Reading and writing files using open , with
• String manipulation methods
• Exception handling with try/except
Practice: - Write a script to read a log file and extract lines containing "ERROR".
Evening: Modules, Comprehensions, Basic OOP
• Importing and using modules
• List and dictionary comprehensions
• Writing simple classes and objects
Practice: - Create a Customer class with methods to load and print customer details.
1
Day 2: Data Engineering Tools & Project
Morning: pandas and Data Processing
• Introduction to pandas
• Loading data from CSV and JSON
• Filtering, selecting, sorting, aggregating
• Handling missing values, merging DataFrames
Practice: - Load a CSV, clean the data, and write it to a new CSV file.
Afternoon: Python + SQL
• Using sqlite3 or sqlalchemy to connect to a database
• Querying data and loading it into pandas
• Writing data back to the database or file
Practice: - Read data from a database, process in pandas, and save results.
Evening: Mini Project
Log Processor App: - Ingest a CSV or log file - Apply custom filtering logic - Save the cleaned data to SQLite
or a new CSV - Print or email a summary report
Tools to Install
• Python 3.10+
• VS Code with Python extension
• Jupyter (optional)
• pandas , openpyxl , sqlalchemy , sqlite3
After the Weekend
• Learn Airflow for workflow orchestration
• Explore PySpark for big data processing
• Get familiar with AWS Glue, Lambda, and S3
• Practice with Docker, Git, and logging libraries
With this foundation, you'll be ready to tackle real data engineering projects and build ETL pipelines.