0% found this document useful (0 votes)
23 views9 pages

Weekly Diary Report-244

Hello

Uploaded by

mamidiharish149
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
23 views9 pages

Weekly Diary Report-244

Hello

Uploaded by

mamidiharish149
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 9

SRM UNIVERSITY, ANDHRA PRADESH

SUMMER INTERNSHIP COURSE, June 2024


WEEKLY DIARY REPORT
(STUDENT BATCH 2022-26)
(To be submitted to Faculty Mentor over mail with CC to Industry mentor)

Name of Student: MAMIDI HARISH


ID NO: AP22110011244
Company Name (For industry internship): EDUNET Foundation
Name of Faculty Mentor: Professor Raja Sekhar

Week 1 From _____JUNE 7_____ to _________JUNE 13___________

S. Field Answer
No

1 Project Title Analysis of Car Dekho Dataset: Trends, Insights, and Anomalies

2 Project Description This project analyses the Car Dekho dataset to uncover trends and insights,
including vehicle manufacturing years, price ranges, record counts, missing data,
vehicle diversity, sales trends, CNG and cost depreciation factors, and
relationships between selling price and mileage.

3 Outline of the Solution The solution starts with exploring and cleaning the Car Dekho dataset to address
missing values and prepare it for analysis. It then centers on answering targeted
questions about manufacturing years, price ranges, vehicle diversity, sales trends,
and depreciation factors. The project wraps up with visualizing key findings and
summarizing insights and recommendations.

4 Design of the Solution The solution involves loading and cleaning the Car Dekho dataset to ensure data
quality in python, followed by performing descriptive analysis to extract insights
on vehicle characteristics, pricing trends, and depreciation factors. Visualizations
will summarize findings for a concise report with actionable recommendations.

5 Hardware and Software Hardware Requirements are Minimum: Intel Core i5 or equivalent, RAM
Requirements to execute minimum: 4 GB, Operating System: Windows 10 or higher.
the project
Software Requirements: Google collab, visual studio or any python ide

6 Environment setup Windows


windows/linux/Raspberry
pi/Arduino

7 Concepts Used 1. Functions


(Functions, header files, • read_csv() is used to load your dataset
data types, and concepts • info() give you a summary of your data
(Loops, arrays, • head() shows the first 5 rows.
conditional statements, • min() and `max() help find the smallest and largest values, like the oldest
etc.). Explanation of the and newest cars.
concepts). For Hardware • groupby() allows you to see trends, like how car prices change based on
projects also explain with the year.
respect to the code being
developed 2. Libraries Used
• Pandas: This tool helps you handle and analyze the data.
• Matplotlib/Seaborn: These are used to create charts and graphs, making it
easier to see patterns and trends.

3. Data Types
• Numbers: Used for values like car prices, mileage, etc.
• Categories or string characters: Used for car brands or fuel types.
• True/False or If conditions: Used to check conditions, like whether a car
has an automatic transmission.

4. Conditional Statements
• If-Else: Used to check, like whether CNG vehicles present in data or not

5. Data Cleaning and Visualization


• Cleaning Data: Making sure the data is correct, and no missing value is
present.
• Charts and Graphs: Used to visually show the results, making it easier to
understand the data.

8 Testing & Validation 1. Boundary Tests


(Boundary tests and • The car price ranges and years of manufacturing are boundaries of the
boundaries of inputs. dataset
Possible inputs and 2. Boundaries of Inputs
corresponding outputs). • For car manufacturing years, check if the data ranges from 2003 to 2018.
• For price ranges, check if the price ranges from 0.10 lakhs to 35 lakhs.
3. Possible Inputs and Outputs
• Inputs: car price, fuel type, Manufacturing Years, vehicle name, boundary
values like the lowest or highest price, etc.
• Outputs: Vehicles Manufacturing Years, Price Range, Seller Analysis,
Depreciation Analysis, Two-Wheelers, etc.

4. Validation
• We can use different conditions to check the analyses we got is correct or
not. It is most important check the data to ensures that the analysis results
are accurate and the data used is complete and reliable.
9 Testing Material
(Screenshots of working
outputs. Images in case of
hardware project
10 User Manual 1. Introduction
• This project aims to analyze the Car Dekho dataset to uncover key trends,
insights, and anomalies in the used car market.
2. Requirements
• Software: python ide or google colab and libraries (Pandas, Matplotlib,
Seaborn).
• Hardware: A computer with at least 4GB RAM.
3. Installation
1. Install Python from python.org you can use google colab.
2. Install pandas, matplotlib, seaborn, libraries
3. Download the project files.
4. How to Use
1. Load Data:
import pandas as pd
df = pd.read_csv('car_Dekho_DA.csv')
2. View Data: Use df.head() to see the first 5 rows.
3. Analyze Data: Run the code to analyse.
5. Testing & Validation
• Check Edge Cases: Test with the lowest and highest car prices.
• Validate Input: Make sure the program handles all types of analyzes
correctly.
7. Conclusion
• This project helps you understand car dataset through analysis and
visualization.

11 Technical Documentation Project Overview


• Objective: To analyze car data from the Car Dekho dataset to uncover
trends, insights, and anomalies.
• Tools Used: Python with Pandas for data manipulation, and Matplotlib &
Seaborn for visualization.

Dataset Description
• Data used: Car Dekho dataset.
• Key elements:
o The manufacturer of the car.
o The specific model of the car.
o The manufacturing year.
o The price of the car.
o The fuel efficiency of the car.
o The type of fuel used like Petrol, Diesel, CNG.
o The type of transmission like Manual, Automatic.

Code Structure
• Data Loading:
Python Code to load data:
o import pandas as pd
o df = pd.read_csv('car_Dekho_DA.csv')
• Data Summary:
df.describe()
• Data Info:
df.info()
• Data Cleaning:
o Must check if any null values are present in the data
• Data Visualization:
o Bar Charts: Transmission type analysis, fuel type distribution,
seller type distribution.
o Pie Charts: Fuel type and seller type distributions.
o Scatter Plots: Cost depreciation vs. kilometers driven, selling
price vs. kilometers driven.
o Box Plot: Selling price distribution of two-wheelers.
o Horizontal Bar Graphs: Most sold vehicle models, overall vehicle
count.
o Purpose: To provide a clear and concise visual summary of key
findings.
o Tools Used: Matplotlib and Seaborn for creating comprehensive
and insightful visualizations.
• Data Analysis:
o Investigate manufacturing years to identify trends over time.
o Analyze price ranges to understand market segments.
o Examine vehicle diversity to assess the variety of cars available.
o Evaluate sales trends to determine peak sales periods.
o Study the impact of CNG on sales and cost depreciation.

Common Issues and Solutions


• Data file not found: Ensure the dataset is correctly placed in the project
directory.
• Data not displaying: Check that all required libraries are installed, and
and no missing value is present.

12 References ▪ Kaggle – https://www.Kaggle.com/


▪ IBM https://skills.yourlearning.ibm.com/activity/PLAN-
D0B733510535

13 Daily Work Breakdown

Day 1 I participated in an EDUNET live session where I set up an IBM website account
and enrolled in the Edunet-TN Data Analytics with Python course. During the
session, I learned some foundational concepts in data analytics and practiced
solving questions with a dataset using Python in Google Colab.

Day 2 I practiced some functions using python in google collab on the dataset
• Head
• Tail
• Describe

Day 3 Completed Module 1 & 2 on Introduction to Data Concepts in IBM course. And
completed 2 quizzes based on module 1 & 2

Day 4 Completed Module 3 & 4 on Introduction to Data Concepts in IBM course. And
completed 2 quizzes based on module 3 & 4

Day 5 Completed Module 5 on Introduction to Data Concepts in IBM course. And


completed quiz based on module 5 and also completed final assignment based on
Introduction to Data Concepts

Week 3 From ___ June 14 _______ to ________ June 20____________

Day 1 Attended EDUNET live session. In this session I have learned about some
functions in python which we can use on dataset. And practiced some questions
based on dataset given.
• How to upload a dataset or file in google collab
• How to read a dataset in our code
• How to import panda library
• And some questions based on Head, Tail, Describe, info, isnull

Day 2 Practiced the questions given on live session

Day 3 Completed Module 1 & 2 on Data Science in Our World in IBM course. And
completed 2 quizzes based on module 1 & 2

Day 4 Completed Module 3 on Data Science in Our World in IBM course. And
completed quiz based on module 3

Day 5 Completed Module 4 & 5 on Data Science in Our World in IBM course. And
completed quizzes based on module 4 & 5 and completed final assignment based
on Data Science in Our World

Week 3 From ____ June 21 ______ to _________ June 27___________

Day 1 Attended EDUNET live session. In this session I have learned about graphs which
we can use on dataset. And practiced some questions based on dataset given.
• Depriciation
• Data visualization using graphs

Day 2 Completed Module 1 on Overview of Data Tools and Languages in IBM course.
And completed quiz based on module 1

Day 3 Completed Module 2 on Overview of Data Tools and Languages in IBM course.
And completed quiz based on module 2

Day 4 Completed Module 3 on Overview of Data Tools and Languages in IBM course.
And completed quiz based on module 3 and completed final assignment based on
Overview of Data Tools and Languages

Day 5 Completed Module 1 & 2 on Clean, Refine, and Visualize Data with IBM Watson
Studio in IBM course. And completed quizzes based on module 1 & 2

Week 4 From ____ June 28______ to _________ July 4___________

Day 1 Attended EDUNET live session. In this session I have practiced some questions
based on dataset given.

Day 2 Completed Module 3 & 4 on Clean, Refine, and Visualize Data with IBM Watson
Studio in IBM course. And completed quizzes based on module 3 & 4

Day 3 Completed Module 5 on Clean, Refine, and Visualize Data with IBM Watson
Studio in IBM course. And completed quiz based on module 5 and completed
final assignment based on on Clean, Refine, and Visualize Data with IBM Watson
Studio
Day 4 Completed Module 1,2 & 3 on Your Future in Data: The Job Landscape in IBM
course.

Day 5 Completed Module 4 on Your Future in Data: The Job Landscape in IBM course
and practiced some questions given in the Telegram group.

Week 5 From ___July 5_______ to _________July 11___________

Day 1 Attended EDUNET live session. In this session I have practiced some questions
which are given based on previous sessions.

Day 2 Started Python for programmer’s course given in the learning plan Data Analytics
with Python on IBM website. And completed modules control flow in python,
getting started with python

Day 3 I completed the modules on basic syntax, control flow, and functions in the Python
for Programmers course on the IBM website. Additionally, I practiced some data
analytics questions provided during the live session.

Day 4 Completed object-oriented programming in python module in Python for


programmer’s course on IBM website.

Day 5 Completed built-in data structures module in Python for programmer’s course on
IBM website. And practiced some questions which are given in the live session.

Week 6 From ___JULY 12_______ to ________JULY 18____________

Day 1 Attended EDUNET live session. In this session we have practiced some questions
and instructors discussed about project

Day 2 Analysed some of the questions related to project

Day 3 I analyzed several project-related questions and began the "Python for Data
Science" course outlined in the course plan on the IBM website, as recommended
by the internship instructors.

Day 4 Some part of Module 1 in python for data science in IBM website has completed

Day 5 Module 1 in python for data science in IBM website has completed and checked
some questions related to project which are previously completed.

Week 7 From ____JULY 19______ to _______JULY 25_____________

Day 1 Completed some of the questions based on data analytics on given dataset for the
project

Day 2 Completed module 2 & 3 in python for data science in IBM website
Day 3 Completed module 4 & 5 in python for data science in IBM website and also
completed some data analytics questions on given dataset related to project

Day 4 I finished all the modules and courses required for the internship, received my
completion certificate, and tackled some data analytics questions related to the
project using the provided dataset.

Day 5 Completed the project and submitted the ppt regarding project to edunet

Week 8 From __July 26________ to _________Aug 1___________

Day 1 Given final presentation regarding project to mentor

Day 2 Gathered some information about project report

Day 3 Started working on project report

Day 4 Completed some work regarding project report

Day 5 Completed some work regarding project report

From ___Aug 2_______ to ________Aug 4____________

Day 1 Completed some work regarding project report

Day 2 Completed some work regarding project report

Day 3 Completed some work regarding project report

You might also like