YashGoel Internship Report

Download as pdf or txt
Download as pdf or txt
You are on page 1of 33

INDRAPRASTHA GAS LIMITED

Internship Report

SUBMITTED IN PARTIAL FULFILLMENT FOR THE REQUIREMENT OF


THE AWARD OF DEGREE OF
BACHELOR OF TECHNOLOGY
IN
COMPUTER SCIENCE

Submitted by
Yash Goel
(2000290120195)

Session 2023-24

DEPARTMENT OF COMPUTER SCIENCE


KIET GROUP OF INSTITUTIONS, GHAZIABAD
(Affiliated to Dr. A. P. J. Abdul Kalam Technical University, Lucknow, U.P.,
India)
CERTIFICATE

This is to certify that the “Internship report” submitted by YASH GOEL is work
done by him and submitted during 2023 – 2024 academic year, in partial fulfillment of
the requirements for the award of the degree of BACHELOR OF TECHNOLOGY in
COMPUTER SCIENCE, at GHAZIABAD..

College Internship Coordinator Mr. Rahul Sir


Department Internship Coordinator Mr. Harshvardhan Sir

Head of the Department of


CS Dr Ajay kumar shrivastava
ACKNOWLEDGEMENT

I would like to thank Mr. Ravi Rashmimant, Chief HR, Indraprastha Gas Limited,
R.K Puram, Delhi for giving me the opportunity to do an internship with the
organization.

I also would like to thank all the people that worked along with me at Indraprastha Gas
Limited, ISID with their patience and openness, they created an enjoyable working
environment.

It is indeed with a great sense of pleasure and immense sense of gratitude that I
acknowledge the help of these individuals.

I am highly indebted to Senior Vice President Mr. S.K Behera and Additional
General Manager Mr. Sudhanshu Sharma, for the facilities provided to accomplish this
internship.

I am extremely grateful to my department staff members who helped me in successful


completion of this internship.

Project Supervisor -
Mr. Sudhanshu Sharma
(Additional General Manager)

Yash Goel

iv
ABSTRACT

This project aims to develop a slice- edge database monitoring dashboard as a summer
internship project at Indraprastha Gas Limited (IGL), a top supplier of renewable energy
results. The aim of this project is to produce a centralized, user-friendly platform exercising
Python's Tkinter GUI toolkit to track and display vital database information, Oracle
database connectivity to extract necessary data, machine learning technique such as ARIMA
to predict and forecast database growth, visualization libraries such as Matplotlib, Pyplot to
visualize the discovered insights.

The dashboard was intended to break the company's present lack of a thorough system to
track specific database sizes, the total size of all databases, and the proportion of storage
each database used. The suggested database dashboard provides a variety of features,
analogous as real- time data changes perceptivity, on- demand database size computation,
and dynamic data visualization using line charts and pie graphs. Database names are buttons
on the dashboard’s user-friendly design, allowing users to snappily pierce full information
about each database with just one click.

The design progressed through organized way of system analysis, design, development, and
evidence during the internship. Stakeholder interviews and demand collecting laid the root
for the creation of an effective GUI that gives users instant access to vital database
information. The result included establishing a MySQL and Oracle connection to the
databases, carrying statistics on data modification, and including dynamic line charts for
data growth pattern.

IGL places a great deal of significance on the accomplishment of this design since the
database dashboard offers perceptivity that may be used to enhance data operation and
decision- making procedures. Organization can effectively understand data dynamics
because of the dashboard's capacity to depict data changes. Also, automatic database size
estimations save time and capitalize while the consolidated platform promotes departmental
orchestration.

The connection and goods of the database dashboard design are shortly described in this
abstract. After the design was executed successfully, IGL was given a strong tool to
meliorate database conservation and use. The dashboard supports IGL’s dedication to
sacrifice sustainable, clean energy results by enabling real- time monitoring and
visualization.

v
Methodologies:

We follow a structured methodology for our design which starts from understanding the
problem to designing the result to the administering the result effectively. Well planned
project reduces the time to deliver the design and any fresh ad- hoc costs to the
association, hence we devote maturity of our time understanding the problem statement
and gather conditions. This ground up approach helps us deliver not only the result to the
problem but also facilitates the association’s functional conditions.

Key parts of the report:

Under each division we further give specific sedulity results on focused disciplines with
cutting edge technologies.

Benefits of the organization through our report:

 Under each division we further give specific sedulousness result on focused


disciplines with cutting edge technologies.

 We emphasize easing effective results to the association by delivering/ systems on


time.

vi
INDEX

S.no CONTENTS Page no


1. Introduction............................................................................................................1

1.1 Modules............................................................................................................ 2

2. System Analysis.......................................................................................................3

3. Software requirements specifications.......................................................................4

4. Technology used .......…………………………………………...…………………5

5. Coding…………………………………………………………...…………………10

6. Screenshots................................................................................................................11

7. Conclusion.................................................................................................................16

8. Bibliography………………………………………………………………………..17

vii
Learning Objectives/Internship Objectives

 Internships are generally allowed to be reserved for college students looking to


gain experience in a particular field. Still, a wide array of people can benefit from
Training externships to admit real world experience and develop their chops.

 An ideal for this position should emphasize the chops you formerly retain in the
area and your interest in learning further.

 Internships are employed in several different career fields, including architecture,


engineering, healthcare, economics, advertising and multitudinous further.

 Some internships are used to allow individuals to perform scientific disquisition


while others are specifically designed to allow people to gain first- hand
experience working.

 Utilizing internships is a great way to make your capsule and develop chops that
can be emphasized in your capsule for future jobs. When you are applying for a
Training internship, make sure to illuminate any special chops or bents that can
make you stand piecemeal from the rest of the applicants so that you have an
bettered chance of landing the position.

viii
1. INTRODUCTION

This internship project at Indraprastha Gas Limited (IGL) revolves around the creation of a
comprehensive and user-friendly database monitoring dashboard, aiming to enhance the company's
data operation and visualization capabilities. IGL, being a prominent provider of clean energy results,
relies on effective data handling to cover and anatomize its databases effectively. Still, the absence
of a centralized platform for database details posed challenges in gaining holistic perceptivity into
the databases performance and storage operation and taking major time in creating a report.

As a summer intern at IGL, I shouldered the responsibility of designing and administering a user
friendly dashboard using Tkinter, a Python GUI toolkit. The primary ideal was to empower IGL with
an intuitive interface that presents vital information about their databases at a regard. The dashboard
offers an array of features, including graphical representation of data differences, database sizes in
tabular form, overall database size. This report provides an overview of the design's methodology,
performance, results, and the precious knowledge exploits gained during the internship.

1
1.1Module Description:

Backend:
======
1) Task Schedule:
a. A python script is scheduled to run every week on Monday at 15:00.
b. The script extracts storage information from remote Oracle Database and stores it in form of a
data frame.
c. The data frame is then saved in local MYSQL Database.
2) ARIMA:
a. ARIMA, which stands for "AutoRegressive Integrated Moving Average," is a widely used
statistical method for time series forecasting and analysis. It's a popular approach for modeling
and predicting future values based on the patterns and trends observed in historical data. Here's a
nutshell explanation of ARIMA:
o AutoRegressive (AR): The "AR" part of ARIMA refers to the autoregressive component.
This component considers the relationship between the current value in a time series and its
previous values.
o Integrated (I): The "I" in ARIMA stands for integrated. This component involves
differencing the time series data to make it stationary. Stationarity is an important concept in
time series analysis, and it means that statistical properties of the data (like mean and
variance) remain constant over time. Differencing helps remove trends and make the data
stationary.
o Moving Average (MA): The "MA" part refers to the moving average component. This
component considers the relationship between the current value in a time series and a linear
combination of past error terms (residuals from previous predictions).

Frontend:

1. Tkinter :

Tkinter is a Python library that serves as a graphical user interface (GUI) toolkit. It
allows developers to create visually appealing and interactive applications with ease.

2. Matplotlib:

Matplotlib is a plotting library for the Python programming language and its numerical
mathematics extension NumPy. It provides an object-oriented API for embedding plots into
applications using general-purpose GUI toolkits like Tkinter, wxPython, Qt, or GTK.

2
3. SYSTEM ANALYSIS

 Requirement

Analysis -

Existing System:

The existing system at Indraprastha Gas Limited (IGL) lacks a centralized platform or
dashboard that provides comprehensive details about the company's databases. Currently,
database monitoring and visualization are manual processes, involving querying databases and
maintaining spreadsheets.
This approach presents several challenges:
 Limited Real-time Insights: The absence of a centralized dashboard hinders real-time
monitoring of database performance and alterations. Stakeholders cannot access up-to-
date information quickly.
 Inefficient Data Visualization: Data alterations within each database are not visualized
effectively, making it difficult for stakeholders to grasp the dynamics of data changes
over time.
 Manual Database Size Calculation: Calculating the size of individual databases and
their cumulative size is a time-consuming task, prone to errors.

Proposed System:

The proposed system addresses the limitations of the existing system by introducing a
comprehensive and user-friendly database dashboard. The primary objective is to create an
efficient platform that empowers IGL to monitor, visualize, and manage database details
seamlessly. The proposed system consists of the following key features:
 Centralized Database Dashboard: The proposed system will have a centralized
dashboard built using Tkinter, a Python GUI toolkit. The dashboard will provide a
unified interface for accessing database details.
 Interactive Data Visualization: To enable stakeholders to understand data alterations
effectively, interactive bar graphs will be incorporated into the dashboard. Clicking on
each database button will display the corresponding bar graph, providing valuable
3
insights into data changes.
 Real-time Database Information: The proposed system will retrieve real-time database
information using Python's SQLAlchemy library. This ensures that stakeholders have
access to the most up-to-date data.
 Automated Database Size Calculation: The proposed system will automate the
calculation of database sizes, eliminating the need for manual intervention and enhancing
accuracy.
 Overall Database Size and Storage Percentage: The dashboard will display the total
size of each database, the overall size of all databases, and the percentage of storage
occupied by each database, providing a holistic view of the company's database
infrastructure.

Benefits of the Proposed System:


The implementation of the proposed system offers several benefits to IGL:
 Enhanced Data Visualization: Interactive bargraphs provide a visual representation of
data alterations, facilitating a deeper understanding of data dynamics.
 Real-time Insights: Real-time database information ensures that stakeholders have
access to the most current data, enabling data-driven decision-making.
 Efficient Data Management: Centralized database details enhance data cohesion and
streamline data management practices across departments.
 Time and Resource Savings: Automated database size calculation reduces manual effort,
leading to time and resource savings.
 Improved Decision-making: The dashboard's comprehensive insights enable
stakeholders to make informed decisions regarding database management and utilization.

4
4. SOFTWARE REQUIREMENTS SPECIFICATIONS

4.1 System configurations

The software requirement specification can be produced at the culmination of the analysis
task. The function and performance allocated to software as part of system engineering are
refinedby established a complete information description, a detailed functional description, a
representation of system behavior, and indication of performance and design constrain,
appropriate validate criteria, and other information pertinent to requirements.

Software Requirements:

• Operating system : Windows 7 and above.


• Coding Language : Python, MYSQL.
• Front-End : Visual Studio Code.
• Data Base : Oracle Database, MYSQL Server.

Hardware Requirement:

• System : 64-bit Operating System, x64-based processor.

• Hard Disk : 1TB.


• Ram : 4GB.

5
5. TECHNOLOGY

5.1 PYTHON

Python is a versatile and high-level programming language known for its simplicity, readability,
and broad applicability across various domains.

It was developed by Guido van Rossum and first released in 1991, Python has gained immense
popularity in the software development community due to its ease of use and robust capabilities.
In this project internship report, we provide an overview of Python and its role in our project.

Working with Python in this project has been a rewarding experience for me. Its user-friendly
syntax and rich ecosystem empowered us to deliver a robust and efficient solution within the
project's timeline.
We acknowledge Python's versatility and potential to drive innovation in various industries.

In conclusion, Python has proven to be an invaluable tool, contributing significantly to the success
of our project. Its wide range of applications and strong community support make it an excellent
choice for future development endeavors.

5.2 TKINTER

Tkinter is a Python library that serves as a graphical user interface (GUI) toolkit. It allows
developers to create visually appealing and interactive applications with ease.

As part of our project internship report, we present an overview of Tkinter and its role in
enhancing the user experience of our project.

Working with Tkinter has been an enriching experience for me. Its simplicity, versatility, and
cross-platform compatibility have empowered us to build a feature-rich and user-friendly
application.
In conclusion, Tkinter has proven to be a valuable technology that significantly enhanced the user
experience of our project.
Its ability to create visually appealing and interactive GUIs, along with its seamless integration
with Python, makes Tkinter an excellent choice for future GUI-based projects and applications.

6
5.3 MACHINE LEARNING

Machine Learning (ML) is a subset of artificial intelligence (AI) that empowers computers to
learn and improve from experience without being explicitly programmed. In this section of our
project internship report, we present an overview of machine learning and its significance in our
project.

Regression Model:

Regression is a fundamental and widely used statistical technique for predicting a numerical
outcome based on one or more input variables. It forms a linear relationship between the input
variables (also called independent or predictor variables) and the outcome variable (also known as
the dependent variable).

Data Visualization:

Data visualization is a crucial aspect of machine learning (ML) that involves representing data
visually through charts, graphs, plots, and other visual elements. It plays a significant role in the
entire ML pipeline, from data exploration and preprocessing to model evaluation and result
interpretation. In this section, we provide an overview of data visualization in machine learning
and its importance in extracting meaningful insights from data.

Working with machine learning has been a transformative experience for our team. Its ability to
harness the power of data and learn from it has elevated the project's functionality and user
experience.
In conclusion, machine learning has played a pivotal role in our project, enabling us to make
accurate predictions, optimize processes, and deliver personalized experiences. Its vast
applications across diverse industries make it a game-changer in the world of technology, and we
believe it will continue to shape the future of intelligent systems and data-driven decision-making.

7
5.4 DATA ANALYTICS

Data analytics is a process that involves examining, cleaning, transforming, and interpreting large
volumes of data to extract valuable insights and support data-driven decision-making. In this
section of our project internship report, we present an overview of data analytics and its
significance in our project.

Data Analytics Process:

Data Collection: The first step in data analytics involves gathering relevant data from various
sources, such as databases, APIs, or files.
Data Preparation: Data is then pre-processed and cleaned to ensure it is in a usable format for
analysis. This step involves handling missing values, removing duplicates, and transforming data
if necessary.
Data Exploration and Visualization: Data visualization techniques, such as charts, graphs, and
plots, are used to explore the data, identify patterns, and gain insights.
Data Analysis: Various statistical and analytical techniques are applied to the data to uncover
patterns, correlations, and trends.
Predictive Modelling and Machine Learning: Advanced data analytics involves building
predictive models using machine learning algorithms to make future predictions based on
historical data.
Interpretation and Reporting: The final step involves interpreting the results and creating
reports or presentations to communicate findings and recommendations to stakeholders.

Working with data analytics has been an enriching experience for our team. It allowed us to make
data- driven opinions, optimize processes, and uncover practicable perceptivity.

In conclusion, data analytics has proven to be a important tool in lodging meaningful information
from data. It serves as a foundation for validation- predicated decision- timber, enabling associations
to make informed choices and gain a competitive edge. The perceptivity gained through data
analytics have positively impacted our design's success, and we recognize its implicit to drive
invention and growth in various disciplines.

8
5.5 MySQL DATABASE

MySQL is an open- source and considerably used relational database operation system (RDBMS)
that stores and manages structured data. In this section of our design internship report, we present
an overview of MySQL Database and its significance in our design.

A database is like a data train in that it's a storage place for data. Like a data file, a database does
not present information directly to a user, the user runs an operation that accesses data from the
database and presents it to the stoners in an untreatable format. A database generally has two
factors the lines holding the physical database access data.

The DBMS is responsible for administering database structure, including maintaining the
connections between data in the database. icing that data is stored correctly, and the rules that
defining data connections are not violated. Recovering all data to a point of known consistency in
case of system failures and recovering all data to a point of known consistency in case of system
failures.

Working with MySQL in our design has been a precious knowledge and experience for our team.
It handed a reliable and scalable database result, easing indefectible data operation and recovery.

In conclusion, MySQL is a important and popular relational database management system,


considerably used in various operations and industriousness. Its open- source nature, scalability,
and effectiveness make it a favored choice for managing structured data. The successful
integration of MySQL in our design significantly contributed to its success, and we admit the
significance of using robust database results to ensure data responsibility and performance in
future systems.

9
6. SOURCE CODE

6.1 Automate Back end Code: (Task Scheduler)

10
SQL Back end Tables:

11
Back end Input:

12
13
14
15
16
17
18
19
Output Screenshots:

20
21
22
7. CHALLENGES FACED

Understanding Figma for GUI Design: I had to devote time to understanding the
functionality and interface of Figma, a collaborative design tool, in order to develop an
understandable and eye-catching dashboard. In order to master Figma's principles,
overcoming this barrier required research into internet resources, courses, and practice.

Learning Tkinter for GUI Development: Using Tkinter, a Python GUI toolkit, to
implement the dashboard provided a learning curve. To create a productive and user-friendly
interface, I had to obtain a complete grasp of Tkinter's capabilities, widgets, and event
handling.

Techniques and libraries for data visualization : It was necessary to be knowledgeable


about data visualization methods and Python libraries like Matplotlib in order to create
interactive bargraphs to visualize data changes. I put a lot of work into learning how to create
dynamic, educational bargraphs.

Access to IGL's Database: Due to security and privacy considerations, accessing the
company's database constituted a serious problem. Coordination with the IT department and
adherence to the company's data regulations were required in order to obtain authorization
and the requisite credentials to connect with the databases.

ARIMA Time Series Analysis: Integrating ARIMA technology into the project involved
comprehending time series analysis from scratch. Overcoming this challenge required a deep
dive into ARIMA concepts, understanding its implementation, and effectively applying it to
forecast data alterations.

23
8.CONCLUSION

The database dashboard design has been a significant achievement during my internship at
IGL. Our project is successfully completed and deployed in the organization’s server. It
has equipped the company with an effective tool to cover and visualize database details,
furnishing essential perceptive for data- driven decision- timber. The design's success would
not have been possible without the support and guidance of IGL's technical team and the
exposure to real- world challenges.

The challenges faced during the internship design presented precious knowledge openings
and strengthened my problem- working and specialized chops. By prostrating these
challenges, I successfully developed an effective database dashboard for IGL, furnishing
stakeholders with critical perceptivity into database details. The experience has not only
enhanced my technical capabilities but also inculcated in me the significance of perseverance
and severity in the face of challenges. The chops and knowledge gained throughout this trip
will really contribute to my future professional trials.

I am immensely thankful for this occasion, and I am confident that the database storage
monitoring dashboard will contribute to IGL’s continued success in plan making and decision
related to database growth.

24
8. BIBLOGRAPHY

The following websites are referred during the analysis and execution phase of the
project:

1. GeeksforGeeks (GfG) - https://www.geeksforgeeks.org/

2. Javatpoint - https://www.javatpoint.com/

3. Stack Overflow - https://stackoverflow.com/

25

You might also like