0% found this document useful (0 votes)
44 views52 pages

Steel

Uploaded by

Sanjay Reddy
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
44 views52 pages

Steel

Uploaded by

Sanjay Reddy
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 52

PREDICTING CRITICAL PARAMETERS ON BLAST FURNACE

AND PREDICTING CO/CO2 RATIO USING MACHINE LEARNING


A report submitted as a part of the Industrial Orientation Training in
IT & ERP Department,
Visakhapatnam Steel Plant
By
BONDALA HARSHA VARDHAN

Trainee No: 100030316


Department of Computer Science and Engineering (AI & ML)
Under the guidance of
Mrs. Sriya Basumallik
(Manager)

RASHTRIYA ISPAT NIGAM LIMITED (RINL), VISAKHAPATNAM


(Duration: 6th May 2024 to 1st June 2024)

CERTIFICATE
This is to certify that the following student of Gayatri Vidya Parishad College of
Engineering (Autonomous), Visakhapatnam, is engaged in the project work titled
PREDICTING CRITICAL PARAMETERS ON BLAST FURNACE AND PRE-
DICTING CO/CO2 RATIO USING MACHINE LEARNING from 6th May 2024
to 1st June 2024.

BONDALA HARSHA VARDHAN TR. NO: 100030316

In partial fulfillment of the degree BACHELORS OF TECHNOLOGY in Com-


puter Science and Engineering stream in Gayatri Vidya Parishad College of Engi-
neering (Autonomous), Visakhapatnam is a record of Bonafide work carried out by
them under my guidance and supervision during the period from 6th May 2024 to
1st June 2024.

Date: 1st June 2024

Place: Visakhapatnam

Mrs. Sriya Basumallik


(Manager)

IT & ERP Department


RINL-VSP

DECLARATION
I am the student under the guidance of Mrs Sriya Basumallik (Manager), IT & ERP
department, Visakhapatnam, steel plant, hereby declare that the project entitled
“PREDICTING CRITICAL PARAMETERS ON BLAST FURNACE AND
PREDICTING CO/CO2 RATIO USING MACHINE LEARNING” is an origi-
nal work done at Rashtriya Ispat Nigam Limited (RINL) Visakhapatnam Steel
Plant, submitted in partial fulfillment of the requirements for the award of Indus-
trial Training project work. I assure you that this project is not submitted in any
university or college.

Place: VISAKHAPATNAM

ACKNOWLEDGEMENT
We would like to express our deep gratitude for the valuable guidance of our
guide, Mrs Sriya Basumallik, ma’am(Manager), IT and ERP department, Visakha-
patnam, Steel plant for all her guidance, help and ongoing support throughout the
course of this work by explaining basic concepts of the yard management system
and their functioning along with industry requirements. We are immensely grateful
to you, madam. Without their inspiration and valuable support, our training would
never have taken off.

We sincerely thank the Training and Development Center (T&DC) for their guid-
ance during safety training and encouragement in the successful completion of our
training.

BONDALA HARSHA VARDHAN

YASWANTH ARREPU

B. SAI SHASHANK

NIRMAL NISHITH PEETHALA

BHARATH VENKAT PADALA

JONNA KOUSHIK

ABSTRACT
This internship project aims to apply machine learning techniques to predict criti-
cal parameters in blast furnace operations, with a specific focus on the CO/CO₂ ra-
tio. The CO/CO₂ ratio is a vital indicator of the efficiency and environmental im-
pact of the blast furnace process. Due to the complexity and interdependence of
variables in blast furnace operations, traditional modeling approaches often fall
short in accuracy and reliability. This project seeks to overcome these challenges
by utilizing advanced machine learning algorithms.

Throughout the internship, data from various sensors and control systems within
the blast furnace will be collected and preprocessed to create a robust dataset. Mul-
tiple machine learning models, including regression algorithms and neural net-
works, will be developed and validated using this dataset. Feature selection tech-
niques will be employed to identify the most influential parameters affecting the
CO/CO₂ ratio, enabling the creation of a more targeted and efficient predictive
model.

The expected outcome of this project is to demonstrate that machine learning mod-
els can significantly improve the accuracy of predictions for the CO/CO₂ ratio and
other critical parameters compared to traditional methods. The successful imple-
mentation of these models in a blast furnace environment has the potential to en-
hance operational efficiency, reduce energy consumption, and lower emissions.
This project highlights the transformative potential of machine learning in indus-
trial processes, contributing to more sustainable and efficient manufacturing prac-
tices

Keywords: Machine Learning Models, Data Preprocessing, Feature Engineering,


Regression Algorithms, Model Training and Validation, Hyperparameter Tuning,
Time Series Analysis, Predictive Analytics, Scikit-Learn, Performance Metrics.
Brief Overview of Steel Plant

Visakhapatnam Steel Plant (VSP) is the integrated steel plant of Rashtriya Ispat Nigam Limited
in Visakhapatnam, founded in 1971. VSP strikes everyone with a tremendous sense of awe, won-
der, and amazement as it presents a wide array of excellence in all its facets in scenic beauty,
technology, human resources, management, and product quality. On the coast of the Bay of Ben-
gal and by the side of scenic Gangavaram beach has tall and massive structures of technological
architecture, the Visakhapatnam Steel Plant. Nevertheless, the vistas of excellence do not rest
with the inherent beauty of location over the sophistication of technology-they march ahead,
parading one aspect after another.

The decision of the Government of India to set up an integrated steel plant at Visakhapatnam was
announced by then Prime Minister Smt. Indira Gandhi in Parliament on 17 January 1971. VSP is
the first coastal- based integrated steel plant in India, 16 km west of the city of destiny, Visakha-
patnam, bestowed with modern technologies; VSP has an installed capacity of 3 million tons per
annum of liquid steel and 2.656 million tons of saleable steel. The saleable steel here is in the
form of wire rod coils, Structural, Special Steel, Rebar, Forged Rounds, etc. At VSP, there lies
emphasis on total automation, seamless integration, and efficient up-gradation. This results in a
wide range of long and structural products to meet stringent demands of discerning customers in
India & abroad; SP product meets exalting international Quality Standards such as JIS, DIN,
BIS, BS, etc.

VSP has become the first integrated steel plant in the country to be certified to all the three inter -
national standards for quality (ISO -9001), Environment Management (ISO-14001) & Occupa-
tional Health & Safety (OHSAS—18001). The certificate covers quality systems of all opera-
tional, maintenance, and service units besides purchase systems, Training, and Marketing func-
tions spreading over 4 Regional Marketing offices, 20 branch offices & 22 stockyards located all
over the country. VSP, by successfully installing & operating Rs.460 crores worth of pollution
and Environment Control Equipment and converting the barren landscape more than 3 million
plants have made the steel plant, steel Township, and VP export Quality Pig Iron & Steel prod-
ucts Sri Lanka, Myanmar, Nepal, Middle East, USA. & South East Asia (Pig iron). RINL—VSP
was awarded ―Star Trading HOUSE‖ status during 1997-2000. Having established a fairly de-
pendable export market, VP Plans to make a continuous presence in the export market.
Different sections at the RINL VSP:

● Coke oven and coal chemicals plant

● Sinter plant

● Blast Furnace

● Steel Melt Shop

● Continuous casting machine

● Light and medium machine mills

● Calcining and refractive materials plant

● Rolling mills

● Thermal power plant

● Chemical power plant


Introduction

Background on Blast Furnace


A blast furnace is a crucial component in the iron and steel industry, used primarily for the ex-
traction of iron from its ores. The process involves the chemical reduction of iron ore (mostly
consisting of iron oxides) to produce molten iron, commonly known as pig iron, which can then
be further refined to produce steel.

Operation and Components


1. Structure and Design:

○ A blast furnace is a large, vertical cylindrical structure lined with refractory materials to
withstand high temperatures. It has a series of layers, each playing a specific role in the smelting
process.

2. Raw Materials:

○ Iron Ore: Typically in the form of hematite (Fe₂O₃) or magnetite (Fe₃O₄).

○ Coke: A form of carbon derived from coal, used as both a fuel and a reducing agent.

○ Limestone: Added as a flux to remove impurities in the form of slag.

○ Hot Blast: Preheated air injected into the furnace to aid combustion and maintain high
temperatures.

3. Process Stages:

○ Charging: Raw materials are added from the top of the furnace in alternating layers.

○ Reduction Zone: As the materials descend, coke burns with the hot blast, producing car-
bon monoxide (CO) which reduces the iron ore to molten iron.

○ Smelting Zone: At higher temperatures, the iron melts and collects at the bottom of the
furnace (hearth).
○ Slag Formation: Limestone reacts with impurities to form slag, which floats on the
molten iron and can be removed.

○ Tapping: Molten iron and slag are periodically tapped from the furnace for further pro-
cessing.

4. Gas Flow:

○ The upward flow of gasses (primarily CO, CO₂, and N₂) generated from the combustion
of coke and reduction reactions plays a critical role in the efficiency of the process.

Importance of Key Parameters

● Temperature: Critical for maintaining the proper chemical reactions. Multiple tempera-
ture measurements (such as CB_TEMP, STEAM_TEMP, TOP_TEMP) help monitor and control
the thermal conditions within different zones of the furnace.

● Pressure: Ensures the proper flow of gases and materials. Measurements like
CB_PRESS, O2_PRESS, and TOP_PRESS are vital for operational stability.

● Gas Composition: Parameters such as CO, CO₂, and O₂ content provide insights into
combustion efficiency and environmental impact.

● Flow Rates: Parameters like CB_FLOW, STEAM_FLOW, and O2_FLOW help in main-
taining the right balance of inputs for optimal operation.

● Auxiliary Conditions: Factors like atmospheric humidity (ATM_HUMID) and PCI rate
(Pulverized Coal Injection) influence the overall efficiency and need to be finely controlled.

Modern Challenges and Innovations


The traditional operation of blast furnaces involves complex interdependencies between various
parameters, making manual control challenging. Modern advancements focus on automating and
optimizing these processes using technology. Machine learning and data analytics are becoming
integral in predicting and controlling these parameters to enhance efficiency, reduce energy con-
sumption, and minimize environmental impact.

By leveraging data from sensors and control systems, machine learning models can provide
deeper insights into the blast furnace operations, predicting critical outcomes and allowing for
proactive adjustments. This project, for example, aims to utilize such techniques to predict the
CO/CO₂ ratio and optimize other operational parameters, paving the way for more sustainable
and efficient iron production.
Analyzing the dataset
This project focuses on enhancing the efficiency and sustainability of blast furnace operations
through the application of machine learning techniques. The primary goal is to predict critical
parameters, particularly the CO/CO₂ ratio, which is a key indicator of furnace performance and
environmental impact. The project involves analyzing a dataset containing various operational
parameters collected from the blast furnace, which will be used to develop robust predictive
models.

The dataset includes the following parameters:

● CB_FLOW (Coke Breeze Flow): The flow rate of coke breeze, which influences the
combustion process.

● CB_PRESS (Coke Breeze Pressure): The pressure of the coke breeze, impacting its dis-
tribution and combustion efficiency.

● CB_TEMP (Coke Breeze Temperature): The temperature of the coke breeze, which af-
fects its reactivity and combustion.

● STEAM_FLOW (Steam Flow): The flow rate of steam, used to control the temperature
and reactions within the furnace.

● STEAM_TEMP (Steam Temperature): The temperature of the steam, which influ-


ences the thermal balance in the furnace.

● STEAM_PRESS (Steam Pressure): The pressure of the steam, affecting its distribution
and efficiency.

● O2_PRESS (Oxygen Pressure): The pressure of oxygen, which is critical for combus-
tion and chemical reactions.

● O2_FLOW (Oxygen Flow): The flow rate of oxygen, directly impacting the combustion
process.

● O2_PER (Oxygen Percentage): The percentage of oxygen in the gas mixture, crucial
for maintaining optimal combustion.
● PCI (Pulverized Coal Injection): The rate at which pulverized coal is injected, affecting
the fuel-to-air ratio.

● ATM_HUMID (Atmospheric Humidity): The humidity of the surrounding air, which


can influence combustion and heat transfer.

● HB_TEMP (Hearth Bottom Temperature): The temperature at the bottom of the


hearth, indicating the heat distribution in the furnace.

● HB_PRESS (Hearth Bottom Pressure): The pressure at the bottom of the hearth, re-
lated to the internal pressure dynamics.

● TOP_PRESS (Top Pressure): The pressure at the top of the furnace, affecting gas flow
and reactions.

● TOP_TEMP1, TOP_TEMP2, TOP_TEMP3, TOP_TEMP4 (Top Temperatures):


Temperatures at various points at the top of the furnace, indicating thermal gradients and effi-
ciency.

● TOP_SPRAY (Top Spray): The application rate of water spray at the top of the furnace,
used for cooling and controlling reactions.

● TOP_TEMP (Overall Top Temperature): The average temperature at the top, reflect-
ing the overall thermal state.

● TOP_PRESS_1 (Alternate Top Pressure): Another measurement of the pressure at the


top, ensuring accuracy and consistency.

● CO (Carbon Monoxide): The concentration of CO, a byproduct of combustion, used to


gauge efficiency and emissions.

● CO2 (Carbon Dioxide): The concentration of CO₂, an important parameter for environ-
mental monitoring.

● H2 (Hydrogen): The concentration of hydrogen, indicating the presence of reducing


gases and efficiency of reduction reactions.

● SKIN_TEMP_AVG (Average Skin Temperature): The average temperature of the fur-


nace's outer surface, reflecting overall heat loss and insulation efficiency.

By analyzing these parameters using machine learning algorithms, the project aims to develop
predictive models that can accurately forecast the CO/CO₂ ratio and other critical parameters.
This will facilitate better decision-making, improve operational efficiency.
Overview of implementation of the solution in the view of ma-
chine learning

Data Cleaning, Analysis, and Pre-Processing


Data Cleaning

Data cleaning is an essential step to ensure the dataset's quality and integrity before performing
any analysis or modeling. In this project, the data cleaning process involved the following steps,
utilizing key Python libraries such as pandas, numpy, and matplotlib:

1. Resampling the Data:

- Original Dataset:The initial dataset consisted of sensor readings recorded every 10 minutes
over a five-month period.

- Resampling to Hourly Intervals: Using the pandas library, the data was remodeled to hourly
intervals. This transformation involved aggregating the 10-minute readings into one-hour periods
using the `resample` function in pandas. This function allowed for the calculation of averages,
sums, or other relevant statistics to ensure that the transformed dataset accurately represented
hourly trends.

2. Handling Missing Values:

-Identification of Missing Data: The dataset was examined for any missing or incomplete data
points using the `isnull` and `sum` functions in pandas. These functions helped in identifying the
extent of missing data across different columns.
- Imputation or Removal: Depending on the extent and pattern of missing data, appropriate
strategies were employed. Missing values were filled using the `fillna` or `interpolate` functions,
or rows with excessive missing data were removed using the `dropna` function in pandas.

3. Feature Scaling and Normalization:

-Standardization: The numpy library, along with the `StandardScaler` from the `sklearn.prepro-
cessing` module, was used for standardizing the data. This ensured that all features contributed
equally to the model training process by scaling features to have a mean of zero and a standard
deviation of one.

4. Outlier Detection and Treatment:

- Identification of Outliers: Outliers in the data were identified using statistical methods or vi-
sualization techniques. The `boxplot` function in the `matplotlib.pyplot` module was used to cre-
ate box plots, which helped in visualizing the spread of the data and detecting any outliers.

- Handling Outliers: Depending on the nature of the outliers, they were either removed or
treated. Treatment could involve capping values or using robust statistical methods to minimize
their impact.

Data Analysis and Pre-Processing

Once the data was cleaned, the next step involved an in-depth analysis to uncover patterns and
relationships within the dataset. The data analysis phase included:

1. Exploratory Data Analysis (EDA):


- Descriptive Statistics: Summary statistics were computed using pandas functions to under-
stand the central tendencies, dispersion, and overall distribution of the data. Metrics such as
mean, median, standard deviation, and percentiles were calculated.

- Data Visualization: Various visualization techniques were employed to explore the data vis-
ually using matplotlib. This included:

- Histograms: Created to understand the distribution of individual features.

- Box Plots: Used to detect outliers and visualize the spread of the data.

- Time Series Plots: Employed to observe trends and patterns over time for the hourly resam-
pled data.

- Scatter Plots: Used to explore relationships between different features.

2. Correlation Analysis:

- Correlation Matrix: A correlation matrix was generated using the `corr` function in pandas to
quantify the linear relationships between different features. This helped in identifying which fea-
tures had strong positive or negative correlations with each other.

- Heatmaps: Visual representations of the correlation matrix were created using the `heatmap`
function from the seaborn library, making it easy to identify significant correlations.

3. Feature Engineering:

- Creating New Features: Based on the insights from the exploratory data analysis, new fea-
tures were engineered. For instance, to predict the CO/CO2 ratio, four new columns representing
the ratio for the 1st, 2nd, 3rd, and 4th hours were created.

- Transformation of Existing Features: Existing features were transformed or combined to cre-


ate more meaningful representations that could improve the model’s predictive power.

4. Pattern and Trend Analysis:

- Temporal Patterns: Analysis was performed to detect temporal patterns, such as daily or
weekly cycles, in the data.
- Anomaly Detection: Any anomalies or unusual patterns were identified and investigated to en-
sure they were understood and appropriately handled.

Modules used:
- pandas:

- Description: Pandas is used for data manipulation and analysis with dataframes, akin to
spreadsheets. It handles creating new columns, handling missing values, and reading/writing data
to/from Excel files.

-matplotlib.pyplot:

- Description: This module is part of the Matplotlib library and is used for creating visualiza-
tions, like the correlation matrix plot in this case.

- seaborn:

- Description: Seaborn enhances Matplotlib's capabilities, providing stylish visualizations.


While not directly called in the code, Matplotlib likely utilizes Seaborn's color palettes for the
correlation matrix plot.

Code for Data Cleaning and Analysis:

This code imports necessary libraries (pandas, matplotlib.pyplot, and seaborn) and reads data
from an Excel file located at "/content/bf3_data_2022_01_07.xlsx" into a pandas dataframe
(data). This sets up the environment for further data analysis and visualization.

python
import pandas as pd

from sklearn.model_selection import train_test_split

from sklearn.ensemble import RandomForestRegressor

from sklearn.metrics import mean_squared_error, mean_absolute_error, r2_score

import matplotlib.pyplot as plt

import numpy as np

data = pd.read_excel('/content/bf3_data_2022_01_07.xlsx')

data.drop(columns=['SKIN_TEMP_AVG'], inplace=True)

data['DATE_TIME'] = pd.to_datetime(data['DATE_TIME'])

data.dropna(inplace=True)

```

This code retrieves the column names of the dataframe (data), removes any rows with missing
values using `dropna()`, and then prints the data types of each column in the dataframe using
`dtypes`. This process ensures the dataset is clean and ready for analysis.

def train_and_predict(X, y, shift_hours):

X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)

model = RandomForestRegressor(n_estimators=100, random_state=42)

model.fit(X_train, y_train)

predictions = model.predict(X)

predicted_df = pd.DataFrame(predictions, columns=X.columns)

predicted_df['DATE_TIME'] = X.index + pd.to_timedelta(shift_hours, unit='h')

mse = mean_squared_error(y_test, model.predict(X_test))


mae = mean_absolute_error(y_test, model.predict(X_test))

r2 = r2_score(y_test, model.predict(X_test))

print(f"Shift {shift_hours} hours - Mean Squared Error: {mse}")

print(f"Shift {shift_hours} hours - Mean Absolute Error: {mae}")

print(f"Shift {shift_hours} hours - R-squared: {r2}")

return model, predicted_df

```

The code creates a new dataframe (predicted_df) by copying the columns from the original
dataframe (data). It then converts the `DATE_TIME` column to datetime format. Next, it calcu-
lates the correlation matrix (corr) of the original dataframe. Finally, it plots the correlation matrix
using matplotlib, displaying the strength of relationships between different features with a color-
coded heatmap.

X = data.drop(columns=['DATE_TIME'])

X.index = data['DATE_TIME']

y = X.copy()

```

The code creates new columns in new_df for predicting the CO/CO2 ratio at future time intervals
(1, 2, 3, and 4 hours ahead) using the `shift` function. These new columns are `Next_1hr`,
`Next_2hr`, `Next_3hr`, and `Next_4hr`. The modified dataframe is then saved to an Excel file
("data/demo3.xlsx"). Finally, the saved Excel file is read back into a new dataframe (df) and
printed.

```

model_shift_1, predicted_df_shift_1 = train_and_predict(X, y, shift_hours=1)

predicted_df_shift_1['CO_CO2_ratio'] = predicted_df_shift_1['CO'] /
predicted_df_shift_1['CO2']
predicted_df_shift_1.to_csv('predicted_data_after_1_hour.csv', index=False)

print(predicted_df_shift_1)

X_shift_2 = predicted_df_shift_1.drop(columns=['DATE_TIME'])

X_shift_2.index = predicted_df_shift_1['DATE_TIME']

model_shift_2, predicted_df_shift_2 = train_and_predict(X_shift_2, X_shift_2, shift_hours=1)

predicted_df_shift_2['CO_CO2_ratio'] = predicted_df_shift_2['CO'] /
predicted_df_shift_2['CO2']

predicted_df_shift_2.to_csv('predicted_data_after_2_hours.csv', index=False)

print(predicted_df_shift_2)

X_shift_3 = predicted_df_shift_2.drop(columns=['DATE_TIME'])

X_shift_3.index = predicted_df_shift_2['DATE_TIME']

model_shift_3, predicted_df_shift_3 = train_and_predict(X_shift_3, X_shift_3, shift_hours=1)

predicted_df_shift_3['CO_CO2_ratio'] = predicted_df_shift_3['CO'] /
predicted_df_shift_3['CO2']

predicted_df_shift_3.to_csv('predicted_data_after_3_hours.csv', index=False)

predicted_df_shift_3

X_shift_4 = predicted_df_shift_3.drop(columns=['DATE_TIME'])

X_shift_4

.index = predicted_df_shift_3['DATE_TIME']

model_shift_4, predicted_df_shift_4 = train_and_predict(X_shift_4, X_shift_4, shift_hours=1)


predicted_df_shift_4['CO_CO2_ratio'] = predicted_df_shift_4['CO'] /
predicted_df_shift_4['CO2']

predicted_df_shift_4.to_csv('predicted_data_after_4_hours.csv', index=False)

predicted_df_shift_4

```

Model Training and Evaluation

Model Training:

For this project, the `RandomForestRegressor` from the `sklearn.ensemble` module was chosen
for its robustness and accuracy in handling regression tasks. The training process included:

1. Splitting the Data:

- Train-Test Split: The dataset was split into training and testing sets using the
`train_test_split` function from `sklearn.model_selection`. Typically, 80% of the data was used
for training and 20% for testing.

- Feature Selection: The relevant features for training were selected, excluding the target vari-
able. Features with high correlation or importance to the target variable were prioritized.
2. Model Initialization and Training:

- Initialization: The `RandomForestRegressor` was initialized with specific parameters, such


as the number of estimators (trees) and random state for reproducibility.

- Training: The model was trained using the training set. The `fit` function was used to train
the model on the input features (X_train) and the target variable (y_train).

Model Evaluation:

Evaluating the model’s performance was critical to ensure it could accurately predict future
CO/CO2 ratios. The evaluation process involved:

1. Prediction:

- Making Predictions: The trained model was used to make predictions on both the training
and testing sets using the `predict` function. This generated predicted values for the CO/CO2 ra-
tio.

2. Performance Metrics:
- Mean Squared Error (MSE): Calculated using the `mean_squared_error` function from
`sklearn.metrics`, MSE measures the average squared difference between actual and predicted
values.

- Mean Absolute Error (MAE): Calculated using the `mean_absolute_error` function from
`sklearn.metrics`, MAE measures the average absolute difference between actual and predicted
values.

- R-squared (R2) Score: Calculated using the `r2_score` function from `sklearn.metrics`, the
R2 score measures the proportion of the variance in the target variable that is predictable from
the input features.

3. Model Performance Interpretation:

- Shifted Predictions: The model was evaluated for different shifts (1, 2, 3, and 4 hours) to as-
sess its performance over varying prediction horizons. The performance metrics for each shift
were calculated and analyzed.

- Visualizing Predictions: The predicted values were plotted against actual values to visually
assess the model’s accuracy. This helped in identifying any patterns or discrepancies.

Insights and Conclusion

The project provided several key insights:

1. Correlation and Feature Importance: The correlation matrix and visualizations highlighted
the relationships between different features and the target variable, guiding the feature selection
process.
2. Model Performance: The `RandomForestRegressor` demonstrated good performance in pre-
dicting CO/CO2 ratios, with reasonable MSE, MAE, and R2 scores across different shifts.

3. Future Work: Potential areas for improvement include experimenting with other regression
models, performing hyperparameter tuning, and incorporating additional features to enhance pre-
dictive accuracy.

Overall, the project successfully utilized data cleaning, analysis, and machine learning tech-
niques to predict future CO/CO2 ratios, providing valuable insights into the behavior of the sys-
tem under study.

This detailed approach, leveraging various data manipulation and machine learning techniques,
ensures the robustness and reliability of the model, contributing to accurate predictions and
deeper understanding of the dataset.

Introduction to Project
The Production Management System is a comprehensive web application designed to streamline
production processes within an organization. It enables efficient tracking of employee activities
and product manufacturing based on specific requirements. The system streamlines production
processes by providing a centralized platform for employees to record their daily activities and
manage product manufacturing. Employees can quickly log their work, reducing paperwork and
manual data entry.

By recording daily activities and product details, the system ensures accurate data capture. Man-
agers and supervisors can monitor employee productivity, identify bottlenecks, and make in-
formed decisions. Employees register using their unique employee IDs, promoting accountabil-
ity. Daily activity records create transparency, allowing supervisors to track progress and address
any issues promptly.

The system generates graphs and histograms using Highcharts. Visualizing production trends,
employee performance, and product quantities helps management make data-driven decisions.
As the organization grows, the system can handle an increasing number of users and data points.
Customizable features allow adaptation to specific requirements. Automated data entry mini-
mizes human errors.

Employees no longer need to duplicate efforts by manually recording activities and product in-
formation. Managers can generate reports based on historical data. These reports aid in perfor-
mance evaluation, resource allocation, and process optimization. User authentication ensures that
only authorized personnel can access the system. Sensitive data (such as employee IDs) remains
secure.

The Production Management System enhances efficiency, transparency, and data-driven deci-
sion-making. By combining frontend technologies (HTML, CSS, JavaScript) with backend logic
(Java, Oracle), we create a robust system that empowers employees and optimizes production
processes.

Software Requirements

★ HTML

Hypertext Markup Language (HTML) is a markup language used to create web pages and web
applications. HTML was created by Tim Berners-Lee in 1990. It is a simple language that is easy
to understand and modify, featuring many formatting tags that facilitate the creation of effective
presentations. HTML is platform-independent, allowing it to be displayed on any device or oper-
ating system.
HTML is called a markup language because it enables users to organize the appearance and pre-
sentation of content on a web page. It instructs the web browser on how to display the page's text
and images to the user. HTML is a text-based approach to describing content structure within an
HTML file. For example, content can be organized into paragraphs, bulleted lists, or structured
using images and data tables.

Some features of HTML include:

➢ - Easy to learn and use


➢ - Platform-independent
➢ - Ability to add images, videos, and audio to a web page
➢ - Capability to include hypertext within the text
➢ - Flexible design options for web pages
➢ - Option to add links to web pages

★ CSS

Cascading Style Sheets (CSS) is a language that allows developers and designers to control the
presentation of HTML documents. CSS can define how a web page should look, including its
colors, fonts, spacing, and the positioning of elements within the browser. CSS was invented by
Håkon Wium Lie.

CSS was introduced to enhance web presentation capabilities by making web page code less
heavy and clunky. Before CSS, HTML on each page contained most attributes that defined a
web page's appearance. CSS offers more attributes and lists than HTML, giving web pages a
more vibrant look and feel. CSS is also easy to learn and understand and is easily readable by
search engines.

Features of CSS include:

➢ - Time-saving: Write CSS once and reuse the same sheet across multiple HTML pages.
➢ - Easy Maintenance: Make a global change by simply altering the style sheet, and all ele-
ments on all web pages will update automatically.
➢ - Search Engine-Friendly: CSS is considered a clean coding technique, meaning search
engines can easily read its content.
➢ - Superior Styles: CSS offers a much wider array of attributes than HTML, allowing for a
far better appearance compared to using HTML attributes alone.
➢ - Offline Browsing: CSS can store web applications locally using an offline cache, en-
abling the viewing of websites offline.

Python

Description: Python is a versatile, high-level programming language known for its readability,
simplicity, and wide range of applications. It supports multiple programming paradigms and is
popular in web development, data analysis, artificial intelligence, scientific computing, and
more.

Key Features:

● Easy to Learn and Use: Python has a simple syntax that is easy to read and write, making
it an excellent choice for beginners and experienced programmers alike.
● Versatile: Python can be used for web development, data analysis, artificial intelligence,
machine learning, scientific computing, automation, and more.
● Extensive Standard Library: Python comes with a vast standard library that includes mod-
ules and packages for virtually any task, from file I/O and system calls to web browsers
and GUI development.
● Cross-Platform Compatibility: Python is compatible with various operating systems, in-
cluding Windows, macOS, and Linux, allowing for seamless development and deploy-
ment across different platforms.
● Integration Capabilities: Python can easily integrate with other languages and technolo-
gies, making it suitable for a wide range of applications and systems.
● Object-Oriented and Procedural: Python supports both object-oriented and procedural
programming paradigms, providing flexibility in how you structure your code.

Flask Framework

Description: Flask is a lightweight, micro web framework for Python that is designed to make it
easy to create simple, yet powerful web applications. It is known for its simplicity, flexibility,
and fine-grained control, making it a popular choice for both beginners and advanced developers.

Key Features:

● Lightweight and Flexible: Flask is a microframework, meaning it provides the essential


components needed to build web applications while allowing developers to choose and
integrate additional tools and libraries as needed.
● Simplicity and Ease of Use: Flask has a simple and intuitive API, making it easy to learn
and use. It allows developers to quickly get a web application up and running with mini-
mal setup.
● Built-in Development Server and Debugger: Flask includes a built-in development server
and debugger, which makes testing and debugging applications easier during the develop-
ment process.
● Modular Design: Flask's modular design allows developers to use extensions to add func-
tionality, such as form validation, authentication, and database integration, without un-
necessary complexity.
● RESTful Request Handling: Flask supports RESTful request handling, making it ideal for
developing RESTful APIs and services.
● Strong Community and Ecosystem: Flask has a vibrant community and a rich ecosystem
of extensions and plugins, providing additional functionality and support.
● Extensible: Flask is designed to be highly extensible, allowing developers to customise
and extend the framework to meet their specific needs.

SQLite3:-
SQLite3 is a C-language library that provides a lightweight, disk-based database. Unlike tradi-
tional database management systems (DBMS), it is an embedded SQL database engine, meaning
it doesn't require a separate server process. SQLite3 is highly portable, easy to use, and well-
suited for small to medium-sized applications, making it an ideal choice for web applications,
embedded systems, and standalone software.

Key Features:-
➢ Self-contained and Serverless: SQLite3 is self-contained, meaning it has no external de-
pendencies, and serverless, meaning it does not require a separate server process, making
it easy to deploy and use.
➢ Lightweight: SQLite3 is a compact library that requires minimal setup and consumes
minimal resources, making it suitable for embedded systems and mobile devices.
➢ Cross-Platform: SQLite3 is cross-platform and runs on various operating systems includ-
ing Windows, Linux, and macOS, providing consistent functionality across different en-
vironments.
➢ SQL Support: SQLite3 supports most of the SQL-92 standard, providing a robust SQL
interface for querying and managing data.
➢ Database File: The entire database, including tables, indexes, and data, is stored in a sin-
gle disk file, simplifying storage and transfer.
➢ High Performance: Despite its lightweight nature, SQLite3 is optimized for high perfor-
mance with efficient query processing and fast read/write operations.
➢ Extensible: SQLite3 can be extended with custom functions, aggregates, and collations
using the user-defined function (UDF) mechanism, allowing for tailored functionality.

Module
1. Login Page:
2. Signup Page:

3. Forget Password Page:


4.

Reset Password Page:

5. Main

dashboard:
6. Add
De- partment
Page:

7. Edit De- partment


Page:
8. De- partment
Page:

9. Graphs
Page:
Flowchart
CODE
CONNECTION.PY
DATABASE.PY
ADD DEPARTMENT.html

DEPARTMENT PAGE.HTML
EDIT DEPARTMENT.HTML
FORGET_PASSWORD.HTML

GRAPHS.HTML
LOGIN.HTML
MAIN PAGE.HTML
SIGNUP.HTML
STYLE.CSS
ADD DEPARTMENT.CSS
Conclusion
In conclusion, both the blast furnace prediction system and the Production Management System
highlight the transformative potential of technology in industrial settings. Leveraging Python
modules such as pandas, numpy, matplotlib, and Scikit-learn, the blast furnace prediction system
effectively cleansed data, trained predictive models, and visualized results, ultimately offering
valuable insights for optimizing operations. Similarly, the Production Management System, em-
ploying a combination of frontend technologies like HTML, CSS, JavaScript, and backend logic
in Java and Oracle, significantly enhances efficiency, transparency, and data-driven decision-
making in production processes.

The blast furnace prediction system's utilization of Gradient Boosting and XGBoost ensemble
methods showcases Python's versatility in handling complex data analysis and machine learning
tasks. Furthermore, Matplotlib's inclusion enables dynamic graph generation, facilitating user
comprehension and interaction with prediction outcomes.

In the case of the Production Management System, the integration of authentication mechanisms
ensures secure access, while a user-friendly interface simplifies employee interaction. Java's ro-
bust backend logic, coupled with Oracle's database management, facilitates seamless CRUD op-
erations and database connectivity, enabling efficient management of employee activities and
product manufacturing.
Both systems underscore the importance of technology in driving operational excellence and pro-
ductivity gains in industrial environments. By harnessing the capabilities of Python and associ-
ated technologies, organizations can make informed decisions, optimize processes, and achieve
tangible business outcomes.

You might also like