0% found this document useful (0 votes)
12 views

Web Application

The document describes the steps to build a web application for heart disease classification using machine learning. It discusses problem definition, data collection, data preparation, modeling, evaluation and deployment. Gradient boosting performed best for classification. The model was saved and deployed in a web app using Flask.

Uploaded by

ouiam ouhdifa
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
12 views

Web Application

The document describes the steps to build a web application for heart disease classification using machine learning. It discusses problem definition, data collection, data preparation, modeling, evaluation and deployment. Gradient boosting performed best for classification. The model was saved and deployed in a web app using Flask.

Uploaded by

ouiam ouhdifa
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 13

The Lifecycle to Build a Web Application for Prediction from

Scratch
BE G I NNE R M A C HI NE LE A RNI NG PRO G RA M M I NG PYT HO N S T RUC T URE D D AT A T E C HNI Q UE UNC AT E G O RI Z E D

The data science lifecycle is designed for big data issues and data science projects. Generally, the data
science project consists of seven steps which are problem definition, data collection, data preparation,
data exploration, data modeling, model evaluation and model deployment. This article goes through the
data science lifecycle in order to build a web application for heart disease classification.

If you would like to look at a specific step in the lifecycle, you can read it without looking deeply at the
other steps.

Problem Definition

Clinical decisions are often made based on doctors’ experience and intuition rather than on the knowledge-
rich hidden in the data. This leads to errors and many costs that affect the quality of medical services.
Using analytic tools and data modeling can help in enhancing the clinical decisions. Thus, the goal here is
to build a web application to help doctors in diagnosing hear t diseases. The full code of is available in my
GitHub repository.

Data Collection
I collected the heart disease dataset from UCI ML. The dataset has the following 14 attributes:

age: age in years.


sex: sex (1=male; 0=female).
cp: chest pain type (0 = typical angina; 1 = atypical angina; 2 = non-anginal pain; 3: asymptomatic).
trestbps: resting blood pressure in mm Hg on admission to the hospital.
chol: serum cholesterol in mg/dl.
fbs: fasting blood sugar > 120 mg/dl (1=true; 0=false).
restecg: resting electrocardiographic results ( 0=normal; 1=having ST-T wave abnormality; 2=probable
or definite left ventricular hypertrophy).
thalach: maximum heart rate achieved.
exang: exercise-induced angina (1=yes; 0=no).
oldpeak: ST depression induced by exercise relative to rest.
slope: the slope of the peak exercise ST segment (0=upsloping; 1=flat; 2=downsloping).
ca: number of major vessels (0–3) colored by fluorosopy.
thal: thalassemia (3=normal; 6=fixed defect; 7=reversable defect).
target: heart disease (1=no, 2=yes).

Data Preparation and Exploration

Here is a snapshot of the data header.

The header of the heart disease dataset

From the first look, the dataset contains 14 columns, 5 of them contain numerical values and 9 of them
contain categorical values.

The dataset is clean and contains all the information needed for each variable. By using info(),
describe(), isnull() functions, no errors, missing values, and inconsistencies values are detected.

#Check null values df.isnull().sum()


Null values in the dataset

By checking the percentage of the persons with and without heart diseases, it was found that 56% of the
persons in the dataset have heart disease. So, the dataset is relatively balanced.

People with and


without heart disease
in the dataset

Attributes Correlation
This heatmap shows the correlations between the dataset attributes, and how the attributes interact with
each other. From the heatmap, we can observe that the chest pain type (cp), exercise-induced angina
(exang), ST depression induced by exercise relative to rest (oldpeak), the slope of the peak exercise ST
segment (slope), number of major vessels (0–3) colored by flourosopy (ca) and thalassemia (thal) are
highly correlated with the heart disease (target). We observe also that there is an inverse proportion
between heart disease and maximum heart rate (thalch).

Moreover, we can see that the age is correlated with number of major vessels (0–3) colored by flourosopy
(ca) and maximum heart rate (thalch). There is also a relation between ST depression induced by exercise
relative to rest (oldpeak) and the slope of the peak exercise ST segment (slope). Moreover, there is a
relation between chest pain type (cp) and exercise-induced angina (exang).

Next, we will analyze these correlations between these features further.

1. Age and Maximum Heart Rate


Heart disease is arising frequently in older people, and the max heart rates are lower for old people with
heart disease.

2. Chest Pain

There are four types of chest pain: typical angina, atypical angina, non-anginal pain, and asymptomatic.
Most of the heart disease patients are found to have asymptomatic chest pain.

2. Chest Pain and Exercise-Induced Angina


The people who have exercise-induced angina; they usually suffer from asymptomatic chest pain, and they
are more likely to have heart disease.

3. Thalassemia

People with reversible defects are likely to have heart disease.

4. ST depression and the Slope of the Peak Exercise ST Segment.


The people who have downsloping ST segment have higher values of ST depression and more chance to be
infected with heart disease. The greater the ST depression, the greater the chance of disease.

5. Age and Number of Major Vessels (0–3) Colored by Flourosopy.

Most of the heart disease patients are old and they have one or more major vessels colored by Flourosopy.

Data Modeling

Let’s create the machine learning model. We are trying to predict whether a person has heart disease. We
will use the ‘target’ column as the class, and all the other columns as features for the model.

# Initialize data and target target = df[‘target’] features = df.drop([‘target’], axis = 1)

– Data Splitting

We will divide the data into a training set and test set. 80% of the data will be for training and 20% for
testing.

# Split the data into training set and testing set X_train, X_test, y_train, y_test =
train_test_split(features, target, test_size = 0.2, random_state = 0)
– Machine Learning Model

Here, we will try the below machine learning algorithms then we will select the best one based on its
classification report.

Support Vector Machine


Random Forest
Ada Boost
Gradient Boosting

The following function for training and evaluating the classifiers.

def fit_eval_model(model, train_features, y_train, test_features, y_test): results = {} # Train the model
model.fit(train_features, y_train) # Test the model train_predicted = model.predict(train_features)
test_predicted = model.predict(test_features) # Classification report and Confusion Matrix

results[‘classification_report’] = classification_report(y_test, test_predicted) results[‘confusion_matrix’]


= confusion_matrix(y_test, test_predicted) return results

Initialize models, train and evaluate.

# Initialize the models sv = SVC(random_state = 1) rf = RandomForestClassifier(random_state = 1) ab =


AdaBoostClassifier(random_state = 1) gb = GradientBoostingClassifier(random_state = 1)# Fit and evaluate

models results = {} for cls in [sv, rf, ab, gb]: cls_name = cls.__class__.__name__ results[cls_name] = {}
results[cls_name] = fit_eval_model(cls, X_train, y_train, X_test, y_test)

Now, we will print the evaluation results.

# Print classifiers results for result in results: print (result) print()for i in results[result]: print (i,
‘:’) print(results[result][i]) print() print (‘ — — -’) print()

The results are below:

Suppor t Vector Machine Result


Random Forest Result

Ada Boost Results

Gradient Boosting Result

From the above results, the best model is Gradient Boosting. So, I will save this model to use it for web
applications.

– Save the Prediction Model

Now, we will pickle the model so that it can be saved on disk.


# Save the model as serialized object pickle with open(‘model.pkl’, ‘wb’) as file: pickle.dump(gb, file)

Model Deployment

It is time to start deploying and building the web application using Flask web application framework. For
the web app, we have to create:

1. Web app python code (API) to load the model, get user input from the HTML template, make the
prediction, and return the result.

2. An HTML template for the front end to allow the user to input heart disease symptoms of the patient and
display if the patient has heart disease or not.

The structure of the files is like the following:

/ ├── model.pkl ├── heart_disease_app.py ├── templates/ └── Heart Disease Classifier.html

Web App Python Code

You can find the full code of the web app here.

As a first step, we have to import the necessary libraries.

import numpy as np import pickle from flask import Flask, request, render_template

Then, we create app object.

# Create application app = Flask(__name__)

After that, we need to load the saved model model.pkl in the app.

# Load machine learning model model = pickle.load(open(‘model.pkl’, ‘rb’))

After that home() function is called when the root endpoint ‘/’ is hit. The function redirects to the home
page Heart Disease Classifier.html of the website.

# Bind home function to URL @app.route(‘/’) def home(): return render_template(‘Heart Disease

Classifier.html’)

Now, create predict() function for the endpoint ‘/predict’. The function is defined as this endpoint
with POST method. When the user submits the form, the API receives a POST request, the API extracts all
data from the form using flask.request.form function. Then, the API uses the model to predict the
result. Finally, the function renders the Heart Disease Classifier.html template and returns the
result.

# Bind predict function to URL @app.route(‘/predict’, methods =[‘POST’]) def predict(): # Put all form
entries values in a list features = [float(i) for i in request.form.values()] # Convert features to array
array_features = [np.array(features)] # Predict features prediction = model.predict(array_features) output =
prediction # Check the output values and retrieve the result with html tag based on the value if output == 1:
return render_template(‘Heart Disease Classifier.html’, result = ‘The patient is not likely to have heart
disease!’) else: return render_template(‘Heart Disease Classifier.html’, result = ‘The patient is likely to
have heart disease!’)

Finally, start the flask server and run our web page locally on the computer by calling app.run() and then
enter http://localhost:5000 on the browser.

if __name__ == ‘__main__’: #Run the application app.run()

HTML Template

The following figure presents the HTML form. You can find the code here.

The form has 13 inputs for the 13 features and a button. The button sends POST request to the/predict
endpoint with the input data. In the form tag, the action attribute calls predict function when the form
is submitted.

<form action = “{{url_for(‘predict’)}}” method =”POST” >


Finally, the HTML page presents the stored result in the result parameter.

<strong style="color:red">{{result}}</strong>

Summary

In this article, you learned how to create a web application for prediction from scratch. Firstly, we started
with the problem definition and data collection. Then, we worked on data preparation, data exploration,
data modeling, and model evaluation. Finally, we deployed the model using a flask.

Now, it is time to practice and apply what you learn in this ar ticle. Define a problem, search for a dataset
on the Internet, and then go through the other steps of the data science lifecycle.

About the Author

Nada Alay

I am working in the data analytics field and passionate about data science, machine learning, and scientific
research. Photograph: I use the attached logo as a personal image on the Internet.

Article Url - https://www.analyticsvidhya.com/blog/2020/09/web-application/

Guest Blog

You might also like