Student Performance Analysis Using Educa
Student Performance Analysis Using Educa
14 ICETCSE 2016 Special Issue International Journal of Computer Science and Information Security (IJCSIS)
ISSN 1947-5500 [https://sites.google.com/site/ijcsis/]
69
Abstract— Software industry is hiring the students from Bayes, K- Nearest neighbor, and many others. Using these
the engineering colleges who are good in communication, techniques many kinds of knowledge can be discovered such as
programming, and also academically performing well. Most of the association rules, classifications and clustering. The discovered
engineering institutions focused on the students performance on knowledge can be used for prediction regarding enrolment of
the above stated factors. The engineering students have to students in a particular course, alienation of traditional
improve their academic performance, programming skills and classroom teaching model, detection of unfair means used in
also communication skills. To help such kind of students, we online examination, detection of abnormal values in the result
designed a project which can predict the students performance sheets of the students, prediction about students‟ performance
before the announcement of their results and before they attend and so on.
their semester exams. By this the students can know their The main aim of this project is to improvise the student performance
performance and can improve their skills by proper planning or in studies based on some important factors. Education is an essential
by making changes in their plans. This can help the students element for the betterment and progress of a country. It enables the
improve in their academics, which eventually leads to a good people of a country civilized and well mannered. Now-a-days
performance in their end examinations. By this the suicide rates developing new methods to discover knowledge from educational
of students will also get reduced since the stress is reduced. This database in order to analyse student's trends and behaviours towards
could help in our country development by providing good and education. To analyse the data from different dimensions categorize
efficient engineers to the country. it and to summarize the relationships. It motivated us to work on
We applying Naive Bayes classification algorithm and student dataset analysation.The data collection, categorization and
Weighted Naïve Bayesian algorithm on the student data set which classification is being performed manually. The main disadvantage
is collected from LBRCE IT department, Mylavaram for building of this process is delay in results, remedial measures are not taken
this model. Based on these results we can classify the weak students properly due to late analysis of student performance. There will be
and take the remedial measures to improve their performance. delay in the results announcements which leads to the poor performance
Keywords: Educational Data Mining, Classification, Prediction. of the students in the next examination due to lack of planning in their
I. INTRODUCTION preparation. When count of students increases, the analysis of
performance of a student becomes difficult. To overcome this difficulty
The advent of information technology in various fields has we now introduce you to educational data mining. When institutes store
lead the large volumes of data storage in various formats like their students details in cloud, it will be difficult to analyse large data
records, files, documents, images, sound, videos, scientific data often called as big data. By applying data mining on the data stored,
and many new data formats. The data collected from different we can easily categories and analyse the results of a student in short
applications require proper method of extracting knowledge time without any difficulties. Here, mainly concentrated on the students
from large repositories for better decision making. Knowledge internal marks, ability to concentrate, attendance, awareness on course
outcomes, tutorials, semester marks, content perception, assignments
discovery in databases (KDD), often called data mining, aims
at the discovery of useful information from large collections of II. DATA MINING DEFINITION AND TECHNIQUES
data [1]. The main functions of data mining are applying Data mining, also popularly known as Knowledge
various methods and algorithms in order to discover and extract Discovery in Database refers to extracting or “mining"
patterns of stored data [2]. Data mining and knowledge knowledge from large amounts of data. Data mining techniques
discovery applications have got a rich focus due to its are used to operate on large volumes of data to discover hidden
significance in decision making and it has become an essential patterns and relationships helpful in decision making. While
component in various organizations. Data mining techniques data mining and knowledge discovery in database are
have been introduced into new fields of Statistics, Databases, frequently treated as synonyms, data mining is actually part of
Machine Learning, Pattern Reorganization, Artificial the knowledge discovery process. The sequences of steps
Intelligence and Computation capabilities etc. identified in extracting knowledge from data are shown in
There are increasing research interests in using data mining Figure 1.
in education. This new emerging field, called Educational Data
Mining, concerns with developing methods that discover
knowledge from data originating from educational
environments [3]. Educational Data Mining uses many
techniques such as Decision Trees, Neural Networks, Naïve
Proceedings of 3rd International Conference on Emerging Technologies in Computer Science & Engineering (ICETCSE 2016)
V. R. Siddhartha Engineering College, Vijayawada, India, October 17-18, 2016
Vol. 14 ICETCSE 2016 Special Issue International Journal of Computer Science and Information Security (IJCSIS)
ISSN 1947-5500 [https://sites.google.com/site/ijcsis/]
70
C. Predication
Regression technique can be adapted for predication.
Regression analysis can be used to model the relationship
between one or more independent variables and dependent
variables. In data mining independent variables are attributes
already known and response variables are what we want to
predict. Unfortunately, many real-world problems are not
simply prediction. Therefore, more complex techniques (e.g.,
logistic regression, decision trees, or neural nets) may be
necessary to forecast future values. The same model types can
often be used for both regression and classification. For
example, the CART (Classification and Regression Trees)
decision tree algorithm can be used to build both classification
trees (to classify categorical response variables) and regression
trees (to forecast continuous response variables). Neural
networks too can create both classification and regression
models.
FIG 1:KDD PROCESS
D. Association rule
Various algorithms and techniques like Classification , Association and correlation is usually to find frequent item findings
Clustering , Regression , Artificial Intelligence , Neural among large data sets. This type of finding helps businesses to
Networks , Association rules , Decision trees , Genetic make certain decisions, such as catalogue design, marketing and
Algorithm, Nearest Neighbor method etc., are used for customer shopping behavior analysis. Association Rule
knowledge discovery from databases. These techniques and algorithms need to be able to generate rules confidence values less
methods in data mining need brief mention to have better than one. However the number of Association Rules for a given
understanding. dataset is generally very large and a high proportion of the rules are
A. Classification usually of little (if
any) value.
The Classification is the one of the most important
technique used in data mining. It is a 2 step process 1.first build E. Neural networks
classification model. 2. Predict the class label, which employs
Neural network is a set of connected input/output units and
a set of pre-classified examples to develop a model that can
each connection has a weight present with it. During the
classify the population of records at large. This approach
learning phase, network learns by adjusting weights so as to be
regularly employs decision tree or neural network-based able to predict the correct class labels of the input tuples. Neural
classification algorithms. The data classification process
networks have the remarkable ability to derive meaning from
involves learning and classification. In Learning the training complicated or imprecise data and can be used to extract
data are analyzed by classification algorithm. In classification patterns and detect trends that are too complex to be noticed by
test data are used to estimate the accuracy of the classification either humans or other computer techniques. These are well
rules. If the accuracy is acceptable the rules can be applied to suited for continuous valued inputs and outputs. Neural
the new data tuples. The classifier-training algorithm uses these networks are best at identifying patterns or trends in data and
pre-classified examples to determine the set of parameters well suited for prediction or forecasting needs.
required for proper discrimination. The algorithm then encodes
these parameters into a model called a classifier. F. Decision Trees
B. Clustering Decision tree is tree-shaped structures that represent sets of
decisions. These decisions generate rules for the classification
Clustering can be defined as discovery of similar classes of of a dataset. Specific decision tree methods include
objects.. By using clustering techniques we can further identify Classification and Regression Trees (CART) and Chi Square
dense and sparse regions in object space and can discover Automatic Interaction Detection (CHAID).
overall distribution pattern and correlations among data
attributes. Classification approach can also be used for effective G. Nearest Neighbor Method
means of distinguishing groups or classes of object A technique that classifies each record in a dataset based on
but it becomes costly so clustering can be used as a combination of the classes of the k record(s) most similar to
preprocessing approach for attribute subset selection and it in a historical dataset (where k is greater than or equal to 1).
classification. Sometimes called the k-nearest neighbor technique.
Proceedings of 3rd International Conference on Emerging Technologies in Computer Science & Engineering (ICETCSE 2016)
V. R. Siddhartha Engineering College, Vijayawada, India, October 17-18, 2016
Vol. 14 ICETCSE 2016 Special Issue International Journal of Computer Science and Information Security (IJCSIS)
ISSN 1947-5500 [https://sites.google.com/site/ijcsis/]
71
Proceedings of 3rd International Conference on Emerging Technologies in Computer Science & Engineering (ICETCSE 2016)
V. R. Siddhartha Engineering College, Vijayawada, India, October 17-18, 2016
Vol. 14 ICETCSE 2016 Special Issue International Journal of Computer Science and Information Security (IJCSIS)
ISSN 1947-5500 [https://sites.google.com/site/ijcsis/]
72
Content perception:
By knowing about the content perception of a student,
the teacher can help the student in understanding the subject FIG2:BLOCK DIAGRAM
further. We can assess whether the student listens or not by
content perception.
Proceedings of 3rd International Conference on Emerging Technologies in Computer Science & Engineering (ICETCSE 2016)
V. R. Siddhartha Engineering College, Vijayawada, India, October 17-18, 2016
Vol. 14 ICETCSE 2016 Special Issue International Journal of Computer Science and Information Security (IJCSIS)
ISSN 1947-5500 [https://sites.google.com/site/ijcsis/]
73
p(Ci)=n(Ci)/m where i=1,2,m; The below table contains Boolean value attribute weights
from scale 0-1. These weights are added in weighted naïve
Step-3: Posterior probabilities Bayesian algorithm, so that to get more accurate results
p(Ci/X)=[p(X/Ci).p(Ci)]/p(X). than that of naïve Bayesian classifier. The Boolean
valued attributes are nothing but having binary values
Step-4: Calculating like yes or no, true or false.
p(X/Ci)=nk=1p(Xk/Ci).
Boolean value Attribute weights from 0-1 scale
Step-5: In order to predict the class label of S.No Awareness Assignments Tutorials
X,p(X/Ci)p(Ci) is evaluated for each class Ci. of CO’s
Yes No Yes No Yes No
p(X/Ci).p(Ci)>p(X/Cj).p(Cj) for 1jm, ji. Professor 0.18 0.0 0.22 0.0 0.22 0.0
1
Professor 0.20 0.0 0.18 0.0 0.18 0.0
2
Professor 0.22 0.0 0.20 0.0 0.20 0.0
3
Average 0.20 0.0 0.20 0.0 0.20 0.0
Proceedings of 3rd International Conference on Emerging Technologies in Computer Science & Engineering (ICETCSE 2016)
V. R. Siddhartha Engineering College, Vijayawada, India, October 17-18, 2016
Vol. 14 ICETCSE 2016 Special Issue International Journal of Computer Science and Information Security (IJCSIS)
ISSN 1947-5500 [https://sites.google.com/site/ijcsis/]
75
Several standard terms have been defined for the 2 class matrix:
The accuracy (AC) is the proportion of the total number of
Multi value Attribute weights from 0-1 scale predictions that were correct.
S.N Basics Ability to Content It is determined using the equation
o Concentrate Perception AC= (a+d)/(a+b+c+d).
in the Class The recall or true positive rate (TP) is the proportion of positive
S A W S A W S A W cases that were correctly
vg vg vg It is determined using the equation
Prof 0. 0. 0. 0. 0. 0. 0. 0. 0. TP=d/(c+d),
esso 50 18 0 60 32 0 75 50 0 The false positive rate (FP) is the proportion of negatives cases
r1 that were incorrectly classified as positive as calculated using the
Prof 0. 0. 0. 0. 0. 0. 0. 0. 0. formula
esso 45 22 0 55 28 0 85 48 0 FP= b/(a+b).
r2 The true negative rate (TN) is defined as the proportion of
Prof 0. 0. 0. 0. 0. 0. 0. 0. 0. negatives cases that were classified correctly as calculated using
esso 65 20 0 65 30 0 80 52 0 the equation.
r3 TN= a/(a+b).
Ave 0. 0. 0. 0. 0. 0. 0. 0. 0. The false negative rate (FN) is the proportion of positives cases
rage 50 20 0 60 30 0 80 50 0 that were incorrectly classified as negative as calculated using the
S:Strong Avg:Average W:Weak equation.
TABLE 4: MULTI VALUE ATTRIBUTE WEIGHTS FN= c/(c+d).
Finally, precision (P) is the proportion of the predicted positive
Multi Value Attribute Weights cases that were correct, as calculated using the equation.
S.No A B C D P=d/ (b+d).
Professor 1 0.90 0.72 0.50 0.0
Professor 2 0.88 0.68 0.8 0.0
Professor 3 0.92 0.70 0.52 0.0
Average 0.90 0.70 0.50 0.0
TABLE 5: MULTI VALUE ATTRIBUTE WEIGHTS
PREDICTED
Negative Positive
Actual NEGATIVE A B
POSITIVE C D
The entries in the confusion matrix have the following Fig 3:Data set of 28 Students
meaning in the context of our study:
a. is the number of correct predictions that an instance is
negative,
b. is the number of incorrect predictions that an instance is
positive,
c. is the number of incorrect of predictions that an instance
negative, and
d. is the number of correct predictions that an instance is positive
[2] U. Fayadd, Piatesky, G. Shapiro, and P. Smyth, From data mining to [14] J. R. Quinlan, “Introduction of decision tree: Machine learn”, 1: pp. 86-
106, 1986.
knowledge discovery in databases, AAAI Press / The MIT Press,
Massachusetts Institute Of Technology. ISBN 0–262 56097–6,1996. [15] Vashishta, S. (2011). Efficient Retrieval of Text for Biomedical Domain
[3] J. Han and M. Kamber, “Data Mining: Concepts and Techniques,” using Data Mining Algorithm. IJACSA - International Journal of
Morgan Kaufmann, 2000. Advanced Computer Science and Applications, 2(4), 77-80.
Alaa el-Halees, “Mining students data to analyze e-Learning behavior: A [16] Kumar, V. (2011). An Empirical Study of the Applications of Data
Case Study”, 2009.. Mining Techniques in Higher Education. IJACSA - International Journal
of Advanced Computer Science and Applications, 2(3), 80-84. Retrieved
[4] U . K. Pandey, and S. Pal, “Data Mining: A prediction of performer or from http://ijacsa.thesai.org.
underperformer using classification”, (IJCSIT) International Journal of
Computer Science and Information Technology, Vol. 2(2), pp.686-690,
ISSN:0975-9646, 2011.
[5] Press, Massachusetts Institute Of Technology. ISBN 0–262 56097–
6,1996.
J. Han and M. Kamber, “Data Mining: Concepts and Techniques,”
Morgan Kaufmann, 2000.
[6] Alaa el-Halees, “Mining students data to analyze e-Learning behavior: A
Proceedings of 3rd International Conference on Emerging Technologies in Computer Science & Engineering (ICETCSE 2016)
V. R. Siddhartha Engineering College, Vijayawada, India, October 17-18, 2016