Mining Educational Data To Predict Student's Academic Performance Using Ensemble Methods
Mining Educational Data To Predict Student's Academic Performance Using Ensemble Methods
net/publication/307968552
CITATIONS READS
344 14,114
3 authors:
Ibrahim Aljarah
University of Jordan
168 PUBLICATIONS 16,342 CITATIONS
SEE PROFILE
All content following this page was uploaded by Ibrahim Aljarah on 10 September 2016.
L.
Abstract
Educational data mining has received considerable attention in the last few years.
A
Many data mining techniques are proposed to extract the hidden knowledge from
EG
educational data. The extracted knowledge helps the institutions to improve their teaching
s I ly.
methods and learning process. All these improvements lead to enhance the performance
of the students and the overall educational outputs. In this paper, we propose a new
ei n
LL
student’s performance prediction model based on data mining techniques with new data
fil O
attributes/features, which are called student’s behavioral features. These type of features
are related to the learner’s interactivity with the e-learning management system. The
is on
performance of student’s predictive model is evaluated by set of classifiers, namely;
Artificial Neural Network, Naïve Bayesian and Decision tree. In addition, we applied
th si
Boosting and Random Forest (RF), which are the common ensemble methods used in the
literature. The obtained results reveal that there is a strong relationship between
eb eV
learner’s behaviors and their academic achievement. The accuracy of the proposed model
using behavioral features achieved up to 22.1% improvement comparing to the results
ad in
when removing such features and it achieved up to 25.8% accuracy improvement using
ensemble methods. By testing the model using newcomer students, the achieved accuracy
m Onl
is more than 80%. This result proves the reliability of the proposed model.
1. Introduction
Recently there is an increasing research interest in educational data mining (EDM).
ok
EDM is an emerging field that uses data-mining (DM) techniques to analyze and extract
the hidden knowledge from educational data context [1]. EDM includes different groups
Bo
of users, these users utilize the knowledge discovered by EDM according to their own
vision and objectives of using DM [2]. For example, the hidden knowledge can help the
educators to improve teaching techniques, to understand learners, to improve learning
process and it could be used by learner to improve their learning activities [3]. It also
helps the administrator taking the right decisions to produce high quality outcomes [4].
The educational data can be collected from different sources such as web-based
education, educational repositories and traditional surveys. EDM can use different DM
techniques, each technique can be used for certain educational problem. As Example, to
predict an educational model the most popular technique is classification. There are
several algorithms under classification such as Decision tree, Neural Networks and
Bayesian networks [5].
This paper introduces a students’ performance model with a new category of features,
which called behavioral features. The educational dataset is collected from learning
management system (LMS) called Kalboard 360 [6]. This model used some data mining
L.
performance of student’s prediction model are Bagging, Boosting and Random Forest
(RF). The remainder of this paper is organized as follows: Section 2 presents the related
A
work in the area of educational data mining algorithms. In Section 3, presents the data
EG
collection and preprocessing. In Section 4 our methodology in predicting students’
s I ly.
performance. The experimental evaluation and results are shown in Section 5, and Section
6 presents our conclusions.
ei n
LL
fil O
2. Related Work
is on
Predicting student’s performance is an important task in web-based educational
environments. To build a predictive model, there are several DM techniques used, which
th si
are classification, regression and clustering. The most popular technique to predict
y er
Romero et al in [29] used DT algorithm to predict students’ final marks based on their
m Onl
usage data in the Moodle system. Moodle is one of the frequently used Learning Content
Management Systems (LCMS). The author has collected real data from seven Moodle
courses with Cordoba University to classify students into two groups: passed and fail. The
objective of this research is to classify students with equal final marks into different
groups based on the activities carried out in a web-based course.
Neural network is another popular technique that has been used in educational data
mining. A neural network is s a biological inspired intelligent technique that consists of
ok
connected elements called neurons that work together to produce an output function [30].
Arsad et al. in [31] used ANN model to predict the academic performance of bachelor
Bo
degree engineering students. The study takes Grade Point (GP) of fundamental subjects
scored by the students as inputs without considering their demographic background, while
it takes Cumulative Grade Point Average (CGPA) as output. Neural Network (NN) trains
engineering Degree students GP to get the targeted output. This research showed that
fundamental subjects have a strong influence in the final CGPA upon graduation.
The authors in [32] used Bayesian networks to predict the CGPA based on applicant
background at the time of admission. Nowadays, educational institutions need a method
to evaluate the qualified applicants graduating from various institutions. This research
presents a novel approach that integrate a case-based component with the prediction
model. The case-based component retrieves the past student most similar to the applicant
being evaluated. The challenge is to define similarity of cases (applicants) in a way that is
consistent with the prediction model. This technique can be applied at any institution that
has a good database of student and applicant information.
L.
learners, monitor student participation, keeping track of their progress across the system
[8]. The LMS allocates and manages learning resources such as registration, classroom
A
and the online learning delivery. In this paper, the educational data set is collected from
EG
learning management system (LMS) called Kalboard 360 Kalboard [6]. Kalboard 360 is a
s I ly.
multi-agent LMS, which has been designed to facilitate learning through the use of
leading-edge technology. Such system provides users with a synchronous access to
ei n
LL
educational resources from any device with Internet connection. In addition to involve
fil O
parents and school management in the learning experience. This makes it a truly extensive
process, which connects and properly engages all parties. The data is collected using a
is on
learner activity tracker tool, which called experience API (xAPI) [9]. The xAPI is a
component of the Training and Learning Architecture (TLA) that enables to monitor
th si
learning progress and learner’s actions like reading an article or watching a training video.
y er
The Experience API helps the learning activity providers to determine the learner, activity
and objects that describe a learning experience.
eb eV
The goal of X-API in this research is to monitor student behavior through the
educational process for evaluating the features that may have an impact on student’s
academic performance. The educational data set that used in the previous work [10]
ad in
contains only 150 student’s records with 11 features. In the current paper that data set
m Onl
extends into 500 students with 16 features. The features are classified into three main
categories: (1) Demographic features such as gender and nationality. (2) Academic
background features such as educational Stage, grade Level and section. (3) Behavioral
features, such as raised hand on class, visited resources, parent Answering Survey and
Parent School Satisfaction. This feature cover learner and parent progress on LMS. Table
1 shows the dataset’s attributes/features and their description. Table 1 was used in the
previous research [10], by reviewing the table we can notice a new feature category which
ok
is a behavioral feature. These features present the learner and the parent participation in
the learning process.
Bo
L.
Academic Educational Stages Stage student belongs such as
A
Background Features (school levels) (primary, middle and high school
levels)
EG
s I ly.
Grade Levels Grade student belongs as (G-01, G-
ei n
LL
fil O 02, G-03, G-04, G-05, G-06, G-07,
G-08, G-09, G-10, G-11, G-12)
is on
C)
y er
eb eV
Viewing
announcements
After the data collection task, we apply some pre-processing mechanisms to improve
the quality of the data set. Data pre-processing is considered an important step in the
knowledge discovery process, which includes data cleaning, feature selection, data
reduction and data transformation.
L.
prove that female students are more satisfied than male students with e-learning systems.
Other researches address that male students have a positive perception of e-learning
A
compared to female students [14]. For the family background feature, different studies
have shown that there is a positive relationship between the parent’s education and
EG
s I ly.
student’s performance [16]. This relation is particularly valid when the learner is being
followed up by their mother. The authors in [17] observed that mothers have a more
ei n
LL
influence on their children academic achievements. Third school attendance feature,
fil O
school attendance is an important feature in educational success [18]. Previous research
[19] has shown a direct relation between good attendance and student achievement. These
is on
researches prove the positive relation between such features: gender, family background
and school attendance students’ performance. This research will shed a light on new
th si
category of features, called behavioral features. This feature related to the learner
y er
engagement with educational system. Student engagement is one of the main researches
in educational psychology field. Student engagement was defined by Gunuc and Kuzu
eb eV
[20] as “the quality and quantity of students’ psychological, cognitive, emotional and
behavioral reactions to the learning process as well as to in-class/out-of-class academic
and social activities to achieve successful learning outcomes”. Kuk [21] refers to the
ad in
student’s engagement by the spent time in classroom. According to Stovall [22], student
m Onl
engagement includes not only the spent time on tasks but also their desire to participate in
some activities. There are various researches that light on student’s engagement and
behavior. All of these researches confirm the positive relationship between students’
behavior and student’s academic achievement.
the step before applying data mining algorithm, it transforms the original data into a
suitable shape to be used by a particular mining algorithm. Data preprocessing includes
Bo
different tasks as data cleaning, feature selection and data transformation [23].
Gender
350 305
300
250
200 175
150
100
50
0
L.
Males Females
A
Gender
EG
s I ly.
Figure 1. Gender Feature Visualization
ei n
LL
As shown in Figure2, students come from different origins such as 179 students are
fil O
from Kuwait, 172 students are from Jordan, 28 students from Palestine, 22 students are
from Iraq, 17 students from Lebanon, 12 students from Tunis, 11 students from Saudi
is on
Arabia, 9 students from Egypt, 7 students from Syria, 6 students from USA, Iran and
Libya, 4 students from Morocco and one student from Venezuela.
th si
y er
Nationality
eb eV
160
140
m Onl
120
100
80
60
40 28 22 17 12 11 9
20 7 6 6 6 4 1
0
ok
Bo
Nationality
Educational Stages
Low
Middle
High
L.
0 50 100 150 200 250 300
A
Educational stages
EG
s I ly.
Figure 3. Educational Stages Visualization
ei n
LL
The student’s data collected through two educational semesters: First and second, in
fil O
which 245 students record collected during the first semester and 235 student’s record
collected during the second semester. Students through these different semesters take
is on
different topics as shown in Figure4, There are 95 students take IT topic, 65 students take
French topic, 59 students take Arabic topic, 51 students take science topic, 45 students
th si
take English topic, 30 students take Biology, 25 students take Spanish, 24 students take
y er
both chemistry and Geology topics, 22 students take Quran topic, 21 students take math
topic, 19 students take History topic. Each student in the data set is followed up by a
eb eV
different parent as follow: 283 students are followed by their fathers and 197 students are
followed by their moms.
ad in
m Onl
Educational Topics
100 95
90
80
70 65
59
60 51
50 45
ok
40 30
30 25 24 24 22 21 19
Bo
20
10
0
Educational topics
The data set includes also the school attendance feature, as shown in Figure5, the
students are visualized into two categories based on their absence days: 191 students
exceed 7 absence days and 289 students their absence days under 7.
L.
A
Figure 5. Students’ Absence Days’ Feature Visualization
EG
s I ly.
This research uses the “student absence days” feature to show the influence of such
ei n
LL
feature on student’s performance. This research also utilizes new category of features; this
fil O
feature is parent parturition in the educational process. Parent parturition feature have two
sub features: Parent Answering Survey and Parent School Satisfaction. There are 270 of
is on
the parents answered survey and 210 are not, 292 of the parents are satisfied from the
school and 188 are not. Data preprocessing used in this research to study the nature of
th si
students’ performance features, and to get the influence ratio of features by defining the
y er
percentage value of each feature. The influence ratio of features will be defined accurately
using feature selection process.
eb eV
Data cleaning is one of the main preprocessing tasks, is applied on this data set to
remove irrelevant items and missing values. The data set contains 20 missing values in
m Onl
various features from 500 records, the records with missing values are removed from the
data set, and the data set after cleaning becomes 480 records.
redundant and irrelevant data [24]. This process can play an important role in improving
the data quality therefore the performance of the learning algorithm. Feature selection
Bo
methods are categorized into wrapper-based and filter-based methods. Filter method is
searching for the minimum set of relevant features while ignoring the rest. It uses variable
ranking techniques to rank the features where the highly ranked features are selected and
applied to the learning algorithm. Different feature ranking techniques have been
proposed for feature evaluations such as information gain and gain ratio.
In this research, we applied filter-method using information gain based selection
algorithm to evaluate the feature ranks, checking which features are most important to
build students’ performance model. Figure6, shows the feature ranks after filter-based
evaluation. During feature selection, each feature assigned a rank value according to their
influence on data classification. The highly ranked features are selected while others are
excluded.
L.
A
EG
s I ly.
Figure 6. Filter-Based Feature Selection Evaluation
ei n
LL
fil O
As shown in Figure6, visited resources feature got the higher rank, then followed by
student absence days, raised the hand on classroom, parent answering survey, nationality,
is on
parent responsible for student, place of birth, discussion groups and parent school
satisfaction features. As we can see the appropriate subset of features consist of ten
th si
features while other ones are excluded. In summary, the features that are related to student
y er
and parent progress during the usage of LMS got the highest ranks, which means the
learner behavior during the educational process have an impact on their academic success.
eb eV
4. Methodology
ad in
problem. In contrast to traditional learning approaches which train data by one learning
model, ensemble methods try to train data using a set of models, then combine them to
take a vote on their results. The predictions made by ensembles are usually more accurate
than predictions made by a single model. The aim of such approach is to provide an
accurate evaluation for the features that may have an impact on student’s academic
success. Figure 7 shows the main steps in the proposed methodology.
ok
Bo
L.
A
EG
s I ly.
ei n
LL
fil O
is on
th si
y er
eb eV
ad in
m Onl
ok
Bo
This methodology starts by collecting data from Kalboard 360 (LMS) system using
experience API (xAPI) as mentioned in Section 3. This step is followed by data
preprocessing step, which concerns with transforming the collected data into a suitable
format. After that, we use discretization mechanism to transform the students’
performance from numerical values into nominal values, which represents the class labels
of the classification problem. To accomplish this step, we divide the data set into three
nominal intervals (High Level, Medium Level and Low Level) based on student’s total
grade/mark such as: Low Level interval includes values from 0 to 69, Middle Level
interval includes values from 70 to 89 and High Level interval includes values from 90-
100. The data set after discretization consists of 127 students with Low Level, 211
students with Middle Level and 142 students with High Level. Then, we use
normalization to scale the attributes values into a small range [0.0 to 1.0]. This process
L.
can speed up the learning process by preventing attributes with large ranges from
outweighing attributes with smaller ranges. After that, feature selection process is applied
A
to choose the best feature set with higher ranks. As shown in Figure7, we applied filter-
EG
based technique for feature selection.
s I ly.
In this paper, ensemble methods are applied to provide an accurate evaluation for the
features that may have an impact on the performance/grade level of the students, and to
ei n
LL
improve the performance of student’s prediction model. Ensemble methods are
fil O
categorized into dependent and independent methods. In a dependent method, the output
of a learner is used in the creation of the next learner. Boosting is an example of
is on
dependent methods. In an independent method, each learner performs independently and
their outputs are combined through a voting process. Bagging and random forest are
th si
example of independent methods. These methods resample the original data into samples
y er
of data, then each sample will be trained by a different classifier. The classifiers used in
student’s prediction model are Decision Trees (DT), Neural Networks (NN) and Naïve
eb eV
Bayesian (NB). Individual classifiers results are then combined through a voting process,
the class chosen by most number of classifiers is the ensemble decision.
ad in
Boosting belongs to a family of algorithms that are capable of converting weak learners
to strong learners. The general boosting procedure is simple, it trains a set of learners
m Onl
sequentially and combine them for prediction, then focus more on the errors of the
previous learner by editing the weights of the weak learner. A specific limitation of
boosting that is used only to solve binary classification problems. This limitation is
eliminated with the AdaBoost algorithm. AdaBoost is an example of boosting algorithm,
which stands for adaptive boost. The main idea behind this algorithm is to pay more
attention to patterns that are hard to classify. The amount of attention is measured by a
weight that is assigned to every subset in the training set. All the subsets are assigned
ok
equal weights. In each iteration, the weights of misclassified instances are increased while
the weights of truly classified instances are decreased. Then the AdaBoost ensemble
Bo
combines the learners to generate a strong learner from weaker classifiers through a
voting process [33].
Bagging is an independent ensemble based methods. The aim of this method is to
increase the accuracy of unstable classifiers by creating a composite classifier, then
combine the outputs of the learned classifiers into a single prediction. The Bagging
algorithm is summarized in Figure8, it starts with resampling the original data into
different training data sets (D1-Dn) which called bootstraps, each bootstrap sample size is
equal to the size of the original training set. All bootstrap samples will be trained using
different classifiers (C1-Cm). Individual classifiers results are then combined through
majority vote process, the class chosen was by the most number of classifiers is the
ensemble decision [33].
In boosting, as contrary to bagging, each classifier is influenced by the performance of
the previous classifier. In bagging, each sample of data is chosen with equal probability,
while in boosting, instances are chosen with a probability that is proportional to their
weight. Furthermore, bagging works best with high variance models which produce
variance generalization behavior with small changes to the training data. Decision trees
and neural networks are examples of high variance models.
L.
A
EG
s I ly.
ei n
LL
fil O
is on
th si
y er
eb eV
Random Forest (RF) is a special modification of bagging where the main difference
ad in
with bagging is the integration of randomized feature selection. Through the decision tree
construction process, RF uses random decision trees to select a random subset of features.
m Onl
Notice that randomness is only performed on the feature selection process, but the choice
of a split point on the selected features is performed by bagging. The combination
between decision tree and bootstrapping makes RF strong enough to overcome the
overfitting problem, and to reduce the correlation between trees which provides an
accurate prediction [33].
All the above classification methods are trained using 10-folds cross validation. This
technique divides the data set into 10 subsets of equal size, nine of the subsets are used for
ok
training, while one is left out and used for testing. The process is iterated for ten times, the
final result is estimated as the average error rate on test examples. Once the classification
Bo
model has been trained, the validation process starts. Validation process is the last phase
to build a predictive model, it used to evaluate the performance of the prediction model by
running the model over real data.
Positive Negative
Positive True positive (TP) False Negative(FN)
L.
Actual
A
Negative False Positive (FP) True Negative (TN)
EG
s I ly.
Accuracy is the proportion of the total number of predictions where correctly
ei n
calculated.
LL
fil O
Precision is the ratio of the correctly classified cases to the total number of
misclassified cases and correctly classified cases. Recall is the ratio of correctly classified
is on
cases to the total number of unclassified cases and correctly classified cases. In addition,
we used the F-measure to combine the recall and precision which is considered a good
th si
(1)
eb eV
(2)
ad in
(3)
m Onl
(4)
There are many features directly or indirectly affecting the effectiveness of student
performance model. In this section, we will evaluate the impact of behavioral features on
Bo
As shown in Table 3, we can notice that the ANN model outperforms other data
mining techniques. ANN model achieved 79.1 accuracy with BF and 57.0 without
L.
behavioral features. The 79.1 accuracy means that 380 of 480 students are correctly
A
classified to the right class labels (High, Medium and Low) and 100 students are
incorrectly classified.
EG
s I ly.
For the recall measure, the results are 79.2 with BF and 57.1 without behavioral
features. The 79.2 recall means that 380 students are correctly classified to the total
ei n
number of unclassified and correctly classified cases.
LL
fil O
For the precision measure, the results are 79.1 with BF and 57.2 without behavioral
features. The 79.1 precision means 380 of 480 students are correctly classified and 100
is on
students are misclassified.
For the F-Measure, the results are 79.1 with BF and 57.1 without behavioral features.
The experimental results prove the strong effect of learner behavior on student’s academic
th si
achievement. We can get more accurate results by training the data set with ensemble
y er
methods.
eb eV
traditional DM methods. Table 3, presents the results of the traditional classifiers and the
m Onl
results of traditional classifiers using ensemble methods (Bagging, Boosting and RF).
As shown in in the Table 3, we can see good results using ensemble methods with
traditional classifiers (ANN, NB and DT). Each ensemble trains the three classifiers, then
combine the results through a majority voting process to achieve the best prediction
performance of student’s model. Boosting method outperform other ensembles methods,
in which the accuracy of DT using boosting is improved from 75.8 to 77.7, which means
that the number of correctly classified students are increased from 363 to 373 of 480.
Recall results are increased from 75.8 to 77.7, which means that 373 students are correctly
ok
classified to the total number of unclassified and correctly classified cases. Precision
results are also increased from 76.0 to 77.8, which means 373 of 480 students are
Bo
L.
the trained model. Table 5, shows the evaluation results using several classification
methods (ANN, NB and DT) through testing process and validation process.
A
EG
Table 5. Classification Methods Results through Testing and Validation
s I ly.
Evaluation Testing results Validation results
ei n
LL
Measure fil O
Classifiers type DT ANN NB DT ANN NB
is on
Accuracy 75.8 79.1 67.7 82.2 80.0 80.0
Recall 75.8 79.2 67.7 82.2 80.0 80.0
th si
As shown in Table 5, we can notice that the evaluation measure results increased for
ad in
the three prediction models through validation process. The three prediction models
achieved accuracy more than 80%, which means that 20 of 25 new students are correctly
m Onl
classified to the right class labels (High, Medium and Low) and 5 students are incorrectly
classified. The results of the validation process prove the reliability of the proposed
model.
6. Conclusion
Academic achievement is being a big concern for academic institutions all over the
ok
world. The wide use of LMS generates large amounts of data about teaching and learning
interactions. This data contains hidden knowledge that could be used to enhance the
academic achievement of students. In this paper, we propose a new student’s performance
Bo
prediction model based on data mining techniques with new data attributes/features,
which called student’s behavioral features. These type of features are related to the learner
interactivity with learning management system. The performance of student’s predictive
model is evaluated by set of classifiers, namely; Artificial Neural Network, Naïve
Bayesian and Decision tree. In addition, we applied ensemble methods to improve the
performance of these classifiers. We used Bagging, Boosting and Random Forest (RF),
which are the common ensemble methods that used in the literature. The obtained results
reveal that there is a strong relationship between learner’s behaviors and their academic
achievement. The accuracy of student’s predictive model using behavioral features
achieved up to 22.1% improvement comparing to the results when removing such
features, and it achieved up to 25.8% accuracy improvement using ensemble methods.
The visited resources feature is the most effective behavioral feature on students’
performance model. In our future work, we will focus more on analyzing this kind of
feature. After completing the training process, the predictive model is tested using
unlabeled newcomer students, the achieved accuracy is more than 80%. This result proves
how realistic the predictive model is. Lastly, this model can help educators to understand
learners, identify weak learners, to improve learning process and trimming down
academic failure rates. It also can help the administrators to improve the learning system
outcomes.
References
[1] C. Romero and S. Ventura, “Educational data mining: A survey from 1995 to 2005”, Expert systems
with applications, vol. 33, no. 1, (2007), pp. 135-146.
[2] M. Hanna, “Data mining in the e-learning domain”, Campus-wide information systems, vol. 21, no. 1,
(2004), pp. 29-34.
L.
[3] C. Romero and S. Ventura, “Educational data mining: a review of the state of the art. Systems, Man, and
Cybernetics”, Part C: Applications and Reviews, IEEE Transactions on, vol. 40, no. 6, (2010), pp. 601-
A
618.
[4] M. E. Zorrilla, E. Menasalvas, D. Marin, E. Mora and J. Segovia, “Web usage mining project for
EG
s I ly.
improving web-based learning sites”, In Computer Aided Systems Theory–EUROCAST 2005, Springer
Berlin Heidelberg, (2005), pp. 205-210.
[5] A. M. Shahiri and W. Husain, “A Review on Predicting Student's Performance Using Data Mining
ei n
LL
Techniques”, Proceeding Computer Science, vol. 72, (2015), pp. 414-422.
fil O
[6] “Kalboard360-E-learning system”, http://kalboard360.com/ (accessed February 28, 2016).
[7] G. Kakasevski, M. Mihajlov, S. Arsenovski and S. Chungurski, “Evaluating usability in learning
management system Moodle”, Information Technology Interfaces, 2008. ITI 2008. 30th International
is on
Conference on IEEE, (2008), pp. 613-618.
[8] S. Rapuano and F. Zoino, “A learning management system including laboratory experiments on
th si
measurement instrumentation”, Instrumentation and Measurement, IEEE Transactions on, vol. 55, no. 5,
(2006), pp. 1757-1766.
y er
[9] V. Moisa, “Adaptive Learning Management System”, Journal of Mobile, Embedded and Distributed
eb eV
[11] S. Putrevu, “Exploring the origins and information processing differences between men and women:
Implications for advertisers”, Academy of marketing science review, vol. 2001, no. 1, (2001).
m Onl
[12] S. S. Meit, N. J. Borges, B. A. Cubic and H. R. Seibel, “Personality differences in incoming male and
female medical students”, Online Submission.
[13] F. G. Gómez, J. Guardiola, O. M. Rodríguez and M. A. M. Alonso, “Gender differences in e-learning
satisfaction”, Computers & Education, vol. 58, no. 1, (2012), pp. 283-290.
[14] C. S. Ong, and J. Y. Lai, “Gender differences in perceptions and relationships among dominants of e-
learning acceptance”, Computers in human behavior, vol. 22, no. 5, (2006), pp. 816-829.
[15] C. Romero, S. Ventura, P. G. Espejo and C. Herv´as, “Data mining algorithms to classify students”, in:
Educational Data Mining, vol. 2008, (2008).
[16] J. Ermisch and M. Francesconi, “Family matter: Impacts of family background on educational
ok
L.
Networks, vol. 6, no. 4, (1993), pp. 525-533.
[31] P. M. Arsad, N. Buniyamin and J. L. A. Manan, “A neural network students' performance prediction
A
model (NNSPPM)”, In Smart Instrumentation, Measurement and Applications (ICSIMA), 2013 IEEE
EG
International Conference on. IEEE, (2013), pp. 1-5.
s I ly.
[32] N. T. N. Hien and P. Haddawy, “A decision support system for evaluating international student
applications”, In Frontiers In Education Conference-Global Engineering: Knowledge Without Borders,
ei n
Opportunities Without Passports, 2007. FIE'07. 37th Annual. IEEE, (2007), pp. F2A-1.
LL
[33] Z. H. Zhou, “Ensemble methods: foundations and algorithms”, CRC Press, (2012).
fil O
is on
th si
y er
eb eV
ad in
m Onl
ok
Bo
L.
A
EG
s I ly.
ei n
LL
fil O
is on
th si
y er
eb eV
ad in
m Onl
ok
Bo