Machine Learning Based Student Grade Prediction A
Machine Learning Based Student Grade Prediction A
net/publication/319350236
CITATIONS READS
31 3,066
4 authors, including:
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Junaid Qadir on 30 September 2017.
Zafar Iqbal* , Junaid Qadir** , Adnan Noor Mian* , and Faisal Kamiran*
arXiv:1708.08744v1 [cs.CY] 17 Aug 2017
*
Department of Computer Science,
**
Department of Electrical Engineering,
Information Technology University,
Lahore, Pakistan
{mscs13039, junaid.qadir, adnan.noor, faisal.kamiran}@itu.edu.pk
In higher educational institutes, many students have to struggle hard to complete different courses since
there is no dedicated support offered to students who need special attention in the registered courses.
Machine learning techniques can be utilized for students’ grades prediction in different courses. Such
techniques would help students to improve their performance based on predicted grades and would enable
instructors to identify such individuals who might need assistance in the courses. In this paper, we use
Collaborative Filtering (CF), Matrix Factorization (MF), and Restricted Boltzmann Machines (RBM)
techniques to systematically analyze a real-world data collected from Information Technology University
(ITU), Lahore, Pakistan. We evaluate the academic performance of ITU students who got admission in
the bachelor’s degree program in ITU’s Electrical Engineering department. The RBM technique is found
to be better than the other techniques used in predicting the students’ performance in the particular course.
1. I NTRODUCTION
Since universities are prestigious places of higher education, students’ retention in these univer-
sities is a matter of high concern (Aud et al., 2013). It has been found that most of the students’
drop-out from the universities during their first year is due to lack of proper support in under-
graduate courses (Callender and Feldman, 2009) (MacDonald, 1992). Due to this reason, the
first year of the undergraduate student is referred as a “make or break” year. Without getting
any support on the course domain and its complexity, it may demotivate a student and can be the
cause to withdraw the course. There is a great need to develop an appropriate solution to assist
students retention at higher education institutions. Early grade prediction is one of the solutions
that have a tendency to monitor students’ progress in the degree courses at the University and
will lead to improving the students’ learning process based on predicted grades.
Using machine learning with Educational Data Mining (EDM) can improve the learning
process of students. Different models can be developed to predict students’ grades in the enrolled
courses, which provide valuable information to facilitate students’ retention in those courses.
This information can be used to early identify students at-risk based on which a system can
1
suggest the instructors to provide special attention to those students (Iraji et al., 2012). This
information can also help in predicting the students’ grades in different courses to monitor their
performance in a better way that can enhance the students’ retention rate of the universities.
Several research studies have been conducted to assess and predict students’ performance
in the universities. In (Iqbal et al., 2016), we analyzed various existing international studies
and examined the admission criterion of ITU to found which admission criterion factor can
predict the GPA in the first semester at the undergraduate level. From the results, we found that
Higher Secondary School Certificate (HSSC) performance and entry test performance are the
most significant factors in predicting academic success of the students in the first semester at
university. In this study, we are further extending this research and examining the effectiveness
of the performance of students of ITU in enrolled courses using machine learning techniques.
In this study, we applied various techniques (CF, SVD, NMF, and RBM) on the real-world
data of ITU students. The CF techniques are one of the most popular techniques for predicting
students’ performance (Sarwar et al., 1998), which works by discovering similar characteris-
tics of users and items in the database; CF, however, does not provide an accurate prediction
for a sparse database. The SVD technique makes better predictions as compared to CF algo-
rithms for sparse databases by capturing the hidden latent features in the dataset while avoid-
ing overfitting (Berry et al., 1995). The NMF technique allows meaningful interpretations of
the possible hidden features compared to other dimensionality reduction algorithms such as
SVD (Golub and Van Loan, 2012). Finally, RBM can also be used for collaborative filtering and
was used for collaborative filtering during the Netflix competition (Salakhutdinov et al., 2007).
(Toscher and Jahrer, 2010) tried to use RBM on the KDD Cup dataset and got promising results.
The contributions of this paper are:
2. We analyzed a real world data collected from 225 undergraduate students of Electrical
Engineering Department at ITU.
3. We evaluated state of the art machine learning techniques (CF, SVD, NMF, and RBM) in
predicting the performance of ITU students.
4. We proposed a feedback model to calculate the student’s knowledge for particular course
domain and provide feedback if the student needs to put more effort in that course based
on the predicted GPA.
5. We proposed a fitting procedure for hidden Markov model to determine the student per-
formance in a particular course with utilizing the knowledge of course domain.
Rest of the paper is organized as follows. In Section 2, we will describe related work pro-
posed in the literature. Different machine learning techniques that can be utilized to predict
students’ GPA are briefly outlined in section 3. The methodology of the study for this paper
and the performance of the ITU students in different courses are described in Section 4. We
present the results and findings of our study in Section 5. We described the insights that hold
for our study in Section 6. We highlight some limitations of this study in Section 7. Finally, we
conclude the paper in Section 8.
2
2. R ELATED W ORK
Numerous research studies have been conducted to predict students’ academic performance ei-
ther to facilitate degree planning or to determine students at risk.
3
of students’ abilities. Their results can be used as a methodological basis for deriving principle
guidelines for admissions committees.
4
procedure for Bayesian Knowledge Tracing and concluded that empirical probabilities had the
comparable predictive accuracy to that of expectation maximization.
In Table 1, we have systematically summarized the studies that are related to our study in a
comprehensive way to present a big picture of literature. Our work is related to grade prediction
systems, recommender systems, and early warning systems within the context of education.
In our study, the approach is to use machine learning techniques to predict course grades of
students. We used the state of the art techniques that are described and implemented in this
section to do a comparative analysis of different techniques that can predict students’ GPA in
registered courses. We also develop a model that can be used in a tutoring system indicating
the weak students in the course to the instructor and providing early warnings to the student if
he/she needs to work hard to complete the course.
5
3. B ACKGROUND
Machine Learning with EDM has gained much more attention in the last few years. Many
machine learning techniques, such as collaborative filtering (Toscher and Jahrer, 2010), matrix
factorization (Thai-Nghe et al., 2011), and artificial neural networks (Wang and Liao, 2011) are
being used to predict students’ GPA or grades. In this section, we will describe these machine
learning techniques and how they are being used to predict students’ GPA in registered courses
within the context of education.
1. The algorithm measures how similar each student in the database to the active student by
calculating the similarity matrix.
3. Predict the GPA of the course of the active user by aggregating the GPA of that course
taken by the most similar students. The aggregation can be a simple mean or weighted
average by taking similarity between students into account.
The k nearest neighbour technique is used to select the neighbourhood for the active user
N(a) ⊂ U. The average rating of the neighbourhood users is calculated using the equation 1,
which becomes the predicted rating for the active use. The grade prediction becomes extremely
challenging for the student with a few courses attended which is a well-known drawback of CF
technique over the sparse dataset.
1 X
r̂aj = rij (1)
|N(a)|
i∈N (a)
6
3.2.1. Singular Value Decomposition
Singular Value Decomposition (SVD) is a matrix factorization technique that decomposes students-
courses matrix R into
R = UΣV T , (2)
where;
7
matrix approximation using only the known ratings of original matrix (Funk, 2006). Stochastic
Gradient Descent (SGD) is a convex optimization technique that gets the most accurate values
of those two featured matrices that are obtained during the decomposition of the original matrix
in the method of SVD. SGD has following steps:
1. Re-construct the target students-courses matrix by multiplying the two lower-ranked ma-
trices.
2. Get the difference between the target matrix and the generated matrix.
3. Adjust the values of the two lower-ranked matrices by distributing the difference to each
matrix according to their contribution to the product target matrix.
Above is a repeated process till the difference is lower than a preset threshold. By reducing
the dimensionality of the students-courses matrix, the execution speed is reduced, and the ac-
curacy of the prediction is increased because of considering only the courses that contribute to
the reduced data. Dimensionality reduction leads to the reduction of noise and over-fitting. This
method is also used in recommender systems for the Netflix challenge (Koren et al., 2009).
V ≈ W H, (3)
where;
• W is a u × k orthogonal matrix,
• H is a k × v orthogonal matrix.
8
3.3. R ESTRICTED B OLTZMANN M ACHINES
The method of Restricted Boltzmann Machines (RBM) is an unsupervised machine learning
method. Unsupervised algorithms are used to find the structural patterns within the dataset. We
have used RBM to predict the students’ performance in different courses. An RBM is in the
form of a bipartite graph that creates two layers of nodes. The first layer is called the visible
layer, which contains the input data (Course Grades). These nodes are connected to the second
layer which is called the hidden layer that contains symmetrically weighted connections. From
the Figure 3 we can see that the graph have five visible nodes (Course Grades) denoted by vi
and four hidden nodes indicated by hj . The weights between the two nodes are wij . Here each
visible node vi represents the grade for course i, for a particular student.
Hidden
Nodes
hj
Weights
wij
Visible
Nodes
vi
Course 1 Course 2 Course 3 Course 4 Course 5
Figure 3: A Restricted Boltzmann Machines (RBM) with five courses and four hidden nodes
for a specific student.
RBM is a form of Markov Random Field (MRF). MRF is a type of probabilistic model
that encodes the structure of the model as an undirected graph for which the energy function is
linear in its free parameters. The energy function E(v, h) for RBM can be calculated using the
equation 4.
9
4. M ETHODS
We used CF (UBCF), MF (SVD and NMF) and RBM techniques to predict GPA of the student
for the courses. A feedback model is developed based on the predicted GPA of the student in a
course.
10
scale with respect to the letter grades A+=4, A=4, A-=3.67, B+=3.33, B=3.0, B-=2.67, C+=2.33,
C=2.0, C-=1.67, D+=1.33, D-=1.0 and F=0.0. Figure 4 shows the frequency distribution of
grades for the students whose grades are available in the dataset. We can see most of the students
have B or B- grades in the courses they have taken.
As prediction algorithm works best with centering predictor variables, so all the data were
transformed by centering (average GPA of a course is subtracted from all GPAs of that course).
The main characteristics of the dataset are shown in the Table 3.
Characteristic Number
Total students 225
Total courses 24
Total cells 5400
Elements (grades) available 1736
Elements (grades) missing 3664
Matrix density 32.14%
11
Linear Circuit Analysis, Islamic Studies, and Signals and System. A student with Id. SB185
have similar GPA in Electronic Circuit and Design course like the student with Id. SB145 and
this student need to enroll into Linear Circuit Analysis, Islamic Studies, Signals and Systems,
and D-Lab courses.
Collaborative Filtering: We have used UBCF to predict the students’ grades in courses.
UBCF do grade prediction of a student s in a course c by identifying student grades in same
courses as s. For prediction of grades, the neighborhood students ns similar to student s are
selected that have taken at least nc courses that were taken by student s. To apply UBCF model
we first converted the students-courses matrix R into a real-valued rating matrix having student
GPA from 0 to 4. To measure the accuracy of this model we have split the data into 70% trainset
and 30% testset. In UBCF model The similarity between students and courses is calculated
using k nearest neighbors.
Matrix Factorization: Matrix factorization is the decomposition of a matrix V into the
product of two matrices W and H, i.e. V ≈ W H T (Koren et al., 2009). In this study, we have
used SVD and NMF matrix factorization techniques to predict the student GPA. The main issue
of MF techniques is to find out the optimized value of matrix cells for W and H.
In SVD approach, the students’ dataset is converted into real-valued rating matrix having
student grades from 0 to 4. The dataset is split into 70% for training the model and 30% for
testing the model accuracy. We used Funk SVD to predict GPA in the courses for which the
students are shown in Table 4 have not yet taken the courses. The largest ten singular values are
191.8012, 18.8545, 14.7946, 13.8048, 12.4328, 11.8258, 11.1058, 10.2583, 9.5020 and 9.1835.
It can be observed from the Figure 5 that the distribution of the singular values of students-
courses matrix diminishes quite fast suggesting that the matrix can be approximated by a low-
rank matrix with high accuracy. This encourages the adoption of low-rank matrix completion
methods for solving our grade/GPA prediction problem.
By applying Funk’s proposed heuristic search technique called Stochastic Gradient Descent
(SGD) gradient to the matrix G we obtained two matrices student and courses dimensional
spaces (with the number of hidden features set to two, so as to ease the task of visualizing the
data). The stochastic gradient descent technique estimates the best approximation matrix of the
problem using greedy improvement approach (Pelánek and Jarušek, 2015).
Table 5 represents the students’ features dimensional space, and Table 6 represents courses’
features dimensional space. With the dot product of these features dimensional space we can
12
Figure 5: Singular vales distribution of students-courses matrix
predict GPA in the courses for which the students are shown in Table 4 needs to enroll. Please
note that we usually do not know the exact meaning of the values of these two-dimensional
space, we are just interested in finding the correlation between the vectors in that dimensional
space. For understanding, take an example of a movie recommender system. After matrix
factorization, each user and each movie are represented by two-dimensional space. The values
of the dimensional space represent the genre, amount of action involved, quality of performers
or any other concept. Even if we do not know what these values represent, but we can find the
correlation between users and movies using the values of dimensional space.
Table 5: Students’ features dimensional space Table 6: Courses’ features dimensional space
Name V1 V2 Name V1 V2
SB145 0.39 0.18 Linear Circuit Analysis 1.19 -0.04
SB161 0.45 0.20 Electronic Circuit and Design 0.94 0.10
SB185 0.42 0.20 Islamic Studies 1.77 -0.03
SB229 -0.31 0.02 Signals and Systems 0.34 0.20
SB304 0.09 0.12 D-Lab 0.46 0.18
13
Figure 6: Rank-k using NMF
Restricted Boltzmann Machines: We have also used RBM an unsupervised learning tech-
nique to predict the student grades in different courses. RBM has been used to fill the missing
data in a students-courses matrix. We have split the data into 70% trainset and 30% testset. We
have trained the RBM method with a learning rate of 0.1, momentum constant of 0.9, the batch
size of 180, and for 1000 epochs.
14
1. Build Student Profile: In the first phase of feedback model; we have to parse students and
courses data into the form of hStudent, Course, GP Ai triplet to built students’ profile.
A students-courses matrix R is created that contains students’ performance in each course
taken. In a matrix R, students are represented in rows and courses are represented in
columns. The value of each cell of matrix R is Rij , that can be calculated using the
equation 7.
student’s i mark on course j, if the student enrolled in course j
Rij =
empty, if the student did not enroll in course j
(7)
For the courses in which a student did not enroll, Rij will be empty. For illustration,
a small chunk of the dataset is presented in matrix given below. This matrix holds the
dataset of five different students and five different courses.
3.67 4
4 3.67
Rij = 3.67 (8)
2 2.67
2. Predict Course GPA: Now we have a matrix R, for which we are interested to find the
unknown GPAs for the courses, which the student has not taken yet. To find the predicted
GPA we have used CF (UBCF), MF (SVD and NMF), and RBM techniques. Detailed
methodology for these techniques is described in section 4.
where;
15
• (1 − P (Lj−1) is the knowledge that is unknown.
Using the equation 9, student knowledge is measured by inferring his knowledge in the
course domain. As we know the probability of the knowledge in the previous step is the
predicted GPA for the student in the subject. To calculate the knowledge gain course
domain average has been converted into the range (0 to 1) and multiplied by the learning
rate 0.005.
5. Feedback: After computing the student knowledge in particular course domain and knowl-
edge inference, the feedback is made. If the student knowledge inference results are less
than 2.67 GPA in a course, then the system generates a warning that the student needs ef-
fort in that course. In this way, feedback results can inform the instructors that the student
is weak in a particular course.
5. R ESULTS
5.1. C ORRELATION A NALYSIS
To find the pre-admission factors (SSC, HSSC, entry test and interview) that can predict stu-
dent performance in the university Pearson Correlation has been applied. The result shows that
there is a positive correlation between entry test and Cumulative Grade Point Average (CGPA)
and also between HSSC and CGPA. The correlation coefficients (r) between the entry test and
CGPA, and HSSC and CGPA are very close (r = 0.29 and r = 0.28 respectively), indicating that
both entry test and HSSC are equally important in predicting the CGPA of a student. Figure
8 shows the correlation between the entry test of the students and their CGPA, and Figure 9
shows the correlation between the higher secondary school performance and the CGPA. These
figures show that the students with a higher score in entry test and a higher percentage in HSSC
performance obtain higher CGPA in the degree program.
4
4
3
3
CGPA
CGPA
2
2
1
50 55 60 65 70 75 80 65 70 75 80 85
Figure 8: Correlation between entry test Figure 9: Correlation between HSSC and
and CGPA CGPA
16
5.2. G RADE P REDICTION
For students, GPA prediction, students-courses matrix G is constructed. The data were trans-
formed by centering the predictor variables by taking average GPA of a course and subtracted
it from all GPAs of that course. 70% of the dataset is used for training the CF MF and RBM
models. Student GPAs for the courses has been predicted and displayed in Table 7.
Table 7: Student GPA prediction in courses based on CF, SVD, NMF and RBM technique
There are several types of measures for evaluating the success of models. However, the eval-
uation of each model depends heavily on the domain and system’s goals. For our system, our
goal is to predict students’ GPA and make decisions if a student needs to work hard to com-
plete the course. These decisions work well when our predictions are accurate. To achieve it,
we have to compare the prediction GPA against the actual GPA for the students-courses pair.
Some of the most used metrics for evaluation of the models are the Root Mean Squared Error
(RMSE), Mean Squared Error (MSE) and Mean Absolute Error (MAE). We evaluated model
predictions by repeated random subsample cross-validation. We performed ten repetitions. In
17
each run, we choose randomly 70% of students data into the train set and 30% of students data
into the test set. We have computed RMSE, MSE, and MAE for each model. From Figure 10
the results show that the RBM model provides a clear improvement over the CF and MF models.
Please note we are not performing student-level cross-validation of predicted results on newly
registered students in this study but the currently enrolled students.
18
6. I NSIGHTS
In this study, we have used CF (UBCF), MF (SVD and NMF) and RBM techniques to predict
the students’ performance in the courses. CF is a popular method to predict the students’ per-
formance due to its simplicity. In this technique, the students’ performance is analyzed by using
the previous data. It provides feedback to enhance the students’ learning process based on the
outcome of the analysis. However, this method has several disadvantages: since it depends upon
the historical data of users or items for predicting the results. It shows poor performance when
there is too much sparsity in the data, due to which we are not able to predict the students’
performance accurately. Comparatively, in SVD technique, the data matrix R is decomposed
into users-features space and items-features space. When SVD technique is used with gradi-
ent descent algorithm to compute the best rank-k matrix approximation using only the known
ratings of R, the accuracy of predicting the students’ performance enhances but it may contain
negative values which are hard to interpret. NMF technique enhances the meaningful inter-
pretations of the possible hidden features that are obtained during matrix factorization. RBM
is an unsupervised machine learning technique that is suitable for modeling tabular data. It
provides efficient learning and inference better prediction accuracy than matrix factorization
techniques. The use of RBM in recommender systems and e-commerce have also shown good
results (Kanagal et al., 2012). From the above discussion, it is clear that the RBM technique out-
performs CF and MF techniques with lesser chances of error. The overall result obtained in this
study also shows that RBM surpasses other techniques in predicting the student’s performance.
7. L IMITATIONS
We note that the reported findings of this study have been based on the dataset of the perfor-
mance of the undergraduate students from ITU. The dataset used in the study is limited with
GPAs available for students in the particular courses. After using CF (UBCF), MF (SVD and
NMF) and RBM techniques on the dataset, we can see that the RMSE for RBM technique is
lower compared to the RMSE of other techniques. RMSE can be estimated with more clear
results if more information of the students’ GPAs is available. Student motivation during stud-
ies also plays a significant role in the prediction of student success which can be considered in
future study related to the grade prediction. Moreover, there is a need to improve the prediction
results by dealing with the cold-start problems. Also, models based on tensor factorization can
be investigated to take the temporal effect into account in the student performance prediction.
Despite these limitations, our research findings have important practical implications for the
universities and institutes in enhancing their students’ retention rate.
8. C ONCLUSION
Early GPA predictions are a valuable source for determining student’s performance in the uni-
versity. In this study, we discussed CF (UBCF), MF and RBM techniques for predicting stu-
dent’s GPA. We use RBM machine learning technique for predicting student’s performance in
the courses. Empirical validation on real-world dataset shows the effectiveness of the used RBM
technique. In a feedback model approach, we measure the students’ knowledge in a particular
course domain, which provides appropriate counseling to them about different courses in a par-
ticular domain by estimating the performance of other students in that course. This feedback
19
model can be used as a component of an early warning system that will lead to students’ motiva-
tion and provides them early warnings if they need to improve their knowledge in the courses. It
also helps the course instructor to determine weak students in the class and to provide necessary
interventions to improve their performance. In this way rate of the students’ retention can be
increased.
R EFERENCES
AUD , S., NACHAZEL , T., W ILKINSON -F LICKER , S., AND D ZIUBA , A. 2013. The condition of educa-
tion 2013. Government Printing Office.
BAKER , R. S. AND YACEF, K. 2009. The state of educational data mining in 2009: A review and future
visions. JEDM-Journal of Educational Data Mining 1, 1, 3–17.
B ERRY, M. W., D UMAIS , S. T., AND O’B RIEN , G. W. 1995. Using linear algebra for intelligent infor-
mation retrieval. SIAM review 37, 4, 573–595.
C ALDERS , T. AND P ECHENIZKIY, M. 2012. Introduction to the special section on educational data
mining. ACM SIGKDD Explorations Newsletter 13, 2, 3–6.
C ALLENDER , C. AND F ELDMAN , R. 2009. Part-time undergraduates in higher education: A literature
review. Prepared for HECSU to inform Futuretrack: Part-time students. London, Birkbeck, University
of London.
D EERWESTER , S., D UMAIS , S. T., F URNAS , G. W., L ANDAUER , T. K., AND H ARSHMAN , R. 1990.
Indexing by latent semantic analysis. Journal of the American society for information science 41, 6,
391.
E LBADRAWY, A. AND K ARYPIS , G. 2016. Domain-aware grade prediction and top-n course recommen-
dation. Boston, MA, Sep.
E LBADRAWY, A., P OLYZOU , A., R EN , Z., S WEENEY, M., K ARYPIS , G., AND R ANGWALA , H. 2016.
Predicting student performance using personalized analytics. Computer 49, 4, 61–69.
F RANC , V., H LAV Á Č , V., AND NAVARA , M. 2005. Sequential coordinate-wise algorithm for the non-
negative least squares problem. In Computer Analysis of Images and Patterns. Springer, 407–414.
F UNK , S. 2006. Netflix update: Try this at home. http://sifter.org/˜simon/journal/20061211.html.
Online; accessed 11 Jan 2017.
G OLUB , G. H. AND VAN L OAN , C. F. 2012. Matrix computations. Vol. 3. JHU Press.
H AWKINS , W. J., H EFFERNAN , N. T., AND BAKER , R. S. 2014. Learning bayesian knowledge tracing
parameters with a knowledge heuristic and empirical probabilities. In International Conference on
Intelligent Tutoring Systems. Springer, 150–155.
H IGHER E DUCATION C OMMISSION OF PAKISTAN. 2012. Cur-
riculum of electrical engineering b.sc./be/bs & m.sc./me/ms.
http://hec.gov.pk/english/services/universities/RevisedCurricula/Documents/2011
Online; accessed 10 Feb 2017.
I QBAL , Z., Q ADIR , J., AND M IAN , A. N. 2016. Admission criteria in pakistani universities: A case
study. In 2016 International Conference on Frontiers of Information Technology (FIT). IEEE, 69–74.
I RAJI , M. S., A BOUTALEBI , M., S EYEDAGHAEE , N. R., AND T OSINIA , A. 2012. Students classifi-
cation with adaptive neuro fuzzy. International Journal of Modern Education and Computer Sci-
ence 4, 7, 42.
20
K ANAGAL , B., A HMED , A., PANDEY, S., J OSIFOVSKI , V., Y UAN , J., AND G ARCIA -P UEYO , L. 2012.
Supercharging recommender systems using taxonomies for learning user purchase behavior. Proceed-
ings of the VLDB Endowment 5, 10, 956–967.
K NOWLES , J. E. 2015. Of needles and haystacks: Building an accurate statewide dropout early warning
system in wisconsin. JEDM-Journal of Educational Data Mining 7, 3, 18–67.
KOREN , Y., B ELL , R., VOLINSKY, C., ET AL . 2009. Matrix factorization techniques for recommender
systems. Computer 42, 8, 30–37.
M AC D ONALD , I. 1992. Meeting the needs of non-traditional students: Challenge or opportunity for
higher education. Scottish Journal of Adult Education 1, 2, 34–46.
M EIER , Y., X U , J., ATAN , O., AND VAN DER S CHAAR , M. 2016. Predicting grades. IEEE Transactions
on Signal Processing 64, 4, 959–972.
P EL ÁNEK , R. AND JARU ŠEK , P. 2015. Student modeling based on problem solving times. International
Journal of Artificial Intelligence in Education 25, 4, 493–519.
S AARELA , M. AND K ÄRKK ÄINEN , T. 2015. Analysing student performance using sparse data of core
bachelor courses. JEDM-Journal of Educational Data Mining 7, 1, 3–32.
S ALAKHUTDINOV, R., M NIH , A., AND H INTON , G. 2007. Restricted boltzmann machines for collab-
orative filtering. In Proceedings of the 24th international conference on Machine learning. ACM,
791–798.
S ARWAR , B. M., KONSTAN , J. A., B ORCHERS , A., H ERLOCKER , J., M ILLER , B., AND R IEDL , J.
1998. Using filtering agents to improve prediction quality in the grouplens research collaborative
filtering system. In Proceedings of the 1998 ACM conference on Computer supported cooperative
work. ACM, 345–354.
S WEENEY, M., L ESTER , J., AND R ANGWALA , H. 2015. Next-term student grade prediction. In Big
Data (Big Data), 2015 IEEE International Conference on. IEEE, 970–975.
S WEENEY, M., R ANGWALA , H., L ESTER , J., AND J OHRI , A. 2016. Next-term student performance
prediction: A recommender systems approach. arXiv preprint arXiv:1604.01840.
T HAI -N GHE , N., D RUMOND , L., H ORV ÁTH , T., K ROHN -G RIMBERGHE , A., NANOPOULOS , A., AND
S CHMIDT-T HIEME , L. 2011. Factorization techniques for predicting student performance. Educa-
tional Recommender Systems and Technologies: Practices and Challenges, 129–153.
T HAI -N GHE , N., D RUMOND , L., H ORV ÁTH , T., NANOPOULOS , A., AND S CHMIDT-T HIEME , L. 2011.
Matrix and tensor factorization for predicting student performance. In CSEDU (1). Citeseer, 69–78.
T HAI -N GHE , N., D RUMOND , L., H ORV ÁTH , T., S CHMIDT-T HIEME , L., ET AL . 2011. Multi-relational
factorization models for predicting student performance. In Proc. of the KDD Workshop on Knowl-
edge Discovery in Educational Data. Citeseer, 27–40.
T OSCHER , A. AND JAHRER , M. 2010. Collaborative filtering applied to educational data mining. KDD
cup.
VAN D E S ANDE , B. 2013. Properties of the bayesian knowledge tracing model. JEDM-Journal of Edu-
cational Data Mining 5, 2, 1–10.
WANG , Y.- H . AND L IAO , H.-C. 2011. Data mining for adaptive learning in a tesl-based e-learning
system. Expert Systems with Applications 38, 6, 6480–6485.
X U , J., M OON , K. H., AND VAN DER S CHAAR , M. 2017. A machine learning approach for tracking
and predicting student performance in degree programs. IEEE Journal of Selected Topics in Signal
Processing.
21
Z IMMERMANN , J., B RODERSEN , K. H., H EINIMANN , H. R., AND B UHMANN , J. M. 2015. A model-
based approach to predicting graduate-level performance using indicators of undergraduate-level per-
formance. JEDM-Journal of Educational Data Mining 7, 3, 151–176.
22