Computers and Education: Arti Ficial Intelligence
Computers and Education: Arti Ficial Intelligence
A R T I C L E I N F O A B S T R A C T
Keywords: Adaptive learning, supported by Information & Communication Technology (TIC), is an important research area
Adaptive recommender systems for educational systems which aim to improve the outcomes of students. Thus, the investigation of what should be
Deep auto encoder adapted and how much to adapt constitute a foundation to Adaptive E-learning Systems (AES). In this paper, we
Artificial neural networks
compared three classes of Deep Auto Encoders and the popularity model to address the problem of learning and
E-learning
Lifelong learning
predicting the preferences of student on AES: Collaborative Denoising Auto Encoders (CDAE), Deep Auto En-
coders for Collaborative Filtering (DAE-CF), and Deep Auto Encoders for Collaborative Filtering using Content
Information (DAE-CI). The results point out that the DAE-CF is more effective providing significant adaptability.
Furthermore, we present the concept named as signature of preference to represent a more granular class of
adaptability. Therefore, this model may be used in e-learning systems to provide adaptability and help to improve
the outcomes of students.
* Corresponding author.
E-mail addresses: [email protected], [email protected] (E. Gomede), [email protected] (R.M. de Barros), [email protected]
(L.S. Mendes).
https://doi.org/10.1016/j.caeai.2021.100009
Received 13 October 2020; Received in revised form 20 January 2021; Accepted 20 January 2021
2666-920X/© 2021 The Author(s). Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-
nc-nd/4.0/).
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
However, an AES should include some features to be considered effec- Discounted Cumulative Gain (NDCG), Personalization, Coverage, and
tive. For example, the system should consider the case in which a student Serendipity (SAUC) are used to compare the results. Thus, we analyze the
A marks an object that causes this object to become available to another evaluation of recommendation systems focusing on the quality of rec-
student B, without this student to be aware of this interaction. Thus, the ommendations rather than on their predictive accuracy of algorithms.
cycle of students (select and rate) and system (group and filter) provides Finally, the split strategy of the dataset is stated and discussed.
the five components, or levels, of an adaptive e-learning recommendation This paper is organized as follows. This initial section presents
system. On the first level, a recommendation system works with collab- fundamental considerations and justifications for this work and defines
orative filtering, i.e., depends on a voluntary and responsible interaction its core objectives. Section 2 presents an overall review of the key topics
between students and objects (Hwang et al., 2013; Krauss, 2016). For treated and the main works upon which this work is based. Section 3
example, an incorrect rating of a book might be generating a wrong presents the main concepts behind the experiments and describes the
recommendation, but on the other hand, a rating done by a group of comparison. Section 4 presents the discussions upon the results obtained
specialized users (professors, tutors, reviewers, and/or researchers) from the data analysis, architectures, hyperparameters, metrics, and
might be increasing the value of the recommendations. In a second, the models. In the concluding section, the conclusions and subsequent de-
recommendation system presents collaborative intelligence. For velopments are presented.
example, a book recommended to student A is not related to the behavior
of a student B, instead, the behavior of some collectivity is responsible for 2. Related works
recommendations (Krauss, 2016; Protasiewicz et al., 2016). Third, a
recommendation system is in user control. For instance, a recommended Many papers about recommendation systems are dedicated to
book, which is part of mandatory bibliography, cannot be rejected by computational issues like performance metrics, mathematical models,
students (although this book might be evaluated by students). As follows, algorithms, and data pre-processing (Strub et al., 2016; Li and She, 2017;
the coverage of recommendations is limited. Fourth, recommendation Chae et al., 2019; Ordway-West et al., 2018). However, there is an
systems offer guidance. For example, a list of recommended articles, increased awareness related to pedagogical issues to improve the out-
which support students in their research, is a guide (Medio et al., 2020). comes of students (Bobadilla et al., 2009; Hwang et al., 2013; Krauss,
This orientation represents a fundamental performance issue in learning 2016; Protasiewicz et al., 2016; Barbieri et al., 2017; Nafea et al., 2019;
science; excess of autonomy can be costly and generate low performance Sedhain et al., 2015; Strub et al., 2016; Chae et al., 2019). Like so, some
results. The key idea is achieving the equilibrium. Finally, recommen- researchers have published approaches to e-learning recommendation
dation systems provide adaptability. Different students do not take equal systems including educational elements (Bobadilla et al., 2009; Hwang
advantage of the same instructions; thus, adaptability is needed. et al., 2013; Krauss, 2016; Protasiewicz et al., 2016; Barbieri et al., 2017;
From a pragmatic point of view, an adaptive e-learning recommen- Nafea et al., 2019).
dation system aims to improve the outcomes of students. However, there Related to computational issues, one of the firsts papers presenting
are some secondary objectives like (a) relevance (an LO needs to be the auto encoder applied to a recommendation system was proposed by
relevant, or it may become useless); (b) novelty (LO needs to present Wang et al. (2015). The auto encoder architecture is a class of feedfor-
innovation or not used in the past); (c) serendipity (an LO which is not ward neural networks which use semi supervised learning and, therefore,
expected, but is still relevant, to the student); and (d) diversity (several does not need a labeled dataset (Strub et al., 2016; Li and She, 2017; Chae
types of LO improving the coverage of the subject to be learned). et al., 2019; Ordway-West et al., 2018). It works with collaborative
Moreover, there are some requirements to make suitable recommenda- filtering based on sparse matrices, composed of implicit and explicit
tions: (a) transparency to explain the recommendation, (b) confidence to feedback, and provides better results compared to other algorithms of
support students to use the LO, (c) satisfaction to improve the outcomes collaborative filtering. The deep hidden layers, responsible for coding
of students, (d) persuasion to convince students to use the LO, (e) effi- and decoding, constitute the latent space which represents the factor-
ciency to assist students in their decisions, and (f) effectiveness to help ization of the sparse matrix. The number of hidden layers varies based on
the students to make suitable decisions (Bobadilla et al., 2009; Hwang the problem. Also, these architectures use many components present in
et al., 2013; Krauss, 2016; Protasiewicz et al., 2016; Barbieri et al., 2017; general artificial neural networks such as loss function (for example, Root
Nafea et al., 2019; Sedhain et al., 2015; Strub et al., 2016; Chae et al., Mean Square Error (RMSE)), activation function, backpropagation algo-
2019). In this manner, the component of intelligent algorithms might be rithm, stochastic gradient descent, learning rate, bias, and splitting
used in this context. dataset. The standard metrics used by evaluating the predicted outputs is
In this paper, we compare three classes of DEA and the popularity the Recall@K, where K represents the number of objects to be recom-
model to address the problem of learning and predicting students’ pref- mended. However, the models built by training of auto encoders might
erences on AES: Collaborative Denoising Auto Encoders (CDAE), Deep be difficult to explain due to the stochastic process of updating the
Auto Encoders for Collaborative Filtering (DAE-CF), and Deep Auto En- weights. Another limitation is the dataset available to train and test. If the
coders for Collaborative Filtering using Content Information (DAE-CI). number of observations is limited, the model may present an overfitting
The hyperparameters, activation function, batch size, dropout, epoch, or underfitting (Wang et al., 2015). Moreover, the splitting of the dataset
layers, loss function, and learning rate are presented and discussed. The needs to be considered developing suitable models. The most known split
performance metrics Mean Average Precision (MAP), Normalized techniques are the holdout, k-fold cross-validation, leave-one-out, and
Table 1
The structure and some example of the raw data.
student_id object_id learning_objecta ip hours timestamp
2
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Fig. 1. Temporal preference of students to access learning objects. It can be noted that such accesses were made during the school period and, most of them, between
the months of March until October and between the hours of 9 a.m. until 6 p.m. The dataset was collected between January of 2018 and December of 2019.
bootstrap (Wang et al., 2015). Broadly speaking, an auto encoder is the name (b) auto described, (c) segmented, (d) based on cognition, and
composed of common elements of neural networks, which build a latent (e) student centered. The first one boils down to identify the character-
space by coding and decoding sparse matrices, being suitable to fill out istics of students, like name, and to present that information on the
the gaps of interaction between student/object, learning and predicting content. The second one, besides the characteristics, collects the prefer-
the behavior of students. ences of students, like the tendency of determined subjects. The third one
With respect to pedagogical issues, one of the firsts papers on the classifies students by similarity and provides adaptability to the group.
subject was proposed by Bobadilla et al. (2009). The authors present The fourth one identifies the cognitive preferences, for example, the
many types of learning objects to be recommended like books, texts, tendency to hear an audio rather than examine a video. Finally, the last
tests, subjects, and teachers, providing a broad vision of e-learning one collects performance of students (in tests, for example) and adjusts
recommendation systems. Also, the learning objects can be standardized the difficulty level of the content. This strategy requires real-time data
by use of Shareable Content Object Reference Model (SCORM) and (Bobadilla et al., 2009; Hwang et al., 2013; Krauss, 2016). Our approach
provided by Learning Objects Repository (LOR). The use of the SCORM can be considered based on cognition, but instead of the set the prefer-
and LOR improves the quality of recommendations and helps teachers, ences of students in a class previously defined, we build a signature of
tutors, and administrators to build valuable learning objects. Likewise, preferences to consume learning objects. This approach allows building
these objects are manipulated from pre-school to post-graduate degree models suitable to each student and predicting the behavior from the use
and reutilized by many students, teachers, and tutors. Also, other papers of a learning object.
present the results of application from recommendation systems to in-
crease the engagement of students; this is achieved by use of the concept 3. Materials and methods
of flow (Bobadilla et al., 2009; Hwang et al., 2013; Krauss, 2016; Shih,
Chuang, & Hwang, 2010). These results were collected from real envi- The essential idea behind the operation of the auto encoder is to fill
ronments. Moreover, other ways of adaptability were proposed by the out the unknown values of the input sparse matrix (Sedhain et al., 2015).
classification of learning styles. These approaches are based on This operation provides 2 features: (a) to learn students’ behavior and (b)
Felder-Silverman Learning Style Model (FSLSM) and identify the ten- to predict the probability of consuming determined learning objects. The
dency of students to acquire, process, comprehend, and store information learning feature is based on the interactions of student/object which
(Nafea et al., 2019). Shih et al. (2010) proposed an adaptability of the provides the representation of how each student interacts with available
learning path of students, selecting the suitable courses, modules, and LOs. The predicting feature is related to the probability, based on pre-
content according to preferences for them. To sum up, the authors pro- vious behavior of each student, to interact with novel LOs. Thus, both are
posed approaches to provide adaptability of many aspects (learning ob- intrinsically dependent.
jects, paths, contents, and so forth) to assist students to improve their
outcomes. 3.1. The dataset
According to Dockterman (2018), there are five strategies of adapt-
ability, summarized from the simplest to the most complex; (a) based on The dataset used is composed of interactions of students with learning
objects from a Massive Open Online Course (MOOC). Examples of these
objects are videos from open courses of many subjects related to com-
puter science. The interaction obtains implicit feedback provided by the
Table 2
time expended per student in a video, for instance. The structure of raw
The sparse matrix built with the transposed raw data.
data is shown in Table 1.
object_id The dataset does not contain restrictions, as explained before in
student_id 1 1 2 … n Section 1, like mandatory learning objects. In addition, the characteristic
2 1.10 ? … ? hours represent the amount of time spent by a student in an interaction
… ? 0.30 … ?
N ? ? … 1.00
3
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Fig. 2. The (a) overall long tail phenomenon, (b) related to students, and (c) related to learning objects.
Fig. 3. The sparse matrix resulting from the transposition of the dataset (the firsts 100 students on y axis and firsts 100 objects on x axis). This matrix indicates that
few students access few learning objects. Thus, there are many objects and students who have not benefited from the contents that are hidden.
with learning objects. Thus, the student A might spend X and Y amounts specific objects. The second states that most students present the mass
of time at the same object A. Furthermore, students can view an object and niche behavior. G€ obel and Mehm (2013) presented empirical evi-
many times and interact with it in different ways (e.g., watching the same dence that both hypotheses are valid but are context dependent. Also, the
object many times, stopping, reloading, forwarding, rewarding, and long tail occurs with students and objects. Fig. 2 shows the (a) overall
jumping). Thus, to model the student interest on a given learning object, long tail phenomenon, (b) related to students, and (c) related to objects.
we aggregate all interactions of the student in an object by a weighted As shown in Fig. 2, the long tail represents, in other words, that few
sum of interaction type and applying a log transformation to smooth the students access few objects often. This is a problem to collaborative
distribution2. The preference to access the learning object is shown in
Fig. 1.
As shown in Fig. 1, students prefer to access the MOOC during the day
and the school year. Moreover, the dataset was collected between 2018
and 2019 and the MOOC was available 24/7/365 (except when down-
time was planned). The raw data, indicated in Table 1, needs to be
transposed to a matrix to be used with the algorithms (Wang et al., 2015).
The result is a bidimensional m x n sparse matrix composed of in-
teractions of students and objects represented by the net amount of time
spent in the object. The sparse matrix is shown in Table 2.
The phenomenon that occurs in a recommendation system based on
collaborative filtering is called long tail3 (Strub et al., 2016; Li and She,
2017; Chae et al., 2019; Ordway-West et al., 2018). This phenomenon is
ruled by two hypotheses fundamentally different. The first states that
most students follow the mass behavior while the minority search
2
In statistics and image processing, to smooth a dataset is to create an
approximating function that attempts to capture important patterns in the data,
while leaving out noise or other fine-scale structures/rapid phenomenon.
3
In probability theory, heavy-tailed distributions are probability distributions Fig. 4. Cluster analysis implementing the PCA technique that indicates the
whose tails are not exponentially bounded: that is, they have heavier tails than concentration of the interactions in some students and learning objects. This
the exponential distribution. In many applications it is the right tail of the dis- visualization complements the long tail phenomenon from another point of
tribution that is of interest, but a distribution may have a heavy left tail, or both view. The scale, in log and normalized form, from 2 to 3 represents the time
tails may be heavy. spent in each interaction student/object.
4
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Table 5
The set of hyperparameters used in the experiments.
Hyperparameter Meaning Range CDAE* DAE-CF DAE-CI
Activation Function used by the activation of neuron Many SELU SELU SELU
Batch Size The size of sampler present to the network [1:N] 64 64 64
Dropout Rate of unused neurons [0:1] 0.2 0.2 0.2
Epoch Number of the iterations to train the network [1:∞] 50 50 50
Factors Max number of neurons in each layer [1:∞] 500 15 15
Layers Number of layers [1:∞] [128,256,128] [512,256,512] [512,256,512]
Learning Rate Rate of updating the synapses weights [0:1] 0.0001 0.0001 0.0001
Regularization Rate to avoiding overfitting [0:1] 0.0001 0.01 0.01
*The CDAE, DAE-CF, and DAE-CI represent the values optimized in class of DAE. The number of layers is chosen according to architecture.
5
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
1 X 1 XN
MAP@K ¼ jUj Pu ðkÞ relu ðkÞ (3) personalization score indicates students’ recommendations are different,
jUj u¼1 M k¼1
meaning the model is offering a personalized experience to each one. The
This metric presents the average precision to recommend k objects to P metric is represented by Equation (5).
a student. Moreover, the average, which is computed to all students,
presents the mean of recommendation of k objects to all students. The XU
ð1 cosine similarityðuÞÞ
P¼ (5)
MAP metric is important to evaluate the ability of the recommendation
u¼1
U
model to achieve the most appropriate recommendation to all students.
As follows, MAP represents both sparse matrix dimensions: students and Thus, P metric offers a way to compare if the recommendation model
objects. is presenting the same recommendations list for each student. The met-
Another key metric is Normalized Discounted Cumulative Gain rics MAP, NDCG, and P present, respectively, the ability of the model to
(NDCG). The NDGC is a measure of the gain of an object based on its recommend relevant objects, recommend top k ordered objects and the
position in the result list. The underlying assumptions are that; relevant level of personalization of the model built. For all these metrics, the
objects are more useful if shown earlier in the result list and highly closer the results get to 1, the better the model becomes.
relevant objects (HRO) are more useful than moderately relevant objects Serendipity is a complex concept which includes other ones: (a)
(MRO), which is more useful than irrelevant objects (IO). Thus, HRO > relevance, (b) novelty, and (c) unexpectedness. We define a relevant LO
MRO > IO. Each recommendation on the list has a relevance score as an LO that a student likes, consumes, or is interested in. The term novel
associated with it; cumulative gain represents the sum of all relevant may have different meanings that can be summarized as: (a) novel item, a
scores in the recommendation set. Moreover, the position of the score is recently added item that users have not yet rated; (b) forgotten item, an
significant because two recommendation sets A ¼ {3, 3, 2, 1} and B ¼ item a user might have forgotten that she/he consumed some time ago;
{3, 2, 1, 3} have the same score, even if they present different sequences (c) unknown item, an item that a user has never consumed in his/her life;
of objects. Thus, the measure involves discounting the relevance score by and (d) unrated item, an item that a user has not rated yet regardless of
dividing it with the log of the corresponding position. However, it is still whether she/he is familiar with the item. We used the term novel item
not complete. Depending on several factors, the number of recommen- corresponding to definition unknown item due to the popularity of the
dations may vary for each student. At that time, normalization is definition in studies on serendipity (Denis et al., 2016). Unexpectedness
required. The NDCG is computed by Equation (4). and surprise are terms frequently used in the literature (Denis et al.,
2016), but to the best of our knowledge there is no consensus on the
XK
2jωðkÞεIμ j 1 terms’ definitions. In this paper, the terms unexpected and surprising
NDCG@Kðμ; ωÞ ¼ (4)
k¼1
log2 ðK þ 1Þ refer to LO that significantly differ from the profile of the student
regardless of how novel or relevant those LO are. Thus, serendipity might
The cut-off version of NDCG is NDCG@K, where K represents the be a way to expand the knowledge of students providing content which is
number of objects evaluated in the recommendation list. Compared to different but relevant. Also, this strategy might keep the focus of students
the MAP metric NDCG carries out an important job at evaluating the avoiding that they become bored with obvious suggestions that they
position of ranked items. It operates beyond the binary relevant/non- might have already discovered.
relevant scenario. Thus, both metrics are complementary, allowing the
interpretation of the results for different dimensions.
Personalization (P) is a metric to assess if a model recommends many 4
These figures were provided and adapted by ML Flow, an open source
of the same objects to different students. It is the dissimilarity (1 cosine platform for the machine learning lifecycle. For more details, please visit
similarity) between students’ lists of recommendations. A high https://mlflow.org/
6
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Fig. 7. The built model (a) and training history (b) of the Collaborative Denoising Auto Encoders (CDAE).
4.1. Popularity model The CDAE assumes that the interactions of student/object represent a
corrupted form of a complete matrix. Thus, this Auto Encoder aims to
This model performs recommendations presenting the most popular reconstruct the matrix considering the missing values as corrupted in-
learning objects. It does not provide adaptability and shows the same teractions. The CDAE is equally represented as a one-hidden-layer neural
recommendation list to all students. The recommendation list is network. In the input layer, there are a total of O þ 1 nodes, where each
composed of the sum of time spent in each object, in descending order; of the firsts O nodes corresponds to an object, and the last node is a
therefore, the most visualized learning object, in time, is the first one, the
second most visualized is the second on the list, and so on. Therefore, this
model is not related to each student, being an overall recommendation of 5
We considered the best case of the all built models with their hyper-
all learning objects. The metrics of this model are shown in Fig. 5. parameters. The worst and average cases were suppressed because they did not
As shown in Fig. 5, the P metric is around 0.334; i.e., this value present expressive differences related to the best case.
7
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Fig. 9. The built model (a) and training history (b) of the Deep Auto Encoders for Collaborative Filtering (DAE-CF).
student-specific node which means the node and its associated weights and the history of training is shown in Fig. 9.
are unique for each student s 2 S in the data. The metrics of this model are The built model includes six layers; UserScore, EncLayer0, Latent-
shown in Fig. 6. Space, Dropout, DecLayer0, and UserScorePred. All of them have the
As shown in Fig. 6, the P metric is around 0.733; i.e., this value input and output sub layers. The UserScore is the same as CDAE. The
represents 73% of adaptability. Also, the metrics MAP, NDCG@5, EncLayer0 is the first layer to encode the previous layer into a factored
NDCG@10, and coverage presented showed important increases. Thus, matrix. The number 0 in the name represents the first layer; thus, the
related to the Popularity Model, the CDAE provides a significant second layer has extension 1, the third extension 2, and so on. The
improvement in adaptability of the learning objects to the preferences of LatentSpace characterizes the finishing stage of efficient compression of
the student. The built model and the history of training are shown in sparse matrix present to the UserScore layer. The Dropout layer repre-
Fig. 7. sents the number of unnecessary neurons that will be removed. Although
The built model allows eight layers: UserScore, UserContent, more precisely, it does not mean that such neurons are unnecessary. The
Dropout, Embedding, Dense, Flatten, LatentSpace, and UserScorePred. pruning process is done at random, that is, there is a probability of
All of them have input and output sub layers. The UserScore is respon- eventual removal of neurons that are useful to the model. The DecLayer0
sible for getting all 4,856 interactions for each student with the learning represents the first layer that starts decoding the LatentSpace into a
objects. For instance, student A has the interaction vector V ¼ {10, NA, subsequent layer. Finally, this last layer decodes the LatentSpace into
…, 5}, where each position of the vector represents the time spent by the UserScorePred, which contains the preference of each student. The
student with the object. NA represents the lack of interaction. The recommendation list, to student 10, is showing in Table 7.
UserContent represents the layer that contains the O þ 1 nodes, as As shown in Table 7, the column score presents the relevancy of the
explained before. In this manner, the default value is 1. The Dropout learning object to the student 10. The first columns represent the position
layer represents the number of unnecessary neurons that will be of the object into the recommendation list.
removed. The Embedding layer groups the O þ 1 nodes into a cluster
based on factors hyperparameter (Table 5). The Dense and Flatten layers 4.4. Deep Auto Encoders for Collaborative Filtering using content
compress the previous layers into the LatentSpace. Finally, this last layer information (DAE-CI)
decodes the LatentSpace into UserScorePred; which contains the pref-
erence of each student. The recommendation list, to student 10, is The DAE-CI is a variation of DAE-CF adding content information. As
showing in Table 6. follows, this architecture is a hybrid implementation. To build the model,
As shown in Table 6, the column score presents the relevancy of the the information store in the learning_object includes collaborative
learning object to student 10. The first columns represent the position of filtering. This information represents the interaction of each student with
the object in the recommendation list. learning objects and is used to include content information to the student
level. The model built by DAE-CI worsens around 5% when compared to
4.3. Deep Auto Encoders for Collaborative Filtering (DAE-CF)
Table 7
The DAE-CF demonstrates the tendency to generalize beyond CDAE The top@10 recommendation list provided by the built model using the Deep
or other kinds of architecture with few layers. Otherwise, this architec- Auto Encoders for Collaborative Filtering (DAE-CF).
ture uses the nonlinear activation functions and techniques to avoid score object_id learning_object
overfitting, such as regularization and dropout. The built model by DAE- 0 0.842951 3188 Learning Object 389
CF presents an improvement around 10% when compared to CDAE. This 1 0.828034 2 Learning Object 10
improvement is due, specially, to the deep layers. The metrics of this 2 0.794385 4675 Learning Object 612
3 0.764501 4572 Learning Object 52
model are shown in Fig. 9.
4 0.728598 1328 Learning Object 22
As shown in Fig. 8, the P metric is around 0.872; i.e., this value 5 0.683430 501 Learning Object 145
represents around 87% of adaptability. Also, the metrics MAP, NDCG@5, 6 0.606181 1262 Learning Object 214
NDCG@10, and coverage presented a significant increase. Thus, related 7 0.583904 1992 Learning Object 28
to CDAE, the DAE-CF provides a significant improvement in adaptability 8 0.568082 972 Learning Object 188
9 0.552525 445 Learning Object 14
of the learning objects to the preferences of the student. The build model
8
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Fig. 11. The built (a) and training history (b) of the Deep Auto Encoders for Collaborative Filtering using Content Information (DAE-CI).
Table 9
Resulting metrics presented in the experiments.
Metric Popularity CDAE DAE-CF DAE-CI
9
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Fig. 12. The results of CDAE, DAE-CF, and DAE-CI related to the metrics NDCG@5, NDCG@10, Personalization, and Coverage.
Fig. 13. The results of Popularity Model, CDAE, DAE-CF, and DAE-CI to the metrics MAP, NDCG@5, NDCG@10, Personalization, Coverage, and SAUC. In (a) the
coverage of each built model. In (b) the comparison between metrics.
Fig. 14. The behaviors of the score related to objects (a) and frequency (b). It is important to note that the long tail phenomenon occur, also, in the score metric. This
an evidence that the resulting recommendation list is based on this phenomenon.
10
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Table 11
The preference signature of the student 10 with the firsts 100 learning objects (this table was filled by the auto encoder with the entry of table 10).
1.2781755 1.1916058 1.0409219 0.9275036 0.9187776 0.8970384 0.8880845 0.8861635 0.8650075 0.8581665
0.8520871 0.8513905 0.7809544 0.7697405 0.7473381 0.7446624 0.7296307 0.7205836 0.7175525 0.7169326
0.7160413 0.7109861 0.7085849 0.7049262 0.7049110 0.7047389 0.6942966 0.6936854 0.6934537 0.6927639
0.6918837 0.6788331 0.6788318 0.6752425 0.6682408 0.6679753 0.6619343 0.6592174 0.6525638 0.6443356
0.6440960 0.6437874 0.6431290 0.6398829 0.6353902 0.6314064 0.6264353 0.6218643 0.6126253 0.6077056
0.6044728 0.6026431 0.6014149 0.5958386 0.5949369 0.5876704 0.5841696 0.5802777 0.5798503 0.5780840
0.5769091 0.5725408 0.5686121 0.5669655 0.5669616 0.5644537 0.5626519 0.5621841 0.5579706 0.5566605
0.5543797 0.5534146 0.5504782 0.5466366 0.5464506 0.5463822 0.5435224 0.5432371 0.5431470 0.5417905
0.5412520 0.5399067 0.5347409 0.5340176 0.5332670 0.5327663 0.5324275 0.5322449 0.5320356 0.5290830
0.5289596 0.5277051 0.5271041 0.5255498 0.5249646 0.5245923 0.5244858 0.5196915 0.5168995 0.5164964
Table 12
The comparison of the previous works. Many of them present endogenous metrics.
Author Year Summary Method Metrics
Bobadilla et al. (2009) 2009 Memory-based method to calculate the proximity of content Memory-based MAE ¼ 2%
of learning objects using KNN
Hwang et al., (2013) 2013 Content-based method to context-aware recommendation of Content-based Specific metric proposed by authors
learning objects
Krauss (2016) 2016 An approach to extend the traditional user-item-matrix of a Cosine similarity Without explicit metrics
recommendation engine by a third dimension – the time
value
Protasiewicz et al. 2016 The developed system helps to collect data characterizing Cosine similarity Precision ¼ 97.96%
(2016) potential reviewers, retrieving information from relational Recall ¼ 54.27%
and unstructured data, and formulating a set of F-score ¼ 69.84%
recommendations
Barbieri et al. (2017) 2017 The authors propose switching SVD to a Stacked Denoising Stacked Denoising Auto encoder MAE and RMSE ¼ 70% in the best case
Auto encoder to avoid the noise of users’ evaluations
Nafea et al. (2019) 2019 The authors combine students actual rating with their Cosine similarity MAE ¼ 22% and RMSE ¼ 38% in the
learning styles to recommend personalized course learning best case
objects (LOs)
Medio et al. (2020) 2020 This paper proposes a hybrid recommender system Content filtering and Collaborative Precision ¼ 22%
filtering Recall ¼ 39%
F1 ¼ 28%
Our 2020 Three classes of deep auto encoder to fill out a sparse matrix Deep Auto Encoder MAE ¼ 81% in the best case
composed by the interactions of student/object in a MOOC RMSE ¼ 86% in the best case
each student. This signature is the output of Deep Auto Encoder. For presented an improvement related to Barbieri (Barbieri et al., 2017) of
example, the vector of all LOs from student 10 which is 0.0 is replaced by around 11% and 15%, respectively. We did not use the metrics Precision,
the probability of that student consuming that LO. This preference Recall, and F-Score because the target output is continuous.
signature, to the firsts 100 learning objects, is shown in Table 11.
Therefore, the use of Deep Auto Encoder (DEA-CF/CI) can represent
an approach to build adaptive e-learning recommendation system 4.6. Educational implications
models. This class of artificial neural networks allows building high
personalized models to recommend relevant objects, covering many From a pragmatic point of view, our approach helps to improve the
objects into the repository, and providing a preference signature to each outcomes of students. The reasons for this improvement can be attributed
student based on his/her previous interactions. Moreover, this class of to some features of our approach. The first feature is relevance; as dis-
architecture is relevant to overcome the challenges to provide a recom- cussed, an LO needs to be relevant or it is useless. Based on students’
mendation list, including objects with relevance (supported by the met- signature of preference it is possible to present relevant LOs provided by
rics MAP and NDCG), and also recommending objects with novelty, LORs and with curricular restrictions (for example, in a lesson A the LOa
serendipity, and diversity (supported by the metric Coverage and SAUC). and LOb are mandatory but LOc and LOd are elective. Besides, the top@k
In conclusion, this class of architectures improves the adaptability, of the LOs can be used to control the volume of content to be assimilated
providing a preference signature for each student (supported by by students. The second feature is novelty; an LO needs to be new or not
Personalization). used in the past. Using the sparse matrix is possible to identify what LOs
Our results can be compared with previous works. Many of them are were used before and to recommend a new one. However, to present
focused on endogenous metrics like Precision, Recall, F-Score, and RMSE. previous LO might be a way to improve cognition by repetition and this
We proposed not only endogenous, but also metrics to provide evaluation can be controlled by our approach. Third one, serendipity; LO, which is,
of more effective models. Mean Absolute Error (MAE) and Root Mean not expect, but relevant, to the student. This can be done to keep the
Squared Error (RMSE) are two of the most common metrics used to focus because something new might be present at any time. The last
measure accuracy for continuous variables, thus, we used them to feature is diversity; diverse uses types of LO for improving the coverage
compare with other authors. The comparison of previous works is present of the subject to be learned. This can be used to increment the knowl-
in Table 12. edge, widely and deeper, from a subject (Hwang et al., 2013; Krauss,
These results are shown in a compiled way, in Fig. 15. Broadly 2016; Protasiewicz et al., 2016; Barbieri et al., 2017; Nafea et al., 2019;
speaking, the metrics are sparse, i.e., used by one work but not done by Medio et al., 2020; Gomede et al., 2020; Sedhain et al., 2015; Strub et al.,
another. The best results, to Precision, Recall, and F-Score were achieved 2016). Also, the learning objects can be standardized by use of Shareable
by Protasiewicz et al. (2016). Content Object Reference Model (SCORM) and provided by a Learning
As shown in Fig. 15, our results, based on metrics MAE and RMSE, Objects Repository (LOR). The learning objects can be created by pro-
fessors (or other professionals) and deployed into a LOR, then, our
11
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
Fig. 15. The comparison of the previous works with our results. Many of them present endogenous metrics.
approach can retrieve them from many LORs. problems in Computational Intelligence. Barros main research fields are
in Governance, Strategic Planning and applied Computational Intelli-
5. Conclusions, practical impacts, and further works gence. Everton Gomede’s main research field is applied Computational
Intelligence for Smart Cities. His main contribution for the paper was in
This work aims at contributing with an adaptive e-learning recom- the processing of data with computational intelligence techniques and
mendation system with which we investigated four approaches to mathematical models. All participants gave important contributions to
develop models. The first one, Popularity Model, represents the base of the formulation of the hypotheses and the definition of the data analytics
comparison with the other ones. The models designed by Auto Encoders for the educational dataset. Based on The CRediT Roles, the contributions
were compared using relevant metrics and with the hyperparameters are Gomede: Conceptualization, Data curation, Formal analysis, Investi-
optimized. The class of Deep Auto Encoders shows better results using gation, Methodology, Software, and Writing – original draft; Barros and
deep layers. The Auto Encoders for Collaborative Filtering (DAE-CF) Mendes: Project administration, Resources, Supervision, Validation,
presented better personalization and coverage of learning objects. Writing – review & editing.
Finally, Deep Auto Encoders for Collaborative Filtering using Content
Information (DAE-CI) presented better recommendation lists of learning Statements on open data and ethics
objects. All these models can be applied to the recommendation and
adaptability problem and to build the preference signature of each The participants were protected by hiding their personal information
student. in this study. They were voluntary and they knew that they could with-
As practical impacts, we may highlight the use of Deep Auto Encoders draw from the experiment at any time. Also, the name of learning objects
to build adaptive e-learning recommendation systems to improve the was suppressed to avoid issues related to privacy (GDPR). The data can
experience and the outcomes of students. This may increase the results of be provided upon requests by sending e-mails to the corresponding
the educational environment, helping teachers, tutors, administrators, author.
researchers, and/or parents (Gomede, Gaffo, Brigan o, Barros, & Mendes,
2018). This approach enables teachers and/or tutors to better recognize Declaration of competing interest
their students and provide more accurate content in an adapted format. It
enables administrators to improve systems management to get high The authors declare that they have no known competing financial
grades for their institutions. Also, it enables researchers to obtain more interests or personal relationships that could have appeared to influence
relevant content for their research and avoid information flooding. the work reported in this paper.
Finally, it enables parents to follow their children’s preferences and help
them achieve better results.
Appendix A. Supplementary data
In conclusion, it is possible to explore additional recommendation
approaches for other educational agents. To teachers, we can search for
Supplementary data to this article can be found online at https://
new adaptive ways to deeply present their subject matter to their stu-
doi.org/10.1016/j.caeai.2021.100009.
dents; to students, we can look for new adaptive forms to have problems
with increased difficult levels presented to them; to researchers, we can
References
look for new open and relevant problems; and to parents, we can search
to present forms to help them follow their children and participate in Baldi, P. (2011). Autoencoders, unsupervised learning, and deep architectures. In
leading them to a more thorough experience with the activity of learning. UTLW’11: Proceedings of the 2011 International Conference on Unsupervised and Transfer
In terms of engineering, we might explore the use of Learning Objects Learning workshop, 27 pp. 37–50).
Barbieri, J., Alvim, L. G. M., Braida, F., & Zimbr~ao, G. (2017). Autoencoders and
Repositories to share many Learning Objects from several subjects, levels, recommender systems: COFILS approach. Expert Syst. Appl., 89, 81–90. https://
idioms, cultures, and so on to improve the quality of recommendations. doi.org/10.1016/j.eswa.2017.07.030
Bobadilla, J., Serradilla, F., & Hernando, A. (2009). Collaborative filtering adapted to
recommender systems of e-learning. Knowl. Base Syst., 22(4), 261–265. https://
Author contributions doi.org/10.1016/j.knosys.2009.01.008
Chae, D. K., Kim, S. W., & Lee, J. T. (2019). Autoencoder-based personalized ranking
framework unifying explicit and implicit feedback for accurate top-N
This paper was developed with the active contribution of the
recommendation. Knowl. Base Syst., 176, 110–121. https://doi.org/10.1016/
doctorate candidate Everton Gomede, the orientation of Leonardo j.knosys.2019.03.026
Mendes, and Rodolfo Barros. Mendes and Barros are both advisors of the Denis, K., Shuaiqiang, W., & Jari, V. (2016). A survey of serendipity in recommender
candidate Gomede, and besides heading the research group, are the main systems. Knowl. Base Syst., 111, 180–192. https://doi.org/10.1016/
j.knosys.2016.08.014
responsible for the propositions and hypotheses being proposed and Dockterman, D. (2018). Insights from 200þ years of personalized learning. NPJ Sci.
tested. Mendes research areas are in Intelligent Cities and applied Learn., 3(1), 1–6. https://doi.org/10.1038/s41539-018-0033-x
12
E. Gomede et al. Computers and Education: Artificial Intelligence 2 (2021) 100009
G€obel, S., & Mehm, F. (2013). Personalized, adaptive digital educational games using Ordway-West, E., Parveen, P., & Henslee, A. (2018). Autoencoder evaluation and hyper-
narrative game-based learning objects. Entertainment for Education Digital Techniques and parameter tuning in an unsupervised setting. In 2018 IEEE International Congress on
Systems-Lecture Notes in Computer Science, 6249. https://doi.org/10.1007/978-3-642- Big Data (Big Data Congress) (pp. 205–209). https://doi.org/10.1109/
14533-9_45 BigDataCongress.2018.00034
Gomede, E., Gaffo, F. H., Brigan o, G. U., Barros, R. M., & Mendes, L. D. S. (2018). Protasiewicz, J., Pedrycz, W., Kozłowski, M., Dadas, S., Stanisławek, T., Kopacz, A., &
Application of computational intelligence to improve education in smart cities. Gałe,zewska, M. (2016). A recommender system of reviewers and experts in
Sensors, 18(1), 267. https://doi.org/10.3390/s18010267 reviewing problems. Knowl. Base Syst., 106, 164–178. https://doi.org/10.1016/
Gomede, E., Barros, R. M., & Mendes, L. S. M. (2020). Use of deep multi-target prediction j.knosys.2016.05.041
to identify learning styles. Appl. Sci., 10(5), 1756. https://doi.org/10.3390/ Sedhain, S., Menon, A. K., Sanner, S., & Xie, L. (2015). Autorec: autoencoders meet
app10051756 collaborative filtering. In WWW ’15 Companion: Proceedings of the 24th International
Hwang, G., Yang, L., & Wang, S. (2013). A concept map-embedded educational computer Conference on World Wide Web (pp. 111–112). https://doi.org/10.1145/
game for improving students’ learning performance in natural science courses. 2740908.2742726
Comput. Educ., 69, 121–130. https://doi.org/10.1016/j.compedu.2013.07.008 Shih, J. L., Chuang, C. W., & Hwang, G. J. (2010). An inquiry-based mobile learning
Krauss, C. (2016). Smart learning: time-dependent context-aware learning object approach to enhancing social science learning effectiveness. Journal of Educational
recommendations. In The 29th AAAI FLAIRS Conference of the Florida Artificial Technology & Society, 13(4), 50–62.
Intelligence Research Society, 5 pp. 501–504). Shih, J. L., Chuang, C. W., & Hwang, G. J. (2010). An inquiry-based mobile learning
Li, X., & She, J. (2017). Collaborative variational autoencoder for recommender systems. approach to enhancing social science learning effectiveness. J. Educ. Technol. Soc.,
In KDD ‘17: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge 13(4), 50–62.
Discovery and Data Mining (pp. 305–314). https://doi.org/10.1145/ Strub, F., Gaudel, R., & Mary, J. (2016). Hybrid recommender system based on
3097983.3098077 autoencoders. In Proceedings of the 1st Workshop on Deep Learning for Recommender
Medio, C., Limongelli, C., Sciarrone, F., & Temperini, M. (2020). MoodleREC: A Systems (pp. 11–16). https://doi.org/10.1145/2988450.2988456
recommendation system for creating courses using the moodle e-learning platform. Wang, H., Wang, N., & Yeung, D. Y. (2015). Collaborative deep learning for recommender
Computers in Human Behavior, 104, 106168 https://doi.org/10.1016/j.chb.2019.10 systems. In KDD ‘15: Proceedings of the 21th ACM SIGKDD International Conference on
6168. Knowledge Discovery and Data Mining (pp. 1235–1244). https://doi.org/10.1145/
Nafea, S. M., Siewe, F., & He, Y. (2019). On recommendation of learning objects using 2783258.2783273
Felder-Silverman Learning Style Model. IEEE Access, 7, 163034–163048. https://
doi.org/10.1109/ACCESS.2019.2935417
13