Embracing Big Data in Complex Educational Systems: The Learning Analytics Imperative and The Policy Challenge
Embracing Big Data in Complex Educational Systems: The Learning Analytics Imperative and The Policy Challenge
Embracing Big Data in Complex Educational Systems: The Learning Analytics Imperative and The Policy Challenge
In the new era of big educational data, learning analytics (LA) offer the
possibility of implementing real–time assessment and feedback systems
and processes at scale that are focused on improvement of learning,
development of self–regulated learning skills, and student success. How-
ever, to realize this promise, the necessary shifts in the culture, techno-
logical infrastructure, and teaching practices of higher education, from
assessment–for–accountability to assessment–for–learning, cannot be
achieved through piecemeal implementation of new tools. We propose
here that the challenge of successful institutional change for learning
analytics implementation is a wicked problem that calls for new adap-
tive forms of leadership, collaboration, policy development and strategic
planning. Higher education institutions are best viewed as complex sys-
tems underpinned by policy, and we introduce two policy and planning
AUTHORS
frameworks developed for complex systems that may offer institutional
Leah P. Macfadyen
teams practical guidance in their project of optimizing their educational
The University of
systems with learning analytics.
British Columbia
Shane Dawson
University of South Australia
Abelardo Pardo
The University of Sydney
Dragan Gašević
Athabasca University
In education, we are awash in data about our learners and educators, our
technologies and activities, achievements and performance. To date these data have
rarely been mined intelligently with the goal of improving learning and informing teaching
practice, although evidence from other sectors such as marketing, sports, retail, health
and technology suggests that the effective use of big data can offer the education sector
the potential to enhance its systems and outcomes (Manyika et al., 2011). Norris and Baer
(2013) have noted that, “Data expands the capacity and ability of organizations to make
sense of complex environments” (p. 13). In this context, learning analytics (LA) offers
the capacity to investigate the rising tide of learner data with the goal of understanding
the activities and behaviors associated with effective learning, and to leverage this
knowledge in optimizing our educational systems (Bienkowski, Feng, & Means, 2012;
Campbell, DeBlois, & Oblinger, 2007). Indeed, in a world of larger and larger data sets,
increasing populations of increasingly diverse learners, constrained education budgets
and greater focus on quality and accountability (Macfadyen & Dawson, 2012), some argue
CORRESPONDENCE that using analytics to optimize learning environments is no longer an option but an
imperative. The value of such analytics is highlighted by the authors of the McKinsey
Email Global Institute (Manyika et al., 2011) noting that, “In a big data world, a competitor that
[email protected] fails to sufficiently develop its capabilities will be left behind…Early movers that secure
access to the data necessary to create value are likely to reap the most benefit” (p. 6).
Education can no longer afford not to use learning analytics. As Slade and Prinsloo (2013)
maintain, “Ignoring information that might actively help to pursue an institution’s goals
seems shortsighted to the extreme” (p. 1521).
Differentiating assess-
ment–for–learning from Provision (to learners and educators) of automated analytics that provide feedback
assessment–for–account- on learner study behaviors, progress and outcomes will not only enhance academic
ability within the educa- performance but also develop student self–regulated learning (SRL) skills, and SRL
tional system forms part proficiency has been demonstrated to be a significant predictor of academic success
of the wicked problem (e.g., Butler & Winne, 1995; Pintrich, 1999; Zimmerman, 2002). Student motivation
of institutional change in
and capacity to undertake accurate self–monitoring had a direct impact on the level
higher education that we
and quality of their study and therefore, their overall learning progression and academic
seek to explore here.
achievement (Dunlosky & Thiede, 1998). Conversely, poor performers are poor at
evaluating their own ability or judging their own learning skills (Kruger & Dunning,
1999). For these reasons, it is argued that a core goal of any effective pedagogical
strategy must include the development of student meta–cognitive skills or judgment of
(own) learning (JOL). Feedback on assessment is one key approach that is often adopted
to assist students in developing meta–cognitive skills, but because provision of relevant
feedback can be labor–intensive, it is often delayed and provided at a time when it is no
longer useful to the student to aid their learning.
Recent research posits that SRL is a process of temporal events that evolve during
learning (Azevedo & Aleven, 2013). This research, alongside recent developments in
learning analytics, data mining and machine learning is providing new methods for
developing novel insights into student learning processes. Historically, assessment
and development of student SRL has made use of tasks associated with JOL which
generally involve asking a student to assess how effectively they have understood a
particular concept (Dunlosky & Lipko, 2007). This self–reported rating is then correlated
against their overall test performance to gain insight into the student’s meta–cognitive
proficiency. While JOL has commonly relied on self–report methodologies such as think
aloud protocols and surveys, these have inherent limitations such as poor recall, and
biased responses (Richardson, 2004).
New options for assessing student learning behaviors are emerging as a result of advances
in learning analytics and natural language processing (NLP), and alternate models have
sought to capture actual learner behavior (in lieu of self–reports) from interactions with
technology–based learning activities. For example, oft–cited SRL researcher Phil Winne
has previously reported that student online interaction data can provide significant
indicators of SRL proficiency (e.g., Winne, 2010; Zhou & Winne, 2012). Winne has
developed the software application nStudy as a web tool that can collect very fine
grained, time stamped data about individual learner interactions with online study
materials. The trace data is then used to provide insight and feedback into the learner’s
cognitive choices as they interact with the online information. Essentially, the tool
makes data for reflection available to both the individual learner and the educator.
comprehensive formative and summative assessment translates into a rich set of requirements
of the current technological infrastructures. Although learning management systems (LMSs)
still host a large percentage of technology–mediated educational activities, educational
institutions are recognizing the need to re–assess the concept of teaching and learning space
to encompass both physical and virtual locations, and adapt learning experiences to this new
context (Thomas, 2010). Thus, together with the need for cultural change and a focus on
pedagogical relevance, an additional sociotechnical factor critical to the adoption of learning
analytics is technology itself (Box 2).
The evolution of technology in recent years offers an unprecedented capacity to store
large data sets, and applications using big data are well established in contexts such as business
intelligence, marketing and scientific research (Dillon, Wu, & Chang, 2010). Education faces a
particular challenge that derives from the rich variety of technological affordances emerging in
20 Volume Nine | Winter 2014
learning environments. From an LMS–centric approach consolidated in the early 2000s, we are
now entering an era in which learning may occur anywhere, at any time, with multiple devices,
over a highly heterogeneous collection of resources, and through multiple types of interactions. In
this new scenario, learning analytics tools need to comply with requirements in the following areas:
1. Diverse and flexible data collection schemes: Tools need to adapt to increasing
data sources, distributed in location, different in scope, and hosted in any platform.
2. Simple connection with institutional objectives at different levels:
information needs to be understood by stakeholders with no extra effort. Upper
management needs insight connected with different organizational aspects
than an educator. User–guided design is of the utmost importance in this area.
3. Simple deployment of effective interventions, and an integrated and
sustained overall refinement procedure allowing reflection.
From the technological point of view, learning analytics is an emerging discipline
(Siemens, 2013) and its connection with assessment remains largely unexplored (Ellis, 2013).
This situation is even more extreme when considering the assessment of competences and
learning dispositions (Buckingham Shum, 2012). Educational institutions need technological
Box 2
Sociotechnical Infrastructure Needs for Effective Learning Analytics
It is unrealistic to
Several initiatives are already tackling the problem of flexible data collection schemes. consider that educators
For example the ADL Experience API3 released in 2013 has been proposed as a solution will adopt time–
that can promote interoperability between data collected in different environments consuming longitudinal
and platforms. The interface offers the possibility of capturing a wide variety of and personalized
assessment models given
events in experiences with heterogeneous scenarios (Glahn, 2013). Similarly, the
the massive increase in
IMS Global Consortium has proposed that the Learning Measurement Framework
resources and workload
IMS Caliper4 – containing descriptions to represent metrics, sensor API and learning
that would be required.
events – will facilitate the representation and processing of big data in the learning
field. In parallel, the concept of a Learning Record Store (LRS) has been proposed as
a framework for storing and manipulating data from distributed events in a learning
environment, encoding not only interaction among stakeholders, but among resources.
This information is then made available through a service–based interface to other
systems within an institution (or across multiple institutions) for further analysis and
processing.
Numerous attempts have been made to meet diverse stakeholder reporting and data
access needs by production of so–called dashboards that show a canvas of multiple
visualizations. Common limitation of these graphical representations, however, are their
actual utility and usability (Verbert, Duval, Klerkx, Govaerts, & Santos, 2013). Adapting
presentation of information to user context, needs and interests is another important
factor that must be taken into account if we wish to facilitate the uptake of learning
analytics solutions.
The third requirement for technology supporting learning analytics is that it can
facilitate the deployment of so–called interventions, where intervention may mean any
change or personalization introduced in the environment to support student success,
and its relevance with respect to the context. This context may range from generic
institutional policies, to pedagogical strategy in a course. Interventions at the level of
institution have been already studied and deployed to address retention, attrition or
graduation rate problems (Ferguson, 2012; Fritz, 2011; Tanes, Arnold, King, & Remnet,
2011). More comprehensive frameworks that widen the scope of interventions and
adopt a more formal approach have been recently proposed, but much research is still
needed in this area (Wise, 2014).
3
http://www.adlnet.gov/tla
4
http:/www.imsglobal.org/IMSLearningAnalyticsWP.pdf Volume Nine | Winter 2014 21
solutions that are deployed in a context of continuous change, with an increasing variety of data
sources, that convey the advantages in a simple way to stakeholders, and allow a connection
with the underpinning pedagogical strategies.
In turn, these technological requirements point to a number of other critical contextual
factors that must form part of any meaningful policy and planning framework for employing
learning analytics in service of improved assessment. Foremost among these is the question
of access to data, which needs must be widespread and open. Careful policy development is
also necessary to ensure that assessment and analytics plans reflect the institution’s vision
for teaching and strategic needs (and are not simply being embraced in a panic to be seen to
be doing something with data), and that LA tools and approaches are embraced as a means of
engaging stakeholders in discussion and facilitating change rather than as tools for measuring
performance or the status quo.
Figure 2. Cause–effect (DPSEEA) framework for institutional assessment and technology policies
(modified from Corvalan et al., 1999).
Second, Corvalán et al.’s (1999) “cause–effect framework” (or DPSEEA framework) usefully
assists in identifying the multiple linkages that may exist between the driving forces underpinning It may not be surprising,
complex systems, illuminating the multiple points in a complex system of relationships where then, that globally,
action may be needed to effect change. Such a framework can, they suggest, “be used to weigh education lags behind
alternatives and to design step–by–step programs for dealing with a particular…problem” (p. 659). all other sectors in
Figure 2 offers a preliminary modification of this framework to represent institutional effects of, harnessing the power of
for example, technology and assessment policies, and may be a useful context mapping tool in the analytics. A preliminary
ROMA process. analysis indicates that
educational institutions
Use of these models for institutional LA policy development is only in the very early
simply lack the practical,
stages, although we have explored elsewhere (Ferguson et al., in press) the ways in which a small
technical and financial
number of apparently successful institutional LA policy and planning processes have pursued
capacity to effectively
change management approaches that map well to such frameworks. In future work, we hope to
gather, manage and mine
be able to present more robust and critical review of real–time application of these frameworks in
big data.
institutional planning, and their possible effectiveness or limitations.
In the meantime, readers may review both frameworks and immediately dispute the
stages, levels, linkages, effects or impacts in relation to their own institutional context. But this is, of
course, the very point of such adaptive models, which can and should be disputed, negotiated and
modified as needed for local institutional contexts, to guide relevant local action. To paraphrase
Head and Alford (2013), when it comes to wicked problems in complex systems, there is no one–
size–fits–all policy solution, and there is no plan that is not provisional.
Rather, the more important role of such frameworks is to continuously remind us of the
need for a holistic understanding of institutional context if the goal is institutional change, including
external and internal influences, political and cultural context, the evidence itself, and the links:
Volume Nine | Winter 2014 25
“All of the other actors and mechanisms that affect how the evidence gets into the policy process” (Young
& Mendizabal, 2009). They can assist in identifying points of intervention (Corvalán et al., 1999) in the
complex adaptive system that is education, to offer leaders and practitioners additional insight and tools in
their project of optimizing the system with learning analytics.
References
Ackoff, R. L. (1974). Redesigning the future. New York, NY: Wiley.
Angelo, T. A. (1999). Doing assessment as if learning matters most. AAHE Bulletin, 51(9), 3–6.
Azevedo, R., & Aleven, V. (Eds.). (2013). International handbook of metacognition and learning technologies.
Amsterdam: Springer.
Berkhout, F., Leach, M., & Scoones, I. (Eds.). (2003). Negotiating environmental change: New perspectives from social
science. Cheltenham, UK: Edward Elgar.
Bichsel, J. (2012). Analytics in Higher Education: Benefits, Barriers, Progress, and Recommendations (Research
Report). Louisville, CO: EDUCAUSE Center for Applied Research. http://net.educause.edu/ir/library/pdf/
ERS1207/ers1207.pdf
Bienkowski, M., Feng, M., & Means, B. (2012). Enhancing teaching and learning through educational data mining
and learning analytics: An issue brief. Washington, DC: U.S. Department of Education Office of Educational
Technology. http://www.ed.gov/edblogs/technology/files/2012/03/edm–la–brief.pdf
Buckingham Shum, S. (2012). Learning dispositions and transferable competencies: Pedagogy, modelling and learning
analytics. In S. Buckingham Shum, D. Gašević, & R. Ferguson (Eds.), International conference on learning
analytics and knowledge (pp. 92–101). New York, NY: ACM Press.
Butler, D. L., & Winne, P. H. (1995). Feedback and self–regulated learning: A theoretical synthesis. Review of
Educational Research, 65(3), 245–281.
Campbell, J. P., DeBlois, P. B., & Oblinger, D. G. (2007). Academic analytics: A new tool for a new era. EDUCAUSE
Review, 42(4), 42–57.
Capra, F. (1996). The web of life. New York, NY: Doubleday.
Churchman, C. W. (1967). Free for all. Management Science, 14, B141–B142.
Cilliers, P. (1998). Complexity and postmodernism: Understanding complex systems. London, UK: Routledge.
Cohen, M. D., & Marsh, J. G. (1986). Leadership and ambiguity: The American college president. New York, NY:
McGraw Hill.
Corvalán, C. F., Kjellström, T., & Smith, K. R. (1999). Health, environment and sustainable development: Identifying
links and indicators to promote action. Epidemiology, 10(5), 656–660.
Dillon, T., Wu, C., & Chang, E. (2010). Cloud computing: Issues and challenges. In IEEE International Conference on
Advanced Information Networking and Applications (pp. 27–33). New York, NY: IEEE Press.
Dumont, H., Istance, D., & Benavides, F. (Eds.). (2010). The nature of learning: Using research to inspire
practice. Educational Research and Innovation series, OECD Publishing. http://www.oecd.org/edu/ceri/
thenatureoflearningusingresearchtoinspirepractice.htm
Dunlosky, J., & Lipko, A. R. (2007). Metacomprehension: A brief history and how to improve its accuracy. Current
Directions in Psychological Science, 16(4), 228–232.
Dunlosky, J., & Thiede, K. W. (1998). What makes people study more? An evaluation of factors that affect self–paced
study. Acta Psychologica, 98(1), 37–56.
Dworkin, G. (2005). The No Child Left Behind act: Accountability, high–stakes testing and roles for sociologists.
Sociology of Education, 78(2), 170–174.