Book - Analytics For Instructional Designers Association For Talent Development
Book - Analytics For Instructional Designers Association For Talent Development
Book - Analytics For Instructional Designers Association For Talent Development
26 25 24 23 12345
No part of this publication may be reproduced, distributed, or transmitted in any form or by any
means, including photocopying, recording, information storage and retrieval systems, or other
electronic or mechanical methods, without the prior written permission of the publisher, except in the
case of brief quotations embodied in critical reviews and certain other noncommercial uses permitted
by copyright law. For permission requests, please go to copyright.com, or contact Copyright
Clearance Center (CCC), 222 Rosewood Drive, Danvers, MA 01923 (telephone: 978.750.8400; fax:
978.646.8600).
ATD Press is an internationally renowned source of insightful and practical information on talent
development, training, and professional development.
ATD Press
1640 King Street
Alexandria, VA 22314 USA
Ordering information: Books published by ATD Press can be purchased by visiting ATD’s website at
td.org/books or by calling 800.628.2783 or 703.683.8100.
ISBN-10: 1953946445
ISBN-13: 9781953946447
e-ISBN: 9781953946454
Emerging cognitive and neuroscience research finds that math anxiety is not
just a response to poor math performance—in fact, four out of five students
with math anxiety are average-to-high math performers. Rather, math anxiety is
linked to higher activity in areas of the brain that relate to fear of failure before
a math task, not during it. This fear takes up mental bandwidth during a math
task…. In turn, that discomfort tends to make those with math anxiety more
reluctant to practice math, which then erodes confidence and skill. In part for
that reason, anxiety has been linked to worse long-term performance in math
than in other academic subjects like reading.
After we lay down the foundations of learning data and analytics, we will
start to take a look at the process for actually getting and using data in part
2. First, we’ll talk about making a plan for what kinds of data you will
gather, including aligning with organizational metrics and many of the
common learning and development frameworks that we use for analysis
(chapters 6 and 7).
Next, we’ll dive into forming your hypothesis from the questions you
need to answer with data (chapter 8). We’ll then take a look at actually
identifying the data needs that will serve those purposes (chapter 9),
building the data capture into our learning experiences so we actually get
the data we need (chapter 10), and collecting and storing it (chapter 11).
At this point many people will arrive at what I have in my own projects
referred to as the moment of “Oh my gosh my baby is ugly!” This is where
you have collected some data, made some analysis, and realized that what
you really wanted to answer was something other than what you just did.
Here’s where the fun begins as you iterate on the learning and data
experience by looking at what you have gathered, and then fine-tuning it
(chapter 12).
Many people conflate the visualization of data with the analysis of data.
And up until this point, we haven’t talked about visualizing data at all!
We’ll spend a little bit of time talking about how we communicate and
visualize data in chapter 13. This is another one of those places for which
dozens upon dozens of wonderful resources exist, so this book will cover it
on just a very high level.
Finally, we’ll take a look at what it means to scale up your analytics
efforts, moving from one or two pilot projects to a future enterprise-wide
state. There are few organizations at this stage as of the writing of this book,
so the future is full of opportunity for you to define it.
How This Book Will Help You
This book takes an “If I can see it, I can be it” approach to learning data and
analytics. In my work helping organizations adopt xAPI, I am frequently
asked for case studies. The questions sound like “Who’s really doing it?”
“How does that actually work?” and “That sounds great, but can you share
an example so I really know I’m getting it?”
My emphasis in this book will be not only on practical what-is and how-
to content, but also real-world examples and longer case studies from
practitioners. In some cases, I’m telling the story. In other cases, the people
who have built it and lived with it share their story in their own words.
Each chapter will conclude with opportunities for you to put these
techniques to work right away, whether you are in a data-rich environment
already, or are just getting started and working on hypotheticals. These
opportunities to give the concepts a try are a valuable part of extending your
learning beyond the pages of this book and into the real world all around
you. If you are learning with your team, these activities can be done in pairs
or in small groups for shared learning impact.
As much as I would love to offer you a book with immediate practical
value to your own work, it’s entirely possible that you don’t yet have the
data necessary to apply these concepts right away. As such, the “give it a
try” ideas at the end of most chapters include reflections and hypotheticals
to let you dig in right away, even though they might not reach your loftiest
aspirations just yet.
And while I aim to be definitive whenever possible, remember there are
very few hard and fast rules. Simply, a lot of it depends. So, at the risk of
sounding like I’m unable to make a firm decision in offering advice, I find
that the very interesting questions in life often have multiple right answers.
The “rightness” depends on your situation, your needs and capabilities,
what you have access to right now, and what your leaders deem makes
sense. And, some of the most complex “right” answers change over time.
Let me also note that this isn’t a book about xAPI. While I believe that
the widespread adoption of a rich and interoperable data specification is
good for the industry, individual professionals, and organizations buying
products and services, I also realize that xAPI is not the only way to work
with learning and performance data.
Whether you’re using xAPI that far extends the capabilities of SCORM
to track learning, practice, and performance activity or another data model,
we have the ability to get our hands on far more plentiful, granular, and
interesting types of data. That’s what this book is about: what data to get,
how to get it, and what to do with it once you have it.
So, let’s get some data!
PART 1
THE FOUNDATIONS
Chapter 1
Sales Metrics
• Annual recurring revenue
• Average revenue per user
• Quota attainment
• Win rate
• Market penetration
• Percentage of revenue from new versus existing customers
• Lifetime value (LTV) of a customer
• Average profit margin
• Conversion rate
• Sales cycle length
• Average deal size
• Year-over-year growth
• Deal slippage
Finance Metrics
• Earnings before interest and taxes (EBIT)
• Economic value added (EVA)
• Current ratio
• Working capital
• Debt-to-equity ratio
• Contribution margin
• Customer satisfaction
• Liquidity ratio
• Return on equity
• Days in accounts receivables
• Net cash flow
• Gross profit margin
• Transactions error rate
Healthcare Metrics
• Number of medication errors
• Complication rate percentage
• Leaving against medical advice
• Post-procedure death rate
• Readmission rate: hospital acquired conditions (HACs)
• Average minutes per surgery
• Average length of stay
• Patient wait times by process step
• Doctor–patient communication frequency
• Overall patient satisfaction
• Patient-to-staff ratio
• Occupancy rate
Hospitality Metrics
• Energy management
• Labor costs as percent of sales
• Employee performance
• Gross operating profit
• Occupancy rate
• Average daily rate (ADR)
• Average room rate (ARR)
• Revenue per available room (RevPAR)
• Net revenue per available room (NRevPAR)
• Revenue per occupied room (RevPOR)
• Gross operating profit per available room
• Marketing ROI
• Online rating
• Customer satisfaction
• Loyalty programs
Manufacturing Metrics
• On-time delivery
• Production schedule attainment
• Total cycle time
• Throughput
• Capacity utilization
• Changeover time
• Yield
• Scrap
• Planned maintenance percentage (PMP)
• Uptime/uptime + downtime
• Customer return rate
• Overall equipment effectiveness (OEE)
Childcare Metrics
• Child abuse indicator
• Immunization indicator
• Staff–child ratio and group size indicator
• Staff training indicator
• Supervision or discipline indicator
• Fire drills indicator
• Medication indicator
• Emergency plan or contact indicator
• Outdoor playground indicator
• Toxic substances indicator
• Handwashing or diapering indicator
Technology Metrics
• Defect density
• First response time
• Function points
• Incidents
• Information security scores
• IT overhead
• IT risk score
• IT security training rates
• Noncompliance events
• Patch rate
• Planned downtime rate
• Project cycle time
• Security overhead
Give It a Try
Here’s a Hypothetical
Leverage the consumerization of data-driven apps to examine the use of
data in your daily life. Choose a piece of software or an application that
you’re familiar with and see how many of these questions you can answer:
• What data is being gathered?
• How is the data gathered and stored?
• What kinds of insights might the company being making from the
data?
• What kind of data is not being gathered but might also be useful?
• What do I, as a user, get from this data?
Do It for Real
It’s entirely likely that if you are reading this book, you are also on the
lookout for experts, articles, podcasts, and conference sessions on data
analytics in L&D. As you do so, take a structured approach to your
learning. Here are some questions you can ask about the case studies you
discover on your hunt (including the ones in this book):
• What data is being gathered?
• How is the data gathered and stored?
• What kinds of insights were derived from the data?
• What kind of data was not gathered but might also have informed
this decision?
• What do the learners get out of this?
• What is the next iteration of this work?
Bonus Points
Reach out to someone in your organization (outside the L&D function) and
ask them how data is used in their work. Use as many of the questions from
the two lists here as are relevant to your conversation.
Using Analytics to Improve Video Design
By Josh Cavalier, Owner, JoshCavalier.com
What Is Data?
The terms data, information, and sometimes even knowledge are frequently
interchanged and misunderstood, but in most casual conversation it doesn’t
really matter. For the purposes of this book, however, we’ll want to be a bit
more specific.
Data
Data consists of symbols and characters that are an abstract representation
of something that exists(ed) or happens(ed) in the world. Without context,
data is meaningless—it is simply a representation of facts, without context,
meaning, or judgment. You could think of a list in a spreadsheet as data.
Here are some examples:
• It rained.
• I am 68 inches.
• Beverly completed a course.
• 94 percent
• 8,287 pounds
When you collect and measure observations, you create data. Since you
can’t collect and measure everything about everything, data is an
incomplete and imperfect representation of what you’re interested in. As
such, data can be remarkably disappointing. There are also plenty of
opportunities for people to poke holes in your analysis if your first step—
your data collection—doesn’t adequately measure what you’re interested in.
(See chapter 9 for more information on identifying data sources.)
Information
Information adds context and meaning to data. It’s the who, what, where, or
when about a particular data point. (Take note: It is not the why.)
Information generally offers some sort of relationship, which may or may
not be true. It can require many data points to create the context that’s
needed. If data is a list in a spreadsheet, you could think of information as a
table with two or more columns. Here are some examples of information:
• It rained on September 20, 2021, in St. Thomas.
• I am 68 inches tall. My best friend is 61 inches tall.
• Beverly completed a course called Introduction to Statistical
Analysis with a score of 94 percent.
• The average dumpster contains 8,287 pounds of garbage.
Information and data aren’t necessarily meaningful. Some information is
meaningful to different people at different times, and some of it never turns
out to be meaningful at all. Other times, you won’t know if it’s meaningful
until you look at it. Here are some pieces of information with different
levels of meaningfulness:
• Beverly completed a course called Introduction to Statistical
Analysis for Marketing with a score of 94 percent.
• Beverly completed a course while wearing a blue t-shirt.
• Beverly completed a course online.
• Beverly completed a course from home.
• Beverly completed a course while drinking a glass of wine.
• Beverly completed a course using the Chrome browser.
• Beverly completed a course and referenced the book The Art of
Statistics, by David Spiegelhalter, while doing so.
• Beverly was wearing pajamas while they completed the course.
Some of this information is very easy to observe, collect, and record.
Some is very difficult, if not outright unethical, to collect. Some of it may
have immediate utility. Some of it may be immediately discardable as
irrelevant. Some if it may be tempting to have on hand just in case you need
it in the future. This is part of the challenge you will face on your data and
analytics journey.
Knowledge
Knowledge is the useful and applicable result of information and the
patterns we can discern from it. With knowledge we can make decisions,
solve problems, and generate predictions. Knowledge influences action.
Here are some examples of knowledge:
• When people reference outside textbooks while taking online
courses, they tend to score higher on the exit tests than people who
only used the course material.
• People who take the online version of the course are more likely to
use outside references and textbooks than people taking the course
with an instructor.
• People who take the online version of the course are more likely to
also be drinking alcohol and wearing pajamas while participating
than people who take the class in person.
Wisdom
Wisdom is generally considered the application of knowledge (along with
other internal resources such as experience, morality, and emotion) to make
good decisions and judgments. It requires use of the principles that underly
the situation at a systemic level to draw novel conclusions. It is an essential
ingredient to intentional innovation.
To sum it up, data can be gathered. Information can be memorized.
Knowledge acquisition (and application) is generally what we consider to
be learning. Wisdom is what we expect from our leaders and experts. Figure
2-1 shows the progression from data to information to knowledge to
wisdom as a function of increased connectedness between the data and our
understanding of it.
Predictive analytics give you the power to use what happened yesterday—that
is, our old ways of measuring and looking at training effectiveness: the
descriptive and diagnostic analytics—to accurately predict what will happen
tomorrow. These data answer the question “What’s going to happen?”
Give It a Try
In each of this chapter’s exercises, you’ll examine a learning program using
most of the definitions we’ve just reviewed. Much of the process of how to
do this will be covered later in this book, so don’t be alarmed if you don’t
have sophisticated answers yet. This is just a starting point!
Here’s a Hypothetical
Explore a learning program for which you may or may not be gathering
data. Answer the following questions about that program to practice using
the terminology. (Note that you are answering questions about the program
delivery, structure, results, and so on, not about the topic of the learning
itself.)
• What data could be gathered from or about this program?
• What information would be relevant about this program?
• What kinds of knowledge about the program would it be insightful
for the L&D team to have?
• What sorts of wisdom would the L&D leadership offer in
this context?
• What kinds of simple math can be performed on the data?
• What kinds of descriptive questions do you have about this program?
• What kinds of diagnoses could you perform based on the data about
this program, if you had it?
• What kinds of things would be helpful to predict based on data about
this program, if you had it?
• What kinds of actions might you prescribe based on data about this
program, if you had it?
• What kinds of information would you put on a report or scorecard
about this program?
• What kinds of visualizations might be useful for this program?
• What kinds of things might a dashboard display? Who would want to
use it?
• Does your organization have a business intelligence function? Do
they use any of this data already? Do they use it for AI or machine
learning?
Do It for Real
Explore a learning program for which you or your organization are already
gathering and using data. Answer the following questions about that
program to practice using the terminology in this chapter. (Note that you are
answering questions about the program delivery, structure, results, not
about the topic of the learning itself.)
If at any point in this list you run into a “Oh, we’re not doing this”
response, feel free to skip back to the previous set of hypothetical questions
to continue on.
• What data is being gathered from or about this program?
• What information is being generated about this program?
• What kinds of knowledge about the program are being created?
• What sort of wisdom is being applied (or perhaps challenged) here?
• What kinds of simple math are being performed on the data?
• What kinds of descriptive analysis are being created about
this program?
• What kinds of diagnostic analysis are being done with this data? Or,
if none is being done now, what could be done with the data?
• What kinds of predictive analysis are being done with this data? Or,
if none is being done now, what could be done with the data?
• What kinds of prescriptive analysis or action are being done with this
data? Or, if none is being done now, what could be done with the
data?
• What kinds of information are being put on a report or scorecard
about this program?
• What kinds of visualizations are being used for this data?
• What is being displayed on a dashboard? Who uses it?
• How is your business intelligence team using this data?
• Is any of this data being analyzed using AI or machine learning?
Bonus Points
Find a stakeholder or expert outside the L&D function in your organization
who is interested in these types of questions. Let them know you’re reading
this book and would like to start a series of conversations about how you
can all better leverage workplace learning analytics. This interested
stakeholder may become a central figure in some of your early analytics
projects.
Using Data to Make Predictions About Future
Performance and Learning Transfer
By Emma Weber, CEO, Lever–Transfer of Learning
Data Specifications in
Workplace Learning
It’s worth taking a moment to discuss the power and value of data
specifications and standards in our learning analytics work. Don’t worry,
we’re not going to spend a lot of time here or get too deep into the weeds.
But having a comfort level with the basics can give you greater insight into
the nature of your data and how easy it will be (or won’t be) to combine it
with other data for analysis. Refer to the appendix for additional reading if
this is an area you’d like to explore even more deeply.
Give It a Try
Here’s a Hypothetical
Consider your organization’s data culture with the following questions:
• Is there a centralized shared team that is responsible for data (often
called the business intelligence function)?
• How easy is it to get data about the organization’s operations?
• How easy is it to get data about learning in the organization?
• How often is data used for decision making?
Do It for Real
Explore the various data specifications in use in your organization’s current
learning ecosystem with the following questions:
• What specifications are supported by your LMSs?
• What specifications do your e-learning authoring tools support?
• Do you have a learning record store (LRS) for xAPI?
• What other learning modalities does your organization use (such as
ILT or performance support)? How is data about usage and efficacy
recorded?
• Do you have learning tools in your ecosystem that do not share their
data with other platforms?
• Is there a place where all the data for learning is stored?
Bonus Points
Add questions to your next learning technology RFP to explore the
vendor’s use of data specifications and whether their data is interoperable
with the rest of your learning ecosystem.
Why Do Data Specifications Matter for Learning
Analytics?
By Tammy Rutherford, Managing Director, Rustici Software
I’ve seen this scenario happen several times—enough to suggest that this is
a phenomenon that others will face as they take on learning analytics: A
team of data experts and L&D professionals comes together and realizes
that there is a lack of overlap in experience and terminology that needs to
be addressed before they can move forward together. The data folks really
know their statistics; the L&D people understand learning. But the data
team may not appreciate all the nuances that come along with a metric like
completion of learning. Many of the terms in L&D have very specific uses,
even though they tend to be described with very commonplace words. And
the L&D people may not appreciate the statistical concepts that will be used
to assess whether the learning is effective. We can’t just throw our data over
the wall to the data scientists and expect to get back something meaningful.
In this chapter we will focus on the kinds of metrics that are relatively
unique to learning and development; these are the concepts that may be new
to your colleagues in data science and analytics. Then in chapter 5 we will
spend time on the statistical concepts commonly used in the analytics and
data science field that L&D professionals may need to brush up on.
Give It a Try
Here’s a Hypothetical
Let’s take testing as an example because there is lots of data available here.
Imagine a test for a topic that you might create a course for:
• What kinds of questions do you have about the results of the test?
• What questions do you have about the experience of taking that test?
• What kinds of data would you want to collect to learn more
about that?
Do It for Real
Again, take testing as an example. Locate a course for which you have
ready access to question-level test data. Download that data in a way that
obscures the names of the individual test takers, because that is both
irrelevant here and a violation of those individuals’ privacy. We haven’t
gotten into the details of analytics yet so for now just scan, sort, and search
through that data and see if any patterns emerge as you go. What kinds of
questions do you have about the results and the experience of that test
itself? What kinds of data would you want to collect to learn more about
that?
Bonus Points
If your industry has an educational association or credentialing and testing
body, consider reaching out and contacting their testing team to see if you
can learn more about the types of test analytics they are using to fine tune
their products and credentials.
Using Learning Data to Discover Insights Into
Student Performance
By Andrew Corbett, Senior Instructional Designer, University of
California–Davis
While you can absolutely get a start in learning analytics using some simple
math that many of us learned in school, the truth is that to answer more
interesting and complex questions you need more interesting and complex
calculations.
It’s for this reason that more mature learning analytics teams include
someone with a statistics or data-science background. Not everyone on the
team needs to have these skills, but having an appreciation and awareness
for them will help you work more effectively.
Basic math and simple statistics will become useful competencies for
instructional designers as we continue to leverage analytics. Perhaps few
have said this as eloquently as Janet Effron of Notre Dame University
(whom you’ll meet later in a sidebar). She frequently says in her
presentations that “Las Vegas is a glowing monument to the fact that people
are ignorant about statistics.”
Fortunately, the analysis of data and statistics is currently a popular topic
in the press. There are several approachable—if not downright entertaining
—books on the subject that I’ve listed in the further reading section.
In this chapter we will take a look at measures of central tendency,
measures of spread, and the sometimes mysterious concepts of statistical
significance and confidence. But first we will explore three types of data.
In Figure 5-3, the mean is skewed lower than you might think because of
the presence of some data in the low end of the range. In such a case, the
median or mode may be a more appropriate measure of central tendency
(Figure 5-4).
Measures of Spread
Measures of spread help us understand the degree of variation—and
therefore risk—in our data. Measures of spread give us a sense for the range
of possible outcomes and can help us make smarter decisions based on the
data. For example, if you found out that your learners rated a course
experience three out of five stars, you might interpret that as a mediocre
rating, and take efforts to improve the course. However, you might find out
that you had had a group of people who were responding very poorly to the
course (rating it a one), and a separate group of people who are responding
very positively to the course (rating it a five), bringing the average of the
two groups to about a three-star rating. Changing the course overall might
actually make things worse for those people who really enjoyed it and rated
it a five. This is why a measure of spread provides more depth than simple
averages. Two measures of spread are range and variance:
• Range: The difference between the highest and lowest value in a
data set. As you can imagine, in situations of high spread, your
ability to rely on the mean, median, or mode may be lower. I like to
think of range as the fast-and-easy look at spread in a data set. For
example, in a batch of test scores, if the lowest scoring individual
earned a 40 percent and the highest scoring individual earned a 100
percent, the range is 60 percentage points. This may indicate that you
should go digging more deeply into why the test had such a large
range of success rates. On the other hand, if your lowest scoring
person earned a 90 percent on the test, a quick look at a very narrow
range on the very high end of test scores indicates that the test is very
easy to pass.
• Variance: The average of the squared differences of each value from
the mean. A high variance indicates that the data in the set are far
from one another and far from the mean. A small variance indicates a
tighter grouping around the mean.
Figure 5-5 shows a data set that demonstrates a lot of range and variance.
The data points are widely scattered from each other and across a wide
range of values. While you could calculate the mean for this number, it
would not be an accurate representation of the nature of the results.
The data set in Figure 5-6 shows less variance but a similar range. There
are two tight groupings of data points, one at the high end and one at the
low end. In this data set the mean would also be a useless number.
And Figure 5-7 shows low variance and a small range. As demonstrated
by a tight grouping around the mean, for this data set the mean would be a
useful number.
What this says is that if I declare that my range of results is one standard
deviation on either side of my mean, 68 percent of my results fall in that
range. I can be 68 percent confident that the true answer is somewhere
within that range. If I go out two standard deviations, I can be 95 percent
sure that the true answer is within that range. And so on.
Of course, most data in the real world does not follow a perfectly smooth
normal distribution—and this is where it’s helpful to engage statistical
analysis.
Side note: You may have heard the term Six Sigma in an organization’s
quality improvement efforts. Sigma is the Greek symbol used for standard
deviation. Six Sigma implies a target quality level within 6 standard
deviations between the average and the acceptable production output, or
99.99966 percent.
Sampling
Sampling is when you only use a subset of the data instead of all of it. In
L&D we may have access to all the learning data for all the learners in our
population, and in these cases sampling may not be relevant. However,
sampling can be useful if our data set is very large, if we are just getting
started, or if we’re conducting an experiment and comparing one group to
another. What we’re doing with sampling is using a subset of the data to
make an inference about the entire population.
It’s absolutely critical when you are sampling to make sure that your
sample is actually representative of your entire population. As you can
imagine, if your sample includes a subset of the population that is different
from others in terms of performance, geographic region, tenure with the
company, and so on, your data and decision making will be skewed.
There are several decisions and techniques that you can use to more
reliably create a sample that is truly representative. Randomization is a
commonly used technique where participation in one or another test group
is randomly based. This tends to be the gold standard. At the opposite end is
convenience sampling, where you have a sample that is made up of the
portion of the population that is easiest to reach; email surveys are a
common example. One of my favorite examples of convenience sampling is
the Level 1 course evaluation as it is typically implemented in
organizations. The course evaluation is voluntary, so we only hear from the
people who chose to fill it out for one reason or another. It is convenient,
but not necessarily representative.
When we do not have a representative sample, we can say that the
sample may be biased. A biased sample gives us results that are not
generalizable across the entire population. A classic example of this is
medical and social science research that, throughout the 20th century, was
often conducted on white male college students because they were
conveniently available to researchers. However, young white males are not
representative of the entire human population, and thus the results of these
studies may not be applicable to other groups.
To draw on an example a little closer to home, consider the number of
times in which instructional designers reach out to their co-workers in L&D
to help them pilot-test a course. The thinking here is that their colleagues do
not know the content that is being taught, therefore they make good test
learners. On the other hand, the instructional designers are not in the job
function of the target audience and will never need to use this content to
improve their performance at work. They are not a truly representative
sample of the population who will use this course, and any results that are
gleaned from a pilot test will need to take that factor into account.
At its very core, sampling bias reduces our ability to draw meaningful
conclusions from data.
As you can see, depending on the nature of the question that you are
analyzing, different data points can serve as the independent or the
dependent variable. This is something we can have a little bit of fun with. In
chapter 8, we will discuss generating the questions that will form the basis
of your analysis.
Give It a Try
Here’s a Hypothetical
Think about a digital learning experience that you’ve either created or
engaged with. What kinds of quantitative, qualitative, and categorical data
could you collect? Which measures of central tendency might be relevant
here? Do you think the spread is large or small? Would you need to use
sampling to gather data or could you use the whole population? If sampling,
what would be a statistically significant sample size to work with?
Do It for Real
Select a learning experience with which you’re familiar that is currently
collecting data. What kinds of quantitative, qualitative, and categorical data
are being collected? Which measures of central tendency might be relevant
here? Is the spread large or small? Does the analysis of this experience rely
on sampling or are you able to use data from the whole population? If
sampling, what is a statistically significant sample size to work with?
Bonus Points
Connect with someone in your organization who analyzes data and discuss
these concepts with them and how they apply in their work. Consider
meeting up with someone in marketing, manufacturing, or research.
Building a Continuous Program Measurement
Instrument: The Learner Adoption Index (LAI)
By Tiffany Jarvis, Edward Jones
Engagement and use of the tool is most important in any mobile learning
environment. We had to develop a metric to represent engagement and use
of the tool. Our stakeholders guided the choice of metrics from DEP and
RTC. What is critical to them is critical to us. From there we conducted the
analysis to show whether it supported our hypothesis that using the tool
would prepare future sailors for RTC and lead to less attrition.
PART 2
DESIGNING FOR DATA
Chapter 6
Before we get too much further into the use of analytics in the L&D space,
let’s take some time to consider that analytics is just one of several uses of
workplace learning data. In this chapter I will suggest a framework that is
less about providing a strict set of rules and buckets, and more about
looking at the variety of things we can be doing with data to gather some
perspective. This is a starting place in an emerging and evolving field, and I
fully expect that as an industry we will improve upon it over time.
Performing Analytics
With data safely gathered and stored, and the basics of reporting and
visualization covered, we now have the ability to perform analytics on the
data. This is obviously the topic of this book, and there is much to come in
the pages that follow. Analytics are typically used to support decision
making within the organization, so they get a lot of attention.
Integrating Workflows
A fifth use for data in this framework is as a marker, milestone, or trigger
within an automated workflow across systems. Using data in this way
doesn’t require visualization, analytics, or even sharing that information
with the individual. This includes and extends the concept known as
“learning in the flow of work” and incorporates what I might call “working
in the flow of learning.”
A simple example is rules-based audience mapping for learning content,
in which data about an individual’s role, activity in systems of work, and
prior learning are all used to assign required learning. It is likely that your
current LMS has the capability to do this. More complex workflows would
involve combining actual performance data and learning data to provide
access, limits, and learning opportunities that are exquisitely useful in the
moment. For example, if I am a salesperson who records an opportunity for
a product that I have never sold before, I could be offered a learning
experience about that product. (Presumably I do not know much about it.)
However, if I had already completed training about that product, and
performed well on it, my capacity constraints for that particular product
may be expanded.
The edges of these categories are left intentionally blurry because each
use of data may blend into the next, and the systems used for them often
accomplish several adjacent uses of data (Figure 6-3). While storing and
securing is a prerequisite to the rest, and reporting and visualizing are
precursors to analyzing data, there is no implied prerequisite, maturity, or
value progression in the other three uses of data. They are simply different
from one another.
Real-world applications in L&D will often include several of these uses
of data, not just one. For example, we at TorranceLearning have designed a
microlearning-based program for one of our manufacturing clients that
includes a significant amount of post-learning action to ingrain the new
skills into the daily workflow. The post-learning interactions use a Mobile
Coach chat bot to continually engage with learners over a 12-week period.
In this case, we first store and secure the data. Data from course interactions
and surveys are analyzed to assess the learner experience and program
engagement. The client can use the chat bot to answer questions about
program usage, engagement over time, and other specific topics. Data that
the learner provides during the chat bot interaction is used to further
respond to the learner in a relevant and personalized way.
Give It a Try
Here’s a Hypothetical
Explore a learning program for which you may or may not be gathering any
data or have access to it yourself. Answer the following questions about that
program to practice using the framework offered in this chapter:
• How could you store and secure the data that could be gathered by
this program? (Assume that it does gather data for the sake of this
exercise.) Where could data be stored? Who will need to access it to
verify credentials?
• What reports could you create with the data? What visualizations
might be useful? Could a dashboard be used?
• What kinds of questions could you answer if you analyzed the data?
• In what ways can the data be used to improve the learning
experience? For personalization? For progress tracking? For
recommendation? How about adaptive learning?
• How could this data be incorporated into a meaningful workflow? Or
conversely, how could performance data be used to influence or
trigger learning?
Do It for Real
Explore a learning program for which you or your organization are already
gathering and using data. Answer the following questions about that
program using the framework offered in this chapter.
• How is the data gathered by this program being stored and secured?
Who has to access it to verify credentials? (I recommend that if you
don’t know the answer to these questions, this is a great opportunity
to connect with someone in your organization to find out.)
• What reports use this data? What visualizations are offered? Is this
data on a dashboard? If yes, what else is on that dashboard?
• What kinds of analytics are being performed on this data?
• How is the data being used to direct in the learning experience? (or,
if it is not, how could it be?) For personalization? For progress
tracking? For recommendation? For adaptive learning?
• How is this data used in the flow of work? (Or, if it is not, how could
it be?)
Bonus Points
What would you add to this framework?
The Total Learning Architecture (TLA) Strategy
By Brent Smith, RD&E Principal (SETA), ADL Initiative
In this chapter, we’ll start the planning process by identifying the challenges
we are trying to solve and the kinds of data that will help us do it. We’ll
offer more ways to start thinking about what data to collect than are
reasonable or feasible for any one program. Your challenge is to find what
resonates with your organization and start there.
All good chapters on planning start with some recognition that if you
don’t have a plan, you won’t get where you’re going. Of course, few things
go exactly according to plan—but if you’re exploring with data, that’s
probably a sign that you’re doing at least some things right! So, with both a
need for a plan and a recognition that it will evolve along the way, let’s get
started.
Generally, your business sponsors will have a set of leading and lagging
metrics that they rely on. If needed, a large list of metrics can be narrowed
and prioritized by asking a few questions:
• Which of these are the most salient to you?
• Which of these are easiest and most timely to obtain?
• Which have the clearest and most agreed-upon definitions?
• Which is most likely to continue being used in the near future?
• Which metrics are you tracking at an individual level? Which do you
share with employees?
• Which of these are you willing to share with the L&D team?
• Where would you like to start?
In addition to strategies and goals, your business sponsors may have
some questions that you can help them answer with learning data. This is a
good time to collect them, too. It might look like Table 7-3.
Table 7-3. Questions, Answer Sources, and Metrics for the Business
Questions Answer Sources Metrics
Learning data, Courses
1. Does training help sales data completed by
improve sales results? rep, sales by
rep
Competency Assessment or
2. Is our engineering team assessment, expert certification
up to date with their analysis testing
materials skills?
Table 7-6. Questions, Answer Sources, and Metrics for the Learning
Organization
Questions Where Are Metrics
the Answers?
• User access • Percent of
1. How quickly will employees rates user log ins
adopt the new LXP? • Content by time
access in period
the LXP • Number
ofcontent
objects
accessed per
user
• Vendor and • Cost to build
2. Can we reduce the cost of internal the refresher
delivering annual refresher costs for training
training? development • Cost (hours x
• Completion pay rate) of
rates learners to
• Time spent take training
in learning
• Help desk • Help desk
3. Does teaching the use of the ticket tickets
performance support during system related to
training reduce the number of • training
help desk calls? Performance topics
support • Performance
usages support
access rates
Give It a Try
Here’s a Hypothetical
Take an existing recent project that you’re familiar with. Challenge yourself
to complete the tables in all or most of the sections of this chapter. What is
your list of metrics? Then work through the hypothetical, whittling down
the questions. Which ones would you prioritize?
Bonus Points
Do this activity with a colleague who is also familiar with the program.
What is their list of metrics? What can you learn about the differences
between the two?
Do It for Real
Work with the business sponsor for a project you’re kicking off now. Work
through their goals and metrics. Work through your instructional strategy
and measurement approach. Work through your own questions. Now
whittle the list down.
Using a Business-Focused Learning Data and
Analytics Plan to Drive Results
By Derek Mitchell, former Head of Insight and Analytics, Sky
At Sky, we were launching a new product, and our use of learning and
performance data enabled us to meet our business goals for it.
Here’s what happened: All 10,000 contact centre staff at Sky needed to
be briefed and trained on a subscription product that was launching in a
regulated market. Training was to commence three weeks prior to the
launch with the training of all staff to take place over a 10-week period
(continuing post launch), meaning that most agents would not be trained in
time for go live. The initially proposed approach for training was to take
groups of agents through the training program in their teams. Delivering
training to the audience in this manner had a high chance of creating a
significant opportunity cost in that many customers might work with a
contact center staff member who had not yet received training. We thus
needed to figure out how to minimize this cost to the organization and
possible friction with customers.
To do this, the learning analytics team at Sky stepped in to question the
order in which agents were being trained and quickly identified that the
historical distribution of sales conversion rates was very broad across the
agent population. We built the argument that simply by changing the
scheduling of training, we could deliver greater revenue for the business
while using the same material, over the same 10 weeks, with the same
group of learners.
We were able to use sales data on existing products to model the sales
rates over the 10-week roll-out period based on both the random selection
of teams attending training (each team consisting of higher and lower
performers), and on front loading the training with the best performers
historically.
By front loading our best performing agents, we were able to route
potential customers of the new product to that group and thus benefit from a
higher conversion rate than otherwise would have been the case. We sold
more products faster because our best agents had additional weeks to sell
than if their training had been distributed across the 10 weeks.
This simple use of analytics generated an additional £300 thousand over
the 10-week period, which otherwise would not have been realized.
Chapter 8
Exploratory Analysis
While hypothesis testing is a rigorous approach to analytics, I should point
out that not all analysis needs to be this hard-edged and focused. In fact,
transforming the questions posed by the business and your L&D team into
discrete, testable hypotheses may require some initial exploratory analysis.
In exploratory data analysis, you will collect a lot of data and dig
through it looking for patterns or a lack of perceivable patterns. This is like
walking into a retail clothing store and not looking for anything particular;
you’re just hoping that a sweater catches your eye and you’ll pick it up, try
it on, and see if it works for you. In this metaphor, the clothes are the data
sets and your casual search is the analysis.
Of course, it’s entirely likely that you had a general idea of the kind of
thing you were looking for when you went into the store (otherwise you
probably wouldn’t have gone into the store). This metaphor continues to
hold true for that type of exploratory data analysis. For example, when I go
into a clothing store, I am only looking for women’s clothing that would fit
me and generally for the season that I happen to be in. I know what clothes
I already have and therefore don’t need to buy, so I kind of have a sense for
what I’m looking for. At the same time, I’m open to new things I might
stumble upon as I shop. The exploratory data analysis is an opportunity to
help find additional questions and create different hypotheses.
One of our nonprofit clients was open to us pursuing an analytics
endeavor for their learning offerings, but didn’t really know what to ask.
They were open to whatever insights we might find and how they could
help them create a strategy for future course development, fundraise more
effectively with donors, or improve the overall experience for the learners.
While we didn’t have any concrete questions, we did have a good sense for
their data because we support their LMS and their course offerings. We had
enrollment dates, user group information, some very limited user data,
SCORM data from their e-learning programs, and very basic video
consumption data from a series of videos they’d released last year.
Knowing what data was available to us helped our team come up with an
initial list of questions that we thought would be useful.
Hypothesis Testing
That gets us to the hypothesis-driven approach to analytics. A hypothesis is
an educated guess or a suggested solution for a particular challenge, where
we don’t know the actual answer. Here are some examples:
• When the weather is rainy or cold, attendance at indoor museums
increases.
• Adults who get at least eight hours of sleep are more alert than those
who get less than eight hours.
• Employees who complete their annual compliance training are less
likely to commit ethics violations.
• When e-learning is designed for mobile access instead of computer-
only, learners will complete the course on their mobile devices more
often than their laptops.
• Technicians who access performance support tools more frequently
make higher-quality repairs than those who do not use performance
support tools.
You may notice that each hypothesis is stated as though it is or could be
true. This is very important. Your hypothesis is simply a testable claim.
What’s more, while your analysis can provide support to confirm your
hypothesis, you cannot claim that your hypothesis is true 100 percent of the
time.
For this reason, we create both a hypothesis and a null hypothesis. The
null hypothesis states that there is no relationship between the variables in
your hypothesis. Table 8-1 offers an example of null hypotheses to the
hypotheses already stated.
And, while you cannot prove that your hypothesis is true 100 percent of
the time, you can (if your data supports it) be confident in rejecting the null
hypothesis.
Now we may want to improve upon this a bit, because maybe everybody
showed a measurable improvement in performance during the period in
which we were testing. In fact, in a “what gets measured gets done”
environment, simply focusing on something and measuring it often
improves performance. So, let’s hone this hypothesis:
Employees who took the training showed more improvement than employees
who did not take the training.
Of course we would like to believe that the training was the cause of the
performance improvement, but let’s leave that aside for right now and just
go for a relationship that we can show. While we cannot prove that training
caused the improvement, what we can do is identify and test a null
hypothesis. (You’ll find with experience that the null hypothesis is actually
the easier of the two to test and verify or disprove.) A null hypothesis, or
the evidence that there is no relationship, would be expressed like this:
Employees who took the training did not have any difference in improvement
relative to those who did not take the training.
So we now have our hypothesis and our null hypothesis and we can
identify the data that we will collect to test them. Let’s pull these two apart
a little bit more:
Give It a Try
Here’s a Hypothetical
Consider one of the questions that you identified in the previous chapter
and form a discreet and testable hypothesis about it. Create the
corresponding null hypothesis. Now parse the two into specific data points
and definitions that you’ll need to be able to have testable data. What are
some of your observations? Would this be easy to implement in your
environment? What would need to change to gather this data (including
your hypothesis)?
Do It for Real
As you work through the analytics process with a current project, consider
the questions that you identified in the previous chapter and form a discreet
and testable hypothesis about one of them. Create the corresponding null
hypothesis. Now parse the two into specific data points and definitions that
you’ll need to be able to have testable data. What are some of your
observations? Would this be easy to implement in your environment? What
would need to change to gather this data (including your hypothesis)?
Bonus Points
Start with an analysis already performed on data in your organization
(learning data or not). Here we’ll work backward through it to help you be a
critical consumer of data and analytics. Based on the analysis (typically
shown as a chart or graph), what might the hypothesis and null hypothesis
have been? What sorts of definitional assumptions or decisions would have
to be made about each of them to arrive at meaningful data?
Identifying Data Needs Requires Deep
Stakeholder Engagement
By Janet Laane Effron, Data Scientist and Instructional
Technology Specialist
We are awash in data. (Whether or not we have ready access to it all in one
place in a usable format is another thing entirely.) The opportunity to collect
and use this data is rich. It is up to us to leverage this opportunity, be
responsible with it, and use it to help drive positive change in our
organizations. In this chapter, let’s take a look at where this data comes
from along what we might refer to as a “learning data supply chain.”
Self-Directed E-Learning
Thanks to SCORM, we’re already tracking some data from our e-learning
experiences. We know who has been assigned training, who has completed
training, and who is currently incomplete in their training. We know how
long they have spent with the course open (which is different from how
long it takes to complete the course!), and which screens they have viewed.
We know the score achieved on one of the tests offered in the course,
usually the most important one. And we may also know their answers to the
individual questions on that test.
If you are using xAPI, you can also track data on every screen, every
interaction, every quiz, every download, every video, and more. With xAPI,
you can see how many attempts it took for an individual to get a question or
an entire test correct.
Video
Video interactions offer another rich source of data. You can think of each
click on the video player as an opportunity to gather data: start, pause, scrub
back to watch something again, skip all the way to the end to get done as
fast as possible, as well as where participants stop watching (abandon) the
video. All of this can be tracked and marked by the timestamp within a
video launched by an LMS or video streaming player.
Observations
In some respects, observations can be thought of as very similar to
assessments. Of note, however, is the fact that observations are conducted
by someone else (an instructor, subject matter expert, auditor, manager, or
peer). What this means is that not only do we have data about the individual
being observed and their performance, we are also able to gather data about
the individual performing the observation, such as a facilitator observing
performance during a training program. This data can include everything
from their usage of the observation tool, to the tendencies of individual
observers to rate others higher or lower than their peers.
Live Classroom
In comparison to e-learning, video, and testing, we are often not gathering
nearly as much data about what happens in a live classroom. In corporate
learning we are often recording manually whether or not somebody has
attended and thus completed a class, generally by the instructor marking an
online roster afterward. In some settings, a hand signed roster is gathered,
and this can be digitized and stored along with the completion record for
added (and auditable) proof. In-class engagement and response technologies
allow us to capture individuals’ answers to questions posed throughout the
class.
Any class work that is performed digitally can be tracked and stored.
Several years ago, I experimented with capturing the images of flipcharts
and activities performed during a live class, recording them using xAPI and
then attaching them to the individuals’ completion record in the learning
management system.
Virtual Classroom
Holding a class in a digital environment offers an entirely different
opportunity to gather data. Who is in the class and when do they leave,
when do they comment and what do they say? How do they answer
questions? Do they raise their hands? All of these events are recorded
digitally, and could be a source of data for you. Since activities in a virtual
classroom can be done on a digital whiteboard, this, too, can be captured as
part of the data stream.
Social Learning
Social media, social platforms, collaboration platforms, and the like gather
and store data each time an individual contributes, reads, reacts, or responds
to content. Many of these platforms allow for the storage of a variety of
types of content (such as text, images, videos, links, and polls) and thus we
are able to gather data not only about who does what, but also which types
of media they consume. We also have the ability to see and record data
about who follows whom, and which content generates the most
engagement. Basically, all the types of data that marketers use on social
platforms are also available to us as we create communities of learning
within our organizations.
Chat Bots
Interactions with chat bots generate learning data at every step. The receipt
of each message, the response to messages, the time it takes to respond to
the messages, and the interaction drop-off rate, in addition to the actual
content of responses and searches, provide a wealth of insight into this type
of learning experience.
Search
The search box is a feature offered on most LMSs, LXPs, and several of the
learning modalities listed earlier, as well as browsers, intranets,
collaboration tools, and more. We don’t often think of search as a data
source, but data we can gather from it includes answers to questions like:
• What are people searching for?
• What results are fed back from the system?
• How many searches does it take before a person finds something to
click on?
• Which results do people pursue?
• Which search terms have no good responses to offer?
All of this can provide insight to the learning team.
Give It a Try
Here’s a Hypothetical
Think about a digital learning experience that you’ve either created or
engaged with, and all the platforms and systems that are involved in
delivering it. What kinds of data could have been recorded about that
experience? And, if you are privy to this, what kinds of data actually were
recorded?
Do It for Real
Select a learning experience with which you’re familiar. Considering all the
platforms and systems involved in storing, delivering, and consuming
learning content, what kinds of data are being recorded about that
experience? What sources and types of data are being missed in this
experience?
Bonus Points
Is data being stored in a single place for all of the points along that learning
delivery supply chain? If not, in how many different places is data being
stored?
Data Ethics
By Stella Lee, PhD, Director, Paradox Learning
Besides making sure that we follow the best data analysis practices, with
people data we have a special responsibility around privacy. And for people
data at Critical Equity, it’s a practice that we’ve had in collaboration with a
legal team to make sure that the personally identifiable information is
always protected. We have a data privacy policy in place, and we also
maintain different levels of data permissions for different team members,
which dictate what they can access and at what level.
And this is where aggregate data can be very helpful. Aggregate data is
when we look at data in more of a sum at a high level rather than individual
levels, and we look at not just one person but a group of individuals’ data.
So, for learning analytics, similar to other people-analytics teams, we
have to make sure we follow all the same sort of data privacy and policy
guidelines. We keep things as much as possible at the aggregate level. After
running an audit of the data and making sure we have the right data sets,
there’s a lot of partnership and investment in relationships before we start
even publishing analysis. This includes cross-functionally building and
enhancing relationships to make sure that we have trust in the data, and that
the business and other stakeholders within the people team trust what comes
out of those conclusions and trends.
All this takes collaboration, consideration, patience, and time to get it
right.
Chapter 10
It probably goes without saying that in order to analyze data, you will need
to collect it and store it, which is a little bit trickier than we might like. For
example, I remember the early days of xAPI when people would tell me
that they had published an e-learning course for xAPI, connected it up with
a free trial of a learning record store, and were very disappointed to find
that the only data they collected looked like SCORM. Of course it did.
That’s all the course was sending! If you want to get more interesting
information, you need to collect more interesting information, and many of
those early tools just could not do that (at least not without additional
programming effort).
As you define your data questions and make a plan for collecting data,
you’ll also need to make a plan for capturing and sending the data.
In this chapter, we will take a look at capturing data from the platforms
you use, building data capture into the learning experience that you
develop, gathering data from non-digital learning experiences, and
improvising when all of the above doesn’t quite work the way you want it
to.
I like to think of data captured from learning systems in the same way as
data captured from sensors such as motion detectors, luxometers,
speedometers, security badges, and so on. These sensors offer a wealth of
information that may be valuable to you in your learning analytics. Devices
such as smartphones and watches, digital assistants, and computers, as well
as the browsers and apps on them, also can provide a lot of information
including location and device information itself. You likely won’t be able to
modify what data is captured, and it may even be difficult—or unethical—
to extract information in a format that you can use, but these may hold
unique opportunities for you and your learning experience.
For a very simple example, at the TorranceLearning headquarters office
we have a sensor on a door to one of our conference rooms. It records an
xAPI statement every time the door is opened or closed. This particular
sensor hardware also collects data about the temperature, altitude, and
barometric pressure at the time the door is opened or closed. Temperature is
sometimes interesting, but altitude and barometric pressure are useless
information to me. Furthermore, the door does not know who opened it, so I
am hard pressed to link that data to any other employee or visitor records.
Similarly, manufacturing systems may be capturing data but without a
connection back to who was operating it at the time the data was captured.
These gaps will provide some challenges to you and your work, but they are
not insurmountable.
To take this one level deeper, if you want to keep track of the number of
times someone attempts a particular interaction on a screen, you will want
to capture data with every click on the screen, rather than design the screen
so that the only data that gets sent back is the learner’s final response. For
example, imagine you want to know how many times people clicked on the
answer options on the screen in Figure 10-1 as indication of their
uncertainty about their response. The best way to capture this data would be
to place data capture on each click, not just the final submitted answer.
Beyond tracking screen clicks, a commonly used instructional strategy in
e-learning is to ask for learners to answer a free-text question. The data that
learners enter is often not able to be used outside of the course as it is not
captured using SCORM. With xAPI, we can capture the answers to that
question, read them, and analyze them. This offers an entirely different
depth of insight into the learners’ experience (Figure 10-2).
Give It a Try
Here’s a Hypothetical
Consider a learning experience that you’re familiar with. What kinds of
data are collected about that experience? What are the triggers or actions
that generate that data? What additional types of data could be collected? If
multiple sources of data are involved, how are they indexed or matched up
so they can be used together? How is personal data kept private? How is the
data kept confidential? Who has access to see it and why? How is the data
kept secure?
Do It for Real
As you’re designing your learning experience, and you’ve identified the
data that you would like to capture from it, what are the triggers or actions
that generate the data you need? If multiple sources of data are involved,
how are they indexed or matched up so they can be used together? How
will personal data be kept private? How will the data be kept confidential?
Who will have access to it and why? How will the data be kept secure?
Bonus Points
Discuss the data privacy, confidentiality, and security protections your
organization has in place with a data expert or the organization’s
legal counsel.
Instrumenting Learning Experiences to Capture
Meaningful Data
By Matt Kliewer, Learning Engineering Team Lead,
TorranceLearning
Of course, the unsung hero of all that we have been discussing so far in this
book are the places that store our data so we can use it. Up until now we
have been discussing sources of data, data providers, and the kinds of
information that we can gather from them. Data storage is often a space in
which instructional designers will need to collaborate with their IT
departments to establish secure and interconnected data stores.
As we proceed through this chapter, you may think, “Wait, we just talked
about xyz tool as a source of data, and now we’re talking about it as a
storage point, too?!” Yes! And while this may seem confusing at points,
some systems will be both generators of data and places to store it.
Learning Experiences
It is not unusual for a learning experience that is delivered outside the LMS
to have its own internal data storage and analytics capabilities. This
includes performance support tools, memory aids, role-play applications,
virtual reality, augmented reality, and many types of gamification. They are
both data providers and datastores all in one component. For many of these
learning experiences, SCORM was never really a relevant concept, so their
adoption of xAPI and other data exchanges may have lagged. More and
more of them are beginning to adopt xAPI, allowing for deep analytics of
the learning experience within the tool, as well as portability of summary-
level data outside to a separate LRS.
This is a good opportunity to take a quick pause and discuss the kinds of
data and analytics at the varying levels we’ve discussed. Within any one
learning experience or learning experience type, the data gathered can be
incredibly, even excruciatingly, detailed. Consider the opportunities for
analytics if you had data about every single button click, when a click is
made, how long it takes to get there, whether or not it’s the one that you
wanted to happen, and how you can influence that. That is exciting stuff
and essential to the analysis and improvement of the learning experience.
Most of the analytics at this level of granularity are situation- and tool-
specific and not very easy or useful to generalize across learning
experiences. By this I mean, a click of a particular button on a particular
page in a particular learning experience may not have much relevance to
another click on a different part of a page in a different learning experience
about a different topic. This layer of data, storage, and analysis is what we
refer to as the “noisy data layer.” In the noisy data layer, there’s a lot of data
about a lot of very little things that allow us to make very specific changes
and adjustments to the learning experience.
As we start to compile data across learning and performance
experiences, we need to be a bit more normalized and at a higher level in
the data that we report and the expectations we have for our analysis. It is at
this more transactional layer that we use summary-level statistics data for
our analysis, while still having access to the noisy layer of very granular
data in case we need to dive in deeper. Data standards such as xAPI allow
us to gather data in one place and analyze it across a wide variety of data
providers. Without unifying data standards, we are left with only the noisy
layer, which may be siloed from learning experience to learning experience,
limiting our ability to make conclusions about the learning experiences
across platforms.
In the example in Figure 11-2, we can see some patterns and insights
emerging, such as:
• The HRIS and performance system are simultaneously a source of
data and a storage point for data, while also offering analytics
capabilities.
• The virtual classroom platform is where a lot of activity may be
happening (it is a potential source of data), but it’s not sending that
data anywhere—so we may be missing out on some value there.
• The competency map and the LRS are sending data to the learning
and employment record (LER), but these systems are not connected
to the HRIS.
• Business systems that house the actual performance of work by
employees are not connected to learning systems for analysis, and
thus they present an opportunity for additional connection.
The other purpose for this tool is to recognize the levels at which we can
expect to use and analyze our data. The closer we are to the activity layer,
the more granular and noisy our data becomes about a particular learning
experience. The further we are from the activity layer, the more that our
data needs to be transactional and interoperable, in order to draw broader
conclusions across learning data types.
The Role of IT
In my experience as a learning designer for more than two decades, I have
often heard the following sentiment: “If we get IT involved, we will never
get this project finished.” However, when we move into a learning analytics
space, and start looking at data outside the granular activity layer, we will
absolutely need to involve IT if we want to achieve our goals. In most
organizations, IT is playing a leading role in data capture, storage, security,
privacy, and transport, all of which you need to do your work in learning
analytics. This is generally not a space in which you will want to work
independently of the infrastructure and support your organization provides.
As you begin your learning analytics journey, be sure to connect with your
IT and BI teams so you can leverage their expertise and the ecosystem they
are already building.
Give It a Try
Here’s a Hypothetical
Refer back to the learning experience that you have been following along
for each chapter, and use the tool provided in Figure 11-1 to map out the
sources of activity level data, where data is stored, where it is analyzed, and
any connections to performance data or other learning experiences that may
be relevant. What patterns do you notice? What gaps do you notice? What
might be an opportunity that warrants further exploration?
Do It for Real
Use the tool provided in Figure 11-1 to map out your organization’s
learning data ecosystem: the sources of activity level data, where data is
stored, where it is analyzed, and connections to business performance data.
It is quite likely that a real-life ecosystem will need a far larger map than
the one provided here, so feel free to expand as wide and as deep as you
need to go. What patterns do you notice? What gaps do you notice? Is there
an opportunity that warrants further exploration?
Bonus Points
Reach out to the IT and business intelligence teams in your organization
and discuss your map and findings with them. What additions would they
make to the map? What insights can they draw from the learning data that
they might not have already included in their data sets?
Working Within a Unique Learning Data
Ecosystem
By Wendy M. Morgan, Learning & Development Senior
Strategist, Frank Porter Graham Child Development Institute,
the University of North Carolina, Chapel Hill
The learning data ecosystem within which the Frank Porter Graham (FPG)
Institute operates is quite unique. Although we’re situated within UNC–
Chapel Hill, the LMS used for student coursework would not work for us.
In fact, we could not find any LMS that would work for us! First, FPG is
home to a many independently funded projects with different audiences,
websites, and data. Second, the common threads among all FPG projects
are that all training is external and community-focused, and research
requires statistical analysis with meaningful variables.
For the FPG Institute, learner activity data is critical not only for
evaluation but for providing effective instruction and support. Each choice
our adult professional learners make within our web-based learning
products is reported through JavaScript as xAPI protocol data and captured
within a learning record store (LRS). This way, our e-learning is truly
portable; it can be hosted on any website, and it will still report the data to
the same database.
Use of the xAPI protocol also allows us to customize the data we collect.
In other words, each learner choice and interaction can be recorded and
operationalized into conceptually meaningful variables that are aligned with
instructional design strategy. We can also query data into custom PDFs
housed within the digital lessons, enhancing the instructional design
strategy. Learners can leave each module with a custom summary
conveying their individual progress, including the areas where they may
need to improve.
Trained as an academic researcher, I’ve always valued making decisions
based on data. Before I accepted my current position, I learned about the
FPG’s data capabilities and raised the subject of the likely need for an xAPI
or LRS learning ecosystem during interviews for the role. I knew that to
provide the most effective instructional strategy and design, I would need
an infrastructure in place to provide custom learner activity data. We started
with an open-source LRS system but eventually found that we were
bumping against the ceiling of our data plan. Once I accepted the role, I
championed the effort to install it. We had a fully functional LRS within
about four years. As we explored our options for a larger system, we were
working with these considerations:
• We needed separate LRSs for each project.
• As a research institute, our needs were very specific as we’re
collecting data to support the unique research objectives of each
principal investigator on the team.
• We would be running our own statistical analyses and creating our
own custom visualizations, so we didn’t want to pay for those add-
ons.
• We needed the power to handle a lot of data.
Eventually, we decided to invest in an on-premises set-up with an LRS
provider, which has worked out very successfully for us.
Our learning ecosystem has become a critical part of our work achieving
FPG’s strategic initiatives. Our ability to collect custom learner activity data
has made us an attractive partner for collaborative work with research
projects in other departments and universities. It is also attractive to
funders.
Considerations for Operating Within Your
Learning Data Ecosystem
By Brent Smith, RD&E Principal (SETA), ADL Initiative
Give It a Try
Here’s a Hypothetical
Using any of the learning programs that you have been working with as
your example throughout this book so far, make a list of the potential
sources of dirtiness in the data that you collect. What kinds of mitigating
factors could you put in place to prevent them?
Next, sketch out some of the questions and the analysis that you may
perform on the data that you collect. And by sketch, I absolutely do suggest
pencil and paper drawings! I often hand draw graphs and charts during my
design process. In such simple and quick sketches I can find the errors in
my logic before I invest a whole lot of time and effort to collect the data.
Since I don’t know which way my data will turn out at this point, I sketch
both a direct and an inverse relationship and see if either of those actually
has any meaning to the work that I’m doing.
Once you have these sketches, you can step back and take a look at what
new questions you might have about your hypothetical learning program.
Do It for Real
Take a look at the raw data being collected by a new or in-development
learning program. You may need to work with a member of your learning
analytics or data team to get this expressed in a raw format you can use,
such as Excel. (You may also need to request a subset of the data so you
don’t overwhelm your personal computing capacity, as some programs may
collect an astonishing amount of data.) Use sorting and filters on your data
to see if you can find any omissions, outliers, or inconsistencies that may
point to dirty data. (Keep in mind that simply because something is an
outlier does not necessarily mean it is dirty or inaccurate, but it might be.)
Next, take a look at some of the existing visualizations and analysis for
the data. Now that you know this, what additional questions do you have
that would take the analysis even deeper?
Bonus Points
There are so many ways to earn bonus points in this chapter!
• Identify ways in which you can mitigate the impact of dirty data or
pre-clean it before it comes into your analysis space.
• Talk to business sponsors about the analysis that you have for the
learning program and find out what new questions they have that
would help you take the analysis even deeper.
• If you are just getting started, be sure to build time and resources into
your project plan to account for any iterative data and analysis work
that you will be doing.
The Ad Hoc Nature of Learning Data Projects
By Janet Laane Effron, Data Scientist and Instructional
Technology Specialist
Purpose
What is your purpose or goal for communicating this data? What do you
want the intended recipient to take away from your message? Being clear
about your purpose helps you decide what to include and what not to
include in your communication. Here are some questions to ask:
• Are you trying to inform the audience about something new or
something they are used to seeing on a regular basis, such as a
quarterly update on the L&D team’s activity?
• Are you evaluating a program’s impact?
• Are you trying to persuade someone to make a decision based on
your data?
• Are you using data to support a needs analysis?
This clarity of purpose can keep you on track and on message.
Audience
Consider what you want the audience to do, then think of what your
audience needs from you in order to do it. What information do they need to
see, andover what time periods? What do they need to feel confident in
your analysis, in your tool set?
As you think about your audience’s needs, consider how familiar they
are with this topic and your terminology. Are they used to working with
data, or will you also need to provide some education as you go? How
much time are they willing and able to spend getting oriented to the
conversation and the way you’re sharing your data? Do they want all the
details and variations in your data, or do they just want you to get to the
point?
In 2020, I worked with the Learning Guild on a research report about the
anticipated stickiness of changes in learning technology use as a result of
the COVID-19 pandemic. We asked a complex grid question of the
respondents: How had their training delivery methods changed as a result of
the pandemic, and how did they anticipate them changing in the future? We
compared their 2019 baseline with March–May 2020 (when the survey was
open), the second half of 2020 (when many people at the time expected to
be back to “normal”), and then two future time periods: 2021, and 2022 and
beyond. We asked about six different training delivery modalities (Torrance
2020).
When it came time to report our data, I was very excited with the
visualization I chose, showing both mode (using a heat map) and mean on
the same graphic for each modality (Figure 13-1).
I thought this was quite clever, and I appreciated the visual designer’s
work in bringing this data to life. The trouble was that every time I showed
this chart to someone (and we have several of these, one for each learning
modality) I had to stop and explain how to read it. In fact, the final version
of the analysis report included an entire page that served as a key to these
charts (Figure 13-2)!
Give It a Try
Here’s a Hypothetical
Become a student of data visualizations in your everyday life. Harvard
Business Publishing, the Federal Reserve Economic Data site (FRED), and
the New York Times have libraries of visualizations you can explore, but
you can also find charts and graphs in all sorts of places. Ask yourself:
What message is being conveyed? Who might the intended audience have
been for this data? See if you can identify any data distortions at work.
Do It for Real
Make a plan for communicating about the data you have collected. Begin
with your purpose and audience, then consider your message and what data
visualizations you will use. Will you use spreadsheet software like
Microsoft Excel or Google Sheets, the visualization functions of a learning
analytics platform, or tools like Microsoft PowerBI, Tableau, Looker, or
other visualization software? Give your report, dashboard, or presentation a
pilot test with people who understand or share the needs of your intended
audience and see how they respond to your work. Adjust accordingly.
Dashboarding Data for an Equity, Diversity, and
Inclusion Program
By Emma Weber, CEO, Lever–Transfer of Learning
2. Leverage Graphics
Never, never, never use a plain bullet slide. This tedious, wall-of-text style
is what people mean when they say “death by PowerPoint.” Instead, use
graphs, frameworks, images, and icons to make your insights
understandable, memorable, and actionable.
To properly communicate your analysis, you should:
• Support your story with data and graphs. This goes without
saying in a book about analytics. The typical problem I see in data
visualization is the equivalent of the wall of text: too much data in
each graph without a clear insight, or too many graphs on one page
without a clear story. Instead, use color, size, enclosures, and arrows
to help your audience see your main point.
• Use frameworks, images, and icons to convey key points. Even if
your audience has the fortitude to read your wall-of-text slides, they
won’t be able to remember all 10 bullets on the page. Instead, chunk
up your ideas into related buckets and create a framework. They’ll
have a better shot at remembering your three categories. And the
three categories will help them retain more of the 10 bullets. Images
and icons help engage your audience and illustrate key points; just
make sure your images and icons are relevant. Adding a stock photo
of employees around a conference table isn’t helping.
3. Reduce Noise
In electrical and audio engineering, the signal-to-noise ratio measures the
relationship between the desired signal and the undesired noise. For
example, imagine listening to a baseball game on AM radio driving out of
town in a storm. (You remember AM radio—that’s the button you hit by
accident when connecting your phone to your car’s Bluetooth.) The signal
is the announcer’s voice, the crack of the bat, and the crowd’s roar. The
noise is the static from the lightning and the fading signal. Anything you
put on your slides that is unnecessary to communicate your key points is
noise. To reduce noise, you should:
• Leverage a standard template with a simple, elegant design.
When collaborators use inconsistent templates, it causes the
“Frankenstein effect,” where differences in design, style, and tone
interfere with communication. Work with your brand team to
develop a clean template with crucial design elements built in, like
color palette, line spacing, page numbers, and corporate logos. Then
build a slide library of standard slides, including cover page, agenda,
graphs, frameworks, and section dividers, so authors don’t have to
reinvent the wheel. Figure 13-4 provides an example of a graph with
common noisy elements called out. Figure 13-5 shows the same
graph with the noise removed. Notice how your brain can focus on
interpreting the data vs. searching for the signal through all the noise.
We’ll finish out this book with an eye to the future: building scale and
maturity in your learning analytics work. To do that, we’ll focus on the five
elements of the foundation discussed in chapter 6—strategy, skill set,
systems and data supply, statistics and data science, and relationships—and
reflect on how they can evolve as you move from early prototype projects
to a full-scale implementation. This is a new space in our industry, and few
organizations are doing deep analytics across all areas of their learning
ecosystem. As the industry matures, you’ll see more and more organizations
taking this on and (hopefully) sharing their case studies and insights for
others to follow.
Prototyping
Very often, early analytics projects are taken on by one or two interested
people in the L&D team—such as you—and it is not unusual for these
projects to be considered side projects, done in “spare time” in a matter of
weeks. The goals of these projects are often to see what is possible, verify
that they actually can get data out of a learning experience, and demonstrate
a proof of concept. Many people who have participated in the xAPI
Learning Cohorts have taken on prototyping projects aiming to extract and
review data from e-learning, LMSs, virtual classroom platforms, apps,
business data, or even voice response systems. In this stage, the data
analysis is often secondary to simply proving that data can be extracted and
reviewed, although if enough data is gathered, some basic analysis can be
performed.
The goal of the prototyping stage is usually to get resources (both time
and budget) and permission to take on a pilot project.
Piloting
Pilot projects are generally one-off projects that receive focused attention
by instructional designers and the business as a first foray into learning data
analytics.
Since the purpose of these early experiments is often to gain support and
funding for using data and analytics in production projects, it is important to
ensure that the analysis you are conducting has real meaning to the
business. It may be a simple analysis, but it should be as accurate as your
data allows it to be. The goal of the pilot projects is to find out what the
organization needs to put in place to move to first production projects.
That said, pilot projects typically begin before an organization has fully
committed to an analytics strategy and before they have fully assembled an
infrastructure, so the data you gather and the analytics you conduct may
serve a short-lived purpose. In some cases, these pilots take place in part to
vet the vendors of their LMS, LRS, and other ecosystem components.
A client of mine, after gaining some basic familiarity with xAPI,
assembled a team to create their first pilot project. The L&D team had
created an e-learning course with a great deal of flexibility in how learners
progressed and how much content they consumed. They were curious to
know how learners actually used the course, and they knew that this
information was not available using SCORM. We shared some best
practices for using JavaScript triggers in their e-learning course to
instrument the data collection, and they set up a trial account with an LRS
provider. Two members of the data science team who were newly dedicated
to L&D participated in the project as well. Because the project was moving
faster than the organization’s overall learning data infrastructure work was
moving, we all went into it knowing that the data we collected and the
analysis we would do would likely not be compatible with the future data
model, and the business sponsor agreed that this would be an acceptable
potential outcome. In this project, the learning developers built skills in
designing and instrumenting for data, and the entire team learned a bit about
what was possible with learning data and analytics at the individual course
level.
Multi-Project Implementation
Between first production projects and full-scale implementation,
organizations will be running several sets of projects, perhaps not all
integrated with one another. This will be a time of experimentation,
learning, and growth as the skills required for working in a data-centric
environment are transmitted through the L&D team, business sponsors
become used to this as normal, and the organization’s learning technology
ecosystem is built out.
Full-Scale Implementation
Full-scale implementation is as much a business initiative as it is an L&D
one. It is a cultural shift in how we approach L&D as a data-driven activity
as well as a technical exercise.
Organizations beginning their journey at full-scale implementation may
find that their L&D teams have skills gaps in data, analytics, visualization,
and technology, and that resolving these can be accomplished with training
and hiring onto the team. Building cross-functional connections with the
business intelligence team can also help.
Full-scale implementation is generally accomplished in phases, working
either by business unit or by learning technology to limit the number of
moving parts in any given phase. Some considerations for full-scale
implementation include:
• Data governance. This includes setting up and communicating the
common ways in which all learning developers will collect data in
their work so that it can be easily analyzed. Data governance
describes the structure and processes by which the organization
keeps its data neat and tidy. It includes common ways to structure
learning asset catalogs, the data standards that will be used, and who
and how new content (and its data) can be added.
• Data privacy, confidentiality, and security. It is likely that your
organization already has these protections in place, and you will want
to make sure that your work falls under this purview.
• Organization design. Some organizations create a learning analytics
team with designated data scientists as members. In other
organizations, the business intelligence team supports the learning
analytics work.
• Conversion strategy. You will need to consider whether to
backtrack and begin collecting data from already existing learning
experiences, or to only focus on future new design work. One
organization I have worked with decided to limit their conversion
effort to only the most-used resources in their current ecosystem, and
focus instead on gathering data from any newly created learning
items in their catalog.
Looking across the foundational elements for each phase of an analytics
implementation gives insight to the work ahead (Table 14-1). Keep in mind
that you may take a somewhat different path through these stages, moving
faster or slower as the specific needs and capabilities of your organization
support.
As you and your organization chart a path to a full-scale implementation
of learning data and analytics, be sure to think across all these foundational
elements because they are linked together. In some cases, the linkages will
come naturally as, for example, your IT team gets involved while you seek
to integrate new software. In other cases, you may need to intentionally
seek out support from other teams to learn what they are doing with their
data and analytics, and perhaps borrow resources.
Give It a Try
Here’s a Hypothetical
Consider the current state of your organization’s learning data and analytics,
its capabilities and needs, and the resources available to you. What could be
a possible path forward? What steps that would you need to take to reach
the next phase in your implementation? Consider whether this is something
you can discuss with your learning leadership or business sponsors.
Do It for Real
Use the table in this chapter to identify where your organization is in its
implementation of learning analytics right now. What might be missing at
your current stage? What could be a possible path forward? What steps will
you need to take to reach the next phase in your implementation? Consider
whether this is something that you can discuss with your learning leadership
or your business sponsors.
Bonus Points
Submit your story as a case study for others to follow, either in the industry
press or as a conference presentation. This emerging space is hungry for
examples of learning data and analytics implementations and the lessons
you have learned along the way.
The Learning Impact Engine
JD Dillon, Chief Learning Architect, Axonify
Correlation is not causation. Two events may happen one after the other, but
that doesn’t mean the first event caused the second. Sales may increase after
the implementation of a new sales training program, but this doesn’t verify
that the program caused the improved result (even if L&D desperately
wants it to be true). Lots of factors can influence business outcomes.
Marketing promotions, changes to employee compensation, a new CRM
platform, and increased customer footfall due to nice weather may have
also influenced sales numbers. Training likely played a role, but every other
department will claim their strategies caused the positive outcome too.
To establish true relationships between learning and performance, L&D
needs a learning impact engine.
L&D can feed the learning impact engine with a variety of workplace
data points, including:
• Engagement: How often employees participate in various learning
activities
• Consumption: Which content objects employees use
• Knowledge: How each person’s knowledge of important topics
changes over time
• Confidence: How their self-reported confidence shifts over time
• Behavior: How they are or are not applying their knowledge on the
job
• Results: Business outcomes associated with employee performance
The learning impact engine applies a machine learning model
(customized to an organization based on its unique data requirements) to
establish relationships between changing data points. Then, L&D can train
the model using historic learning and performance data. This allows the
technology to learn how changes in different data points relate to one
another over an extended period. As L&D gathers new data, we can test the
model’s predictive ability and accuracy.
The engine applies its relational understanding of the organization’s data
to isolate the impact of learning activities on business outcomes,
represented as a percentage. For example, the engine may determine that a
specific training program caused 15 percent of the company’s quarterly
sales result. L&D can then convert this percentage to a monetary value and
calculate the return on investment for the training initiative. If provided
with the right data, the engine may also be able to determine how other
factors, such as marketing, coaching, or seasonality, influenced the business
outcome. Otherwise, the engine will still isolate these influences but label
them as “unknown.”
L&D can take this process a step further by applying the learning impact
engine to proactively adjust learning strategies. Consider an example of the
engine used to measure the impact of training within a warehouse
operation. It determined a training program meant to accelerate shipments
was not having the desired impact. Rather than waiting for the flawed
program to complete and making assumptions about its impact, L&D
evaluated the training, found that it was focused on the wrong employee
behaviors, and adjusted the content in-flight. The engine’s insight proved
critical to recording the desired improvement in business results upon
program completion.
A learning impact engine can also enable hyper-personalized learning
experiences. For example, it may recommend a topic, activity, or content
object to an individual based on its proven ability to affect their
performance goals. Rather than modeling our approach to personalized
learning on Netflix and relying on subjective peer ratings and content
consumption patterns, L&D can take an outcome-based approach and adapt
learning to focus on tactics that really work.
Learning impact engine technology is available right now. We can finally
measure the impact of learning! So why is L&D still struggling to prove
our value?
Technology isn’t the problem. It’s the data needed to power the system
that’s lacking. L&D doesn’t just need more data. We need a greater variety
of data to establish connections between our solutions and the results we’re
trying to achieve. To fix measurement, we must first evolve our learning
strategies. Traditional training is data-poor. There’s only so much insight
you can gather from course completions, test scores, and reaction surveys.
To understand the relationship between learning and performance, L&D
must assess how people change (or do not change) over time as a result of
our programs. This starts with adopting continuous learning tactics, such as:
• Microlearning: Targeted content provides more granular insight into
specific concepts
• Reinforcement: Practice activities provide insight into how people’s
knowledge and skills change over time
• Observation: Behavior data provides a critical connection between
what people know and if or how they apply that knowledge on
the job
Once we’ve expanded our data collection practices, L&D must partner
with stakeholders to access the business data needed to power a learning
impact engine. This may include data related to sales, safety, productivity,
customer satisfaction, or anything else the organization measures associated
with our learning programs. When partnering, L&D must explain how this
sensitive data will be used to improve support and drive business results.
This data is essential to establishing the relationship between learning
activities and performance outcomes, so we must ensure our stakeholders
trust us with it.
Measuring learning impact is difficult, but it’s not impossible.
Technology plays a critical role. Good data is essential. But fixing L&D
measurement begins with making it a priority. If we want to move L&D
forward and get that “seat at the table,” improving our data practices must
be our top priority. Otherwise, causation will remain out of reach, and
“butts in seats” will be the measure of our contribution to the organization.
Scaling the Learning Analytics Team
By Tiffany Jarvis, Director of Learning, Edward Jones
When I joined the learning analytics team at Caterpillar, the team had
already begun their data and analytics journey. The situation when I started
was to establish the learning measurement and evaluation processes to
enable learning analytics across our dealer-facing programs. The first
milestone was to gather all the historical data across our different learning
systems and provide an overview of our dealers’ learning activity across the
learning ecosystem. With more than 160 independent dealers that account
for 150,000 learners worldwide, the ability to track and analyze learning
data at scale is critical to ensure the success of our learning program goals.
From a metrics standpoint, our leaders needed to identify our “true”
active user population and their adoption patterns. Therefore, the focus was
to identify what learning modality our dealer learners were using (formal
versus informal activities) and gauge the adoption of our key learning
programs (leadership, sales, and service career development).
Achieving this goal required a deep understanding of how the historical
data and the current data were generated, so data assumptions could be
developed to avoid misinterpreting it. Once completed, we used those
assumptions to develop our data import and normalization plan, which
consisted of importing the historical data from our LMS into our external
LRS, as well as integrating each of our learning systems (including third-
party content, videos, and assessments) so we could normalize the data
using the xAPI profile.
The use of ingestion templates helped us properly import the data to our
LRS and then use our learning analytics platform to analyze the data and
find patterns across it. In this way, we could check the validity of our
assumptions and act on which data was needed and which was not. The data
normalization through xAPI ensured we could evaluate the entire data lake
regardless of having different data sources.
This provided the foundation of our learning analytics strategy, which
encircled three types of standard dashboards for our learning programs’
stakeholders: program readiness, program skills and knowledge transfer
effectiveness, and learning experience effectiveness.
• The program readiness dashboards focused on using descriptive
analytics around target population adoption, providing insight to the
question “Is our target audience completing the critical learning
programs on time for product or services deployment?” This
question goes from a global view to a dealer-by-dealer view.
• The program skills and knowledge transfer effectiveness
dashboards focused on using diagnostic analytics, which combines
formative assessments (such as gap assessments or self-assessments)
with summative assessments (such as tests, exams, or performance
assessments) associated with the critical learning programs. The gap
assessments determine who should be going to what learning
program or competency development plan, and the performance
assessments validate the competency level. This type of dashboard
focused on a specific learning program in a given timeframe and
requires a holistic approach rather than focusing just on specific
learning activities. The questions center on gauging the effectiveness
of a given learning program to improve the performance of a target
audience.
• The learning experience effectiveness dashboards focused on the
quality of the learning program, mostly looking at engagement
behavior displayed through multiple visualizations using trending
and correlation analysis to visualize digital learning activities
patterns (such as in e-learning and videos engagement). This one
helps our learning experience design team identify what modalities
have the highest reach based on the target audience and topics (such
as leadership, sales, and service programs).
This structure enables us to provide a common framework to share
insights into our learning programs, by showing adoption rates (total people
engaging with the learning activity versus total people completing the
activity or program).
My advice for others looking to get started and build a similar structure
is first to figure out what data you are already collecting. Sometimes we
may be eager to start from scratch, but if your organization is already
collecting data from learning activities, start with them and identify what
data is valuable or not.
From a deployment planning perspective, the approach that worked for
me was:
• Normalize your data. For us, the key was having an external LRS
that could ingest both learning data and business data and consolidate
it into a single data lake. In this way, the data can be easily structured
to show the insights we were looking for.
• Experiment with your stakeholders. A big success for adoption
was including our stakeholders on the development phase of our
analytics dashboard. This helped us ensure alignment with our
business requirements as well as explore new ways to use the data.
And finally, discuss frequently with your stakeholders how they intend
to use the data. Ask questions about their business requirements, because
they may shift often. Also, encourage them to brainstorm on what success
and failure look like from a data perspective, so standard metrics can be
developed based on those scenarios.
Acknowledgments
“We are uncovering better ways of developing software by doing it and helping
others do it.”
About the time I kicked off the writing of this book in earnest (or should
have kicked it off), I got some fantastic advice that has helped me push
through some of the big blockers in life—including writers’ block: “Just
skip to the good parts.” So many people—Michelle, Matt, Maria, the entire
team at TorranceLearning—have helped me see the “good parts” of life
along the way. And a well-placed “you’ve got this” goes a long way (I’m
looking at you, Emmet). Thank you.
Are you ready for more? I found these books interesting and useful along
my own journey, some of which provided insight and direction for this book.
Others were interesting beach reading.
The Data Detective: Ten Easy Rules to Make Sense of Statistics Tim
Harford (Riverhead Books, 2020)
Behind Every Good Decision: How Anyone Can Use Business Analytics to
Turn Data into Profitable Insight Piyanka Jain and Puneet Sharma
(American Management Association, 2015)
The Art of Statistics: How to Learn From Data David Spiegelhalter (Basic
Books, 2019)
The Functional Art: An Introduction to Information Graphics Alberto Cairo
(New Riders, 2013)
The Big Book of Dashboards: Visualizing Your Data Using Real-World
Business Scenarios Steve Wexler, Jeffrey Shaffer, and Andy Cotgreave
(Wiley, 2017)
Show Me the Numbers: Designing Tables and Graphs to Enlighten Stephen
Few (Analytics Press, 2012)
Learning Analytics: Using Talent Data to Improve Business Outcomes John
R. Mattox II, Peggy Parskey, and Cristina Hall (KoganPage, 2020)
Investigating Performance: Design and Outcomes With xAPI Janet Laane
Effron and Sean Putman (MakingBetter, 2017)
Measurement Demystified: Creating Your L&D Measurement, Analytics,
and Reporting Strategy David Vance and Peggy Parksey (ATD Press,
2021)
Making Sense of xAPI Megan Torrance and Rob Houck (ATD Press, 2017)
Data Story: Explain Data and Inspire Action Through Story Nancy Duarte
(IdeaPress, 2019)
The Visual Display of Quantitative Information Edward R. Tufte (Graphics
Press, 2001)
References
A
Action Mapping, 95, 95t–96t
adaptive learning, 80
ADDIE, 165–166, 165f, 166f
Advanced Distributed Learning (ADL), vi, 36, 86–87, 162–163
aggregate data, 136
Agile approach, 105–106, 165
AI (artificial intelligence), 26
AICC specification, 36
all-in-one learning platforms, 7
analytics, 21. See also data and analytics
categories of, 77, 78f
levels of, 22–23
performing, 80
Anderson, Chris, 3
Ann Arbor Hands-On Museum Digitally Enhanced Exhibit Program
(DEEP), v
answer sources, 103, 103t
artificial intelligence (AI), 26
assessments, 46–47, 126
audience, 176–178, 177f, 178f
augmented reality, 128–129
Axonify, 200–202
Azar, Ulduz Berenjforoush, 121–122, 136
B
Bersin, Josh, 70
Betts, Ben, 33
BI (business intelligence), 7, 25
big data, 20–21
blended scenarios, using data across, 88–89
Bottom Line Up Front (BLUF), 179
Boulton, C., 4
Brant, Caroline, 22–23
Britz, Mark, 92
browser data, 130
building scale and maturity, 189–208
case studies, 200–208
first production projects in, 191–194
foundational elements in, 195–198, 196t–197t
and framework for using learning data, 188–189, 189f
full-scale implementation in, 194–195
multi-project implementation in, 194
pilot projects in, 190–191
prototyping in, 190
business acumen, 106
business intelligence (BI), 7, 25
C
categorical data, 57
Caterpillar, 206–208
Caulkin, Simon, 3
causation, 66, 200
Cavalier, Josh, 14
Cavoukian, Ann, 134
chart design, 179–180
chat bot data, 128
childcare metrics, 11
“cleaning” data, 167
cmi5 specification, 36, 40
communicating data, 175–188. See also visualization(s)
audience for, 176–178, 177f, 178f
case studies, 182–188
with dashboards, 175
with data presentations, 176
design of charts and graphs for, 179–180
framing message in, 178–179
purpose in, 176
through reports, 175
completion of learning metric, 43–46
confidence-based testing, 46
consumer-focused software, 4
continuous improvement, 5–6, 170
control groups, 121–122
convenience sampling, 64
conversion strategy, 195
Corbett, Andrew, 50–51
correlation, 66, 200
COVID-19 pandemic, vi, 177
Crayton, Kimberly, 73–74
Critical Equity Consulting, 121–122, 136
customer service metrics, 9
D
dashboards. See data dashboards
data, 17–18, 20f. See also data and analytics; workplace learning data
categories of, 77, 78f
communicating (See communicating data)
meaningfulness of, 18–19
storing (See storing data)
types of, 56–58
uses of, 78–82, 78f, 82f
data analysis
in data science, 23
iterating on, 168–170
storing data from, 155
data and analytics, 3–13
case studies, 14
to drive decision making, 8–12
for L&D function, 7–8
reasons for caring about, 4–6
data capture, 137–149
case study, 147–149
designing experiences for, 6
improvising with, 143
from learning experiences, 140–142, 141f, 142f, 147–149
from non-digital experiences, 142–143
from pre-built systems, 137–140, 139t
privacy, confidentiality, and security with, 143–145
data collection, 6, 167–168
data confidentiality, 144, 145, 195
data dashboards, 25
communicating data with, 175
for equity, diversity, and inclusion program, 182–184, 184f
iterations of, 169–170
data ethics, 134–135
data gathering, 23. See also planning
data governance, 195
data lakes, 154–155
data literacy, 5
data needs, 115–118, 118t, 120, 122
data preparation, 23
data presentation, 23, 176
data privacy, 136, 143–145, 195
data projects, ad hoc nature of, 173
data protection and privacy, 136
data science, 23
data security, 78–79, 144–145, 195
data sources, 123–136
case studies, 134–136
HR systems, 123–125
learning delivery platforms, 129
learning experiences, 125–129
moving from core question to, 118f
for things that don’t happen online, 131–132
upstream, storing data from, 151–152
data specifications and standards, 35–41
AICC, 36
case study, 39–41
choosing, 37
HR Open Standards, 36
IEEE standards, 87
for learning analytics, 39–41
1EdTech Caliper Analytics, 36
SCORM, 35–36
usefulness of, 37
xAPI, 36
DataStory (Duarte), 176
data visualization(s), 25, 79–80
audience for, 177–178
cautions about, 180–181
dashboards, 175
design of charts and graphs for, 179–180
iterating on, 168–170
purpose of, 175
storing data from tools for, 155
data warehouses, 154–155
decision making, data for, 8–12
DEEP (Ann Arbor Hands-On Museum Digitally Enhanced Exhibit
Program), v
definitions, 17–29
case studies, 30–33
evolution of, 26
of terms, 17–26 (See also individual terms)
Department of Defense (DoD), 86–87
dependent variables, 65–66, 65f
descriptive analytics, 22
devices, data from, 130
diagnostic analytics, 23
digital dexterity, 5
digital transformation, 4
Dillon, JD, 200–202
DoD (Department of Defense), 86–87
Duarte, Nancy, 176
E
EDI (equity, diversity, and inclusion) dashboard, 182–184, 184f
EDLM (Enterprise Digital Learning Modernization), 86
Edward Jones, 70–72, 203–205
effects of learning on performance, 48
Effron, Janet Laane, 55, 120, 173
engagement
to influence, 185–188, 187f
with learning experience, 47–48
Enterprise Digital Learning Modernization (EDLM), 86
equity, diversity, and inclusion (EDI) dashboard, 182–184, 184f
Experience API (xAPI), v, vi, xii, 7
flexibility of, 40
learning record stores, vi, 80, 153, 155–156
xAPI Learning Cohorts, vi, 36, 40, 190
exploratory analysis, 112–113
F
finance metrics, 8–9
first production projects, 191–194
Fox, Marina, x
framing messages, 178–179
Frank Porter Graham Child Development Institute, 88–89, 160–161
“Fuel Business Strategies With L&D Analytics” (Pease and Brant), 22–23
full-scale implementation, 194–195
G
goals
aligning data gathering/use with, 91–94, 93t, 94t
prioritizing, 103–105
Goldberg, Becky, 15
graph design, 179–180
H
Hall, Cristina, vii, 22, 174
healthcare metrics, 10
hospitality metrics, 10
HR Open Standards, 36
HR systems data, 123–125
HubSpot, 57
human resources metrics, 9
hypothesis formation, 111–122
case studies, 122
exploratory analysis in, 112–113
and identification of specific needs, 115–118, 118t
scientific method in, 111, 112f
testing hypotheses, 113–115, 114t–115t
I
immersive games, 128–129
improvising data capture, 143
independent variables, 65–66, 65f
influence, 185–188, 187f
information, 18–19, 20f
instructional design frameworks metrics, 95–97, 95t–97t
integrating workflows, 81–82, 82f
inter-rater reliability, 48
intranet data, 127–128
intuition, 106
IT, data storage role of, 158
item difficulty, 46
item discrimination, 46–47
iterating, 165–173
and ADDIE approach, 165–166, 165f, 166f
case study, 173
for continuous improvement, 170
on data analysis and visualizations, 168–170
on data collection, 167–168
LLAMA approach to ADDIE, 166, 166f
J
Jarvis, Tiffany, 70–72, 203–205
job performance data, 130–131
johnpolkandassociates.com, 185–188
JoshCavalier.com, 14
K
Katzell, Raymond, vii
key performance indicators (KPIs), 24
Kirkpatrick, Donald, vii
Kirkpatrick Model, 97–98
Kirkpatrick Partners, 97–98
Kliewer, Matt, 147–149
knowledge, 19, 20f
knowledge base data, 127–128
KPIs (key performance indicators), 24
L
LAI (Learner Adoption Index), 70–72
LAMM (Learning Analytics Maturity Model), 33
LAPs (learning analytics platforms), 80, 155
Learner Adoption Index (LAI), 70–72
learner analytics, 23
learning analytics, 22
data specifications for, 39–41
to drive results, 109
levels of, 23
scientific method applied to, 111, 112f
Learning Analytics (Mattox, Parskey, and Hall), vii, 22
Learning Analytics Maturity Model (LAMM), 33
learning analytics platforms (LAPs), 80, 155
learning analytics team, scaling, 203–205
learning data
to drive results, 109
framework for using, 82–83, 83f, 189, 189f
lack of, vii–viii
from the workplace (See workplace learning data)
learning data ecosystem
considerations for operating within, 162–163
mapping, 156–158, 156f, 157f
working within, 160–161
learning delivery platform data, 129
learning experience(s)
data capture from, 140–142, 141f, 142f, 147–149
as data source, 125–129
incorporating data in, 80
storing data from, 152–153
learning experience platforms (LXPs), 129, 152
learning function, continuous improvement of, 5–6
Learning Guild, 177
learning impact engine, 200–202
Learning Locker, 57
learning management systems (LMSs), v, viii, 7, 129, 152
learning measurement framework metrics, 97–102
learning organization, 103, 103t
Learning Pool, 33
learning program analytics, 23
learning record stores (LRSs), vi, 7, 80, 153
learning-specific metrics, 43–53
case studies, 50–53
completion of training, 43–46
effects of learning on performance, 48
engagement with learning experience, 47–48
tests and assessments, 46–47
learning transfer analytics, 30–32
Learning-Transfer Evaluation Model (LTEM), 98–102
Lee, Stella, 134–135
LeMoyne Institute, vi
leveraging graphics, 186
Lever-Transfer of Learning, 30–32, 182–184
live classroom data, 127
LMSs. See learning management systems
Lot Like Agile Management Approach (LLAMA), 166, 166f
LRSs (learning record stores). See learning record stores
LTEM (Learning-Transfer Evaluation Model), 98–102
LXPs (learning experience platforms), 129, 152
M
machine learning, 26
manufacturing metrics, 11
math anxiety, ix
Mattox, John, vii, 22, 174
maturity. See building scale and maturity
mean, 68–70, 69f, 70f
measurement and evaluation processes, establishing, 206–208
Measurement Demystified (Vance and Parskey), vii, 5–6, 21, 79
measures, 23
of central tendency, 58–60, 59f, 60f
for decision making, 8–12
of spread, 60–63, 61f–63f
median, 68, 69, 69f
metrics, 23
aligning data gathering/use with, 91–94, 93t, 94t
for decision making, 8–12
from instructional design frameworks, 95–97, 95t–97t
from learning measurement frameworks, 97–102
for learning organizations, 103, 103t
learning-specific (See learning-specific metrics)
prioritizing, 103–105
related to learning organization outcomes, 103, 103t
Mitchell, Derek, 52, 53f, 109
mode, 68, 69f, 70f
Moore, Cathy, 95
Morgan, Wendy M., 88–89, 160–161
multi-project implementation, 194
Myers, Rodney, 73–74
N
National Institute of Standards and Technology, 24
noise reduction, 186–187, 187f
noisy data layer, 154
non-digital experiences
data capture from, 142–143
as data source, 131–132
Novo Nordisk, 52, 53f
null hypothesis, 114–116, 114f–115f
O
objectives and key results (OKRs), 24
observation data, 126–127
1EdTech Caliper Analytics, 36
organization team designs, 195
outcome metrics, 103, 103t
P
Panetta, K., 5
Paradox Learning, 134–135
Parskey, Peggy, vii, 5–6, 21, 22, 79, 174
Pease, Gene, 22–23
people analytics, 121
performance
data from support tools for, 127–128
effects of learning on, 48
predictive analytics for, 30–32
student, learning data for insights into, 50–51
using analytics to evaluate, 15–16
personalized learning, 80
Phillips, Jack, vii
pilot projects, 190–191
planning, 91–109
Agile approach in, 105–106
to align with business goals and metrics, 91–94, 93t, 94t
case study, 109
metrics from instructional design frameworks, 95–97, 95t–97t
metrics from learning measurement frameworks, 97–102
metrics related to learning organization outcomes, 103, 103t
prioritizing list of goals, questions, and metrics, 103–105
using business acumen and intuition in, 106
Polk, John, 185–188
pre-built system data capture, 137–140, 139t
predictive analytics, 23, 30–32
prescriptive analytics, 23
PriceWaterhouseCoopers, vi
Privacy by Design Framework, 134–135
program rollouts, testing, 121–122
Project CATAPULT, 40
Project Tin Can, v
prototyping, 190
purpose, in communicating data, 176
Q
qualitative data, 56–57
QuantHub, vi
quantitative data, 56
questions, prioritizing, 103–105, 103t
R
range, 60–63, 61f–63f
reporting, 24–25, 79–80. See also communicating data
reporting strategy, 175
retrospectives, 170
Riley, Alfonso, 206–208
Rustici Software, 39–41
Rutherford, Tammy, 39–41
S
sales metrics, 8
sampling, 63–65
sampling bias, 65
scaling, 83. See also building scale and maturity
scientific method, 111, 112f. See also hypothesis formation
scorecards, 25
SCORM (Sharable Content Object Reference Model), 35–36, 40
SCORMs. See sharable content object reference models
search feature data, 129–130
self-directed e-learning data, 125–126
sentiment analysis, 56–57
70-20-10, 95, 96t
Sharable Content Object Reference Model (SCORM), 35–36, 40
sharable content object reference models (SCORMs), v, viii, 7
simulation data, 128–129
Six Sigma, 63
skills needs identification, 52, 53f
Sky, 109
Slack, 148
Smith, Brent, 86–87, 162–163
Social by Design (Britz), 92
social learning data, 128
Sparks, Sarah, ix
spread, measures of, 60–63, 61f–63f
Spurious Correlations (Vigen), 66
stakeholders, 92, 120
standard deviation, 62–63, 63f
standards. See data specifications and standards
statistical significance, 66–67
statistics, 55–72
case studies, 70–74
correlation and causation, 66
measures of central tendency, 58–60, 59f, 60f
measures of spread, 60–63, 61f–63f
relationships between variables, 65–66, 65t
sampling, 63–65
statistical significance, 66–67
types of data, 56–58
storing data, 78–79, 151–163
case studies, 160–163
from data analysis and visualization tools, 155
from data warehouses and data lakes, 154–155
from learning analytics platforms, 155
from learning experiences, 152–153
from learning management systems and learning experience platforms,
152
from learning record stores, 153
options for and approaches to, 155–158, 156f, 157f
role of IT in, 158
tools and services for, 151–155
from upstream data sources, 151–152
storytelling, 184
survey data, 130
T
Talent Development Capability Model, 5
technology metrics, 11–12
terminology. See definitions
tests and testing
data from, 126
of hypotheses, 113–115, 114t–115t
learning-specific metrics from, 46–47
of program rollouts, 121–122
when iterating on data, 168
Thalheimer, Will, 98
Total Learning Architecture (TLA), 79, 86–87, 162–163
Trane technologies, vi
Travelers Insurance, 15–16
Turning Learning into Action, 182
U
Underwriting Professional Development Program (UPDP), 15–16
University of California–Davis, 50–51
UPDP (Underwriting Professional Development Program), 15–16
upstream data sources, storing data from, 151–152
US Navy program effectiveness, 73–74, 74f
V
Vance, David, vii, 5–6, 21, 79
variables, 65–66, 65t
variance, 61–61, 62f
video design, 14
video interaction data, 126
Vigen, Tyler, 66
virtual classroom data, 127
virtual reality data, 128–129
visualization. See data visualization(s)
“Vs of data,” 104–105
W
Weber, Emma, 30–32, 182–184
wisdom, 19, 20f
workflows, 81–82
workplace learning data, 77–89
case studies, 86–89
categories of data and analytics in, 77, 78f
foundational elements for using, 82–83, 82f
uses of, 78–82, 78f, 82f
X
xAPI. See Experience API
xAPI Learning Cohorts, vi, 36, 40, 190
xAPI Learning Record Stores, vi, 7, 80, 153
Y
YouTube, 14
About the Author