Data Analytics and AI
Data Analytics and AI
Edited by
Jay Liebowitz
Distinguished Chair of Applied Business and Finance
Harrisburg University of Science and Technology
First edition published 2021
by CRC Press
6000 Broken Sound Parkway NW, Suite 300, Boca Raton, FL 33487-2742
Reasonable efforts have been made to publish reliable data and information, but the author and publisher cannot
assume responsibility for the validity of all materials or the consequences of their use. The authors and publish-
ers have attempted to trace the copyright holders of all material reproduced in this publication and apologize to
copyright holders if permission to publish in this form has not been obtained. If any copyright material has not
been acknowledged please write and let us know so we may rectify in any future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced, transmitted,
or utilized in any form by any electronic, mechanical, or other means, now known or hereafter invented, including
photocopying, microfilming, and recording, or in any information storage or retrieval system, without written
permission from the publishers.
For permission to photocopy or use material electronically from this work, access www.copyright.com or contact
the Copyright Clearance Center, Inc. (CCC), 222 Rosewood Drive, Danvers, MA 01923, 978-750-8400. For
works that are not available on CCC please contact mpkbookspermissions@tandf.co.uk
Trademark notice: Product or corporate names may be trademarks or registered trademarks, and are used only for
identification and explanation without intent to infringe.
Typeset in Garamond
by Deanta global Publishing Services, Chennai, India
Foreword..............................................................................................................ix
Preface..............................................................................................................xvii
List of Contributors..........................................................................................xix
Editor...............................................................................................................xxiii
1 Unraveling Data Science, Artificial Intelligence, and Autonomy............1
JOHN PIORKOWSKI
vii
viii ◾ Contents
Index............................................................................................................235
Foreword
Introduction
Analytics and artificial intelligence (AI), what is it good for? The bandwagon keeps
answering: absolutely everything! Analytics and artificial intelligence have captured
the attention of everyone from top executives to the person in the street. While
these disciplines have a relatively long history, within the last ten or so years, they
have exploded into corporate business and public consciousness. Organizations
have rushed to embrace data-driven decision-making. Companies everywhere
are turning out products boasting that “artificial intelligence is included.” We are
indeed living in exciting times. The question we need to ask is, do we really know
how to get business value from these exciting tools?
The core of analytics hasn’t changed much since Tukey wrote “The Future of
Data Analysis” in 1962.* Statistics is still statistics. Linear and non-linear regres-
sion models, various flavors of classification models, hypothesis testing, and so on
have been around for many years. What has changed is the vast amount of data
that is available both inside and outside of organizations and the dramatic increase
in computing power that can be called upon to process it. Moving well beyond the
rigid structures of corporate data in relational databases, organizations are collect-
ing, processing, and analyzing unstructured text, image, audio, and video data.
Today’s deeply technical analytical tools and techniques are truly wondrous. Are
these wondrous techniques being translated to business solutions?
* Tukey (1962) The Future of Data Analysis. Ann. Math. Statist., Volume 33, Number 1, 1–67.
ix
x ◾ Foreword
(what some companies call “quants”*) who analyze and model data. Analysis is a
small part of the much larger data science life cycle† (DSLC) shown in Figure F.1.
I’ve divided the DSLC into what is currently performed by data scientists and what
is currently performed by database and information technology (IT) personnel.
“Quants” retrieve data from company data repositories and analyze it. They use
statistics to build descriptive, prescriptive, predictive, diagnostic, and other kinds of
models. This is analysis. Data science covers analysis plus everything else that has to
do with the data. It is truly the science of data.
Data science starts at data generation. Today, data generation happens on a
massive scale. We’re at the point where data storage is so inexpensive (fractions of
a cent per gigabyte per month on Amazon cloud storage‡) that organizations are
storing anything and everything just in case it’s needed someday. Video cameras are
recording 24/7. Calls, emails, and documents are recorded and stored. With every
smartphone having a camera, consumers are taking pictures and storing them at a
furious rate. It’s estimated that by 2025, 463 exabytes of data will be added to the
* IBM: The Quant Crunch: How the Demand for Data Science Skills is Disrupting the Job
Market.
† The Data Life Cycle. MIT Press 2019. https://hd sr.mit press.mit.edu/pub/577rq08d
‡ https://aws.amazon.com/s3/pricing/
Foreword ◾ xi
global data storage swamp daily.* That’s 463 million terabytes. Little thought is
given to what happens next to all this data. How will the data be processed? What
will it be used for? How can we model this data? AI will help us address these ques-
tions more every day.
We’re just now trying to tackle the immense task of collecting the data that
might hold something interesting for analysis. Entire industries are dedicated to
collecting and organizing semi-structured data (for example, search engine indus-
tries, social media industries, and so on). For the most part, data collection relies on
the data generator to tag and index just what is in that data. Hashtags must be used
to figure out what a tweet means. Audio, video, and still pictures must have index-
ing keywords. However, vast quantities of data remain uncategorized. This is a big
problem and will only get worse as companies realize the value of unstructured
data. Deep learning might be a solution.
Deep learning is an advanced form of AI that is optimized for feature extrac-
tion in unstructured data and especially useful for classification tasks. For example,
when trained on a variety of images, deep learning can classify what kind of leaves
are present in a pile, can extract and identify various objects in a picture, and can
perform facial recognition tasks, to name a few. In unstructured text, deep learning
can identify messages and determine what kind of message it is, what the subject is,
and relate that message to other similar messages.
Data must be processed. Missing, duplicate, and/or conflicting data needs to
be resolved, known as “cleaning.” Data scientists must be able to trust the data
provenance. They need to know where the data came from. “Data wrangling” is the
new data science buzzword. The tools and techniques and theories on how to do it
right are rarely taught and therefore must be learned “on the job,” something most
data scientists are ill-prepared for. They are even less prepared to model data that
has been gathered through AI (such as facial recognition) or analyzed by AI (for
example, predictive search patterns).
Once we have clean data, data scientists must decide what is needed for analysis
and, more importantly, what is missing. This is where data management comes in.
How can we find the data that tells us what we need to know? Decades of work with
structured data in relational databases has produced good results with optimized
and provably correct data structures. However, unstructured data is proving dif-
ficult to manage. Not only is it difficult to wrangle, it’s difficult to store, index, and
optimize. The problem is made worse when unstructured data comes in at different
velocities. AI can assist with this type of data by performing some of the wrangling
by processing data through a “data lake.”† The data lake, compared with the usual
data silos, is a highly scalable, massively parallel data hub that smooths the torrent
of data input. Unstructured data has incredible richness with far more detail than
* https://www.weforum.org /agenda/2019/04/ how-muc h- dat a-is-gener ated- each- day-c f4bd
df29f/
† https://www.cio.com/a rticle/340 6806/ai-un leashes-the-power- of-unst r uctu red-d ata.html
xii ◾ Foreword
what to do with them, so they were stored on a shelf in case they proved useful
someday. Most of these models are still sitting on shelves gathering dust.
Today, there is a surplus of data scientists turning out models that are never
used. Companies are reducing their hiring, and universities that were on the lead-
ing edge of data science education are closing their programs. All of this is caused
by a basic misunderstanding: Data Science is not Analytics. Data science is data and
analysis. Analytics is the bridge between data science and business.
Over the past year, I’ve interviewed data scientists and business managers in
companies across the United States. Data scientists consistently say that managers
can’t ask the right questions and don’t understand statistical models. Managers say
that data scientists don’t understand the business and can’t give them what they
need. Some companies are stepping up and creating a “translator” role to bridge the
gap between data science and management. This is a step in the right direction, but
only a baby step. Translators must understand both data science and the business.
Period. Very soon they will also need to understand AI.
Today, in the early 2020s, this idea is spreading. The data science life cycle now
includes a very important step: “Understand the business.”* The problem is that
this is not currently the data scientist’s job. Data science departments are usually
set apart from the rest of the organization. They have very little experience with
the organization as a whole and even less with specific divisions or departments.
Similarly, managers and executives are experts in their part of the organization but
have very little experience with statistical models or the intricacies of data. Once
again, business schools need to step up and provide training and re-training for
both data scientists and managers.
Analytics must be recast not as a technical statistical field but rather as a disci-
pline unto itself. Look at any current university’s “Business Analytics” program and
you’ll see a lot of data science and very little business. This is not analytics as the
term should be used. Whereas analysis is a small part of the whole of data science,
data science itself is a relatively small part of analytics. The entire analytics life cycle
is shown in Figure F.2.
The analytics life cycle (ALC) covers everything from start to finish all within a
business context. The steps are straightforward and relatively intuitive. It’s useful to
go through a checklist of all the steps even if you know that nothing has changed.
The littlest things may end up having tremendous impacts. The steps are:
6. Develop questions
7. Use the data science life cycle
8. Understand the analytic model through visualization
9. Evaluate the analytic model
10. Apply the model back to the questions
(send unanswered questions by the model back through the process)
11. Apply the answered question(s) to the business problem
The ALC is far more than data science. It requires a deep understanding of the
business, including what the business is, the business environment, its goals and its
strategies, and what problems are being addressed. All this is required for generat-
ing questions that can be answered by data science and analysis. Only then will the
two questions posed earlier be addressed: obtaining business value and business
solutions from analytics.
What business manager or executive has the time or the skill for all this? If they
could do it, they would be doing it. Managers and even top executives think they
know what’s going on in the organization and where the organization is headed.
Many times, they are very, very wrong. What’s needed is an AI application that can
scan the organization’s environment, internal processes, and structures to deter-
mine if the organization’s strategy is indeed leading them in the right direction
toward the desired goals. This is very similar in theory to DSAI. Enter management
AI (MAI).
MAI can advise management and top executives about where the organization
is actually headed. It can suggest changes to strategy or even suggest adjusting
organizational goals. A continuously operating MAI can suggest organizational
Foreword ◾ xv
adjustments in real time. With a good idea of the proper strategy, MAI can scan
for problems that are keeping the organization from executing strategy and develop
questions that can be passed along to the DSAI for answering. MAI has the poten-
tial to mine things we currently view as ephemeral such as managerial intuition
and hunches.*
DSAI has the potential to produce models that are accurate, relevant, and
understandable in the business context. This can help managers and executives do
what they do best.
Today, analytics is still in its infancy. It’s a loose conglomeration of statistics,
database/IT, and organizational management. Analytics brings all this together
into a coherent whole. Supplemented with AI in the form of DSAI and MAI, we’re
moving into a very exciting future.
* Nelson, Clark, and Stewart. (2018) Hunch Mining. Hawaii International Conference on System
Sciences, Waikoloa, HI.
Preface
xvii
Finally, I am very lucky to have such a fantastic, loving supportive family who
wonders whether this might be my last book! I’ll keep them guessing…
Jay Liebowitz
Washington, DC/Philadelphia
List of Contributors
xix
xx ◾ List of Contributors
Dr. Jay Liebowitz is the Distinguished Chair of Applied Business and Finance at
Harrisburg University of Science and Technology. He previously was the Orkand
Endowed Chair of Management and Technology in the Graduate School at the
University of Maryland University College (UMUC). He served as a Professor in
the Carey Business School at Johns Hopkins University. He was ranked one of the
top 10 knowledge management researchers/practitioners out of 11,000 worldwide,
and was ranked second in KM Strategy worldwide according to the January 2010
Journal of Knowledge Management. At Johns Hopkins University, he was the found-
ing Program Director for the Graduate Certificate in Competitive Intelligence and
the Capstone Director of the MS-Information and Telecommunications Systems
for Business Program, where he engaged over 30 organizations in industry, govern-
ment, and not-for-profits in capstone projects.
xxiii
Chapter 1
Contents
1.1 The Beginnings of Data Science....................................................................2
1.2 The Beginnings of Artificial Intelligence.......................................................3
1.3 The Beginnings of Autonomy........................................................................4
1.4 The Convergence of Data Availability and Computing.................................6
1.5 Machine Learning the Common Bond.........................................................6
1.5.1 Supervised Learning..........................................................................7
1.5.2 Unsupervised Learning....................................................................10
1.5.3 Reinforcement Learning..................................................................10
1.6 Data Science Today.....................................................................................11
1.7 Artificial Intelligence Today........................................................................13
1.8 Autonomy Today.........................................................................................16
1.9 Summary....................................................................................................17
References............................................................................................................18
Often in discussions on data science, artificial intelligence (AI), and autonomy, the
terms become conflated. Recently, we have experienced hype cycles in data science,
artificial intelligence, and autonomy. Although these fields share common tech-
nologies and algorithms, their history has evolved independently, and they employ
different frameworks and address different real-world applications. This chapter
1
2 ◾ Data Analytics and AI
explains the commonalities and differences in the fields of data science, artificial
intelligence, and autonomy. First, we will provide a historical perspective for each
of these fields, followed by an exploration of common technologies and current
trends in each field.
Figure 1.1 “Trade wind chart of the Atlantic Ocean,” by Matthew Fontaine
Maury, 1851. Geography and Map Division, Library of Congress.
* https://blogs.loc.gov/maps/2018/07/scientist-of-the-seas-the-legacy-of-matthew-fontaine-
maury/
Unraveling Data Science, Artificial Intelligence, and Autonomy ◾ 3
◾◾ Multidisciplinary investigations
◾◾ Models and methods for data
◾◾ Computational systems
◾◾ Pedagogy for education
◾◾ Evaluation of tools
◾◾ Theoretical foundations
Cleveland’s paper is cited as the seminal data paper; however, the field did not gain
popularity until the explosion of internet connectivity, the low cost of data storage,
and the “Big Data” era. The Big Data term refers to large and complex data that
cannot be addressed with traditional relational database tool sets.*
* https://en.m.wikibooks.org/wiki/Data_Science:_An_Introduction/A_History_of_Data_
Science
4 ◾ Data Analytics and AI
* https://towardsdatascience.com/history-of-the-first-ai-winter-6f8c2186f80b
† https://towardsdatascience.com/history-of-the-second-ai-winter-406f18789d45
Unraveling Data Science, Artificial Intelligence, and Autonomy ◾ 5
The term “automatons” was used for centuries. It was not until 1920 that the
term “robot” appeared in the title of a play, Rossum’s Universal Robots, written by
Karel Čapek, a Czech writer.
In the 1960s, industrial robots emerged with the development of the Unimate
robotic arm for factory automation. One of the early robots that involved computa-
tion was Shakey, which was developed in the late 1960s at the Artificial Research
Institute at the Stanford Research Center. Shakey combined research in robotics
with computer vision and natural language processing.
The current field of robotics and autonomy is enabled by sensors, actuators, and
AI.* In 2009, the United States published the first national roadmap for robotics
(Computing Community Consortium, 2009). The roadmap described broad criti-
cal capabilities to include robust 3D perception, planning and navigation, human-
like dexterous manipulation, intuitive human-robot interaction, and safe robot
behavior. For robotic systems to perform autonomously, the capabilities of percep-
tion and planning and navigation are critical. These are areas where AI has become
an enabler. An update to the 2009 roadmap was published in 2016, and deep learn-
ing and reinforcement learning were highly emphasized (Computing Community
Consortium, 2016).
* https://www.wired.com/story/wired-guide-to-robots/
6 ◾ Data Analytics and AI
◾◾ Supervised learning
◾◾ Unsupervised learning
◾◾ Reinforcement learning
patterns in data without labeled data. Finally, reinforcement learning employs risk
versus reward models to predict actions.
1.5.1 Supervised Learning
The process of supervised learning consists of two phases (Figure 1.2). The first is
the training phase for the machine learning model. Training data must be curated
and engineered, and this is generally the most resource-intensive part of creating the
machine learning classifier. Once the training data are created, a supervised machine
learning algorithm is applied to the data to create a machine learning model. The
second phase is the classification phase, where new data are presented to the trained
machine learning model and the new data are classified as one of the trained categories.
Popular supervised machine learning algorithms are Naive Bayes, support vec-
tor machines (SVMs), random forest, and deep neural networks. In addition, there
are ensemble methods in supervised machine learning where different algorithms
are combined in various ways to improve the classification results. The ensemble
classification is illustrated in Figure 1.3. Combining the outputs of the individual
classifiers can be achieved using different combination rules. Examples may include
majority vote or probability-based solutions.
As previously mentioned, a boom in machine learning occurred with the results of
the 2012 ImageNet competition (Krizhevsky et al., 2012). This work transformed the
area of supervised machine learning by the application of deep neural networks (aka deep
learning). Deep learning combines feature extraction with classification. In machine
learning, a feature is measurable or observable data that can be used for analysis.
Figure 1.4 uses a simple example of feature engineering for classifying bicycles
with machine learning.
Traditional supervised learning would require human feature engineering to
create the machine learning classifier. In the case of designing a bicycle classifier,
the designer would need to select features such as circles and triangles. For deep
learning, the computer algorithm would discover the features by providing the
algorithm with examples of pictures with bicycles and no bicycles. This comparison
to the traditional approach is described in Figure 1.5.
Deep learning algorithms are based on artificial neural networks, which is com-
putation modeled on brain functions (e.g., neurons). A neural network is built upon
a set of neurons or units that take multiple inputs and provide a single output
(Figure 1.6). The layered units are combined into a neural network with several
layers and neurons per layer. When the neural network contains more than two
hidden layers, it is considered a deep neural network (Figure 1.7). Hidden layers are
isolated from the input and output layers of the neural network. When designing
Unraveling Data Science, Artificial Intelligence, and Autonomy ◾ 9
deep neural networks, the designer must decide on the number of layers, the num-
ber of neurons per layer, and the activation functions for the neurons. Common
activation functions include the following:
◾◾ Step
◾◾ Linear
◾◾ Sigmoid (logistic)
◾◾ Tangent hyperbolic (TanH)
◾◾ Rectified linear unit (ReLU)
◾◾ Softsign
10 ◾ Data Analytics and AI
D ( X ,Y ) = å(x - y )
2
Euclidean distance i i
i =1
å ( x - X ) (y - Y )
n
i i
Pearson linear correlation C ( X ,Y ) = i =1
å (x - X ) å (y - Y )
n n 2
2
i i
i =1 i =1
X Y
Cosine similarity C ( X ,Y ) =
X Y
1.5.2 Unsupervised Learning
Unsupervised learning does not rely on labeled data for training. Instead, data are
processed by the unsupervised algorithm, which seeks to cluster like data based on
various distance measures. “Unsupervised” refers to a learning algorithm that does
not assume any prior knowledge of data labels. The key for unsupervised learning
is to represent the data into a structure that allows for the application of distance
measures. For example, the data will be structured into feature vectors. A feature
vector is a vector of numerical features that represent some object. The common dis-
tance measures are summarized in Table 1.1. Common distance functions include
Euclidean, Pearson linear correlation, and cosine. In each of these equations, X and
Y represent the feature vectors.
1.5.3 Reinforcement Learning
Supervised and unsupervised learning both rely on data sets to learn patterns.
Reinforcement learning uses a framework to explore alternatives to achieve a
reward signal, which is the goal of the reinforcement learning problem. Another
way to think about reinforcement learning is in situations when you can measure
the goodness of a solution but do not have prior knowledge of the solution.
The architecture for reinforcement learning is illustrated in Figure 1.8.
The components of reinforcement learning include the following:
The name reinforcement learning comes from the idea of reinforcing desired
behavior. If the AI does something desirable, that behavior can be reinforced via a
“reward signal.” This reward is what the AI is trying to optimize. Specifically, it is
trying to learn how to behave such that it gets as much reward as possible over time.
One of the limitations of reinforcement learning is when environment states
become too large and computation becomes significant. Recently combining deep
learning with reinforcement learning has advanced the field. By using a deep neural
network to represent the state and observation space the learning performance is
increased. The field of deep reinforcement learning (DRL) is rising in popularity
and has been applied to robotics, natural language processing, and financial man-
agement (Yang et al., 2018).
* https://builtin.com/data-science/data-science-applications-examples
12 ◾ Data Analytics and AI
Neil Lawrence offers a framework to assess data readiness for analytics (Lawrence,
2017). He uses three levels of data readiness (Figure 1.10). The lowest level
(C-Level) describes the challenges with data engineering and wrangling. As
Lawrence explains, many organizations claim they have data, but the data have
not been made available for analytic use. He refers to this type of data as “hearsay”
data. B-Level data require an understanding of the faithfulness and representation
of the data. Finally, A-Level data are about data in context. With A-Level data, it
is understood whether the data can answer organizational questions. Once data
are made available in a data warehouse or data lake, reporting can be performed.
Many organizations create reports using spreadsheets or text documents. This
approach looks backward, reflecting what has happened in the past. The promise
of data science is to move beyond backward-looking reporting to become forward
looking. In the field of data science, analytics are typically described as descriptive,
predictive, and prescriptive, although some include diagnostic after descriptive
analytics. Descriptive analytics involves understanding the characteristics of the
data. For numerical data, descriptive analytics includes statistical measures such
as means, standard deviations, modes, and medians. Other analytics may include
histograms. Descriptive analytics helps to discover anomalous and missing data
examples. Descriptive analytics are backward looking too.
Moving from backward looking to forward looking can be achieved with pre-
dictive analytics. Predictive analytics uses data about the past to make predictions
about the future. Supervised machine learning provides an analytical tool for pre-
dictive analytics. As previously described, supervised machine learning uses train-
ing data to create a machine learning model. The machine learning model can then
be used to make predictions about new data sets.
Unraveling Data Science, Artificial Intelligence, and Autonomy ◾ 13
The third type of data science analytic is prescriptive analytics, which addresses
the human intervention because it provides for decision options. Moving beyond
predictive analytics, which describe a future state, prescriptive analytics offers
courses of action to bring value to an organizational objective. Reinforcement
learning is a machine learning approach that provides a foundation for prescriptive
analytics.
* https://medium.com/politics-ai/an-overview-of-national-ai-strategies-2a70ec6edfd
14 ◾ Data Analytics and AI
active research and will be critical for national security as the real world
is constantly changing, including the tactics of adversarial agents in the
operating environment.
Creativity and Innovation. This layer is the highest aspiration in AI—
emergent machine intelligence that autonomously realizes capabilities well
beyond initial programming. In the lower layers, the AI algorithms are
largely shaped by preprogrammed models and allowed to make adjustments
from these models. Unlike machines, humans possess creativity and imagi-
nation and can think outside the constraints of models. For this layer, the
AI algorithms are demonstrating some degree of creativity. An interesting
tangent here is the recent work of Pearl and Mackenzie (2018) on causality.
He asserts causal discovery and the ability to reason about how “counterfac-
tuals” represent key aspects of imagination in an intelligent agent.
Autonomous Perception
Systems that reason about their environment, focus on the critical aspects of the
scene, understand the intent of humans and other machines, and learn through
exploration.
Superhuman Decision-Making and Autonomous Action
Systems that identify, evaluate, select, and execute effective courses of action with
superhuman speed and accuracy for real-world challenges.
Human-Machine Teaming at the Speed of Thought
Systems that understand human intent and work in collaboration with humans to
perform tasks that are difficult or dangerous for humans to carry out with speed
and accuracy.
Safe and Assured Operation
Systems that are robust to real-world perturbation and resilient to adversarial
attacks with ethical reasoning and goals that are guaranteed to remain aligned
with human intent.
It should be noted that the technical areas will require a combination of data-
driven and symbolic AI techniques. Reasoning becomes more prevalent as machine
abilities advance to perform pattern recognition. DARPA describes the next wave
of AI as systems constructing contextual explanatory models* which will require
* https://www.darpa.mil/attachments/AIFull.pdf
16 ◾ Data Analytics and AI
1.8 Autonomy Today
The field of autonomy today is dominated by modern robotics, driverless cars, and
drones. Three main components of an autonomous system include sensors, reason-
ing, and actuators. Modern autonomous systems are enabled by AI. Figure 1.12
maps AI functions (i.e., perceive, decide, and act) to the components of an autono-
mous system.
In the field of robotics, there are numerous frameworks for developing robotic
systems. Tsardoulias and Mitkas (2017) provide a comprehensive survey of robotic
frameworks. Generally, these frameworks address the following:
1.9 Summary
This chapter describes the fields of data science, AI, and autonomy. Each of these
fields is experiencing significant growth and interest as practical applications
emerge. Machine learning is a critical enabler and provides tremendous capabili-
ties with the availability of data and computing power. All three fields leverage
the advances in machine learning with an emphasis on deep neural networks and
reinforcement learning. Finally, there is a large amount of overlap in the fields of
data science, AI, and autonomy. However, by considering the history of these fields
and thematically what they seek to achieve, differences can be established, as sum-
marized in Figure 1.13.
References
Anderson, C. (2015). Creating a Data-Driven Organization (1st ed.). Sebastopol, CA:
O’Reilly Media, Inc.
Cleveland, W. (2001). Data science: An action plan for expanding the technical areas of the
field of statistics. International Statistical Review/Revue Internationale De Statistique,
69(1), 21–26.
Computing Community Consortium (2016, October 31). A Roadmap for US Robotics—
From Internet to Robotics (2016 ed.). Washington, DC: Computing Research
Association.
Computing Community Consortium (2009, May 21). A Roadmap for US Robotics—From
Internet to Robotics. Washington, DC: Computing Research Association.
Computing Community Consortium and Association for the Advancement of Artificial
Intelligence (AAAI) (2019, August 19). A 20-Year Community Roadmap for Artificial
Intelligence Research in the US.
Deng, J., Dong, W., Socher, R., Li, L., Li, K., & Fei-Fei, L. (2009). ImageNet: A large-scale
hierarchical image database. 2009 IEEE Conference on Computer Vision and Pattern
Recognition, Miami, FL, 2009, pp. 248–255.
Garcez, A. S., Gori, M., Lamb, L. C., Serafini, L., Spranger, M., & Tran, S. N. (2019).
Neural-symbolic computing: An effective methodology for principled integration
of machine learning and reasoning. IfColog Journal of Logics and their Applications
(FLAP), 6, 611–632.
Garnelo, M., & Shanahan, M. (2019). Reconciling deep learning with symbolic artifi-
cial intelligence: Representing objects and relations. Current Opinion in Behavioral
Sciences, 29, 17–23.
Graham, B., & Dodd, D. (1934). Security Analysis. New York: McGraw-Hill.
Gunning, D. (2017, October 10). The promise of artificial intelligence—Again. Presented
at the MIT Washington Seminar Series: Artificial Intelligence and Machine Learning.
Hopfield, J. J. (1982, April). Neural networks and physical systems with emergent collective
computational abilities. Proceedings of the National Academy of Sciences of the United
States of America, 79(8), 2554–2558.
Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). ImageNet classification with deep
convolutional neural networks. Proceedings of the 25th International Conference on
Neural Information Processing Systems (NIPS 2012).
Lawrence, N. D. (2017). Data Readiness Levels., arXiv.
Mallows, C. (2006). Tukey’s paper after 40 years. Technometrics, 48(3), 319–325.
Morris, R. J. T., & Truskowski, B. J. (2003). The evolution of storage systems. IBM Systems
Journal, 42(2), 205–217.
Pearl, J., & Mackenzie, D. (2018). The Book of Why: The New Science of Cause and Effect (1st
ed.). New York, NY: Basic Books, Inc.
Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1986, October 9). Learning represen-
tations by back-propagating errors. Nature, 323(6088), 533–536.
Silver, D., Huang, A., Maddison, C., Guez, A., Sifre, L., van den Driessche, G., Schrittwieser,
J., Antonoglou, I., Panneershelvam, V., Lanctot, M., Dieleman, S., Grewe, D.,
Nham, J., Kalchbrenner, N., Sutskever, I., Lillicrap, T., Leach, M., Kavukcuoglu,
K., Graepel, T., & Hassabis, D. (2016). Mastering the game of Go with deep neural
networks and tree search. Nature, 529(7587), 484–489.
Unraveling Data Science, Artificial Intelligence, and Autonomy ◾ 19
Contents
2.1 Introduction................................................................................................21
2.2 Situation Overview.....................................................................................22
2.2.1 Data Age..........................................................................................23
2.2.2 Data Analytics.................................................................................23
2.2.3 Marriage of Artificial Intelligence and Analytics.............................24
2.2.4 AI-Powered Analytics Examples......................................................27
2.3 The Way Forward........................................................................................28
2.4 Conclusion..................................................................................................29
References............................................................................................................30
2.1 Introduction
Artificial intelligence (AI) is the study and research of software and hardware
that attempts to emulate a human being. AI is poised to be more prolific and
span every aspect of our daily lives. The International Data Corporation (IDC)
forecasts that worldwide spending on AI will grow at a compound annual growth
rate (CAGR) of 28.4% for 2018–2023 will reach over $97.9 billion. AI disrupters
will drive better customer engagements and have accelerated rates of innovation,
21
22 ◾ Data Analytics and AI
2.2 Situation Overview
AI-powered analytics is a relatively young, though already proven, technological
advancement. Many companies have invested heavily in deep learning, a subset of
machine learning (ML), which is itself a subset of AI. Deep learning and artificial
neural networks enable image recognition, voice recognition, natural language
processing (NLP), and other recent advancements. We have already come to take
these for granted in our personal lives in the age of the internet and big data, but
such features are hardly commonplace in analytics software. Modern analytics
solutions have just started supporting NLP that enable end-users to use everyday
language to ask questions of the data. But in just a few years, AI-powered analyt-
ics will help organizations create and deliver on all their most important strate-
gies. With AI-powered analytics, businesses can concentrate on the initiatives that
really matter instead of burying themselves in reports. Basic analytics packages
report what has happened and helps to investigate why. But to stay relevant today,
businesses must do more. They need to get on top of the mountains of data their
customers generate as they engage through an ever-growing array of channels and
devices. And then they must turn around and use that data to create compelling
experiences on every channel that delight their customers and exceed even their
highest expectations. Action is the goal. Analysis and insights are means to an end,
not an end in themselves. AI and machine learning can identify the best action,
and businesses can decide on situations where they want to automate business
processes. AI and machine learning assist, augment, and amplify work. And when
done right, these insights can empower everyone in the company, not just a few
specialists.
Unlock the True Power of Data Analytics with Artificial Intelligence ◾ 23
2.2.1 Data Age
Successful DX relies on converting data into actionable insights, and this reliance
on data is contributing to a new era of the data age. As per the IDC’s Data Age
2025 study (Doc #US44413318) sponsored by Seagate, the IDC predicts that the
global “datasphere” will grow from 33 zettabytes (that’s one trillion gigabytes) in
2018 to 175 zettabytes in 2025. All this data will unlock unique user experiences
and a new world of business opportunities. While we see an explosion in data cre-
ation, the sad reality is that a very small percentage of global datasphere is fed into
AI, as they are not useful unless they are tagged, labeled, or annotated. We have an
acute shortage of training data sets. In fact, AI is being used to create synthetic data
to feed into AI training.
2.2.2 Data Analytics
As we know, data analytics is a broad field. There are four primary types of data
analytics: descriptive, diagnostic, predictive, and prescriptive analytics. Each type
of data analytics has a different goal and a different place in the data analysis pro-
cess (see Figure 2.1).
◾◾ Descriptive analytics analyzes raw data from multiple data sources to give
valuable insights into the past—what’s happening or what happened in my
business. However, these findings simply signal that something is wrong or
right, without explaining why.
◾◾ Diagnostic analytics gives in-depth insights into a problem—why is it
happening.
◾◾ Predictive analytics tells us what is likely to happen. It uses the findings of
descriptive and diagnostic analytics to detect clusters and exceptions, and to
predict future trends.
◾◾ Prescriptive analytics tells us what we need to do. It is to literally prescribe
what action to take to eliminate a future problem or take full advantage of
a promising trend. The prescriptive model utilizes an understanding of what
has happened, why it has happened, and a variety of “what-might-happen”
analysis to help the user determine the best course of action to take. A pre-
scriptive analysis is typically not just with one individual response but is, in
fact, a host of other actions.
The applications of data analytics are seemingly endless. One of the earliest
adopters is the financial sector. Data analytics has a significant role in the bank-
ing and finance industries, used to predict market trends and assess risk. Credit
scores are an example of data analytics that affects everyone. These scores use
many data points to determine lending risk. Data analytics is also used to detect
24 ◾ Data Analytics and AI
and prevent fraud to improve efficiency and reduce risk for financial institutions.
The use of data analytics goes beyond maximizing profits and return on invest-
ment (ROI), however. Predicting patient outcomes, efficiently allocating fund-
ing, and improving diagnostic techniques are just a few examples of how data
analytics is revolutionizing healthcare. The Internet of Things (IoT) is a field that
is exploding. These devices provide a great opportunity for data analytics. IoT
devices often contain many sensors that collect meaningful data points for their
operation. Devices like the Nest thermostat track movement and temperature to
regulate heating and cooling. Smart devices like this can use data to learn from
and predict your behavior. This will provide advance home automation that can
adapt to the way we live.
otherwise-interested parties. Now in the age of big data and nanosecond process-
ing, machines can rapidly mimic aspects of human reasoning and decision-making
across massive volumes of data. Through neural networks and deep learning, com-
puters can even recognize speech and images.
When first introduced to AI-powered analytics, most people tend to jump right
to the potential for predictive analytics. However, there are far more potential use
cases for front-line knowledge workers surrounding the diagnostic capabilities of
AI-powered analytics—“why did something happen?” Because data is constantly
moving and growing, augmented analytics is crucial to more quickly diagnose root
causes and understand trends. An analyst may need hours, days, or even weeks or
months to develop and evaluate hypotheses and separate correlations from causal-
ity to explain sudden anomalies or nascent trends. With AI-powered analytics and
machine learning techniques, they can sort through billions of rows in seconds to
diagnose the indicators and root causes in their data, guiding and augmenting their
work to deliver consistent, accurate, trustworthy results. By leveraging AI-powered
diagnostic analytics, you get valuable insights on the current state of the world
faster. This yields a competitive advantage to businesses trying to stay ahead in
dynamic marketplaces.
Beyond opening new opportunities outside of descriptive and diagnostic ana-
lytics, AI and machine learning bring other significant benefits to an analytics prac-
tice. For example, they can take over tedious tasks that deflect their attention from
strategy. Many of these tasks involve building and maintaining rules that guide
the analysis of data. These tasks are critical, but by automating them businesses
can focus on the message, the creative, and the content, as well as responding to
what is happening. Businesses can move from rules-based to AI-powered analytics
and reap significant benefits. With rules-based analytics, the investigation of why
an event happened and possible actions are manual. With AI-powered analytics,
evaluation of what factors contributed to the event and suggestion of a cause and
action is automatic.
AI and machine learning analyze data, makes assumptions, and learns and pro-
vides predictions at a scale and depth of detail impossible for individual human
analysts. AI and machine learning systems work with fuzziness. They predict. They
will consider a path but can abandon it if new data negates a line of reasoning—
then begin looking at a new direction. Machine learning systems get smarter as
more data is given to them and are well-suited to identifying anomalies over time.
For example, digital businesses with high-risk exposure are adopting AI-based
online fraud analysis and detection platforms. First, supervised machine learning
algorithms are used to finetune decades worth of transaction data to minimize false
positives and provide extremely fast responses to inquiries. Second, unsupervised
machine learning is used to find emerging anomalies that may signal entirely new,
more sophisticated forms of online fraud.
Advanced analytics provides companies new means and abilities to better
understand the past and to predict the future. Innovations in machine learning,
26 ◾ Data Analytics and AI
advanced learning, and NLP make it even easier for business users to understand
quicker than before about impacts to the future. AI can enable you to sort through
mountains of data, even uncovering insights to questions that you didn’t know to
ask—revealing the proverbial needle in the haystack. It can increase data literacy,
provide timely insights more quickly, and make analytics tools more user-friendly.
These capabilities can help organizations grow revenue, improve customer service
and loyalty, drive efficiencies, increase compliance, and reduce risk—all require-
ments for competing in the digital world. AI-powered analytics can help automate
tasks in the analytics lifecycle, augment human intelligence, or support human-
machine collaboration (see Figure 2.2).
Here are examples of how AI-powered analytics can transform businesses:
any attribute, generating facts about their data, including the drivers of the
results, key segments that influence behavior, and anomalies where the data
is not aligned with expected patterns. Smart data discovery expands the
explaining capability of AI-powered analytics, beyond analyzing an attri-
bute or measure, examples include explaining an entire data set, explaining
specific data events, or automatically understanding the links to related
data.
◾◾ Auto Suggest. Business analysts and consumers often don’t know where
to start their insights initiatives. Auto Suggest embeds suggestions intel-
ligently throughout the analytics life cycle, using ML behind the scenes,
personalized to user needs, and fast-tracks the cycle to quicker actions. It’s
a valuable time-saver that helps users get to the data, visualizations, or con-
tent they are looking for. Auto-suggestions are just that—suggestions, not
mandates, and users are still able to override the suggestions with the click
of a mouse.
◾◾ Automated insight-generation. Insight-generation is a key application of AI in
analytics. Rather than relying on a user to ask the right questions, AI-powered
analytics solutions can traverse massive data sets at the click of a button—or
via a scheduler—and find interesting insights on their own. With user feed-
back, machine learning can help determine which insights are interesting and
which are just noise to individual users and groups.
industries. Businesses can see between a 23% and 50% sizeable increase over
their current conversion rates, which translates into millions of dollars in rev-
enue growth.
We are at a market disruption point. With modern analytics platforms, insights
and actions are moving to real time. AI-powered analytics will provide organiza-
tions with knowledge on how to develop, invest, monetize, and release products
more than previously done before. Enterprise economies and the nature of compe-
tition have changed. While still important, economy of scale has been augmented
with economies of scope and economies of learning. Now, leading companies are
pursuing “economies of intelligence,” the continual improvement, innovation,
and variation based on leveraging data and AI technologies to identify and fulfill
changing needs to enhance scale, scope, and customer engagement. This is chang-
ing the nature of intellectual property, whose value has shifted to where it’s created
rather than where it’s realized and contributing to an asymmetrical accumulation
of capital and innovation where an organization’s capacity to learn has a distinct
competitive advantage.
2.4 Conclusion
AI is not new. In fact, it is over 60 years old, since John McCarthy together with
Alan Turing, Marvin Minsky, Allen Newell, and Herbert A. Simon coined the term
“artificial intelligence” in 1955 and organized the famous Dartmouth conference in
the summer of 1956. This conference started AI as a field. After decades of experi-
encing a slow burn, artificial intelligence innovation has caught fire to become the
hottest item on the agendas of the world’s top technology firms. Today’s flurry of
AI advances wouldn’t have been possible without the confluence of three factors
that combined to create the right equation for AI growth: the pervasiveness of data
(from IoT, social media, and mobile devices), seemingly infinite scalability of cloud-
based compute power combined with the emergence of powerful graphics process-
ing units (GPUs) for complex computations, and the re-emergence of a decades-old
AI computation model—deep learning. Understanding and sophistication of the
algorithms have expanded as well. While many ML algorithms have been around
for years, the ability to apply complex mathematical calculations to data, and pro-
cess them quicker than ever before, is a recent development.
DX is reaching macroeconomic scale and spanning all aspects of business.
Artificial intelligence is at the heart of digital disruption and business strategies
for companies of all sizes. Today, AI is embraced by businesses in silos. By 2024,
innovations in the digital economy and newer AI-powered enterprise architectures
and solutions will lead to the employment of AI across the business holistically
and much beyond basic ML models. AI will transform companies to a business
outcomes approach. Increasing productivity and capacity, improving efficiency,
and reducing the cost of operations are some of the first business outcomes to be
30 ◾ Data Analytics and AI
expected from AI. Ongoing sophistication of AI technologies will drive the devel-
opment of AI solutions that can be leveraged to increase the accuracy and effective-
ness of business processes.
It’s time for AI-powered analytics. To be successful, businesses must ensure
that their analytics tool includes AI capabilities. AI-powered capabilities should be
usable across the organization, not just by specialists. With AI-powered analytics,
businesses can pull ahead of competition and win the hearts and minds of every
customer.
The benefits of AI-powered analytics are many. Data analysts have more time to
focus on deep data insights (that they might not have uncovered previously) rather
than data preparation and report development. Decision-makers can explore much
deeper and faster than they were ever able to with predefined dashboards. Perhaps
the most important ingredient to adopting AI as part of your analytics strategy is
trust. For AI-powered analytics to gain people’s trust, there are three key consid-
erations: accuracy, relevance, and transparency. These are paramount concerns if
business leaders are to make decisions and act based on the results of AI-powered
analyses.
Technology providers are working on innovative tools and technologies to
address each of these concerns. With these bases covered, expect to see more and
more companies adopting analytics strategies that take advantage of artificial intel-
ligence as a significant component.
References
Artificial Intelligence Global Adoption Trends and Strategies (IDC Doc #US45120919,
June 2019).
IDC FutureScape: Worldwide Artificial Intelligence 2020 Predictions (IDC Doc #US455
76319, Oct 2019).
https://www.seagate.com/files/www-content/our-story/trends/files/idc-seagate-dataage-
whitepaper.pdf (An IDC White Paper I Doc #US44413318 I November 2018).
Chapter 3
Machine Intelligence
and Managerial
Decision-Making
Lee Schlenker and Mohamed Minhaj
Contents
3.1 Managerial Decision-Making......................................................................32
3.1.1 What Is Decision-Making?..............................................................33
3.1.2 The Decision-Making Conundrum.................................................33
3.1.3 The Decision-Making Process......................................................... 34
3.1.4 Types of Decisions and Decision-Making Styles............................. 34
3.1.5 Intuition and Reasoning in Decision-Making.................................35
3.1.6 Bounded Rationality........................................................................36
3.2 Human Intelligence....................................................................................36
3.2.1 Defining What Makes Us Human..................................................36
3.2.2 The Analytical Method....................................................................38
3.2.3 “Data-Driven” Decision-Making.....................................................39
3.3 Are Machines Intelligent?............................................................................41
3.4 Artificial Intelligence...................................................................................41
3.4.1 What Is Machine Learning?........................................................... 42
3.4.2 How Do Machines Learn?.............................................................. 42
3.4.3 Weak, General, and Super AI..........................................................43
3.4.3.1 Narrow AI........................................................................ 44
3.4.3.2 General AI........................................................................ 44
3.4.3.3 Super AI........................................................................... 44
31
32 ◾ Data Analytics and AI
3.1 Managerial Decision-Making
Every decision has a consequence.
—Damon Darrell
Decision-making is an essential component of management and plays a pivotal role
in achieving the goals of any organization. Irrespective of the nature of manage-
ment and the rungs of an organization, one critical success factor in any organiza-
tion is the quality of its managerial decision-making.
Machine Intelligence and Managerial Decision-Making ◾ 33
Good decisions save time, money, and stress (“The Anatomy of a Great
Decision,” 2019). People take multiple decisions daily in response to both personal
and organizational challenges. The outcome of these decisions is either good or
bad for the decision-maker, the organization, and society. A “good” decision is
a course of action that will capture the most value or get the most of what you
are seeking given the uncertainties and complexities of the real world (Neal and
Spetzler, 2015). While every decision-maker would like to take good decisions,
is there a secret formula for that? With so many factors influencing the decision-
making process, what should be an ideal approach to make better decisions in the
limited time available? This section of the chapter endeavors to address some of
these questions.
3.1.1 What Is Decision-Making?
Decision-making involves the selection of a course of action from among two
or more possible alternatives in order to arrive at a solution for a given problem
(Trewatha and Newport, 1976). In both personal and organizational decision-
making, the objective is to choose the best available option within a given time
constraint. it is generally observed that when it comes to personal decisions, most
people go with their gut feeling, hunch, or intuition. However, when it comes to
organizational decisions, many managers are very careful and slow: they discuss,
deliberate, get into finer details, and only then arrive at a decision.
Why do people react differently when faced with organizational decisions?
Obviously, they are worried about taking a poor decision, and its repercussions on
them, the business environment, and society at large. Each manager’s approach to
decision-making is marked by their own personality: some are quick in their deci-
sions while others are risk averse and are very hesitant in taking any concrete deci-
sion. Some are astute at arriving at a logical conclusion from the available figures
and facts, while others struggle with data-based decision-making.
• Alternaves that are developed in the design phase are evaluated and
Choice one of them is chosen
Phase
Strategic
Taccal
Operaonal
3.1.6 Bounded Rationality
Humans have different personality traits and deal with situations in different ways
according to their natural instincts. Our emotions drive our behavior and can
cloud our reasoning when making decisions. Is it possible for all decision-makers in
all situations to assess all decisions rationally?
According to Herbert A. Simon’s theory of bounded rationality, the rationality
is by nature limited: managerial decision-making is always limited by the tracta-
bility of the decision problem, the cognitive limitations of the mind, and the time
available to make the decision. Decision-makers, in this view, act as satisficers,
seeking a satisfactory solution rather than an optimal one. Herbert A. Simon pro-
posed bounded rationality as an alternative basis for the mathematical modeling
of decision-making widely prescribed in political science, economics, and the hard
sciences. It complements “rationality as optimization,” which views decision-mak-
ing as a fully rational process of finding an optimal choice given the information
available (Gigerenzer and Selten, 2002). This limited rationality leads to systematic
errors in the decisions that we make. Kahneman, in his extensive work on this
domain, refers to these limitations as cognitive biases and has experimentally proved
how cognitive biases like framing or anchoring can lead to poor decision-making
(Kahneman, 2011).
3.2 Human Intelligence
Can the essence of humanity be captured in an algorithm? Before discussing how
human and machine intelligence can potentially work together, it can be useful to
ask what separates humanity from machines. If the visionaries of artificial intelli-
gence promise to use technology to augment what makes us human, the proponents
of Dataism argue that little differentiates us from the technologies we produce to
mimic our behavior (Briggs, 2016). Are human thought and feelings nothing more
than programmed biochemical responses to environmental stimuli? Let’s consider
four unique human traits before asking how machine intelligence can make us
better: the notion of human agency, the capacity for abstraction and empathy, the
multiple forms of human intelligence, and the capability to differentiate right from
wrong.
3.2.3 “Data-Driven” Decision-Making
People don’t act on data, but on their perceptions of what the data represents. The
human brain isn’t wired to store dates, formulas, or statistics but rather its syn-
apses process colors, feelings, and sensations that are tied to human experience.
When presented with new data, human beings compare what they “see” with these
records of experience in recognizing the need or not to act. Understanding how
customers, employees, and managers interpret the data they see is fundamental in
understanding how human and machine intelligence can work together to inspire
action. In managerial decision-making there is no direct link between data and
action, but rather four steps of perception, prediction, evaluation, and action wor-
thy of our attention. Using the AI Roadmap, we can map out the challenges and
opportunities of matching human and machine intelligence based on specific busi-
ness contexts (Figure 3.3).
Perception refers to the way in which data is regarded, assimilated, and inter-
preted. No two customers, employees, and managers interpret data in exactly the
same manner, for their prior experience is used in accessing, selecting, qualifying,
and organizing sensory information. Using the example of a business dashboard of
last quarter’s sales figures, the human brain uses this information to build a mental
picture or story of what is going on in the organization or in the market. Perception
is thus much more than just “information coming in,” it is an interactive process
in which prior experience is used to interpret the facts, and the facts are eventually
used to reinterpret experience.
Prediction is a second step in the decision-making process. A prediction is a
conclusion that a decision-maker makes about what they believe is going to hap-
pen. Predictions are based on the hypothesis that if we understand the roots of an
opportunity or a challenge, history most likely will repeat itself. Unfortunately,
predictions are made using imperfect information: we may not have access to all the
40 ◾ Data Analytics and AI
3.4 Artificial Intelligence
Machine learning (ML) is no longer a subject of fiction depicted in movies and sci-
ence novels. It has become a reality and has provided remarkable business applica-
tions. Besides Siri, Alexa, or Cortana becoming household names, we notice that
most industries have embarked on the road to artificial intelligence (AI). Personal
assistants used in healthcare, the recommendation engines used in retail portals,
and Internet of Things (IoT) data to detect quality issues in manufacturing are all
examples of how businesses are leveraging machine intelligence.
The notion of machine intelligence is as almost old as the computer itself. In
1950, in his seminal paper “Computing Machinery and Intelligence,” computing
pioneer Alan Turing laid out several criteria, known since as the “Turing test,”
to assess whether a machine could be qualified as intelligent (Turing, 1950). The
term Artificial Intelligence was first used by John McCarthy in 1956 to describe
the research focused on problem-solving and symbolic methods. Subsequently, the
domain of AI has grown in both breadth and depth and has contributed signifi-
cantly to the world of business. Some notable outcomes of AI research and develop-
ment include Deep Blue and Watson. Deep Blue was the first chess engine to beat
reigning world chess champion, Garry Kasparov, on May 11, 1997. Watson is a
question answering system from IBM which defeated champions of the challenging
television quiz show Jeopardy! Another notable AI system is AlphaGo, which can
play a board game considered to be more complex than chess and in fact defeated
the world’s top ranked Go player.
Many industries like healthcare, logistics, retail, and financial services are using
AI to enhance the speed, precision, and efficiency of human interactions in business
42 ◾ Data Analytics and AI
processes. The key point to observe is that all these industries who have embarked
on AI are dealing with large amounts of data which is spawned from their digital
initiatives. The bottom-line is AI has become a mechanism to create value from
data by aiding organizations in either tapping new opportunities or solving their
business problems.
On the other hand, there are some business problems which cannot be auto-
mated using traditional computer programs. Historically, algorithms were pro-
grammed to follow a fixed set of instructions in a specific context. If there is change
in the inputs or conditions, the algorithm cannot modify the instructions or rules
on its own and requires human intervention to correct the program. For example,
what if we would like to develop a system to detect fraudulent transactions on an
e-commerce portal. In a traditional approach, a rule-based system would be devel-
oped to test each transaction against all the rules defined in the algorithm. The
system would not be capable of detecting any implicit correlations nor incorporate
new rules deducted from the online transaction’s environment. The ML approach
on the other hand can detect subtle and hidden events in consumer behavior which
may be a signal of possible fraud. This is possible because the ML algorithms are
designed to process large data sets to detect hidden correlations between user
behavior and the likelihood of fraudulent actions. After identifying a transaction as
fraudulent, this is fed back to the system for factoring these features into the future
processing of fraudulent transactions. This is what is referred to as machine learn-
ing: the correlations or relationships are recorded as new data points to improve the
working of the machine/system.
Based on the different approaches used to discover the patters from data,
machine learning can be broadly classified as Supervised and Unsupervised.
Machine learning requires past figures, facts, and experiences given in the form
of data for it to predict or act on the new set of data. The idea behind supervised
learning is to make the system or machine learn a task by giving it a set of questions
along with an answer. The training process is continued until the model achieves
a desired level of accuracy on the training data. The supervised learning methods
include regression, decision tree, KNN, and so on.
In unsupervised learning, unlike supervised learning, we do not give the
answer to the questions, and the model is expected to find the inherent patterns
or answers from the given data. Clustering and association are popular examples
of unsupervised learning methods. Without any labeled prior knowledge, the
unsupervised algorithms explore and discover interesting hidden structures in
the data.
The choice between a supervised and unsupervised machine learning algorithm
typically depends on factors related to the structure and volume of data available
and the specific purpose. In most cases, data scientists use supervised learning and
unsupervised learning approaches together to address a use case.
3.4.3.1 Narrow AI
The term “Narrow Intelligence” stems from the fact that this kind of intelligent
system is explicitly created for a specific task. They are limited and are focused on
doing only one or one set of tasks well. Owing to its ability to perform specific
tasks automatically, narrow AI is meant to stimulate, rather than replace human
intelligence. AI-assisted chess engines, smart digital assistants, recommendation
engines, and spam filtering systems are some examples of narrow AI. In a variety
of markets, businesses can employ narrow AI today to automate and integrate busi-
ness processes more efficiently than ever.
3.4.3.2 General AI
Artificial General Intelligence (AGI) is broad and adaptable and is meant to repli-
cate human-level intelligence. So, it can understand, think, and act the same way
a human might in any given situation. This includes tasks such as vision and lan-
guage processing, cognitive tasks such as processing, contextual understanding,
and a more generalized approach to thinking as a whole.
With extensive efforts from the global research community, AGI has reached
to a relatively higher degree of maturity. But unlike narrow AI, there are few major
functional use cases today. The key constraint for implementing AGI is Transfer
Learning, a term used to denote applying knowledge learned in one domain
(attending a business school) to another (managing a business unit). While Transfer
Learning is naturally imbued in humans, such attributes in a machine have yet to
see the light of day.
3.4.3.3 Super AI
Super Artificial Intelligence is a term used to denote an AI that exceeds human intel-
ligence and capabilities. Nick Bostrom defines Super AI as any intellect that greatly
exceeds the cognitive performance of humans in virtually all domains of interest
(2014). Super AI is more conjecture today than of practice, though the adversaries
of AI often refer to a form of super AI that will one day enslave humanity.
3.4.4 The Limitations of AI
Today, organizations are investing heavily in AI in the hope of automating and
optimizing a long list of time-consuming and costly manual tasks. Does this suc-
cess in chess, smart cars, and drones mean that AI can replicate human intelli-
gence? AI’s superior and ever-improving capabilities in different fields has evoked a
hope from the global tech community as well as the general public. Specifically, the
use cases like AI detecting cancers better than human doctors can lead us to believe
that AI can massively benefit humanity by raising our standard of living.
Machine Intelligence and Managerial Decision-Making ◾ 45
For the organizations that are procuring and implementing AI, in addition to
limitations like availability of data, cost, and lengthy implementation cycles, sev-
eral ethical issues have arisen. As we will see in the next section, privacy, security,
responsibility, and dealing with the unintended consequences of machine intel-
ligence are all concerns worthy of a manager’s attention.
3.5.1 Human Singularity
We have argued that one of the characteristics that separates man from machine
is the notion of human agency, the very human capability of thinking and activ-
ity independently. Although this affirmation has been debated over the years, our
ability to act of our own accord is fundamental to our economic, social, and politi-
cal systems. Nonetheless, because such decision-making is both energy and time
consuming, we have naturally developed mental shortcuts, in the form of rules of
thumb, mental models, and heuristics to save both time and effort. Machine intel-
ligence is little more than an extension of this practice, a translation of our mental
models into algorithms that can be processes with little human intervention.
Slowly but surely, these algorithms are becoming increasingly pervasive in our
daily lives. Google‘s PageRank algorithm conditions how we search and rank infor-
mation on the internet, our GPS guides our sense of orientation and direction, and
Facebook and LinkedIn provide structure in how we see our social and professional
relationships. In an ever-increasing frequency, algorithms are used to evaluate our
aptitude for employment, for credit, and medical care. Will this push to support
human decision-making lead us to one day believe that algorithms will one day
rule the world? Dan Brown, the author of the novel Origin investigates this trend in
exploring the duality of Where we come from? / Where we are going (Brown, 2018)?
His protagonist, Edmond Kirsch, preaches the end of human singularity, where we
will be replaced by a different form of intelligence based on some blend of human
and artificial intelligence.
How plausible is this vision? As we have seen previously, human decision-
making relies on two distinct forms of cognition: intuition and bounded rational
46 ◾ Data Analytics and AI
3.5.2 Implicit Bias
As the recent example of Microsoft’s chatbot Tay demonstrates, one of the major
concerns of artificial intelligence is the institutionalization of implicit bias that can
seriously comprise the purported advantages of machine intelligence (Vincent,
2016). When we evoke implicit bias, we are referring to unconscionable attitudes or
stereotypes that influence our understanding, actions, and decisions in both repeti-
tive and exceptional events situations. Although this bias can be potentially either
favorable or unfavorable, these implicit evaluations or beliefs can severely limit our
ability to be objective. In reference to machine intelligence, there are four potential
sources of implicit bias: in the data, in the algorithms, in our own logic, and in our
definitions of ethical behavior. Let’s explore each in turn.
We have argued that machine learning involves specifying a data set that is used
to train (test) the algorithm that will subsequently be used in real-world conditions.
Although as a rule the more data you collect to train the model, the better the
results, several sources of implicit bias can significantly compromise the pertinence
of your work. Sample bias occurs when the data used to train the model does not
accurately represent the problem space in which the model will operate. Prejudicial
bias occurs when training data content is influenced by stereotypes or prejudice
coming from the population. Finally, measurement bias can result from faulty mea-
surement of the outcomes. The algorithms at the heart of machine learning will
never be better than the data used to test them.
A second source of implicit bias is found in the way in which algorithms are
constructed to explain or to predict real-world phenomena. For the data scientist,
bias, along with variance, describe an algorithm property that influences prediction
performance. Since bias and variance are interdependent, data scientists typically
seek a balance between the two. Models with high variance tend to flex to fit the
training data very well but may not generalize well to data outside the training data
set. Finding the appropriate balance between these two properties for a given model
in each environment is a critical data science skill set.
A third source of bias can be found in the means in which human beings induce
and deduce logical arguments. A cognitive bias is a systematic pattern of devia-
tion from the norm or rationality in judgment. Some cognitive biases are presum-
ably adaptive. Cognitive biases may influence perceptions of similarity, judgment,
Machine Intelligence and Managerial Decision-Making ◾ 47
memory, and action. Although cognitive biases may lead to more effective actions in
each context, they often lead to misleading representations of personal and business
challenges. Furthermore, allowing cognitive biases enable faster decisions which
can be desirable when timeliness is more valuable than accuracy, as illustrated in
heuristics. Other cognitive biases are a “by-product” of human processing limita-
tions, resulting from a lack of appropriate mental mechanisms (bounded rational-
ity) or simply from a limited capacity for information processing. Gigerenzer argues
that human rationality isn’t riddled with irrational cognitive biases, but rather that
human intelligence is an adaptive tool that defies the rules of formal logic or math-
ematics (Gigerenzer, 1991).
A final source of bias comes from our views of ethics. Ethics can be under-
stood as a set of moral principles that guide our individual behavior. Ethics reflect
our perceptions of what is right and wrong, our beliefs about justice, as well as
our convections of what is acceptable in terms of human behavior. According
to the theory of cultural relativism, there is no singular truth on which to base
ethical behavior, our interpretations of the limits of what is acceptable are condi-
tions of existing social norms, cultural practices, and religious influences. Ethical
considerations are often in conflict with our perceptions of self-interest: and as a
result, people often cross the line without being aware that they are doing any-
thing wrong (Bazerman, 2011).
3.5.3 Managerial Responsibility
A manager’s job is to take decisions. On one hand, machine intelligence should
greatly assist management in providing timelier, more pertinent, and more insight-
ful data to understand the past, analyze the present, and predict future performance.
The introduction of machine intelligence in core business processes challenges the
traditional definitions of managerial command and control. On the other, manag-
ers at all levels will have to adapt to the world of smart machines. What exactly
are these challenges and how can management adapt to the needs of human and
machine intelligence?
Machine intelligence will soon be able to execute many administrative tasks
automatically, more quickly, more efficiently, and at lower cost than their human
counterparts. It is estimated that managers at all levels of an organization spend
more than half of their time working on administrative coordination and con-
trol processes (Kolbjørnsrud, 2016). According to the consulting group Accenture,
middle management will need to focus on “judgment-oriented” skills beyond
the scope of AI to enhance their value to the organization: creative thinking and
experimentation, data analysis and interpretation, brand management, and strategy
development (Accenture, 2019).
Managers needs to have some familiarization of data analytics to ensure they
are asking the right questions to address the potential ethical issues when apply-
ing machine intelligence to business. The potential dangers of machine learning
48 ◾ Data Analytics and AI
3.5.4 Semantic Drift
What does truth mean when business is reduced to “facts” and figures? Charles
Filmore suggested that the primary function of language is communication, we use
words, idioms, and expressions to convey meaning in exchanging ideas, products,
and services (Clark, 1996). Shared meaning in business communication depends
upon the context in which we work: descriptions, opinions, and ideas are woven
together using technology (paper, text, video, etc.) to form the setting or back-
ground for an event, statement, or idea. Semantic drift refers to the evolution of
language over time, often to the point in which the current meaning of a concept
is radically different from its original usage. How can we communicate with our
colleagues and customers if they there is little agreement on the very meaning of
the words we use?
Consider the notion of privacy in its relationship to personal identity. Privacy
has been defined as an individual’s “right to be free from unwarranted intrusion
and to keep certain matters from public view” (Law 2015). In this light, the con-
cept of privacy “is an indispensable feature of a democracy … (that) protects our
subjectivity from the pervasive efforts of commercial and government actors to
Machine Intelligence and Managerial Decision-Making ◾ 49
3.6 Conclusion
Can machine intelligence improve managerial decision-making? We began our
analysis based on the assumption that human decision-making can be analyzed,
modeled, and improved. We posited two potential use scenarios for machine
50 ◾ Data Analytics and AI
References
Accenture. The Promise of Artificial Intelligence. accessed 12/12/2019 at https://www.accen-
ture.com/_acnmedia/PDF-32/AI_in_Management_Report.pdf.
Bandura, A., (2001). Social Cognitive Theory: An Agentic Perspective. Annual Review of
Psychology, 52(1):1–26.
Bazerman, Max H., (2011). Blind Spots: Why We Fail to Do What’s Right and What to Do
about It, Princeton University Press.
Briggs, W.M., (2016). The New Religion of Dataism [blog]. accessed 15/10/2019 at https://
wmbriggs.com/post/19677/.
Brown, Dan, (2018). Origin, Anchor Publishing.
Camillus, J., (2008). Strategy as a Wicked Problem. HBR, 86(5):98–101.
Clark, H.H., (1996). Using Language, Cambridge University Press
Cohen, J.E., (2012). What privacy is for (2013). Harvard Law Review, 126.
Gigerenzer, G. and R. Selten eds., (2002). Bounded Rationality: The Adaptive Toolbox, MIT
Press.
Harris, R., (2012). Introduction to Decision Making. accessed on 19/10/2019 at https://www.
virtualsalt.com/crebook5.htm.
Henn, S., (2014). If There’s Privacy in the Digital Age, It Has A New Definition. accessed
12/10/2019 at https://www.npr.org/sections/alltechconsidered/2014/03/03/28533
4820/if-theres-privacy-in-the-digital-age-it-has-a-new-definition.
Kahneman, D., (2011). Thinking, Fast and Slow, Macmillan.
Klein, G., (2007). The Power of Intuition: How to Use Your Gut Feelings to Make Better
Decisions at Work, Crown Publishing Group.
Law, Jonathan, (2015). A Dictionary of Law, Oxford University Press.
Liebowitz, J. (ed.), (2019). Developing Informed Intuition for Decision Making, Taylor &
Francis.
Liebowitz, J., Y. Chan, T. Jenkin, D. Spicker, J. Paliszkiewicz, and F. Babiloni (eds.), (2019).
How Well Do Executives Trust Their Intuition, Taylor & Francis.
Mathieu, Joseph, (2019). Technology and Truth in the Digital Age. accessed 10/19/2018 at
https://newsroom.carleton.ca/story/technology-truth-digital-age/.
Neal, L. and C. Spetzler, (2015). An Organization-Wide Approach to Good Decision-
Making. Harvard Business Review.
Simon, H.A. and J. March, (1976). Administrative Behavior: A Study of Decision-Making
Processes in Administrative Organizations, Free Press.
The Anatomy of a Great Decision [WWW Document], (2019), Farnam Street. accessed
1/6/2020 at https://fs.blog/2019/04/decision-anatomy/.
Trewatha, R.L. and M.G. Newport, (1976). Management: Functions and Behavior, Business
Publications, Dallas.
Turing, A.M., (1950). I.—Computing Machinery and Intelligence. Mind:LIX, 433–460.
doi:10.1093/mind/LIX.236.433.
Chapter 4
Measurement Issues
in the Uncanny Valley:
The Interaction between
Artificial Intelligence
and Data Analytics
Douglas A. Samuelson
Contents
4.1 A Momentous Night in the Cold War.........................................................54
4.2 Cybersecurity..............................................................................................55
4.3 Measuring AI/ML Performance..................................................................56
4.4 Data Input to AI Systems............................................................................58
4.5 Defining Objectives....................................................................................59
4.6 Ethics..........................................................................................................60
4.7 Sharing Data—or Not................................................................................61
4.8 Developing an AI-Aware Culture................................................................62
4.9 Conclusion..................................................................................................62
References............................................................................................................62
53
54 ◾ Data Analytics and AI
Artificial intelligence (AI) researchers call it “the uncanny valley”: numerous stud-
ies indicate that people fear and distrust machines that seem nearly human, but
not quite. There are many theories about this, but nothing conclusive (“Uncanny
Valley,” Wikipedia, 2020).
The phenomenon can be extended, admittedly without strong empirical evi-
dence of its existence and extent, to many subject areas relating to AI—or machine
learning (ML). The latter term has become more popular than the original among
researchers still queasy about the history, over the past 50 years or so, of AI being
oversold and over-promised, with disappointing outcomes. In particular, it is worth-
while to address the question of how well we know whether an AI/ML system has
performed satisfactorily. If it did not work exactly as its developers predicted, how
good was it? How sure are we about that assessment? In short, the performance of
AI/ML systems is a subject area that clearly requires better measurement and assess-
ment—which, of course, is exactly what good data analytics is about.
The dependence operates in the opposite direction, as well. There are many
efforts, recent and ongoing, to improve data analysis using AI/ML techniques.
Reviewing and assessing these efforts is well beyond the scope of this chapter—and
the author’s expertise. What is relevant, however, is the resulting paradox: if AI/ML
methods can quickly solve data analysis problems that defy traditional inference
techniques, how sure can we be that the AI/ML solution is correct? How can we
better assess whether to trust the AI/ML answer?
These are not merely theoretical issues. AI/ML systems are increasingly in use
in a number of application areas, some of which are literally life or death decisions.
AI/ML systems are contemplated to direct swarm and counter-swarm warfare
involving thousands of supersonic unmanned vehicles. This is a situation in which
humans would be incapable of making judgments sufficiently quickly, much less
then translating those judgments into thousands of movement and action orders in
seconds. Therefore, despite the best intentions and admonitions from the AI/ML
research community, we could soon have AI/ML systems making kill decisions.
Knowing how much to trust the machines is, therefore, critically important and
becoming even more so.
All eyes in the command center were on LTC Petrov. He recounted later, “I felt
as if I was sitting in a hot frying pan.” In a real attack, a few minutes’ delay could
cost millions of the lives he was there to protect. However, reacting to a false alarm
would precipitate an immense catastrophe, literally ending human civilization as
we have known it.
Fortunately, he had another warning system, based on ground-based radars,
that he could check. He decided to wait for the ground-based radars to confirm the
launches. “I just had this intuition,” he explained later, “that the U.S. would not
launch five missiles. Either they would launch one or two, to show they were seri-
ous, and give us an ultimatum, or they would launch all 1,053. So I did—nothing. I
was afraid that informing my superiors, in accordance with my orders, would start
a process that would acquire a momentum of its own” (Petrov obituary, New York
Times, September 18, 2017).
LTC Petrov’s story has many implications, but one in particular is noteworthy
here: he made the right decision because he acted counter to his orders and refused to
trust the machine’s conclusion. His intuition was correct. He had contextual informa-
tion the machine-based system did not. (The Soviets had analyzed and wargamed
what attack profiles the US might employ in various situations.) But suppose the
Soviets, or the Americans, or whoever else developed a new detection and warning
system, more powerful, more reliable, arguably more trustworthy. How could such
a system be taught the intuition on which LTC Petrov relied? How would we know
that the system had enough such intuition to be trusted? At this time, these ques-
tions are totally unanswered in the AI/ML research.
4.2 Cybersecurity
At least it is possible in kinetic combat to assess some results quickly. In non-
kinetic conflict, such as economic and diplomatic confrontations, effects take
much longer to appear and are then much harder to link back to causes. In infor-
mation systems conflicts, the difficulty is even greater. Douglas W. Hubbard, a
well-known expert on measurement and risk assessment, declares, “The biggest
risk in cybersecurity is not measuring cybersecurity risk correctly” (Hubbard,
2016). What he meant by this is that the threats are mostly events that have
never happened, so estimating their probability of occurrence becomes a highly
judgmental exercise. The use of Bayesian methods is promising, but then the
analyst faces the danger of introducing overly influential biases in the choice of
prior probability distributions and in the selection of the presumed process to be
modeled. Training analysts to calibrate their estimates of uncertainty—that is, to
have a much better understanding of how uncertain they are about their conjec-
tures—improves the resulting risk assessments. In contrast, many popular meth-
ods and techniques increase estimators’ confidence in their estimates without
actual improvement. If, as in cybersecurity, we cannot avoid relying on opinions,
56 ◾ Data Analytics and AI
we can at least train the people forming those opinions to be more realistic about
how likely they are to be right.
It is also useful to get senior decision-makers accustomed to the fact that analy-
ses based on guesses do not produce certainty or even near-certainty. Moreover,
analyses based on highly imprecise data cannot produce conclusions any more pre-
cise than the least precise input. Better trained analysts can more effectively insist
on these principles when senior decision-makers push for unrealistic assurances.
This is one of the areas in which better data analytics can drive better AI.
If anything, Hubbard seems to have been optimistic: it is not clear that many
organizations with major responsibilities in the field can even define cyberse-
curity risk—or even just what cybersecurity is. Some influential organizations
advocate developing and applying a maturity model to information manage-
ment: if everything is forced into well-defined processes, and those processes
are rigorously followed, then perfect security will ensue. This approach has a
fundamental flaw. The most important fact anyone can know about a detection sys-
tem is what it can’t detect. However, no metrics of observed adherence to defined
processes yield any information on this all-important subject. Only challenge
testing does. This, in turn, has the limitation that one cannot test the system’s
response to challenges one never imagined. Still, with metrics depicting the
range of challenges the system has detected, it is at least possible to compare
systems and rank them in terms of demonstrated responsiveness to these speci-
fied types of threats.
Another method can be inferred from one of the most famous philosophical
problems in AI: Searle’s Chinese Room (Cole, 2004). He posited an experiment
in which two translators are in opaque, soundproof cells, and we pass in Chinese
scripts and they pass out English translations. If the two translations seem equally
good, to our inexpert view, we conclude that the translators are also equally good.
Now, if it turns out that one cell actually contained an expert Chinese translator
and the other contained a novice with an excellent Chinese-English dictionary,
Searle argues that any translating ability in the second cell was in the dictionary,
not the person using it. From this, he argues that AI is a nonsensical concept, as the
real intelligence lies in the software, not the machine. This argument has spawned
lively debate.
Often lost in the debate is that the setup is what is nonsensical, or at least highly
forced and unrealistic. Searle explicitly excludes from the experiment what any
reasonable evaluator would do: ask each translator to critique the other’s transla-
tion! The difference between naively relying on the literal meanings of words and,
in contrast, understanding implied context and the use of idiom would readily
emerge. For example, in 2009, actress Sharon Stone posed nude for Paris Match
(Huffington Post, 2009). (Lest any sensibilities be offended, please note that this
reference tells the story but does not display the nude photos.) The headline read,
“J’ai 50 ans, et – alors!” Some American headline writer with a dictionary and
absolutely no knowledge of idiomatic French rendered this as, “I’m 50 years old,
and then some!” This version—usually without the photo, in the US—got wide
circulation. But the correct meaning is, “I’m 50 years old, and look here!”
Having AI systems critique each other to correct errors is also not a new idea.
The Space Shuttle had four computers. They had to agree before taking a critical
action, such as a course change. If one disagreed with the others some number of
times, the other three ignored it and eventually shut it down or advised ground-
based control to do so. So an important concept in evaluation is already in practice
to some extent: AI/ML systems supporting or advising critical decision-making
should be required and designed to critique each other’s performance and help
human decision-makers understand how solid the AI systems think their recom-
mendations are. (This, of course, is exactly what human decision-makers are also
usually advised and trained to do. Somehow, however, this ancient wisdom is some-
times forgotten in the design and use of AI/ML systems.)
Another useful approach is—again—challenge testing. One of the tasks AI
systems are often assigned is to reason out the likely behavior of another party, espe-
cially an adversary. There has been much research and even more speculation about
how to do this. But, to this author’s knowledge, nobody has ever implemented
what I call the Trogdor Test. Trogdor is a rather simple, silly online game in which
a dragon, Trogdor, moves through a landscape stomping on peasants (“Trogdor,”
2020). When he stomps on a specified number of peasants, he acquires the power
to burn huts. Burning all the huts in a village advances him to the next level. The
player moves Trogdor around using the arrow keys on the keyboard.
58 ◾ Data Analytics and AI
Trogdor must face deadly attacks from archers shooting arrows at him from
the side boundaries, and from two sword-wielding knights who prowl the land-
scape. Avoiding the arrows is simply a matter of staying alert, seeing them coming,
and getting out of the way. But a player needs to know how to move to avoid the
knights, or else Trogdor dies by the sword. Simply staying as far as possible from
the knights doesn’t work, as they eventually converge systematically on Trogdor. A
good approach is to recognize occasions, when the knights are some distance apart,
when Trogdor can safely move between them from one region to another. This
apparently resets the knights’ movements to random, halting their more effective
search-and-seek movements.
Now, clearly these knights’ movements are dictated—entirely and invari-
ably—by a computer program. It is therefore possible to ascertain exactly what
their behavior is, for all situations. So the test is: can the AI, by playing the game,
deduce the knights’ behavioral programming? Unlike most situations in real life,
we have documented truth against which to compare any conjectures an AI pro-
duces. So here is a modest proposal: any AI for which the developers claim an
ability to deduce opponents’ behavior should be required to pass the Trogdor Test.
(The present author is aware that widespread adoption of this proposal could bring
huge riches to the company that produced and runs the Trogdor game. The present
author hastens to add that he has no financial ties to that company or its owners.)
In the interests of fairness and impartiality, it is also proper to concede that
other games could furnish an equally valid test. The point is that such games do
exist, and it is simply ludicrous not to require AI systems to demonstrate, by means
we have now shown to be readily available, that they can reproduce documented
reality for at least one of the rare instances in which it exists.
If you asked a human sixth grader to identify the 16th most populous city in
the US, probably the sixth grader would find a list of cities and their populations,
rank them, and count down to 16. As of a few years ago, it is my understanding
that Watson couldn’t deduce that procedure. (Private communication from a friend
at IBM, 2016). It needed to find a document stating, “The sixteenth most popu-
lous city in the US is…” In short, without quite a bit of structuring of data input,
AI/ML systems can prove remarkably lacking in what humans would deem to be
simple cognitive tasks.
The difficulties AI/ML systems have with data are compounded by some cur-
rent practices in data analytics. Many people have had the experience of having a
library moved by professional movers. The movers are concerned with which books
will pack together tightly in the boxes. Size, not content, matters. The result is that
the move goes smoothly, the books arrive undamaged, and the owner may then
need weeks to put books of different sizes but related content back together on the
bookshelves.
In big data analytics, much efficiency in input processing can be obtained
by “sharding” the data, breaking the data into more manageable smaller chunks
(“Sharding,” 2020). In some cases, this might consist of something sensible, like
separating the data set into geographic regions, in which there is little interaction
among data elements in different regions. In other cases, however, the process is
more like what the movers did, matching the next incoming chunk of data to the
storage area closest to it in size. This achieves near-maximum efficiency in the use
of storage space at the cost of vast inefficiencies in retrieval (Samuelson, 2014). “The
computer is fast, it will find it soon enough” is a recipe for disaster when the data,
stripped of all metadata and contextual information, are thus stored willy-nilly—in
terabytes. The question of how much metadata to keep, and how much content
relationships can or should influence storage locations, looks like a promising area
for research. In general, research is called for into many issues of how best to pre-
pare data to be used by AI/ML systems—focusing on facility of use rather than
ease of storage.
4.5 Defining Objectives
Another issue too little addressed, in both AI/ML and data analytics, is—as in the
discussion of cybersecurity risk, above—defining the objective. One of the more
trenchant criticisms that came to be raised about US policy in Vietnam in the
1960s was that nobody seemed sure of how success was defined, hence how it was
to be recognized. This was part of why the administration lost credibility by repeat-
edly assuring the American public that success was near at hand—and then rede-
fining what that meant when the conflict dragged on.
Even at that tender age, many of my contemporaries had learned, some from
relying on insight gained from board games such as Avalon Hill “D-Day,” “Afrika
60 ◾ Data Analytics and AI
Korps,” and “Panzer Blitz,” that a clear and unambiguous definition of victory
conditions is essential to success in any competition. Attempting to construct a
game and define victory conditions quickly impresses on the designer the practical
necessity of specifying a criterion less than “total victory,” whatever that means. The
Allies did not need to kill every German soldier in Europe to reclaim the continent
from Nazi domination. In any poorly structured debate over policy and strategy,
there will always be some fool who insists that nothing short of total victory is
acceptable. A cursory overview of history indicates that pursuing this insistence
leads to disaster. To cite just one example, Thucydides, in assessing the outcome of
the Pelopponesian War, attributed the ultimate defeat of the Athenians to “want
of more,” the pursuit of additional gain after the major objectives had been mostly
won (Thucydides, trans. Strassler, 2008).
Of course, these historical observations are still subject to dispute. The reader
will also readily recognize that the appropriate definition of success is complicated,
controversial, and most likely unending. But this ambiguity is precisely the point
for AI: it is inconceivable that humans can clearly and unarguably define success
and teach that definition to a machine, when no such definition can be constructed
by humans. Moreover, machines tend to have a much lower tolerance for ambiguity
than humans. What humans explain as “feels close enough” calls upon intuition
that machines do not have. Even with human subordinates, but especially with
machines, much more specificity and detail in instructions are necessary to avoid
damaging failures. In particular, we are challenged to develop metrics that char-
acterize the outcomes, including the associated uncertainties, much more precisely
than we have been accustomed to doing.
4.6 Ethics
The ethics of constructing, using, limiting, counteracting, and generally trying
to live with AI/ML systems constitute a subject area far beyond the scope of this
chapter. A few observations seem in order, however.
First, it is widely agreed that machines should not be trusted with a kill deci-
sion. One proposed formulation of the principle is, “Never trust a kill decision or
recommendation to any entity—whether an individual human, a group of humans,
a trained animal, a machine, an organization—that is incapable of remorse.” Like
Asimov’s Laws of Robotics, this seems to encompass all the meaningful situa-
tions. Also like the Laws of Robotics, there are many ways to bend this principle,
get around it, misuse it, and create chaos and mishaps. Still, it seems useful as a
guideline.
One of the problems is, as mentioned earlier, that there are serious capa-
bilities in development, in many advanced countries, to use AI/ML to direct
swarms of extremely fast-moving unmanned vehicles against opposing swarms
of extremely fast-moving unmanned vehicles, generating a combat decision space
Measurement Issues in the Uncanny Valley ◾ 61
far beyond the attention and control span of humans. This raises, among other
serious questions, grave ethical concerns about how such combat systems can be
properly limited—or perhaps whether they should be prohibited, as poison gas
and toxic organisms are, as too dangerous. Such a prohibition, of course, raises
new questions about detection of and response to the now-outlawed creation of
such systems—with attendant measurement and risk assessment challenges that
data analytics can address.
How to evaluate and report uncertainty also has ethical aspects. Decision-
makers want to be sure, even when the supporting analysis indicates that certainty
is unattainable. When the decision could have grave consequences, “how sure are
we really?” becomes more than an analytical disagreement. Add to this the trend of
modern science: as methods and techniques progress, we tend to drive uncertainty from
the computation to the assumptions, where it is much more difficult to detect. Similarly,
as complex systems are designed to drive the probability of minor failures lower and
lower, there is an unintended consequence that the only possible failures remaining are
big, rare, multiple-cause events that are more likely, when they happen, to overwhelm
the respondents.
This trend has a counterpart in the management of organizations that analyze
data and make and support decisions. Management is becoming more and more
adept at insisting on what they want and ignoring or marginalizing objections from
subordinates. Ironically, the US, both in the government and in the private sector,
is spending billions of dollars to teach machines to think like humans, while spend-
ing trillions of dollars to teach and incentivize humans to act more like automata.
4.9 Conclusion
AI/ML and data analytics have much to contribute to each other, especially in
evaluating the performance of tools, technologies, and methodologies that sup-
port critical decision-making. Data analytics can and should be used much more
broadly and creatively to assess the performance and risks of AI/ML systems. AI/
ML systems can extend and improve data analytics. These efforts will also sup-
port needed analyses of ethical and policy considerations. The development and
sustenance of a workforce that can better understand, use, explain, and critique
AI/ML and data analytics is critical to our society, as these technologies offer great
potential for both good and harm.
References
Cole (2004). https://plato.stanford.edu › entries › chinese-room retrieved January 9, 2020.
Hubbard, D.W, Seierson R. (2016). How to Measure Anything in Cybersecurity Risk, New
York: Wiley.
Measurement Issues in the Uncanny Valley ◾ 63
An Overview of Deep
Learning in Industry
Quan Le, Luis Miralles-Pechuán,
Shridhar Kulkarni, Jing Su, and Oisín Boydell
Contents
5.1 Introduction............................................................................................... 66
5.1.1 An Overview of Deep Learning...................................................... 66
5.1.1.1 Deep Learning Architectures.............................................67
5.1.2 Deep Generative Models..................................................................69
5.1.3 Deep Reinforcement Learning.........................................................69
5.2 Applications of Deep Learning....................................................................70
5.2.1 Recognition.....................................................................................70
5.2.1.1 Recognition in Text...........................................................71
5.2.1.2 Recognition in Audio........................................................72
5.2.1.3 Recognition in Video and Images......................................72
5.2.2 Content Generation.........................................................................76
5.2.2.1 Text Generation.................................................................76
5.2.2.2 Audio Generation..............................................................78
5.2.2.3 Image and Video Generation.............................................79
5.2.3 Decision-Making.............................................................................81
5.2.3.1 Autonomous Driving.........................................................82
5.2.3.2 Automatic Game Playing...................................................83
5.2.3.3 Robotics............................................................................85
5.2.3.4 Energy Consumption........................................................86
5.2.3.5 Online Advertising............................................................87
65
66 ◾ Data Analytics and AI
5.2.4 Forecasting......................................................................................88
5.2.4.1 Forecasting Physical Signals...............................................88
5.2.4.2 Forecasting Financial Data............................................... 90
5.3 Conclusion..................................................................................................91
References............................................................................................................92
5.1 Introduction
Applications driven by deep learning are transforming our society. To name but
a few examples: Google Assistant supports real-time speech-to-speech translation
between 44 languages—helping to break down human communication barriers,
Amazon Alexa understands human voice commands and assists in many everyday
tasks such as ordering products and interacting with our smart homes, autonomous
driving is assessed to be safer than human drivers, and deep learning models are
helping us to understand our genome and to develop precision medicine.* Deep
learning’s ability to learn highly accurate representations of the task at hand, given
enough annotated training data, helps it achieve better accuracy than traditional
machine learning methods in many cases. This capability has opened up opportu-
nities for its application to many new problems where traditional machine learning
methods have fallen short.
In this chapter, we introduce the reader to many important applications of
deep learning in industry. To begin, we provide a high-level overview of deep
learning and its key architectures. We follow with a survey and discussion of the
main applications of deep learning, categorized into four general tasks: recogni-
tion; generation; decision-making; and forecasting. We conclude the chapter with
a discussion on the strengths and weakness, as well as the future applications of
deep learning.
As such, deep learning models do two things at the same time: learning an effective
feature representation from the raw inputs and learning a model that maps from
the feature representation to the outputs. The early layers of the models may only
capture simple features calculated from the raw inputs, but the later layers combine
these simple features to learn more abstract features which are optimized for the
task at hand.
maximize the cumulative reward based on the experience obtained when interact-
ing with the environment (e.g., playing a game, manipulating a robot, or activating
and deactivating the heater). Some examples of RL objectives are playing online
games at a human level, driving cars without human intervention, or managing
traffic lights to reduce traffic.
RL was developed a few decades ago but, because of some limitations, it was
unable to reach its full potential. This situation changed in recent years with the
development of deep learning. Deep learning algorithms, for example, the popular
deep Q-learning algorithm is able to approximate the Q-Table with an artificial
neural network very effectively in terms of memory and processing requirements.
The combination of deep learning and RL is called deep reinforcement learning
(DRL) and has multiplied the potential of RL, boosting the general interest in RL
of the industry and the scientific community.
5.2.1 Recognition
In the context of machine learning, and specifically deep learning, recognition is
defined as a task of identifying predefined labels from unannotated inputs—for
5.2.1.1 Recognition in Text
Much of the digital data we collect and analyze is text-based (e.g., scientific litera-
ture, news articles, contracts, and medical reports), and the popularity of social
media has led to an online text explosion. Deep learning has been applied suc-
cessfully in natural language processing (Collobert et al. 2011) where deep feed-
forward neural networks (often with the use of convolution layers) are used to learn
features automatically and perform competitively in key tasks along the natural
language processing pipeline.
Deep text models are the dominant approach for sentiment analysis, usually
a part of any text analytics pipeline. Modern sentiment analysis systems can be
decomposed into different subtasks: target detection and polarity classification
(Zhang et al. 2018). Li et al. (2017) proposed deep memory networks for both
of these tasks. Other text information extraction tasks that have benefited from
the application of deep learning include topic classification, semantic role label-
ing, part of speech tagging, named entity recognition, and chunking (Young
et al. 2018).
In the era of fake news, important work is being done on using deep neural
networks—usually a combination of CNN and RNN models—to detect decep-
tive opinions (Sharma et al. 2019). In a recent development, Graph CNN was
used to detect fake news on social media with high accuracy (Monti et al. 2019).
FakeNewsAI* is an example of a commercial news verification service based on a
deep learning architecture.
Another commercial application of a deep text model is in the Google search
engine, where the attention based BERT model is used to rank results returned
from a search query.† In an example of a more end-to-end application of deep learn-
ing, Amazon’s Alexa personal assistant uses LSTM to classify text and recognize
the commands of the users (Naik et al. 2018). Apple’s Siri‡ and Google’s Assistant§
also use similar techniques.
* www.fakenewsai.com
† https://www.blog.google/products/search/search-language-understanding-bert/
‡ www.apple.com/ios/siri/
§ assistant.google.com
72 ◾ Data Analytics and AI
5.2.1.2 Recognition in Audio
Making sense of audio has long been a focus of the machine learning commu-
nity, and deep learning has made significant progress in this field which leads to
important commercial applications. Conversational speech recognition (CSR) is
one of the most important such applications. In 2016, a series of breakthroughs
were reported in CSR based on the advances of deep neural networks. For example,
IBM researchers (Saon et al. 2016) achieved a 6.6% error rate with a large vocabu-
lary English conversational telephone CSR system using a combination of CNNs
and RNNs. In 2016, Microsoft achieved the milestone of reaching human parity
in speech recognition for English with a 5.9% error rate on a similar task (Xiong
et al. 2016). For commercial use cases, Google uses speech recognition to generate
subtitles for videos automatically as well as to carry out voice searches.
Speech recognition is a more challenging task in noisy environments or in con-
versations. Different from single-speaker speech recognition, a step of vocal source
separation is needed. For example, in order to process recordings from a cocktail
party, we need to identify the voice from one speaker out of other speakers as well
as background noise. Simpson et al. (2015) have applied CNNs to this problem to
very good effect.
Moving away from speech, identifying the characteristics of music has been
widely studied in machine learning and again deep learning is making inroads
here. For music recommendation systems, automatic music genre recognition is
a key step and deep learning brings a new approach in this area. Julien Despois
demonstrated how to classify the genre of a piece of music or a complete song with
CNNs* achieving genre recognition accuracies in the high 90% range. Niland is
an important commercial player in the application of deep learning to music and in
2017 was acquired by Spotify.†
* chatbotslife.com/finding-the-genre-of-a-song-with-deep-learning-da8f59a61194
† www.spotify.com
An Overview of Deep Learning in Industry ◾ 73
Network (He et al. 2015) achieved the landmark of performing better than human
performance. Figure 5.3 shows an example of image classification on a CNN model
trained on the ImageNet data set.
Similar CNN-based approaches have been used for image recognition tasks
such as Google photo search for medical application (Hegde et al. 2019), image
caption generation, and video frame classification.* Salesforce research arm† is an
example of a commercial application that uses deep models for textual sentiment
analysis, as well as image classification tasks. There are also good examples of deep
learning-based image recognition solutions being used to drive revenue in niche
areas. For example, Deepomatic‡ have leveraged deep learning to build commer-
cially successful image tagging services in domains including fashion, security, and
interior design.§ Similarly, Tractable¶ are using deep learning to estimate the cost
of repair for insurance claims by recognizing the amount of damage in images of
cars involved in incidents.
Face recognition is a long-standing image processing challenge. Prior to the
introduction of deep learning models, state-of-the-art approaches to face recogni-
tion in images relied on first recognizing a set of carefully selected hand-crafted
features within an image using image processing techniques, and then using these
as an input to a machine learning model. CNNs enable very accurate facial recog-
nition in an end-to-end system. For example, the DeepFace system from Facebook
(Taigman et al. 2014) is a nine-layer deep CNN-like neural network used for face
recognition. The DeepFace system was shown to achieve an accuracy of 97.35%
on the Labeled Faces in the Wild (LFW) data set (Huang Erik Learned-Miller,
2014) (a well-known and challenging face recognition benchmark), which is sig-
nificantly better than the state-of-the-art non deep learning approaches prior
to 2014. Subsequent to the release of DeepFace, Google researchers introduced
FaceNet (Schroff et al. 2015), which was also based on CNNs and achieved 99.63%
* cs.stanford.edu/people/karpathy/deepvideo/
† https://einstein.ai/
‡ www.deepomatic.com
§ www.deepomatic.com/demos
¶ www.tractable.ai/technology
74 ◾ Data Analytics and AI
accuracy on the LFW data set. Face recognition is an active area of research with
many new deep learning models being proposed recently; for example, SphereFace
and ArcFace (Wang and Deng, 2018). It has also been widely applied in business:
from mobile systems that implement the technique to log users in automatically, to
many law enforcement agencies who use it to detect criminals.
The success of deep learning models in image classification has translated to
many successful applications of deep models in medical images, often through
the use of the transfer learning technique. One notable example is the system
described by Ciresan et al. (2013) that won the MICCAI 2013 Grand Challenge on
Mitosis Detection. The mitosis detection task is particularly interesting as unlike
the ILSVRC image classification task (Russakovsky et al. 2015), the goal is not to
classify an entire image as belonging to a category, but rather to identify the por-
tions of a large image that belong to a particular category—in this case, examples
of mitosis in an image. This is referred to as image segmentation and is at the
core of many image processing tasks. In recent years, deep models have attained
human expert level performance in multiple tasks, including melanoma screening
and detection, identifying diabetic retinopathy, cardiovascular risk analysis, and
pneumonia detection on chest X-rays (Esteva et al. 2019).
The vast amounts of aerial imagery enabled by the lowering costs in satellite
technology and the prevalence of low-cost aerial drones underpin another area in
which deep learning-based image recognition is being widely applied. In 2010,
Mnih and Hinton (Mnih and Hinton, 2015) produced a pioneering work in which
deep learning methods were used to identify roads in aerial images. This is an
example of an image segmentation problem (similar to the medical image recogni-
tion problem described previously) as the system not only recognizes that a road is
present in an image but also the specific pixels in the image that contain the road.
Marcu (Marcu, 2016) used CNNs to very accurately segment portions of aerial
images into semantically meaningful categories (e.g., roads, buildings, parks).
Recently, Facebook used deep learning to automatically segment roads from sat-
ellite images and generate accurate maps.* Their work automates the costly and
time-consuming process of annotating maps manually, which will be helpful
in many unmapped regions of the world—especially the regions in developing
countries.
Aiming for the insurance industry, TensorFlight† uses these techniques to
analyze aerial images and to provide automatic annotation on construction type,
building footprint, and roof characteristics. Terrapattern‡is an interesting exam-
ple of a group adopting similar deep learning-based approaches to build an aerial
photo search engine that will find common patterns in massive collections of aerial
* ai.facebook.com/blog/mapping-roads-through-deep-learning-and-weakly-supervised-train-
ing/
† www.tensorflight.com
‡ www.terrapattern.com/about
An Overview of Deep Learning in Industry ◾ 75
imagery. Figure 5.4* shows an example set of search results for wastewater treat-
ment plants.
Extending systems that recognize objects in images to systems that recognize
objects in video is an obvious step and modifications to the core deep learning
approaches (e.g., CNNs) to work on video have been shown to work well (Girshick,
2015; Ren et al. 2015). Clarifai† is an interesting startup working on automatic
object recognition in video using deep learning for a wide range of tasks in indus-
try. They are especially focused on the advertising industry and they use their tech-
nology to find appropriate videos in which to place ads.
Deep learning has allowed a step change in the performance of systems built to
recognize objects in images and are now the de facto standard for that task. CNNs
can be used both to classify entire images or to segment objects within images. It
is worth noting that object recognition in non-standard images remains very chal-
lenging. For example, while it is possible (e.g., Valdenegro-Toro, 2016), object rec-
ognition in the sonar images collected by autonomous underwater vehicles (AUVs),
widely used in the oil and gas industry, remains very difficult for automated systems.
Similarly, it is worth noting that almost real-time object recognition is required in
certain applications (e.g., autonomous vehicle control). While this can be achieved
in some cases (e.g., Iandola et al. [2016] for traffic light recognition), it remains a
significant challenge in using deep learning as significant computation is required
to use a deep network to make a prediction. GANs are also beginning to become
Figure 5.4 GeoVisual search results for wastewater treatment plants in satellite
imagery.
* www.medium.com/descartestech/geovisual-search-using-computer-vision-to-explore-the-
earth-275d970c60cf
† www.clarifai.com
76 ◾ Data Analytics and AI
widely adopted for image recognition tasks but only for specialist applications and
CNNs remain much more popular.
5.2.2 Content Generation
Rather than recognizing what is available in the data, in generation tasks, the
objective is to output novel or additive content based on input data. Examples
include generating captions for images, converting a piece of music into a new
style (Hadjeres and Pachet, 2016), or composing entire documents. This section
surveys key applications of deep learning for generating novel content. The ability
to build machine learning systems that generate new content is something that did
not really exist before the advent of deep learning approaches and has spurred a
renewed interest in the area of computational creativity.
5.2.2.1 Text Generation
In terms of using deep learning for generation, text has attracted more attention than
any other data format. Deep learning approaches have been successfully applied
to many different tasks including generating captions for images, generating con-
versational responses for chatbots, generating screenplays, novels or speeches, and
machine translation. In this section, we describe a selection of interesting examples
of this application of deep learning technology.
Image and video captioning techniques are created to address the multimodal-
ity challenge of visual data and human language. A good captioning model needs to
identify key objects in view and output a fluent sentence showing correct relations
between the identified objects. Basically, there are two approaches in the research
community including end-to-end pipelines and stepwise captioning models.
The Show and Tell caption generator from Google researchers (Vinyals et al.
2016) gives an early example of end-to-end captioning pipelines. A CNN network
is employed to encode an input image to a fixed-length vector. Thereafter, this vector
is taken as the initial hidden state of an RNN network. RNN decodes the vector
into a sentence. There is no object detection step in this pipeline; loss of caption
error is counted from each generation step. The Show and Tell model features ease
of manipulation and quality of reading experience.
It was the winner of the Microsoft COCO 2015 Image Captioning Challenge.*
The NeuralTalk model sets a milestone of stepwise captioning practice (Karpathy
and Li, 2015). This model is not a single CNN plus an RNN. Instead, a more com-
plicated module is applied to extract visual features. The authors use a region con-
volutional neural network (R-CNN) (Girshick et al. 2014) to detect object regions
from an input image. This R-CNN was pre-trained on ImageNet. Thereafter, an
image-sentence score metric is introduced to find the maximum correspondence
* mscoco.org/dataset/#captions-leaderboard
An Overview of Deep Learning in Industry ◾ 77
between each object region and a word in caption sequence. In 2016, Karpathy
et al. released NeuralTalk2, a revision of the original system capable of more believ-
able captions.* While the current state-of-the-art of these captioning systems is not
yet capable of human-level performance, these systems are already being applied in
commercial offerings; for example, automatic captioning of images in the Facebook
newsfeed.†
Going further than simple image caption generation, the movie director Oscar
Sharp and the AI researcher Ross Goodwin developed Benjamin,‡ an LSTM-based
system that can generate original screenplays automatically. This is achieved by
training it with dozens of science fiction screenplays and then asking it to gener-
ate its own. Their system was capable of generating long sections of novel movie
script—one of which was actually filmed and released as the short film Sunspring.§
Although most of the text generation systems described so far are commercially
interesting, they have not yet seen wide industrial adoption. Machine translation
of texts from one language to another, on the other hand, is of massive commer-
cial value. Deep learning approaches to the machine translation task, commonly
referred to as neural machine translation (NMT), have led to a step change in the
performance of automated machine translation systems. Instead of using phrase-
level matching between two languages (as is done in older approaches to machine
translation), the NMT model works on entire sentences which provide NMT sys-
tems with the opportunity to model more contextual information than is possible
in other approaches. Google’s NMT system is a good example of a modern NMT
engine and it has three main components: the encoder LSTMs, the decoder LSTMs,
and an attention module (Wu et al. 2016). The encoder LSTMs transforms an input
sentence to a list of vector representations with one vector per symbol. The decoder
LSTMs takes the vectors from the encoders and generates one language symbol at a
time. The attention module regulates the decoders to focus on specific regions dur-
ing decoding to drive increased accuracy of translations, and their addition was an
important step in driving translation accuracy.
NMT systems reach translation error rates significantly below those statistical
machine translation (SMT) approaches. As a result, Facebook have moved their
entire translation system to an NMT-based solution based on LSTMs which will
handle more than 2,000 translation directions and six billion translations per day.
Skype by Microsoft has also deployed an NMT-based translation system. In this
case, speech is automatically translated from one language to another. The system
first performs speech recognition on the original language, then translates the text
to the destination language, before finally using a text-to-speech system to gener-
ate speech in the destination language, where all of these components rely on deep
* cs.stanford.edu/people/karpathy/neuraltalk2/demo.html
† www.wired.com/2016/04/facebook-using-ai-write-photo-captions-blind-users/
‡ bigcloud.io/filming-the-future-how-ai-directed-a-sci-fi-short/
§ www.arstechnica.com/the-multiverse/2016/06/an-ai-wrote-this-movie-and-its-strangely-
moving/
78 ◾ Data Analytics and AI
5.2.2.2 Audio Generation
It is also possible to use deep learning approaches to generate audio. Speech synthe-
sis is by far the most studied application but approaches to music composition and
sound effect generation have also been proposed. In this section, we describe some
of the most interesting applications of deep learning approaches to audio generation.
Generating authentic sounding artificial speech, or speech synthesis, has long
been a focus of artificial intelligence researchers. Deep neural networks, however,
bring new approaches to this long-standing challenge. WaveNet (van den Oord
et al. 2016), a deep autoregressive model developed by Google DeepMind, achieves
state-of-the-art performance on text-to-speech generation and the generated speech
audio is rated as subjectively natural by human raters. This performance is achieved
with a dilated CNN model that manages to model long-term temporal dependen-
cies with a much lower computational load than LSTM models.
Recently, text-to-speech synthesis techniques reached a new milestone after the
landmark of WaveNet (van den Oord et al. 2016), and Google researchers introduced
Tacotron 2 (Shen et al. 2017). This system employs a sequence-to-sequence model to
project textual character embeddings to spectrograms in the frequency domain. Then
a modified WaveNet model generates time-domain waveform samples from spectro-
gram features. Compared with WaveNet, Tacotron 2 has a better performance in
learning human pronunciations and its model size is significantly smaller.
Beyond text-to-speech (TTS) techniques, speech-to-speech (STS) has drawn
attention in recent years. Google researchers introduced a direct STS translation
tool, named as Translatotron (Jia et al. 2019). Traditionally, speech-to-speech trans-
lation is achieved in three steps (or models) including speech-to-text transcription
on the source language, text-to-text translation, and text-to-speech synthesis to
generate audio in the target language. This routine is well established with convinc-
ing accuracy, also it is widely deployed in commercial applications. Translatotron
is the first trial to merge the aforementioned three steps in one model and show its
value. Although the benchmark of Translatotron is slightly below a baseline model
on the Spanish-to-English translation task, this direct translation approach is able
to mimic the voice of the source speaker in the synthesized target speech.
As a side-effect of the advances on TTS, it is now easy to generate a fake voice
or speech toward a target person. An AI startup Dessa released a speech synthesis
model called RealTalk which creates the human voice perfectly.† Currently, details
of data set, models, and benchmarks are not publicly available, but people can try
to tell the real voice from the fake on this page.‡
* www.skype.com/en/features/skype-translator/
† medium.com/dessa-news/real-talk-speech-synthesis-5dd0897eef7f
‡ http://fakejoerogan.com/
An Overview of Deep Learning in Industry ◾ 79
Rather than generating speech from text, deep learning approaches have also
been used to generate sound effects based on video inputs. An artificial foley art-
ist* described by Owens et al. (2015) can reproduce sound effects for simple silent
videos based on an ensemble of CNN and LSTM models. A CNN model is trained
to extract high-level image features from each video frame. A sequence of these
image features (color and motion) is taken as input to an LSTM model, and the
LSTM model is trained to create an intermediate sound representation known as a
cochleagram. In the final step, the cochleagram is converted to waveforms through
an LSTM-based sound synthesis procedure. Although only applied in very simple
environments, the results are impressive.
Deep learning models can also be used to generate original music. DeepBach
(Hadjeres and Pachet, 2016), for example, uses an LSTM-based approach to
compose original chorales in the style of Bach. The model is composed of mul-
tiple LSTM and CNN models that are combined in an ensemble which given a
melody can produce harmonies for the alto, tenor, and bass voices. Similar sys-
tems based on RNNs that generate original music in other styles have also been
demonstrated—for example, music in the style of Mozart†,‡ or traditional Irish
music.§,¶
* vis.csail.mit.edu
† www.wise.io/tech/asking-rnn-and-ltsm-what-would-mozart-write
‡ www.hochart.fr/rnn/
§ highnoongmt.wordpress.com/2015/08/07/the-infinite-irish-trad-session/
¶ h ighnoongmt.wordpress.com/2015/05/22/lisls-stis-recurrent-neural-networks-for-folk-
music-generation/
** https://hyperconnect.com/?lang=en
80 ◾ Data Analytics and AI
seamlessly and naturally into scenes from hundreds of movies and TV shows using
just a single photograph.*
Deepfake techniques are developing fast and this is becoming a challenge for per-
sonal privacy and public security. It is not only humans that cannot tell a faked portrait
or video clip from the original, but advanced face recognition software is also being
cheated. Korshunov and Marcel (Korshunov and Marcel, 2018) performed a study
where the results showed that state-of-the-art recognition systems based on VGG and
Facenet neural networks are vulnerable to Deepfake videos, with 85.62% and 95.00%
false acceptance rates respectively. The best fake detection method is based on visual
quality metrics which shows an 8.97% error rate on high-quality Deepfakes.
In order to improve fake detection techniques, Korshunov and Marcel (2018)
released the first public available fake video data set, vidTIMIT. Tech giants also
joined this campaign. Recently, Google and collaborators released a deep fake detec-
tion data set with over 3,000 manipulated videos (Rössler et al. 2019). Facebook
and partner organizations started the Deepfake Detection Challenge (DFDC) and
funded over US$10 million to support this industry-wide effort.†
Image generation refers to the process of automatically creating new images
based on existing information sources. Deep learning has been applied in many
image generation tasks, including image (and video) super-resolution, image color-
ization, image generation from text or other images, and so-called neural art.
Image super-resolution (ISR) is an image generation problem in which the
resolution of a digital image is vastly increased through the application of algo-
rithms. In recent years, Microsoft researchers have applied CNNs to this problem
and achieved state-of-the-art restoration quality (Dong et al. 2014). Although deep
CNNs significantly improve the accuracy and speed of ISR, there still remains
a challenge of restoring the finer texture details. Ledig et al. (Ledig et al. 2016)
proposed SRGAN for image super-resolution. The SRGAN is capable of restor-
ing photo-realistic natural images for 4× upscaling factors. Recently, a team from
ElementAI developed HighRes-net,‡ a deep learning model capable of stitching
multiple low-resolution satellite images to create a super-resolution image. Unlike
other super-resolution models which could add fake details to the final image, their
model recovers the original details in the super-resolution version after aggregating
the information from multiple low-resolution ones. As such, their model has wide
applications: from automatic land management to mapping road networks.
While it remains a very challenging task (and performing it at a human level
is well beyond the current state-of-the-art), deep learning has led to advances in
the ability of systems to automatically generate images based on textual descrip-
tions. Systems that can do this can be helpful in graphic design, animation, and
* www.theverge.com/2019/9/2/20844338/zao-deepfake-app-movie-tv-show-face-replace-pri-
vacy-policy-concerns
† ai.facebook.com/blog/deepfake-detection-challenge/
‡ www.elementai.com/news/2019/computer-enhance-please
An Overview of Deep Learning in Industry ◾ 81
5.2.3 Decision-Making
The recognition and generation systems described in previous sections perform
niche tasks that are often embedded in larger systems. It is, however, also possible
to build end-to-end control systems using deep learning, in particular, DRL which
is, as described in Section 5.1.3, the combination of deep learning and reinforce-
ment learning.
In this subsection, we introduced systems mainly based in DRL that make deci-
sions continuously in dynamic and static environments. As it has been described by
Yuxi Li (2017), there is a wide range of applications areas where DRL can be effec-
tively applied such as dialogue systems, education, healthcare, or computer vision.
In the following sections, we focus on applications in autonomous driving, game
playing, robotics, energy consumption, online advertising, and finance.
5.2.3.1 Autonomous Driving
An autonomous car or self-driving car is “a vehicle that is capable of sensing its
environment and navigating without any human input” (Hussain, 2016). Deep
learning approaches are often used for object recognition as part of an autonomous
driving pipeline, and technologies based on these systems dominate the current
commercial autonomous driving efforts. Google’s self-driving car unit, for exam-
ple, started in 2009 and in the next seven years drove over two million miles of
test journeys on open roads. This car implements deep learning models extensively
for object recognition. Similarly, the Tesla Autopilot system incorporates Tesla-
developed deep learning systems based on CNNs for the tasks of recognizing
objects through vision, sonar, and radar sensors. There are also examples of smaller
startup self-driving car companies such as Drive.ai, which created a deep learning-
based software for autonomous vehicles, or Tealdrones.com, a startup that equips
drones with onboard deep learning modules for image recognition and navigation.
In this section, however, we are more interested in DRL based end-to-end con-
trol systems in which deep learning models are used not only for recognition tasks
An Overview of Deep Learning in Industry ◾ 83
but also to actually make control decisions based on inputs from cameras and other
sensors. The use of neural networks for end-to-end autonomous driving control has
a history that stretches back to the late 1980s. The ALVINN system (Pomerleau
and Pomerleau, 1989) or Nvidia’s DAVE-2 system are good examples of a modern
deep learning approach to controlling autonomous cars. In the case of the DAVE-2
system, CNN models generate steering commands based on the video input from
three cameras placed at the front of the car. To train the model, long segments of
recorded video, together with human steering, were used to link the correct steering
commands with the camera inputs. The DAVE-2 system has achieved impressive
results in simulation and has driven long journeys in a fully autonomous mode.
Deep Tesla is another interesting example of an end-to-end autonomous car con-
trol system based on CNNs. However, due to some limitations, it is much more
likely that, for the moment, deep learning models will be used for developing spe-
cific components of self-driving car control systems such as pedestrian detection or
road-sign recognition.
As described in Section 5.1.3, RL creates a table that associates states to actions.
In such a way that the agent driving the car is constantly looking up the table to see
which is the best action for each state. RL cannot be applied in training mode in
real scenarios because the agent will take random actions to learn and this can be
very dangerous (Sallab et al. 2017). This is a big handicap for real-world applications
because it is not possible in real life to have accidents in order to learn. However,
there are simulations of the environment in which the car can learn how to behave
and once the car learns the right actions, it can be deployed in the real world. For
example, in an investigation carried out by Dong Li et al. (2018), neural networks
were trained with the images obtained from a car simulator called TORCS (The
Open Racing Car Simulator). The implemented model used to drive the car was
composed of two modules: one based on multitasking neural networks, which were
responsible for taking the driver’s vision as an input, and another one based on
DRL, which was responsible for making decisions from the extracted features of
the neural network. This proposed model was capable of driving the car with great
precision and was also able to adapt to new screens (situations) not previously seen
by the system. The lack of real-world simulators is one of the main limitations of
implementing DRL end-to-end systems in cars, nevertheless, the number of cars
driving with multiple cameras makes it much more feasible to create a real-world
simulator where DRL can be trained.
action with the highest expected reward based on an image of the current game
state. The reward is the mechanism by which RL algorithms learn. For example, in
Atari 2600 games, the agent gets a positive reward when it increases the score and
a negative reward when the agent loses a game. In the case of the DQNs applied
to the Atari 2600 (Van Hasselt et al. 2016), it was notable that it only utilized an
image of the game screen as the input. The agent was able to learn by itself and it
achieved superhuman levels after playing the game for several hours.
Deep Q-Networks can be distinguished from other deep learning architectures
by the fact that they do not require labeled training sets. Rather, training DQNs
involves multiple iterations of experimentation, the success of which is measured
using an appropriate reward function. For the DeepMind Atari game playing sys-
tem, the reward function was the score achieved in a round of playing the game.
By playing hundreds of thousands of rounds, the system used the reward to modify
the parameters of a deep network and to guide it to a version that could achieve
superhuman performance at the game.
The DQN algorithm inspired many researchers to develop control systems for
other games. For example, MIT researchers developed DeepTraffic, a gamification
of highway traffic, Lin applied a DQN to play FlappyBird,* and Tampuu et al.
(2015) applied a DQN to a multiagent Pong game where each agent was controlled
by a DQN.
It is also worth including board games among the applications discussed here
for two reasons: first, the level of difficulty that they entail, and second if a DQN is
able to solve difficult problems, it will also be able to solve the easy ones. The prob-
lem of making a machine able to play chess or the game of Go better than a human
has been a challenge for AI since its beginning (Silver et al. 2018). Some remarkable
authors such as Alan Turing and John Von Neumann tried to develop hardware
and software to enable computers to play board games. One of the milestones of AI
was achieved when the Deep Blue program beat the world chess champion in 1997.
However, the development of these programs required great levels of supervision by
experts in both chess and coding (Figure 5.6).
The most widely known example of a deep learning approach to play board
games is probably that of DeepMind’s AlphaGo, which is an autonomous agent
for the game of Go. AlphaGo defeated the world Go champion, Lee Sedol, 4-1
in March 2016 (Chouard, 2016) and continues to beat world-class human play-
ers. The AlphaGo model uses deep convolutional neural networks and a general
tree search algorithm (Silver et al. 2017). The architecture of AlphaGo is especially
interesting as it is a hybrid system incorporating Monte Carlo tree search algo-
rithms, supervised learning (SL) policy networks, reinforcement learning policy
networks, and value networks (Silver et al. 2016). The first of these components,
Monte Carlo tree search, has been a mainstay of automated Go playing systems
since the 1990s (Brugmann, 1993). The latter three components are implemented
* https://github.com/yenchenlin/DeepLearningFlappyBird
An Overview of Deep Learning in Industry ◾ 85
as CNN models with slightly different objectives. This makes AlphaGo an interest-
ing mix of traditional and cutting edge AI techniques.
AlphaZero, which was developed following on from the work on AlphaGo,
leverages RL so that it does not need any human telling the agent what the best
movements are or which are the best strategies. Through RL, the agent is able to
learn by itself on the basis of trial and error. Playing against itself over many mil-
lions of games, the agent is able to learn which are the best moves in many different
games, including chess and the game of Go.
5.2.3.3 Robotics
While autonomous vehicles can be thought of as robots, there is a much broader set
of robotic systems that also implement deep learning approaches to control their
actions. We are primarily interested in systems that use deep learning for end-to-
end robot control rather than systems that use deep learning components for spe-
cific tasks such as object recognition or speed control. For example, BRETT* from
UC Berkeley can learn to perform tasks such as stacking LEGO blocks, putting
together a toy plane, and screwing bottle caps onto bottles using deep reinforce-
ment learning (Levine et al. 2015). CNN models are used to process the image
input and to convert them into motor control signals. The model is trained using a
technique similar to the DQN approach previously described (Levine et al. 2015).
Other similar systems include the Google Brain grasping robot (Levine et al. 2016)
and the systems developed by Delft Robotics that used deep learning to win the
Amazon Picking Challenge.†
Moving away from fixed-base grasping robots, Peng et al. (2016) introduced
a deep reinforcement learning approach for a robot locomotion control policy in
a physical simulation. The objective in this simulation setting was to navigate a
simulated robot dog through different types of terrain obstacles, such as gaps, walls,
and slopes. This control task shares similar challenges to those faced in the Atari
* https://news.berkeley.edu/2015/05/21/deep-learning-robot-masters-skills-via-trial-and-error/
† http://amazonpickingchallenge.org/
86 ◾ Data Analytics and AI
2600 games previously described. The terrain descriptions, as well as the robot
state descriptions, are high dimensional, therefore they are not suitable to be used
directly for traditional reinforcement learning systems (Figure 5.7).
In this case, a mixture of actor-critic experts (MACE) approach is introduced to
learn the right output, the parameterized actions, for complex input data (e.g., leaps
or steps) (Peng et al. 2016). The actor-critic approach in RL consists of creating two
models: the first takes the actions and the latter evaluates how good the action was,
so the agent can learn. The MACE system is a type of CNN in which the inputs
are terrain features and the robot state features, and the outputs are the actions and
estimated rewards.
As we already said for self-driving cars, it is unlikely that in the short term, end-
to-end deep learning systems will be used for the control of autonomous robots that
interact closely with humans. Rather, it is more likely that deep learning models
will be used as components in these systems for specific tasks such as object rec-
ognition or text detection. For robotic applications that do not involve significant
interactions with people, however, end-to-end control systems based on deep rein-
forcement learning are feasible. Another interesting potential for the application to
robotics is using simulations to train models that are deployed in real robots (Rusu
et al. 2016). This overcomes the complication of performing hundreds of thousands
of experiment iterations in a real environment.
5.2.3.4 Energy Consumption
One of the applications that called the attention of many companies has been the
reduction by 40% of the cooling bill of Google data centers.* These data centers,
which can be seen in Figure 5.9, are in charge of storing all the information col-
lected in Google applications such as emails in Gmail, photos in Google Maps, or
Figure 5.7 Deep reinforcement learning techniques have been successfully used
in robotic arms (left) and in drones (right).
* https://deepmind.com/blog/article/deepmind-ai-reduces-google-data-centre-cooling-bill-40
An Overview of Deep Learning in Industry ◾ 87
the documents in Google Drive. This information is stored in large rooms with
around 2.5 million servers, according to the company Gartner Inc. To prevent
these servers from overheating, it is necessary to cool them down every so often
using a cooling system usually based on pumps, chillers, and cooling towers.
Developing an algorithm to optimize energy consumption to cool down the
servers is very complex because each data center has its own characteristics. These
include factors such as the local climate, the demand for computing power, and the
cost of energy. However, any improvement in the reduction of energy consumption
can result in a large economic saving and a potential reduction in energy generation
derived carbon emissions.
Google decided to address the problem of maximizing energy consumption
by implementing an algorithm based on deep reinforcement learning. The objec-
tive of the algorithm is to maximize the power usage effectiveness (PUE), which
is a metric obtained by dividing the total energy consumed by the data center by
the energy consumed to run the computer infrastructure. Google does not usually
give information about the techniques used to achieve their achievements, and the
case of maximizing the PUE was not an exception. However, Google itself pub-
lished information showing that its algorithm was trained with historical data from
thousands of sensors that captured information from several factors (temperature,
consumption, pump speed, etc.) to create a system able to automatically adapt to
dynamic environments by using deep neural networks (Figure 5.8).
5.2.3.5 Online Advertising
Bidding optimally in internet advertising is a very complex task where RL has
been successfully applied (Du et al. 2017). In the paper presented by Han Cai
et al. (2017), a case is presented of success where an agent is able to bid intel-
ligently for each impression in the real-time bidding (RTB) model. RTB is like
a massive worldwide auction where publishers offer visits (from users who access
their pages) to advertisers so that they can display their ads. Generally, in online
Figure 5.8 An approach based on reinforcement learning has reduced the elec-
tricity bill of Google Data Centers by 40%.
88 ◾ Data Analytics and AI
5.2.4 Forecasting
In this section, we introduce the application of deep learning to forecasting
future values of a time series. We distinguish between three main categories of
application—forecasting physical signals, forecasting financial data, and forecast-
ing wind speed and power.
that the model is capable of mapping solar irradiance, which is usually highly
non-linear and time changeable, as it combines advantages of recurrent and wave-
let neural networks (Figure 5.9).
Another application is in wind forecasting in relation to wind energy produc-
tion. Wind prediction is complex due to the wind’s high degree of volatility and
deviation. Therefore, in real electricity markets, system operators have barely begun
to factor wind forecast information into their daily operations and reserve deter-
mination. However, in terms of academic research, many publications have intro-
duced short-term or long-term wind forecasting technologies and experience based
on deep learning approaches. In Kariniotakis et al. (1996), a recurrent higher-order
neural network (RHONN) model was developed for wind power forecasting in a
wind park. This model can be used to predict wind speed or power in time scales
from some seconds to three hours. The work of More and Deo (2003) employs the
technique of neural networks (feed-forward and recurrent networks) and statistical
time series respectively to forecast daily and monthly wind speeds in India. The
results show that the neural networks perform better than the baseline ARIMA
models. However, the average of daily and monthly wind speed could be smoother
than that of hourly wind speed, which implies that it is not difficult to obtain a
more accurate forecasting result for daily and monthly wind speed. In Rohrig and
Lange (2006), a method based on artificial neural networks was utilized to predict
the average hourly wind speed. A multilayer perceptron neural network with three-
layer feed-forward architecture was adopted as their forecasting system. The input
selection was determined on the basis of correlation coefficients between previous
wind speed observations.
Other authors of deep learning applications to physical system forecasting
include: Romeu et al. (2013) who applied stacked denoising autoencoders (Vincent
et al. 2010) (a specific type of feed-forward network) to the task of indoor tem-
perature forecasting and James et al. (2017) who applied deep learning models to
forecast wave conditions. These techniques are also being applied in industry. For
Figure 5.9 Deep learning has been widely applied in both domains, wind speed
estimation (left) and forecasting financial data (right).
90 ◾ Data Analytics and AI
example, Nervana* leveraged deep learning for extreme weather forecasting and
climate change modeling and Descartes Labs† applied deep learning over satellite
images for predicting crop production.
* www.nervanasys.com/
† www.descarteslabs.com/
An Overview of Deep Learning in Industry ◾ 91
5.3 Conclusion
In this chapter, we have reviewed a broad range of different applications of deep
learning, categorized into four general types of task: recognition, generation, deci-
sion-making, and forecasting, with a particular focus on relevance to industry.
As we have shown, deep learning has produced impressive achievements both in
terms of improvements in accuracy, compared with traditional machine learning
approaches, as well as enabling completely new AI applications. We are already
benefiting from many of these in our everyday lives, and as applications of deep
learning continue to improve and expand, we can expect to experience many more
benefits in the future.
However, recent applications of deep learning in industry have also raised con-
cerns such as hidden biases in training data, adversarial manipulation of trained
models, and the difficulty in understanding the rationale behind decisions made by
* www.sentient.ai
† www.bloomberg.com/news/articles/2017-02-06/silicon-valley-hedge-fund-takes-on-wall-
street-with-ai-trader
‡ www.man.com/
§ www.bloomberg.com/news/features/2017-09-27/the-massive-hedge-fund-betting-on-ai
¶ www.eurekahedge.com/Research/News/1639/Quantitative-Hedge-Funds-Strategy-Profile
** www.numer.ai
†† w ww.wired.com/2017/02/ai-hedge-fund-created-new-currency-make-wall-street-work-like-
open-source/
92 ◾ Data Analytics and AI
deep learning models due to their black box nature. Further research is needed to
make deep learning applications safer and more trustworthy in society.
References
D. Bahdanau, K. Cho and Y. Bengio. “Neural machine translation by jointly learning to
align and translate”. arXiv preprint arXiv:1409.0473 (2014).
T. Baltrušaitis, C. Ahuja and L.-P. Morency. “Multimodal machine learning: A survey and
taxonomy”. In: IEEE, 2018.
W. Bao, J. Yue and Y. Rao. “A deep learning framework for financial time series using
stacked autoencoders and long-short term memory”. PloS one 12.7 (2017), e0180944.
J. Beirne, J. Hunter and M. Simpson. “Is the real exchange rate stationary? – A similar sized
test approach for the univariate and panel cases”. In: Economics and Finance, Dept
of Economics and Finance Research Papers, Brunel University (2007). url: http://bura.
brunel.ac.uk/ handle/2438/1024.
B. Brugmann. Monte Carlo go. Tech. rep. Technical report, Physics Department, Syracuse
University, Syracuse, NY, 1993.
H. Cai et al. “Real-time bidding by reinforcement learning in display advertising”. In:
Proceedings of the Tenth ACM International Conference on Web Search and Data
Mining. ACM, 2017, pp. 661–670.
J. Cao and X. Lin. “Application of the diagonal recurrent wavelet neural network to
solar irradiation forecast assisted with fuzzy technique”. Engineering Applications of
Artificial Intelligence 21.8 (2008), pp. 1255–1263.
J. Chao, F. Shen and J. Zhao. “Forecasting exchange rate with deep belief networks”. In: The
2011 International Joint Conference on Neural Networks. July 2011, pp. 1259–1266.
doi: 10.1109/IJCNN. 2011.6033368.
T. Chouard. “The Go Files: AI computer wraps up 4-1 victory against human champion”.
Nature News (2016).
Dan C. Cireşan et al. “Mitosis detection in breast cancer histology images with deep neural
networks”. In: D. Image Computing and Computer-Assisted Intervention – MICCAI
2013: 16th International Conference, Nagoya, Japan, September 22–26, 2013,
Proceedings, Part II, Springer, Berlin Heidelberg, 2013, pp. 411–418. isbn: 978-3-
642-40763-5. doi: 10.1007/978-3642-40763-5 51.
R. Collobert et al. “Natural language processing (almost) from scratch”. Journal of Machine
Learning Research 12(Aug 2011), pp. 2493–2537.
J. Devlin et al. “Bert: Pre-training of deep bidirectional transformers for language under-
standing”. arXiv preprint arXiv:1810.04805 (2018).
X. Ding et al. “Deep learning for event-driven stock prediction”. In: Proceedings of the 24th
International Conference on Artificial Intelligence. IJCAI’15. AAAI Press, Buenos
Aires, Argentina, 2015, pp. 2327–2333. isbn: 978-1-57735-738-4. url: http://dl.acm.
org/citation.cfm?id=2832415.2832572.
X. Ding et al. “Knowledge-driven event embedding for stock prediction”. In: COLING
2016, 26th International Conference on Computational Linguistics, Proceedings of the
Conference: Technical Papers, December 11-16, 2016, Osaka, Japan. 2016, pp. 2133–
2142. url: http://aclweb.org/anthology/ C/C16/C16-1201.pdf.
An Overview of Deep Learning in Industry ◾ 93
C. Dong et al. “Learning a deep convolutional network for image super-resolution”. In:
Computer Vision – ECCV 2014: 13th European Conference, Zurich, Switzerland,
September 6–12, 2014, Proceedings, Part IV. Ed. by David Fleet et al. Springer
International Publishing, Cham, 2014, pp. 184–199. isbn: 978-3-319-10593-2. doi:
10.1007/978-3-319-10593-2 13.
M. Du et al. “Improving real-time bidding using a constrained markov decision process”.
In: International Conference on Advanced Data Mining and Applications. Springer,
2017, pp. 711–726.
A. Esteva et al. “A guide to deep learning in healthcare”. Nature Medicine 25.1 (2019),
pp. 24–29.
E. F. Fama. “The behavior of stock-market prices”. The Journal of Business 38.1 (1965),
pp. 34–105. issn: 00219398, 15375374. url: http://www.jstor.org/stable/2350752.
T. Fischer and C. Krauss. Deep Learning with Long Short-Term Memory Networks for
Financial Market Predictions. Tech. rep. FAU Discussion Papers in Economics, 2017.
J. Flynn et al. “DeepStereo: Learning to predict new views from the world’s imagery”. ArXiv
abs/1506.06825 (2015). url: http://arxiv.org/abs/1506.06825.
S. Galeshchuk and S. Mukherjee. “Deep learning for predictions in emerging currency
markets.” ICAART 2. 2017a, pp. 681–686.
S. Galeshchuk and S. Mukherjee. “Deep networks for predicting direction of change in
foreign exchange rates”. Intelligent Systems in Accounting, Finance and Management
24.4, (2017b), pp. 100–110.
L. A. Gatys, A. S. Ecker and M. Bethge. “A neural algorithm of artistic style”. ArXiv e-prints
(Aug. 2015). arXiv: 1508.06576 cs.CV.
R. B. Girshick. “Fast R-CNN”. arXiv abs/1504.08083 (2015). url: http://arxiv.org/abs/1504.
08083.
R. Girshick et al. “Rich feature hierarchies for accurate object detection and semantic
segmentation”. In: Proceedings of the 2014 IEEE Conference on Computer Vision and
Pattern Recognition. CVPR ’14. IEEE Computer Society, Washington, DC, 2014,
pp. 580–587. isbn: 978-1-4799-51185. doi: 10.1109/CVPR.2014.81.
I. Goodfellow, Y. Bengio and A. Courville. Deep Learning. MIT Press, 2016. http://www.
deeplearningbook. org.
I. Goodfellow et al. “Generative adversarial nets”. In: Advances in Neural Information
Processing Systems 27. Ed. by Z. Ghahramani et al. Curran Associates, Inc., 2014,
pp. 2672–2680. url: http: //papers.nips.cc/paper/5423-generative-adversarial-nets.pdf.
S. Ha et al. “MarioNETte: Few-shot face reenactment preserving identity of unseen tar-
gets”. arXiv e-prints (Nov. 2019), arXiv:1911.08139. arXiv: 1911.08139 cs.CV.
G. Hadjeres and F. Pachet. “DeepBach: A steerable model for bach chorales generation”.
ArXiv e-prints (Dec. 2016). arXiv: 1612.01010 cs.AI.
K. He et al. “Delving deep into rectifiers: Surpassing human-level performance on ima-
genet classification”. In: The IEEE International Conference on Computer Vision
(ICCV). Dec. 2015.
N. Hegde et al. “Similar image search for histopathology: SMILY”. NPJ Digital Medicine
2.1 (2019), p. 56.
S. Hochreiter and J. Schmidhuber. “Long Short-Term Memory”. Neural Computation 9.8
(Nov. 1997), pp. 1735–1780. issn: 0899-7667. doi: 10.1162/neco.1997.9.8.1735. url:
http://dx.doi.org/10. 1162/neco.1997.9.8.1735.
94 ◾ Data Analytics and AI
G. B. Huang E. Learned-Miller. Labeled Faces in the Wild: Updates and New Reporting
Procedures. Tech. rep. UM-CS-2014-003. University of Massachusetts, Amherst,
May 2014.
M. Hussain. “Security in connected cars”. In: Proceedings of the European Automotive
Congress EAEC-ESFA 2015. Springer International Publishing, Cham, 2016,
pp. 267–275. isbn: 978-3-31927276-4. doi: 10.1007/978-3-319-27276-4 24.
F. N. Iandola et al. “SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and
¡1MB model size”. arXiv abs/1602.07360 (2016). url: http://arxiv.org/abs/1602.07360.
P. Isola et al. “Image-to-image translation with conditional adversarial networks”. ArXiv
e-prints (Nov. 2016). arXiv: 1611.07004 cs.CV.
H. Jaeger. “The ‘‘echo state’’ approach to analysing and training recurrent neural net-
works with an Erratum note”. German National Research Center for Information
Technology (2001). url: http://www.faculty.jacobs-university.de/hjaeger/pubs/
EchoStatesTechRep.pdf.
S. C. James, Yushan Zhang and Fearghal O’Donncha. “A machine learning framework to
forecast wave conditions”. arXiv preprint arXiv:1709.08725 (2017).
Y. Jia et al. “Direct speech-to-speech translation with a sequence-to-sequence model”. CoRR
abs/1904.06037 (2019). arXiv: 1904.06037. url: http://arxiv.org/abs/1904.06037.
G. N. Kariniotakis, G. S. Stavrakakis and E. F. Nogaret. “Wind power forecasting using
advanced neural networks models”. IEEE transactions on Energy conversion 11.4
(1996), pp. 762–767.
A. Karpathy and F-F Li. “Deep visual-semantic alignments for generating image descrip-
tions”. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR
2015, Boston, MA, USA, June 7–12, 2015, 2015, pp. 3128–3137. doi: 10.1109/
CVPR.2015.7298932.
T. Kmet and M. Kmetova. “A 24H forecast of solar irradiance using echo state neural net-
works”. In: Proceedings of the 16th International Conference on Engineering Applications
of Neural Networks (INNS). EANN ’15. ACM, Rhodes, Island, Greece, 2015, 6:1–
6:5. isbn: 978-1-4503-3580-5. doi: 10.1145/2797143.2797166.
P. Korshunov and S. Marcel. “DeepFakes: A new threat to face recognition? Assessment
and detection”. CoRR abs/1812.08685 (2018). arXiv: 1812.08685. url: http://arxiv.
org/abs/1812. 08685.
W. Kristjanpoller, A. Fadic and M. C. Minutolo. “Volatility forecast using hybrid neural
network models”. Expert Systems with Appllication 41.5 (Apr. 2014), pp. 2437–2442.
issn: 0957-4174. doi: 10. 1016/j.eswa.2013.09.043.
Y. LeCun et al. “Generalization and network design strategies”. Connectionism in Perspective
(1989), pp. 143–155.
C. Ledig et al. “Photo-realistic single image super-resolution using a generative adversarial
network”. arXiv abs/1609.04802 (2016). url: http://arxiv.org/abs/1609.04802.
S. Levine, N. Wagener and P. Abbeel. “Learning contact-rich manipulation skills with
guided policy search”. In: IEEE International Conference on Robotics and Automation,
ICRA 2015, Seattle, WA, USA, 26–30 May, 2015. 2015, pp. 156–163. doi: 10.1109/
ICRA.2015.7138994.
S. Levine et al. “End-to-end training of deep visuomotor policies”. ArXiv abs/1504.00702
(2015). url: http://arxiv.org/abs/1504.00702.
S. Levine et al. “Learning hand-eye coordination for robotic grasping with deep learning
and large-scale data collection”. ArXiv abs/1603.02199 (2016). url: http://arxiv.org/
abs/1603.02199.
An Overview of Deep Learning in Industry ◾ 95
I. Sutskever, O. Vinyals and Q. V. Le. “Sequence to sequence learning with neural net-
works”. ArXiv e-prints (Sept. 2014). arXiv: 1409.3215 cs.CL.
Y. Taigman et al. “DeepFace: Closing the gap to human-level performance in face verifica-
tion”. In: Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern
Recognition. CVPR ’14. IEEE Computer Society, Washington, DC, 2014, pp. 1701–
1708. isbn: 978-1-4799-5118-5. doi: 10.1109/CVPR.2014.220.
A. Tampuu et al. “Multiagent cooperation and competition with deep reinforcement learn-
ing”. ArXiv abs/1511.08779 (2015). url: http://arxiv.org/abs/1511.08779.
M. Valdenegro-Toro. “Objectness scoring and detection proposals in forward-looking sonar
images with convolutional neural networks”. In: Artificial Neural Networks in Pattern
Recognition: 7th IAPR TC3 Workshop, ANNPR 2016, Ulm, Germany, September
28–30, 2016, Proceedings. Ed. by Friedhelm Schwenker et al. Springer International
Publishing, Cham, 2016, pp. 209–219. isbn: 978-3319-46182-3. doi: 10.1007/978-
3-319-46182-3 18.
A. van den Oord et al. “WaveNet: A generative model for raw audio”. ArXiv eprints
abs/1609.03499 (2016). url: http://arxiv.org/abs/1609.03499.
H. Van Hasselt, A. Guez and D. Silver. “Deep reinforcement learning with double qlearn-
ing”. In: Thirtieth AAAI Conference on Artificial Intelligence, 2016.
P. Vincent et al. “Stacked denoising autoencoders: Learning useful representations in a
deep network with a local denoising criterion”. Journal of Machine Learning. Research
11 (Dec. 2010), pp. 3371–3408. issn: 1532-4435. url: http://dl.acm.org/citation.
cfm?id=1756006.1953039.
O. Vinyals et al. “Show and tell: Lessons learned from the 2015 mscoco image captioning
challenge”. In: IEEE Transactions on Pattern Analysis and Machine Intelligence, 2016.
M. Wang and W. Deng. “Deep face recognition: A survey”. arXiv preprint arXiv:1804.06655
(2018).
Y. Wu et al. “Google’s neural machine translation system: Bridging the gap between
human and machine translation”. ArXiv abs/1609.08144 (2016). url: http://arxiv.org/
abs/1609.08144.
W. Xiong et al. “Achieving human parity in conversational speech recognition”. ArXiv
e-prints (Oct. 2016). arXiv: 1610.05256 cs.CL.
T. Young et al. “Recent trends in deep learning based natural language processing”. IEEE
Computational intelligenCe Magazine 13.3 (2018), pp. 55–75.
R. L. Zeff and B. Aronson. Advertising on the Internet. John Wiley & Sons, Inc., 1999.
L. Zhang, S. Wang and B. Liu. “Deep learning for sentiment analysis: A survey”. Wiley
Interdisciplinary Reviews: Data Mining and Knowledge Discovery 8.4 (2018), e1253.
J-Y Zhu et al. “Generative visual manipulation on the natural image manifold”. In: Computer
Vision – ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October
11–14, 2016, Proceedings, Part V. Ed. by Bastian Leibe et al. Springer International
Publishing, Cham, 2016, pp. 597–613. isbn: 978-3-319-46454-1. doi: 10.1007/978-
3-319-46454-1 36.
Chapter 6
Contents
6.1 The Chinese Perspective on Innovation and AI.........................................101
6.2 AI with Chinese Characteristics................................................................103
6.3 National Security in AI.............................................................................104
6.4 “Security” or “Protection”.........................................................................106
6.5 A(Eye).......................................................................................................109
6.6 Conclusions...............................................................................................112
Bibliography.......................................................................................................113
* A great many thanks to Jared Adams, Executive at the Association of Public Sector Executives
(Canada) for his wit, keen eye, and thoughtful support in the drafting of this chapter.
99
100 ◾ Data Analytics and AI
Once viewed as a niche endeavor of digital specialists and researchers, the twin
issues of data production and analysis have been vaulted to the forefront of policy
in the emerging global “artificial intelligence (AI) race.” With the growing signifi-
cance of data to realms of economic, social, and political endeavor, there has been
a commensurate growth in both political and popular attention to data and its
downstream applications. Of all those involved, it may well have been Vladimir
Putin who did the most to push the idea of a global AI race forward when he stated
in 2017 that “whoever leads in (artificial intelligence) will rule the world.”*
While many have pointed out that alluding to an “AI race”—drawing com-
parison with the “space race” of the last century—may serve to overstretch the
analogy,† it is nonetheless remarkable to note how comparative standing in AI has
become such a coveted commodity among geopolitical powers. Unlike the space
race of the 1950s and 1960s, however, it is not Russia, but China who has done
the most to draw the world’s attention to state-backed leadership in AI through its
challenge to American primacy. China’s strategic pursuit of leadership in data and
AI may lack the rhetorical flourish of a Vladimir Putin, but it lacks none of the
audacity. In 2017, China announced a bold “New Generation Artificial Intelligence
Development Plan,” which has as its goal positioning the country as the indisput-
able global leader in artificial intelligence by 2030.‡
Certainly, political leaders are accustomed to proclaiming ambitious targets,
especially when long-term time horizons provide a degree of insulation from the
potential for failure. However, this type of announcement takes on a special sig-
nificance in a place like China where close adherence to publicly shared plans are
held as sacrosanct by the political leadership. As an authoritarian country where
continued political rule depends on the regime’s ongoing credibility in delivering
programs and services, failure to adhere to such a plan has the potential to strike at
the core of the government’s legitimacy and hold on power. That is to say that such
an announcement is not one that is made lightly, and should be taken by observers
as a credible statement of China’s ambition and self-assessed capacity.§ China’s con-
fidence in its growing AI capacity stems from some fairly linear observations about
fields of endeavor which contribute to success in the AI sector.
While historically China has had a weak environment for innovation due to a
lack of venture capital, few startups, and a weak tertiary education system, it has
been systematically addressing these weaknesses for years. Huge government ini-
tiatives to build the university system and attract global talent have had a marked
impact on the performance of this sector.¶ China has also had a great deal of success
in supporting a tech startup culture, including through the founding of innovation
* Meyer, 2017.
† Biddle, 2019.
‡ Notice of the New Generation Artificial Intelligence Development Plan (trans.), 2017.
§ Lee, 2018.
¶ Robbins, 2016.
Chinese AI Policy and the Path to Global Leadership ◾ 101
parks and business incubators. Although at the time of writing there appears to be
growing softness in the Chinese venture capital market, there has otherwise been
significant year-over-year improvement in access to venture capital for the better
part of a decade.* Improvements in all of these sectors contribute to the country’s
strength in AI specifically (and the technology sector more broadly) and work to
bolster AI-specific programs that have been launched to support China’s AI leader-
ship drive.
This is to say that much of China’s leadership declaration for AI is underpinned
by improvements in the weakest parts of China’s innovation system, in areas that
China should have the undebatable capacity to improve. Its global leadership bid is
not based on a recent program or policy drive that is specific to AI and viewed to be
a silver bullet of sorts. Rather, much of the Chinese story in AI is that of a country
which has been frantically pursuing its own development for decades and has con-
cluded more recently that sustained growth will only be supported by an economy
that is innovative. Investing in the core drivers of innovation will certainly sup-
port China’s AI sector along with many others, but having meaningful capacity
in AI is viewed by the central government as much more than simply a symptom
of wider progress. China views AI as an area worthy of its own dedicated policy
approach, much of which is outlined specifically in the “New Generation Artificial
Intelligence Development Plan” (AIDP).
the work of Kissinger (2011). A selected works of Xi Jinping was also released in 2014 on the
subject of governance in China which, although somewhat one-sided, provides interesting
insights into the Chinese perspective (full references in the Bibliography).
102 ◾ Data Analytics and AI
Canada’s justice system.” The National Post. May 28, 2019. https://nationalpost.com/news/politics/
canada-calls-out-on-china-at-wto-council-meeting-for-evidence-to-back-canola-ban
‡ Robbins, 2019.
§ Schell and Delury, 2013.
¶ Zenglin and Holzmann, 2019.
* Allen, 2019.
† Notice of the New Generation Artificial Development Intelligence Plan, (trans.) 2017.
‡ Notice of the New Generation Artificial Development Intelligence Plan, (trans.) 2017.
104 ◾ Data Analytics and AI
statements take on a different significance and merit close attention to the details
which indicate an important shift in perspective. With the impact of AI being more
broadly felt and widespread than in the past, the AIDP effectively represents a gov-
ernment recognition that AI will have everyday impacts for the Chinese citizenry, a
realization that reflects a sort of crossing of the Rubicon in Chinese AI governance.
The emerging everyday impact of AI very dramatically increases the importance
of the field to the state and in turn, denotes an expansion of the state’s role in the
governance of AI.
This crucial recognition that AI requires a more comprehensive and active
approach from the government signals an end to the status quo which had relegated
AI to more or less a niche research area under the purview of siloed experts. The
AIDP is, in this sense, explicit in its view of AI as being key to Chinese economic
prosperity, which, for a government with a long history of economic intervention-
ism—even at times when laissez-faire economic management has been in vogue—
suggests more active state involvement in AI R&D and industrial policy. While the
AIDP makes the perspective of there being an explicit connection between AI and
overall economic prosperity clear enough, what is perhaps less immediately obvious
is the emphasis on the connection between AI and national security.
6.3 National Security in AI
The AIDP clearly and repeatedly establishes an explicit connection between AI for
economic development and AI for national security.* This can be confusing to the
casual observer. Certainly, there are relevant military applications for AI, as there
are relevant applications for AI in most realms of activity, but penetration of AI into
military applications is relatively low.† Various forms of data science have featured
in the operations of the military for some time, and more sophisticated uses of AI
have been key to complex defense platforms as well, such as drones, aircraft, mis-
siles, and the like. Some research suggests that the ability to leverage AI in defense
applications might provide players like China an asymmetric advantage in capa-
bilities, which could provide a partial explanation for the emphasis placed on the
security implications of AI.‡
Certainly, the impact of technological progress in AI on military affairs can-
not be left unstated yet it would be shortsighted to interpret the explicit mention
* Notice of the New Generation Artificial Development Intelligence Plan (trans.), 2017.
† By contrast, in the early days of the nuclear age, nearly all activity in nuclear energy was being
conducted by the military in one part of the supply chain or another, if for no other reason
than that the security imperative crowded-out private R&D. The same cannot be said for
AI, where the vast majority of economic activity occurs in the civilian private sector, making
AI applications that need to be reconverted to be compatible with military uses (Artificial
Intelligence and National Security, 2019).
‡ Artificial Intelligence and National Security, 2019.
Chinese AI Policy and the Path to Global Leadership ◾ 105
of national security as only referring to the military.* While the use of analytics
and AI in military systems will continue to be important, the progress being made
in these use cases do not encompass the full sea change being described in the
AIDP. Many observers have noted that the most significant security concerns for
the Chinese political leadership and its hold on power may not be from the exter-
nal geopolitical threats for which a military prepares so much as the government’s
ability to retain meaningful political control over the Chinese populous. Indeed,
China has had an overriding preoccupation with its territorial integrity and inter-
nal security for centuries.†
Given China’s longstanding preoccupation with internal stability, it is of lit-
tle surprise to note that widely spread advancements in AI and proliferation of
advanced technologies among the Chinese citizenry are viewed with a degree of
caution. In this sense, the stated recognition in the AIDP of AI as having broad-
based impacts returns as an important one. One reading of the AIDP’s emphasis on
national security is to note the potential of AI’s proliferation among the citizenry
to represent a security concern in the eyes of the Chinese leadership. Indeed, much
of the Chinese’s leadership’s decision-making related to digital technologies has
been formed by its experience of the Arab Spring in the Middle East and Color
Revolutions in the post-Soviet sphere, whereby the proliferation of new technolo-
gies (namely social media) were viewed as the cause of instability, and in some cases
even revolution, in these countries.
China’s political leadership has been inclined to view these recent waves of
political instability as not simply as grassroots uprisings emboldened by open access
to digital communications, but as disturbances that were deliberately fostered by
rival powers using technology companies as surrogates to advance their geopolitical
interests.‡ This again harkens back to the core principle of the Chinese perspec-
tive on governance which views political and economic activities as being closely
related, if not inseparable. On that basis, China has demonstrated reluctance, and
even hostility at times, to Western-originated technology companies that are trying
to enter the Chinese market. This reluctance is especially pronounced when these
companies have eschewed Chinese political censorship by refusing to adjust their
treatment of topics like “human rights” and “free speech.”§
China is understandably cautious (if perhaps somewhat paranoid) when it
comes to the ability of new technologies like AI to foment domestic unrest, and
this tinges its relationship with AI. Complicating matters further is that the line
between world-class technology companies and their products, like “AI,” is blurred
* There are several excellent works on this subject beginning to emerge, and while an important
issue, the military application of AI in Chinese falls outside of the scope of this article. For
more information, a valuable resource and good starting point can be found in Cory (2019).
† Kissinger, 2011.
‡ Chen, 2010.
§ Tan and Tan, 2012.
106 ◾ Data Analytics and AI
and perhaps superficial in many cases, much as are political and economic issues
blurred in the Chinese perspective on governance. As technological leadership in
the 21st century overlaps greatly with proficiency in AI, so too does being a tech-
nology company with an economic footprint in China entail increased political
scrutiny from the government of China. In this sense, the security references of the
AIDP can be viewed as stating a concern that exists with both the technology itself
and the large (and often foreign) firms which employ it.
This can be viewed as a clear continuation and reiteration of Chinese policy
toward foreign technology firms and their influence in China, which has been
described in terms ranging in tone from measured hesitancy about foreign tech-
nology firms to outright persecution of them. As a result of the Chinese govern-
ment’s perspective on AI governance and technology governance more broadly, the
relationship between the Chinese government and foreign technology companies
has often been a troubled one. China has blocked the entrance of many foreign
technology firms, such as Google, Facebook, and Twitter, and produced fatal bar-
riers to success for others, such as Uber and Amazon. At the same time, China
has promoted and nourished domestically grown technology mega-firms, such as
Baidu, Tencent, and Alibaba, which now hold primacy in the Chinese market and
compete in others.*
There are cogent arguments to the effect that China applies a double standard
that allows native Chinese technology companies to flourish while preventing mul-
tinational firms from competing in the Chinese market. There is certainly some
measure of truth to this perceived approach of economic nationalism, but it is
far from the whole truth. A more nuanced interpretation notes that the Chinese
approach to technology companies demands of them a level of coziness and integra-
tion with the Chinese government that proves unacceptable to many multinational
firms. Those technology firms that are able to successfully operate in China are
thus doing so in a way which is compatible with the very involved approach sought
by the Chinese government, one which often requires compromises on issues of
liberty, free speech, and human rights for the sake of security—as defined by the
government of China.
6.4 “Security” or “Protection”
The relationship between free competition in AI industries and the restrictions
placed on competition due to Chinese security concerns is a complicated one.
Especially so when considering the clash in viewpoints on political economy
between China and liberal-democratic countries. In this light, there are a num-
ber of different ways to interpret the explicit statement of the AIDP of connec-
tions between national security, AI, and economics. Certainly, it is possible to take
at face value the idea that AI represents a security concern to the government of
China, but the very invocation of “security” as an issue raises wider implications for
trade and innovation policy.
The reference to security can be understood, for one, as a reiteration of the
Chinese government’s commitment to economic protectionism.* Rather than
being a serious attempt to address security concerns, the “security” references to
AI could be interpreted as a way for the government to ensure preferential treat-
ment of Chinese firms, who, given the contours of the Chinese political system,
are more receptive to accommodating the demands for censorship and control that
may come from the Chinese central government. Firms associated with “security”
are given certain privileges in terms of how they report taxes which provides them
more insulation from scrutiny. Again, there is certainly some truth to this, as mul-
tinational firms’ non-compliance with the government of China’s restrictions have
been used as justifications for their persecution and ejection from the country.
Foreign firms may also be uncomfortable with having such a close relation-
ship with the Chinese government as a basis for operating there because of what it
defines as security issues, namely issues of free speech, repression, and censorship.
On that basis, foreign firms that are willing to plug their noses on security issues
to entertain collaboration with the Chinese government face pressure from their
home country to forego such relationships, as Google has over its plans to launch a
censored search engine for the Chinese market.† All of this helps to ensure that the
Chinese domestic market continues to be captured almost exclusively by Chinese
firms
Announcing connections between AI and security also offers additional protec-
tions to Chinese AI firms seeking a strong footing to export to foreign markets. A
government that gives preferential treatment to domestic firms, whether overtly,
covertly, or through systematic advantages in policy treatment, risks legitimate
sanction of itself or of those firms in trade negotiations and international economic
forums. This is something which could very well present a risk to China. However,
as a result of historical circumstances which embed the primacy of state sovereignty
in international law, issues deemed by a country to be of importance to “security”
are given special treatment in disputes about free trade.
A government declaration that a certain industry constitutes a matter of national
security provides a country immunity from World Trade Organization (WTO)
challenges that the industry is receiving unfair support from the state or undue
protection from foreign competition.‡ These kinds of national security exemptions
have been legitimately used to insulate military industries from the pressures of the
global marketplace which might otherwise see important industrial components of
the defense sector move offshore. Such a move offshore could constitute a very real
* Cory, 2019.
† Bastone, June 16th 2019.
‡ Trebilcock, 2011.
108 ◾ Data Analytics and AI
threat to national security in the event of a military conflict, and so offering some
legalized protectionism in this space helps to ensure the overarching viability of the
global trade regime.
It should be clearly stated that this line of argument has yet to be employed
in a major trade dispute regarding AI, and while it would be highly abnormal, we
live in interesting times. Recent years have seen some very creative interpretations
of the national security provision in trade disputes. Perhaps most notable was the
invocation of security privileges by the United States, which used it in an attempt to
reshore* steel and aluminum production during the United States, Mexico, Canada
Agreement (USMCA) negotiations in 2018.† With AI having been declared a secu-
rity issue in China, ostensibly any Chinese company with a claim to be involved in
“AI”—a definition which casts a very wide net as a result of AI’s horizontality across
industry sectors—could also be afforded protections from international trade law
disputes.
This is the theory, at least; it would be quite another thing to see how far this
line of argument can be successfully applied in practice. While the security jus-
tification for discriminating against foreign steel and aluminum production was
widely viewed as a cynical abuse of the security exemption, were China to employ
this same security provision for AI industries writ large, it would be a significant
escalation in gaming the rules. With the obvious importance of AI to the eco-
nomic activities of the 21st century, invoking a point of privilege in such a manner
might understandably raise some skepticism about how genuine the connection is
between “national security” and “AI.” It’s worth noting that while emerging tech-
nologies in the past have a significant overlap with military applications, such as
nuclear energy or even aircraft writ large, this same degree of overlap does not exist
for AI in the military—at least, not yet.‡
There has nonetheless historically been a close connection between national
security issues and cutting-edge emerging technologies. While this is often attrib-
uted to the strategic need for the military to have technological supremacy, the truth
of the matter is much more complicated, as there is seldom a clear-cut case where
national security was invoked in a sector and the motivation for protectionism and
* What is often highlighted in this story is how Washington was attempting to reshore these
industries to the United States from its allies, Canada and Mexico. While accurate, what can
be lost in the details is that this was a threat to Canada and Mexico to persuade those countries
to be more restrictive on Chinese steel imports. In some ways it can be better understood as
part of an ongoing China–US trade war.
† Globerman, July 18, 2018.
‡ By contrast, in the early days of the nuclear age, nearly all activity in nuclear energy was being
conducted by the military in one part of the supply chain or another, if for no other reason
than that the security imperative crowded-out private R&D. The same cannot be said for
AI, where the vast majority of economic activity occurs in the civilian private sector, making
AI applications that need to be reconverted to be compatible with military uses (Artificial
Intelligence and National Security, 2019).
Chinese AI Policy and the Path to Global Leadership ◾ 109
state supports was entirely absent, especially when considering the lack of trans-
parency that exists around defense spending practices.* Certainly, emerging tech-
nologies are often of military significance, but as they also represent a significant
economic interest which may bear much less relation to the security imperative.
While the trend is undeniable, the parsing out of economic value from political and
security related considerations is a much more challenging task.
6.5 A(Eye)
Trade law and economic policy interpretations of these security declarations aside,
it’s worth noting the operational connection that exists between the development
of domestic Chinese AI industries and the operations of the state security sector,
especially with regard to surveillance technologies. China has long employed an
extensive state surveillance system, and it has been very proactive in its incorpo-
ration of technological solutions to this system. China is today the world’s most
significant user and producer of public closed circuit television (CCTV) cameras.†
China at some point began to approach the human limitations of its domestic sur-
veillance system which contained 176 million cameras in 2017, a number projected
to increase to 626 million cameras by 2020.‡
A clear impediment to these systems is the staggering resource limitations to
conduct successful human-only monitoring of the CCTV feeds. Given the band-
width issue for human-centric observation, there was a clear incentive to support
the development of AI technologies which could automate much of the process.
This very quickly spilled over into AI software (computer vision) being applied in
cameras, making China a global leader in AI-based surveillance today.§ It really is
no coincidence that the country with the highest number of CCTV cameras by
very far, also holds a leading edge in AI technologies related to surveillance, namely
facial recognition. In fact, the largest acquisition of an AI company in history was
the purchase of SenseTime, a Chinese AI that got its start in facial recognition from
working with the Chinese government on Made in China 2025.¶
This lead in AI security surveillance has been networked with other localization
technologies, financial records, and other pieces of information, such as purchasing
* Robbins, 2015.
† There is an interesting conversation to be had about the proliferation of CCTV cameras and
what it may signify. China certainly dwarves all other countries in terms of the sheer number
of cameras, but given its population this can be misleading. By cameras per person, and when
focusing at the city level, many Chinese cities continue to lead but are in good company with
cities like London, UK and Atlanta, USA. That is to say that there is some room for nuanced
interpretation (Bishoff, 2019).
‡ Bishoff, 2019.
§ Perryer, October 29, 2019.
¶ Marr, June 17, 2019.
110 ◾ Data Analytics and AI
habits, into what is called the “Social Credit System.” This system aggregates infor-
mation about an individual and uses it to assign a moving “social credit score”
which increases as a result of socially desirable behaviors and decreases as a result
of “undesirable behaviour.”* The potential for political censorship in such a system
is very significant, and it has already been employed with the greatest enthusiasm
in the politically restless regions of Tibet and Xinjiang. As such, there are few
AI-based projects of the scope, magnitude, and potential for controversy as the
Social Credit System.
The Social Credit System is an initiative of the Chinese government that uses
this centralized system as a tool for accomplishing its broader objectives, and the
government makes no apologies for doing so. Aside from the political instrumenta-
tion for security purposes, it raises objections for clearly violating the norms held in
many countries surrounding individual privacy by widely and indiscriminately col-
lecting and analyzing data, including violating the principle of informed consent
for data collection and use, limitations on storage and collection, and the right to a
private sphere of life more generally. While China offers few natural privacy protec-
tions and seems to have light social expectations of privacy, the thought of such an
intrusive system is jarring to many elsewhere. At a more visceral level, it raises fears
of state control and social engineering, with many observers describing the system
as “dystopian” or “Orwellian.”†
The system however appears to be quite popular in China itself where it is
viewed as a way for re-establishing the potential for trust in strangers against the
backdrop of rapid urbanization and creeping social anomie.‡ Indeed, the system is
designed not simply to collect information for the state but makes each individual’s
social credit score openly and transparently available to anyone else who might
want to access it. While it is possible to imagine nefarious and questionable uses,
some of the common uses of the system are fairly humdrum, such as a reference
point for vetting potential dating partners.§ Chinese government officials them-
selves are often quick to point out that the country has plenty of methods for social
control without the Social Credit System, which they insist represents a different
endeavor all together.¶
It may be tempting to pass judgement on the ethical or moral worth of such
a system, but that is an exercise that will be left for another author. Instead, it is
important here to note that the Social Credit System and use of AI in security is
closely related to Chinese innovation strategy, governance customs, and many of
* There is some nuance to this that is often missed, namely that a universal score does not yet
exist in a functional capacity, as much as there are several “scores” being collected and used for
different purposes, although policy documents refer to making greater progress in amalgamat-
ing this into a single score.
† Campbell, November 21, 2019.
‡ Minter, January 24, 2019.
§ Robbins, 2019.
¶ Creemers, 2018.
Chinese AI Policy and the Path to Global Leadership ◾ 111
China’s core competitive advantages in AI, while also representing a divisive wedge
issue that has deep seated cultural roots and sensitives elsewhere. Observers have
noted that these specific conditions in China give AI researchers a major advantage
since successful R&D and testing of AI is fed by access to huge amounts of data.
Having the world’s largest population, few privacy restrictions, and a supportive
government certainly lends to China’s competitive advantage as a jurisdiction for
AI industries.*
Other leading jurisdictions for AI, such as the United States, are simply unable
to offer the same type of beneficial arrangement to AI firms at a structural level;
both governance norms and social expectations of privacy forbid it. This is not to
say that China has an incontestable advantage in the AI space—that is certainly not
the case—however, some of China’s most clear and decisive advantages as a juris-
diction are difficult for other political-economic systems to counter or match. In
this light, the invocation of “security” in the AIDP and related documents may not
only be technically accurate, but also a necessary measure for trying to avoid the
penalization of Chinese AI industries in a manner that would ebb away at a pur-
portedly unnatural competitive advantage. Indeed, this is far from a hypothetical
concern, as critics and human rights activists have lambasted China and Chinese
companies selling surveillance and AI-based facial recognition technology abroad
for “exporting authoritarianism.”†
Many jurisdictions openly discuss banning facial recognition technology and
related surveillance for posing a threat to liberty,‡ privacy, or other closely held
values, and some governments are actively pursuing a ban at present.§ Others
still, however, suggested that the Chinese approach to AI, particularly those enti-
ties which have had a close connection to state surveillance and the Social Credit
System, crosses moral and ethical boundaries and needs to be systematically
eschewed. The United States government, for one, has imposed stark restrictions
on a list of Chinese AI startups (including SenseTime) that have a relationship with
the security sector for their involvement in the repression of minorities in Tibet and
Xinjiang.¶ Without passing judgement on the issue of whether or not a ban on these
technologies should occur, these discussions do represent a move to impose limita-
tions on China’s AI specialty and a growing area of national competitive advantage.
* Lee, 2018.
† Yang and Murgia, December 9, 2019.
‡ Samuel, November 15, 2019.
§ Lam, December 3, 2019.
¶ Shepardson and Horwitz, October 7, 2019.
112 ◾ Data Analytics and AI
6.6 Conclusions
The government of China’s strategic objectives, governance conventions, and the
cultural norms of the Chinese people closely impact the contours of the economic
and technological leadership that China demonstrates in AI today. The free-for-all
in data collection in the Chinese “Wild, Wild East” of AI development is only
possible due to the country’s much more permissive norms surrounding privacy*
combined with the Chinese model of governance which permits more integrated
collaboration between state institutions and private industry. As China establishes
AI as a national priority and stakes important political capital on global leadership
in AI, it is clear that the issue will continue to be taken very seriously.
The security sector in China proves to be a central hub of activity in AI for a
variety of reasons, and likely will continue to be. Some of this is deeply rooted in
Chinese governance practices, such as the perceived need to keep close watch on
foreign technology firms or the standing concern about the security risk posed by
China’s own citizenry. Other reasons are much more pragmatic, such as the ability
to mount a legal defense in the event of the sanction of industries of national inter-
est or the opportunity to support innovation in AI in a way that was convenient
given other security objectives. This has allowed not only for the rapid development
of hard surveillance infrastructure and data collection, but also for these to be inte-
grated with parallel efforts to support AI development in that space.
This strategy has given China a vast gap in leadership in AI applications related
to security specifically and related areas like facial recognition more generally. The
Chinese approach to AI and security has been a major force in China’s overall level
of progress with AI and is certainly helping the country’s pursuit of global leader-
ship. Kai Fu Lee has famously suggested that this makes the United States and
China (as the two leading jurisdictions for AI) natural partners that can find syn-
ergies in one another’s structural strengths and weakness.† This represents a fairly
optimistic view, and one which seems increasingly unlikely given the times. The
relationship between AI and security is proving not only to be a core part of China’s
AI strategy, but also one which is deeply divisive in liberal-democratic countries
like the United States.
* It’s worth pointing out that the permissiveness of Chinese privacy law is sometimes contested
as the country has released some policies that appear to suggest even higher personal privacy
standards than employed by the EU’s General Data Protection Regulation (GDPR) which is
often viewed as a protective standard. On close examination however, many of these policies
seem to be superficial, lacking enforcement mechanisms or much by way of serious commit-
ment. For more see: Sacks, January 29, 2018.
† Lee, 2018.
Chinese AI Policy and the Path to Global Leadership ◾ 113
Bibliography
Allen, Gregory C. Understanding China’s AI Strategy: Clues to Chinese Strategic Thinking
on Artificial Intelligence and National Security. Washington, DC: Centre for a New
American Security. February 6, 2019.
Bastone, Nick. “Google Executive Finally Confirms That the Tech Giant Is No Longer
Working on a Censored Internet Search Engine for China.” Business Insider. June 16,
2019. https://ww w.businessinsider.com/google-kills-censored-search-engine-for-chi
na-2019-7.
Biddle, Sam. “An ‘AI Race’ Between the US and China Is a Terrible, Terrible Idea.” The
Intercept. July 21, 2019. https://theintercept.com/2019/07/21/ai-race-china-artifi cial-
intelligence/.
Bishoff, Paul. “The World’s Most Surveilled Cities.” Comparitec. August 15, 2019. https://
www.comparitech.com/vpn-privacy/the-worlds-most-surveilled-cities/.
Blanchfield, Mike. “Chrystia Freeland Was Urged to Tell China to Stop the ‘Inaccurate’
Talk about Canada’s Justice System.” The National Post. May 28, 2019. https://na
tionalpost.com/news/politics/ca nada-calls-out-on-china-at-wto-council-meeting-for-
evidence-to-back-ca nola-ban.
Campbell, Charlie. “‘The Entire System Is Designed to Suppress Us.’ What the Chinese
Surveillance State Means for the Rest of the World.” Time. November 21, 2019. https://
time.com/5735411/china-surveillance-privacy-issues/.
Castro, Daniel, McLaughlin, Michael and Chivot, Eline. Who Is Winning the AI Race:
China, the EU or the United States? Washington, DC: The Centre for Data Innovation.
August 19 2019.
Chen, Titus. “China’s Reaction to the Colour Revolutions: Adaptive Authoritarianism in
Full Swing.” Asian Perspective 34(2) 5–51. 2010.
Cory, Nigel. The Ten Worst Protectionism and Mercantilist Innovation Policies of 2019.
Washington, DC: The Information Technology and Innovation Foundation. 2019.
Creemers, Rogier. China’s Social Credit Score System: An Evolving Practice of Control. Leiden:
Leiden University. May 9, 2018. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=
3175792.
Elstrom, Peter. “China’s Venture Capital Boom Shows Signs of Turning into a Bust”
Bloomberg. July 9, 2019. https://ww w.bloomberg.com/news/articles/2019-07- 09/chin
a-s-venture-capital-boom-shows-signs-of-turning-into-a-bust.
Ferracane, Martina and Lee-Makiyama, Hosuk. China’s Technology Protectionism and It’s
Non-Negotiable Rationales. Brussels: European Centre for International Political
Economy. 2017.
Fu, Xiaolan, Woo, Wing Thye and Hou, Jun. “Technological Innovation Policy in China:
The Lessons and the Necessary Changes Ahead.” Economic Change and Restructuring
49(2–3) 139–157. 2016.
Globerman, Steven. Trump Administration Summons ‘National Security’ to Justify Tariffs.
Vancouver: The Fraser Institute. July 18, 2018.
Jinping, Xi. Governance in China. Beijing: Foreign Languages Press. 2014.
Kissinger, Henry. On China. New York: Penguin Books. 2011.
Lam, Kristin. “Portland, the Largest City in Oregon, Plans to Propose First Facial
Recognition Ban Affecting Private Companies.” USA Today. December 3, 2019.
https://ww w.usatoday.com/story/tech/2019/12/03/facial-recognition-portland-orego
n-ban/2601966001/.
114 ◾ Data Analytics and AI
Lee, Kai-Fu. AI Superpowers: China, Silicon Valley and the New World Order. Boston:
Houghton Mifflin Harcourt. 2018.
Marr, Bernard. “Meet the World’s Most Valuable Startup: China’s SenseTime.” Forbes.
June 17, 2019. https://www.forbes.com/sites/bernardmarr/2019/06/17/meet-the-worl
ds-most-valuable-ai-startup-chinas-sensetime/#341320c0309f.
Meyer, David. “Vladimir Putin Says Whoever Leads in Artificial Intelligence Will Rule
the World.” Fortune. September 4, 2017. http://fortune.com/2017/09/04/a i-artificial-
intelligence-putin-rule-world/.
Minter, Adam. “Why Big Brother doesn’t bother most Chinese.” Bloomberg. January 24,
2019. https://ww w.bloomberg.com/opinion/articles/2019- 01-24/why-china-s-socia l-
credit-systems-are-surprisingly-popular.
N.s. Artificial Intelligence and National Security, Congressional Research Service. November
21, 2019. https://fas.org/sgp/crs/natsec/R45178.pdf.
N.s. Notice of the New Generation Artificial Intelligence Development Plan. (trans.). State
Council of the People’s Republic of China. July 8, 2017. http://www.gov.cn/zhengce/
content/2017- 07/20/content_5211996.htm.
(PDF) Here Be Dragons: Lessons from China’s AI Strategy. https://www.researchgate.net/
publication/330937827_ Here_ Be_ Dragons_ Lessons_f rom_China.s_ AI_ Strategy
[Accessed December 12, 2019].
Perryer, Sophie. “Surveillance Cameras Have Become One of China’s Most Valuable
Exports- Heres Why.” World Finance. October 29, 2019. https://www.worldfina nce.
com/f e atur ed/su r veil lance - came ras-h ave-b ecome - one- of- ch i nas- most- valua ble- e
xports-heres-why.
Robbins, Mark. Here Be Dragons: Lessons from China’s AI Strategy. Ottawa: The Institute
on Governance. 2019.
Robbins, Mark. The Thousand Talents Program: Lesson From China About Faculty Recruitment
and Retention. Ottawa: The Conference Board of Canada. 2016.
Robbins, Mark. Towards a Political Economy of Military Spending. Dissertation. Carleton
University. 2015.
Sacks, Samm. New China Data Privacy Standard Looks More Far Reaching than GDPR.
Washington, DC: Center for Strategic and International Studies. January 29, 2018.
https://www.csis.org/analysis/new-china-data-privac y-sta ndard-look s-more-far-rea
ching-gdpr.
Samuel, Sigal. “Activists want Congress to ban facial recognition. So they scanned lawmak-
ers’ faces.” Vox. November 15, 2019. https://www.vox.com/f uture-perfect/2019/11/15/
20965325/facial-recognition-ba n-congress-activism.
Schell, Orville and Delury, John. Wealth and Power: China’s Long March to the Twenty-First
Century. New York: Random House. 2013.
Shepardson, David and Horwitz, Josh. “U.S. Expands Blacklist to Include China’s Top AI
Startups Ahead of Trade Talks.” Reuters. October 7, 2019. https://www.reuters.com/a
rticle/us-usa-trade-china-exclusive/u-s-expands-blacklist-to-include-chinas-top-a i-st
artups-ahead-of-trade-talk s-idUSKBN1WM25M.
Tan, Justin and Tan, Anna E. “Business Under Threat, Technology Under Attack, Ethics
Under Fire: The Experience of Google in China.” Journal of Business Ethics 110(4)
469–479. 2012.
Trebilcock, Michael. Understanding Trade Law. Cheltenham: Edward Elgar Publications.
2011.
Chinese AI Policy and the Path to Global Leadership ◾ 115
Yang, Yuan and Murgia, Madhumita. “How China Cornered the Facial Recognition
Market.” The Los Angeles Times. December 9, 2019. https://ww w.latimes.com/business/
story/2019-12- 09/china-facia l-recognition-surveillance.
Zenglin, Max and Holzmann, Anna. Evolving Made in China 2025: China’s Industrial
Policy in the Quest for Global Leadership. Washington, DC: Mercator Institute for
China Studies. July 2, 2019.
Chapter 7
Natural Language
Processing in Data
Analytics
Yudong Liu
Contents
7.1 Background and Introduction: Era of Big Data......................................... 118
7.1.1 Use Cases of Unstructured Data.................................................... 118
7.1.2 The Challenge of Unstructured Data............................................. 119
7.1.3 Big Data and Artificial Intelligence................................................120
7.2 Data Analytics and AI...............................................................................121
7.2.1 Data Analytics: Descriptive vs. Predictive vs. Prescriptive..............121
7.2.2 Advanced Analytics toward Machine Learning and Artificial
Intelligence....................................................................................122
7.2.2.1 Machine Learning Approaches........................................122
7.3 Natural Language Processing in Data Analytics.......................................124
7.3.1 Introduction to Natural Language Processing...............................124
7.3.2 Sentiment Analysis........................................................................125
7.3.3 Information Extraction..................................................................127
7.3.4 Other NLP Applications in Data Analytics...................................128
7.3.5 NLP Text Preprocessing................................................................129
7.3.6 Basic NLP Text Enrichment Techniques.......................................130
7.4 Summary..................................................................................................130
References..........................................................................................................131
117
118 ◾ Data Analytics and AI
* https://www.searchtechnologies.com/blog/big-data-use-cases-for-business
† https://ww w.brig htloc al.com/research/loca l- consumer-review-su rvey/
Natural Language Processing in Data Analytics ◾ 119
* https://ww w.statista.com/statistics/278890/us-apparel-and-accessories-retail-e-commerce-re
venue/
120 ◾ Data Analytics and AI
Optimization and decision modeling technologies are used to solve complex deci-
sions with millions of decision variables, constraints, and tradeoffs.*
As the starting step in data analytics, descriptive analytics has been the major
form in traditional business intelligence. However, with the availability of big vol-
ume data and advanced analysis techniques, more effort is going toward predictive
and prescriptive analytics.
Classification has been one of the most common machine learning tasks. A
supervised machine learning algorithm is “trained” using the training data
set, which means a statistical model is built through modeling the “correla-
tion” among the label and features. The model will be evaluated with a “test”
data set. A good test data set should contain a sufficient number of “unseen”
examples (data points that did not occur in the training data set) to avoid an
“overfitting” issue. This way can make the model built more generalizable.
Some of the most widely used supervised learning methods include lin-
ear regression, decision trees, random forests, naive Bayes, logistic regression,
and support vector machines (SVM), and have broadly applied to a variety
of business applications, including fraud detection, recommendation system,
risk analysis, and sentiment analysis.
◾◾ Unsupervised Learning. Unsupervised learning applies when not enough
labeled data is available. In other words, no well-understood data set is
available or labeling a new data set is just too costly. By applying some well-
designed distance metrics, an unsupervised learning algorithm can strategi-
cally group the data into clusters. Data points within the distance threshold
will be grouped together. This process iterates until the number of clusters
converges or no more change takes place to each cluster. Some of the widely
used clustering algorithms include k-means and k-nearest neighbors (KNN).
◾◾ Reinforcement Learning. Unlike supervised learning where the labeled data
is given to “teach” or “train” the system to learn, in reinforcement learning,
learning takes place as a result of an interaction between an agent and the
environment through trial and error. This learning process fundamentally
mimics how humans (and animals) learn. As humans, we perform actions
and observe the results of these actions on the environment. The idea is com-
monly known as “cause and effect” which can be translated into the following
steps in reinforcement learning:*
1) The agent observes a state in the environment.
2) An action is selected using a decision-making function (policy).
3) The action is performed.
4) The agent receives a scalar reward (reinforcement) or penalty from the
environment.
5) The policy is updated/fine-tuned for that particular state accordingly
(learning step).
6) The above steps are iterated until an optimal policy is found.
Reinforcement learning has been widely applied in robotics and game play-
ing. It is also used for complex tasks such as autonomous vehicles.
◾◾ Neural Networks and Deep Learning. Neural networks and deep learning cur-
rently provide the cutting-edge solutions to many problems in traditional
AI fields such image recognition and speech recognition. Neural networks
* https://www.cse.unsw.edu.au/~cs9417ml/RL1/introduction.html
124 ◾ Data Analytics and AI
on translating 60 Russian sentences into English in 1954. With the idea and inter-
est of AI emerging, the NLP work from the late 1960s to late 1970s moved into
a new phase of more sophistication, with more emphasis on the representation
and manipulation of world knowledge and the construction of a knowledge base.
Stimulated by the development of computational grammar theory in linguistics and
the use of logic in knowledge representation and reasoning in AI, NLP moved to
be more practically motivated from failing to build practical systems in the 1980s.
In the 1990s, statistical NLP (Manning & Schütze, 1999) started flourishing due
to the arrival of a large amount of data on the internet and the computing power
of handling it. Corpus data, together with machine learning, has greatly driven the
development of statistical parsing, which is one of the core NLP tasks, whereby
NLP has made significant progress during this period in other subtasks such as
machine translation and information extraction. Presently, NLP research and
development have entered another new and exciting era due to the availability of
rich computational resources, vast quantities of data, rapid development of machine
learning techniques and tools, and the emergence of many new application oppor-
tunities and challenges. From IBM’s “Watson”* (2006) to Apple’s “Siri”† (2011),
from Amazon’s “Alexa”‡ (2014) to Google’s “Google Assistant”§ (2016), powered
by Natural Language Processing and other technologies, AI has gradually become
part of our daily life and will continue to transform our lives in every conceivable
field. The applications of NLP today are incredibly diverse. It spreads in various
fields such as machine translation (e.g., Google Translate), question answering (e.g.,
Apple Siri), information extraction, natural language generation (document sum-
marization, chatbot), writing assistance, video scripting, text categorization, senti-
ment analysis, speech technologies (speech recognition and synthesis), hate speech
detection, fake news detection, etc. In the following, we will introduce some NLP
applications in data analytics.
7.3.2 Sentiment Analysis
Sentiment analysis (SA, also named opinion mining, opinion extraction, subjectiv-
ity analysis) focuses on the extraction of people’s sentiments, opinions, emotions,
and attitudes toward entities such as services, products, organizations, individu-
als, issues, events, topics, and their attributes (Liu, 2015). With the rapid growth
of social media data on the web, the amount of opinionated data in digital form
has become unprecedentedly available. Since early 2000, sentiment analysis has
grown to be one of the most active research areas in NLP. Due to its important role
in decision-making, its applications have been widely spread across fields such as
* https://en.wikipedia.org/wiki/Watson _(computer)
† https://en.wikipedia.org/wiki/Siri
‡ https://en.wikipedia.org/w iki/A mazon_ Alex a
§ https://en.wikipedia.org/w iki/Google _ Assistant
126 ◾ Data Analytics and AI
marketing, finance, political science and health science. Many big corporations as
well as small startups are dedicated to providing sentiment analysis services.
Sentiment analysis can be performed at three levels: document-level, sentence-
level, and aspect-level. Document-level SA assumes each document expresses either
a positive or a negative opinion on a single aspect and aims to determine the overall
opinion of the document. Sentence-level SA aims to extract sentiment expressed
in each sentence. There is no fundamental difference between document and sen-
tence-level SA because a sentence can be viewed as a short document. Aspect-level
SA* aims to answer the question of “What aspects of the entity do people love/
hate.” This requires extracting aspects (attributes) of an object and allocating each
aspect a sentiment (positive or negative). This task is particularly useful for a cus-
tomer review report generation.
Document- and sentence-level SA is usually formulated into a binary classifica-
tion task. Supervised and unsupervised techniques are applied. Common super-
vised learning methods include SVM, naive Bayes, maximum entropy, and so on.
And the common features that are used include n-grams (n = 1, 2, 3) and their
frequencies, part-of-speech tags, opinion words (or sentiment words) and phrases,
negotiations, and syntactic dependencies. Opinion words/phrases are the words/
phrases that carry commendatory or derogatory senses themselves. For instance,
good, fantastic, amazing, and excellent are commendatory terms and bad, slow, flimsy,
worst, and poor are derogatory words. Although most opinion words are adjec-
tive or adverbs, nouns (“rubbish”) and verbs (“hate,” “like”) can also often indicate
opinion. Negations are undoubtedly important in SA because the sentiment of a
sentence can be completely reversed due to their presence. Syntactic dependency
features are useful when multiple opinion words exist in text thus the interactions
among these words need to be taken into account. Unsupervised methods use the
(co)occurrences of positive and negative words in the sentence/document to calcu-
late their weighted sum/average, respectively. The one with the higher score will be
the final sentiment of the text.
Sentiment lexicon plays an important role in SA, so the construction of senti-
ment lexicon has also been studied extensively in NLP. In general, there are three
methods: manual construction, corpus-based methods, and dictionary-based
methods. Manual construction is time-consuming and costly which is rarely used
or used only when combined with other methods. Both corpus-based methods
and dictionary-based method use an initial seed set of opinion words. Corpus-
based methods use co-occurrence metrics to iteratively gather more opinion words
and dictionary-based methods construct the lexicon by propagating the seed words
through the WordNet† synonym/antonym graph until a threshold is reached.
* http://alt.qcri.org/semeval2016/task5/
† https://wordnet.princeton.edu
Natural Language Processing in Data Analytics ◾ 127
7.3.3 Information Extraction
IE is the process of extracting useful information from unstructured or semi-struc-
tured text. The extracted information is either readily usable or requires some addi-
tional processing and analyses. Although there is no strict rule on what information
an IE system should extract, there are three types of information that are com-
monly extracted by most IE systems. They are named entities, relations, and events.
* https://www.ldc.upenn.edu/collaborations/past-projects/ace
128 ◾ Data Analytics and AI
language and are more interactive due to the application of NLP techniques.
Compared with a traditional rule-based chatbot, an NLP-based chatbot can
continue to learn from every interaction and automatically evolves to improve
the quality of the support they offer in the future. A social chatbot such as
Microsoft Xiaoice (Zhou L. et al., 2018) involves far more sophisticated and
advanced NLP techniques, among many other techniques, and represents the
highest-level achievement of a chatbot in this generation.
* https://bitbucket.org/kganes2/text-mining-resources/downloads/A%20Brief %20Note%20on
%20Stop%20Words%20for%20Text%20Mining%20and%20Retrieval.pdf
130 ◾ Data Analytics and AI
7.4 Summary
As described in this chapter, due to the availability of big data, computational
resources, and the advancement of machine learning techniques, there are many
remarkable uses of NLP in data analytics today. As NLP continues to make data
* https://en.wikipedia.org/wiki/Edit_distance
† https://ai.googleblog.com/2016/05/an nouncing-s yntax net-worlds-most.html
Natural Language Processing in Data Analytics ◾ 131
more “user-friendly” and “insightful,” it will be more and more widely adopted in
all types of data analytics platforms. In spite of its wide application, NLP is still
in its infancy compared with people’s expectations for AI. Languages are complex,
subtle, and ambiguous. Processing natural language itself is an extremely chal-
lenging task. No matter it is a low-level processing step such as tokenization or a
high-level task such as machine translation, the existing NLP systems are still far
from perfect. Though there is still a long way to go, NLP is rapidly advancing today
along with deep learning techniques and revolutionizing the way people interact
with computers and data. Looking to the future, it is clear that NLP-empowered
applications will become even more capable, and the improvement will continue to
take place. As NLP is embracing a new renaissance that is unprecedented, NLP will
play an indispensable part in the next generation of AI-empowered applications
and NLP applications will become ubiquitous in our lives.
References
Christensen, J. et al. (2010). Semantic Role Labeling for Open Information Extraction.
Proceedings of the NAACL HLT 2010 First International Workshop on Formalisms
and Methodology for Learning by Reading, pp. 52–61.
Cotterell, R., Duh, K. (2017). Low-Resource Named Entity Recognition with Cross-Lingual,
Character-Level Neural Conditional Random Fields. Proceedings of the Eighth
International Joint Conference on Natural Language Processing (Volume 2: Short
Papers), pp. 91–96.
Eisenstein, J. (2019). Introduction to Natural Language Processing. Cambridge, MA: The
MIT Press.
Hearst, M.A. (1992). Automatic Acquisition of Hyponyms from Large Text Corpora. Pro
ceedings of the 14th Conference on Computational Linguistics, vol. 2, pp. 539–545.
Li, J. et al. (2018). A Survey on Deep Learning for Named Entity Recognition. arXiv preprint
arXiv:1812.09449, 2018.
Liu, B. (2015). Sentiment Analysis: Mining Opinions, Sentiments, and Emotions. Cambridge,
UK: Cambridge University Press.
Manning, C.D., Schütze, H. (1999). Foundations of Statistical Natural LanguageProcessing.
Cambridge, MA: The MIT Press.
Miwa, M., Sasaki, Y. (2014). Modeling Joint Entity and Relation Extraction with Table
Representation. Proceedings of the 2014 Conference on Empirical Methods in Natural
Language Processing, pp. 1858–1869. Doha, Qatar: Association for Computational
Linguistics.
Yang, B., Mitchell, T. (2016). Joint Extraction of Events and Entities within a Document
Context. Proceedings of the 2016 Conference of the North American Chapter of
the Association for Computational Linguistics: Human Language Technologies,
pp. 289–299.
Zhou, Li. et al. (2018). The Design and Implementation of Xiaoice, an Empathetic Social
Chatbot. arXiv preprint arXiv:1812.08989, 2018.
Chapter 8
AI in Smart Cities
Development: A
Perspective of Strategic
Risk Management
Eduardo Rodriguez and John S. Edwards
Contents
8.1 Introduction..............................................................................................134
8.2 Concepts and Definitions..........................................................................134
8.2.1 How Are AI, Smart Cities, and Strategic Risk Connected?............137
8.3 Methodology and Approach......................................................................137
8.4 Examples of Creating KPIs and KRIs Based on Open Data......................141
8.4.1 Stakeholder Perspective..................................................................142
8.4.2 Financial Resources Management Perspective............................... 145
8.4.3 Internal Process Perspective...........................................................146
8.4.4 Trained Public Servant Perspective................................................146
8.5 Discussion.................................................................................................147
8.6 Conclusion................................................................................................148
References..........................................................................................................148
133
134 ◾ Data Analytics and AI
8.1 Introduction
This chapter is about the creation of artificial intelligence (AI) systems to support
the management of strategic risk in smart cities. The use of AI in smart cities is
growing and the applications to several sectors are booming: from analyzing data
from sensors in bridges for the maintenance process to health care at home or secu-
rity and safety of the citizenry.
Some events affecting the performance of cities can be predicted and controlled.
However, there exist events that we will call “Black Swans” after Taleb (2007) that
cannot be accurately predicted or controlled. For predictable events, societies can take
proactive measures to mitigate and control their risks. The prediction of possible results
allows the realization of the range of random variables, identification of groups by
impact levels/areas in the city, resources used, and time of occurrence. It also enables the
administration of the city to proactively intervene to improve the indicators they would
like to manage. For non-predictable events, the actions must instead revolve around
preparedness to react when the event appears. Nevertheless, possible analytics-based
observations and risk management attributes can provide clues to help be prepared.
In this chapter, we present examples of various events affecting a city’s perfor-
mance and show how, with AI use, it is possible to provide a better understanding
of the impact of decisions in managing strategic risk. In a city’s performance, the
allocation of budget is related to protection, hedging, mitigation, and control of
risks related to crime, accidents, healthcare, education, and so on. Proactive risk
control helps balance the costs of pre-event, event, and after-event actions, thereby
reducing them. Natural disasters such as flooding, hurricanes, and other Black
Swan events often have the highest impact in societies, but preparedness can reduce
the severity of the event’s impact.
The analysis of the performance of a city can follow several types of framework.
In this chapter, the organization of the performance measurement and strategic
objectives set is based on an adaptation of the balanced scorecard (BSC) to munici-
pal administrations. The factors under a city’s control are similar to those of for-
profit organizations but with a different main purpose, the benefit/satisfaction of
citizens rather than financial performance. We will use the stakeholders’ quality of
life as the driver to align financial, public servant, and services perspectives. The
next section presents a brief explanation of the terms that we use in this chapter,
and then we describe how to use analytics tools to develop an AI system that can
dynamically and continuously guide the city administration. Finally, we offer some
examples by way of illustration, but we can only hint at the possibilities in the space
available here.
systems do things that would require intelligence if done by humans” [our change
in italics: the original said “men”]. This is developed in the context of smart cities
by Agarwal et al. (2015) providing us with the aspects to consider in AI applica-
tions: “The main objective of the AI system is to create intelligent machines and
through this, to understand the principles of intelligence.” In the context of this
chapter, we consider AI application as part of the process of creating knowledge
from data and developing the systems to maintain a continuous analysis of data
that can drive actions in the solution, prevention, and orientation of problems in
a city.
The concept of a smart city is evolving, as the use of technology emerges to
connect citizens and their city’s administration. Musa (2016) provides a detailed
definition:
A smart city is defined as a city that engages its citizens and connects
its infrastructure electronically. A smart city has the ability to integrate
multiple technological solutions, in a secure fashion, to manage the
city’s assets—the city’s assets include, but not limited to, local depart-
ments’ information systems, schools, libraries, transportation sys-
tems, hospitals, power plants, law enforcement, and other community
services.
The technologies that might provide these solutions are described in a report from the
International Standards Organization (ISO/IEC JTC 1 Information Technology,
2015): “Smart Cities is a new concept and a new model, which applies the new
generation of information technologies, such as the internet of things, cloud com-
puting, big data and space/geographical information integration, to facilitate the
planning, construction, management and smart services of cities.” Furthermore,
the SmartCitiesWorld/Phillips Lighting survey (Simpson, 2017) identifies the
three most important requirements as open/sharing data, citizen engagement, and
enhanced services for citizens. Sodhi, Flatt, and Landry (2018) observe that “The
path to creating smart cities should be through a culture of openness, co-creation,
and shared resources.”
As Frigo and Anderson (2011) explained, “Strategic risks [our italics] are those
risks that are most consequential to the organization’s ability to execute its strate-
gies and achieve its business objectives.” There are many variables that might be
used as KPIs (key performance indicators) and KRIs (key risk indicators) for a
smart city. Albino, Beardi, and Dangelico (2015) identify aspects related to the
environment, finance, people, administration, living, and so on. Whatever factors
a city chooses to consider become strategic in the sense of defining the path for the
city administration to follow. The aggregated risk related to those factors is then
the strategic risk.
The measurement process is crucial for managing the city and keeping stakehold-
ers satisfied. Measurement generates metrics and indicators. These are potentially
136 ◾ Data Analytics and AI
affected by uncertainty, and the variation of the expected results is what strategic
risk management is looking to assess, mitigate, and control.
Many types of risk can affect citizens’ lives. There are financial risks associated with
the city’s capacity to cover its obligations through resources such as taxes. Moody’s
(2019) indicated risk mitigation actions and allocation of the budget resources that
smart cities are taking: credit risk control is one of the areas of concern in a city’s risk
management. Other types of risk are related to external factors and conditions of the
environment that can affect the performance of the city, for example:
Possible controllable risks
it and learning as we go. We use open data from three cities (Baltimore, Toronto,
and Chicago) to represent performance aspects where risk events can exist. We
illustrate potential components of an AI system by indicating possible KPIs, KRIs,
and models. There is a selection of outputs related to crime, traffic management/
accidents, budgeting, education, and customer service.
Figure 8.1 represents our proposed process of building an AI system for a smart
city. The first component of the system is the definition of KPIs and KRIs. They are
the set of goals that the AI for the smart city must help achieve.
Suppose the city intends a strategic initiative to develop programs educating
families for awareness and protection. A suitable KPI might be to increase citizen
satisfaction by 2%. A potential risk is that the community does not want to engage
with such a top-down initiative, so a KRI might be the change in the number of
people participating compared with local crime rates. A risk-mitigating strategic
response would be to develop local leaders and support the community’s activities.
Once the KPIs and KRIs are selected, the next step is to gather data for the
analytics of the specific target. The source of data is expected to be the open data
that the smart city is building. Data quality is crucial; open data must be seen as
more than just generating raw data for citizens’ consultation. Table 8.1 indicates
general KPIs and KRIs in the domains of the examples in this chapter. They are
divided into the four perspectives of the modified BSC: Stakeholder (the driver),
Financial Resources Management, Internal Process, and Trained Public Servants.
The logic is that influential systems to offer good results to citizens are based on
trained people dealing with municipal administration, processes in place that are
effective and efficient, and the use of financial resources according to the defined
Figure 8.1 The process to design an AI System for strategic risk management.
Table 8.1 Examples of KPIs and KRIs According to BSC Perspectives
Perspective
vs. actions in Traffic management/
smart cities Crime control accidents Customer service requests Education
Stakeholder perspective
KPI Time to respond to Ratio pedestrian Ratio calls/emergency level Ratio expenses/
inquiries volume/accidents graduated students
KRI Variability of time for Time to repair traffic Changes in level of cases Changes in the number
attention signs solved of students graduating:
churn analysis
Variation of appropriate Location: intersections Time for closing cases with Changes in expenditure
resources with high accident solutions by enrolment
probability
KRI Level of effectiveness Changes in trend in Goal achievement variation Changes in number and
accidents of the solutions amounts of benefits
Level of satisfaction Level of areas with high Cases closed with a good Compensation
accident rate evaluation per dollar education providers/
invested quality level
(Continued)
AI in Smart Cities Development ◾ 139
Table 8.1 (Continued) Examples of KPIs and KRIs According to BSC Perspectives
Perspective
vs. actions in Traffic management/
smart cities Crime control accidents Customer service requests Education
KRI Changes in investment/ Variation in the ratio Changes in the ratio cases Change in enrolment/
140 ◾ Data Analytics and AI
strategic priorities for risk and expected performance. Figure 8.2 shows how cause
and effect flows from the fourth perspective through to the first.
The third step uses models to provide actionable knowledge in the adminis-
tration of the city. The examples provided in Section 8.4 indicate possible more
detailed KPIs and KRIs, which can be studied with statistical and machine learn-
ing algorithms. The process of selecting models that fit and provide better value to
prediction and prescription will feed the creation of rules for action, and in parallel,
there is feedback to review if the KPIs and KRIs are the ones the city administra-
tion should use in future.
The fourth step brings in risk, creating rules, parameters, and thresholds to
guide actions such as budgeting and use of resources, and providing awareness and
warning systems. AI benchmarking with other smart cities will help “tune up” the
KPIs and KRIs.
In the fifth step, the analytics results are translated into policies and actions to
manage strategic risk. Human input from management is essential at this stage. No
AI model can ever be complete.
The dynamic BSC review then represents the preparedness and orientation to
take action based on changes in KPIs and KRIs, finally reviewing the results of
those actions and also considering changing the KPIs and KRIs if appropriate.
From Table 8.1, we present examples to show the interaction between KPI, KRI,
Model, and Outcome. We give the most examples for the top-level Stakeholder
perspective.
8.4.1 Stakeholder Perspective
Crime. The analysis is based on training a model to detect law enforcement resources’
capacity to attend events and be proactive in reducing/controlling expected events, based
on patterns over time and severity measured by number of arrests. Data from Chicago.
Time of day/week, arrest or not, domestic or not, type of crime, and city area are all
factors relating to the use of law enforcement resources. Risk appears as the change
in probability of an event’s occurrence. Expected higher probabilities could lead
changes in available resources and training. Figure 8.3 shows, perhaps surprisingly,
that arrests for theft peak around 4 pm and that there is little variation between
days of the week. Figure 8.4 displays a forecast with a confidence interval, show-
ing results are not limited to a single value. A fuller analysis, including seasonality
(again we have no room for this) would suggest possible law enforcement attributes
to consider in scheduling and training resources for operational planning.
Traffic management/accidents.
Changes in accident rates because of traffic signs is just one example of how infra-
structure can affect the occurrence of events. The risk assessment is based on the
changes in the expected probabilities of occurrence of accidents. Cities’ open data
sources can include such variables as age, car type, hour, road conditions, cell phone
use, light, number of passengers, alcohol/drug use, and so on. The risk assessment
is based on the changes of the probability of accidents at the main points of control
in the city. (Data from Toronto and Chicago.)
The outcome measure is based on accident severity using categories of incapac-
ity or fatality. Analysis suggests that traffic signage is in fact only a second level
contributory factor, whereas “cell phone use other than texting” is a top-level factor,
as is “distraction from inside vehicle.” Driver (and passenger?) education may be
more important than traffic signage.
Education.
The relationship of various factors with the level of graduation of students from
public schools/colleges can be a basis for budget allocation to schools and programs
to improve education performance in the city. This analysis is related to the churn
situation. (Data from Chicago.) Figure 8.5 shows a classification tree including
the factors that most affect graduation rates. The most important factor is teacher
attendance rate. Below 95.3% has a clear negative effect on graduation, although
care needs to be taken with causality. Possibly there is another factor affecting some
schools that demoralizes both teachers and students. Again, this shows the need for
human interpretation.
Customer service requests.
311 (non-emergency)
KPI Ratio calls/service type
KRI Changes in number of cases solved
911 (emergency)
KPI Ratio calls/emergency levels
KRI Changes in number of cases solved by emergency level
This shows how staffing requirements can be managed according to service type
and factors providing indications of expected call numbers. The prediction of calls
is possible under certain conditions, and there are proactive measures that the city
can take in order to reduce some occurrences. The probability of occurrence gen-
erates the indicator to control. The risk is assessed based on the changes of the
probabilities of occurrence of different events. (Data from Baltimore.) Table 8.2
shows how the time taken to deal with the most common 311 requests varies by
AI in Smart Cities Development ◾ 145
type, from quick-to-handle information requests (what was the “odd one out”?) to
harder-to-solve water leaks.
From this perspective, the concept is how resources are used to support the plan
and modify the risk level. For example, the budget allocation (Table 8.3) should
146 ◾ Data Analytics and AI
Table 8.2 Sample of 311 Requests and Solution Times (Baltimore Data)
Number Handled that Handled Average
Service type of calls day (number) that day (%) time (days)
match the risk priorities. Many of these analyses will not be new, but their scope
may be.
This analysis helps prepare the city’s administration to maintain the risk indicators
under control at the level of financial and stakeholders’ requirements.
People, technology, and processes (Edwards, 2005) need to work together to deliver
the city’s services. Control of risk based on knowledge and development of people’s
AI in Smart Cities Development ◾ 147
Local $2,887,830,967
Grant $99,372,723
8.5 Discussion
Based on our necessarily brief review, the following are the key aspects of the com-
ponents of an AI system for a smart city and their integration:
1. Progressing from just providing data and information for citizens to the pre-
diction and review of possible actions to serve citizens better.
2. Embracing the analysis of the smart city under the strategic risk review of what
the city needs to accomplish, according to a BSC. Moving from KPIs to KRIs to
actions to implement municipal strategies, working on the bases of multiple risks,
multiple indicators, and multiple models to create components of the AI system.
3. Moving from fixed rules to dynamic rules in the development and use of
BSC, and actions to assess, monitor, and control risks. Use of a Strategic
Risk Management Dashboard—an extension of what Kourtit and Nijkamp
(2018) describe—that includes current performance and an early warning
system element.
4. Defining and providing details of each area of performance in the city using
statistical and machine learning models to incorporate the outcomes as com-
ponents of the AI system.
148 ◾ Data Analytics and AI
8.6 Conclusion
We have presented illustrative examples of the uses of AI systems to support the
management of a smart city. The design of the systems follows an iterative seven-
step approach, driven by a focus on strategic risks as represented in a balanced
scorecard.
We believe these serve as a proof of concept for using the proposed approach to
develop an integrated set of AI systems supporting all aspects of the city’s strategic
management, though this is still some time away.
References
Agarwal, P.K., Gurjar, J., Agarwal, A.K. and Birla, R. (2015) “Application of Artificial
Intelligence for Development of Intelligent Transport System in Smart Cities”,
International Journal of Transportation Engineering and Traffic Systems, 1(1).
Albino, V., Beardi, U. and Dangelico, R.M. (2015) “Smart Cities: Definitions, Performance,
and Initiatives”, Journal of Urban Technology, 22(1), pp 3–21.
Allam, Z. and Dhunny, Z.A. (2019) “On Big Data, Artificial Intelligence and Smart
Cities”, Cities, 89(June), pp 80–91.
Baloh, P., Desouza, K.C. and Hackney, R. (2012) “Contextualizing Organizational
Interventions of Knowledge Management Systems: A Design Science Perspective”,
Journal of the American Society for Information Science and Technology, 63(5),
pp 948–966.
Dasgupta, S. and Tam, E. (2005) “Indicators and Framework for Assessing Sustainable
Infrastructure”, Canadian Journal of Civil Engineering, 32(1), pp 30–44.
Eden, C. and Huxham, C. (1996) “Action Research for Management Research”, British
Journal of Management, 7(1), pp 75–86.
Edwards, J.S. (2005) “Business Processes and Knowledge Management”. In: M. Khosrow-
Pour (Ed.), Encyclopedia of Information Science and Technology (Vol. I, pp. 350-355),
Idea Group, Hershey, PA.
Frigo, M.L. and Anderson, R.J. (2011) “Strategic Risk Management: A Foundation for
Improving Enterprise Risk Management and Governance”, The Journal of Corporate
Accounting and Finance, 22(3), pp 81–88.
AI in Smart Cities Development ◾ 149
ISO/IEC JTC 1 Information Technology (2015) Smart Cities – Preliminary Report 2014,
ISO, Geneva, Switzerland.
Kaplan, R.S. and Norton, D.P. (1992) “The Balanced Scorecard—Measures That Drive
Performance”, Harvard Business Review, 70(1), pp 71–79.
Kourtit, K. and Nijkamp, P. (2018) “Big Data Dashboards as Smart Decision Support
Tools for I-Cities – An Experiment on Stockholm”, Land Use Policy, 71, pp 24–35.
Krishnamoorthy, C.S. and Rajeev, S. (1996) Artificial Intelligence and Expert Systems for
Engineers, CRC Press/Taylor & Francis, Boca Raton, FL.
McKinsey Global Institute (2018), Smart Cities: Digital Solutions for a More Livable Future,
McKinsey Global Institute, Brussels, San Francisco and Shanghai.
McNichols, T.J. (1977) Executive Policy and Strategic Planning, McGraw Hill, New York.
Moody’s (2019) Largest US Cities Take Proactive Steps to Mitigate Credit Risk from Climate
Change [Online], https://ww w.moodys.com/research/Moodys-Largest-US-cities-tak
e-proactive-steps-to-mitigate-credit--PBM_1158519.
Musa, S. (2016) “Smart City Roadmap” [Online], http://www.academia.edu/21181336/Sm
art_City_ Roadmap.
Ng, K.-S., Hung, W.-T.. and Wong, W.-G. (2002) “An Algorithm for Assessing the Risk of
Traffic Accidents”, Journal of Safety Research, 33(3), pp 387–410.
Simpson, P. (2017) “Smart Cities: Understanding the Challenges and Opportunities”
[Online] SmartCitiesWorld/Philips Lighting, https://smartcitiesworld.net/AcuCus
tom/Sitename/DA M/012/Understanding_the_Challenges_a nd_Opportunities_of_
Smart_Citi.pdf.
Sodhi, Z., Flatt, J. and Landry, J.N. (2018) Getting to the Open Smart City [Online]
Evergreen and Open North for Future Cities Canada. https://futurecitiescanada.ca/
downloads/2018/Getting _to_Open_ Smart_Cit y.pdf.
Taleb, N. (2007) The Black Swan: The Impact of the Highly Improbable, Random House,
New York.
Ullah, I., Fayaz, M. and Kim, D.-H. (2018) “Analytical Modeling for Underground Risk
Assessment in Smart Cities”, Applied Sciences, 8(6), pp 921–939.
Van Aken, J.E. and Romme, A.G.L. (2012) “A Design Science Approach to Evidence-Based
Management”. In: D.M. Rousseau (Ed.), The Oxford Handbook of Evidence-Based
Management (pp. 43–57), Oxford University Press, Oxford.
Chapter 9
Contents
9.1 Introduction.............................................................................................. 152
9.2 Electronic Dental Records and Analytics.................................................. 153
9.3 Impact of Missed Dental Appointments.................................................... 155
9.4 Patient Responses to Fear and Pain........................................................... 157
9.4.1 Dental Anxiety.............................................................................. 157
9.4.2 Dental Avoidance.......................................................................... 158
9.5 Potential Data Sources..............................................................................160
9.5.1 Dental Anxiety Assessments..........................................................160
9.5.2 Clinical Notes................................................................................ 161
9.5.3 Staff and Patient Reporting............................................................ 161
9.6 Conclusions...............................................................................................162
References..........................................................................................................163
151
152 ◾ Data Analytics and AI
9.1 Introduction
The use of information technology (IT) and electronic dental record (EDR)
systems has produced a sea change in how dentists are trained and how den-
tistry is practiced in dental schools. Since the 1980s, dental schools across the
United States have seen the incremental introduction of IT in various functional
areas. Initially, the focus of IT implementation was on dental billing systems
and insurance processing. Adoption of EDRs in dental schools has increased
as technology administrators, faculty, and clinicians realize the value of using
information technology in additional areas such as dental practice and student
progress tracking.
EDRs provide administrators, clinical faculty, and others involved in dental
education with several capabilities (Schleyer et al., 2012; Schleyer, 2004). These
systems can help link patient data with evidence-based resources to provide recom-
mendations with the help of clinical decision support systems. The patient data can
be analyzed to identify errors in medical decision-making and highlight inefficien-
cies in clinical operations thereby reducing the cost of providing care. Electronic
health data can be exchanged much more easily with other care providers compared
with paper records and can potentially reduce the need for repeated tests and dental
imaging studies such as radiographs (Stewart et al., 2010). EDR data can be reused
by researchers to identify therapies and treatments that work in real life and modify
treatments that do not work outside of lab settings (Schleyer et al., 2012). This
secondary analysis and data mining of healthcare data is one of the biggest driv-
ers for the adoption of electronic health records (EHRs) and EDRs in healthcare
organizations.
Analytics in medicine and healthcare generally falls into three main categories:
descriptive, diagnostic, and predictive (others have a fourth category dealing with
prescriptive to bring in optimization techniques). Descriptive analytics tools and
methods are used to depict an event or outcome in a way that helps document
anything that may be known about the underlying data. This type of analytics
generally tries to answer simply what is occurring. For example, EDR data can be
used to analyze patient care utilization patterns, such as preventive dental services,
across various demographic groups. Diagnostic analytics tools and methods try to
resolve why certain events occur to uncover the cause of observed outcomes. For
example, self-reported patient data could be tracked at specific intervals and ana-
lyzed to identify why certain patients are unhappy with the treatment. These and
other sources of data could be used together to implement measures that improve
overall patient engagement and satisfaction.
The greatest value of any healthcare analytics project involves trying to answer
when a future event is likely to occur using patterns in historical data. Proactively
identifying patients who have the highest risk of poor outcomes and will benefit
from clinical intervention is a prime example of the use of predictive analytics in
healthcare (Thyvalikakath et al., 2015).
Predicting Patient Missed Appointments ◾ 153
Data mining and text analytics of EDR data, combined with other sources, can
help the dental school become more proactive and anticipate outcomes based on data.
This allows faculty, students, and staff to take appropriate actions before a negative
outcome occurs. Risk scores are commonly used in dental practice. These denote the
future risk of an outcome such as periodontal disease which helps the dentist take
aggressive measures to prevent further aggravation of the disease. Data from the EDR
can be used to predict other medical issues as well, provided the dental and medical
data sets can be linked in a meaningful manner (Acharya et al., 2018).
Predictive analytics is also used to identify patients who have the highest risk
of treatment failures, such as patients who need dental implants or dental pros-
thesis (Chuang et al., 2002). It also is used to improve and enhance clinical deci-
sion support tools that personalize the treatment for each patient and recommend
treatments that can provide the greatest benefit. Furthermore, these three main
analytical approaches are used in tandem to target the same problem from different
perspectives. In a dental school setting, descriptive analytics can be used to iden-
tify current spending patterns and resource utilization. Diagnostic analytics might
identify areas where resource utilization can be optimized and make recommenda-
tions to improve the efficiency of providing dental care. Predictive analytics can be
used to predict resource requirements and allocating resources proactively.
Identifying patients who have a history of missed appointments and reliably
flagging patients with a future risk is a key use case for analytics in the dental
school environment. Using analytics to identify patients at risk of missing appoint-
ments and creating predictive models from EHR data has been attempted in other
domains (Lenzi et al., 2019). However, few studies have been done to systematically
identify and predict patients who are at risk of missing appointments in a dental
school setting (Gustafsson et al., 2010).
The goal of this chapter is to present a broad overview of the reasons for and
impact of missed dental appointments in the academic dental clinic, and how cer-
tain approaches might be used to reduce missed dental appointments. Features
of EDRs will be discussed, particularly regarding the capture of data for analytic
purposes and inherent challenges regarding data collection and data quality. An
exploration of the research on dental avoidance serves to identify valuable poten-
tial risk factors that may not currently be captured consistently in dental school
EDRs. Future directions must necessarily be based upon the current ability of den-
tal schools to leverage the functionality of information systems; however, students,
faculty, staff, and patients all might play a role in identifying potentially effective
interventions to reduce missed appointments.
treatments, tools, and techniques to help manage or reduce pain experienced dur-
ing a dental procedure. These new treatment modalities combined with improved
pain-reducing medication have benefited patients. Several behavioral and pharma-
cological interventions have also been developed to help patients better manage
their anxiety and fear of dental treatments (Appukuttan, 2016). While general
anxiety levels have increased in the United States, studies across several decades
have shown that levels of dental anxiety have remained relatively stable (Smith &
Heaton, 2003). Research has shown that approximately 10–20% of the adult pop-
ulation has a moderate to high degree of dental anxiety (Armfield et al., 2006;
Ragnarsson, 1998; Schwarz & Birn, 1995).
Dental anxiety can vary depending on the type of dental procedure being
performed. Invasive procedures such as extraction of teeth, drilling, periodontal
therapy, and dental injections tend to invoke greater anxiety (Moore et al.,1993;
Stabholz & Peretz, 1999). Dental anxiety can vary based on age group, gender,
socio-economic status, and educational levels. Age is associated with increased
dental anxiety, with anxiety increasing across age groups (Armfield et al., 2007)
and middle-aged patients showing a higher prevalence of dental anxiety (Armfield
et al., 2006).
Women tend to have higher anxiety levels when compared with men (Armfield
et al., 2006; Moore et al., 1993). Low education and income levels are correlated
with moderate levels of dental anxiety (Moore et al., 1993), with the highest prev-
alence of dental fear found among unemployed people (Armfield et al., 2006;
Armfield et al., 2007). The behavior of the dentist toward the patient plays a major
role in increased anxiety levels, especially when patients are faced with a dentist
that displays hostile behavior such as condescension (Moore et al., 1993).
9.4.2 Dental Avoidance
Low income, lack of dental insurance, and lack of access to dental care are some
of the major barriers that prevent patients from receiving the dental care that they
require. However, even when these barriers are not a primary factor in accessing
dental care, a significant proportion of the population avoids accessing dental ser-
vices. Dental anxiety is considered to play a key role in contributing to this dental
avoidance behavior (Armfield et al., 2007; Moore et al., 1993; Skaret et al., 2003).
The presence of pain may cause a patient to seek emergency dental treatment to
alleviate the discomfort (Armfield et al., 2007). This motivation to stop pain may
be the primary driver in the short term. However, once the pain and other symp-
toms such as swelling subside, the fear of sitting through multiple appointments of
complex dental procedures can result in dental avoidance behavior. This behavior
includes a delay in making dental appointments, missing appointments, canceling
appointments without rescheduling, or avoiding some or all recommended dental
treatments. Patients with high dental anxiety avoid visiting the dentist for several
Predicting Patient Missed Appointments ◾ 159
years (Moore et al., 1993; Schuller et al., 2003). These individuals have a signifi-
cantly higher number of decayed and missing teeth and a significantly lower num-
ber of teeth that have been filled and are functional (Armfield et al., 2007; Schuller
et al., 2003).
Only a small proportion (1.5–2%) of adult patients report fear of treatment as a
reason for canceling or missing dental appointments (Shabbir et al., 2018). Studies
with younger patients have shown that children with dental anxiety are more likely
to have two or more missed appointments (Wogelius & Poulsen, 2005). The pres-
ence of general anxiety and dental fear among parents or caregivers may make it
harder for parents to motivate, educate, and persuade their child to attend a dental
appointment (Gustafsson et al., 2010). In particular, the dental anxiety of mothers
is strongly correlated with dental anxiety in children (Wigen et al., 2009).
Moderate to high levels of dental anxiety lead to a pattern of behavior that
can further exacerbate existing dental problems. If increased levels of dental anxi-
ety lead to children or adults missing or canceling appointments, it may further
deteriorate their existing dental condition and result in poor treatment outcomes.
A poor treatment outcome can result in a further need for corrective measures
to rectify the problem. For example, preventive measures such as frequent oral
prophylaxis appointments and application of dental sealants where necessary can
prevent the occurrence of gum problems or tooth decay, especially in children. A
fear of dental treatment brought on by previous negative dental experiences such
as pain may result in the patient avoiding the dental appointment to receive pre-
ventive measures. Since patient dental education is part of the treatment process
in most dental clinics, the patient is therefore deprived of both the preventive
dental treatment and the educational reinforcement required to maintain proper
oral health status.
Missing or canceling appointments can prevent dentists and dental students
from identifying pathology in the early stages. For example, early-stage dental car-
ies (tooth decay) can be treated using conservative dental restoration techniques.
These procedures normally result in very little pain and removal of dental tissues.
In some cases, a dental anesthetic may not be required. However, patients with high
anxiety are associated with an increased average time between visits (Armfield et al.,
2007). Therefore, a problem that could have been rectified easily becomes a bigger
issue between visits. If the dental decay extends deep into the tooth, then injecting
a dental anesthetic and considerable drilling is required to remove decayed tooth
tissue. If the decay extends to the pulp, then removal of the pulp and filling of the
root canals will be required. All these procedures are expensive, more complex, and
potentially more painful. Studies have shown that patients with most dental fear
have significantly more missing teeth than patients with little or no fear (Armfield
et al., 2007). Although it has been hypothesized that the fear of dental treatment
can motivate patients to maintain good oral health status, studies have shown that
this is not the case (Schuller et al., 2003).
160 ◾ Data Analytics and AI
9.5.2 Clinical Notes
Since it has been established that EDR features and appointment scheduling pro-
cesses differ among dental schools, and that academic EDRs may differ in func-
tionality from those in other dental clinics, it is reasonable to expect that dental
students may not encounter a consistent format to enter clinical notes during their
training. In this way, EDR systems reflect the challenges faced with entering valu-
able clinical data into EHRs (Strome & Liefer, 2013). Defined fields may exist
but may be inconsistently used (Weir et al., 2003). If fields do not exist, capturing
qualitative data in clinical notes—even in unstructured text form—remains valu-
able for predicting potential future missed appointments. Enhanced student EDR
training regarding clinical documentation could emphasize where delineators or
notes templates should be used in a consistent manner. This enhances data quality
on the structured data in the EDR and improves the ability to use artificial intel-
ligence techniques such as natural language processing (NLP) on unstructured
data.
In an academic dental clinic, faculty concurrently oversee multiple students.
This presents a unique opportunity for student-faculty communication and faculty-
patient communication that may not be available in a conventional dental clinic or
primary care setting. These encounters may be natural places to add data collection
points to the clinical workflow and enhance the dental record with additional clini-
cal observations that can be used as factors for risk modeling. For example, several
students may report to a faculty member that they have seen an increase in the
number of missed appointments for a specific reason. Students may not see this as a
trend across a certain patient population, and they may not document the findings
in clinical notes in a way that would reveal a pattern if researched. Faculty may
identify the need to enter the occurrences into the EDR. Furthermore, if faculty
regularly interact with multiple patients in the course of supervising many students,
they serve as additional clinical resources that could detect predictors of future
missed appointments and document accordingly.
or easily introduced into a workflow, depending on how such changes are usually
documented across the clinic.
Composite reports based on patient satisfaction surveys or other quality mea-
sures may already be used for operational and financial purposes in the dental
school. Even if these reports are maintained in a system outside the EDR, their
raw data might be used as a source to compare outcomes of patients who success-
fully complete appointments compared with those who miss or reschedule appoint-
ments. Over time, patient satisfaction surveys might be enhanced with additional
questions about the relative level of difficulty involved in keeping appointments,
the top reasons patients give for needing to reschedule or cancel appointments, and
the ease with which they remember or track future appointments.
Other methods to improve data collection include simple operational changes
that focus on a few proven techniques for improving the process of appoint-
ment rescheduling, reminder management tools, and missed appointment rea-
son tracking. An additional contact from dental school students or support staff
may increase patient appointment compliance, although some methods are more
effective than others, particularly when a long appointment lead time is a factor
(Goffman et al., 2017).
If the dental school has already implemented a patient portal (such as
PatientAccess by Exan), the capability may already exist for patients to contrib-
ute data to their dental records. For example, if patients can cancel or reschedule
through a self-service tool, the system might provide a list of reason codes before
the new appointment can be confirmed or the original appointment canceled.
Identifying patients who tend to reschedule presents an opportunity to intervene
before the appointment is missed altogether, while gathering data about the dif-
ficulties faced by patients to adhere to their dental treatment plans. Finally, if the
patient portal includes features for messaging to providers or clinic staff, the text
transmitted by the patient could be as valuable as the text found in clinical notes.
9.6 Conclusions
Examining causes of appointment avoidance behavior in the academic dental clinic
is an opportunity to improve health outcomes while building on analytic successes
in other healthcare settings. From a provider and operational perspective, there are
important similarities between dental health clinics that serve a cross-section of the
community and primary care clinics (Chariatte et al., 2008; Schmalzried & Liszak,
2012). From the patient perspective, the primary drivers of missed appointments
for dental care may have the same root causes as missed appointments for other
kinds of medical care.
Since EDR and EMR systems have yet to reach a practical level of interoperabil-
ity, a complete patient profile across all sites of care remains difficult to maintain
Predicting Patient Missed Appointments ◾ 163
(Burris, 2017). Ideally, the dental student would know causes of missed appoint-
ments or the risk of future missed appointments and develop intervention strategies
to limit the negative impact to their own training and their patients’ dental health
outcomes. In practice, oral health providers often only see part of the full patient
picture.
Even where information is limited due to missing or inconsistent data, den-
tal schools can still leverage existing systems to collect likely factors to include in
predictive modeling. Such analytic methods might combine known sociodemo-
graphic factors such as employment status, age, and caregiving status with common
assessments used to measure dental avoidance. If a dental school is in the earliest
phases of using an EDR to enhance data collection, perhaps the most practical
way to begin is with a basic set of scenarios that illustrate a few likely variables that
would lead to the increased risk of missed appointments. Then, it might consider
the methods that work best for staff, students, and faculty to maintain consistently
high data quality when capturing data that can be used to test these scenarios.
Once processes are in place to reinforce the importance of including such data in
the patient’s dental record, other quantitative and qualitative data that can aid in
predictive modeling may become more readily apparent.
References
Aartman, I. H., van Everdingen, T., Hoogstraten, J., & Schuurs, A. H. (1998). Self-report
measurements of dental anxiety and fear in children: A critical assessment. ASDC
Journal of Dentistry for Children, 65(4), 252–230.
Acharya, A., Cheng, B., Koralkar, R., Olson, B., Lamster, I. B., Kunzel, C., & Lalla,
E. (2018). Screening for diabetes risk using integrated dental and medical elec-
tronic health record data. JDR Clinical and Translational Research, 3(2), 188–194.
doi:10.1177/2380084418759496
Appukuttan, D. P. (2016). Strategies to manage patients with dental anxiety and dental
phobia: Literature review. Clinical, Cosmetic and Investigational Dentistry, 8, 35–50.
doi:10.2147/CCIDE.S63626
Armfield, J. M., Spencer, A. J., & Stewart, J. F. (2006). Dental fear in Australia: Who’s afraid
of the dentist? Australian Dental Journal, 51(1), 78–85. doi:10.1111/j.1834-7819.2006.
tb00405.x
Armfield, J. M., Stewart, J. F., & Spencer, A. J. (2007). The vicious cycle of dental fear:
Exploring the interplay between oral health, service utilization and dental fear. BMC
Oral Health, 7, 1. doi:10.1186/1472-6831-7-1
Beckwith, F. R., Ackerman, R. J., Cobb, C. M., & Tira, D. E. (1999). An evaluation of
factors affecting duration of orthodontic treatment. American Journal of Orthodontics
and Dentofacial Orthopedics: Official Publication of the American Association of
Orthodontists, its Constituent Societies, and the American Board of Orthodontics, 115(4),
439–447. doi:10.1016/s0889-5406(99)70265-9
Burris, C. (2017). Connecting oral and physical health via the health information exchange.
North Carolina Medical Journal, 78(6), 410–412. doi:10.18043/ncm.78.6.410
164 ◾ Data Analytics and AI
Chariatte, V., Berchtold, A., Akré, C., Michaud, P., & Suris, J. (2008). Missed appoint-
ments in an outpatient clinic for adolescents, an approach to predict the risk of miss-
ing. The Journal of Adolescent Health: Official Publication of the Society for Adolescent
Medicine, 43(1), 38–45. doi:10.1016/j.jadohealth.2007.12.017
Chuang, S. K., Tian, L., Wei, L. J., & Dodson, T. B. (2002). Predicting dental implant survival by
use of the marginal approach of the semi-parametric survival methods for clustered obser-
vations. Journal of Dental Research, 81(12), 851–855. doi:10.1177/154405910208101211
Corah, N. L. (1969). Development of a dental anxiety scale. Journal of Dental Research,
48(4), 596. doi:10.1177/00220345690480041801
Corah, N. L., Gale, E. N., & Illig, S. J. (1978). Assessment of a dental anxiety scale.
Journal of the American Dental Association (1939), 97(5), 816–819. doi:10.14219/jada.
archive.1978.0394
Dailey, Y. M., Humphris, G. M., & Lennon, M. A. (2001). The use of dental anxiety ques-
tionnaires: A survey of a group of UK dental practitioners. British Dental Journal,
190(8), 450–453. doi:10.1038/sj.bdj.4801000
George, A. C., Hoshing, A., & Joshi, N. V. (2007). A study of the reasons for irregular
dental attendance in a private dental college in a rural setup. Indian Journal of Dental
Research: Official Publication of Indian Society for Dental Research, 18(2), 78–81.
doi:10.4103/0970-9290.32425
Goffman, R. M., Harris, S. L., May, J. H., Milicevic, A. S., Monte, R. J., Myaskovsky,
L., … Vargas, D. L. (2017). Modeling patient no-show history and predicting future
outpatient appointment behavior in the Veterans health administration. Military
Medicine, 182(5). doi:10.7205/milmed-d-16-00345
Gustafsson, A., Broberg, A. G., Bodin, L., Berggren, U., & Arnrup, K. (2010). Possible
predictors of discontinuation of specialized dental treatment among children and
adolescents with dental behaviour management problems. European Journal of Oral
Sciences, 118(3), 270–277. doi:10.1111/j.1600-0722.2010.00734.x
Guzek, L. M., Gentry, S. D., & Golomb, M. R. (2015). The estimated cost of “no-shows”
in an academic pediatric neurology clinic. Pediatric Neurology, 52(2), 198–201.
doi:10.1016/j.pediatrneurol.2014.10.020
Holmes, D. C., Trombly, R. M., Garcia, L. T., Kluender, R. L., & Keith, C. R. (2000).
Student productivity in a comprehensive care program without numeric require-
ments. Journal of Dental Education, 64(11), 745–754. Retrieved from http://www.
jdentaled.org/content/64/11/745.abstract
Holtshousen, W. S. J., & Coetzee, E. (2012). The effect of overbooking on idle dental chair
capacity in the pretoria region of the gauteng oral health services. SADJ: Journal of the
South African Dental Association = Tydskrif Van Die Suid-Afrikaanse Tandheelkundige
Vereniging, 67(8), 460–464.
Humphris, G. M., Morrison, T., & Lindsay, S. J. (1995). The modified dental anxiety scale:
Validation and United Kingdom norms. Community Dental Health, 12(3), 143–150.
Ilgüy, D., Ilgüy, M., Dinçer, S., & Bayirli, G. (2005). Reliability and validity of the modi-
fied dental anxiety scale in Turkish patients. The Journal of International Medical
Research, 33(2), 252–259. doi:10.1177/147323000503300214
Lenzi, H., Ben, Â J., & Stein, A. T. (2019). Development and validation of a patient no-
show predictive model at a primary care setting in southern Brazil. PLoS ONE, 14(4).
doi:10.1371/journal.pone.0214869
Lin, K. C. (2009). Behavior-associated self-report items in patient charts as predictors of
dental appointment avoidance. Journal of Dental Education, 73(2), 218–224.
Predicting Patient Missed Appointments ◾ 165
Moore, R., Birn, H., Kirkegaard, E., Brødsgaard, I., & Scheutz, F. (1993). Prevalence and
characteristics of dental anxiety in Danish adults. Community Dentistry and Oral
Epidemiology, 21(5), 292–296. doi:10.1111/j.1600-0528.1993.tb00777.x
Moresca, R. (2018). Orthodontic treatment time: Can it be shortened? Dental Press Journal
of Orthodontics, 23(6), 90–105. doi:10.1590/2177-6709.23.6.090-105.sar
Ragnarsson, E. (1998). Dental fear and anxiety in an adult Icelandic population. Acta
Odontologica Scandinavica, 56(2), 100–104. doi:10.1080/00016359850136067
Ramoni, R. B., Etolue, J., Tokede, O., McClellan, L., Simmons, K., Yansane, A., …
Kalenderian, E. (2017). Adoption of dental innovations: The case of a standardized
dental diagnostic terminology. Journal of the American Dental Association (1939),
148(5), 319–327. doi:10.1016/j.adaj.2017.01.024
Sachdeo, A., Konfino, S., Icyda, R. U., Finkelman, M. D., Gulati, H., Arsenault, P., &
Hanley, J. B. (2012). An analysis of patient grievances in a dental school clinical
environment. Journal of Dental Education, 76(10), 1317–1322.
Safi, Y., Khami, M. R., Razeghi, S., Shamloo, N., Soroush, M., Akhgari, E., & Moscowchi,
A. (2015). Designing and implementation of a course on successful dental practice for
dentists. Journal of Dentistry (Tehran, Iran), 12(6), 447–455. Retrieved from https://
www.ncbi.nlm.nih.gov/pmc/articles/PMC4754571/
Schwarz, E., & Birn, H. (1995). Dental anxiety in Danish and Chinese adults--a
cross-cultural perspective. Social Science & Medicine (1982), 41(1), 123–130.
doi:10.1016/0277-9536(94)00288-5
Schleyer, T. K. L. (2004). Should dentistry be part of the national health information infra-
structure? Journal of the American Dental Association (1939), 135(12), 1687–1695.
doi:10.14219/jada.archive.2004.0120
Schleyer, T. K., Thyvalikakath, T. P., Spallek, H., Dziabiak, M. P., & Johnson, L. A. (2012).
From information technology to informatics: The information revolution in dental
education. Journal of Dental Education, 76(1), 142–153.
Schmalzried, H. D., & Liszak, J. (2012). A model program to reduce patient failure to
keep scheduled medical appointments. Journal of Community Health, 37(3), 715–718.
doi:10.1007/s10900-011-9505-0
Schuller, A. A., Willumsen, T., & Holst, D. (2003). Are there differences in oral
health and oral health behavior between individuals with high and low dental
fear? Community Dentistry and Oral Epidemiology, 31(2), 116–121. doi:10.1034/
j.1600-0528.2003.00026.x
Shabbir, A., Alzahrani, M., & Abu Khalid, A. (2018). Why do patients miss dental appoint-
ments in eastern province military hospitals, kingdom of Saudi Arabia? Cureus, 10(3),
e2355. doi:10.7759/cureus.2355
Shimpi, N., Ye, Z., Koralkar, R., Glurich, I., & Acharya, A. (2018). Need for a diagnos-
tic centric care in dentistry: A case study from Marshfield clinic health system.
Journal of the American Dental Association (1939), 149(2), 122–131. doi:10.1016/j.
adaj.2017.09.030
Smith, T. A., & Heaton, L. J. (2003). Fear of dental care: Are we making any progress?
Journal of the American Dental Association (1939), 134(8), 1101–1108. doi:10.14219/
jada.archive.2003.0326
Skaret, E., Raadal, M., Kvale, G., & Berg, E. (2003). Gender-based differences in factors related
to non-utilization of dental care in young Norwegians: A longitudinal study. European
Journal of Oral Sciences, 111(5), 377–382. doi:10.1034/j.1600-0722.2003.00072.x
166 ◾ Data Analytics and AI
Stabholz, A., & Peretz, B. (1999). Dental anxiety among patients prior to different den-
tal treatments. International Dental Journal, 49(2), 90–94. doi:10.1111/j.1875-
595x.1999.tb00514.x
Stewart, B. A., Fernandes, S., Rodriguez-Huertas, E., & Landzberg, M. (2010). A pre-
liminary look at duplicate testing associated with lack of electronic health record
interoperability for transferred patients. Journal of the American Medical Informatics
Association: JAMIA, 17(3), 341–344. doi:10.1136/jamia.2009.001750
Strome, T. L., & Liefer, A. (2013). Healthcare analytics for quality and performance improve-
ment. Wiley: Hoboken, NJ.
Teich, S. T., Wan, Z., & Faddoul, F. F. (2012). Relationship between broken appointments
and dental students’ clinical experience level. Journal of Dental Education, 76(9),
1167–1174.
Thyvalikakath, T. P., Padman, R., Vyawahare, K., Darade, P., & Paranjape, R. (2015).
Utilizing dental electronic health records data to predict risk for periodontal disease.
Studies in Health Technology and Informatics, 216, 1081.
Trenouth, M. J. (2003). Do failed appointments lead to discontinuation of orthodontic
treatment? The Angle Orthodontist, 73(1), 51–55. doi:10.1043/0003-3219(2003)07
3<0051:DFA LTD>2.0.CO;2
Walji, M. F., Kalenderian, E., Stark, P. C., White, J. M., Kookal, K. K., Phan, D., …
Ramoni, R. (2014). BigMouth: A multi-institutional dental data repository. Journal of
the American Medical Informatics Association: JAMIA, 21(6), 1136–1140. doi:10.1136/
amiajnl-2013-002230
Weir, C. R., Hurdle, J. F., Felgar, M. A., Hoffman, J. M., Roth, B., & Nebeker, J. R. (2003).
Direct text entry in electronic progress notes: An evaluation of input errors. Methods
of Information in Medicine, 42(1), 61–67.
Wigen, T. I., Skaret, E., & Wang, N. J. (2009). Dental avoidance behaviour in parent
and child as risk indicators for caries in 5-year-old children. International Journal of
Paediatric Dentistry, 19(6), 431–437. doi:10.1111/j.1365-263X.2009.01014.x
Wogelius, P., & Poulsen, S. (2005). Associations between dental anxiety, dental treat-
ment due to toothache, and missed dental appointments among six to eight-year-
old Danish children: A cross-sectional study. Acta Odontologica Scandinavica, 63(3),
179–182. doi:10.1080/00016350510019829
Chapter 10
Machine Learning in
Cognitive Neuroimaging
Siamak Aram, Denis Kornev, Ye Han,
Mina Ekramnia, Roozbeh Sadeghian,
Saeed Esmaili Sardari, Hadis Dashtestani,
Sagar Kora Venu, and Amir Gandjbakhche
Contents
10.1 Introduction............................................................................................168
10.1.1 Overview of AI, Machine Learning, and Deep Learning in
Neuroimaging.............................................................................168
10.1.2 Cognitive Neuroimaging............................................................170
10.1.3 Functional Near-Infrared Spectroscopy......................................171
10.2 Machine Learning and Cognitive Neuroimaging....................................172
10.2.1 Challenges...................................................................................173
10.3 Identifying Functional Biomarkers in Traumatic Brain Injury
Patients Using fNIRS and Machine Learning.........................................173
10.4 Finding the Correlation between Addiction Behavior in Gaming and
Brain Activation Using fNIRS................................................................ 174
10.5 Current Research on Machine Learning Applications in
Neuroimaging......................................................................................... 176
10.6 Summary.................................................................................................178
References..........................................................................................................179
167
168 ◾ Data Analytics and AI
10.1 Introduction
In this chapter, we will review how neuroimaging models have used predictive
models to ask new questions and uncover brand new aspects of organization in
cognition. This chapter will provide an overview of current, up-to-date information
on how these approaches can be used for brain disorders. This includes two differ-
ent brain applications:
modality from existing data without performing actual scans can help avoid radia-
tion (Zhu et al., 2019).
10.1.2 Cognitive Neuroimaging
Cognitive neuroimaging is the science that studies the relationship between human
brain activity and other perspectives of the nervous system with cognitive processes
and behavior. It uses common approaches and methods of visualization of the
structure, function, and biochemical characteristics of the brain. As a study about
the human brain, cognitive neuroimaging is the co-discipline of neurology, neuro-
surgery, and psychiatry on one hand, and as a branch of both neuroscience and psy-
chology intersecting with cognitive psychology and neuropsychology on the other.
Since ancient times, people have tried to penetrate into the depths of consciousness,
and Hippocrates was one of the first thinkers who claimed that thinking was car-
ried out in the brain (O’Shea, 2013). The modern history of neuroimaging began
in the 19th century from a hypothesis Italian neuroscientist Angelo Mosso had that
“an attentional or cognitive task can locally increase cerebral blood flow” (Sandrone
et al., 2013). Based on this theory, Mosso developed the “Mosso method” (Mosso,
1881) and “human circulation balance” (Mosso, 1882).
Nowadays, in the realm of cognitive neuroimaging, different studies about vari-
ous forms of cognitive disorders have been provided and the methods of structural
and functional neuroimaging are an essential part of these studies (Odinak et al.,
2012). These methods are functional near-infrared spectroscopy (fNIRS), fMRI,
electroencephalogram (EEG), PET, and magnetoencephalography (MEG), which
could be used by researchers for the investigation of the hemodynamic response
patterns in a cognitive disorder characterized by memory impairment (Li et al.,
2018), vascular dementia diagnosis (Van Straaten, Scheltens, & Barkhof, 2004;
O’Brien, 2007), evaluating, classifying, and following some of the cognitive disor-
ders (De Medeiros Kanda et al., 2009), as a marker of the early stage of Alzheimer’s
disease (Alexander et al., 2002), and as a method for measuring neuronal events in
a direct and non-invasive way (Uhlhaas et al., 2017) respectively.
At the same time, affordability and ability empirically predict the result
allowing cognitive scientists to use neuroimaging methods to build test models
for experiments in spoken word production and research into episodic memory
(De Zubicaray, 2006). Based on the current stage of cognitive neuroimaging
development, researchers have discussed the present and future perspectives of
brain imaging (Dolan, 2008). Now they have the opportunity to address unbe-
lievable challenges in human brain study, scientists can operate, monitor, and
predict performance in the human brain without direct impact. They can set
questions and provide responses about the reaction of brain regions for external
manipulations in terms of correlation with different disorders, such as addictive
behavior in gaming using f NIRS, which is one of the reviewed projects in this
chapter.
Machine Learning in Cognitive Neuroimaging ◾ 171
Figure 10.1 The fNIRS device was used to find the correlation between addic-
tion behavior in gaming and brain activation.
172 ◾ Data Analytics and AI
10.2.1 Challenges
As the applications of machine learning methods are rising, so are the new chal-
lenges. Some of the aspects of machine learning are needed to be kept in mind in
order to reliably apply them in neuroscience (Davatzikos, 2019). In the era of big
data, deep learning models fit the data outstandingly well. However, generalization
seems to still be an issue to ponder on. It is called “overfitting.” Proper split-sample,
cross-validation, and replication to new samples would be helpful to evaluate the
accuracy of our model. One of the major studies of classical machine learning
applications in clinical neuroscience is to classify patients vs. control/healthy par-
ticipants. Although this is a great starting point, a more significant value would
come from detecting subtle changes in the early stages before the disease is clini-
cally detectable. This is a much tougher problem. The other challenge would be
the heterogeneity of the patterns and sources. Data acquired from different sites
and channels increase the problem of overfitting and generalization. This problem
can be treated by appropriate statistical harmonization methods which are conse-
quently in a growing field (Fortin et al., 2018).
Figure 10.2 The average activity maps for the left slope of the activity curve
(CSL) and variance of the HbO signal (HV) features for the healthy (A) and TBI
(B) subjects are illustrated.
and scientists. This research study concentrates on data acquired by fNIRS and the
Iowa gambling task (IGT) (Aram et al., 2019) to study how prefrontal cortex brain
activation and gaming addiction are interrelated, i.e., the goal of this research is to
show the interrelation of activity in the human prefrontal cortex with individual
ability to make decisions and how hemodynamic changes react to gaming tasks.
Thirty participants were involved in the experiment. They performed the IGT test
while connected to the fNIRS instrument for measuring activity in their brains.
The task for detecting patients with damage to the ventromedial sector of prefron-
tal cortices was developed by Bechara et al. (1994). This method simulated the
real-life decision-making process and allowed identifying the elusive deviation of
patients with different disorders in the laboratory environment. In recent studies,
the IGT was used for investigations into whether brain activity increased during a
video game (Dong et al., 2017), examining internet addiction and internet gam-
ing disorder (Lin et al., 2019), and others. The same approach was executed in this
176 ◾ Data Analytics and AI
study. Each participant was offered the choice of cards one by one from one of the
four decks on the digital table. Each card deck had a different gain-loss structure
and thus the ability to lose or win money. During the decision-making process in
the participant’s brain, fNIRS measured the level of oxyhemoglobin (HbO) of the
prefrontal cortex and eventually brain activation and characteristics of addictive
behavior in the video game playing process. As an additional task, the same strat-
egy as in Section 10.3 was used to extract the features using a machine learning
approach and the results confirmed the initial findings. Acquired data in the results
of experiments with IGT and fNIRS has been analyzed using repeated measures
ANOVA (Girden, 1992), and post-hoc comparisons were analyzed using dependent
t-tests with Bonferroni correction (Bonferroni, 1936). The outcome shows that the
HbO level of the left hemisphere of participants’ brains had a significantly different
value than the right (see Figure 10.3). This result correlates with an earlier study
about relations between motivation intensity and fNIRS activity in the prefrontal
cortex (Fairclough, Burns, & Kreplin, 2018), and depicts the relation of HbO levels
and activity in the brain during a memory task as a reaction on probable financial
gain or loss.
In the IGT test, the task was divided into five blocks of 20 trails each, and the
score was calculated based on the difference between the number of advantageous
deck selections with the number of disadvantageous selections. Data was screened
for errors, missing data (none), outliers (none found with Mahalanobis distance),
and assumptions. Normality, linearity, homogeneity, and Mauchly’s test (Mauchly,
1940) indicated that assumptions were met.
In spite of the fact that the IGT score using Huynh-Feldt correction (Huynh &
Feldt, 1976) for p-value showed different results, no group differences were revealed
by post-hoc comparisons analyzed using dependent pairwise t-tests with Bonferroni
correction. Based on achieved results, the conclusion could be made that the ele-
ments of a deep learning approach used in this study allow us to judge their effec-
tiveness in identifying the correlation between addiction behavior in gaming and
brain activity. The results of these experiments do not contradict the previously
conducted experiments (Fairclough, Burns, & Kreplin, 2018; Lin et al., 2019) in
the field of psychology and can be used as a basis for further research.
Figure 10.3 Mean oxygenated hemoglobin with 95% confidence intervals for
left and right hemispheres of the prefrontal cortex of the brain.
10.6 Summary
In recent years, AI, machine learning, and deep learning have become increas-
ingly popular as methodologies not only for analyzing research results in cognitive
neuroscience and brain research but also for confirming diagnoses and support-
ing the treatment of diseases. In turn, these methods, as based on mathematical
algorithms and problems, need reliable tools and approaches for collecting and
Machine Learning in Cognitive Neuroimaging ◾ 179
References
Alexander, G.F., Chen, K., Pietrini, P., Rapoport, S.I., & Reiman, E.M. (2002).
Longitudinal PET evaluation of cerebral metabolic decline in dementia: a potential
outcome measure in Alzheimer`s disease treatment studies. The American Journal of
Psychiatry, 159(5), 738–745.
American Medical Association (2007). AMA takes action on video games. Retrieved October
15, 2018, https://www.ama-assn.org/sites/default/fi les/media-browser/public/about-
ama/councils/Council%20Reports/council-on-science-public-health/a07-c saph-eff e
cts-video-games-internet.pdf.
American Psychiatric Association (2013). Diagnostic and Statistical Manual of Mental
Disorders. DSM-5® (5th Ed.). American Psychiatric Publishing.
American Psychiatric Association (2018). Internet Gaming. Retrieved from https://ww w.psy
chiatr y.org/patients-families/internet-ga ming.
Aram, S., Levy, L., Patel, J.B., Anderson, A.A., Zaragoza, R., Dashtestani, H., … Tracy,
J.K. (2019). The Iowa Gambling Task: a review of the historical evolution, scientific
basis, and use in functional neuroimaging. SAGE Open, I–I2. Retrieved from https://
journals.sagepub.com/doi/full/10.1177/2158244019856911
Bechara, A., Damasio, A., Damasio, H., & Anderson, S. (1994). Insensitivity to future
consequences following damage to the human prefrontal cortex. Cognition, 50, 7–15.
Benerradi, J., Maior, H.A., Marinescu, A., Clos, J., & Wilson, M.L. (2019). Exploring
Machine Learning Approaches for Classifying Mental Workload Using fNIRS Data From
HCI Tasks. In Proceedings of the Halfway to the Future Symposium, ACM.
180 ◾ Data Analytics and AI
Bhambhani, Y., Maikala, R., Farag, M., & Rowland, G. (2006). Reliability of near‐infrared
spectroscopy measures of cerebral oxygenation and blood volume during handgrip
exercise in nondisabled and traumatic brain‐injured subjects. Journal of Rehabilitation
Research and Development, 43, 845.
Bonferroni, C.E. (1936). Teoria Statistica Delle Classi e Calcolo Delle Probabilità. Libreria
internazionale Seeber, 8, 1–62.
Bunce, S., Izzetoglu, K., Izzetoglu, M., Onaral, K., Banu O., & Kambiz P. (2006).
Functional near-infrared spectroscopy. IEEE Engineering in Medicine and Biology
Magazine, 25(4), 54–62.
Chen, H., Zhang, Y., Zhang, W., Liao, P., Li, K., Zhou, J., & Wang, G. (2017). Low-dose
CT via a convolutional neural network. Biomedical Optics Express, 8, 679–694.
Chow, D. (2018). Machine Learning in neuroimaging. American Journal of Neuroradiology.
Retrieved from http://ajnrdigest.org/machine-learning-neuroimaging/.
Convolutional Neural Networks (CNNs / ConvNets). Convolutional Neural Networks for
Visual Recognition. Retrieved from http://cs231n.github.io/convolutiona l-net works/.
Cortes, C. & Vapnik, V. (1995). Support-vector networks. Machine Learning, 20, 273–297.
Davatzikos, C. (2019). Machine Learning in neuroimaging: progress and challenges.
NeuroImage, 197, 652.
De Medeiros Kanda, P.A., Anghinah, R., Smidth, M.T, & Silva, J.M. (2009). The clinical use
of quantitative EEG in cognitive disorders. Dementia Neuropsychologia, 3(3), 195–203.
De Zubicaray, G.I. (2006). Cognitive neuroimaging: cognitive science out of the armchair.
Brain and Cognition, 60, 272–281.
Dolan, L.J. (2008). Neuroimaging of cognition: past, present, and future. Neuron, 60(3),
496–502.
Dong, G., Wang, L., Du, X., & Potenza, M.N. (2017). Gaming increases the craving
for gaming-related stimuli in individuals with internet gaming disorder. Biological
Psychiatry: Cognitive Neuroscience and Neuroimaging, 2(5), 404–412.
Entertainment Software Association (2019). Essential Facts About the Computer and Video
Game Industry. Retrieved from https://ww w.theesa.com/esa-research/2019-essential
-facts-about-the-computer-and-video-game-industry/.
Fairclough, S.H., Burns, C., & Kreplin, U. (2018). fNIRS activity in the prefrontal cortex
and motivational intensity: impact of working memory load, financial reward, and
correlation-based signal improvement. Neurophotonics, 5(3), 035001-1-10.
Ferguson, C.J., Coulson, M., & Barnett, J. (2011). A meta-analysis of pathological gam-
ing prevalence and comorbidity with mental health, academic and social problems.
Journal of Psychiatric Research, 45,1573–1578.
Fortin, J.P., Cullen, N., Sheline, Y.I., Taylor, W.D., Aselcioglu, I., Cook, P.A., … &
McInnis, M. (2018). Harmonization of cortical thickness measurements across scan-
ners and sites. Neuroimage, 167, 104–120.
Friston, K.J. (1994). Functional and effective connectivity in neuroimaging: a synthesis.
Human Brain Mapping, 2(1–2), 56–78.
Girden, E.R. (1992). ANOVA: Repeated Measures. Quantitative Applications in the Social
Sciences. Newbury Park, CA: Sage Publications.
Griffis, J.C., Allendorfer, J.B., & Szaflarski, J.P. (2016). Voxel-based Gaussian naïve Bayes
classification of ischemic stroke lesions in individual T1-weighted MRI scans. Journal
of Neuroscience Methods, 257, 97–108.
Haller, S., Badoud, S., Nguyen, D., Garibotto, V., Lovblad, K.O., & Burkhard, P.R.
(2012). Individual detection of patients with Parkinson disease using support vector
machine analysis of diffusion tensor imaging data: initial results. American Journal of
Neuroradiology, 33(11), 2123–2128.
Machine Learning in Cognitive Neuroimaging ◾ 181
Hibino, S., Mase, M., Shirataki, T., Nagano, Y., Fukagawa, K., Abe, A., … Ogawa, T.
(2013). Oxyhemoglobin changes during cognitive rehabilitation after traumatic brain
injury using near infrared spectroscopy. Neurologia Medico‐chirurgica, 53, 299–303.
Ho, T.K.K., Gwak, J., Park, C.M., & Song, J.I. (2019). Discrimination of mental work-
load levels from multi-channel fNIRS using Deep Learning-based approaches. IEEE
Access, 7, 24392–24403.
Huynh, H. & Feldt, L.S. (1976). Estimation of the box correction for degrees of freedom
from sample data in randomized block and split-plot designs. Journal of Educational
Statistics, 1(1), 69–82.
Jiang, F., Jiang, Y., Zhi, H., Dong, Y, Li, H., Ma, S., … Wang, Y. (2017). Artificial intel-
ligence in healthcare: past, present, and future. Stroke and Vascular Neurology, 2(4),
230–243.
Kamal, H., Lopez, V. & Sheth, S.A. (2018). Machine learning in acute ischemic stroke
neuroimaging. Frontiers in Neurology, 9, 945.
Kamnitsas, K., Castro, D.C., Folgoc, L.L., Walker, I., Tanno, R., Rueckert, D., … &
Nori, A. (2018). Semi-supervised learning via compact latent space clustering. arXiv
preprint arXiv:1806.02679.
Kang, E., Min, J., & Ye, J.C. (2017). A deep convolutional neural network using directional
wavelets for low-dose X-ray CT reconstruction. Medical Physics, 44(10), 360–375.
Karamzadeh, N., Amyot, F., Kenney, K., Anderson, A., Chowdhry, F., Dashtestani, H., … &
Diaz‐Arrastia, R. (2016). A Machine Learning approach to identify functional bio-
markers in the human prefrontal cortex for individuals with traumatic brain injury
using functional near‐infrared spectroscopy. Brain and Behavior, 6(11), 1–14.
Li, R., Rui, G., Chen, W., Li, S., Schulz, P.E., & Zhang, Y. (2018). Early detection of
Alzheimer’s disease using non-invasive near-infrared spectroscopy. Frontiers in Aging
Neuroscience, 10, 366.
Li, R., Zhang, W, Suk, H., Wang, L., Li, J., Shen, D., & Ji, S. (2014). Deep Learning-
Based imaging data completion for improved brain disease diagnosis. Medical Image
Computing and Computer-Assisted Intervention, 17(Pt 3), 305–312.
Lin, C.H., Wang, C.C., Jia-Huang Sun, J.H., Ko, C.H., & Chiu, Y.C. (2019). Is the clinical
version of the Iowa Gambling Task relevant for assessing choice behavior in cases of
Internet addiction? Frontiers in Psychiatry, 10, 232.
Mauchly, J.W. (1940). Significance test for sphericity of a normal n-variate distribution. The
Annals of Mathematical Statistics. 11(2), 204–209.
Mazurowski, M.A., Buda, M., Saha, A., & Bashir, M.R. (2018). Deep Learning in radi-
ology: an overview of the concepts and a survey of the state of the art. Journal of
Magnetic Resonance Imaging, 49(4), 1–27.
McCarthy, J. (2017). What is Artificial Intelligence? Retrieved from http://jmc.stanford.edu
/articles/whatisai/whatisai.pdf.
Merzagora, A.C.R., Izzetoglu, M., Onaral, B., & Schultheis, M.T. (2014). Verbal work-
ing memory impairments following traumatic brain injury: an fNIRS investigation.
Brain Imaging and Behavior, 8, 446–459.
Merzagora, A.C., Schultheis, M.T., Onaral, B., & Izzetoglu, M. (2011). Functional near‐
infrared spectroscopy-based assessment of attention impairments after traumatic
brain injury. Journal of Innovative Optical Health Sciences, 4, 251–260.
Mosso, A. (1881). Concerning the Circulation of the Blood in the Human Brain. Leipzig,
Germany: Verlag von Viet & Company.
Mosso, A. (1882). Applicazione della bilancia allo studio della circolazione del sangue
nell’uomo. Atti della R. Accademia delle scienze di Torino, XVII, 534–535.
182 ◾ Data Analytics and AI
Nie, D., Cao, X., Gao, Y., Wang, L., & Shen, D. (2016). Estimating CT image from MRI
data using 3D fully convolutional networks. Deep Learning and Data Labeling for
Medical Applications, 2016, 170–178.
O’Brien, J.T. (2007). Role of imaging techniques in the diagnosis of dementia. The British
Journal of Radiology 80(2) Spec No 2, 71–77.
Odinak, M.M., Emelin, A.Y., Lobzin, V.Y., Vorobyev, S.V., & Kiselev, V.N. (2012). Current
capacities for neuroimaging in the differential diagnosis of cognitive impairments.
Neurology, Neuropsychiatry, Psychosomatics 4(2S), 51–55.
O’Shea, M. (2013). The brain: milestones of neuroscience. The New Scientist, 218(2911),
ii–iii.
Overmars, O. (2012). A Brief History of Computer Games. Retrieved from https://ww w.sti
chtingspel.org/sites/default/fi les/histor y_of_ games.pdf.
Przybylski, A.K, Weinstein, N., & Murayama, K. (2017). Internet gaming disorder:
investigating the clinical relevance of a new phenomenon. The American Journal of
Psychiatry, 174, 230–236.
Quaresima, V. & Ferrari, M. (2019). A Mini-Review on functional near-infrared spectros-
copy (fNIRS): where do we stand, and where should we go? Photonics, 6(3), 87.
Rehme, A.K., Volz, L.J., Feis, D.L., Bomilcar-Focke, I., Liebig, T., Eickhoff, S.B., Fink,
G.R., Grefkes, C. (2015). Identifying neuroimaging markers of Motor Disability in
acute stroke by Machine Learning Techniques. Cerebral Cortex, 25, 3046–3056.
Rojas, R.F., Huang, X., & Ou, K.L. (2019). A Machine Learning approach for the iden-
tification of a biomarker of human pain using fNIRS. Scientific Reports, 9(1), 5645.
Sadaghiani S, Hesselmann G, Friston KJ, Kleinschmidt A. (2010). The relation of ongo-
ing brain activity evoked neural responses and cognition. Frontiers in the Systems
Neuroscience, 4, 20.
Sandrone, S., Bacigaluppi, M., Galloni, M.R., Cappa, S.F., Moro, A., Catani, M. …
Martino, G. (2013). Weighing brain activity with the balance: Angelo Mosso’s origi-
nal manuscripts come to light. Brain, 259(11), 1–13.
Schlemper, J., Caballero, J., Hajnal, J.V., Price, A., & Rueckert, D. (2017). A deep cas-
cade of convolutional neural networks for dynamic MR image reconstruction. IEEE
Transactions on Medical Imaging, 37(2), 491–503.
Sirpal, P., Kassab, A., Pouliot, P., Nguyen, D.K., & Lesage, F. (2019). fNIRS improves sei-
zure detection in multimodal EEG-fNIRS recordings. Journal of Biomedical Optics,
24(5), 1–9.
The World Health Organization (2018). Gaming Disorder. Retrieved from https://www
.who.int/features/qa/gaming-disorder/en/.
Uhlhaas, P.J., Liddle, P., Linden, D.E.J., Nobre, A.C., Singh, K.D., & Grossa J. (2017).
Magnetoencephalography as a tool in psychiatric research: current status and perspec-
tive. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging, 2(3), 235–244.
Van Straaten, E.C., Scheltens, P., & Barkhof, F. (2004). MRI and CT in the diagnosis of
vascular dementia. Journal of the Neurological Sciences, 226(1/2), 9–12.
Varoquaux, G., & Thirion, B. (2014). How Machine Learning is shaping cognitive neuro-
imaging. GigaScience, 3(1), 28.
Zhu, G., Jiang, B., Tong, L., Xie, Y., Zaharchuk, G. & Wintermark, M. (2019). Applications
of Deep Learning to neuro-imaging techniques. Frontiers in Neurology, 10, 869.
Chapter 11
People, Competencies,
and Capabilities Are
Core Elements in
Digital Transformation:
A Case Study of a
Digital Transformation
Project at ABB
Ismo Laukkanen
Contents
11.1 Introduction............................................................................................184
11.1.1 Objectives and Research Approach.............................................186
11.1.2 Challenges Related to the Use of Digitalization and AI..............187
11.2 Theoretical Framework............................................................................187
11.2.1 From Data Collection into Knowledge Management and
Learning Agility..........................................................................187
11.2.2 Knowledge Processes in Organizations.......................................189
11.2.3 Framework for Competency, Capability, and Organizational
Development...............................................................................190
183
184 ◾ Data Analytics and AI
11.1 Introduction
Digitalization or the widespread adoption of digital technologies, industrial
Internet of Things (IIoT), artificial intelligence (AI), and advanced analytics is
driving digital transformation in various industrial sectors. The major drivers for
this development have been energy and industrial revolutions, as illustrated in
Figure 11.1. Industrial revolution is a major driver for digitalization, IoT, and AI.
The transformation will disrupt several industrial sectors including, for example,
energy production, industrial manufacturing, transportation, and infrastructure.
As one real-life example, electric utilities are facing both opportunities and
challenges in the emerging energy revolution. Some key questions are how to boost
productivity out of aging and regulated assets and how to manage distributed
renewable energy resources, along with the future massive integration of electric
plug-in vehicles and the complications and changes in the regulatory landscape. At
the same time, the decline in demand and revenue from the regulated asset base
has them thinking about digitalization, monetization of data, and alternate sources
of revenue (ABB, 2018a).
Based on our analysis related to the use of digital technologies in industry, there
is a remarkable potential in the use of digital, industrial IoT, analytics, and AI
Core Elements in Digital Transformation ◾ 185
Figure 11.1 Industrial revolution is a major driver for digitalization, IoT, and AI
(ABB, 2019a).
are clustered among industry leaders. These early adopters saw a 70% productivity
increase, compared with just 30% for industry followers.
Several national level R&D and innovation programs (e.g., Industry 4.0 in
Germany, Made in China 2025, and Manufacturing USA) have a strong focus
in the development in IIoT, AI, advanced analytics, and intelligent technologies
to enable the fourth industrial revolution (World Economic Forum, 2018; CSIS,
2019; Wübbeke et al., 2016; Manufacturing USA, 2019).
Based on McKinsey, there will be remarkable value migration opportunities in
digital services, which are illustrated in Figure 11.3.
The research and development efforts reported here wish to serve both as useful
theoretical insights and as practical solutions to the described overall challenge and
to the problems of particular cases. The report is based on the results from the proj-
ect “TAKEOFF innovation, talent and competence development for industry 4.0.”
This research has been partly funded by EIT Digital and done in collaboration with
Aalto University, Aalto Executive Education, and some partner companies of ABB.
11.2 Theoretical Framework
11.2.1 From Data Collection into Knowledge
Management and Learning Agility
The role of knowledge and knowledge work as a competitive advantage has received
considerable attention in management theories (e.g., Drucker, 1994; Liebowitz,
1999, 2000, 2012), Davenport, 2005). The common hierarchy of knowledge, as a
part of organizational information and knowledge management, can be defined as
follows: data, information, knowledge, expertise/intelligence, and wisdom as illus-
trated in Figure 11.4. The key challenge in advanced analytics and AI solutions is
how to identify the critical data variables to be collected, how to get enough rich
188 ◾ Data Analytics and AI
Digital will replace physical Physical product and digital solution will
products be integrated solutions
Adapted from Furr and Shipolov [2019] and Fountaine, McCarthy, and Saleh [2019].
Core Elements in Digital Transformation ◾ 189
and reliable data for advanced analytics algorithms, and how to refine collected
data into useful knowledge and intelligence.
The creation and diffusion of knowledge have also become increasingly impor-
tant factors in competitiveness, especially when knowledge as a high valuable
commodity is embedded in products (especially high technology) and knowledge
services (knowledge intensive and digital services) (Hertog, 2000; Dalkir, 2017).
At the same time, the competitive advantage and value of knowledge is becoming
shorter, emphasizing the need for combining knowledge management and trans-
formational leadership to enable the transient competitive advantage which will be
introduced in Section 11.3.
leveraging short and temporary advantages, which are illustrated in Figure 11.6.
According to McGrawth (2013), the wave of transient advantage consists of five
phases:
Figure 11.6 Ramp-up and reconfiguration of competencies and capabilities in management of transient advantage. (Adapted
from McGrawth, 2013.)
Core Elements in Digital Transformation ◾ 193
(e.g., 70:20:10 model). Transformational leadership aims for the full commitment of
the cornerstones of leadership seen like shared practical values. The focus on learning
is to confirm single loop learning which can be a challenge in complex and dynamic
work environments having a focus on troubleshooting and firefighting. The presence
of the idea and practices of deep learning is vital. At the team level, the focus is on
the development of cross-functional team and process capabilities which are often
dynamic capabilities (Teece et al., 1997). The learning process is based on double
loop learning where the best practices of processes are systematically identified and
improved. Transformational leadership is based on a team level learning process.
Understanding and using the unique personalities, personal skills, and knowledge
are the source of creative and motivating team processes.
At the organizational level, the focus of development is on organizational capital
including human resources, organizational structures, and networks. The learn-
ing process is organizational learning where strategic capabilities are identified and
developed based on strategic targets. Transformational leadership focuses on con-
tinuous improvement of the leadership system and the development of the culture.
At the inter-organizational level, the development of capabilities focuses on busi-
ness ecosystems or supplier network capabilities which often are complementing
capabilities. The learning process is inter-organizational, combining benchmark-
ing and learning from the best partners. Deep leadership focuses on creating open
interaction and accelerating open or semi-open innovation in a supplier network
type of organizational structure.
194 ◾ Data Analytics and AI
Systematic organizational development and change are the key areas when ana-
lyzing, planning, and executing transformation programs. A typical challenge is
that the focus of the programs covers mainly individual competences and team,
organizational, and inter-organizational aspects are neglected. Beyond this, the
synchronization of these levels and the respective learning loops is mostly forgot-
ten. Detailed theoretical background of this model can be found in Nissinen and
Laukkanen (2018).
goal was to develop a systematic concept for a country level ramp-up of digital busi-
ness. The concept is illustrated in Figure 11.10. Phased and customer focused ramp-
up of digital services and the ramp-up was executed using scaled and agile methods:
Execution of digital transformation starts from digital business strategy and targets.
The strategic targets are cascaded down into selected industrial focus segments, for
example, intelligent marine industry and prioritized customer projects. These verti-
cals and projects will prioritize development of the solution portfolio, capabilities,
and operating model needed for customer delivery projects. Finally, competencies
and capabilities are mapped on the basis of competence gap analysis in different levels.
The execution in Finland was based on the use of the scaled and agile model
(SAFe) in the planning and execution of the project as sprint type increments as
illustrated in Figure 11.11.
The framework based on customer solutions is illustrated in Figure 11.12 and
examples of competence and capability mapping for these customer solutions in
different levels are illustrated in Figure 11.13. Critical elements in the develop-
ment of digital business capability are competencies, skills, and capabilities. In
Figure 11.13, competencies and capabilities needed in the delivery of industrial IoT
solutions are illustrated for different hierarchy levels.
Core Elements in Digital Transformation ◾ 197
Product 2.0 The next level is called Product 2.0 and it is a by-product of
continuous monitoring giving manufacturers insight into
Solution:
the features that their industrial customers use most. For
Digital
R&D and software development, this data gives very
Powertrain
important insight and helps to decide how to make
products better based on usage data.
Figure 11.11 Scaled and agile model for ramp-up of digital solutions.
Figure 11.12 IoT solution framework and examples of customer solutions in ramp-up.
Figure 11.13 Competencies and capabilities needed in the delivery of IoT solutions.
Core Elements in Digital Transformation ◾ 201
202 ◾ Data Analytics and AI
Figure 11.14 ABB Ability Marine Pilot vision system for ships (Moratella, 2019).
204 ◾ Data Analytics and AI
◾◾ It improves the safety of the passengers and the crew by improved situational
awareness, real time docking assistance, and incident and damage prevention
for infrastructure and other ships.
◾◾ There are also significant time savings in docking operations, which will also
result in significant annual fuel savings.
In addition to ABB Marine Pilot vision, there is an ABB diagnostic tool providing
troubleshooting and condition monitoring for electrical and mechanical systems
installed onboard the vessel with possibilities for remote assistance from the ABB
Collaborative Operations Center. To address the individual needs of each vessel,
the RDS is completely modular. This allows for the expansion of the diagnostic
capability from monitoring a few critical components or subsystems, to a fully
multi-disciplinary diagnostic solution for an entire vessel’s equipment, enabling
remote support from different manufacturers.
Summary: Impacts of digitalization and smart asset management in ABB
Marine Pilot Vision:
11.5.2 Future Recommendations
ABB Marine solutions is one of the leading businesses in the utilization of technolo-
gies and other industrial sectors will follow the steps toward autonomous operations
which are illustrated in Figure 11.16. An industrial case pilot for autonomous ships
has already been piloted (ABB, 2019c) and a concept for collaborative operations
has been launched.
The journey and lessons learned from autonomous shops show that the time
needed for digital transformation from product business into solution and service
can take several years as illustrated in Figure 11.17. Integration of different digital,
mobile, AI, and advanced analytics technologies will speed up this development,
and ABB Finland had the first industrial case pilot of using a 5G network, AI,
and the ABB Ability platform in industrial operations. The use of scaled and agile
methods in this transformation will accelerate the change.
References
ABB, 2018a, Digital road to the energy revolution, Rob Massoudi, https://www.abb-
conversations.c om/2018/06/digital-road-to-the - ener gy-re volut ion/ (Retrieved
10.12.2019).
ABB, 2019a, ABB group web page, http://www.abb.com (Retrieved 10.12.2019).
ABB, 2019b, ABB Ability web page, https://ability.abb.com (Retrieved 10.12.2019).
ABB, 2019c, ABB enables groundbreaking trial of remotely operated passenger ferry, https://
new.abb.com/news/detail/11632/abb-enables-groundbreak ing-trial-of-remotely-oper
ated-passenger-ferr y (Retrieved 10.12.2019).
ABB 2019d, ABB Customer case: Metsä Group, Äänekoski bioproduct mill https://new.abb.
com/motors-generators/iec-low-voltage-motors/references/metsa-group (Retrieved
10.12.2019).
ABB, 2019e, ABB Reference case: World’s first industrial application utilizing 5G and AI
adopted at ABB drives factory in Helsinki, https://new.abb.com/news/detail/25464/
worlds-first-industrial-application-utilizing-5g-and-ai-adopted-at-abb-drives-factory-
in-helsinki (Retrieved 10.12.2019).
ABB, 2019f, ABB Ability Marine Pilot Vision, https://new.abb.com/marine/systems-a nd-
solutions/digital/abb-ability-marine-pilot/abb-abilit y-marine-pilot-vision (Retrieved
10.12.2019).
ABB, 2019g, ABB Ability Marine Pilot Vision, https://new.abb.com/marine/systems-a nd-
solutions/digital/abb-ability-marine-pilot/abb-abilit y-marine-pilot-vision (Retrieved
10.12.2019).
208 ◾ Data Analytics and AI
ABB Review, 2019, Unleashing the electrical powertrain through digital monitoring, ABB
Review 4, 52–55.
Accenture, 2019, Client case study – ABB Making machinery speak volume, https://
ww w.accenture.com/us-en/success-making-machiner y-spea k-volumes (Retrieved
10.12.2019).
Bontis, N., Crossan, M.M., and Hulland, J., 2002, Managing and organizational learning
system by aligning stocks and flows, Journal of Management Studies 39(4), 437–469.
CSIS, 2015, Made in China 2025, Center for strategic and international studies, https://ww
w.csis.org/analysis/made-china-2025 (Retrieved 10.12.2019).,
Dalkir, K., 2017, Knowledge Management in Theory and Practice, MIT Press., Cambridge, MA.
Davenport, T., 2005, Thinking for a Living, How to Get Better Performance and Results from
Knowledge Workers. Boston, MA: Harvard Business School Press.
Dixon, N.M., 2000, Common Knowledge – How Companies Thrive by Sharing What They
Know, Boston: Harvard Business School Press,188 p.
Drucker, P., 1994, Post-Capitalist Society, New York: Collins, 240 p.
Fountaine, T., McCarthy, B., and Saleh, T., 2019, Building the AI-poweredorganization,
Harvard Bussines Review 97(July–August), 63–73.
Furr, N., and Shipilov, A., 2019, Digital doesn’t have to be disruptive, Harvard Bussines
Review 97(July–August), 94–103, https://hbr.org/2019/ 07/digital-doesnt-have-to-
be-disruptive.
Hertog, P., 2000, Knowledge intensive business services as co-producers of innovation,
International Journal of Innovation Management 4(04), 491–528.
Laukkanen, I., 2008, Knowledge transfer and competence development in complex paper
production environments, D.Sc. Thesis, TKK Dissertations 103, Helsinki University
of Technology, Espoo.
Li- L., 2018, China’s manufacturing locus in 2025: With a comparison of “Made-in-China
2025” and “Industry 4.0,” Technological Forecasting and Social Change, 135, 66–74.
Liebowitz, J. (ed.) 1999, Knowledge Management Handbook, Boca Raton, FL: CRC Press.
Liebowitz, J. (ed.) 2012, Knowledge Management Handbook: Collaboration and Social
Networking, 2nd edition, CRC Press, Boca Raton, FL.
Liebowitz, J., 2000, Building Organizational Intelligence – A Knowledge Management
Primer, Taylor & Francis., Boca Raton, FL.
Manufacturing USA programme, https://www.manufacturingusa.com/ (Retrieved
10.12.2019).
McGrath, R.G. 2013, The End of Competitive Advantage: How to Keep Your Strategy Moving
as Fast as Your Business, Boston, MA: Harvard Business Review Press.
Moratella, L., 2019, Electric. Digital. Connected., The future platform for the shipping
industry, Mari-Tech conference, Ottawa, Canada.
Nissinen, V., and Laukkanen, I., 2018, Deep leadership and knowledge management:
Conceptual framework and a case study, In KMO 2018 - International Conference on
Knowledge Management in organizations, Springer, p. 48–59.
Nonaka, I., and Takeuchi, H., 1995, The Knowledge-Creating Company: How Japanese
Companies Create the Dynamics of Innovation, New York: Oxford University Press,
304 p.
O’Halloran, D, and Kvochko E., 2015, Industrial internet of things: Unleashing the
potential of connected products and services, World Economic Forum, Switzerland:
Davos-Klosters.
Core Elements in Digital Transformation ◾ 209
Quinn, J.B., Anderson, P., and Finkelstein, S., 1996, Managing professional intellect:
Making the most of the best, Harvard Business Review 74(2), 71–80.
Teece, D.J., Pisano, G., and Shuen, A., 1997, Dynamic capabilities and strategic manage-
ment, Strategic Management Journal 18(7), 509–533.
World Economic Forum 2018, in collaboration with Accenture, “Maximizing the Return
on Digital Investments,” May 2018, http://www3.weforum.org/docs/DTI_ Maximiz
ing _ Return_Digital_W P.pdf (Retrieved 10.12.2019).
World Economic Forum 2019, “HR4.0: Shaping people strategies in the fourth industrial
revolution,” December 2018, http://www3.weforum.org/docs/ WEF_ NES _Whi
tepaper_ HR4.0.pdf (Retrieved 10.12.2019).
Wübbeke, J., Meissner, M., Zenglein, M., Ives, J., and Conrad, B., MADE IN CHINA
2025 The making of a high-tech superpower and consequences for industrial coun-
tries, Merics Papers on China, https://www.merics.org/sites/default/files/201807/
MPOC_No.2_ MadeinChina 2025_web.pdf (Retrieved 10.12.2019).
Chapter 12
AI-Informed Analytics
Cycle: Reinforcing
Concepts
Rosina O. Weber and Maureen P. Kinkela
Contents
12.1 Decision-Making.....................................................................................212
12.1.1 Data, Knowledge, and Information............................................ 214
12.1.2 Decision-Making and Problem-Solving...................................... 215
12.2 Artificial Intelligence...............................................................................216
12.2.1 The Three Waves of AI................................................................216
12.3 Analytics.................................................................................................. 218
12.3.1 Analytics Cycle...........................................................................223
12.4 The Role of AI in Analytics.....................................................................224
12.5 Applications in Scholarly Data.................................................................229
12.5.1 Query Refinement.......................................................................230
12.5.2 Complex Task and AI Method....................................................231
12.6 Concluding Remarks...............................................................................232
References..........................................................................................................233
211
212 ◾ Data Analytics and AI
12.1 Decision-Making
Decisions are the essence of dynamics of all entities. Organizations, people, ani-
mals, and agents all progress into new states through decisions. When you get up
from bed, it is because you made a decision to do so. A soccer player makes a choice
to run a certain direction because of a decision. Organizations rise and fall based on
daily decisions that go from simple to immensely complex. Decisions are triggered
by goals. Decisions lead agents to move closer to goals because they change states
of facts. Productive humans get up to live a fulfilled life, so they make decisions to
get up and go to work. Soccer players want to score, so they decide to seek the ball
and predict where it will be next to reach it.
We adopt the simplified model of decision-making proposed by Simon (1957)
that describes decisions through three steps, namely, intelligence, design, and choice.
These three steps are executed in a cycle. The intelligence step is when the decision-
maker obtains information about the decision to be made. For example, weather
and traffic conditions are information points obtained during the intelligence step
for a decision on how to get to work. Information gathered in the intelligence step
describes attributes of entities associated to the decision context such as the train
is late.
In the design step, the decision-maker enumerates alternative decisions such as
driving, getting a scooter, and taking the train. Finally, the decision-maker selects
one decision such as getting a scooter—this is the choice step.
Figure 12.1 shows these three steps in a cycle. We particularly chose to depict
the decision-making process in a gear cycle because the decision-maker may go
back and forth through these steps in no particular order; we want to emphasize
that changes in any step may impact all other steps.
The implications of the decision-making process for analytics and AI require
further exploration of the process. Figure 12.2 depicts a semantic network of con-
cepts related to the decision-making model originally proposed by Simon (1957)
and later complemented by Huber (1980) who included implementation and moni-
toring. Contextualizing decision-making within problem-solving provides moti-
vation to decisions. Even if not all motivations are problematic, referring to the
context as a problem emphasizes how decisions are part of a process that requires
an action. The action is taken when the decision is made and implementing the
decision means that the information points of the decision may change. After one
or more decisions are made, you may be on your way to work and the soccer player
may be at a different position. These changes in the values of state variables and
attributes of entities represent new information, which characterizes a new context
to trigger new decisions depending on an agent’s goals.
In Figure 12.2, we extend the decision-making and problem-solving process to
give prominence to the concepts that decision-makers take into account when mak-
ing decisions. On the right-side of the network, we show concepts that are directly
considered in each of the steps. This semantic network of concepts should facilitate
AI-Informed Analytics Cycle ◾ 213
Figure 12.1 Proposed gear cycle view of the decision-making process proposed
by Simon (1957).
the comprehension of the process, and, for the purposes of this chapter, helps us
identify all pertinent elements that one has to consider for analytics, as will become
clear later on.
the decision-making steps and these two last steps within problem-solving. Huber
defined decision-making as “the process through which a course of action is cho-
sen” (Huber, 1980, p. 9). The definition of knowledge given by Alavi and Leidner
(2001) is a “justified belief that increases an entity’s capacity for effective action”
(2001, p. 109). Note that both consider the idea that a decision implies an action.
As we did in Figure 12.2, we label this action a process. This process is the one that
has to be effective and this is what decision-makers need to envision to decide on
which alternative decision to adopt.
A process has a goal, a purpose, and an end. The concept of processes is typical in
business organizations and ubiquitous in mundane life. Taking a shower or prepar-
ing a meal are processes. Processes are subsets of projects, i.e., a project entails mul-
tiple processes. For example, organizing a party is a project that entails a series of
processes such as inviting guests, deciding what to serve, planning the venue and its
layout, and so on. These processes require the execution of different complex tasks
such as planning, design, configuration, classification, and so on. These complex
tasks serve as frameworks to execute these processes. They provide context for the
decisions that lead to their execution. These complex tasks are also the core of the
relationship between analytics and artificial intelligence (AI) that we discuss next.
12.2 Artificial Intelligence
AI is a field of study dedicated to the advancement of intelligent agents, which
are those entities that exhibit rational behavior when making decisions through
complex tasks such as planning, classification, recommendation, interpretation,
prescription, and so on. It is very important not to confuse the name of this field
of study with the adjective and noun it contains because “artificial intelligences”
are not really the focus of this computing-related field. The name of the field was
coined at a workshop in 1956 (Moor, 2006). The vision for the field was never to
create any “artificial intelligences,” but simply to replicate the execution of tasks.
Many of the complex tasks they proposed to replicate computationally are typically
performed by humans. This vision of AI would later be perceived as its first wave
(DARPA, 2019).
transactions, filtering spam emails, and enabling robotic vacuum cleaners to func-
tion autonomously.
The first wave AI methods can roughly be categorized as data-oriented, knowl-
edge-oriented, natural language, and evolutionary, as shown in Figure 12.4. Data-
oriented methods typically learn from data and characterize the subfield of AI called
machine learning. Machine learning methods (Russell and Norvig, 2014) include
neural networks, decision trees, naive Bayes and support vector machines, and so on.
These methods require large data sets to work well. Knowledge-based methods include
rule-based reasoning, constraint satisfaction, and ontologies. Case-based reasoning
combines both knowledge-based and machine learning aspects. Knowledge-based
methods require some type of knowledge engineering and maintenance. Natural lan-
guage processing combines techniques to treat sentences and grammar (ibid.), whereas
natural language understanding theorizes the existence of a language model that can
be used to process natural language segments (Allen, 1995). An example of an evolu-
tionary method is genetic algorithms (Russell and Norvig, 2014), which evolve solu-
tion representations. This type of method is usually combined with others for best
results. It requires a reference of quality as a fitness function to guide the evolution.
The Second Wave of AI starts with the increased volume of data produced by
social media, electronic sensors, widespread use of the web, and reduced memory
storage costs. In May 2019, a private statistics company revealed that they com-
puted an average of 500 hours of video uploaded to YouTube by minute (Statista,
2019). This increased volume of data pushes the bar of data-oriented AI methods.
In response to the need to deal with big data, these data-oriented methods become
more robust and start achieving never-before-seen results. Machine learning meth-
ods, particularly neural networks, evolve into much more complex functional net-
works to deal with big data that became known as deep learning.
Figure 12.4 Artificial intelligence and its main subfields as established in the
first wave.
218 ◾ Data Analytics and AI
12.3 Analytics
A widely adopted (e.g., Edwards and Rodriguez, 2018) framework to classify ana-
lytics characterizes it through three technical types (Robinson et al., 2010):
AI-Informed Analytics Cycle ◾ 219
◾◾ Descriptive Analytics. The use of data to figure out what happened in the past.
◾◾ Predictive Analytics. Uses data to find out what could happen in the future.
◾◾ Prescriptive Analytics. Uses data to prescribe the best course of action to
increase the chances of realizing the best outcome.
1 Data are used to record Descriptive Sales and procurement How much did we sell? How much did
transactions. listings: sorted and/or we buy?
subtotaled.
2 Data are used to make Descriptive Budget-versus-actual How much did we sell compared with
comparisons of activity from reports; quarterly and what we thought we would sell?
period to period. annual reports. How much did we buy compared with
what we thought we would buy?
220 ◾ Data Analytics and AI
3 Comparisons are used to make Predictive Comparisons of same fields How much did we sell (or buy) in each of
predictions of future activity. (Study #1) for multiple years with the ten years? Based on this, what will
predictive methods. happen this year?
4 Predictions are used to adjust Prescriptive Budgets, schedules, Since Study #1 predicts that sales will
plans. estimates. drop dramatically in the summer this
year, how do we adjust revenues, costs,
estimates, and production schedules?
5 Data are used to make Descriptive Listings of sales by Do we have data that demonstrate certain
comparisons of activity from (Study #2) demographic and time of demographics tending to buy these
period to period and year. products at unusual times of the year?
demographic to demographic.
(Continued)
Table 12.1 (Continued) Examples of Analytics Considering Motivating Questions, Purpose, and Type of Analytics
Technical
Row Purpose of data use type Products Motivating question
6 Studies are combined in order to Prescriptive Advertising campaigns. Since Study #1 predicts that sales will drop
arrive at an advertising strategy. dramatically in the summer this year and
Study #2 demonstrates that certain
demographics buy our product at unusual
times, what is our advertising strategy?
7 Studies are combined in order to Predictive Predictive report. What do we predict will be the results of
predict the results of the (Study #3) this advertising strategy?
advertising strategy.
8 Data are used to identify Predictive Predictive report. Study #3 demonstrates a 25% uptick in
demographic factors which (Study #4) sales in areas where this type of
might affect the success of our advertising strategy is employed. What
advertising campaign in various factors will determine if this will be the
areas. effect of an advertising campaign in
certain demographics? Based on studies
of these factors, what do we predict will
be the effect of our advertising strategy
in various areas?
9 Heuristics are applied to ascertain Predictive Predictive report. Based on politics, world conditions, and
the wider effect of current (Study #5) the economy, do you think this
events on these results. prediction will come true this year?
10 New predictions used to adjust Prescriptive Updated budgets, Based on the results of Study #5, adjust
planning documents. schedules, estimates. budgets, schedules, and estimates to
account for a predicted increase over
AI-Informed Analytics Cycle ◾ 221
ability to simulate budgets and business activity through plans can provide a more
realistic measure of the impact of variations. Finally, even more important than
answering the decisions, “How much did we sell compared with what we thought
we would sell?” and “How much did we buy compared with what we thought we
would buy?”, would be to answer, “How much more or less do we have to sell or
buy to meet our goals?”
The step described in the third row is predictive. It answers the decision, “What
will this year or next year look like?” By examining historical data, regression mod-
els can suggest what the future may look like. For truly effective prediction, it
may be beneficial to combine purely data-oriented methods based on statistics with
knowledge-based methods that incorporate judgement. This activity is more com-
plex in the sense that it usually involves combining data from multiple periods.
The logical next step in the evolution is prescriptive. Based on results from the
previous step, the goal is to update plans such as budgets, schedules, and estimates.
The decision is, “How do we use the prediction in the previous step to make these
adjustments?” This step is more complex than the previous ones, both from a pro-
cessing standpoint and from a human judgement standpoint. Here again analytics
benefits from automated plans to simulate business activity.
These analytics discussed in the first four rows in Table 12.1 are a type of cycle
that we often see in business: examine the past, predict the future, simulate its
impact, and adjust plans to accommodate the prediction and steer the business
to its goals. Another cycle can be: examine the past and make plans to accommo-
date the information, without real predicting being necessary. For example, when
a budget-versus-actual report shows that we are spending too much, often the next
step is to figure out where to cut spending. Predictive informatics may not be used
in this case. We don’t have to predict—we already know we are over budget, and
by exactly how much. It would, however, be fundamental to find out why. This is
where methods from the third wave will help.
On the fifth row, we go back to a descriptive analytics goal. A two-step cycle is
what happens next in Table 12.1. This goal arises because planners looking at the
prediction, which came out of the step in the third row, and arrived at another piece
of descriptive information needed. The decision associated with this query was,
“Are there customers in any demographics who make purchases at an unusual time
of year?” This report can be produced through data warehousing, or pivot tables,
which sets up the data to count based on an examination of more than one field
simultaneously. The goal of finding this information arose from a problem-solving
activity of human planners trying to figure out how to advertise during the low
sales months. With the advances of the second wave, deep learning networks can
combine multiple predictions and conditional functions to automate such analyses.
As a result of the step in the fifth row, the step in the sixth row is prescriptive
and answers the decision, “How do we change our advertising to accommodate
the demographical information arrived at in the fifth row?” Planning the advertis-
ing strategy is a complex human problem-solving activity and could likely involve
AI-Informed Analytics Cycle ◾ 223
examining other data elements in addition to the sales and procurement tables we
have looked at thus far. It could spin off a whole new cycle of descriptive, predictive,
and prescriptive analytics. The simulation of an advertising strategy requires careful
contemplation of highly sensitive variables. With the speed at which information
spreads with social media, automated methods are required to keep track of their
impact. One area to consider for this tracking used in textual analytics is sentiment
analysis (Liu, 2015).
Steps in rows 7, 8, and 9 are predictive steps which examine the new advertis-
ing plan in increasingly complex ways. The decisions are, “Will this new strategy
be effective? Will the effectiveness of the new strategy be altered in various demo-
graphics? Will the effectiveness be altered based on possible future changes in the
status quo?” These steps are complex because they can involve a combination of
steps considering what has happened in the past as indicated by the data, predict-
ing what it implies using both human and statistical analysis, and using human
intuition to analyze the overall meaning of this combination of information items.
It is always important to remember that prediction methods are only as good as
the data they use. For effective predictions, high quality data has to be used that
contemplates all variations of trends and directions.
The users having arrived at some decisions as a result of the previous steps, Step
10 is again prescriptive. The results can help us update plans to simulate budgets,
schedules, and estimates, now with greater confidence based on the analytics we
have engaged in.
12.3.1 Analytics Cycle
Process cycles represent steps to follow in order to implement methodologies. Such
cycles are useful to guide practitioners and to help create mental models of such
methodologies. As disciplines evolve, they earn new names and their process cycles
evolve. Analytics is no exception. It evolved directly from the knowledge discovery
from databases (KDD) process, which was popularized by publications such as
Fayyad et al. (1996). Despite the goal of their vastly referenced article, which was
to distinguish data mining from KDD, the term data mining became the catch-all
term for KDD.
As the world and its technologies evolved and the amount of data involved grew
exponentially, we entered a new era, the era of data science and data analytics. The
business world and academia have advanced along very similar lines in their respec-
tive adoption and investigation of data technologies. The business world adopted
the term business intelligence (Pearlson et al., 2019) for the general set of methods
and capabilities that enable businesses to learn and advance from data. In aca-
demia, the analogous term is data science, the field of study dedicated to advancing
the set of methodologies and capabilities that produce new knowledge from data.
On the other hand, the term analytics is rather reserved for the process that
is driven by a question or query, similar to the motivation of KDD. Earlier, we
224 ◾ Data Analytics and AI
explained that decisions are what drives business. Business analytics (Pearlson
et al., 2019) is hence the analytics process driven by the need for decision-making.
We refer to the process as analytics because we describe it to be used beyond the
business context.
The KDD process proposed by Fayyad et al. (1996) consists of these steps:
Selection, Preprocessing, Transformation, Data Mining, and Interpretation/
Evaluation. Delen (2018) adopts the Cross-Industry Standard Process for Data
Mining (CRISP-DM) for data analytics, which includes Business Understanding,
Data Understanding, Data Preparation, Model Building, Testing and Evaluation,
and Deployment. We could keep going as there are multiple variations of analyt-
ics cycles. Based on a thorough examination of various cycles (EMC, 2015), an
overview cycle consists of the steps: Discovery, Data Preparation, Model Planning,
Model Building, Communicate Results, and Operationalize. EMC (2015) appro-
priately suggests arrows going back to the previous step of the cycle as a way to
remind data analysts to move back a step and reconsider the results of the previous
steps in order to promote success. Due to our educational focus, in our proposed
cycle, we once again utilize the gear cycle representation to emphasize that all previ-
ous steps may be revisited if any change in direction becomes necessary. Given the
presence of a motivating question or query in analytics, we rename Discovery with
Query Refinement. We also adopt the final two steps from the KDD process and
use Testing and Evaluation, and Deployment, as shown in Figure 12.5.
when a query is well defined; to build this needed experience, we propose some steps
that can be considered for refining queries. One step is to eliminate ambiguous or
uncertain terms. In question-answering systems such as IBM Watson (Lally et al.,
2012), questions are analyzed, and their focus is identified. The focus characterizes
the object of the answer such as questions starting with who have a person as focus
and questions starting with when have a date as focus. Referring to Table 12.1,
questions starting with how much have amounts as focus. See the motivating ques-
tion in the fifth row in Table 12.1. That question starts with do we, but it is actually
asking about trends of demographics in association with time periods. Other ques-
tions such as those in the sixth to eighth rows imply the need of a plan. When we
propose a query refinement step, we suggest that these questions are stated in ways
that make the focus clear. The other components that need to be clearly inferred are
the technical type of analytics, the purpose of the data, and the products required
to answer the questions.
226 ◾ Data Analytics and AI
Complex Task. This step is crucial for successfully using AI methods. From
the motivating question, the data, the technical type, and products, and a clear
complex task, must be identified. Figure 12.7 shows a hierarchy of analysis and
synthesis tasks as they are usually discussed in AI. Note, however, that these tasks
and their descriptions within this section are limited to structured data. Later on,
we discuss unstructured data that relies on textual analytics.
The main categorization of complex tasks groups them as analysis or synthesis
tasks. The general idea of analysis tasks is that they analyze an input to produce a
result. Analysis types could be further categorized as classification and regression.
Classification tasks examine the input to determine the class to which the input
belongs. For example, the input is an image and the classification result is that
the input is an image of a cat. The clue to recognize a classification task is that the
classes are given in the input. This means that the input given includes the result,
also called the label. For this reason, this classification task will be later imple-
mented as supervised learning, i.e., because it includes the label.
AI-Informed Analytics Cycle ◾ 227
that includes the most popular methods includes decision trees, support vector
machines, random forest, case-based reasoning (Richter and Weber, 2013), and
naive Bayes (Russell and Norvig, 2014). Note that these methods are known to
execute supervised learning, which means that the input data used includes infor-
mation of the class to which each instance belongs. Classification can also be per-
formed with rule-based reasoning, but this is not a data-oriented method, which
means it is not amenable for large data sets and is not considered within machine
learning.
When data is not labeled with a class, then the learning is called unsupervised. The
most typical method is known as clustering (Aggarwal and Reddy, 2014). There are
various types of clustering algorithms to choose from and selecting one will depend
on specific characteristics of the problem. One of these characteristics is whether the
number of classes is known or can be guessed and tested. Note that when clustering is
used, the complex task may not be classification, but a step toward it. The actual tasks
are called clustering or categorization, which means to identify categories.
Various AI methods execute the planning task (Ghallab et al., 2004; LaValle,
2006; Richter and Weber, 2013; Russell and Norvig, 2014). Some of the methods
that execute planning are informed search algorithms such as A* and Dijkstra, and
methods such as rule-based reasoning, case-based reasoning, and hierarchical task
networks. Informed and uninformed search algorithms are also used for design,
scheduling, and configuration tasks, as well as rule-based reasoning, case-based
reasoning, and constraint satisfaction.
Recommendation can be executed with multiple methods. There are five major
types of recommender systems, of which content-based and collaborative filtering
are the most popular. There are also utility-based, knowledge-based, and demo-
graphic recommender systems. Content-based methods are typically implemented
with case-based reasoning (Richter and Weber, 2013). Collaborative filtering rec-
ommenders are implemented with matrix factorization, which may use neural net-
works (Koren and Volinsky, 2009).
This step of the analytics cycle does not really end with the selection of the AI
method. Once the method is selected, it becomes clear as to how the data are to
be processed. Each of these methods require very different data preparation and
processing. We recommend further reading for those methods.
12.5.1 Query Refinement
We start by looking for ambiguity and uncertainty in the question, and then we
analyze its focus, analytics technical types, and products. The adjective new is
rather vague. Keeping this in the question forces us to define what new encom-
passes. The rule-of-thumb here is to determine the implications of removing the
uncertain term. By keeping the adjective new, the question becomes about jobs that
will exist and that did not exist before. By removing the adjective new, the question
Question What are the new jobs that will be most frequently offered
Editing within ten years or so?
What are the jobs that will be most frequently offered within
ten years or so?
What are the jobs that will be most frequently offered within
five to ten years?
Products What are the advances in all fields in the last five to ten years?
What are the topics contextualizing the advances in all fields
in the last five to ten years? What are the most frequent
topics contextualizing the advances in the last five to ten
years?
AI-Informed Analytics Cycle ◾ 231
is simply about jobs that will be sought in the future, regardless of whether they
existed before or not. Of course, one alternative would be to conduct analytics for
both, but for simplicity, let us remove the uncertain word.
The second uncertain expression is ten years or so. Based on comprehension of
the data, which is one of the requirements for analytics, we use the assumption
that advances reported in scholarly publications are about five to ten years ahead of
their effective use in society as a job skill. Consequently, this information finds the
match between the data and the question, which is that at the core of what we need
to determine is in the advances discussed in scholarly data. We propose therefore to
replace the expression ten years or so with five to ten years.
Considering that there are no uncertain expressions remaining in the ques-
tion, we move to its focus. When asking for jobs, what level of specificity are we
looking for? Is the question asking for job titles or industries that will be offering
them? Again, we rely on comprehension of the data and consider that although
scientific advances can give us details on what the processes would entail in differ-
ent jobs, they do not tell us job titles. Consequently, we can affirm the focus to be
about fields or areas of study, and not information specific to the detailed processes
entailed in the advances. In fact, in order to make any use of the answer, it would
not be beneficial to get a detailed list of job descriptions, but areas of study or topics
that advances may have in common.
The question we pose requires predictive analytics, as it is a prediction task. In
order to answer it, we need to consider which products we need for this prediction.
The relationship between jobs and advances suggests that we need a descriptive
analytics report with the advances in all fields in the last five to ten years. Note that
the products require analytics of other technical types. We are interested in the top-
ics contextualizing those advances, so we can determine which advances share the
areas and will ultimately become the most frequently offered. Therefore, the other
product needed is a report of the topics contextualizing those advances. Then, we
will need to find common topics across various advances so we can identify the
most frequent fields. The idea is to find common trends of topics across different
fields that would suggest a direction or trend. It would be our assumption based on
the comprehension of scholarly data that the most frequent jobs in the next five to
ten years will be related to the most frequent topics contextualizing the advances in
the last five to ten years.
understanding, which follows a very different structure. In NLU, the primary con-
cern is to learn a language model. The language model is what guides the decisions
on how to execute multiple tasks such as search, or comparisons between words,
segments, or documents. This chapter is limited in scope and does not cover textual
methods, which would require another entire chapter. We briefly mention some of
these methods in this illustration for the reader to appreciate the extent as well as
the limitations of AI methods at the time this chapter was published.
We discuss the tasks following the sequence of questions. The first is, “What are
the advances in all fields in the last five to ten years?” This question still needs to
be further broken down to the question, “What are the advances described in this
article?” With the answer from each published article, then the advances from the
last five to ten years can be collected, and then their topics can be identified, fol-
lowed by the advances that share topics. Unfortunately, this reaches a limit to NLU.
Very recently, Achakulvisut et al. (2019) reported on an extractor that demonstrates
this is not a task that methods from NLU are ready to perform.
Notwithstanding, we will continue our analysis as if this first step were feasible.
Assuming we had a report with all the advances from all publications in the last five
to ten years, this report would be a corpus of sentences, where each article’s contri-
bution would be likely reported in a set of sentences. To examine the feasibility of
this task, we start from the estimate that all scholarly peer-reviewed journals pub-
lish a total of 2.5 million articles a year (Ware and Mabe, 2015). We estimate that
each description of an article’s core advance, commonly referred to as the article’s
contribution, could, on average, be described with a number of words similar to
that of an abstract, which ranges from 15 to 400 words. This way we can consider
an average of 300 words for each contribution for five years of 2.5 million articles
per year, which amounts to 3 billion words. Fortunately, this volume is similar to
the volume of words used to learn a language model in BERT (2018). BERT is the
first of a class of algorithms that can learn contextual word representations in a
language model. Note that there are other algorithms that learn contextual repre-
sentations (e.g., Yang et al., 2019). In fact, progress in NLU comes quickly, making
it difficult to keep up with this field.
In summary, if it were possible to effectively extract the advances described in
scholarly publications, then we would be able to categorize their topics and find out
what the most frequently offered jobs would be in ten years or so. In fact, had we
overcome this impediment, the industry today would not have unfilled positions
for data scientists.
12.6 Concluding Remarks
This chapter presents the role of AI in analytics and suggests applications in schol-
arly data. The chapter aims to emphasize the role of AI methods which are so far
unexplored in analytics. With this purpose, it proposes an AI-informed analytics
AI-Informed Analytics Cycle ◾ 233
cycle. The analytics cycles provided in this chapter are depicted with gear cycle dia-
grams because the authors would like to emphasize the fact that changes in any of
the steps are likely to impact all other steps. In practice, it means that any change
in direction will require practitioners to revisit questions, purposes, and products.
Analytics is currently being practiced using AI methods from the second wave
of AI, namely, machine learning methods. These methods are recommended to
deal with large and complex data sets and are required to learn knowledge from
data. There are, however, more methods in AI to be used in the creation and auto-
mation of the intermediary products required in analytics, such as those to cre-
ate plans, budgets, and simulate potential results of different business strategies
to improve decision-making. Harnessing AI methods beyond machine learning is
likely to improve the potential of analytics.
This chapter presents the application of analytics in scholarly data, which is
unstructured data. However, the descriptions of AI methods and complex tasks are
focused on structured data. The combination of contents should provide a good
overview of AI-informed analytics, but readers are encouraged to expand their
knowledge with the various sources referenced herein.
References
Achakulvisut, T., C. Bhagavatula, D. Acuna and K. Kording. 2019. Claim extraction in
biomedical publications using deep discourse model and transfer learning. arXiv pre-
print. https://arxiv.org/abs/1907.00962.
Aggarwal, C. C. Ed. 2015. Data classification: algorithms and applications. Boca Raton:
CRC Press.
Aggarwal, C. C. and C. K. Reddy. 2014. Data clustering: algorithms and application. Boca
Raton: CRC Press.
Alavi, M. and D. E. Leidner. 2001. Knowledge management and knowledge management
systems: Conceptual foundations and research issues. MIS Quarterly 25(1):107–136.
Allen, J. 1995. Natural language understanding. New York: Pearson.
Bishop, C. M. 1995. Neural networks for pattern recognition. New York: Oxford University
Press.
DARPA. 2019. Defense advanced research projects agency. AI Next Campaign. https://www.
darpa.mil/work-with-us/a i-next-campaign (accessed: February 18, 2019).
Delen, D. 2018. Data analytics process: An application case on predicting student attri-
tion. In Analytics and knowledge management, ed. S. Hawamdeh and H. C. Chang,
pp. 31–65. Boca Raton: CRC Press.
Devlin, J., M-W. Chang, K. Lee and K. Toutanova. 2018. Bert: Pre-training of deep bidi-
rectional transformers for language understanding. arXiv preprint. https://arxiv.org/
abs/1810.04805.
Edwards, J. S. and E. Rodriguez. 2018. Knowledge management for action-oriented ana-
lytics. In Analytics and knowledge management, ed. S. Hawamdeh and H. C. Chang,
pp. 1–30. Boca Raton: CRC Press.
EMC Education Services. 2015. Data science & big data analytics: discovering, analyzing,
visualizing, and presenting data. Indianapolis: John Wiley & Sons.
234 ◾ Data Analytics and AI
Fayyad, U., G. Piatetsky-Shapiro and P. Smyth. 1996. From data mining to knowledge
discovery in databases. AI Magazine 17(3):37–54.
Galloway, S. 2017. The four: the hidden DNA of Amazon, Apple, Facebook and Google. New
York: Portfolio/Penguin.
Ghallab, M., D. Nau and P. Traverso. 2004. Automated planning: theory and practice. New
York: Elsevier.
Goodfellow, I., Y. Bengio and A. Courville. 2016. Deep learning. Cambridge: MIT press.
Holmen, M. 2018. Blockchain and scholarly publishing could be best friends. Information
Services & Use 38(3):1–10.
Huber, G. P. 1980. Managerial decision making. Glenview: Scott Foresman & Co.
Koren, Y., R. Bell and C. Volinsky. 2009. Matrix factorization techniques for recommender
systems. Computer 8:30–37.
Lally, A., J. M. Prager, M. C. McCord, B. K. Boguraev, S. Patwardhan, J. Fan, P. Fodor
and J. Chu-Carroll. 2012. Question analysis: How Watson reads a clue. IBM Journal
of Research and Development 56(3.4):1–14.
LaValle, S. M. 2006. Planning algorithms. Cambridge: Cambridge University Press.
Lim, A., A. J. Chin, H. W. Kit and O. W. Chong. 2000. A campus-wide university exami-
nation timetabling application. In AAAI/IAAI, ed. R. Engelmore and H. Hirsh, pp.
1020–1015. Menlo Park: AAAI.
Liu, B. 2015. Sentiment analysis: mining opinions, sentiments, and emotions. Cambridge:
Cambridge University Press.
Martins, J. P., E. M. Morgado and R. Haugen. 2003. TPO: A system for scheduling and
managing train crew in Norway. In IAAI, ed. J. Riedl and R. Hill, pp. 25–34. Menlo
Park: AAAI.
Moor, J. 2006. The Dartmouth College artificial intelligence conference: The next fifty
years. AI Magazine 27(4): 87–91.
Pearlson, K. E., C. S. Saunders and D. F. Galletta. 2019. Managing and using information
systems: A strategic approach. Hoboken: John Wiley & Sons.
Richter, M. M. and R. O. Weber. 2013. Case-based reasoning: a textbook. Berlin:
Springer-Verlag.
Russell, S. J. and P. Norvig. 2014. Artificial intelligence: a modern approach. Upper Saddle
River: Prentice Hall.
Robinson, A., J., Levis, and G. Bennet. 2010. INFORMS news: INFORMS to officially
join analytics movement. OR/MS Today 37(5).
Simon, H. A. 1957. Models of man; social and rational. Hoboken: John Wiley & Sons.
Statista. 2019. Hours of video uploaded to YouTube every minute as of May 2019. https://
ww w.stat ista.com/statistics/259477/hou rs-of-video-uploaded-to-youtube-every-m
inute/ (accessed: August 16, 2019).
Ware, M. and M. Mabe. 2015. The STM report: An overview of scientific and scholarly
journal publishing. University of Nebraska – Lincoln. Technical Report. https://
di g ital c ommo n s.un l.edu /cgi/ v iewc onten t.cgi ?arti c le=1 0 08& c ontex t= sch olcom
(accessed March 16, 2018).
Weber, R. O. 2018. Objectivistic knowledge artifacts. Data Technologies and Applications
52(1):105–129.
Yang, Z., Z. Dai, Y. Yang, J. Carbonell, R. Salakhutdinov and Q. V. Le. 2019. XLNet:
Generalized autoregressive pretraining for language understanding. arXiv preprint.
https://arxiv.org/abs/1906.08237.
Index
235
236 ◾ Index
magnetic resonance imaging (MRI), 169 Diagnostic analytics, 23, 152, 153
positron emission tomographic (PET) Dictionary-based methods, 126
patterns, 169 Diffusion tensor imaging (DTI), 169
Deep learning, in industry Digital powertrain, 202–203
architectures, 67–68 Digital transformation, 184–186
content generation ABB Smartsensor, 199, 202
audio generation, 78–79 ABB takeoff program, Industry 4.0
image and video generation, 79–81 background, 194, 195
text generation, 76–78 IIoT and digital solutions, value
decision-making, 81–82 framework, 194, 196, 197
automatic game playing, 83–85 intelligent industry takeoff, 194, 196,
autonomous driving, 82–83 198–201
energy consumption, 86–87 autonomous ships, 203–204
online advertising, 87–88 digitalization and AI, challenges of, 187, 188
robotics, 85–86 digital powertrain, 202–203
deep generative models, 69 future recommendations, 205–207
deep reinforcement learning, 69–70 integrated model, 191, 193–194
forecasting objectives, 186–187
financial data, 90–91 people, competency, and capability
physical signals, 88–90 development, critical roles, 207
overview of, 66–67 research approach, 186–187
recognition, 70–71 theoretical framework
in audio, 72 framework competency, capability, and
in text, 71 organizational development, 190
in video and image, 72–76 knowledge management and learning
DeepMind Atari game, 84 agility, data collection, 187, 189
Deep neural networks (DNNs), 7–9, 16, 72, 87, knowledge processes in
124, 169 organizations, 189
Deep Q-networks (DQNs), 83–85 transient advantages, 190–192
Deep reinforcement learning (DRL), 11, 69–70, Digital transformation (DX) initiatives,
81–83, 86 22, 28, 29
DeepStereo, 81 Disengage, 191
Deep Tesla, 83 DNNs, see Deep neural networks
Deep text models, 71 Document-level SA, 126
DeepTraffic, 84 DQNs, see Deep Q-networks
Dental anxiety DRL, see Deep reinforcement learning
assessments, 160 DSAI, see Data science artificial intelligence
odontophobia, 157 DSLC, see Data science life cycle
procedure type, 158 DTI, see Diffusion tensor imaging
treatment modalities, 158 DX, see Digital transformation initiatives
women vs. men, 158
Dental avoidance Echo state network (ESN), 88
barriers, 158 EHRs, see Electronic health records
conservative dental restoration Electronic dental records (EDRs)
techniques, 159 BigMouth Dental Data Repository, 155
pain, 158 clinical decision support systems, 152
poor treatment outcome, 159 data errors, 154
treatment fear, 159 data quality, 154
Depot of Charts and Instruments, 2 dental schools, 152, 154, 155
Descriptive analytics, 12, 23, 152, 153, 219 diagnostic and predictive analytics, 154
Deterministic environments, 38 electronic health records (EHRs), 152
DFDC, see Deepfake detection challenge healthcare data, 154
Index ◾ 239