Managing The Machines: AI Is Making Prediction Cheap, Posing New Challenges For Managers
Managing The Machines: AI Is Making Prediction Cheap, Posing New Challenges For Managers
Managing The Machines: AI Is Making Prediction Cheap, Posing New Challenges For Managers
7 October 2016
Introduction
We are living through a renaissance in artificial intelligence (AI). The major tech
companies (e.g. Google, Facebook, Amazon, Microsoft, Tesla, and Apple) are
prominently including AI in their product launches. They and others are acquiring AI-
based startups by the dozen. Depending on how you look at it, this conjures up images
that range from augmentation of human labor to replacement with the consequent
follow-on logic of excitement to fear. In each case, proponents point to an advance —
an AI that plays Go or one that can drive a truck — and define a future path by
extrapolating from the human tasks enhanced or replaced. But economic history has
taught us that this is a flawed approach. Instead, when looking to assess the impact of
radical technological change, one approach stands out: ask yourself, what is this
reducing the cost of? It is only then that you can figure out what really might change.
To understand how important this framing can be, let’s step back one technological
revolution ago and ask that same question. Moore’s Law — the 18-month doubling of
transistor intensity on microprocessors — has dominated information technology for the
past four decades. What did those advances reduce the cost of? The answer:
arithmetic.
This answer may seem surprising as computers appear to do much more. They
allow us to communicate, to play games and music, to design and to create art. This is
all true but at their heart computers are direct descendants of electronic calculators.
That they appear to do more is testament to the power of arithmetic. In their earliest
1 Rotman School of Management, University of Toronto and NBER. We thank James Bergstra,
Tim Bresnahan, and Graham Taylor for helpful discussions. All views remain our own.
2
days, this relationship was more obvious. Computers focused on arithmetic operations
related to the census, artillery tables, and other largely military applications. Prior to the
invention of the digital computer, “computers” were humans who spent their days
solving arithmetic problems related to these applications. What digital computers did
was make arithmetic so inexpensive that thousands of new applications for arithmetic
were implemented and discovered; data storage, word processing, and digital
photography are all novel applications of arithmetic.
AI presents a similar opportunity: to make something that was once expensive very
cheap, and with it, to make resources that were once scarce now abundant. For AI, that
particular task is prediction. That is, the ability to take information you have and
generate information you do not have.
To reduce the economics of AI down to a single thing may seem like a tall order. To
be sure, for many who have looked at the emerging applications of AI, our take may
seem far from obvious. The purpose of this article is, however, to convince you
otherwise. What we will argue here is that AI — in its modern incarnation — is all about
advances in prediction. Based on this thesis, we will then explore the impact of
advances in AI on the nature and direction of applications coming from it: how AI will be
used as an input for traditionally non-prediction-oriented problems, how the value of
some human skills will rise while others fall as AI advances, and the implications for
managers. These speculations will be informed by earlier radical technological changes
that similarly involved the near elimination of the costs of particular tasks. As we will
demonstrate, this allows us to understand how AI is likely to change what workers and
managers do.
This approach to learning is not optimal for all knowledge. But it does turn out to be
a highly efficient process for understanding what outcomes will result from a large
number of quantifiable observations. Consider the job of identifying objects in a basket
of groceries. If we humans could describe what, say, an apple looks like and compare
that to an orange, then we could program a computer to recognize both objects; say
based on some color and shape classifications. But a basket of groceries includes other
objects some of which are apple-like in colour and shape. It may be possible to continue
encoding our knowledge of apples in finer detail but as we move towards environments
in the real world, potential complexity increases exponentially. In this respect, one starts
to appreciate just how hard these jobs can be and how amazing it is that humans can
do them so easily.
It is in environments with this type of complexity that machine learning is most
useful. In one type of training, the machine starts with a set of objects that it is shown
pictures with names attached to them. It is then shown millions of pictures that may or
may not include those objects but each with objects named. As a result, it notices
correlations. For instance, that objects named “apples” tend to be red. Of course, if the
apple is sometimes eaten or half-eaten that may be trickier. But other objects may be
red so the machine looks for other correlates within its range of classifiers — shapes,
texture, and then most importantly context. The final one requires some sophisticated
machine learners, but a red round object is more likely to be an apple in a basket of fruit
than in a ball pit at a play centre.
What is happening under the hood is that the machine uses information from past
images of apples to predict whether the current image contains an apple. Why use the
word “predict”? Prediction uses information you have to generate information you do not
have. Machine learning uses data collected from sensors, images, videos, typed notes,
or anything else that can be represented in bits. This is information you have. It uses
this information to fill in missing information, to recognize objects, and to predict what
will happen next. This is information you do not have. In other words, machine learning
is a prediction technology.
There are a large variety of possible predictions. We most often think about
prediction as determining what will happen in the future. For example, machine learning
can be used to predict whether a bank customer will default on a loan. There is also a
great deal of useful data we do not have that is not about the future. For example, mass
retailer Target predicted which of its customers were pregnant based on their
purchasing behavior. This was not predicting the future: The customers were already
pregnant. Instead, it was filling in missing data in a way that proved useful to the
company. Similarly, medical diagnosis is a prediction problem: using data on symptoms
4
images can now be recognized with much less time and money required. Image
classification is thus an automated task. Similarly, fraud detection in banking is moving
from a costly human-based process to a much less expensive and more easily scaled
machine-based process. In this sense, recent advances in machine learning have led to
a dramatic decrease in the cost of prediction.
Hawkins’ work is controversial. His ideas are debated in the psychology literature and many
computer scientists flatly reject his emphasis on the cortex as a model for prediction machines.
Irrespective of whether the underlying model is appropriate, his emphasis on prediction as the basis for
intelligence is useful for understanding the impact of recent changes in AI. In particular, the falling cost
of prediction due to advances in machine learning has meant that many tasks previously considered
unique to humans can now be done by machines, including driving, language translation, playing Go,
and image classification.
To the extent that our brains really are “organs of prediction,” then advances in AI may indeed lead
to increased substitution of humans for machines. However, to the extent that Jeff Hawkins’ model of
humans is incorrect - that instinct, or some other process besides prediction, drives human behavior -
then humans have an advantage in efficiently undertaking some tasks because they seem to have
better algorithms for converting sensory input (data) into actions. It requires much more data for
machines to overcome their disadvantage in algorithms. Machine learning remains far less efficient at
converting data into predicted actions than the human brain. A teenager learns to drive with tens or
hundreds of miles of practice. Automated driving systems are given millions of miles of data and still
periodically require human intervention.
6
doing they apply judgment in combination with the prediction. While an experienced
driver may not learn much from their own behavior, an inexperienced one will see how
their action led to an outcome and then, through a process of feedback, use that
information to inform their future predictions.
Seen in this light, it is useful to distinguish between the value and the cost of
prediction. As we have mentioned, AI advances have lowered the cost of prediction.
Given data and something to predict, it is now much less costly to derive an accurate
prediction. But of equal importance is what has happened to the value of prediction. Put
simply, prediction has more value in a task if data is more widely available and
accessible. The many decades-long improvements in the ability of computers to copy,
transmit, and store information have meant that the data available for such predictions
has also improved, leading to a higher value of prediction in a wider variety of tasks.
Consider autonomous driving. Engineers have been able to make cars that could
drive themselves in specific environments for decades. The modern era of autonomous
driving began in the 1980s. The US and Germany were the two nations at the forefront
in this line of research. In the US, the research was largely funded by DARPA (Defense
Advanced Research Projects Agency). In Germany, large automotive companies such
as Mercedes-Benz funded research. The leading projects utilized computer vision-
based systems, lidar, and autonomous robotic control. The decision-making systems
were essentially driven by optimizing if-then-else algorithms (e.g., optimize speed
subject to not exceeding the speed limit and not hitting anything; “if is raining, then slow
down…”, “if a pedestrian approaches within 5 feet on the left, then swerve right”). In
other words, the systems were algorithmic, that is, codifying an algorithm describing the
connection between road conditions and decisions related to speed and steering.
However, to be able to drive on roads in unstructured and unpredictable
environments, including ones where other drivers (human or autonomous) also exist,
requires predicting the outcomes of a large number of possible actions. Codifying the
set of possible outcomes proved too challenging. In the early 2000s, however, several
groups began using primitive versions of modern machine-learning techniques. The key
difference between the new method and the old method is that while the old method
centered around optimizing a long list of if-then-else statements, the new method
instead predicts what a human driver would do given the set of inputs (e.g., camera
images, lidar information, mapping data, etc). This facilitated significant improvements in
autonomous driving performance.
8
Who Judges?
The ultimate outcome of any task comes from the action that is taken. As already
emphasized, one input to what action gets taken is prediction. However, as depicted in
Figure 2, what we call “judgment” has a distinct role in determining an action. Judgment
is the ability to make considered decisions. In other words, to determine the impact
different actions have on outcomes given predictions. As it turns out, this hinges directly
on how clear outcomes themselves are. For some tasks, the precise outcome that is
desired can be easily described. For instance, in terms of labeling images, you want the
label to be accurate, which it either is or is not. In these situations, the separate need for
judgment, as might be applied by a human, is limited and the task can be largely
automated.
In other situations, describing the precise outcome is difficult. It resides in the mind
of humans and cannot be translated into a form a machine can understand. This is why
it is often argued that AIs may be less adept at handling emotional tasks. However,
machine prediction can significantly impact more pedestrian tasks. For instance, an AI
that maps out the optimal route to take to minimize travel time (such as might occur in
apps like Waze) cannot easily take into account the preferences of drivers who have to
do the work of driving. For instance, the AI may find a route with many turns that
reduces travel by a few seconds but a human driver may prefer a slower route with
fewer turns or one that takes them closer the dry cleaning store where they
remembered they had clothing to pick up. Of course, a clever engineer will resolve
these particular issues. But the point here is that the outcome that is desired often has
indescribable elements that require the exercise of judgment to mitigate and absorb.
Judgment takes predictions and uses them as information that is useful to determine
actions. Hence, alongside prediction, judgment is a critical input into many actions.
Whether a machine can undertake the action depends on whether the outcome can be
described in such a way that the machine can exercise suitable judgment.
This is not to say that our understanding of human judgment cannot evolve and be
automated. One of the features of new modes of machine learning is that they can
examine the relationship between actions and outcomes and use this feedback to
further refine predictions. In other words, prediction machines can learn. This dynamic
aspect drives improvements in how a task is performed. For instance, machines may
learn to predict better by observing how humans perform in tasks. This is what
DeepMind’s AI AlphaGo did when learning the game of Go. It analyzed thousands of
human-to-human games and then played itself millions more games, each time
9
In other contexts, however, more abundant prediction could lead to increased value
of human-led tasks, that is, where the machine provides predictions but the human
exercises judgment and undertakes the action component of a task. For example,
Google Inbox’s autoreply processes your incoming email and predicts several short
responses. It then lets the human judge which is the most appropriate. Selecting
between one of a handful of actions is much quicker than composing a reply, enabling
the human to respond to more emails in less time.
Autonomous driving is another example. While autonomous driving is an area
where new developments in AI are seen as taking humans out of the picture, many of
the most imminent advances still keep humans in the driver’s seat. For instance, many
car companies are using AI to predict when a collision might occur. Most of the time, all
is clear. However, when an issue arises, the machine alerts the human who then
exercises judgment as to the course of action they will take to avoid the collision. While
many have argued that machine reactions may be superior in such situations, it is
precisely because the full dimensions of the outcome cannot be easily described that
humans remain in control. In particular, human control arises when judgment requires
decisions not of whether to avoid harm altogether but what type of harm to avoid and
when situations are sufficiently unusual that judgment is required to correct for
unreliable prediction. Combined with human judgment, prediction has made human-led
driving safer and so increased the value of human-led driving overall.
Another example is medical diagnosis. Artificial intelligence is improving diagnosis.
While this may reduce the human role in diagnosis, treatment and patient care require
judgment. Each patient may have different needs that are difficult to define for the
machine. More effective diagnosis is likely to lead to increased use of effective
treatment and better patient care.
Overall, human judgment is particularly important when it is difficult to describe
outcomes in a manner that can be understood by machines. Part of this inability comes
from a lack of fundamental and precise knowledge of what those outcomes involve. But
in other situations, it may never be possible to describe the relevant trade-offs to
machines in a way that is acceptable to humans.
2. The most valuable workforce skills involve judgment. For many activities,
prediction was the bottleneck to automation. This meant a role for human
workers in aiding a variety of prediction tasks, including pick-and-place and
driving. Going forward, this role for human workers will diminish. Instead,
employers will look for workers who augment the value of prediction. In the
language of economics, just as the returns to numeracy and literacy increased
with the diffusion of computing and the demand for golf balls rises with lower
prices of golf clubs, the most valuable skills in the future will be those that are
complementary to prediction: those related to judgment. We can only
speculate on the aspects of judgment that will be most valuable: ethical
judgement, emotional intelligence, artistic ability, task definition, etc. This
diverse group of skills suggests that the set of activities in which the human
role may increase in value is potentially wide-ranging. Nursing skills related to
physical intervention and emotional comfort may be more valuable if prediction
leads to better diagnosis of disease. Retail store greeters may become
increasingly effective if social interactions help differentiate stores in the
presence of reliable predictions on purchase behavior. Private security guard
skills related to ethical judgement and use of force may be of increased value
as a complement to better prediction of crime. For many activities, human
judgment becomes more valuable when paired with effective prediction. It is
these skills that will be most valuable as artificial intelligence technology
13
improves and diffuses. Still, as noted above, the parts of tasks that need
human judgment can change over time. In particular, when you have enough
observations of judgment, it becomes a prediction problem. The AI can use the
data to predict the human judgment. Thus the judgement aspect of a task is a
moving target, requiring humans to adapt to new situations in which judgment
is required.
3. Managing may require a new set of talents and expertise. Managerial tasks
may also change. First, many managerial tasks are predictive. For example,
many managers emphasize their skills in identifying which applicants to hire
and which workers to promote. These are prediction skills: predicting which
workers will succeed. As machines improve their performance in prediction,
human prediction skills will become less valuable compared to judgement
skills, such as mentoring, emotional support, and ethics. Second, managing a
workforce whose skills are complementary to prediction may be different from
managing a workforce for which a core skill is prediction. For example,
promotion cannot be based on the (often well-measured) success of past
predictions. Instead, alternative metrics must be developed. Third, and
perhaps most importantly, managing a firm with artificial intelligence
capabilities will involve managing the artificial intelligence itself. What are the
opportunities for prediction? What should be predicted? How should the
artificial intelligence agent learn to improve those predictions over time?
Managing in this context requires judgment in identifying and applying the most
useful predictions and weighing the relative costs of different types of errors.
The usefulness of the technology is limited by the usefulness of the object to
be predicted. Sometimes such predictions are clear in the sense that there is a
well-acknowledged objective, such as object recognition to identify people from
their faces. Other times, however, the objective is less clear and requires
judgment to determine. For example, should the team managing Facebook’s
Newsfeed, maximize clicks, likes, content sharing, time on Facebook, or
profitability to advertisers? In such cases, managers’ judgment becomes a
particularly valuable complement to prediction technology.
At the dawn of the twenty-first century, the set of recognized prediction problems
were classic statistical questions: inventory management, demand forecasting, etc.
However, over the last ten years, researchers learned that image recognition, driving,
and translation may also be framed as prediction problems. As the range of tasks that
are recast as prediction problems continues to grow, we believe the scope of new
applications will be extraordinary. The key managerial challenges will be: 1) shifting the
14