Understanding Explaination Patterns 2
Understanding Explaination Patterns 2
ROGER C. SCHANK
Yale University
Copyrighted Material
Copyright © 1986 by Lawrence Erlbaum Associates, Inc.
All rights reserved. No part of this book may be reproduced in
any form, by photostat, microform, retrieval system, or any other
means, without the prior written permission of the publisher.
Bibliography: p.
1. Artificial intelligence. 2. Cognition.I.Title.
Q335.S3863 1986 006.3 86-19872
ISBN 0-89859-768-4
Copyrighted Material
Contents
Preface v
Chapter 1: The Explanation Game 1
Chapter 2: What Needs To Be Explained 25
Chapter 3: Explanation Goals 51
Chapter 4: The Process of Explanation:
Explanation Questions 80
Chapter 5: The Process of Explanation:
Explanation Patterns 109
Chapter 6: Creativity Through the
Misapplication of Explanation Patterns 140
Chapter 7: Question-Driven Understanding 178
Chapter 8: Some Final Questions 208
Appendix: SWALE: A Program That Explains 232
References 255
iii
Copyrighted Material
Copyrighted Material
Preface
In the age of the computer, conjecture about things mechanical has natu-
rally led to the question of whether machines can think. As the emphasis
on Artificial Intelligence (AI) has grown rapidly, questions about machine
intelligence have begun to have a certain urgency. There is, at the same
time, an opposite, almost whimsical type of question that arises when one
begins to consider whether machines can think, namely, Can people think?
People seem willing enough to grant machines the possibility of doing
calculations of tremendous complexity with utmost ease. They seem willing
to believe that machines will be able to talk, walk, hear, and evaluate, but
when it comes to the possibility that machines might be creative, on their
own, most people deny the possibility? Why? What magic power do we
believe people to possess that we also believe machines to be inherently
incapable of possessing?
This is why the question of whether machines can think is actually one of
whether people can think, or, to put this another way, if the present pursuit
of algorithms that describe the process of thinking in sufficient detail
actually enables a machine to emulate that process, we can expect, based
on past reactions to machine intelligence, that people will begin to worry
about the nature of their own humanness. If a machine can do something
"human," then that "something" has been reduced to a mere mechanical
process that no longer seems worth doing.
The question we are concerned with in this book is: If we can find a set
of processes that machines can slavishly follow, and if by so doing, these
machines can come up with creative thoughts, what would that tell us
about human beings? If the machine's procedure was adapted from a
v
Copyrighted Material
vi PREFACE
human procedure, that is, if all the machine was doing was what we know
people are doing, would we abandon our inherent skepticism about the
abilities of machines, or would we demystify our inherent admiration for
things human?
In a sense, these are the issues I deal with in this book. I say "in a sense"
because this book is no way a philosophical treatise. Rather it is an exercise
in Artificial Intelligence and in Cognitive Science. It is an attempt to come
to understand one of the most complex problems of mind by examining
some of the mechanisms of mind: to define the apparatus that underlies
our ability to think. In other words, this book is about people. But, it is not
only about people, it is about simulating people doing what they consider
to be their most human-specific act, being creative. Thus, it is also a book
about machines and machine-capabilities.
Why discuss creativity at all? Why, in an era of great difficulty in getting
machines to do the simplest understanding tasks, should one even begin to
consider creativity, certainly one of the most complex issues of all? My
premise is that the simplest mental tasks and the most complex of mental
tasks are all interrelated. Understanding the simplest sentence is a creative
act. Seen as a simple process involving merely the syntactic parsing of a
sentence, it is hard to see where the creativity might lie. But, understanding
a sentence also involves relating what you have understood to what you
already knew, and that task involves memory search. Memory search, as I
attempted to show in Schank (1982) involves learning since memory is
never quite the same as it was the last time it was used and will not be the
same after the new sentence is processed. And, learning involves creativity
since it isn't always so clear what should be learned. All these things are
really part of the same process. Thus it follows that AI will never exist in
any important way until machines are capable of some basic creativity.
This premise has been at the base of most of the work being done at the
Yale AI laboratory in recent years. I wish to thank my graduate students
who have helped me formulate these ideas by serving as a sounding board,
by writing programs based on these ideas, by revising some of these ideas
with their own better worked out ones, and by reading and commenting on
the manuscript. I also wish to thank the Advanced Research Projects
Agency of the Department of Defense, the Air Force Office of Scientific
Research, the Office of Naval Research, and the National Science Founda-
tion for its support of this work.
Copyrighted Material
To Joshua
who is more creative
than today's computers
Copyrighted Material
Copyrighted Material
1 The Explanation Game
The question of whether machines can think is perhaps a bit tired at this
point in the history of computing, but the question of what would be
required in order to make machines think is just beginning to be explored.
Of course, each question relates to the other, so, in order to begin the
process of building a thinking machine, we must consider not only if such a
project is possible, but also what elements of the human thinking process
are inextricably bound up with our sense of what we mean when we talk
about a thinking machine. To build a machine with Artificial Intelligence,
we must come to understand, in a profound way, what it means to have
intelligence. For years, researchers in AI could content themselves with
adding yet another cute feature to their programs, allowing machines to do
new things, one by one. But, it is becoming increasingly clear that to really
make intelligent machines, as opposed to machines that exhibit one or two
aspects of intelligence, one must attack the basic issues of the nature of
human thought and intelligence head on.
There have, of course, been many discussions of machine thought, by far
the most famous of these by the British mathematician, Alan Turing
(Turing, 1963), when computers were in their infancy. Turing proposed a
test, which he called the Imitation Game, that has become a common but
often misunderstood way of judging the ability of a machine to understand.
Turing's major argument was that the question of whether a machine can
think is meaningless. He suggested the following alternative: If a person
failed to distinguish between a man imitating a woman (via teletype) and a
1
Copyrighted Material
2 1. THE EXPLANATION GAME
Copyrighted Material
ON MEN AND WOMEN 3
computer to do as well at the Imitation Game as the man did, and assuming
that there is no difference between men and women that would be recogniz-
able via teletype, the task of the machine is to duplicate a human in its
answers. Thus, Turing's test doesn't actually depend upon men and women
being discernably different. But, the nature of what it means to understand
may best be illustrated by that distinction.
To see what I mean, let us consider the question of whether men can
really understand women, (or alternatively, whether women can really
understand men). It is common enough in everyday experience for men
and women to both claim that they really do not understand their opposite
number. What can they mean by this? And, most importantly, how is what
they mean by it related to the problem of determining whether computers
can understand?
When the claim is made that men and women are really quite different
(mentally, not physically), what is presumably meant is that they have
different beliefs, different methods of processing information, different
styles of reasoning, different value systems, and so on. (It is not my point
here to comment on the validity of these assertions. I am simply attempting
to use the principle of these assertions in my argument. These same
assertions might be made about different ethnic groups, cultures, nations
and so on; I am simply using Turing's domain.)
The claim that I assume is not being made by such assertions is that men
and women have different physical instantiations of their mental processes.
(Of course, it is possible that men and women do have brains that differ
physically in important respects, but that would be irrelevant for this
argument.) So, what is it that makes men and women feel they have difficulty
understanding each other? Empathy. Understanding involves empathy. It is
easier to understand someone who has had similar experiences—and who,
because of those experiences, has developed similar values, beliefs, mem-
ory structures, rules-of-thumb, goals, and ideologies—than to understand
someone with very different types of experiences.
Understanding consists of processing incoming experiences in terms of
the cognitive apparatus one has available. This cognitive apparatus has a
physical instantiation (the brain, or the hardware of the computer) and a
mental instantiation (the mind, or the software of the computer). When an
episode is being processed, a person brings to bear the totality of his
cognitive apparatus to attempt to understand it. What this means in prac-
tice is that people understand things in terms of their particular memories
and experiences. People who have different goals, beliefs, expectations and
general life styles will understand identical episodes quite differently.
Therefore, no two people understand in exactly the same way or with
exactly the same result. The more different people are from one another,
the more their perception of their experiences will differ. On the other
Copyrighted Material
4 1. THE EXPLANATION GAME
hand, when people share certain dimensions of experience, they will tend
to perceive similar experiences in similar ways. Thus, men tend to under-
stand certain classes of experiences in ways that are different from women.
It is unlikely that an experience that in no way bears upon one's sex will
be understood differently by men and women. Recall that the assumption
here is that the baseline cognitive apparatus is the same regardless of sex.
Any experience that does relate to the sex of the observer in some way will
be processed differently. This can involve obvious issues, such as the
observance of an argument between a man and a woman. There, we would
expect a man to observe the episode from the point of view of the man and
a woman to observe it from the point of view of the woman. In addition,
such identification with different characters in a situation can extend to
observations of situations where the feuding characters are of the same sex,
but one displays attributes more traditionally male and the other displays
traditional female behavior. Identification, and thus perception, can thus
be altered by one's understanding of the goals, beliefs, or attitudes underly-
ing or perceived to underlie the behavior of the characters in an episode
one is observing.
Thus, for example, one's perception of the validity and purpose behind a
war can be altered by whether one is the mother of a son who is about to be
drafted or whether one previously fought in a war and found it an enno-
bling experience. In general, one's sense of what is important in life affects
every aspect of one's understanding of events.
The claim then, is that men and women, as examples of one division of
human beings, do not, and really cannot, understand each other. The same
argument can be put forward, with more or less success, depending upon
the issue under consideration, with respect to Arabs and Israelis or intellec-
tuals and blue collar workers. In each of these cases, differing values can
cause differing perceptions of the world.
Copyrighted Material
COMPUTERS AND PEOPLE 5
Copyrighted Material
6 1. THE EXPLANATION GAME
Copyrighted Material
THE NATURE OF UNDERSTANDING 7
for example (Schank & Abelson, 1977), was used in an attempt to understand
restaurant stories, and it was that script which prompted Weizenbaum's
criticism about understanding love in a restaurant. Since that earlier research
we have come to realize that these knowledge structures function best if
they are dynamic. That is, they must be able to change as a result of new
experiences. In other words, we expect that as knowledge is used, it
changes. Or, to put this another way, as we undergo experiences we learn
from them.
At the core of such an hypothesis is the notion that in attempting to
understand we are attempting to relate our new experiences to our prior
experiences by utilizing knowledge structures that contain those previous
experiences. Consider, for example, the following situation. Imagine that
you are hungry and that someone suggests a marvelous restaurant to you
called Burger King. You happily go to this restaurant, armed as you always
are, with a set of expectations about what will happen there. Specifically,
you expect that you will: ENTER; BE SEATED; GET & READ MENU;
ORDER; BE SERVED; EAT; PAY; and EXIT. The items written in bold
face are called scenes and can be best understood as bundles of expecta-
tions themselves that concern the particulars of how those scenes will
actually take place. The assertion put forth in Schank (1982) is that such
scenes are derived from experience and are subject to constant change by
experience.
You were told that you were going to a restaurant, so you brought out your
Memory Organization Package (or MOP) for restaurants (M-RESTAURANT)
that told you what scenes to expect. What you find in Burger King,
however, is a different sort of thing altogether. The order of scenes is
ENTER; ORDER (without a menu exactly); PAY; SERVE (but it doesn't
look much like the other SERVE); BE SEATED (but on your own); EAT;
LEAVE. So what is a processor to do?
The obvious thought is that you have been fooled. This is not a
restaurant at all. Or maybe this is just a weird restaurant. Without
other evidence, it is hard to know what to do. What a processor can
do is mark the expectation failures. That is, we expected that scenes
would come in a certain order, and we expected that scenes would
proceed in a certain manner, but they didn't. What we do in this case
is to index the exceptions so that we will be able to recognize them when
they occur again.
Now suppose that after you complain to your friend about Burger King,
he suggests that you try another restaurant, called McDonald's, instead.
You are confronted with a new situation which you must attempt to
"understand." And it is reasonable to say that you have understood
McDonald's if you recall the Burger King experience (in other words, are
reminded of Burger King) and use it to help create expectations about what
Copyrighted Material
8 1. THE EXPLANATION GAME
will happen next. The key point, however, is what happens to your memory
as a result of these two experiences.
You have now encountered two exceptions to a MOP that were them¬
selves quite similar. It is reasonable, therefore, to create a new MOP that
has expectations in it that correspond to these new experiences. We might
call that new MOP, M-FAST-FOOD. We can index this MOP as an excep-
tion to M-RESTAURANT so that it will be available to change and grow
as new experiences relevant to it occur.
The important point here then is that when we are reminded of some
event or experience in the course of undergoing a different experience, this
reminding behavior is not random. We are reminded of this experience
because the structures we are using to process this new experience are
the same structures we are using to organize memory. Thus, we cannot
help but pass through the old memories while processing a new input.
There are an extremely large number of such high level memory struc-
tures. Finding the right one of these (that is, the one that is most
specific to the experience at hand) is one of the things that we mean by
understanding.
In other words, an important part of what we mean by understanding is
the accessing of extant knowledge structures in which to place our new
experiences. We feel that we have understood when we know where a new
input belongs in memory. Sometimes understanding is more profound than
that, however. The creation of new knowledge structures in memory is also
part of understanding. Such structures are created in terms of old ones.
The more these new structures differ from the old, the more complicated
understanding can be.
In this view then, understanding is finding the closest higher level structure
available to explain an input and creating a new memory structure for that
input that is in terms of the old closely related higher level structure.
Understanding is a process that has its basis in memory then, particularly
memory for closely related experiences accessible through reminding and
expressible through analogy. Further, the depth of understanding will
increase if there are many available relevant personal experiences in terms
of which inputs can be processed. Lastly, understanding means finding
some memory, any memory at all sometimes, that will help one cope with
an experience that initially seems novel. We want to feel that we have
understood, and we feel that way to the extent that we can believe that
what we have just seen really is like something else we have seen previously.
With this definition of the nature of understanding then, let us now
return to our understanding spectrum.
Copyrighted Material
THE SPECTRUM OF UNDERSTANDING 9
today's work on computer understanding can only reasonably claim the left
half of the spectrum as its proper domain. It is a legitimate argument that
computers will never understand, if what is meant by that is that they
would be unlikely to understand much at the right hand side of the
spectrum. Computers are not likely to feel the same way as people do, are
not likely to have the same goals as people do, and will never, in fact, be
people. Given that most people fail to understand each other at points on
the right hand side of this spectrum, it seems no great loss to admit to the
likely failure of computers to achieve this level of understanding. Com-
Copyrighted Material
10 1. THE EXPLANATION GAME
Copyrighted Material
THE REVISED TURING TEST 1 1
question for AI is not whether the levels we have already achieved can be
construed to be examples of understanding, but rather how we can achieve
the levels of understanding that are still eluding us. MAKING SENSE has
been achieved by computers in certain domains. On the other side of the
spectrum, it may well be the case that computers cannot, in principle,
completely understand people, any more than men can completely under-
stand women or vice versa. COMPLETE EMPATHY is just an end point
on the continuum. There is a lot of space in between the end points. The
real issue is not whether computers can understand, but how to make them
understand better. To do this, we must get a better picture of what it means
to understand.
How different are these three points on the spectrum in terms of what
processes would be necessary in order to construct machines that met
those criteria? What does a machine do in order to achieve the ability to
MAKE SENSE, and what would it have to do in order to COGNITIVELY
UNDERSTAND or show COMPLETE EMPATHY?
Turing's Imitation Game has not left everyone in AI thrilled by the pros-
pects of having to meet its criteria as a measure of success. Colby (1973)
argued that his paranoid simulation program did indeed pass Turing's test.
He found that psychiatrists who were presented with output from Colby's
PARRY program and output in a similar form from an actual human
patient, were unable to effectively distinguish between them. This passing
of the Turing test failed to convince AI people that Colby's program
understood or thought, nor should it have done so. Despite whatever
validity Colby's program might have as a model of paranoia, it would seem
that the failure of experts to distinguish between imitations and the real
thing should not be taken as much more than a statement of the compe-
tence of the experts. In fact, in the case of the psychiatrists, the Imitation
Game was a particularly poor test since a psychiatrist's choices upon facing
a non-normal patient are not that extensive. PARRY was not brain-damaged
or schizophrenic, so "paranoid" was, given the presence of a few paranoid
signs, a reasonable diagnosis. What Colby seems to have done is effectively
simulated various aspects of the output of a paranoid, which may or may
not reflect accurate processes underlying the production of that output.
The issue then, is one of the distinction between good output and good
simulation.
Ideally, the test of an effective understanding system, if I may use that
word, is not the realism of the output it produces, as Turing would have it,
but rather the validity of the method by which that output is produced.
Copyrighted Material
12 1. THE EXPLANATION GAME
Copyrighted Material
THE REVISED TURING TEST 13
output: Do you remember the way your father used to treat you
on holidays when he made you call all your relatives? He
meant no harm either.
input: I see what you mean.
output: I thought you might; there's no friend like an old friend.
Assuming that these input/output pairs are not entirely fanciful, I would
now like to draw some conclusions about them that are reflective of my
view of what a reasonable test should comprise. My conclusions are effectively
summarized with the following words:
ACCURACY; SURPRISE; EMOTION
The claim I am making is, to the extent that output is an effective way of
characterizing degree of understanding (although that is to a very limited
extent indeed, it may well be our only choice), we can judge the signifi¬
cance of that output in terms of its place on the understanding spectrum
with respect to the following features:
The extent that that output accurately accomplishes a task that a
competent human could do.
The extent that that output characterizes an especially original or
important result that most humans cannot easily accomplish.
The extent that that output effectively seems to replicate a real
live human being with whom someone is familiar.
The above three standards reflect the essence of the three points on the
understanding spectrum that I have been discussing.
Turing's Imitation Game then, can be seen, as a result of further insights
into AI since the time that Turing first concerned himself with these issues,
to be somewhat outdated. Merely fooling people is not the true task in AI.
Rather we must, in addressing fundamental questions about the nature of
intelligence, devise a set of requirements that any intelligent entity ought to
be able to meet. And, if we are talking about intelligent beings with
linguistic capabilities, any test we devise can have an added level of
complexity. That is, the true value of language in an intelligence-testing
framework is that we can transcend the limitations of only being able to get
solutions to problems or answers to questions. Linguistically equipped
intelligent beings can explain their own actions.
Copyrighted Material
14 1. THE EXPLANATION GAME
Copyrighted Material
THE EXPLANATION TEST 15
well as a human would. Now, it would not have passed Turing's test,
because it understood nothing but blocks. But, I claim, it should be
recognized as having passed the Explanation Test at the level of MAKING
SENSE.
Each point on the understanding spectrum has essentially the same
requirements in the Explanation Test. For COGNITIVE UNDERSTAND-
ING, the program must be able to explain why it came to the conclusions it
did, what hypotheses it rejected and why, how previous experiences influ-
enced it to come up with its hypotheses and so on. We do not let a human
being come up with innovative ideas, generalizations, correlations and so
on, unless he can explain himself. Creative scientists are supposed to be
able to tell us what they do—not what they do while they are being creative,
but what the reasoning is behind what they have discovered. We may not
expect Einstein to know how he came up with his creative ideas, but we
certainly expect him to be able to explain the physics behind them. We will
put up with unexplained brilliance for a while, but eventually we object, we
believe that we are somehow being fooled. We cannot put up with a
scientist who says: E = MC2,I don't know what that means, but I am sure it
is true. There is no reason why we should let a machine off more easily.
The machine must be able to answer the question, How do you know? to
the satisfaction of an examiner in a school who would expect no less from
his students on a similar accomplishment in a similar domain.
The last point we have presented on our spectrum, that of COMPLETE
EMPATHY, has no easier test. Any system purporting to satisfy this level of
understanding must satisfy its examiner in the same way that a human
would satisfy him in a similar situation. I am claiming that this is not likely
in its fullest sense, and that this improbability is what all the uproar is about
with respect to the assessment of the possibilities for Artificial Intelligence
by laymen. No machine will have undergone enough experiences, and
reacted to them in the same was as you did, to satisfy you that it really
understands you. One is extremely lucky if one meets one person in one's
lifetime who satisfies that criterion.
The second aspect of explanation is much more complex. What would it
mean for a system to explain something to itself, and why does that matter?
The argument put forward in this book is that what is important about
understanding is how it changes over time. A machine that understands at
the level of making sense fails to convince us that it really has understood
because it continues to understand in the same way every time. The
computer programs that we built at Yale during the 1970's, all had the
feature, as do most AI programs, of behaving exactly the same way every
time.
But, people, especially intelligent people, do not behave this way. They
adapt to new circumstances by changing their behavior. How do they do
Copyrighted Material
16 1. THE EXPLANATION GAME
this? The argument put forward in this book is that the method that they
use for this adaptation is also based upon explanation, but explanation to
oneself. Explaining something to yourself means, in essence, attempting to
correct an initial misunderstanding by finding relevant experiences in one's
memory that might account for the incomprehensible event. The result of
such explanation can be a new understanding of old information as well. In
other words, explanation, learning, and also creativity, are inextricably
bound together.
To explain something to yourself at the making sense level, one need
only find a relevant knowledge structure in which to place a new input.
Seeing a new event as an instance of an old event constitutes understanding
at a very simple level. But, trying to find the theory that explains a group of
events can be a much more complicated and more profound type of
understanding. Explaining a group of events to yourself can thus mean
coming up with a theory of those events. Thus, cognitive understanding
really requires a good grasp of what would constitute a reasonable explana-
tion for something in general.
From the perspective of a test of a computer's understanding ability
then, this spectrum allows us to set some criteria that will allow us to
examine whether an understanding system is understanding by looking at
that systems's powers of explanation. But, more important is a system's
ability of self-explanation. That is, one cannot easily examine how a system
changes itself as a result of its experiences, one can just observe it over
time. But the claim is that the concept of explanation is key in bothe cases.
Explaining the world, to others and to oneself, is, in essence, the heart of
understanding.
Copyrighted Material
QUESTIONS AND EXPLANATIONS 17
left untestable, however. Was the program changed in some way by the
joke? This is not an irrelevant question for a person, nor should it be for a
machine, though it might a bit difficult to answer in either case.
What will it take to get a machine to the level where it satisfies people
that it understands at the level of COGNITIVE UNDERSTANDING? Or,
to put this question another way, what is it that people do with each other
that makes us believe that they are understanding at that level? The answer
can, I believe be reduced to two words: questions and explanations.
People generate questions about those things that they did not fully
understand. Sometimes those questions cause people to find explanations,
sometimes novel and sometimes mundane, for the odd, unexpected, situa-
tions that they encounter. Sometimes those questions remain, to be answered
at a later date, or to be combined with new explanations as the seeds of
more creative explanations.
This book is about what it would take to achieve COGNITIVE UNDER-
STANDING. But it is not a philosophical work. I am interested in how to
push the level of intelligence that computers have already achieved to the
next step. I am also interested in exactly what it is that people must be
doing when they do intelligent things.
Because of that, this book is about questions and explanations. It is a
book about people and machines. It addresses the questions of how humans
formulate questions and how humans formulate explanations. The idea
here is that if we can find out what it is that humans do when they are
thinking and learning, then maybe we can model our machines on them.
The real intent of Artificial Intelligence, is, I claim, to find out what
intelligence is all about. We tend to say that a person is intelligent to the
extent that he is insightful, creative, and in general, able to relate apparently
unrelated pieces of information to come up with a new way of looking at
things. We tend to claim that a person is unintelligent to the extent that his
behavior is thoroughly predictable with reference to what we know that he
knows. Thus, when a person does things the way he was told to do them,
never questioning and thus never creating new methods, we tend to see him
as unintelligent.
I mention this here because I see the Explanation Test as a kind of
intelligence test. We are not asking the computer to simply replicate
intelligent behavior because we have no knowledge of which aspects of
such behavior are more intelligent than others. Is composing a sonnet a
more or less an intelligent act than playing chess? There is no way to
answer this objectively because it isn't the acts themselves that are at issue
here, but rather the quality of those acts. Turing could not differentiate
between these feats because he did not have the experience of trying to
build programs to do each task. But now, as a result of years of AI research,
such a question is easier to answer.
Copyrighted Material
18 1. THE EXPLANATION GAME
We can make a program write bad sonnets or play poor chess fairly
easily. Neither of these feats seem much of a mark of intelligence. Indeed,
working on either of them would not be considered AI any more, although
such work was done by AI researchers not so long ago. Today, work on
computer poetry or computer chess falls within the domain of AI only to
the extent that it mimics the complex cognitive processes associated with
the creativity inherent in both acts. Thus, if the computer poetry program
started with a set of feelings and was able, by relating such feelings to its
personal experiences, to create poetry, particularly poetry of some new
type, we would be legitimately impressed. Similarly, if our computer chess
program was capable of improving its playing ability by inventing a new
strategy or employing an old one that it recalled having seen in a match it
knew about, that would be an AI-type feat.
We have come to understand in AI that it isn't the tasks themselves that
are interesting. What matters is how they are done. Thus, I claim, the only
way to know if our machines are intelligent is to make them do what we
would expect a human to do in a similar situation. We must expect them to
be able to explain how they did it. Furthermore, those explanations should
have some connection with how the task in question actually was performed.
Often this is a difficult task for people to perform. We do not always know
where our creative powers come from or how they were employed in any
given instance. But, we can attempt to give rational explanations. We
should demand no less from machines.
Copyrighted Material
CHANGING MACHINES 19
CHANGING MACHINES
Copyrighted Material
20 1. THE EXPLANATION GAME
Copyrighted Material
CHANGING MACHINES 21
Copyrighted Material
22 1. THE EXPLANATION GAME
EXPLANATION IS UNDERSTANDING
A first assumption in this book, then (first stated in Schank, 1982), is that
when our expectations are found to be in error we must attempt to explain
why. Failures lead to attempts to explain failures. We want to avoid making
the same mistake twice. How? By understanding what principles underlie
the expectations that we had in the first place. We must understand them
so we can fix them.
In the remindings that we examined, one thing that we found (Schank,
1982) was that each pair of events that we connected by a reminding had in
common not only an identical expectation failure, but an identical explana-
tion of that failure as well.
Explanation is critical to the understanding process. As understanders,
we want to know what others around us are doing, and will do. Furthermore,
we want to know why others do what they do. Thus, people are constantly
seeking explanations for the behavior of others. One of the simplest kinds
of explanations is a script (Schank & Abelson, 1977). We believe that we
understand why a person is doing what he is doing if we can point to a
script that he is following. Knowing that someone is doing a particular
action because it is part of the behavior prescribed by a commonly adhered
to set of rules makes us feel comfortable that we have explained his
behavior. We can, of course, look for explanations that are more profound
than saying "he is doing that because he (or someone in his situation)
always does that." We usually feel that we have understood without going
that deep, however. And, it is this script-type of understanding that we are
referring to when we talk about Making Sense. But, when script-based
understanding is not enough, when we really want to understand something
in a deep way, the explanation process must come into play. It is the
creation of new explanations that produces Cognitive Understanding.
At the root of our understanding ability is our ability to seek and create
explanations. What we are trying to do when we seek an explanation for an
actor's behavior, is to find a set of beliefs that the actor could hold with
which the actions that he took would be consistent. So, what is understand-
ing then? In this view, understanding can be seen to be no more than, and
no less than, explanation.
We have, in AI, gotten used to speaking of "computer understanding,"
meaning by that term that an input sentence has been combined with a
memory structure of some sort. To put this graphically, we have believed
that:
Copyrighted Material
EXPLANATION IS UNDERSTANDING 23
AN INPUT SENTENCE
+
A MEMORY STRUCTURE THAT INDICATES WHERE TO
PLACE THAT SENTENCE
results in
AN UNDERSTOOD SENTENCE
In other words, the process of understanding, for computers in the early
stages of AI research, was embodied in the notion that understanding
something means putting it in its proper context, or Making Sense of it.
The major problem with this point of view is that, on occasion, the
proper context either does not previously exist, or, it is difficult to deter-
mine exactly what that context is. Graphically:
AN INPUT SENTENCE
but
NO MEMORY STRUCTURE IS PREPARED TO ACCOMMODATE
THAT SENTENCE
results in
A SENTENCE THAT NEEDS EXPLANATION
The premise here is that, in principle, what we are doing when we
understand is explaining what we have heard. Most of the time those
explanations are very mundane, so we don't feel as if we are explaining
anything. Rather, we are simply placing the new information in a previously
extant structure. In other words, we don't have to explain why someone is
reading a menu in a restaurant since we know exactly when and where such
actions take place. Explanations are necessary when an action fails to come
in the way we expected, or cannot be placed into an existing structure.
But, the point is that we are always needing to explain every action we
encounter. We feel that we are explaining when we are doing something
other than accessing an everyday, run-of-the-mill, structure, or when, while
using such a structure, we have still found ourselves incapable of deciding
exactly what to do. Thus, explaining and understanding are the same thing,
one is just more conscious a process than the other. And, explanation must
be a more conscious process, because it occurs when something has gone
wrong and needs to be fixed.
Copyrighted Material
References
Abelson, R. , Carroll, J. (1965). Computer Simulation of Individual Belief Systems, American
Behavioral Scientist. 8, pp. 24–30.
Carbonell, J. (1979). Subjective Understanding: Computer Models of Belief Systems, Ph.D. Thesis,
Yale University.
Colby, K. M. (1973). The Structure of Belief Systems, Schank, R.C. , and Colby, K.M. eds., Computer
Models of Thought and Language, W.H. Freeman and Company, San Francisco, pages 251–286.
Colby, K.M. , & Smith, D.C. (1969). Dialogues between Humans and an Artificial Belief System,
Walker, D.E. and Norton, L.M. eds., Proceedings of the International Joint Conference on Artificial
Intelligence, IJCAI, Boston.
Cullingford, R. (1978). Script Application: Computer Understanding of Newspaper Stories, Ph.D.
Thesis, Yale University. Research Report #116.
DeJong, G. (1979). Skimming Stories in Real Time: An Experiment in Integrated Understanding, Ph.D.
Thesis, Yale University.
Dreyfuss, H.L. (1972). What Computers Can't Do: A Critique of Artificial Reason, Harper & Row, New
York.
Hammond, K. (1984). Indexing and Causality: The organization of plans and strategies in memory.
Technical Report 351, Yale University Department of Computer Science.
Kogos, F. (1970). Book of Yiddish Proverbs. Poplar Books, Secaucus, NJ.
Lebowitz, M. (1980). Generalization and Memory in an Integrated Understanding System. Ph.D.
Thesis, Yale University.
Lytinen, S. (1984). The Organization of Knowledge In a Multi-lingual, Integrated Parser, Ph.D. Thesis,
Yale University. Research Report #340.
Newell, A. , & Simon, H. (1972). Human Problem Solving. Prentice-Hall, Englewood Cliffs, NJ.
Restak, R.M. (1979). The Brain: The Last Frontier. Doubleday, Garden City, NY.
Rieger, C. (1975). Conceptual Memory and Inference, Conceptual Information Processing. North-
Holland, Amsterdam.
Schank, R.C. (1972). Conceptual Dependency: A Theory of Natural Language-Understanding.
Cognitive Psychology. 3/4, pp. 552–631.
Schank, R.C. (1975). Fundamental Studies in Computer Science, Volume 3: Conceptual Information
Processing, North-Holland, Amsterdam.
Schank, R.C. (1982). Dynamic Memory: A Theory of Learning in Computers and People, Cambridge
University Press.
Schank, R.C. , & Abelson, R. (1977). Scripts, Plans, Goals and Understanding, Lawrence Erlbaum
Associates, Hillsdale, New Jersey.
Schank, R.C. , & Riesbeck, C. (1981). Inside Computer Understanding: Five Programs with
Miniatures, Lawrence Erlbaum Associates, Hillsdale, New Jersey.
Searle, J.R. (1980). Minds, Brains and Programs, The Behavioral and Brain Sciences, 3, pp. 417–424.
Shortliffe, E.H. , Axline, S.G. , Buchanan, B.G. , Merigan, T.C. , & Cohen, N.S. (1973). An Artificial
Intelligence Program to Advise Physicians Regarding Antimicrobal Therapy, Computers and
Biomedical Research, 6, pp. 544–560.
Turing, A.M. (1963). Computing Machinery and Intelligence, Feigenbaum, E.A. , and Feldman, J. eds.,
Computers and Thought, McGraw-Hill, New York.
Weizenbaum, J. (1976). Computer Power and Human Reason: From Judgement to Calculation,
Freeman, San Francisco.
Wilensky, R. (1978). Understanding Goal-Based Stories, Ph.D. Thesis, Yale University. Research
Report #140.
Winograd, T. (1972). Understanding Natural Language, Academic Press, New York.