Dennis Yi Tenen - Literary Theory For Robots - How Computers Learned To Write-W.W. Norton (2024)
Dennis Yi Tenen - Literary Theory For Robots - How Computers Learned To Write-W.W. Norton (2024)
Dennis Yi Tenen - Literary Theory For Robots - How Computers Learned To Write-W.W. Norton (2024)
DENNIS YI TENEN
CONTENTS
CHAPTER 1:
Intelligence as Metaphor (An Introduction)
CHAPTER 2:
Letter Magic
CHAPTER 3:
Smart Cabinets
CHAPTER 4:
CHAPTER 5:
Template Culture
CHAPTER 6:
Airplane Stories
CHAPTER 7:
Markov’s Pushkin
CHAPTER 8:
9 Big Ideas for an Effective Conclusion
NOTES
INDEX
CHAPTER 1
C OMPUTERS LOVE TO READ. And it isn’t just fiction before going to bed.
They read greedily: all literature, all of the time—novels,
encyclopedias, academic articles, private messages, advertisements, love
letters, news stories, hate speech, and crime reports—everything written
and transmitted, no matter how insignificant.
This ingested printed matter contains the messiness of human wisdom and
emotion—the information and disinformation, fact and metaphor. While we
were building railroads, fighting wars, and buying shoes online, the
machine child went to school.
Literary computers scribble everywhere now in the background, powering
search engines, recommendations systems, and customer service chatbots.
They flag offensive content on social networks and delete spam from our
inboxes. At the hospital, they help convert patient-doctor conversations into
insurance billing codes. Sometimes, they alert law enforcement to potential
terrorist plots and predict (poorly) the threat of violence on social media.
Legal professionals use them to hide or discover evidence of corporate
fraud. Students are writing their next school paper with the aid of a smart
word processor, capable not just of completing sentences, but generating
entire essays on any topic.
In the industrial age, automation came for the shoemaker and the factory-‐
line worker. Today, it has come for the writer, the professor, the physician,
and the attorney. All human activity now passes through a computational
pipeline—even the sanitation worker transforms effluence into data. Like it
or not, we have all become subjects to automation. To survive intact, we
must also learn to become part software engineers and part, well—whatever
you are doing is great!
If any of the above comes as a surprise to you, my job, I feel, is mostly
done. Curiosity piqued, you will now start noticing literary robots
everywhere and join me in pondering their origins. Those not surprised
perhaps believe (erroneously!) that these siliconites have learned to chat
only recently, somewhere on the fields of computer science or software
engineering. I am here to tell you that machines have been getting smarter
in this way for centuries, long before computers, advancing on far more
arcane grounds, like rhetoric, linguistics, hermeneutics, literary theory,
semiotics, and philology.
So that we may hear them speak—to read and understand a vast library of
machine texts—I want to introduce several essential ideas underpinning the
ordinary magic of literary computers. Hidden deep within the circuitry of
everyday devices—yes, even “smart” light bulbs and refrigerators—we will
find tiny poems that have yet to name their genre. By the end of the book, I
hope you can see this technology anew, replete not just with instrumental
capacity (to keep food cold or to give light) but with potential for creativity
and collaboration.
Throughout, we will be tempted to ask existential questions about the
nature of artificially intelligent things. “How smart are they?” “Do they
really ‘think’ or ‘understand’ our language?” “Will they ever—have they
already—become sentient?”
Such questions are impossible to answer (in the way asked), because the
very categories of consciousness derive from human experience. To
understand alien life forms, we must think in alien ways. And rather than
argue about definitions (“Are they smart or not?”), we can begin to describe
the ways in which the meaning of intelligence continues to evolve.
Not long ago, one way of appearing smart involved memorizing a bunch
of obscure facts—to become a walking encyclopedia. Today, that way of
knowing seems like a waste of precious mental space. Vast online databases
make effective search habits more important than rote memorization.
Intelligence changes. The puzzle of its essence cannot therefore be
assembled from sharp, binary attributes, laid out always and everywhere in
the same way: “Can machines think: yes or no?” Rather, we can start
putting the pieces together contextually, at specific times and places, and
from the view of an evolving, shared capacity: “How do they think?” and
“How do we think with them?” and “How does that change the meaning of
thinking?”
In answering the “how” questions, we will discover a strange sort of a
twinned history, spanning the arts and sciences. Humans have been thinking
in this way—with and through machines—for centuries, just as they have
been thinking with and through us. The mind, hand, and tool move at once,
in unison. But the way we train minds, hands, or tools treats them almost
like entirely separate appendages, located in different buildings, in
unrelated fields on a university campus. Such an educational model isolates
ends from means and means from ends, disempowering its publics. In this
volume, I would like to imagine an alternative, more integrated curriculum,
offered to poets and engineers alike—bound, eventually, for a machine
reader as part of another training corpus.
Next time you pick up a “smart” device, like a book or a phone, pause
mid-use to reflect on your body posture. You are watching a video or
writing an email perhaps. The mind moves, requiring mental prowess like
perception and interpretation. But the hand moves, too, animating body in
concert with technology. Pay attention to the posture of the intellect—the
way your head tilts, the movement of individual fingers, pressing buttons or
pointing in particular ways. Feel the glass of the screen, the roughness of
paper. Leaf and swipe. Such physical rituals—incantations manifesting
thought, body, and tool—bring forth the artifice of intellect. The whole
thing is “it.” And that’s already kind of the point: Thinking happens in the
mind, by the hand, with a tool—and, by extension, with the help of others.
Thought moves by mental powers, alongside the physical, the instrumental,
and the social.
What separates natural from artificial forces in that chain? Does natural
intelligence end where I think something to myself, silently, alone? How
about using a notebook or calling a friend for advice? What about going to
the library or consulting an encyclopedia? Or in conversation with a
machine? None of the boundaries seem convincing. Intelligence demands
artifice. Webster’s dictionary defines intelligence as the “skilled use of
reason.” Artifice itself stems from the Latin ars, signifying skilled work,
and facere, meaning “to make.” In other words, artificial intelligence just
means “reason + skill.” There are no hard boundaries here—only synergy,
between the human mind and its extensions.
What about smart objects? First thing in the morning, I stretch and, at the
same time, reach for my phone: to check my schedule, read the news, and
bask in the faint glow of kudos, hearts, and likes from various social apps.
How did I get into this position? I ask alongside the beetle from Kafka’s
The Metamorphosis. Who taught me to move like this?
It wasn’t planned, really. Nor are we actually beetles living in our natural
habitat, an ancient forest floor. Our intimate rituals morph organically in
response to a changing environment. We dwell inside crafted spaces,
containing the designs for a purposeful living. The room says, Eat here,
sleep there; the bed, Lie on me this way; the screen, Hold me like this. Smart
objects further change in response to our inputs. To do that, they must be
able to communicate: to contain a layer of written instructions. Somewhere
in the linkage between the tap of my finger and the responding pixel on-‐
screen, an algorithm has registered my preference for a morning routine. I
am the input and output: the tools evolve as they transform me in return.
And so, I go back to bed.
THE CHAPTERS OF THIS book correspond to what I see as the three major gaps
in our collective thinking about artificial intelligence. First, I will argue that
our understanding of AI can and must become more grounded in the history
of the humanities. Second, that a software-engineering education
necessarily involves a number of inescapable philosophical problems,
related to several nerdy-but-I’m-sure-you’ll-find-them-fascinating kinds of
textual matters. Finally, I propose to consider AI as part of the broader
social and political dynamics arising, more generally, from the automation
of labor, particularly as it concerns intellectual labor or “knowledge work.”
History will remain my first and primary concern. Insert something clever
here about forgetting the past to the detriment of the future. Any field
developing rapidly, as computer science has been, can be forgiven for
orienting itself toward the future. But not for long. Yes, one can learn to
type quickly or to write a spreadsheet macro without a second thought.
Reflect, however, on the immense frustration of commandeering all of our
clumsy, impoverished digital avatars. You want me to save a what, where?
File? Folder? Server? Cloud? The pervasive presence of metaphors should
hint at the extent of our collective alienation. What is a file? Wherefore a
folder? Why are they even called that? Who thought of putting exactly eight
bits into a byte, and why not nine? How does a keystroke transform itself
into a configuration of pixels on screen, forming a letter? Explain a
common algorithm used to search, finish sentences when typing, filter
spam, or to recommend products. No way. Too much. Such questions fall
out of serious consideration because, paradoxically, they are “too simple”
for computer science and way, way too complex for baseline literacy.
But first, why do history? One can simply launch into the future from the
state-of-the-art, the state-of-the-present. At least two points are needed to
make a line, however. History’s subtle advantage lies in the revealing of
trajectories through time. AI and indeed “artifice” or “intellect” (and even
the toasting of bread) represent a cluster of related objects, ideas,
techniques, practices, and institutions. A static snapshot contains their
anatomy in three dimensions—this is how modern toasters work and this is
how a mind thinks today. History expands these arrangements into the
fourth dimension of time. Like any technology, the mind changes over time.
Its anatomical hardware stays the same, but the software receives regular
updates: all those encyclopedias, calculators, search tools, and other smart
devices that amplify the powers to reason.
Yet this history of AI should have nothing to do with the recent
resurgence of covertly neo-Hegelian thought so popular with futurists and
disruptive entrepreneurs, who believe in some sort of a teleological “event
horizon,” past which humankind will attain a posthuman singularity (all
challenges met, competitors bested, obstacles overcome). A singularity that
leads to the end of history sounds both contemptible and horrific—I’m not
sure if humans would survive it. History cannot be personified in that way.
It isn’t leading us to salvation or transcendence. It has no post-human goals
as such, other than those we make for ourselves.
On a less apocalyptic scale, history allows for extrapolation. It helps
anticipate changes in the direction of travel. Without history, our ideals and
our policies lack sufficient foresight to prepare adequately and to alter
course if needed. Without history, society is doomed to struggle with long-‐
term effects of technology, such as pollution and urban sprawl, and soon, if
not already, AI.
History ultimately makes for better engineers and users of technology.
Any complicated gadget contains within it a number of contingencies that
appear senseless today because they were attained through a compromise
with past limitations.
QWERTY, the layout of modern English keyboards, for example, was
famously arranged in that order to slow down typing on typewriters that
jammed otherwise. Similarly, eight bits (ones and zeroes) make a byte (a
unit of information) by convention. The number eight holds little
significance. The arbitrary length always confused me, until I learned that
early telegraph systems also experimented with variable-bit encoding, such
as the Morse code (which can have anywhere between one and five bits), as
well as fixed encoding, like the five-bit Baudot code, deriving from earlier
cryptographic cyphers proposed by Francis Bacon and others in the early
seventeenth century. The linear ticker tape used in telegraphs subsequently
could break a string of punches into any arbitrary length. It just so happens
that the rectangular punch card in use by early tabulating census machines,
and later by IBM, would not physically fit more than eight characters
across, in rows of seventy to eighty columns stacked. The convention of an
eight-bit byte (and the eighty-character code line) persisted, even though
magnetic tape and later “hard drive” media were no longer constrained by
those limitations. The physical properties of cardboard thus continue to
inform the design of modern computers. The abstract and alien concept of
“bits” became concrete to me by imagining holes punched through
cardboard. On a recent trip to Kyrgyzstan, a friendly librarian gifted me a
punch card used by Soviet computers in 1977. I use it as a bookmark now,
but its neat rows and columns still help conjure the rather elusive “byte” as
an object.
History gives meaning to such inherited structures. In following the
evolution of technology, we are able to make sense of legacy technological
debt. Whatever is meant by “innovation” consists of realizing which
features of the inherited design remain necessary and which can be
discarded or improved. Without history, the present becomes invariable,
sliding into orthodoxy (“it is what it is”). History’s milestones mark a path
into a possible future.
It’s okay, then, to sometimes struggle with basic file operations alongside
some of my best data-science students. Believe in yourself, Dennis! How
can I consider myself literate if I don’t fully understand how writing works?
A technical answer isn’t enough, either. “How writing works” cannot be
reduced to the present technological moment, because that moment will
pass, and rapidly. Instead, writing “has come to mean” this specific
arrangement of mind, body, tool, circuit, bit, gate . . . through its historical
development. An obtuse technological concept, such as solid-state memory
storage (the quantum shimmer behind words on screen), obtains meaning
only through a specific genealogy: from ticker tape to punch card and to the
NOT-AND gates flickering silently in the palm of your hand.
At the very least, the erasure of such realities stunts intellectual growth.
Good engineers need history, because the progress of their education can
follow the gradual development of their subject matter. History leads to
deep understanding because it explains why and how something has come
to be the way it is. Without it, we passively inherit a bundle of meaningless
facts: it’s just the way things are.
Poets steeped in engineering likewise stand to understand something vital
about the nature of writing (sign, symbol, inscription, representation, if you
want to get fancy). From that perspective, writing machines can be seen as
the culmination of a long, mythical tradition: In the beginning, there was
the letter. Then, a byte came to encode eight bits, because eight was enough
to represent a character. Characters formed words and commands in a
string. The string ate its own tail and learned to rewrite itself. These were
the early days of autopoiesis, the writing that writes itself. What we teach
today as the history of literature is also but an episode of a larger emerging
storyline. Its continuing evolution must involve the people of the word.
Each turn of the historical screw brings wonder, quickly fading into the
ordinary. Who remembers the ninth-century Persian mathematician
Muh˙ammad ibn Musa al-Khwarizmi, still with us in his al-gorithms? What
of Ramon Llull, the Majorcan monk, who studied rhetorical combination in
the thirteenth century? Did he invent one the earliest chatbots with his
rotating paper charts? Or do their origins lie in ancient divination charts,
like the zairajah described by the great medieval historian Ibn Khaldun, or
the Yi Jing (Book of Changes), written in the Western Zhou period?
Can I tell you also that Llull (pronounced yooy) inspired Francis Bacon
and Gottfried Leibniz to propose their binary cypher systems, influenced
also by the Chinese tradition? What about the Wunderkammern of the
German baroque poets in the seventeenth century? These curious drawer
cabinets—literally furniture—could be pulled in any arrangement to
produce beautiful music and poetry, to the delight and horror of the
audience. A long tradition of language machines also includes An Essay
Towards a Real Character, and a Philosophical Language by John Wilkins.
Written in 1668, the book imagines a new system of writing that would
function as a kind of a translation protocol between cultures, commercial
interests, and religions. Later, the mathematical notion of a Markov chain,
central to contemporary AI, developed out of a paper on Pushkin’s poetry.
These mostly forgotten artifacts remain in the background of
contemporary computation. Later, we’ll have the chance to pluck a few of
them out—unplug and hold them for a bit to see how the whole thing fits
together.
In writing this book, I was surprised to discover the history of computing
permeated by distinctly literary puzzles. But to see a past shared between
literature and technology has implied, also, the recognition of their altered
present. Technology continues to reconfigure human orientation to the
word. To study the history of literature today is to go beyond the familiar
categories, like fact or fiction, or even science and the humanities, art and
technology. Instead, our thread leads to an entirely separate branch of the
symbolic arts, practiced on a massive scale, on the work floors of
intellectual labor, across diverse industries. Here, we can find stories
produced by the MITRE Corporation, one of the largest military contractors
in the world, woven deep inside the Cold War–era missile defense systems;
encounter fairy tales written by airplanes as part of Boeing’s effort to
automate their incident-reporting systems; or visit the “vector space”
supporting everyday search-engine recommendations. To what genre do
these texts belong? On which bookshelf should we file them? How do we
read something so fundamentally unfit for human consumption?
The advance of artificial intellect obviously implicates a number of
open-ended metaphysical questions. From these, I want to highlight a
subset related to the subject of my expertise: literature, poetics, semiotics,
philology, textual studies, the study of inscription and incantation, e.g., et
al., etc.—the arcane arts.
Except not really that arcane, because though people these days read
fewer things labeled “Literature,” they consume more literature than ever.
Text weaves through the online digital world, stitching it together. It bonds
and bounds whatever is meant by artificial intelligence, in the way an
encyclopedia or a library holds the sum total of human knowledge. The
search engine or chatbot is just an outer garment, draped over the
complication within.
A body of written works sustains all organized human activity. I don’t
mean to suggest that all things are textual. The world becomes truly visible
to a literary scholar only through the looking glass of inscription. For
example, unlike your well-adjusted physician, I know little about and am
uncomfortable around bodies. A patient exists for physicians as a living
body first, but that’s not all a patient becomes in a hospital: Patient histories
(stories) may be recorded to supplement a physician’s notes. A recording of
their consultation will be transcribed and transcoded into a digital file. It
will be translated from one system to another, in multiple formats. It will be
redacted, edited, compressed, processed, and mined for missing billable
codes. This record will be sold and resold and used to train automated
diagnostic bots of the future. It will be summarized, cataloged, tagged, and
archived for later use. A body, in the sense of the literal Latin corpus, has
thus been transformed into the English corpus: figuratively, a collection of
textual records.
Much of this churn implicates automated, artificial, and otherwise
intelligent work. But neither the physician nor the hospital administrator
may consider it as such. Though a patient’s journey from body to record
leads directly to health outcomes (and to a hospital’s finances), it remains
without a guide, meandering through multiple, inherited legacy systems.
Here, a sharp-eyed arcanist may find a worthwhile mystery: Finally, a
puzzle for those who have studied exactly these sorts of things, like
corpuses and taxonomies, editions and redactions, catalogs, summaries,
translations, and archives—a rich trove of textual goodies!
In this book, we will similarly observe the history of artificial intelligence
through the looking glass of inscription. History tells us that computers
compute not only in the mathematical sense but universally. The number
was incidental to the symbol. In the 1840s, Ada Lovelace, daughter of Lord
Byron and one of the first “programmers” in the modern sense, imagined an
engine that could manipulate any symbolic information whatsoever (not just
numbers). A century later, Alan Turing extended that blueprint to imagine a
“universal machine,” capable of reading from and writing to an infinite
string of characters. The children of Turing and Lovelace occupying our
homes are therefore expert generic symbol manipulators. They are smart
because they are capable of interpreting abstract variables, containing
values, representing anything and everything.
Modern machines read and write. And because they carry language, we
readily imagine them capable of higher-order functions like sympathy or
sentience. But how do machines grow from handling inert textual matter to
living language? Why do they seem to exceed their programming by telling
jokes, making complex logical inferences, or writing poetry?
Metaphor holds the key to their slow awakening. Algebraic values can
signify human values: x = y arranges two variable entities into a familiar
relation; x can also be assigned to ideas like “cat” or “dog,” or an action like
“call home” or “order milk.” The art of manipulating symbols leads to a
game of metaphors in which certain inscriptions or gestures stand for
something else. Children, spies, and poets love to play games like these.
When I call (after Shakespeare) the night “thick,” I mean it to be
particularly dark, humid, and full of anticipation. The metaphor packs a lot
of meaningful baggage into a compact transport. Similarly, a machine
becomes smart by “rising higher and higher through the circles of spiraling
prose” (thanks, Winfried Georg Sebald!). Symbol emerges from inert
hardware, gesturing toward things that we care about, like flagging
offensive content, deleting spam, fighting terrorism, or indeed shopping for
shoes.
Care is also a value. But can machines really soar (crawl?) from simple
values like x and y to complex ones like care, equity, freedom, or justice?
Reader, are you human? Is there anything inside? Can machines think, or
are you just trolling me?
Letter Magic
WHEEL
Let me let you in on a little secret: koldun, the word meaning “sorcerer” in
several Slavic languages, arguably originates from the name of one of the
most prominent medieval (fourteenth-century) scholars, Abd ar-Rah ˙ man
ibn Muh ˙ ammad ibn Khaldun al-Hadhrami—the son of eternity from
Hadhram—or simply Ibn Khaldun. In his epic history of the world,
modestly titled, Muqaddimah or “Introduction,” Ibn Khaldun documented
the use of zairajah—“a remarkable technical procedure [. . .] for alleged
discovery of the supernatural”:
The form of the zairajah they use is a large circle that encloses other
concentric circles for the spheres, the elements, the created things, the
spiritualia, as well as other types of beings and sciences. Each circle is
divided into sections, the areas of which represent the signs of the
zodiac, or the elements, or other things. The lines dividing each section
run to the center. They are called chords. Along each chord, there are
sets of letters that have conventional (numerical) value [. . .] Inside the
zairajah, between the circles, are found the names of the sciences and
of topics of the created (world). On the back of (the page containing)
the circles, there’s a table with many squares, fifty-five horizontally and
one hundred and thirty-one vertically. Some of the squares are filled in,
partly with numbers, and partly with letters. Others are empty.
TABLE
So far, we’ve managed to dip our toes (fall off a cliff?) into the magical
waters of soothsaying and divination, in considering the medieval Arabic
Q&A bot—a set of “remarkable techniques” for language manipulation.
A careful examination of the zairajah and similar devices reveals several
distinct attributes, each worthy of its own little genealogy.
Note, for example, that despite its circular shape, the divination wheel can
also be seen as a type of table or chart. Charts aren’t as simple as they
appear. Unlike the messiness of presented information, on any topic, a chart
boils things down to their essentials. Charts are therefore by nature more
compact and more portable than the source of their original, non-processed
data. This property will become important later, when we consider their
capacity to encapsulate and transfer intelligence removed from its origins.
At the very least, a table presents information in a systematic way, usually
containing rows and columns. A chart implies order, however simple.
Here’s my fake shopping list, for example:
Product Price Quantity Type
Oatmeal $4.99 1 package Steel-cut
Chocolate $3.48 3 bars Dark
Blueberries $9.75 2 bags Frozen
Milk $3.15 ½ gallon Whole
Aside from being impressed with my organizational skills and passion for
antioxidants, an alien visiting our planet for the first time would learn that
foodstuffs fall under the category of “products,” and that products contain
such attributes as “price,” “quantity,” and “type.” A table therefore already
implies a sense of hierarchical labels, or a taxonomy. Moreover, the
taxonomy is “controlled,” in the sense of containing a limited number of
allowed values. Of course, I am not so formal with my shopping lists. You
would be surprised, however, to see milk requested in units of “cow”
instead of “gallons,” for instance, or blueberries of the type “decorative.”
By limiting the available language choices, a taxonomy attempts to capture
the relationship between things in the world, where products have prices
and chocolate is sold by the bar.
Taxonomies of all types were synonymous with medieval science, as well
as being an important topic in contemporary logic, medicine, law, and
computer science. Consider a field like medicine, where the infinity of
analog bodily sensations must be distilled into a controlled set of named
and billable conditions—a diagnosis. The American Psychiatric Association
publishes the infamous Diagnostic and Statistical Manual of Mental
Disorders, now in its fifth edition (DSM-5). Your insurance is unlikely to
cover a health issue not named in the manual. Conversely, aspects of the
human condition now considered in the range of normal (neurosis,
homosexuality) were once unjustly included, to the detriment of the people
they labeled. A part of ancient arcane arts, taxonomies still rule the world.
Think of electronic databases working in the background of every hospital
as giant zairajah circles, spinning a yarn that connects patients, physicians,
pharmacies, and insurance companies.
The greatest popularizer of divination circles in the medieval Western
world was Ramon Llull, who, in the wake of the Christian reconquest of
Muslim Spain, both borrowed and extended the system to create his “new
logic.” The Art, as he called it, made extensive use of rotating spheres that
could, by the use of a complex notation, represent knowledge about
anything and everything. In addition to being portable and categorical,
Llulian logic was symbolic, rule-based, and combinatorial.
Let’s take these individual properties slowly and by turn.
Start by examining the following partial table transcribed from Ars
Brevis,
2nd
1st Wheel Questions Subjects Virtues Vices
Wheel
B goodness difference whether? God justice avarice
C greatness concordance what? angel prudence gluttony
E power beginning why? man temperance pride
H virtue majority when? vegetative charity ire
Note that, initially, the symbols BCEH in the first column could be
“overloaded” or have multiple meanings, symbolizing questions, subjects,
virtues, or vices. For instance, B could, depending on its position, stand for
“goodness,” “difference,” “whether?”, “God,” “justice,” or “avarice.” Llull
explained that:
General Collection, Beinecke Rare Book and Manuscript Library, Yale University.
Smart Cabinets
T WO GERMAN BAROQUE POETS WALKED into a bar. Well, more like, one was
a poet and the other a scholar. And they corresponded by letter.
Athanasius Kircher, a neat man in his forties, spoke and dressed plainly. An
established figure by the time of their virtual meeting, he was slated to take
Johannes Kepler’s place at Habsburg court as the royal mathematician. In
years past, he had taught mathematics and ancient languages at Wurzburg.
Fate found him at the Collegio Romano in Italy, where his research program
expanded to cover geology, chemistry, and biology.
The eager attention from a flamboyant younger admirer clearly annoyed
him. The gaunt Quirinus Kuhlmann struck an odd figure. A childhood
speech impediment left him to pursue a career in letters with fervor. At the
University of Jena, known as a bit of a party school in the early sixteen
hundreds, he thought little of his peers or teachers. Instead of studying law
as planned, he cultivated an image of a brooding poet, claiming to receive
divine visions through illness and hallucination. Rumors of heresy
circulated, as well as of his “colossal egotism.”
At Jena, Kuhlmann’s scholarly brooding included the study of Llull’s
cryptics, as they appeared in Kircher’s work. By this time, and despite its
own brush with the Inquisition, Llull’s Art had spawned numerous
adherents across Europe. The emerging science of combinatorics promised
shortcuts to divine revelation. Under the spell of Llull, Kuhlmann wrote a
series of unprecedented technical sonnets, among them “Heavenly Love-‐
Kiss XLI,” meant to showcase his understanding of combinatorial
composition. I’ll translate a snippet for you here:
The sonnet’s concluding lines invite the reader to rearrange the words at
will—a radical gesture. The many resulting combinations produce
individual meaning, amounting to the sum total of human knowledge,
Kuhlmann explained. Despite such grandiose claims, Kuhlmann abandoned
the technique in other compositions, making me think he saw this particular
“love-kiss” as a technical experiment, not much more.
At stake in the virtual meeting between poet and scientist was Kircher’s
recent invention called the Mathematical Organ. Waiting for their ale, both
men contemplated the device in front of them. Made of polished, “artfully
painted” wood, the Organ resembled a large box. Opening its lid revealed a
row of labeled wooden slates, filed vertically: four columns of fifteen
narrow slates, four columns of seven wider ones, and four columns of five
of the largest pieces. When pulled out, each of the slates contained a string
of letters. By consulting the included booklets—which he called, wait for it,
applications—any combination of the planks cohered into a complete,
harmonious composition. Cleverly, depending on which manual was
consulted, the same organ could be used to compose music, write poetry or
secret messages, and even do advanced math, such as reckoning the Easter
calendar.
The conversation grew more heated with ale. Kircher took the same
systematic approach to all his research projects. The box was made
according to the latest scientific principles. He had recently sold a version
of the device to the young Archduke Charles Joseph of Austria. The child
would use it in his studies, with the help of a trusted tutor. Such devices,
Kircher argued, were easy, delightful, and instructive. They were, above all,
useful, allowing untrained operators to create and to perform. Technology
was there to serve, making difficult knowledge accessible to a larger
audience.
Kuhlmann objected. The path to knowledge should remain tortuous,
accessible only to those (like him!) willing to walk it properly. The box is
just an ingenious game, my ingenious Kircher! Kuhlmann was starting to
slur his words a little. The intelligence lies with you. Without the box, the
young duke remains an idiotic parrot (I’m sorry, he really did write that,
only in Latin). Nothing is retained on the inside. And without inward
understanding, the child captures neither knowledge nor intelligence.
A part of the Mathematical Organ, as described by one of
Kircher’s students, P. Gaspar Schott. Printed by Johann Andreas
Endter in Würzburg (1668), 135.
One of the “application” tables included with the device. Schott
(1668), 93.
One would hope, in another timeline, the young Kuhlmann was not
planning to travel Europe to say such things about royal children out loud,
especially not in the presence of, let’s say, a certain fledgling and repressive
Eastern regime. Though his talents were recognized in Germany, he would
later burn at the stake in Moscow, by the order of Ivan V, denounced by his
fellow Lutherans for iconoclasm.
We arrive then, as always, at the ghost in the shell—of smart furniture.
Clearly, neither poet nor scientist thought of the Mathematical Organ as an
animated entity. They differed rather in the idea of the human involved. For
Kuhlmann, the tooling edge of combinatorial poetry aimed inward, toward
the reconfiguration of the spirit. Combinatorial sonnets helped precipitate a
mercurial spiritual transformation, difficult for us to recreate sober and in
my translation. For example, the sonic parallelism between Haze and Blaze
(which I mirrored from the German), encouraged readers to seek further,
conceptual similarities between smoke and fire. Name and physical
property momentarily aligned. New, unexpected connections were meant to
be made through private internal reconfiguration.
By contrast, Kircher’s machine con-figurated externally, producing a
specified rational effect—be it the calculation of numbers or composition of
poetry—which also happened to be useful, elegant, or delightful—all
outward qualities.
The two views on the machine tell us something important about
intelligence, which manifests both as a private experience of inward
understanding and its outward, instrumental effects on the world. Let’s take
a moment to consider both. On the former view, we’ll call it Platonic,
intelligence means something like “the correct internal alignment of
thoughts and feelings with universal truth.” On the latter, we can call it
Aristotelian, it implies “a universal capacity to achieve specific results.”
Notice something strange?
In the Platonic model, intelligence works as the SOURCE of action. It
therefore necessarily exhibits qualities of being private and local, ineffable
even. Therefore, paradoxically, Platonic universal rationalism tolerates
contingent, nonrational paths to enlightenment. Of course! I think to myself.
There is no one general way of being smart. How would that work?
Sometimes I can’t even articulate how I’ve come to understand certain
things. It just happens or doesn’t. In that sense, our intelligence is bound by
our individual bodies, education, and personal history. It therefore resists
one single universal description.
Private experience doesn’t have to be wholly subjective or arbitrary.
Outside access tends to be difficult, however. How can a teacher glimpse a
student’s view from the inside, so as to better ascertain whether they’ve
learned something properly? Plato—well actually, the fictional character
called Socrates—famously mistrusted all writing in general and rote
memorization in particular (despite being a memorable writer). An often-‐
cited dialogue recalls Socrates encountering Phaedrus, a student of his, who
was on the way back from another lecture. When Socrates asked him about
the content of that lecture, Phaedrus retold the words from memory.
Socrates suspected a parrot, however, and proceeded to question Phaedrus
poignantly, exposing the limits of his student’s understanding. Similarly,
any fixed, mechanized, external trappings of acumen risk being
disconnected from the source. Instead of smarts, you get parrots, who, also
according to Descartes in his Mediations, merely repeat without
understanding.
There can be no shortcuts on the road to knowledge. A student achieves
internal change through mental toil. The Platonic is therefore necessarily
also a human-centric view, because, on the inside, I don’t know how to be a
bat or a rock or a smart table or anything else. Platonic intelligence implies
appropriate personal transformation.
In the Aristotelian model, intelligence is the GOAL of thought. It doesn’t
matter how it is achieved at the source. One jumps into the collective pool
called “intelligence” through appropriate action. And all sorts of stuff
besides humans float in that pool: bats, tools, automatic transmissions,
smart tables, mathematical organs, smart phones, etc. This makes sense to
me also, because when I think about stuff, I usually don’t do it on the inside
only: I pace, take notes, read books, and talk to friends. Looking things up
on Wikipedia, consulting a table, following a certain procedure I learned
from a textbook, or evoking a “smart assistant”—none of these detract from
intelligence in the Aristotelian sense.
It would be foolish in fact to refuse such outside assistance. I cannot
imagine knowledge that develops in complete isolation from the world.
(Incidentally, the twelfth-century Andalusian treatise on the “Self-Taught
Philosopher,” titled Hayy ibn Yaqzan in the original Arabic by Ibn Tufail,
was one such attempt, prefiguring the modern “castaway” genre,
represented by Daniel Defoe’s Robinson Crusoe (1719) or television series
like Lost (2004–2010) or the film Cast Away (2000), starring Tom Hanks.)
Science itself would be impossible in isolation because it produces
intelligence in a way that’s fundamentally social and collective. By
developing novel tools and techniques for externalized thought, like
Aristotle or Kircher, a scientist contributes to the overall shared capacity for
reason. In that sense, something becomes right or true by its measure of
accordance to an established body of work. Intelligence in the Aristotelian
sense is subject to external review. We don’t have to concern ourselves with
“what’s really going on, on the inside.” If it shimmies and quacks like a
smart duck, it’s a smart duck.
That the two senses of intelligence—Platonic and Aristotelian, internal
and external, private and public—both use the same word causes much
confusion. It would be simpler if we had at least two distinct words for the
Platonic and the Aristotelian definitions, capturing the difference between
intelligence as the SOURCE of and the GOAL of action.
Baroque birds of all sorts bring us a step closer to the story of modern
digital computers, able to mimic any other “discrete-state machine”
universally. Called a “universal machine” by Alan Turing, the computer
promises to model general intelligence, regardless of the task it was
originally designed to perform. Recall the obsolescence of single-purpose
machines like calculators, cash registers, and digital music players. The
universal computer ate them all. Similarly, general intelligence refers to the
ability of an individual or machine to learn, reason, and problem-solve
across a wide range of domains and contexts. In this sense, general
intelligence is often presented to us in the guise of “universal intelligence,”
intelligence capable of performing any cognitive task.
Given the discussion above, the words general and universal should be
treated with suspicion. Universalism, for all its strengths, comes with a host
of problems, already apparent in our case studies.
From the Platonic, internalist perspective, how smarts are achieved
matters. I might value good grades on an exam, for example, but not if they
were the result of cheating. Conversely, I don’t usually care how my
refrigerator thinks, only that it does its job. Anything goes for a smart
machine, evaluated on efficacy alone. A machine may perform intelligence
by any obscure means. My students must show their work.
The “private” ways of machine intelligence also may not necessarily
agree with my own, or that of my family or community. Efficacy sometimes
comes at a cost. For instance, it may be that my television gets smarter by
continually recording conversations in my living room, so as to better tailor
its recommendations to my family. These means do not justify the ends for
me. The so-called alignment problem stems from an impassible paradox:
generally universal intelligence does not always align with local values.
Smart in the general sense might not be so smart in the specific context. We
should therefore always be cautious with claims for intelligence in the
general, universal sense.
From the Aristotelian, externalist perspective, the idea of an intelligence
considered “universal” can be equated to “generally applicable.” General
application contains no impassable paradoxes, with one huge caveat: where
it gains in universal application, it loses on smarts.
Tasks that can be automated become devalued. Witness numerous smart
devices, like toasters and thermostats, that modern humans treat with
disdain. While inspiring awe initially, smart technology of the past becomes
mindlessly assistive, no matter the expanded ingenuity. The same can be
said for the tools of intellectual trade. Dictionaries, grammars, thesauruses,
and encyclopedias were once hailed as monumental national achievements.
Today, they are silently integrated into digital autocompletion or
autocorrection tools. We don’t even think of them as sophisticated
technology.
As our tooling improves, the functional definition of intelligent action
changes as well. Because the externalist perspective is grounded in the
world, general intelligence recedes beneath the rising tide of average
achievement. For instance, at one time, the mere fact of being literate would
have been considered exceptional. One could build a career of simply
reading and interpreting texts for an illiterate public. With mass literacy,
skills of reading and writing migrate from the upper to the lower bounds of
intellectual work. Similarly, while the ability to memorize complicated
calendar patterns may have been impressive a few centuries ago, spending
time on unaided astronomical reckoning today would be considered foolish,
with the most advanced astronomers being those able to use assistive
technology (like powerful computing) most effectively. In the applied
world, intelligence bubbles above shared capacity. But no matter how
cleverly effervescent, once a smart tool reaches general adoption, it
dissipates into the surface of baseline intelligence.
Here we reach the paradox of the functional idea: Exceptional
performance cannot become universal, by definition. A “smart” device is
merely smarter than the previous generation. Once adopted widely, it passes
into the average.
UNIVERSAL CHARACTER
The alphabet, as the author admitted after transcribing the Lord’s Prayer
into his system, might not have turned out as practical as he had planned.
Still, he hoped it delivered a distinct advantage over European alphabets,
mistakenly believing Chinese or Arabic writing systems to be somehow
more intrinsically closer to his ideals. The Real Character would render
each word precisely, according to the classification of things and
relationships located in the main table. And though Wilkins could not
achieve a perfect analogy between word and thing, he thought that further
improvements of his system would someday enable great advances in
scientific communication, commerce, and even facilitate peace between
nations.
Judging by the many wars and disagreements since, this experiment
failed. Still, Wilkins sowed yet another seed of universal language. And it
would bear fruit in the universal machines made by Leibniz, Babbage,
Lovelace, and, later, Turing.
STEPPED RATIOCINATOR
In 1679, Gottfried Wilhelm Leibniz wore a frizzy wig. Like other great
men, he was modest about his accomplishments. “I’m not sure why
Aristotle or Descartes didn’t consider this amazing thing I did while still a
boy,” was a common refrain in his letters. Though, to be fair, he did
accomplish impressive feats like developing calculus (contentiously, in
competition with Isaac Newton), inventing a novel binary notation (under
the spell of Chinese and Egyptian sources), and proposing the blueprints for
his own Characteristica universalis. The latter combined Kircher’s
Mathematical Organ and Wilkins’s Real Character into a single, unified
system.
We stand at a crossroads with Leibniz because, from here, the history of
artificial reason splits into at least two diverging paths. The road more
traveled leads to the development of modern calculus (literally “the small
pebble” in Latin) through the work of Maria Agnesi, Augustin-Louis
Cauchy, Louis Arbogast, Bernhard Riemann, and John von Neumann. The
larger, though now somewhat neglected, pebble rolled down the road of
universal language, leading directly to modern conversational AIs.
In 1679, Leibniz revived an earlier, monumental project, a universal
encyclopedia, aimed at healing the rift between Protestant and Catholic
churches. He called it Plus Ultra (Further Beyond), in which, like Wilkins,
he aimed to assemble (though never completed) a “precise inventory of all
the available but dispersed and ill-ordered pieces of knowledge.” Unlike
Wilkins, Leibniz concentrated his efforts on the logic of composition—the
middle, least-developed part of Real Character by Wilkins—imagining a
kind of a linguistic calculus, which he called the “General Characteristic.”
“Learned men,” he wrote, “have long since thought of some kind of
language or universal characteristic by which all concepts and things can be
put into beautiful order.” In that task, he hoped to succeed where Wilkins
failed.
Thus, in parallel to his writings on mathematical calculus, Leibniz
nurtured the far more ambitious dream of a language calculus, capable of
both “discovery and judgment, that is, one whose signs or characters serve
the same purpose that arithmetical signs serve for numbers.” This new
language would increase the power of the human mind, far more so than
microscopes or telescopes amplify sight. Without it, we are like merchants,
he wrote—in debt to each other vaguely for various items mentioned in
passing, but never willing to strike an exact balance of meaning.
For the aims of striking an exact balance, Leibniz proposed a kind of a
“complication,” or a mechanism, amounting to something greater than the
sum of its parts. “When the tables or categories of our art of complication
have been formed,” he wrote, “something greater will emerge. For let the
first terms, of the combination of which all others consist, be designated by
signs; these signs will be a kind of alphabet [. . .] If these are correctly and
ingeniously established, this universal writing will be as easy as it is
common, and will be capable of being read without any dictionary; at the
same time, a fundamental knowledge of all things will be obtained.” More
precise than modern-era artificial languages such as Esperanto, Leibniz
imagined a truth-telling mechanism, able to automate philosophic thought
itself.
Let’s pause here to breathe and reflect on the story so far.
Medieval rhetorical wheels worked by chance combination. For some,
like Ibn Khaldun, they were nothing more than clever parlor tricks, at best
generating a few surprising twists of the tongue. Consider that today’s most
advanced language AIs continue to face the same kind of skepticism:
Indeed, a mechanical parrot reading from a script written by a typing
monkey may occasionally articulate something profound.
Llull and later his followers in the Age of Enlightenment saw the
possibility of something greater, beyond mere combination. Given logic,
language could be brought under a rational system, and therefore
automated, in a way that benefits humanity. Metaphorically, this device
would be called science, or calculus, or logic. But the dream of a universal
language manifested itself also as a family of literal devices, made to
amplify the power of reasoning through automated language production. Its
blueprint, extant in the most advanced of AI algorithms today, indicates,
initially, learning—the systematic assimilation of all human knowledge.
The device then attempts to translate the mess of human expression into
some sort of a comprehensive middle language, expressed by rigidly
defined characteristics, in the manner of Wilkins or Leibniz. Finally, the
logic of composition could be derived to produce novel discourse—new
ideas and answers to questions not encountered in the original learning
phase.
That was the dream, at least. As we will see going forward, major
metaphysical points of failure persisted despite technical advances.
Contemporary machines assimilate and compute over vast amounts of text.
They are also still tied to the paradoxes inherent in the idea of a universal
language.
Language retains its whimsical connection to the physical world. Its rules
can be fanciful or arbitrary. They change, as what it means to be human
changes in response to a changing world. Intelligence struggles with
universal application, in addition, because the world cannot be the same
everywhere at once. The world itself attains universality only from a great
distance, described in broad strokes by physics or theology. Its particulars
often differ depending on the local context.
Yet universals are useful also, insofar as they force us to abstract away
from private points of reference, toward shared ideals. An automated
sentence-completion gadget may make you a “better” writer. And one may
even playfully call it a “good” writer. But let’s not forget that any content of
“goodness” resides within a community, interested not only in functional
outcomes (the production of more prose), but in the implicit character of
good writers and readers. That’s why we still ask our children to learn math
without a calculator or to compose by hand. Beyond general intelligence,
we want our children to become exceptional.
The instrumental goal of calculation or composition stands secondary to
the intrinsic value of personal growth, achieved in the struggle of learning.
The hand carries the load of value through lived experience. And
experience cannot be automated, even as we witness the emergence of
clever devices—back at the inn, with our baroque friends, tinkering with the
intricate cabinetry of language manipulation.
CHAPTER 4
Template Culture
S TOP READING HERE IF YOU love literature but hate thinking about how it
is made. This might be an unsavory story about making sausage.
Though perhaps it’s not so bad, because great art, like great cuisine, will
always remain the purview of exceptional talent. Average chefs do most of
the cooking, however, just as most writing is done by ordinary authors. I
don’t mean to diminish the study of great literature in any way. We ignore
the average at our peril, however. The failure to understand the mechanisms
of ordinary authorship holds dire consequences for the quality of our
collective intellectual experience. Artificial intellect thrives in the gap
between the average and the exceptional. There’d be no need for calculators
if we were all mathematical geniuses. AI was created specifically to make
us smarter. Spell-checkers and sentence autocompletion tools make better
(at least, more literate) writers.
In considering the amplification of average human capacity for thought,
both Babbage and Lovelace circled around the idea of the Jacquard loom,
an innovation in weaving manufacture that used perforated “operation
cards” to pattern its designs. Within the gears of the Analytical Engine,
which Lovelace called “the mill,” an operation card would rearrange or
“throw” the mechanism “into a series of different states,” “determining the
succession of operation in a general manner.” A different type of a card,
called the “supply card,” held the data—which, recall, could represent any
object—“furnishing the mill with its proper food.” The mill processed its
supplies according to instructions punched into operation cards. These
could even be swapped mid-operation to perform nested functions—an
eclipse within the regular moon calendar—Babbage’s exceptional miracle.
The connection between symbolic logic and commercial weaving wasn’t
accidental.
In 1832, still in the early days of the Difference Engine, Babbage issued a
missive from Dorset Street. He called it a “direct consequence” of his work
on the machine, compiled in preparation for a proposed lecture series at
Cambridge. Titled On the Economy of Machinery and Manufactures, this
lengthy tome contained a comprehensive study of all known industrial
manufacturing processes. Reading it has significantly changed my
perspective on the Analytical Engine.
Where at first I saw it in the lineage of thought-manipulating machines,
via Kircher and Leibniz, I could now also perceive its ambition on the scale
of macroeconomics, in the broader context of the industrial age.
The Analytical Engine was to achieve for the world of letters what the
Jacquard loom has done for the commercial weaving of fabrics. This was no
mere metaphor for Babbage. Nor was it solely his invention. From the
nineteenth century onward, the notion of template-based manufacturing
permeated all human industry, and especially the manufacture of consumer
and capital goods, like clothing, furniture, machinery, or equipment.
Yet seldom do we consider templating in the production of symbolic
goods, like literature, film, music, philosophy, or journalism. There’s just
something unsavory about the thought of individual human genius being
diminished by mechanical reproduction. Several modern high-art or avant-‐
garde movements have even defined themselves explicitly in opposition to
industry. For the Romantics, like George Byron or William Wordsworth,
and the modernists, like Bertolt Brecht, Franz Kafka, Walter Benjamin, or
Virginia Woolf, to be human was to raise an exception. There, in the
rarefied heights of individual genius, the automated sank beneath the
surface of creativity, much less intelligence, in favor of the handcrafted and
the extraordinary.
Scholarship in the humanities inherited the Romantic emphasis on the
exception at the expense of the automated, and therefore the instrumental
and the collective. Many everyday practices of reading, writing, and
interpreting texts—besides fiction—have fallen out of scholarly purview as
a consequence. Not much research is being done on medical literature for
instance, self-help, or on collaborative writing practices in the television
industry.
If there has to be one, take it as this book’s punchline—intellect requires
artifice, and therefore labor. Innate genius can neither be explained nor
imparted. The artificial seems autonomous because it arrives removed from
its singular, inimitable point of origin. Artifice identifies precisely those
aspects of an art subject to be explained, documented, and transferred to
others. Consequently, intellectual work participates in the history of
automation, affecting all trades at some point of its development. Like the
making of shoes or automobiles, the production of intellectual goods has
long been moving from bespoke workshops to the factory floor. Once we
understand intellectual labor as labor, we should not be surprised to find
tools, templates, and machines near it. But because mind and language are
special to us, we like to pretend they are exempt from labor history. This
book so far has told and will tell a story about labor, not just robots or
literature.
Reams have been written on the minute-by-minute writing schedules of
exceptional authors. Yet little is known of the tools augmenting much of
everyday knowledge work. Templates support the labor of ordinary intellect
everywhere, from the physician’s office to the photography studio and the
TV writers’ room. Their use was developed systematically at the time of
Babbage, in every human industry, including the intellectual. Tables,
schemas, skeleton forms, molds, patterns, matrices, and frames lie at the
basis of all content production. The templates driving engagement on
modern social media are just the most recent of a long-standing historical
phenomenon. The end of the nineteenth century in fact saw an explosion of
template-based art. Though their use may have been a cause of personal
embarrassment or even professional taboo.
Few artists like to admit to painting by the numbers. Nobody wants to
seem ordinary. The occasional visibility of artifice—portable, explainable,
documented, transferable, automated—therefore tends to startle or repulse
audiences acculturated into the privilege of exceptional human genius.
But it also shouldn’t, because an emphasis on those aspects of an art that
can be transferred, instead of inherent talent, lies at the basis of a
democratic education.
SKELETON FORMS
In his Economy, Babbage surveyed the state of automation in the first half
of the nineteenth century, without nostalgia for special handcrafts of times
past.
Here, “skeleton forms” performed the same function on the
macroeconomic level, as did “operation cards” on the level of the machine.
The difference between “making” and “manufacturing,” for Babbage, was
found in the ability of makers to rationalize their creative process, step by
step, in a way that could be optimized and replicated.
In the author’s usual diligent fashion, major portions of the book were
dedicated to the exhaustive enumeration of various copying processes,
including: the printing from cavities, copperplate printing, engraving on
steel, music printing, calico, stenciling, printing of red cotton handkerchiefs
(?), printing from surface, block printing, moveable type, stereotype,
lithography, iron casting, casts in plaster, casts of vegetable productions in
bronze (!), casting in wax, imitations of plants, the molding of bricks, tiles
and cornice of the Church of St. Stefano, embossed china, glass seals,
square glass bottles with names, wooden snuff-boxes, horn knife and
umbrella handles, swagging, engraving by pressure in steel, forgery of bank
notes, copying by stamping, coins and medals, military ornaments, cliché,
copying by punching, boilerplates, tinned iron, buhl-work, steel chains,
copying with elongations, wire drawing, brass tubing, leaden pipes,
vermicelli, copying with altered dimensions, pentagraph, rose engine
turning, lathe, shoe lasts, and veils made by caterpillars (!?).
In conclusion, Babbage suggested his readers undertake a systemic study
of automation by using the provided “skeleton form,” binding hundreds of
responses into a pocket book of surveyed manufacturing templates. This
skeleton of skeletons was made of questions, documenting the bare-bones
outline of any manufacturing process. Its blanks left space for the number
of workers employed, hours worked, tools and maintenance involved,
maintenance, the division of labor, the list of operations and the number of
times each was repeated, among other exacting elements of business logic.
Among the industries queried, Babbage made special note of the printing
process itself, in the “copying through six Stages in Printing this volume.”
“It is here that the union of the intellectual and the mechanical departments
takes place,” he wrote, recalling Lovelace—in a pattern, on a template,
along the cells of a table—“down to the cavity punched through the letters
a, b, d, e, g, &c.”—“stamped, died, and formed”—“these movable types,
the obedient messengers of the most opposite thoughts,” where the most
conflicting of theories are “themselves copied by casting from moulds of
copper, called the matrix.”
Templates were everywhere in the industrial age. Viewed in their light,
the novelty of analytical engines lay in the application of manufacturing
templates to the domain of symbolic logic, philosophy, and art. Templates
used in the production of symbolic goods necessarily share a history with
physical manufacturing, influenced by the same industrial forces affecting
the manufacture of material goods—such as the division of labor,
mechanization, assembly line production, lowered barrier to entry,
decreased cost, increased output, efficiency, and standardization. (But also,
the rise of middle management, globalization, union busting, and lame
corporate team-building events.)
Literary production in the West accelerated with the advance of the
printing press, alongside rising literacy rates. Increasingly, folks were
beginning to write things for fun and profit, not just edification. The
eighteenth and nineteenth centuries saw an expansion of literary markets,
with the corresponding emergence of popular genres—from self-help to
guidebooks and travelogues to books on homemaking and home
improvement, farmers’ almanacs, how-to manuals, children’s literature,
pamphlets, true crime, detective fiction, and pornography. Magazine and
journal subscription was also becoming a trend at home, increasing both the
demand for and the supply of literary content.
Allow me to bore you with some numbers, for a sense of scale. Authors
prolific by eighteenth century standards wrote on the order of a dozen or so
major literary works. Among them let’s count Daniel Defoe, author of
Robinson Crusoe, who wrote, not counting the pamphlets, around eight
novels and sixteen works of longer nonfiction; Jonathan Swift—no stranger
to text generators found his Gulliver’s Travels—published fifteen; Goethe,
six novel-length pieces alongside as many theater plays; Jane Austen,
authored nine long novels; Laurence Sterne, under ten, depending on how
you count Tristram Shandy; and Sir Walter Scott, of Waverley fame—seven
novels. You can probably find an outlying graphomaniac or two somewhere
in addition, without affecting these averages.
The word count began to climb dramatically in the nineteenth century.
Dickens and Dostoevsky, whose outputs were the envy of contemporaries,
and who both wrote serially for periodicals, penned fewer than twenty
novels each. The Soviets published Count Tolstoy’s collected works in
ninety volumes. Alexandre Dumas wrote more than a hundred plays and
novels, likely with the help of ghostwriters.
By the first half of the twentieth century, such exceptional output became
the norm. Georges Simenon, the Belgian writer of detective fiction,
published several hundred novels, 75 of them about French police detective
Jules Maigret. Elizabeth Meade Smith, the author of A World of Girls
(1886), wrote more than 300 mysteries, romance novels, and works of
young-adult fiction. Paul Little, more than 700 novels. Kathleen Lindsay
(Mary Faulkner), more than 900, under many pseudonyms. Edward L.
Stratemeyer, the American author-producer behind the writer syndicates
responsible for such popular series as The Rover Boys and Nancy Drew, in
the thousands. The Spanish author of romance novels Corin Tellado, a fast
4,000.
Consider a few additional points of reference, quoted here from my
published research, where you can check my receipts. The total number of
titles published in the United Kingdom grew from almost 3,000 in the
1840s to more than 10,000 in the 1900s. (Yes, that’s right: individual
authors in the twentieth century eclipsed the production of whole countries
a few decades prior!)
According to my own figures, based on the American Catalogue of Books
in Print, the total number of titles printed in the United States (including
translations and reprints) grew tenfold, from around 2,000 in 1876 to
17,000 in the 1930s. True figures were likely higher, because the official
record omitted popular “low” genres, like true crime and pornography. A
reference I found in the Writer’s Digest from 1928 mentioned one pulp-‐
fiction publisher, Street & Smith, receiving “close to 900,000 manuscripts a
year.” Another was reported to buy “nearly a million words a month for
their various pulp fiction magazines.” “Fiction House authors draw real pay
from this office,” an editor reported, citing figures as high as $5,000 a year.
At the going rate of one to two pennies per word, the sums translate to an
average author placing on the order of 250,000 to 500,000 words per year
—a volume equivalent to five lengthy novels annually, not counting
rejections.
Something clearly changed on the supply side of literary production in the
short time between Jane Austen and Writer’s Digest.
Discounting quality, concentrate for a moment on the sheer textual output
in terms of classical economics. The rise in the demand for printed
materials can be explained, among other causes, by the printing press,
Protestant Reformation, the Enlightenment, rising literacy rates,
urbanization, and educational reform. This wedge of the economic lemon
—the demand side—has been squeezed well in other studies.
Seldom do we consider the supply. Though it changes through time, the
matter of literary composition remains roughly static by volume, at least
since Aristotle (whose Poetics is still taught widely today). As my slow
progress on these chapters can attest, natural human mental physiology
(body and brain) place hard limits on an individual’s capacity for creativity
(I need a break).
Under the conditions of constrained supply, one expects the rapid increase
in demand to push prices and wages up. A good word should, by economic
logic, cost the same or more with the expansion of the market. That’s not
what we observe. Author wages (per word) instead fell dramatically toward
the end of the nineteenth century. Where, in the 1850s, writing for serial
journal publication, the young Charles Dickens garnered around a dollar per
word (according to my rough estimates normalized to American currency
today)—in the 1920s, a decent American author would be lucky to expect a
few cents. And today most authors write for free, online.
The fall of wages, instead of the expected rise, implies the increase in
productivity, beyond baseline physiological capacity. As with other labor
dynamics of the time—remember the Jacquard loom!—we must also look
to the methods of production for an explanation, on the side of the supply.
A DRAMATIC SITUATION
Meet another one of our lovely weirdos, then, in the spirit of democratic
hospitality. History was not kind to Georges Polti. In 1895, this little-known
French writer published a small book meant to help playwrights compose
new plays, called The Thirty-Six Dramatic Situations. As you’d expect, it
contained a list of thirty-six dramatic situations, such as “supplication,
deliverance, vengeance, pursuit, disaster, revolt, the enigma, rivalry,
adultery,” etc. Each situation was introduced briefly, in broad strokes,
containing several further examples Polti collected from the classics.
For example, in the section on pursuit, Polti explained that our interest
should be “held by the fugitive alone; sometimes innocent, always
excusable, for the fault—if there was one—appears to be inevitable,
ordained; we do not inquire into it or blame it, which would be idle, but
sympathetically suffer the consequences with our hero, who, whatever he
may once have been, is now but a fellow-man in danger.” Four subtypes
were further enumerated, including: (a) “fugitives from justice pursued for
political Offences, etc.”; (b) “pursuit for a fault of love”; (c) “a hero
struggling against a power”; and (d) “a pseudo-madman struggling against
an Iago-like alienist.”
Anticipating how this might sound to his high-minded colleagues, Polti
didn’t apologize. “They will accuse me of killing imagination,” he wrote.
“They will say I have destroyed wonder and imagination. They will call me
‘the assassin of prodigy.’ ” Far from it! Without method, the aimless
imitation of the past stifles creativity. “All the old marionettes reappear,
inflated with empty philosophic ideas.” Only a systematic exploration of the
unknown could produce truly novel artistic forms. But novelty for novelty’s
sake was also meaningless. His method, in addition, trivialized the very
idea of novelty. Authors would now concentrate on ideals higher than mere
innovation—ideals like beauty, balance, and harmony.
Armed with the scientific method, playwrights of the future could employ
countless possible dramatic combinations, “ranged according to their
probabilities,” Polti predicted. Further, rigorous experimentation was bound
to change art in fantastical ways, by “analyzing orders, systems and groups
of systems”—as yet unseen methods for literary production would emerge
—“measured and classified with precision.”
By the book’s end, Polti revealed the practical impact of his project.
Many situational arrangements, “their classes and sub-classes,” were at
present ignored, he wrote. Modern plays are too boring and too familiar to
the audience. Many more situations “remain to be exploited” in
contemporary art. Fuzzy Romantic terms like “imagination,” “invention,”
and “composition,” must spring naturally from the “art of combination,”
furnished by the composite building blocks of his system.
And then the surprising conclusion: “Thus, from the first edition of this
little book, I might offer (speaking not ironically but seriously) to dramatic
authors and theatrical managers, ten thousand scenarios, totally different
from those used repeatedly upon our stage in the last fifty years. The
scenarios will be, needless to say, of a realistic and effective character. I will
contract to deliver a thousand in eight days. For the production of a single
gross, but twenty-four hours are required. Prices are quoted in single
dozens. Write or call, No. 109, Passage de l’Elysee des Beaux Arts.”
I don’t know if anybody took Polti up on the offer to buy scenarios by the
dozen (people were still wearing striped long-pants to the beach), though I
combed the archives for months to find out. Given his lack of commercial
success as a playwright, Polti’s name was rarely mentioned in scholarly
literature, and when it was, was quickly dismissed. This thread seemed to
lead to a dead-end. I was beginning to lose hope when an important clue
soon unraveled the case.
The foreword to the English edition of Polti’s work was written by one
William R. Kane—publisher and editor of the Editor: Journal of
Information for Literary Workers. The title caught my attention, because
“literary workers” was not the way I was taught to discuss great authors.
Published first in 1896 out of Franklin, Ohio, the magazine opens a view
into the American literary market at the turn of twentieth century. Full of
adverts for typewriters and filing systems, the Editor published insider
pieces with titles like “500 Places to Sell Manuscripts!,” “What Young
Writers Should Read,” “Consideration for Editors,” “Does Poetry Pay?,”
“The Historical Short Story,” among other topics of interest for professional
authors. Advertisements for Polti’s slim volume ran here for decades, along
with many other, similar, “how to write a best-seller” schemes. Though
rarely mentioned in the annals of high modernism (the literary movement
usually associated with this period), the practice of “writing by the
template” prevailed among literary workers. Their output would fuel a mass
phenomenon, spreading across numerous pulp-fiction magazines and soon
through the Hollywood blockbuster machine.
The Editor sat on top of an iceberg of similar publications. Weeks of
tedious microfiche trawling at the New York Public Library produced many
others, including the Author’s Journal (1895) out of New York; the Author
out of Boston, Massachusetts, founded in 1889 and described as “a monthly
magazine to interest and help all literary workers”; the Writer also out of
Boston, established in 1887 and still active today; the Publisher (1891); the
Writer’s Monthly (1925), affiliated with the Home Correspondence School
in Springfield, Massachusetts; Pennsylvania’s Writer’s Review (1925);
Markets and Methods for Writers (1925) by the Palmer Institute of
Authorship, a correspondence school out of Hollywood, California; and
New York’s own Author’s Digest (1927).
Further, advertisements in the above journals led me to dozens of “how-‐
to” books similar to Polti’s. Remarkably, these were mostly lost to literary
history, hardly mentioned in research, nor collected by university libraries.
The record was mostly silent on their existence.
The following titles give us a good sense of their content. Consider,
within the fullness of primary sources, a list of selected works, including:
Skeleton Essays, or Authorship in Outline (1890) by Thomas English;
James Knapp Reeves’s Practical Authorship (1900); The Technique of the
Novel (1908) by Charles Horne; Writing the Short-Story: A Practical
Handbook on the Rise, Structure, Writing, and Sale of the Modern Short-‐
Story (1909) by Joseph Esenwein; Harriott Fansler’s Types of Prose
Narratives: A Text-Book for the Story Writer (1911); Henry Phillips’s The
Plot of the Short Story (1912); The Technique of the Mystery Story (1913)
by Carolyn Wells; The Technique of Play Writing (1915) by Charlton
Andrews; The Technique of Fiction Writing (1918) by Robert Saunders
Dowst; Plots and Personalities (1922) by Edwin Slosson and June
Downey; and William Cook’s Plotto: A New Method of Plot Suggestion for
Writers of Creative Fiction (1928), among many other examples.
This was finally the machinery of literary production on an industrial
scale. Keywords repeated consistently across the above archive recall those
surveyed by Babbage in his Economy: template, method, technique,
skeleton form, mold, schema, system. These tools were missing from the
scholarly record because authors who made it out of the “fiction factory”
rarely revealed the dirty secrets of their trade. Painting by the numbers
diminished the myth (and the earning potential) of any artist. It wasn’t good
for one’s career. And so the world marveled at Vladimir Nabokov’s
exceptional use of index cards in the making of his novels, while the pages
of the Editor routinely pushed ordinary articles about “using index cards to
author your bestseller,” alongside advertisements for home card-filing
furniture.
Now, I did hunt down and read most of the books above, so you don’t
have to. Some of them were pretty bad. Others offered reasonable advice.
The techniques proposed within—the machinery, so to speak—generally
fell into one of several categories.
The first included “skeleton forms” proper—consisting of near-finished,
prefabricated pieces that could be easily assembled and modified lightly to
produce reasonable outputs, such as letters, essays, or short stories. These
were sometimes meant to—ahem—“aid” in the production of graded
university papers, as was the case with Skeleton Essays, or Authorship in
Outline (1890) by the aptly named Thomas English.
The second involved the revival of general Aristotelian formulas,
procedures, and “rules of thumb”—derived for the construction of
“balanced” compositions, whether in fiction or for general rhetorical
impact.
From these, it may be useful to highlight The Technique of Drama (1892)
and The Analysis of Play Construction (1908) by William Thompson Price,
a self-proclaimed “Aristotle from Kentucky.” Price was especially
entertaining to read because his advice for young writers was backed by his
experience as a play-reader for several major Broadway theater production
houses. He therefore peppers his rules of composition with negative
examples from “a bottomless pit” of rejected manuscripts on his desk: how
NOT to write a best-seller. Price’s manuals were adopted by several
influential theater programs, just then coming into existence, such as the
one at Harvard, taught by Professor George Pierce Baker from 1905
through the 1920s. While met with skepticism at first—the very idea of
teaching art systematically was controversial—Baker’s students found
commercial success on Broadway and in Hollywood. (Among them,
Believe Me, Xantippe [1918], a silent-screen comedy starring John
Barrymore, Drew Barrymore’s grandfather.)
Apart from technique-based manuals, we find also the various visual or
graphical aids pressed into service for narrative ends.
For example, Carolyn Wells, a successful author of detective fiction,
concluded her Technique of the Mystery Story (1913) with advice on
diagramming complex story structures, so as to better handle their
trajectories. Harry Keeler’s Web-Work, published first in the Student-Writer
in 1917, represented a more advanced system, anticipating our
contemporary developments in network theory. Instead of singular narrative
plots, Keeler emphasized a multiplicity of character arcs, which often
developed outside the visible plot structure. For example, a soldier
transformed by the war meets an old friend at dinner, where they share
elements of their backstory, which happened “offstage.” Thus, in addition to
the story told, an author was advised to track multiple backstories, so as to
more accurately represent character development. Keeler’s work urged
authors to document and visualize large fictional worlds, connected by a
complex web of social relations—where a single story plot represented one
path among many possible paths, through a multifaceted world, full of other
plots and entanglements.
Yet another type of literary device instructed authors in the systemic
gathering of realistic detail—a database—culled from newspaper clippings,
conversation snippets, and notes from real observed situations.
Mark Twain patented and sold “self-gumming” scrapbooks for authors as
early as the 1870s. Writing for the Editor in 1903, Jack London (author of
White Fang and Iron Heel) urged authors to keep a notebook at all times.
Authors in the habit of collecting such notes would quickly find themselves
overwhelmed by collected information. A single notebook wasn’t enough
for an author submitting five manuscripts a year. Notebooks grew into filing
cabinets. Devices such as the Eureka Pocket Scrap Book were routinely
advertised on the pages of the American Stationer, the Dial, the Athenaeum,
and Popular Mechanics. The Educational Specialty Company out of
Detroit, Michigan, sold the Chautauqua Literary File—a large piece of
home-office furniture involving a complex color-coding system for the
quick retrieval of reference information.
The Phillips Automatic Plot File Collector consisted of some two hundred
custom-made containers, referenced to the included plot classification
system. Advertised on the pages of magazines and in the back of published
books, the file cabinet promised to “hold many thousands of uniform items
of plot material, designed to contain Plot Material, Plot Germs and
Complete Plots—as well as statistical and reference data—in the form of
Notes, Newspaper Clippings Excerpts, Photographs, Pictures, Pages, and
Complete Articles.”
Henry Phillips explained how he used his Automatic File Collector in the
writing of a fictional character named Pod:
Where Plot Genie was simpler than Plotto, it also added an element of
probability. The book came with an elaborate insert—a wheel sandwiched
between two cardboard plates called the “Plot Robot”—picturing a
stereotyped genie in the process of inspiring a young author, crouched
beside a magic bottle. Several spins of the wheel produced a string of
numbers, which could be transferred onto the included blank form. Hill sold
multiple such forms, alongside reference books for romance, comedy, short
story, and action-adventure. The system culminated with the appropriately
titled Ten Million Photoplay Plots, containing sections on the “Twelve
Reasons for the Non-Sale of Stories,” “Censorship Regulations,” “The Cost
of Production,” and even “Plagiarism.”
SKELETON FORMS, VISUAL AIDS, filing systems, and algorithmic text generators
contributed to the increased productivity of American authors, writing for
an expanding market.
Together, these documents provide ample evidence for the widespread use
of templating tools in the production of textual goods, made in the USA.
(The Soviet Union was another hot spot for the industrialization of literary
output.) In the recollection of one contemporary author: “In those days, I
was rigidly following the rules of what I call the [name of the journal
omitted] school of the American short story . . . Stories of the school which
it dominated were all like Fords. They were of limited horsepower, neat,
trim, and shiny, taking up very little road space, structurally correct and all
following the blueprint without the slightest deviation.”
As we will see in the next chapter, the soon-to-be first wave of computer
scientists spearheading research in artificial intelligence through text
generation were initially aware of their industrial-era literary predecessors.
Earliest documented examples of AI text generators implemented
rudimentary versions of these paper-and-ink systems, using similar
techniques such as story grammars, event databases, multiple-pass
expanders, random tree walks, randomized-event engines, network
traversals, and background-world generators.
It would take some time for machines to gain the computational strength
to outperform paper systems (roughly when digital file systems began to
rival the filing cabinet in storage capacity). The future held a few surprises
yet for aspiring “author-manipulators,” in a timeline full of intrigue, rivalry,
pursuit, conflict, deadly misfortune, and deliverance. Crucially, at this point
our story also begins to splinter into many diverging paths—among
departments of linguistics, anthropology, literary study, and computer
science—each losing sight of the other’s trajectory.
CHAPTER 6
Airplane Stories
1. A = B + C [addition]
2. A = B, C, D, E . . . [randomization]
3. A = B + . . . + C [insertion]
(090) WHEN A FIRE-BOX HAS FOUR WHISTLES, HIS BELLS, SMALL AND
ENGINEER SMALL, HE HAS HEATED, LITTLE, LITTLE, OILED, OILED AND OILED
WHISTLES.
TALE-SPIN
Chomsky’s grammars were increasing in complexity, too, running into the
limits of syntax, at the boundary of meaning. Syntax, as we saw, governed
the rule of sentence composition. Sentences hold singular thoughts. To
compose more sophisticated, chained units, such as paragraphs or stories,
higher-order rulesets were needed.
And where does one find a ruleset for the composition of stories?
Folklore studies proved an obvious ally. The reappearance of Propp’s
Morphology of the Folktale, translated into English in 1968, coincided with
the expansion of the AI program, now in its search for story grammars. The
random generator of English sentences would graduate to produce random
English tales.
In 1976, James Meehan deposited his dissertation at Yale, under the
direction of Roger Schank, whose previous work on the SAM storyteller
took a tack different from that of generative grammar. Schank and Meehan
were both interested in production of “schemas,” so prominent in the work
of industrial-era authors at the turn of the twentieth century.
The computer program complying Meehan’s dissertation was called
TALE-SPIN, made to produce Aesop-like fables in English, “interacting
with the user, who specified characters, personality characteristics, and
relationships between characters.”
Apart from how-to manuals like Plotto and Plot Genie, schemas also
emerged as a concept in structuralist psychology, most prominently in the
work of Jean Piaget (who defined structuralism for us earlier in the
chapter). Similar to an algorithm, grammars represented a set of rules. But
unlike generalized language grammars, a schema contained the bare-bones
outline of a specific real-world situation—a template.
Piaget thought schemas played a key role in a child’s development.
According to his research, humans recognized complex situations based on
just a few outstanding qualities. How can you tell a birthday party apart
from a funeral? Observing just the party hats and balloons, one can be sure
to infer a birthday event. No further information is required (unless we’re at
a funeral for a clown, I guess). The mind fills the rest in.
Where grammars contained terse formulas, schemas described skeleton
“scripts” for common situations, containing typical details: lists of expected
characters, locations, goals, and activities. Scriptwriters like Polti sold
schemas for use in Hollywood scripts. Psychologists speculated that
children accumulated scripts in memory as a kind of cognitive shorthand for
later recognition. The two types of scripts, dramatic and psychological,
were intrinsically related.
Crucially, schemas described the relation of things out there in the world
—party hat, balloon, cake—where grammars described the arrangement of
words within language. Grammar-based generators produced grammatically
correct but often meaningless sentences. Schemas arranged words in a way
that also made sense in context. Taken together, generating new text ruled
by both grammar and schema promised stories both grammatically correct
and meaningful.
By the time Meehan started his dissertation, grammar-based language
generation seemed like a solved problem. Schank, his advisor, established
the efficacy of script schemas. Meehan’s TALE-SPIN combined both
approaches in one program, citing both Polti—the French playwright—and
Propp—the Soviet folklorist—as influences in addition.
TALE-SPIN wrote stories by first generating a world and then charting a
path through it, similar to Keeler’s social “web-work.” On program startup,
the reader was asked to select several options from a list of pre-defined
characters: a bear, bee, boy, girl, fox, crow, or an ant, among others. The
corresponding character schema instantiated a specimen template, using
randomly chosen type attributes: Will our bear be tall or short? Male or
female? Brown or black? A bear template specified that bears live in homes
and that homes were to be located somewhere on a map, another template.
Bears also owned things, perhaps dishware and furniture. The program
dutifully conjured randomly appointed caves, meadows, hills, and forests,
each according to its own schema. A story could now begin with a bear
lounging comfortably in its hilltop cave.
Given several characters, the program would also attempt to model a
character’s state of mind—what it knew, wanted, and believed—as well as
its goals and motivations. Our Bear may be hungry, for instance, but doesn’t
know where to find honey. His friend the Bee knows where to find honey.
Their mind-states differ, presenting diverging narrative possibilities. The
potential for narrative resolution of the plan HUNGER swings into motion. A
template indicates that HUNGER can be resolved by action EATING.
Here the stage is finally set for a heartwarming tale about Bee and Bear
sharing a meal on an alpine meadow. At least in theory.
Given the short time to dissertation completion, parts of Meehan’s
TALE-SPIN remained speculative. The included code produced output
similar to the following skeletal structure:
> JOHN BEAR WALKS FROM A CAVE ENTRANCE TO THE BUSH BY GOING
THROUGH A PASS THROUGH A VALLEY
THROUGH A MEADOW.
> JOHN BEAR TAKES THE BLUEBERRIES. JOHN BEAR EATS THE BLUEBERRIES.
THE BLUEBERRIES ARE GONE.
JOHN BEAR IS NOT VERY HUNGRY. THE END.
> -- DECIDE: DO YOU WANT ANOTHER STORY ABOUT THOSE CHARACTERS?
Some sixteen years after the birth of Yngve’s Random English Sentence
Generator, and more than forty years after Propp’s folktale Morphology,
machines thus learned to tell simple tales beyond the sentence, both
grammatically correct and mostly meaningful.
961211044319C
PASSENGER CUSSED OUT FLIGHT ATTENDANT TAXIING TO RUNWAY.
PIC RETURNED TO GATE.
PASSENGER REMOVED.
961216043479C
TURBINE RIGHT ENGINE FAILED. DIVERTED TO RIC.
OVERWEIGHT LANDING.
HAD CONTAINED TURBINE FAILURE.
960712045359C
PIC BECAME INCAPACITATED AFTER LANDING.
LANDING WAS ERRATIC AS WAS TAXIING.
FIRST OFFICER TOOK OVER.
PIC HAD STROKE.
Clark’s program was intelligent in the sense that it modeled the “rational
pursuit of goals.” It “understood” random events to the extent that it
contained a rudimentary theory of other minds, where fictional pilots and
passengers expressed certain goals, knew how to accomplish them, and
were transformed by their achievements. Once such schemas were in place,
Clark hoped that the reverse would be possible as well. A system for story
generation was also a system of story analysis: Why did the pilots divert
their course? Was it because a passenger fell ill onboard? A smart machine
could traverse the causal chain both ways, finding patterns among millions
of messages generated by airplanes daily. And perhaps one day, such a
system could fly a plane on its own.
Down but not out, the number of research papers related to schema-‐
generators continued to climb through the 1990s, in part due to their success
under controlled conditions—in fields like education, game design,
customer service, and medicine, where scripts circulated long before the
computer.
Educators were beginning to modernize their assignments, syllabi,
evaluation methods, and lesson plans for digital use. Schemas powered the
creation of immersive computer-game worlds, in cult classics like Baldur’s
Gate and Dwarf Fortress. Similarly, the interaction between patients and
physicians has been templated into digital forms and drop-down menus
found in medicine. Schemas were increasingly starting to mediate between
billing and diagnostics, to the point of becoming an essential part of every
medical institution’s information infrastructure.
Today’s patients and passengers lose themselves entering a maze of
options in conversation with automated customer service representatives,
used widely by airlines and insurance companies. Anyone observing their
physician struggle with clinical notes can appreciate the pervasive nature of
narrative schemas in large modern bureaucracies. Stories of health and
illness are exchanged within the hive of a hospital, where a chance turn
down the wrong hallway—a missed click of a doctor’s mouse, registering a
wrong diagnosis from a drop-down menu—can forever alter a patient’s
outcomes.
CHAPTER 7
Markov’s Pushkin
Imagine for a moment attempting to guess what letter may follow the
initial letter Z to continue a valid English word (similar to the game of
Wordle). The letter A would be a good guess, given its occurrence in words
like zag or zap. An O would start us on a path to zoo or zombie. However,
certain letters would never follow Z, as no words contain the combination
of ZK or ZP, for example. You can try experimenting with Markov chains
on paper by writing down a few random letters and then guessing at the
possible words that could result from their beginnings.
Here we must also embark on a little historical side quest. The elephant in
the room we’ve managed to ignore so far has been the development of the
telegraph, happening roughly from the time of Babbage—past Markov—to
MIT’s first research computer. The history of telecommunications and
computing merge definitively after that. The telegraph is important for us
because many clever folks working on language at the time came to the
question of word meaning from the perspective of machine communication,
not human texts like Eugene Onegin or even The Little Train.
One of the basic practical problems in telegraphy was simply the volume
of information that could be sent over a single wire. The volume of water
flowing through a pipe could easily be calculated according to a formula.
No such formula existed to reckon the information capacity of telegraph
cables. Noise and information loss posed additional concerns. Yet the
volume of machine communication was growing exponentially, dwarfing
human output. For over a century, everything—from love messages to
stock-purchase orders and machine instructions—was conveyed by
telegraph. Cables spanned continents, struggling to keep pace with
increasing traffic. A way to calculate their capacity, similar to water volume
in pipes, was needed to manage the flow of information.
Markov’s model of language fit this problem well because telegraphs
processed information in chains, as strings of text, entered letter by letter.
Engineers working on increasing the sheer capacity of communication in
Markov’s vein didn’t much care what kind of information was sent through
the wires, nor what it meant. A string of text moved through wires like city
water, providing utility to its users. What senders or receivers did with that
information was outside an engineer’s concern. “Frequently the messages
have meaning, that is they refer to or are correlated according to some
system with certain physical or conceptual entities,” Claude Shannon wrote
in his groundbreaking 1948 essay, titled “A Theory of Mathematical
Communication.” “These semantic aspects of communication are irrelevant
to the engineering problem,” he concluded sharply.
For Shannon and his colleagues, it sufficed to imagine human
communication as a probabilistic Markov-chain process, “generating
messages, symbol by symbol.” For instance, Shannon explained that a
message could denote a sequence of letters chosen at random, such as
“XFOML RXKHRJFFJUJ ZLPWCFWKCYJ FFJEYVKCQSGHYD
QPAAMKBZAACIBZLHJQD.”
Given the probabilities of how often each single letter occurs in the
English language independently, a string could be approximated as “OCRO
HLI RGWR NMIELWIS EU LL NBNESEBYA TH EEI ALHENHTTPA
OOBTTVA NAH BRL.”
Advancing by all probable two-letter combinations, or bigrams, produced
“ON IE ANTSOUTINYS ARE T INCTORE ST BE S DEAMY ACHIN D
ILONASIVE TUCOOWE AT TEASONARE FUSO TIZIN ANDY TOBE
SEACE CTISBE.” Here, we can see that certain correct words begin to
occur by chance, including on, are, and Andy.
Given the probability of any letter occurring after any previous two
letters, in combinations of three, or trigrams, Shannon composed “IN NO
IST LAT WHEY CRATICT FROURE BIRS GROCID PONDENOME OF
DEMONSTURES OF THE REPTAGIN IS REGOACTIONA OF CRE.”
This begins to resemble English, randomly producing correct words in, no,
whey, of, and the. Other combinations seem plausible: I had to look up birs
and pondenome in the dictionary in case they chanced on actual words (they
did not).
Things get more interesting when we compute the probabilities of word
combinations, instead of letters. A distribution of any word following, given
the previous, creates a list of all two-word units, or bigrams. To make your
own, take the first two words of any sentence and advance forward, word
by word. Given the previous sentence, this yields, from the left to right:
to make,
make your
your own
own take
take the
the first
of any first two
any sentence two words
sentence and words of
and advance
advance forward
forward word
word by
by word
Note how each of the bigrams connects to the previous, making it a linked
chain. Now imagine calculating the frequency—how often each
combination occurs—over a long text, such as Harry Potter. With modern
computers, this can be done with every word string encountered online,
producing a giant table of prior observed probabilities.
Once these priors are computed, we can use them to complete sentences.
“When I grow up, I want to be _____.” Statistically speaking, a plausible
continuation of the sentence could be “an astronaut,” or “the president of
the United States,” or “a doctor.” My grandmother tells me I wanted to be a
street sweeper because I was fascinated by autumn leaves. If we collected a
bunch of such responses from children, mine would be less probable than
others. Some word combinations, we would encounter rarely if at all.
Virtually nobody says they want to be “a tree,” or “a shoe,” or “makes a
magnificent asparagus”—that doesn’t even make grammatical sense. Priors
give us only the plausible continuations of a chain.
Given the raw frequency of words in the English language, without regard
to the previous link, Shannon approximated (by hand, remember):
“REPRESENTING AND SPEEDILY IS AN GOOD APT OR COME CAN
DIFFERENT NATURAL HERE HE THE A IN CAME THE TO OF TO
EXPERT GRAY COME TO FURNISHES THE LINE MESSAGE HAD
BE THESE.” The results based on single-word probabilities look even more
convincing than our character simulations before!
A distribution of likely two-word combinations, or bigrams, linked,
produces: “THE HEAD AND IN FRONTAL ATTACK ON AN ENGLISH
WRITER THAT THE CHARACTER OF THIS POINT IS THEREFORE
ANOTHER METHOD FOR THE LETTERS THAT THE TIME OF WHO
EVER TOLD THE PROBLEM FOR AN UNEXPECTED.” You can
imagine that three, four, and longer word linkages would work even better,
though, too long of a link becomes unique and not useful for further
extrapolation.
While still mostly meaningless, Shannon’s output rivaled that of Yngve’s
grammar-based English Sentence Generator. The surprising effectiveness of
the probabilistic approach lay in the absence of any grammatical or
semantic assumptions. The method worked purely by statistics, powered by
the diligence of the original tabulation. The greater the number of statistical
combinations calculated in advance, the better the expected results.
Markov chains were hungry for observed probabilities. To be effective,
they needed to read, process, and retain a large volume of texts. The same
hunger also limited their reach. Recall that in the 1960s, Yngve’s random
generator required the application of some 108 multiword rules alongside
several dozen vocabulary words from The Little Train in order to operate.
IBM’s 709 machine used in that experiment barely had the capacity to hold
the program in its memory, limited as it was to 30,000 words in total,
including those reserved for the operating system. Markov calculated
20,000 of his “letter chains” by hand, half a century prior. In other words,
machines did not yet perform better than humans. The derivation of every
bigram (two-word combination) in the English language, as suggested by
Shannon, would be impossible by hand or by computer at the time,
requiring data points in the billions. The computation of trigrams (every
three-word combination) would need more storage and more processing
power, by orders of magnitude. It would take almost a century after Markov
for computers to get powerful enough for that task.
Brute reading force scaled with processing power. The problem of
digitizing printed text, instead of doing the tabulations by hand, also had to
be solved. The technology for scanning and digitizing documents was still
in its infancy. For these reasons, actual (instead of hypothetical) working
Markov-chain language generators did not appear in earnest until later.
Their use in the 1960s to 1980s was limited to more contained systems,
whether musical composition, table games, calculating insurance
probabilities, where the number of potential combinations was smaller.
Andrey Andreyevich Markov continued to publish research in mathematics,
actuary science, and chess—never again returning to literature.
Though Shannon never implemented Markov’s chains digitally, the idea
made a lasting impact. In theory, the algorithm worked. Markov’s and
Shannon’s mathematical models of communication made no assumptions
about the world or the brain. Given the number of times I wrote “given,”
they followed only the precepts of observed probability. Everything hinged
on how much the computer could tabulate—or “learn,” again
metaphorically speaking—before the task of generating sentences, by
ingesting as much text as possible. The brute force required to process most
everything published in the English language would not be available until
the twenty-first century.
Physical limitations of mid-twentieth-century computing did not
completely preclude the evolution of probability-based text machines,
though these appeared in somewhat exotic new fields, like optical character
recognition (extracting text from images), spell-checking, and document
retrieval.
In 1959, Bell Telephone Labs patented a device for the Automated
Reading of Cursive Script, US3127588A. The patent proposed using shape
frequencies to detect possible errors in the recognition of cursive script. The
work by W. W. Bledsoe and I. Browning from the same year pioneered the
use of “word context” for the purpose of “pattern recognition and reading
by machine.” Thus in the scanning of a text, machines could try to guess an
illegible word based on the probability of its individual characters within
the sentence.
A team from Cornell’s Aeronautical Laboratory—airplanes again!—‐
proposed a novel spell-checking algorithm in 1964. Lockstep with our
statisticians, it relied on dictionaries and probable “word contexts” for the
task of “correcting garbled English text.” Highlighting the absence of
assumptions about meaning, the authors wrote, “We have been concerned
instead with the recognition of the identity of a word or of letter sequences
by making use of the fact that only certain letter sequences occur in English
with any appreciable probability.” The program was capable of correcting
sentences like “TAAT SOUIDS LIKE A AIVTDE KITTEN,” SAUD
TOMYN—“ ‘That sounds like a little kitten,’ said Tommy.”
With the passing of the telegraph and linear ticker tape, the metaphor of a
chain gave way to the related notion of “word vector space,” more
appropriate for the representation of printed documents, where words
expanded along two dimensions, instead of a straight line.
I love the idea of a vector space because it evokes the night sky,
preferably somewhere up in the mountains, away from light and
civilization.
Imagine the universe of language as a constellation of concepts scattered
across the starry skies: each concept in its place and each following a
different trajectory in relation to others. This picture changes the way we
think about word meaning in general. Rather than looking for meaning in
some fixed dictionary, we can agree that words “mean” by their common
locations or contexts. The word sunny, for example, often occurs in the
context of weather; the words speed record next to race or the Olympics. In
this way, the words fast and speedy are similar because they occur in similar
contexts—they have similar locations and trajectories, or vectors.
Of course, the state of the world itself isn’t always so sunny. For instance,
large, complicated social problems such as abusive language or racism can
manifest in the constellation of certain often-repeated expletives or negative
stereotypes. Cleverly, the computer can distinguish the two meanings of
race based on its contextual location. Race words found near speed records
and distances mean something different from those found near disparity or
mass incarceration. Not so cleverly, a computer that “learns” from biased
language—everything ever written—may continue to spread abuse or
racism it has learned from humans.
Despite its shortcomings, a vectorized model of language has allowed
machines to begin making complex inferences, related to the meaning of
the words they don’t really “know.” Machine “understanding” is instead
woven out of the statistical context. The “correct” answer to any given
question represents also the “likely” continuation of a string.
Words in the shape of mathematical context-vectors further enable the
computation of semantic distance, capturing shades of difference between
fuzzy negative attributes, such as bad and its even worse cousin terrible.
With word embeddings we can also attempt to calculate those words
furthest from any point in our starry skies, finding, automatically and
without teaching the computer anything about the world, that awesome
occupies reaches of space far away from awful.
Vectors can also be used to find related clusters or constellations of
words, which signify topics, literary genres, or themes in a conversation.
Just as contexts define words, a constellation of contexts can define whole
systems of thinking. A certain conceptual word cloud can be highly
correlated with the domain of physics, for instance, where another cluster of
words corresponds to politics. Similar words, occurring together frequently,
produce patterns in three dimensions, not just on a page but in whole books
and libraries, within a vector space.
The idea of word embeddings and vector spaces found another harbor in
the literature on document retrieval, written by scholars working at the
intersection of computer and library science. Modern online search giants
were born out of the necessity to organize and retrieve the world’s
information, long the purview of truly arcane arts like filing, cataloging,
and indexing. Before computers, librarians filed things according to hand-‐
labeled classification schemas, in which a book by Markov could be
classified under subjects such as Science, Mathematics, Probabilities
(QA273–280, in the US Library of Congress classification schema). A
book’s indexer reached into the content of the work to pull out notable
themes, names, places, or references for reader convenience.
By the 1960s, researchers proposed ways to extract keywords
automatically.
Among them, Peter Luhn extended the notion of “words in context,” to
“keywords in context” at IBM. Gerald Salton at Harvard used word
frequencies to derive a metric of document similarity: You could now
request a “similar” book from your university library—the similarity
computed automatically using word distances.
Karen Spärck Jones built on her work with Margaret Masterman at
Cambridge to propose “weighting” keywords statistically, as a function of
term frequency rather than meaning. Matches on rare word combinations
would therefore count for more than common words. Querying a search
engine for “recipes for Beef Stroganoff” would amplify the importance of
the rare term Stroganoff while discounting the common for, recipe, and
beef.
Sergey Brin, Lawrence Page, Hector Garcia-Molina, and Robin Li (the
founder of Baidu) further built systems that weighted mutual citation.
Documents that referenced other documents counted for more than those
that were filed in isolation. Search engine technology was thus built out of
parts found in library science, philology, and linguistics, using techniques
like text alignment, indexing, cataloging, term frequency comparison, and
citation analysis.
The rest is history, well documented—you can google it or ask your
friendly neighborhood chatbot for assistance. Words in context brought us
into the modern age of literary intelligence.
S O FAR, WE’VE BEEN PATIENTLY assembling the modern chatbot from parts
found on the workbench of history. I have tried to animate that device
for you, by strings of anecdote and light philosophical commentary.
We didn’t neglect to feed this Pinocchio in addition. Even before
publication, my text will enter the collective machine unconscious—in the
way all digital texts today are vacuumed into datasets—to be processed for
the use of training of future text generators. What will it think of its
parents? Will there be a sense of pride or shame in family history? Or
perhaps nothing, because all outputs were already anticipated? The
connection between Ibn Khaldun and Llull, for instance, though obscure, is
well documented. The same can be said of Leibniz, Babbage, and Lovelace.
A few entirely novel synapses were sparked in the linkages between
industrial manufacture and the rise of mass literary markets. The work on
pulp fiction, structuralism, and early computer science is also entirely my
own novel contribution, based on original research.
A few loose leads remain, poking out of the seams between chapters.
Let’s trim them and draw toward a conclusion.
Several important sources lay too far afield to cover adequately. For
instance: Histories of the Turing machine and the Turing test often neglect
the direct influence owed to Ludwig Wittgenstein’s lectures, along with the
presence of Margaret Masterman, the pioneer of machine translation, in the
same classroom. Masterman’s universal thesaurus harkens back to Wilkins
and other universal-language makers, central to a whole separate and
important branch of AI—machine translation. It would require more than a
chapter of its own, orthogonal to the direction of our travel. Aspects of
encryption used for diplomacy or military communications would also lead
to entirely different exit points. Regionally, an emphasis on German or
Soviet computing would make mine a far less Anglocentric endeavor. The
rich traditions of Arabic, Indian, and East Asian philology are mostly out of
my range of expertise. There’s ample room in these gaps for machine or
human readers to amble on their own.
Another alley not fully explored would lead us to I. A. Richards, a part of
the linguistic oatmeal duo figuring briefly in the last chapter. Aside from his
work on linguistics, Richards was one of the major thinkers at the
foundation of the New Criticism movement, which emphasized close
reading and formal analysis of text.
This book was conceived as both a tribute and a rejoinder to Richards’s
Science and Poetry. The original volume helped launch W. W. Norton as a
publishing house in the 1920s. Reading it fondly a century later, I was
surprised to find a pervasive nostalgia for poetry, besieged by the
“onslaught” of science. Elsewhere, I knew Richards to be an unsentimental
scholar, writing with clarity and sharp insight. Yet in Science and Poetry,
the thought wanders. In the conclusion to his book, he writes of “getting the
guns into position,” “sciences that progressively invade every province,”
science that will “force [other myths] to surrender,” and “the Hindenburg
Line [. . .] held” but also, in the same sentence, “abandoned as worth neither
defence nor attack.” Who these enemies are exactly, we don’t know. One
just gets the feeling of general embattlement.
Long a student of Richards, I can’t help but read war between the lines.
Written not long after the end of World War I, in 1925, and expanded and
reissued on the cusp of World War II, in 1935, Science and Poetry is
haunted by conflict.
Almost a century later, the possibility of another world war again hangs in
the air. As I write these lines, more than 200,000 civilians and soldiers have
perished as a result of the Russian aggression in Ukraine. More deaths will
surely follow. An aging dictator, out of touch with reality, daily threatens
total nuclear annihilation.
Armies of chatbots have been conscripted on both sides, but especially on
the side of Russia. My days are occupied in real-time observation of troop
movements, on the ground and online. The swings of public opinion and
soldier morale have become matters of life and death. I am contributing in
my own small ways, by translating war communiques, filing take-down
notices, and writing code to automate certain language tasks useful for the
war effort. The battle between Western support and the unwieldy machine
of Russian state propaganda rages in countless online forums, where
information and disinformation merge into a rush of war commentary,
fueled by AI authors.
Poetry did not, however, protect Russia from the myths of imperialism or
fascism, in the way that Richards had hoped for humanity in his book.
Literature was instead mobilized as one of its last remaining weapons. The
Russian invasion of Ukraine in 2022 happened under the pretext of
protecting a Russian linguistic minority in the Ukrainian Donbas region, “in
danger of cultural erasure,” alongside the alleged removal of Pushkin,
Tolstoy, Lermontov, and Dostoevsky from school curriculums. So dear
leader appealed to great literature, doting on his own nostalgia for an
empire whose time has passed. Left out of his grandiose nationalism were
Pushkin’s Cameroonian ancestry and Lermontov’s glorification of Russian
imperial war crimes in the Caucasus.
I cannot therefore share in Richards’s enthusiasm for the protective
bulwarks of literary imagination. Nor in his stark separation between
science and poetry. My thesis runs contrary to his. Many languages other
than English, like German or Russian, do not honor the distinction between
the sciences and the humanities. There, science just means “the organized
study of,” a synonym to the English scholarship. The science of poetry
holds no contradictions in that sense. It indicates simply the systematic
study of the thing, beyond casual reading.
While romping through the centuries, we found the roots of computer
science to be inextricably entwined with literary and linguistic concerns.
Viewing them apart has impoverished both communities: poets, in terms of
financial and cultural capital, and programmers, in terms of belonging to a
deep intellectual tradition. Worse yet, the division gave both communities a
kind of a shallow myopia, where AI seems to entail either the death of us all
or a cure for all ills.
Which brings us to the evergreen question: What is to be done? How can
we think of robots and literature in a more deeply holistic way?
In response, I leave you with the following nine important ideas that will
totally change your life in addition to explaining why AI will neither
destroy humanity nor solve all its problems.
1. ARTIFICIAL INTELLIGENCE IS
COLLECTIVE LABOR.
2. INTELLIGENCE IS DISTRIBUTED.
Collective labor involves work distributed across time and space. For
instance: To write for me means not just to think but also to fidget, to
consult dictionaries, to take walks, to draw diagrams, to consult with
colleagues, and to take editorial input. The hypothesis of distributed
cognition holds that cognitive tasks don’t just happen in the mind. We think
with our bodies, with tools, with texts, within environments, and with other
people.
But if human intelligence is distributed, more so the machine.
The disassembly of a chatbot, a smart mobile phone assistant for instance,
reveals multiple, diffuse locations where thought happens. Some “smarts”
are baked into the circuit. Some happen in the cloud. Some lead to large-‐
scale feats of electrical engineering, software development, and project
management. More threads connect to institutional decision-making,
marketing, corporate governance. Together, these linkages constitute the
“smarts” of a mobile phone assistant, within a single massively
collaborative act of intelligence.
As before, multiplicity confounds. We are at pains to acknowledge a long
list of collaborators baked into the technology. A crowd of ghostly
assistants surrounds the act of writing. A view of intelligence from the
perspective of labor admits a long list of contributors. Please take this point
along with my gratitude to numerous collaborators participating in the
writing of this book.
Despite representing diffuse, distributed pluralities, AI often assumes the
singular. Pay attention to the headlines: “AI Is About to Take Your Job.”
What does that mean literally? Who exactly is about to take your job?
Reason it out. Chances are, it will become plural under scrutiny. You will
find multiple specific entities, agents, persons, and organizations
responsible. Like the passive voice, AI can be demystified in more careful
phrasing.
Chatbots and self-driving cars don’t originate from a single source. You
will often hear something like “AI is bad at recognizing giraffes.” Ask:
Which algorithm exactly? Does the “sort by date” function in my
spreadsheet editor know or care about giraffes? Does the smart vacuum?
Why are we considering all of them together, in the singular, as one
intelligence?
Isn’t that also a strange way of putting things? You don’t expect humans
to work like that. I happen to be bad at playing chess and at predicting
weather. My own localized abilities are on the spectrum of what is generally
possible for an average person. AI does a bad job recognizing giraffes in the
way “humanity is bad at dealing with global warming.” It’s true in a general
sense, ignoring local differences between specific polluters. “Humanity”
offers a broad generalization about distinct humans. “AI” similarly offers a
broad generalization about distinct technologies.
3. AI HOLDS A METAPHOR.
The value of labor diminishes with group size. And so brave heroes climb
the Mount Everest of intelligence, while the unnamed Sherpas carry their
luggage. But how is that fair? The luggage grows heavier near the summit.
And though I might thank my editors, it’s just not customary to
acknowledge all the folks actually involved in the production of intellectual
artifacts, like books. The sources alone probably took hundreds of people to
preserve, digitize, and make available for instant reference. Then there’s the
work of word-processing, spell-checking, copyediting, marketing, printing,
and distribution.
How would I go about acknowledging the aid of thousands? Does my
effort in the composing of individual sentences equal or exceed their
combined contribution? Individually, in terms of hours spent—maybe. In
aggregate, probably not. The ego insists on its own exceptionalism,
however. A less egocentric understanding of collective intelligence would
require the careful attribution of labor involved. In another, more radical
version of this book, our analysis of any AI would consist of forensic labor
attribution only—roll the credits—the way air accident investigations use
forensic techniques to assign causes proportionally. In a better world, we
might also reconsider our inflated sense of individual capacity for
independent intellectual achievement. Under scrutiny, intelligence almost
always unpacks into a collective endeavor.
Work that can be automated loses its economic value. The diminishment
of previously creative labor that comes with AI entails enormous economic
consequences. Consider the impact of automation in the nineteenth century,
at the time when the mass manufacture of goods began to supersede
artisanal production. Factories made trades like blacksmithing or
shoemaking obsolete. In the twentieth century, the automation of the supply
and distribution chains killed the small-town mom-and-pop shop. Robots
supplanted factory workers, leaving whole regions to rust in the wake of a
changing labor market.
A similar fate awaits intellectual laborers—legal professionals, writers,
physicians, and software engineers—to the extent that their labor can be
automated. At the time of writing this book, AIs pass bar exams in the 90th
percentile of performance. Writers are striking in Hollywood, in part
because studios might replace them with machines. Modern AIs are capable
of passing the most difficult of technical job interviews with excellent
marks. And they outperform physicians in most verbal and visual diagnostic
tasks not involving surgical intervention. The ranks of American ghost
towns ravaged by industrialization may soon be joined by former tech
centers.
Not all our prospects are so glum.
Human intelligence continually expands to cover new publics. Another
way of saying “automation cheapens labor,” would be to say “automation
reduces barriers to entry, increasing the supply of goods for all. Learning,
for example—something that required special training and immense effort
in times past—has entered the purview of the public at large, and at low
cost. The tide of encyclopedias and search engines has already lifted all
smart boats.
Consequently, the mere fact of having a large vocabulary or memorizing
many facts no longer suffices to maintain a profession. We have no more
need for scribes or scholars who merely regurgitate facts. Instead, freed
from the bondage of erudition, today’s scribes and scholars can challenge
themselves with more creative tasks.
Recent advances in AI promise a similar transformation for many
professions. It is true that in the future markets may require far fewer
doctors or software engineers. But those that remain will also find their
work enriched. Tasks that are tedious have been outsourced to the
machines.
Paradoxically, the computerization of human intellect also implies the
humanization of computer sciences.
Conversational AI was built on the legacy of the language arts. For over a
century, computer science has diverged from that lineage to finish building
something incredible: a machine that writes as well as most humans. It also
codes. Which means that the technical component now becomes the lesser
obstacle to the inception of any project.
For instance, twenty years ago, the making of a social network involved
first a significant engineering component, and social analysis only second.
The same can be said about word-processing or electronic health-‐
management systems. In each case, engineering difficulty necessitated a
disproportionate investment in technology. Judging by the results, this
sometimes left little time for research. But just communicating with our
loved ones or patients is not enough. We want to do it in a way that makes
our relationships better. Physicians don’t want to just manage their patients.
They want to do it in a way that promotes health. Yet, until now, because of
the technical complexity of the task, technology companies invested heavily
in the instrumental, engineering component: how something can be done.
They can now better concentrate on the why and to what ends.
An anthropologist or a historian that studies the formation of modern
families has insight into the values motivating people to create social bonds.
Similarly, a linguist or a textual scholar entering the hospital can study how
human stories about health are told, so that they can be better translated into
medical diagnostics. The work of encoding values into software systems
can now be more trivially automated. The lowering of barriers to technical
expertise allows the humanities to fully integrate into the practice of
engineering. The two fields, as we have seen, were never that far apart to
begin with.
8. TECHNOLOGY ENCODES POLITICS.
At each stage of its development, from Ibn Khaldun to the early versions of
the generative pre-trained transformers (GPT), we have often found our
rather narrow set of technical concerns to veer toward the spiritual. As the
gold rush around this latest wave of AI research subsides, I cannot help but
wonder about this noninstrumental excess.
Even the most rudimentary of word mechanisms has the immediate
capacity to delight, in the way children still play with simple origami
fortune tellers. Contemporary machines do much more than delight; they
portend—what some see as the coming of a new era. “Are you a modern or
an ancient science?” I ask the machine after Ibn Khaldun, hoping to glimpse
the divine. But no matter how the prompt is engineered or how often, the
child of eternity repeats at best a corporate answer:
Abelson, Robert, 92
Aeronautical Laboratory (Cornell University), 110
Aesop, 94
agency, 127, 131–32, 141
Agnesi, Maria, 44
AI, See artificial intelligence
AI and Data Ethics Institute, 131
Air Research and Development Command (US Air Force), 87
algebraic patterns, 55
algorithms, 9, 131
alignment problem, 38
alphabets, 43
Alphaville (film), 93
American Catalogue of Books in Print, 66
American Psychiatric Association, 23
American Stationer, 74
Analytical Engine (analytical engines), 48–52, 54–56, 60–61, 64
Andrews, Charlton
The Technique of the Play Writing, 71
Appelbaum, Matthew, 92
applications, 32, 33
application tables, 48
Arabic language, 43
Arbogast, Louis, 44
Aristotelianism, 34, 36–38, 72
Aristotle, 36, 44
Poetics, 50–51, 67
Ars Brevis (Llull), 24, 31
artifice, 4, 61, 123
artificial intelligence (AI)
in academia, 137–38
and agency, 141
as collective labor, 122–23
conversational AI, 135
and creative process, 133–34
dangers of, 127, 129, 137
definitions of, 4, 11
demystification of, 124
economic consequences of, 133–35
ethical, 22
gaps in thinking about, 5–7
history of, 12
“intelligence” aspect of, 14–16, 21, 125
language-based, 21, 46
and machine translation, 119
participants in, 132
personification of, 127, 130
purpose of, 59
and responsibility, 132
artificial intelligence (AI) (continued)
scope of, 5, 16, 93, 128, 129
in social sphere, 127, 136, 139
and template culture, 83
assistive technology, 15, 28, 38–39, 123–24, 138
Athenaeum, 74
Austen, Jane, 65, 67
Author, 70
Author’s Digest, 71
Author’s Journal, 70
author wages, 67
automated assistants, 28, 138
Automated Reading of Cursive Script, 110
automated tasks, devaluation of, 38
automatic transmissions, 14–16
automation
in industrial age, 2
of reason, 40
of work, 133–34
Babbage, Charles, 43, 48–54, 56, 59–60, 62–64, 71, 105, 118
On the Economy of Machinery and Manufactures, 60, 63–64, 71
Passages from a Life of a Philosopher, 49–50
backstories, 73
Bacon, Francis, 7, 10
Baidu, 113
Baker, George Pierce, 72–73
Baldur’s Gate, 100
Barrymore, John, 73
BASEBALL (story generator), 92
basic drives, 128
Baudot code, 7
Believe Me, Xantippe (film), 73
Bell Telephone Labs, 110
Benjamin, Walter, 61
Bibles, 39
Bibliothèque universelle de Genève, 54
bigrams, 106–7, 109
bits, 6–9
Blackburn, Simon, 84
Bledsoe, W. W., 110
Bloomfield, Leonard, 83
Bobrow, Daniel, 92
body posture, 4
Boeing, 11, 97
Boston, Mass., 70
bots, 136–37
Brecht, Bertolt, 61
Brin, Sergey, 113
Browning, I., 110
brute force computing, 110
Byron, Lady Anne Isabella, 51, 52
Byron, Lord George, 12, 51, 61
bytes, 6–9, 88
Darwin, Charles, 51
databases, 3, 24
Defoe, Daniel
Robinson Crusoe, 36, 65
De Morgan, Sophia Elizabeth, 51
Descartes, Réné, 44
Meditations, 35
de Stael, Germaine, 80
Detroit, Mich., 74
Diagnostic and Statistical Manual of Mental Disorders (DSM-5), 23–24
Dial, 74
Dickens, Charles, 51, 65, 67
dictionaries, 29
Diderot, Denis
Encyclopédie, 122
Difference Engine, 48–49, 52, 60
disinformation, 132
distributed intelligence, 123–24
distributed thought, 123–25
divination circles, 24
document retrieval, 110
Donbas region, 121
Dostoevsky, Fyodor, 65, 121
Douglas, Mary
How Institutions Think, 127
Downey, June
Plots and Personalities, 71
Dowst, Robert Saunders
The Technique of the Fiction Writing, 71
Dumas, Alexandre, 65
Dwarf Fortress, 100
eclipses, 52
Editor: Journal of Information for Literary Workers, 70, 72, 74
educational reform, 67
Educational Specialty Company, 74
ELIZA (chatbox therapist), 20, 92, 93
emergence, 16
encryption, 119
Encyclopédie (Diderot), 122
English, Thomas, 71
Skeleton Essays, or Authorship in Outline, 71, 72
Enlightenment, 46, 67
epistemology, 84
equality, concept of, 116
Esenwein, Joseph
Writing the Short Story, 71
Esperanto, 45
Essay Towards a Real Character, and a Philosophical Language, An (Wilkins), 10, 41–45
Eugene Onegin (Pushkin), 104, 117
Eureka Pocket Scrap Book, 74
event horizons, 7
exceptions and exceptionalism, 38, 39, 52, 59, 61, 62, 79–80, 133
Habsburgs, 30
Hanks, Tom, 36
Harris, Zellig, 83, 86, 114
Harry Potter, 107
Harvard University, 72–73, 113
Hayy ibn Yaqzan (Ibn Tufail), 36
“Heavenly Love-Kiss XLI” (Kuhlmann), 31, 40
Hegelianism, 6–7
Herder, Johann Gottfried von, 80
hermeneutics, 2
Hill, Wycliffe, 75, 77
Plot Genie, 75–77, 81, 94
history, 5–9, 12, 91
Hobbes, Thomas
Leviathan, 127
Hollywood, 71, 94–95, 134
Home Correspondence School, 70–71
Hopi, 80
Horne, Charles
The Technique of the Novel, 71
How Institutions Think (Douglas), 127
human intelligence, 116, 122, 123, 134
humanities, 121
Humboldt, Wilhelm von, 50
Kafka, Franz, 61
Kane, William R., 70
Kaplan, Ronald, 92
Kay, Martin, 92
Kazakhstan, 139
Keeler, Harry
Web-Work, 73, 95
Kepler, Johannes, 30
keywords, weighted, 113
keywords in context, 113
Khwarizmi, Muhammad ibn Musa al-, 9
Kircher, Athanasius, 30–34, 36, 39–40, 44, 48, 49, 56, 60, 75
Kissinger, Henry, 16
Klein, Sheldon, 92
knowledge, 11, 21, 31, 32, 35–36, 46, 132
knowledge work, 5, 62, 133–36
KPMG, 131
Kubrick, Stanley, 93
Kuhlmann, Quirinus, 30–32, 34, 40
“Heavenly Love-Kiss XLI,” 31, 40
Kyrgyzstan, 8
labor, 123
collective, 123
intellectual, 61–62
and knowledge work, 133–36
Latin, 39
Latin Word Study Tool, 28
Laughery, Kenneth, 92
law, 127
Law & Order (television series), 79
“Laws of Literary Invention,” 81
learning, 115–16, 125–26
Lehert, Wendy, 93
Leibniz, Gottfried, 10, 43–46, 48, 55, 60, 118
Characteristica universalis, 44
Plus Ultra, 44–45
Leningrad State University, 82
Lenski, Lois
The Little Train, 87, 88, 90, 109
Lermontov, Mikhail, 121
letter magic, 20–22
Leviathan (Hobbes), 127
Li, Robin, 113
Lindsay, Kathleen, 66
linguistic intelligence, 101
linguistic proficiency, 115–16
linguistics, 2, 83–87, 92–93, 101, 103, 113, 114, 119
literacy, 38, 64, 67
literary markets, 65, 70
literary production, 64–67, 69, 71–72, 79
literature, 1, 9–11, 59, 61, 65, 80–82, 120–21
Little, Paul, 66
Little Train, The (Lenski), 87, 88, 90, 109
Llull, Ramon, 9, 10, 24–27, 31, 46, 48, 91, 118
Ars Brevis, 24, 31
local values, 38
London, Jack, 74
Longfellow, Henry, 80
Lost (television series), 36
Lovelace, Ada, 12, 43, 48, 51–52, 54–60, 64, 91, 118
Lucas, George, 93
Luhn, Peter, 113
Nabokov, Vladimir, 72
Nancy Drew, 66
national personhood, 125, 127
neural networks, 115, 125–26
New Criticism, 119
Newton, Isaac, 44
New York City, 70, 71
New York Public Library, 70
Nietzsche, Friedrich, 21
Norman, Donald, 92
notation, mechanical, 53–54
notebooks, 74
tables
in Analytical Engine, 52, 64
application, 33, 48
in Mathematical Organ, 75
reference, 40
smart, 29, 36, 41
table of, 25
in zairajah, 19, 22–27
TALE-SPIN (story generator), 92, 95–98
tasks, devaluation of automated, 38
taxonomies, 12, 23–24
Technique of Drama, The (Price), 72–73
Technique of Fiction Writing, The (Dowst), 71
Technique of the Mystery Story, The (Wells), 71, 73
Technique of the Novel, The (Horne), 71
Technique of Play Writing, The (Andrews), 71
technological change, 131
telecommunications, 105
telegraphs, 7–8, 105, 111
Tellado, Corin, 66
template-based manufacturing, 60–61, 64
templates, 62, 64, 70, 71, 79–80, 83–84, 94–97, 100
Ten Million Photoplay Plots, 77
text generators, 20–21, 65, 77, 78
“Theory of Mathematical Communication, A” (Shannon), 105–6
thinking and thought, 3, 4, 14, 36, 45, 59–61, 93, 123–25, 132, 138
Thirty-Six Dramatic Situations, The (Polti), 68–71
Thompson, Henry, 93
Tolstoy, Leo, 65, 121
TRAC (story generator), 92
translation, machine, 119
transmissions, 14–16
trees, 55
trickster, figure of the, 80
Tristram Shandy (Sterne), 65
true crime, 66
truth, ground, 40
Tufts University, 28
Turin, Italy, 54
Turing, Alan, 12, 37, 43, 93
Turing machine, 119
Turing test, 119
Twain, Mark, 73–74
2001: A Space Odyssey (film), 93
Types of Prose Narratives (Fansler), 71
values
encoding of, 135–36
local, 38
vector spaces, 11, 111–12
von Neumann, John, 44
zairajah, 18–22, 26
Norton Shorts
B R IL L IA N C E WIT H B R E V IT Y
Forthcoming
Mehrsa Baradaran on the racial wealth gap
Rina Bliss on the “reality” of race
Merlin Chowkwanyum on the social determinants of health
Daniel Aldana Cohen on eco-apartheid
Jim Downs on cultural healing
Reginald K. Ellis on Black education versus Black freedom
Brooke Harrington on offshore finance
Justene Hill Edwards on the history of inequality in America
Destin Jenkins on a short history of debt
Quill Kukla on a new vision of consent
Barry Lam on discretion
Matthew Lockwood on a new history of exploration
Natalia Molina on the myth of assimilation
Rhacel Salazar Parreñas on human trafficking
Tony Perry on water in African American culture and history
Beth Piatote on living with history
Ashanté Reese on the transformative possibilities of food
Jeff Sebo on the moral circle
Tracy K. Smith on poetry in an age of technology
Onaje X. O. Woodbine on transcendence in sports
ALSO BY DENNIS YI TENEN