Ai: The Tumultuous History Of The Search For Artificial Intelligence [1 ed.] 0465029973, 9780465029976

In the summer of 1956, ten young scientists, some barely out of their doctoral studies, sat down to consider the astound

1,366 168 49MB

English Pages 400 [404] Year 1993

Report DMCA / Copyright

DOWNLOAD FILE

Polecaj historie

Ai: The Tumultuous History Of The Search For Artificial Intelligence [1 ed.]
 0465029973, 9780465029976

Citation preview

"Superbly written and utterly engrossing."

— Wall

Street Journal

TUMULTUOUS HISTORV OF THE

2

ARTIFICIAL INTF

DANIEL "A

superb history ...

it

was

like

C

R

going back to see

it

2

E all

V

FOR

H E

NCE

I

with other eyes."

— Marvin Minsky

I

/ft

i

7- u~

-f

\y*> "?

•»

THE

TUMULTUOUS HISTORY

OF THE SEARCH FOR

ARTIFICIAL INTELLIGENCE

DANIEL CREVIER

BasicBooks A Division of WzvperGoWmsPublishers

Copyright

©

1993 by Daniel Crevier.

Published by BasicBooks,

A

Division of HarperCollins Publishers, Inc.

All rights reserved. Printed in the United States of America.

of

this

book may be reproduced

in

written permission except in the case of brief quotations in critical articles

No

part

any manner whatsoever without

embodied

and reviews. For information, address BasicBooks,

10 East 53rd Street,

New

York,

NY

10022-5299.

Designed by Ellen Levine

Library of Congress Cataioging-in-Publication Data Crevier, Daniel,

AI

:

1947-

the tumultuous history of the search for

artificial

intelligence / Daniel Crevier. p.

cm.

Includes bibliographical references and index.

ISBN 0-465-02997-3

ISBN 0-465-00104-1 1.

Artificial intelligence

(cloth)

(paper)



History.

I.

Title.

Q335.C66 1993 006.3'09— dc20

91-55461 CIP

94 95 96 97

CC/HC

987654321

To

Celine

Contents

ix

Preface

Acknowledgments Introduction:

Human 1

xiii

Probing the Mystery of

Intelligence

1

Engineering Intelligence: Computers and

Programming

9

2

The

3

The Dawn of

Golden Years: 1956-63

51

4

The Conquest of Micro Worlds: 1963-70

73

5

Clouds on the AI Horizon

108

6

The Tree of Knowledge

145

7

Coming of Age

163

8

The

197

9

Game

First

AI Program: Defining the

the Field

Rollercoaster of the 1980s Playing:

Checkmate

for

Machines?

26

217

Ill

CONTENTS

1

Souls of Silicon

237

1 1

How Many

281

12

The

Bulldozers for an Ant Colony?

Silicon Challengers in

Our

Future

312

Notes

342

Index

369

Preface

\\ ot too long ago,

ll

if

you walked into the computer room of MIT's

Artificial Intelligence

Laboratory, the

winding yellow brick path painted on the

first

thing you noticed was a

floor.

At the end of this

path,

dangling from the ceiling above a large computer, was a resplendent

rainbow. In case you missed the

first

two

references, a poster of Judy

Garland as Dorothy was taped to the computer's side and the computer

had been given the nickname Oz.

itself

In

reality,

Oz was

nothing more sophisticated than a mainframe

computer controlling earlier traveler

a

MIT's

network of smaller computers. But

just as

over the rainbow had hoped that a wizard named

might be able to fashion ers at

1

an

Oz

a brain for a straw-stuffed friend, the research-

Artificial Intelligence

Lab hoped

that their

Oz

might be

used to create computer-generated intelligence.

From

its

inception the Artificial Intelligence Laboratory at

occupied most of a a section

I

building in

what is known

as

MIT

has

Technology Square,

of campus a short jog away from the heart of the Massachu-

setts Institute

that

tall

of Technology on Mass Avenue.

spent far too

many

It

was on

this

campus

days and nights in the early 1970s busily

assembling a Ph.D. thesis on a topic related to AI, but different enough to keep

me away from

the lab

observed the goings on over

at

itself.

While doing

Tech Square with

my own

work,

a perplexed

I

and

— x

PIEFACE

somewhat envious out about some

For every so often information would leak

interest.

new and

exotic breakthrough that suggested that

where the computer world's future was being created and being loosed on the

demonstration

vividly a

TV camera,

us.

so that those of us interested in a

a

demonstration of the piling

image of a model block structure

to the camera. In another demonstration, an operator conversed

with a computer in English, ordering tion displayed

about

first

computer manipulate the arm of a robot

children's blocks, building a mirror

shown

remember

I

which the AI computer was hooked up to

in

future could observe the

up

of

rest

followed these developments with intense interest.

I

AI was

tested before

its

on

it

to manipulate a block construc-

and even had the machine answer questions

a screen,

Not knowing how to evaluate all this, many of us whether we had just witnessed the dawning coming age of AI or a bit of theater.

motivations.

in attendance weren't sure

experiment of the

The answer was not long the

1

960s and early

able to create,

would not be

it

1

coming. While these AI experiments of

in

970s were fun to watch and probably even enjoy-

soon became

lems in restricted

areas.

main sponsors of

early

clear that the techniques they

from dealing with

useful apart

Not AI

surprisingly, the U.S. military,

was

research,

employed

carefully simplified prob-

also

one of the

one of the

first

to have

second thoughts. Tech Square, and other centers that had sprung up

around the world with high hopes of

early success,

selves fighting for their very survival.

Human

was, was not about to yield

In the early 1980s, after

work

I

its

had

in Al-related research,

left

it

MIT

to return

home

to

Canada

to

read about the next breakthrough

I

years away, the idea that

whatever

secrets quickly.

expert systems. While broad-based

many

soon found them-

intelligence,

artificial intelligence

one could create systems

the decision-making processes of

human

might

still

be

that captured

experts, even if those systems

operated only on narrowly focused tasks, suddenly seemed on the

horizon sort

or, at

most,

of nonhuman

without

guilt

just

over

idiot savant,

or shame.

It

humans and machines, or

it.

The

expert system was promoted as a

but one that could be exploited for profit

could be tailored to diagnose the ailments of to select the optimal site for oil exploration

or the best wine for dinner.

The prospects were

exciting

and appeared

boundless. Bankrolled primarily by a great deal of private capital, expert

systems

moved out of

the laboratory and into the marketplace, where

si

PREFACE

many of

the

new companies

have always been part of a community that has a

artificial intelligence

common

foundered. But those people involved in

which numerous approaches

ideal, in respect to

rival

The

other for attention, funding, and the chance to succeed.

expert systems rekindled interest in the rival technology of

webs of interconnected,

neural networks, those

each

failure

specially

of

artificial

designed

processors that could be "trained" (through the application of varying degrees of voltage) to respond

At

that point,

on

own

their

began gathering material for

I

following years, which

I

Al-related business,

we may soon have

that

is

and running

my own

a

computer for

a

world

taking hold of the marketplace with litde noise or

Meanwhile, large AI projects

computers with

field,

A second generation of expert systems, this time more

savant than idiot, fanfare.

book. Over the

observed the next stage of the rollercoaster ride

I

now seems

chess champion.

this

spent renewing contacts, attending conferences,

interviewing the founders and stars of the

begin. It

to external stimuli.

common

worldly wise machines

CyC and SOAR seek to endow

like

sense in the not-too-distant future; such

may

acquire knowledge and refine their reason-

power by themselves, much as humans do, but with the advantages attendant on greater computer speed and memory.

ing

The

story of

naysayers,

AI

consists of successes

new hardware and

lots

and

failures, visionaries

of programs.

It is also

and

the story of a

how humans think. means uncovering the

slow but steady acquisition of knowledge about After

all,

the pursuit of

artificial intelligence

unbelievably complex layers of our

own

thought processes. The quest some crucial philosophical ishumans if researchers in artifi-

for artificial intelligence therefore raises sues.

What will be

cial intelligence

the consequences for

succeed and force us to share the world with

we

smarter than ourselves? Are

of the species that will replace to

make not only business and

legal,

and moral choices?

scientific decisions for us,

raised these

I

the events

I will relate

happened

emphasis stems only in part from

where ing

I

studied and keep so

AI research occurred

in

many

now

social,

in the

may

my own

friends:

bias

it is

surprise you.

United

States.

This

toward the country

a fact that

most pioneer-

America, probably because of the overbear-

ing interest of the U.S. military. are

but also

and similar questions with the

leaders of AI: their answers (and your conclusions)

Most of

entities

new Renaissance or the creation us? And should we rely on these creations facing a

My apologies

catching up: future accounts of

AI

to Japan will say

and Europe,

who

more about them!

Acknowledgments

am I me

deeply grateful to those participants in the history of AI

and help

their time

Particular thanks

in preparing this account

must go

to

of

who

lent

their activities.

Marvin Minsky, Herbert Simon, Allen

Newell, Gerald Sussman, Daniel Dennett, Joseph Weizenbaum, Ber-

John McClelland, Hans MoraDavid Waltz, Patrick Winston, Guy Lapalme, and

thold Horn, Randall Davis, Carl Hewitt, vec, Raj Reddy,

Marion

Finley,

Roger Schank,

who

granted

also read

me

interviews. Several of them, as well as

and commented on relevant sections of the

manuscript.

Many

thanks also to those friends and relatives

who

reviewed the

manuscript and supplied their comments: Simon Pare, Jacques Martineau, Francis Dupuis, Marie-Josee Rousseau, Laurens R. Schwartz,

Dowey, and, of course, Paul-Henri Crevier, my father, who helped out on the neurology and philosophy. Basic Books's Susan Rabiner must be acknowledged for her prodding help throughout; and Phoebe Hoss, for her enlightened revision of the Claire

manuscript.

Many

thanks also to those publishers and authors

who

granted

me

permission to quote from their copyrighted material: Academic Press, the I

American Council on Education, and Patrick Winston. could not have completed the project without the understanding

xlv

acknowledgments

attitude

toward book writing of my employer, the Ecole de Technologie

Superieure, which final

I

thank for providing time and secretarial help

months. Jocelyne

Hall, in particular,

bringing the illustrations into

My

final

deepest gratitude goes to

tience, love,

in

shape.

my

wife, Celine, for her unfailing pa-

and support throughout the long years

the project through.

in the

was of invaluable help

it

has taken to carry

Introduction

PROBING THE MYSTERY

HUMAN INTELLIGENCE

OF

It is not

my aim

or shock you

to surprise

now

to say that there are

Moreover, their ability to do these things future the





but the simplest way I can summarise

in the world machines that think, that learn is

going

and



to increase rapidly until

is

that create. in

a

visible

the range ofproblems they can handle will be coextensive with the range to which

human mind has

— Herbert Simon, 1957

been applied.

driven by some invisible hand, humans have always yearned to Asunderstand what makes them think, and be, and have tried to if

feel,

re-create that interior silicon chips

Long before

life artificially.

of the modern

digital

computer, mythologies and

the

vacuum

computer, long before the

literature

tubes and

first

mate the inanimate, from Pygmalion's attempt to bring to perfecdy sculpted Galatea to Gepetto's desire that the

Pinocchio be a

An

early

the royal city of

until the

in the

Napata

machine" was the in ancient

eligible heirs

god made known

grab the successor.

A priest,

the

life

wooden puppet

real boy.

"man

demise of a pharaoh,

Amon

analog

recorded a timeless need to ani-

Amon

statue

of the god

Egypt around 800

b.c.

Amon Upon

in

the

were marched past the statue of

his choice

by extending

his

arm

then "delivered" a consecrating speech.

to

1

of course, controlled the statue with levers and uttered the

I

2

A

sacred words through an opening on the back of the statue. Although likely that

those

who

and onlookers

alike,

knew

it is

took part

procedure was taken seriously: statue system

added up

to

of theater was being put on, the

in the

more than

would-be pharaohs

in the process,

that a bit

Egyptian mind, the priest-cum-

the

sum of its

and embodied

parts,

the god.

The

more sophisticated self-activated automata, atwork of the divine blacksmith Hephaestus. Among

Iliad describes

tributing

them

to the

the golden female attendants were gangly walking tripods, forerunners

of today's many-legged

how

field robots.

In the

molded

The truvius

such

likely inspiration for

tells

city

literary

automata were

of Alexandria. The

us that between the third and

and amusing things of

Icarus,

first

kinds

all

.

.

.

ones being

real

Roman

architect Vi-

centuries B.C., a school

of early engineers founded by Ctesibius "conceived vices

also recounts

woe of his son

which patrolled the shores of Crete. 2

the copper giant Talos,

engineered in the Greek

Homer

Iliad,

Daedalus, wing maker extraordinaire to the

.

.

.

automatic de-

ravens singing through

contrivances using the force of water, and figures that drank and

moved

3

about." Similarly, Hero, author of a treatise called Automata? treated his readers to a description of a steam-propelled carousel;

it

now

stands as

the first-known description of a steam engine!

In Europe, the late Middle Ages and the Renaissance saw a resur-

gence of automata. Roger Bacon reportedly spent seven years constructing talking figures. 5

automaton

in the

To honor

shape of a

Louis XII, Leonardo da Vinci built an

lion.

During the sixteenth and seventeenth

and French designers, such as Gio Battista Aleotti and Salomon de Caus, elaborated on designs of the Alexandrian school:

centuries, Italian

gardens and grottoes resonated with the songs of flutists,

pepped up

aristocratic receptions.

automaton

that

while

fully

artificial

birds and

animated nymphs, dragons, and satyrs

mechanical

6

Rene Descartes

built in

1649 an

he called "my daughter Francine." But on one of Des-

cartes' trips, a superstitious ship captain

happened

containing Francine and, frightened by her

lifelike

to

open the case

movements, threw

case and contents overboard.

By

the eighteenth century, the French artisan Jacques de

Vaucanson

could assemble his celebrated duck, which provided an almost complete imitation of its model.

One

prospectus described

it

as

an

of golden Copper that Drinks, Eats, Croaks, Splashes Digests as a live duck." Life-sized, the animal rested

on

"artificial

in

duck

water and

a waist-high case

INTRODUCTION: PROBING THE MYSTERY OF HUMAN INTELLIGENCE containing a

drum engraved with cogs and

passing through

the duck's legs caused

grammed in the Memory (ROM)

device.

Although

we would

cogs: today

artisans like Pierre

had reached

call

grooves.

3

Control rods

move in a manner prothe drum a Read Only

and Louis Jaquet-Droz kept producing

automata for the pleasure of the rich cal arts

to

it

7

their limits.

late into the century, the

The next

mechani-

leap required the automatic

switch, or electromechanical relay. This device contains an iron kernel

which can change positions under the influence of a magnetic erated by a current fed to the relay.

Depending on the

field

gen-

polarity

of

the current, the kernel opens or closes electrical contacts to motors, lights,

or even other relays. Interconnected relays can be used to build

mechanisms more complex than those allowed by interlocking gears and cams.

At the turn of the century, the Spaniard Leonardo Torres y Quevedo built a relay-activated automaton that played end-games in chess. The philosophers Leibnitz and Pascal had constructed mechanical computing devices centuries before, but those were perceived as soulless contraptions intended to alleviate the drudgery of addition 7

De

and subtraction.

Vaucanson's duck was a superb emulation of a living creature. But

a device that

would play chess

completely different:

it

against

seemed

to have the ability to think.

In early chapters of this book,

vacuum tubes and

transistors,

human opponents was something

I

shall describe

how relays

which formed the

computer, the invention that made

evolved into

basis for the digital

artificial intelligence

possible. Digital

computers are simply devices for manipulating discrete pieces of information,

of

initially

taken to correspond to numbers.

artificial intelligence

was

The

insight at the root

that these "bits" could just as well stand as

symbols for concepts that the machines would combine by the rules

strict

of logic or the looser associations of psychology. European philos-

ophers and mathematicians had been probing the issues involved in representing the world through abstract concepts for millennia; and they

had been asking

in the process

fundamental questions about the nature

of mind and thought. The American pioneers of AI,

and

later in

at first unwittingly

enlightened fashion, began to tap the insights of these

humanist predecessors. Expressions such

as "experimental epistemol-

ogy" and "applied ontology" then started to describe systematic, downto-earth research projects in

Emerging

as

it

computer

does from

many

science.

fields



philosophy, mathematics,

4

\|

psychology, even neurology

about

human



intelligence raises basic issues

artificial

memory,

intelligence,

mind/body problem, the

the

origins

of language, symbolic reasoning, information processing, and so

— from base metal — seeking

AI researchers

are

who

alchemists of old

like the

to create thinking

forth.

sought to create gold

machines from

infinitisi-

mally small bits of silicon oxyde.

The

birth

of AI was

tied to the efforts

trical

engineers, psychologists, and even

became

these figures never really ideas,

developed

mining the tions as

of a variety of

talented, in-

well-educated budding mathematicians, elec-

tellectually self-confident,

of the

early direction

political scientist.

of AI research per

a part

in other contexts,

AI researchers and then

one

were enormously

se; yet their

influential in deter-

Others made seminal contribu-

field.

left

Some of

the field to pursue other work.

few, there at the creation, remain there to this day.

The

A

various origins

of the creators of AI and the enormous influence of their work explains

some of

the colorful aspect of their pronouncements, and the roller-

coaster evolution of their trated in 1989,

even the

field.

As

the hoopla over cold fusion

of physics

staid science

is

not

illus-

immune

to

exaggeration and false claims. Yet excesses of optimism seem to occur

with particular frequency in AI. There are several reasons that AI

workers were, and often

still

are,

more

likely to

make exaggerated

claims

than their colleagues of other disciplines. First, there its

were plausible reasons

in AI's early years for believing in

rapid progress, and these induced early researchers to display the

excessive optimism that in using

came

to characterize the field. Early progresses

computers for arithmetic were

truly breathtaking. In a

few

years,

technology went from cranky mechanical calculators to machines that could perform thousands of operations per second.

It

was thus not

unreasonable to expect similar progress in using computers as manipulators

of symbols to imitate human reasoning.

One

misconception further enhanced

this temptation. Psychological

experiments in the 1950s and 1960s pointed to an oversimplified picture

of the mind. The Carnegie Mellon researcher and AI pioneer Herbert its

way around complex

obstacles, saying in effect that complexity lay in the

environment and not

Simon compared in the

mind

itself.

it

8

to a dim-witted ant

There

between ant and human

is

winding

truth in this statement, but bridging the gap

still

requires a giant leap in complexity. Yet, in

the postwar decades, controlled studies

reason showed their basic limitations.

on our ability to remember and The next time you look up a

INTRODUCTION: PROBING THE MYSTERY OF HUMAN INTELLIGENCE seven-digit

phone number,

seconds before

try thinking

dialing. If you're like

forget the number.

else for a

this will

few

make you

We can't keep more than five to nine items at a time

our short-term memory; and

in

of something

most people,

5

Our long-term memory

soon

as

as

we look

away, they vanish.

has an almost unlimited capacity, but

very slowly. Transferring an item from short-term to long-term takes several seconds.

(It

takes

me

it

learns

memory

about two minutes to learn by heart

phone number.) When we rate alternatives in any complicated problem, like a number puzzle, we need pencil and paper to make up for these deficiencies of our memory. the seven digits of a

Early

AI

from these

researchers reasoned that their computers did not suffer limitations.

Even

in those days, the

machines had memories

with capacities of thousands of items and access times of microseconds.

They could shuffle data much faster than fumbling, pencil-pushing humans can. Computers, thought Simon and his colleagues, should be able to take advantage of these capabilities to overtake humans: it was only a matter of a few years before suitable programming would let them do

it.

These researchers had not

realized that, in activities other than

purely logical thought, our minds function

puter yet devised.

They

much faster we are

are so fast, in fact, that

than any com-

not even con-

scious of their work. Pattern recognition and association

make up

the

core of our thought. These activities involve millions of operations carried in parallel, outside the field to hit a wall after inability to

of our consciousness. If AI appeared

earning a few quick victories,

it

did so owing to

its

emulate these processes.

Already deluded by

false expectations, early

further onto the path of exaggerated claims

by

a

AI workers were drawn myriad of other

One was

the recent and sudden emergence of

discipline.

Like

all

new frontiers, AI

AI

as

factors.

an identifiable

attracted a particular kind of person:

of academic security in the new

one

willing to face the lack

live

with the haphazard financing of the early years. So novel were the

insights offered

by the new technology that

early researchers

field

and to

looked

like

As some branches of deeply affecting modern philoso-

elephants in the well-tended flower beds of conventional sciences.

we

shall see,

AI brought about major

psychology and mathematics. phy.

To one

It is

also

revisions in

used to bringing forth such innovations, moderation

is

a

hard virtue to learn. I

have already mentioned that AI

is

a multidisciplinary science.

As

Herbert Simon told me: "AI has had problems from the beginning.

It

I

6 is

A

a

new

field in

which people came from many

know where many people imported them. And we still has meant they didn't always

norms of wheels

responsibility,

the time." 9

all

One

haven't established a set of

of Herbert Simon's best-known students, his

illustrates the incompatibility

Ph.D.

engineering graduate, Feigenbaum was

and worked on

registered at the business school,

ously belonged to psychology: the modeling of fields

computer

from which AI researchers emerge nowadays (psychology,

its

they are often at odds with each other.

common

own

accepted methodology, and

The

different branches

language, values, or standards of achievements.

discipline acts as a

moderator on other

fields

and

The need

it

of AI lack

A

uniform

of science and enables

AI

research communities to police themselves. fluence,

a subject that previ-

human memory. Each

science, linguistics, physics, philosophy, mathematics, neurol-

ogy, or electrical engineering) has

a

of disciplines that well

in a tangle

of AI with the academic structures of the

An

1950s and early 1960s.

of the

came from, because

of referencing, that keeps us from reinventing

Edward Feigenbaum, earned late

That

different directions.

things

their

lacks that sobering in-

shows.

to attract research

in researchers, especially in a

money can

young

space research, and astronomy,

also induce careless behavior

science. Like nuclear physics, aero-

AI

gets

its

funds from government

AI researchers found way they could channel money away from these traditional was to proclaim their merits, and the louder the better. Even

sources. Davids against Goliaths, the

young

early

that the only disciplines

AI

today,

researchers have a vested interest in AI's appearing solid and

confirmed. Public discussion of the failures and are against their interest.

predictions and

Simon

They now do

empty promises

likes to say that

AI

is

difficulties

of the

field

however, that rash

realize,

are not to their advantage. Herbert

"hankering for respectability."

It is

perhaps

symptomatic that many younger researchers do not repeat the errors of their elders in this respect. "I refuse to

Horn

thold

researcher

was

told me.

He compared

make

predictions,"

his attitude to that

MIT's Berof an older

who

in charge

of the AI conference

in

Boston ten years ago. There

were reporters swarming around, and he was saying things years

up

from

now

we'll

the things that

like

"Five

have robots going around your house picking

you dropped on the

floor."

I

dragged him into a

corner and told him, "Don't make these predictions! People have

INTRODUCTION: PROBING THE MYSTERY OF HUNAN INTELLIGENCE

done time

before and gotten into trouble. You're underestimating the

this

He said, "I don't were after my retirement

will take."

it

chosen

and people

retired

7

care.

date!"

come back and

will

Notice that

all

the dates I've

said, "Well, I

I

me why

ask

won't be

they don't have

10 robots picking up socks in their bedrooms!"

A

AI

further spur to boastfulness about

much of

that reporters can explain

is

The

in words understandable to anyone.

it

and

final goals

A commore imme-

everyday successes of AI are close to our everyday concerns. puter that beats chess champions or diagnoses diseases has diate

impact than the discovery of another elementary particle or prog-

ress in gene- splicing technology. a temptation

is

hard to

been above making a

resist,

little

For

a reclusive scientist, the limelight

and some

more of

have probably not

scientists

their discoveries to get front-page

coverage. Finally, fast-evolving fields like

AI

are

more

subject than others to a

perennial problem in technological forecasting. In

all

domains, research-

ers invariably overestimate the short-term potential

the speed of

its

One

progress.

result

is

more vulnerable than other high-technology grams

that, like aerospace,

of

their

cost overruns, to

work and

which AI

is

projects. Contrary to pro-

AI

involve national prestige or security,

cannot overcome errors of judgment through sheer budgetary excesses.

Often AI researchers helplessly watch a reputation for not delivering

their finances

Let's not forget, though, that the other

technological forecasting

onlookers

at Kitty

Hawk

is

most common mistake

to underestimate long-term achievements.

never imagined today's

airliners.

never thought of Nagasaki. Believing today in the

be

like deciding, after the

run dry, and gain

what they promise.

Vanguard

in

The

Marie Curie

of AI would

failure

flops in the 1950s, that space travel

was impossible. I shall, in

forecast.

the a

first

probing the past of AI research,

Having recounted the

four chapters,

I shall

origins

I

try to arrive at a better

early

golden years of AI in

demonstrate in chapters 5 to 8

few amazing successes, AI has not so

promised.

and

shall investigate the

far delivered

that, despite

what the pioneers

reasons for this state of

affairs

and

examine more recent developments to see whether there are reasons to believe that early promises might be fulfilled. I shall first

establish the conviction

To answer

this question,

of most modern philosophers that

our minds are essentially a product of the complex physical processes

.

8 in

Al

our brains.

Do

I

shall

then ask,

How

is

thought generated

our machines accomplish anything similar to

gree? This will be the subject of chapters 9 to

that,

in the brain?

and to what de-

1 1

Recent research and the trends of the past decades indicate that

machines

just as clever as

too-distant future.

Such

a

humanity's self-esteem, to and, indeed, to

its

concluding chapter.

human

beings

development its

meaning

very survival.

I

may indeed emerge will raise

in the overall

shall

in a not-

deep challenges to

scheme of

things,

examine these questions

in a

1

ENGINEERING

INTELLIGENCE: COMPUTERS

AND PROGRAMMING

I .

about

believe that in .

to

.

make them play

fifty

years' time

the imitation

it

game

will be possible to

so well that

programme computers

an average

interrogator will not

have more than 70 per cent chance of making the right identification after five minutes

— Alan

of questioning.

A

definition

of

artificial intelligence

this art is that

accepted by

of MIT's Marvin Minsky: "AI

is

many

Turing, 1950

practitioners

of

the science of making

machines do things that would require intelligence

if

done by men."

1

The machines involved are usually digital computers, and they can be "made" to do things by programming them in certain ways. "Computers" and "programming" will make up the two themes of this chapter. I shall first review how computers came about, and provide an overview of

how

they work. Next,

programming,

early philosophers

embodiment

I

shall

logic, calculation,

in

examine the relationships between

and thought, and

how

the inquiries of

and logicians into the nature of thought found

contemporary computer programs.

their

10

Al

COMPUTERS Early Devices It

can be argued that computing devices emerged

much

the

same reason

have only ten fingers and ten

and

larger

in

our century for

we

that the abacus did centuries earlier: because

and computations involving

toes,

numbers required devices

larger

that can handle these greater

sums

with better accuracy and speed. Stone-age calendars are the external

mechanisms

evidence of

first

on

this desire to rely

to alleviate mental burdens. Rather than tediously

counting the days to crop-planting time, our prehistoric ancestors used for alarm clocks the coincidence

of

celestial

bodies with lines of sight

defined by carefully positioned ground markers. In the Orient, the

abacus helped out in numerical calculations even before the invention

of long-hand arithmetic.

With technology showing

its

metallic head, the lack of extensive

number-crunching power became a serious

on

tion tables sent ships crashing

imprecise calculations, tumbled down.

manual

liability.

coastlines;

Error-filled naviga-

and bridges,

To weed

calculations, several inventors, including Leibnitz

tried their

hands

at building

and Pascal, 2

mechanical calculators. They met with

limited success, partly because of the cost of their complex, devices,

and

partly because

of their specialized natures:

could perform only one basic arithmetic operation a sequence

upon

built

out mistakes from

hand-made

early calculators

at a time.

Performing

of operations involving several numbers involved many

lengthy manipulations by the users. In other words, one could not

program these machines to perform several operations.

Oddly enough, with numbers.

the

first truly

programmable device had nothing

The Frenchman Joseph-Marie Jacquard invented

1805 to drive looms: removable punched cards

weave

different patterns.

Some

let

the

do

to it

in

same machine

forty years later, the British inventor

Charles Babbage picked up the idea of punched cards to feed instructions to his ill-fated "analytical engine." This steam-driven contraption

would have contained, had

it

ever been built to specifications,

elements of a modern computer, including a unit

(I'll

define these

words

memory and

all

the

processing

shortly).

Babbage teamed up with Augusta Ada, countess of Lovelace and

ENGINEERING INTELLIGENCE: COMPUTERS AND PROGRAMMING daughter of the poet Lord Byron. She

is

computer programming, the science of tions to

perform on what pieces of

11

often credited with inventing

computer what opera-

telling a

data. Unfortunately,

Ada never

had a chance to run her programs because Babbage's grand ideas crashed against the same technological limit that had put a cap on the progress of automata a century

earlier.

Nineteenth-century mechanics

couldn't produce sufficiently accurate parts, and the analytical en-

still

gine was never completed. Babbage and Lovelace later tried to recoup their losses

by inventing

a chess-playing

machine and

machine. They even devised a system for winning

which sions. tary,

in fact forced the countess to

Nowadays

the

perpetuates the

By

1

memory of

on two occa-

they nevertheless

made

filled a

mili-

the countess of Lovelace.

890, hand-driven mechanical calculators

one, would have cost

at the race track,

jewels

computer language Ada, favored by the U.S.

Babbage's visionary device sive,

pawn her

a ticktacktoe

much more modest than

their appearance.

crying need.

more than $100,000 in

Clumsy and expen-

The Ohdner calculator, today's money and took

for

ten

minutes to perform a multiplication! In the same year, the American

Herman

a tabulating maon punched cards. Hollerith's Tabulating Machine Company eventually merged into a conglomerate that became IBM. Such machines were called "digital calculators," because they represented numbers by cogged wheels or related devices, which could take only a fixed number of positions corresponding to the digits in the numbers they denoted. For example, if a wheel had ten possible positions, it could represent the digits from to 9. Even then, though, the cheapest and most efficient way to speed up a calculation

Hollerith invented for the U.S.

chine that processed census data fed to

government

it

remained the abacus. For technical work, the alternative

when

it

didn't matter

slide rule

offered an

whether the answer was off by

a

few

percentage points.

Claude Shannon and His Switches In the cal.

first

third

of this century, calculators remained

essentially

mechani-

In 1 93 1 Vannevar Bush at MIT brought to its pinnacle the technology ,

of computing through mechanical devices. His

much more

than add or multiply numbers:

it

differential analyzer did

actually solved differential

equations, using rotating shafts as integrating devices. (Bush's machine

was an

analog calculator: the angular position

of a shaft could take any

12

Al

value,

and the machine's accuracy was limited only by how precisely one

could manufacture the shafts and measure their positions. By contrast, the accuracy of a

how many

depends on

digital calculator

[cogged

digits

wheels] are used to represent numbers.)

Bush hired

In 1938,

Shannon, to run the

a

twenty-two-year-old research assistant, Claude 3

differential analyzer.

Even though

the heart of the

device was mechanical and analogical, a complicated digital circuit using

electromechanical relays controlled

Couldn't one make the circuit

it.

This fact set Shannon thinking.

complicated than

less

it

was?

And how

about using the relays themselves for computing instead of the spinning disks?

These considerations

Shannon

led

to

show

that

one could

build,

using only interconnected switches that could turn each other on or a digital calculating

off,

machine performing any imaginable operation on

numbers. Further, Shannon's theory showed how, by examining the nature of the operations to perform, one could arrive at a configuration

of switches that embodied of operating much digital

The

this operation.

faster than the

switches had the potential

cogged wheels previously used

in

machines. Since, however, they could only take two positions (on

or off), only two values, taken to be

and

1,

were

representing numbers in the machines. This

available for the digits

why computers

is

started

using binary arithmetic.

And

so

it

was

that in the late 1930s, electromechanical relay switches

started replacing gears

and cogs

World War, Howard Aiken

at

in calculators. Just before the

Harvard

human

being.

soon replaced electromechanical

relays.

calculate twelve times faster than a transistors

driven switches embodied no faster than relays,

moving

tubes and containing no

World War.

It

parts

moving

Vacuum These

could

tubes and

electronically

and could operate much

which were limited by how

could change positions. Thus the

Second

built a calculator that

fast their iron kernels

electronic

computer, based on vacuum

parts,

appeared during the Second

was invented not once, but three

times: in

Germany,

airplane design; in the United States, for calculating artillery tables; in

England, for breaking

German

secret codes.

for

and

The American machine much power as a

contained 18,000 tubes, weighed 30 tons, used as locomotive, and would have

filled a

tions per minute, the Electronic

tor

(ENIAC) was

competition.

a

ballroom. But at 20,000 multiplica-

Numerical Integrator

thousand times

faster

than

its

And

Calcula-

relay-operated

13

ENGINEERING INTELLIGENCE: COMPUTERS AND PROGRAMMING

Von Neumann and

His Architecture

ENIAC

to uses other than war, a

When

put

scientists

major design

To change the sequence of operations per(what we now call the "program"), engineers had the data on formed to rewire hundreds of connections in the machine. John von Neumann

weakness became

is

clear.

way to

usually credited for pointing the

a better

computer architecture

in 1945.

Names can

mislead: as

German. 4 Born

Chopin wasn't French, von Neumann wasn't

Budapest

in

family

moved

mann

could rightly count

game

in 1903, this scion

of an upper-class Jewish

to Princeton in the 1930s. In the mid-1 940s,

among

von Neu-

his contributions the invention

of

theory; the theory of automata (which discusses the possibility that

machines might be able to reproduce themselves); and, in the

field

of

hydrodynamics, calculations of shock-wave propagation, which were used during the Manhattan Project to help trigger and control the chain reaction of a nuclear explosion in

its

early phases.

He was

also the author

of a celebrated essay on the mathematics of quantum theory, which was said to

have inspired Alan Turing

(whom

I shall

come a mathematician. 5 Remembering the difficulty he had had

discuss shortly) to be-

earlier

using a mechanical

desk machine to calculate shock-wave propagation, von Neumann, hearing of the platform,

ENIAC

became

what

was

one could to

do

store the sequence

in the

same

also the first

of instructions

one to use the term memory for

computer: Babbage had called

He

telling the

used to hold the data. (Von

circuitry

it

a train

answer to the problem of changing the

computer's instructions appears in retrospect very simple. that

on

project in a casual conversation

fascinated. His

the "store.")

The

realized

machine

Neumann

this part

so-called

of the

von Neu-

mann computer the Second

The

architecture, embodied in virtually all computers World War, breaks a computer into two parts.

since

(CPU) operates on the data items

to be

central processing unit

manipulated. These data items (numbers or symbols) are stored in the part of the computer. The memvon Neumann machine operates in well-defined cycles: Fetch the first instruction from memory. Fetch the data item to operate upon from another part of memory. Perform the operation. Fetch the next instruction from memory, and so on.

memory, which makes up the second ory also contains the program.

A

14

Al

The

Electronic Discrete Variable

Computer (EDVAC)

first

embod-

RAND

Corporation's

JOHNNIAC, so named in honor of John von Neumann.

Following von

ied this architecture.

Neumann's penchant

It

was followed by the

for puns, the creators

of another machine couldn't

new machine Mathematical Analyzer, Numerical Inte(MANIAC). von Neumann's colleagues did not enjoy his Old World charm

resist calling their

grator

And

Alas,

Calculator

and

jokes, or

war.

He

endure

his impossible driving habits, for

long after the

died in 1957 of cancer, perhaps induced by radiation exposure

during the Manhattan Project.

COGNITION AS COMPUTATION ror

several years following their invention,

computers were generally

perceived as devices for manipulating numbers and straightforward items of data such as names in a telephone directory. However,

became

clear to

some of

switch positions inside the machines could take particular, they

it

soon

the computers' inventors and users that the

on other meanings. In

could stand for symbols representing concepts more

abstract than straightforward data. If the

these symbols as specified in

its

computer then manipulated

program, perhaps then

it

could be said

to "think." This concept of cognition as computation had been the subject

of much debate throughout the history of philosophy and mathematics.

Could one represent

Could thought

result

all

things under the sun through a set of symbols?

from the manipulation of these symbols accord-

ing to a set of predefined rules?

symbols be? As we

AI

shall see,

And

if so,

what should the

rules

and

such questions found their echoes in early

efforts.

Early Attempts to Formalize Thought The

thirteenth-century Spanish missionary, philosopher, and theologian

Ramon

Lull

artificially

is

often credited with making the

first

systematic effort at

generating ideas by mechanical means. Lull's method, crude by

today's standards, simply consisted in

randomly combining concepts

through an instrument called a "Zairja," which the missionary had

brought back from his travels in the Orient. A Zairja consisted of a circular

15

ENGINEERING INTELLIGENCE: COMPUTERS AND PROGRAMMING

with concentric disks on which appeared letters and philosophi-

slide rule cal

symbols. 6

The combinations obtained by spinning the

to provide metaphysical insights. Rechristening it without

the

Ars Magna (Great

disks

were

said

undue modesty

Art), Lull generalized the Zairja to a host

of other

beyond metaphysics and turned the instrument into the Middle

fields

Ages equivalent of a computer for blending books dealing with various applications of

wrote dozens of

ideas. Lull

Great Art, ranging from

his

morals to medicine and astrology. For every subject the method was the same: identify basic concepts; then combine them mechanically with

themselves or ideas pertaining to a related

field.

Yet merely generating random combinations was only a small step toward mechanizing thought: to interpret

one

also required systematic

first

means

and evaluate the combinations. In the seventeenth century,

the diplomat, mathematician,

and philosopher Gottfried- Wilhelm Leib-

nitz suggested the possibility

of a

lus, to

calculus ratiocinator,

or reasoning calcu-

achieve this goal. Apparently following Lull's lead, 7 Leibnitz

number 8 and

hoped

to assign to every concept a

issues

by formally manipulating these numbers. The diplomat

nitz

foresaw such an instrument as a

common

to resolve the thorniest

language

among

in Leib-

nations.

Leibnitz never achieved his objective of completely formalizing

thought and in time became keenly aware of the taking.

of

all

One major

concepts: "There

no

contains

difficulty

stumbling block, he noted, lay in the

relations

other things or even

is

no term so absolute or detached

and of which

all

a perfect analysis

other things." 9 Three centuries

researchers, trying to carve

of the under-

interconnectedness

up

reality into

"micro worlds," would also founder on

this

that

it

does not lead to later,

modern AI

convenient niches called very issue.

Boole and the "Laws of Thought" The

first

recognizable glimmerings of the logic that would later be

implemented into computers emerged from the work of

a self-taught

Englishman, the son of a shoemaker, named George Boole. Boole eventually century.

became one of the most

To

influential thinkers

of the nineteenth

help support his parents, he became an elementary school-

work occurred before and after the room he plowed through advanced mono-

teacher at age sixteen. But his real

school day

when

alone in his

graphs in mathematics, learning them with the same thoroughness that

had

earlier

marked

his

mastery of Greek and Latin.

A few years later, he

16

Al

was publishing

mathematical journals. By the time he was thirty-four,

in

even though he did not have

a university degree,

Boole was appointed

professor of mathematics of the newly founded Queen's College at Cork in Ireland.

There he attempted nothing

than a mathematical formu-

less

He started by investihow one could combine classes and subclasses of objects, and then how such classes intersected with other classes. Boole showed how one could draw useful conclusions from such analysis. He assigned of the fundamental processes of reasoning.

lation

gating

symbols to the operations of combining either (which he

elements of two sets

all

named "union"), elements belonging

to

both

sets ("intersec-

tion"), or objects falling outside a given set ("complement"). In this

way, operations on sets could be represented in a crisp shorthand. For

example,

A u B

meant "the union of

sets

A

and B." Using these

symbols, Boole could analyse and simplify complicated operations volving

many

sets,

much

as his fellow

in-

mathematicians could manipulate

ordinary algebraic equations. Boole formulated simple and well-defined

laws to perform these simplifications (see figure

1.1).

854 book, The Laws of Thought, Boole stated that these principles were fundamental descriptions of thought. In the tide of his celebrated

He was

pardy

right.

oranges, one should

or

of

class)

not

all.

After

know

fruits. It also

Further, not

some form of

sugar

all

all,

that

1

to talk intelligendy about apples

both belong to the wider

helps to

know

that

some

set (or category

apples are red, but

red fruits are apples, although they

when

ripe,

and

and so on. For the

all

time,

first

contain

Boolean

algebra enabled a rigorous and quasimechanical manipulation of categories,

an

activity basic to

human

thinking.

Boole could claim universality for Replace the concept of true or false.

That

is,

sets

by

his laws in yet

another way. 10

logical propositions that

instead of the set of

sentence "Boole was born in 1815," which

is

all

can be either

apples, consider the

true. Further, replace the

operators union, intersection, and complement by the logical operators

OR, AND, and NOT. These can combine

logical propositions to

form

other propositions. For example, the combined proposition "Boole was

born

in

1815

AND

he died

in

1816"

propositions "Boole was born in 1815

was born

in

1815

is false.

On

the other hand, the

OR he died in

AND he did NOT die in

1816" and "Boole

1816" are both

true. It turns

out that these operators and propositions combine together in a manner exactly analogous to the set-theoretic operators union, intersection,

complement. Thus, assumed Boole,

if

the

and

mind works according

to

A 17

ENGINEERING INTELLIGENCE: COMPUTERS AND PROGRAMMING

FIGURE The Here

1.1

Postulates for Boole's

Laws of Thought

are Boole's laws as they apply to logical propositions

case, the objects studied, basic operations,

and

set theory. In either

and identity elements are

as follows:

Logic

Sets

logical propositions

Objects Studied

collections of objects

b)

(a,

(A,B)

AND:

Operations

O:

equivalent to the

intersection

preposition "and"

Elements

Identity

In these two fields of study,

it

OR: "or"

U: union

NOT:

~"

"not"

complement

:

1:

true

I:

0:

false

(J):

universal set

empty

set

can be shown by inspection that the following basic

facts, called postulates, are true.

Operations are

a

commutative.

a

There are identity elements

a

for the

two operations.

Each operation

AND b = b AND OR b = b OR a OR AND

a

distributes

a

over the other.

(a

= 1

AOB

a

A

a

=

OR (b AND c) = OR b) AND (a OR

AND (b OR c) AND b) OR (a AND c)

=

a

a

AND (NOT

=

a

OR (NOT

From

these basic facts, and others that follow

logical expressions, or sentences

about

manipulate algebraic expressions.

One

a)

=

a)

sets, in

U (B n Q = U B) n (A U Q a n (B u Q = (a n b) u (a n Q

could do

which elements

and multiplications,

is

+

AU""A

much

it is

is, it is

=

(|)

I

possible to manipulate

same way as one would more easily, in fact, because

the

this a little

sets.

More

specifically,

numbers and the operations

not a Boolean algebra because addition

over multiplication. That {a

are the real

""A =

API

1

ordinary algebra offers less freedom than does logic or algebra, in

A

(A

from them, very

A c)

(a

complement.

=

H UA

Am = a

a

a

Each element has

= B B = B

U U

A

not generally true that a

+

(b

not distributive

is

x

ordinary

are additions

c)

=

{a

+

b)

x

,).

these laws,

it

performs

logical operations in the

same way

it

manipulates

sets (see figure 1.1).

In

fact,

Boole had

laid the

foundation for analyzing thought in more

ways than even he had foreseen. Ninety years

after their publication,

18

Al

Boole's ideas supplied the basis for Claude Shannon's analysis of switchI have described, makes up the modern computers. Shannon's intuitive

ing circuits, which, as

theoretical founda-

tion for

leap

all

was

to realize

that switches resembled logical propositions in that they could take only

two true

and

positions, open

If

closed.

one took these positions

to stand for

and false, one could then analyze combinations of switches with the

same mathematical machinery illustrate this point,

sitions

and

their

I

Boole had used for propositions.

that

To

have drawn up examples of simple Boolean propo-

embodiments

in switching circuits in figure 1.2.

Reasoning Calculuses, or the Fundamental Impotence of Logic Yet Boole's laws did not shape up to

a

complete calculus for reasoning.

"Laws of

Essential elements were missing. Boolean algebra, as his

Thought"

are

now known,

could not serve as a complete generic tool

for expressing logical sentences because of

you assign true or

false values to basic

its

lack of flexibility

7

propositions such as "I

It lets

.

own my

house" or "Mary owns her house," but cannot express statements such as

"Every house has an owner." Boole's formalism prevents the creation

and manipulation of statements about general or Further, each Boolean proposition totally

beyond

ability to

reach.

A

false logical propositions,

sentences that could be true or

Then

came up with such

and combine these into

system in 1879. 11

mathematics

at the

on Boole's system by introducing

the

contains arguments that are not logical variables: the predicate

true ifj really

The

insides

A predicate is a logical entity with a true-false value.

OWNS(x,j,) could mean

false.

its

require the

The German mathematician Gott-

thirty-one, Frege, an assistant professor of

concept of predicates. it

false.

a

University of Jena, improved

But

an unbreakable atom,

more powerful formalism would

define basic elements (such as house and owner) that are not

themselves true or

lob Frege

is

indefinite objects.

owns

x,

that person j

x and j,

but

owns house

x.

OWNS has value

in themselves, are neither true

nor

A further refinement of Frege's system introduced two quantifiers. universal quantifier

logical proposition

is

V

x,

which means "For

true for

quantifier 3 j ("There exists

all

all

x," denotes that a

values of variable x.

aj such

that")

means

The

that at least

existential

one value

FIGURE

1.2

Three Switching Networks and Their Corresponding Boolean Propositions

Switch Switch

k

a:

If

you

Switch

are driving

and there

is

Current: then you

Switch

a stop sign

must stop

If

a:

you

and there

b:

Switch

c:

are driving is

a traffic light

or a stop sign

Current: then you must stop

W

(*)

Switch Switch

The two

d:

b:

If the light

you see

is

then the cross-street light

green

is

not green

switches are connected by the Boolean operator

as a string:

when

b

is

closed,

d

is

NOT,

here portrayed

forced open and vice versa. In computers,

electronic connections replace strings.

20

||

of j

which the proposition

exists for

that follows

is

true.

x 3 y OWNSfojj ("For all x there exists OWNSfoj/*) means "All houses have an owner."* sentence V

So powerful was mathematics,

this idea that for the first

became

it

a

y

Thus, the such that

time in the history of

possible to prove general theorems by simply

applying typographical rules to sets of predefined symbols.

was

It

still

necessary to think in order to decide which rules to apply, but the written

proof required no intermediate reasoning expressed

in natural language.

In that sense, Frege had finally realized a true reasoning calculus.

But

in

language.

another sense, Frege had

still

The word predicate stems from

"to proclaim." Thus, a predicate

proclaiming to be what

its

is

not

fully freed

reason from

the Latin predicare, which

means

nothing but a mnemonic label

user defined, through the use of language,

beforehand. Hence, the meaning of an argument in formal logic entirely in the

mind of

lies

the beholder. Because of this subjective bias,

Frege emphasized that his reasoning calculus worked well only in very restricted

domains. 12 Even so,

many

early efforts at

programming com-

puters for general-purpose intelligent behavior relied in great measure

on

this calculus, steering

begun

to

emerge only

Some of

AI

from which

into an equivocal path

it

has

recently.

the problems inherent to reasoning calculuses were

first

noted by the British mathematician Bertrand Russell. These observations, while disconcerting to those

hoping to

reason-

finally create a true

ing calculus, provided unexpected insights into the nonlogical nature of

thought and consciousness. In June 1902, Russell wrote to Frege to disclose his discovery of a contradiction in the metic.

The

gist

latter's

theory about the fundamental laws of arith-

of Russell's argument was

with two central catalogs: catalog selves,

and catalog B

lists all

assume that the books central catalogs

question

is,

On

A

A

books

that

do

books

Consider

a library

that refer to

them-

not. (For simplicity, let us

in the library are also catalogs, thus

and B catalogs of

which

as follows.

lists all

catalogs, or sets

central catalog should

we

list

of

making

sets.)

The

B? Either choice

*Actually, Frege used a different notation for the universal and existential quantifiers.

notations V and 3, as well as many other mathematical symbols in use today, were introduced by the Italian mathematician Giuseppe Peano in the late nineteenth century, and later finalized by Bertrand Russell and Alfred North Whitehead in Prinapia Math-

The

ematica (1910).

21

ENGINEERING INTELLIGENCE: COMPUTERS AND PROGRAMMING sounds wrong. to

A

We

we

but

itself,

few years

cannot

also

later,

list

cannot

B

list

in catalog

B

overcame the problems

raised

B does

A, because

because then

Russell and his associate Alfred

wrote a three-volume work entitled thors

B

in

it

North Whitehead

Principia Mathematica.

by

classes

not refer

refers to itself.*

In

it,

the au-

of classes through what

they called the "theory of types." Individuals, sets (or classes), and

of classes belong to different

classes

belong to a

belong to a

class,

class

and

logical types.

of

a class to a class

of objects and, in

particular,

classes,

individual can

but a class cannot

cannot belong to

in effect outlawing self-reference in their reasoning

and Whitehead managed to avoid the

An

logical traps

about

By

itself.

sets, Russell

on which

Frege's

work had foundered. The prospects for completely formalizing mathematics appeared excellent, until in 1931 a paper by an unknown twenty- five-year-old Austrian

mathematician brought

optimism to an end. Entitled

this

"On

Formally Undecidable Propositions in Principia Mathematica and Related Systems I," It

13

the article shattered Russell and Whitehead's system.

demonstrated that using their very axioms and notation, one could

state true

theorems that no amount of manipulation would ever prove.

Kurt Godel, the paper's author, went even further and claimed that every consistent logical system would suffer from a similar weakness.

proved

this result

up and acquire

Through

logical

about,

Godel

over, he

a level

a clever

tween

by coaxing the

logical

of meaning

formalism in

its

authors had never foreseen.

encoding scheme establishing a correspondence be-

symbols and the very numbers they were supposed to built logical sentences that referred to themselves.

showed

and consistent

He

Principia to stand

that

one can always encode, in any

logical system, a sentence that

sufficiently

talk

More-

powerful

means: "This sentence

cannot be proved using the system's formalism." The intriguing result

of such a construction

is

that the sentence has to be true!

To

see why,

remember that the logical system in question is assumed consistent, which means that it does not allow us to prove false statements. But first

suppose

now

that the sentence

proved," being false means that

is it

false.

Since

it

can, in fact, be proved. If

proved, our system would not be consistent, since

*The

original (and entirely equivalent) formulation

the set of

all

sets

which

are not

says "I cannot be

members of

it

of Russell's paradox

themselves.

Is

R

it

can be

would enable us

a

is

to

"Consider R,

member of

itself?"

22

41

prove a

false statement.

Thus, the sentence has to be

that

demonstration

really

its

beyond the

is

true,

which means

of our

capabilities

logical

system.

As

the philosopher

J.

R. Lucas pointed out in 1961, 14 an even

surprising fact about this result

of the sentence, but the

truth

is

human

that

logical

system cannot.

We

realize

by reflecting upon the meaning of the sentence and deducing consequences. As

I

pointed out

more

reasoning recognizes the

its

its

truth

obvious

the logical system cannot recog-

earlier,

nize the truth of the sentence since the symbols in the sentence have

meaning

for

no

it.

Alan Turing and His "Machine" The

British

mathematician Alan Turing came to conclusions similar to

Godel's, but in an entirely different manner. For the

approach brought together, into

first

time, Turing's

a reasoning calculus, the theoretical

investigations with the builders of automata's hands-on yearnings to create

life.

Born

in 1912,

strange mixture of blunt,

clumsy

man

Alan Turing remained throughout

his adult life a

boy genius and bemused professor. In appearance with

little

a

care for social graces, Turing disconcerted

auditors with his high, stammering voice and nervous crowing laugh.

He

way of making strange screeching sounds when lost in thought, his mind almost audibly churning away at concepts. He was also renowned for his absent-mindedness. His younger colleague Donald Michie recalls how Turing, fearing a German invasion during the Second World War, tried to provide against the confiscation of his bank ac-

had

a

count: converting his savings into silver bullion, he buried

woods of Buckinghamshire, only

to lose track

Turing had a knack for solving

at a

befuddle engineers for days. time

visit to

A

typical

glance problems that tended to

example occurred during a war-

when he won over AT&T Bell Laborafiguring out how many combinations a

the United States,

tory personnel by instantly

special voice-encoding device provided a Bell Lab's technician to arrive

Mathematicians, computer ing for

in the

it

of the spot forever. 15 But

at.



a result

it

had taken a week for

16

scientists,

and AI researchers revere Tur-

two major ideas he had: the Turing machine and the Turing

A Turing

machine

abstract device

is

not, in fact, a physical

mechanism;

rather,

test.

it is

which enjoys many of the properties of a modern

an

digital

23

ENGINEERING INTELLIGENCE: COMPUTERS AND PROGRAMMING

computer. "Imagine," said Turing in substance, "a reading and writing

head which processes a tape of divided into

which

on

the reading head could be in any states defined

the tape.

or a

1

.

how

the machine

For example,

state

number of

would

tape

would be

Like a typewriter,

lower- or uppercase letters depending

will print

in,

it is

These

The

infinite length."

squares, each containing a

little

on which mode

"states

of mind."

symbol

react to a given

number 73 might correspond

to the

statements:

change to

If the square contains a 0, to the left

on

If the square contains a 1,

square by a

0,

state 32,

and move one square

the tape.

change to

state 57, replace the

and move one square to the

1

in the

right.

number 32 and 57, in turn, might correspond to other statements similar to these. In modern parlance, we would call the sequence of states controlling the head a "program." The initial writing on the tape would be the "data" on which the program acts. Turing showed that these elementary steps could be used to write a States

program performing any sequence of well-defined operations. For example,

if

the l's and 0's

on

the tape represent binary numbers, one could

write programs to extract their square roots, to divide them, or to

combine them

in

any manner imaginable. The idea that any systematic

procedure for operating on numbers could be encoded as a sequence

of elementary, machinelike operations has since become

"Church-Turing

thesis."

known

as the

(The American mathematician Alonzo Church

independently reached conclusions similar to Turing's.) Turing even

demonstrated the existence of one machine that could mimic the operation of any other of his machines: this he called the "universal machine."

One

could aptly describe modern

digital

computers

as practical

embodi-

ments of universal Turing machines. Like Godel, Turing noted his abstract

(in his case, in

regard to the capabilities of

machines) that there exist certain kinds of calculations,

which sometimes appear

trivial

to

humans,

that

no Turing machine can more defeat

ever perform. Although this result could be viewed as one

means of dealing with the world, Turing himself did be reason enough to doubt the possibility of making

for pure logic as a

not believe

computers

it

to

think.

When computers had become a reality in

1950, Turing

24 discussed this question in a celebrated paper entided

"Computing Ma-

chinery and Intelligence":

[T]his [weakness Is this

of Turing machinesl gives us

feeling illusory? It

much importance

is

should be attached to

it.

of superiority. do not think too

a certain feeling

no doubt quite genuine, but

I

We too often give wrong answers

to questions ourselves to be justified in being very pleased at such evidence

on the parts of the machines. Further, our superiority can only on such an occasion in relation to the one machine over which we have scored our petty triumph. There would be no question of tnumphing of

fallibility

be

felt

simultaneously over

all

machines.' 7

Mathematical arguments, claimed Turing in the same paper, are no

He

help in deciding whether a machine can think.

real

argued that the

question could be settled only experimentally, and proposed the following test to this

effect.

you might put

to

Suppose

just as a

it

communicating through tell

a

computer could answer any question

human would.

a terminal with

In

fact,

two hidden

suppose you were

parties

by questioning them which was human and which was

Wouldn't you then have to grant the computer

and couldn't a

computer.

this evasive quality

we

call intelligence?

This procedure, which has the advantage of neady sidestepping the

thorny issue of defining intelligence, has become test."*

come

known

as the

"Turing

Turing firmly believed that thinking machines would one day

about.

He

would "play the

predicted in his paper that, by the year 2000, a machine imitation

game

so well that an average interrogator will

not have more than 70 percent chance of making the right identification after five

minutes of questioning." 18

such an early date; but as off

we

by more than twenty-five

Alan Turing

will

well have been in cracking

years of the war, while

people would

still

agree with

Turing may not have been

years.

down

probably go

about thinking machines. Yet in

may

Few

shall see later,

working

in history for his seminal ideas

his lifetime, his

German at

most important work

intelligence codes. In the early

Bletchley Park, a suburb of London,

Turing designed a machine called the "Bombe," which explored the *In fact, the procedure as Turing defined it was a little more elaborate: the computer was supposed to pretend it was a woman, the other party being a real woman trying to convince you of her identity. If the computer could fool you as often as a man would in

its

position, then

it

passed the

the simplified procedure

I

test.

Nowadays

have described.

the term Turing

test

usually refers to

ENGINEERING INTELLIGENCE: COMPUTERS AND PROGRAMMING possible combinations generated by the

Enigma. The

Bombe was

25

German encoding machine

a special-purpose calculating

machine based

on electromechanical relays. Turing's efforts eventually led to the development of Colossus, a machine many consider the first electronic computer. Colossus, which relied on vacuum tubes rather than on relays, laid German communications bare to British eavesdropping. The British could now direct transatlantic supply ships to steer away from German U-boats, which

made

possible the buildup leading to the

Normandy

landing.

Turing did not reap any reward from the society he had helped so generously during individualistic,

its

time of need. Atheist, homosexual, and fiercely

he did not

fit

into the conformist British society of the

day, nor even into the organizations in

which he worked. After the war,

unable to deal with politics and bureaucracy, Turing Physical Laboratory

where he had participated

chine called the Automatic

left

the National

in the design

Computing Engine,

in

many ways

of a mathe suc-

cessor to Colossus. Prosecuted for his homosexuality, he was convicted in cal

1953 of "gross indecency" and sentenced to a one-year pharmaceutitreatment tantamount to chemical castration.

Turing ended his

life

by eating an apple dipped

On

7 June 1954, Alan

in cyanide.

2 THE FIRST Al PROGRAM: DEFINING THE FIELD

Every aspect of learning or any otherfeature of intelligence can in principle be described that

a machine can be made

to

simulate

— Organizers of Symbol-processing AI

as

we know it

tual field in the years following the

to 1956. First,

Three

critical

AI had

intelligent

the

Dartmouth conference, 1956

today defined

this

processes could be emulated

new

digital

who

in

Al-computer work to

The

is,

neurons.

The

more

efficiendy

by the

did not develop a fascina-

machines either quickly

intelligence or stayed in the parallel field

to develop: that

efforts at replicat-

artificial

approach when they decided that

emerging computer technology. Those

mass had

intellec-

events had to occur during this period.

AI broke away from

tion with the

an

— from 1945

to be tied to the computer. Early investigations about

machines centered on feedback theory and

human thought

critical

itself as

Second World War

ing the workings of the brain in networks of

pioneers of

so precisely

it.

lost interest in artificial

of neural networks. Second,

enough people had

create an intellectual

a

to start to dabble

community

for such ideas.

kernel of this group was formed by Marvin Minsky, John McCarthy,

Allen Newell, Herbert Simon, and their students. Third and most important, these individuals

had to find each

other. This gathering process

THE FIRST

27

PROGRAM: DEFINING THE FIELD

AI

started with the

emergence of two independent, informal groups around

Boston and Pittsburgh, and culminated in the 1956 Dartmouth conference, where the first AI program was presented and discussed.

POSTWAR EFFORTS In the period immediately following the Second World War, the study

of

intelligent

machines blended

disciplines: artificial

cessing in digital computers.

human

The fundamental

brain, feedback systems,

and

this period,

it

fields that later

congealed into distinct

neural networks, control theory, and symbol pro-

and

belonged to the

digital

first

differences

between the

computers were not

clear in

generation of researchers to bring

out the disparities between them.

Norbert Wiener and Feedback Theory One of

the

Americans

first

common

to observe

points between the

mind and engineered devices was the MIT professor of engineering and mathematics Norbert Wiener. The embodiment of the distracted genius (pictorial caricatures of his cigar-toting rotund figure still hang proudly today on the walls of the main hallway of MIT's Building 10), 1

Wiener was MIT's

star

sometimes also the

institute's chief

died in 1964). difficult to

A

performer as all-round

intellectual gadfly,

embarrassment, for

and

thirty years (he

speaker of several languages, he was

known

to be

follow in any of them.

Prior to joining the mathematics department of MIT shortly after the

Second World War, Wiener received formed postdoctoral work

in

Ph.D. from Harvard and per-

his

England, where he managed to displease

such eminent mathematicians as Bertrand Russell and David Hilbert. 2 Seeing himself as too broad an a single field

intellect,

however, to confine himself to

of study, Wiener wandered in what he called the "frontier

areas" between disciplines. While roaming along the borders of engi-

neering and biology, Wiener created the science of cybernetics.

Feedback

is

a

well-known mechanism

in biology.

Warm-bodied

ani-

mals keep themselves within a certain temperature range through biological

feedback mechanisms; predatory animals adjust their move-

28

Al

ments

stat:

catching their prey through scent and visual feedback mech-

in

The feedback system we

anisms.

most

are

familiar with

is

by assessing the actual room temperature, comparing temperature, and then responding

on or

ditioner) either

it

to a desired

the heater (or air con-

whether the existing tempera-

value.

The word feedback

describes

the process returns (feeds back) the result of the control action

compensating mechanism. Cybernetics

(the temperature) to the

was

science of control) ics

— by turning

off according to

below or above the desired

ture lies

how

the thermo-

achieves a constant temperature in any enclosed environment

it

(the

mathematical theory of feedback. 3 Cybernet-

why feedback mechanisms,

explained

teracting

a

complex

especially in

in-

sometimes become unstable. This breakthrough

systems,

allowed Wiener, with

the

of others, to develop procedures

help

used during the Second World

War

to stabilize radar-controlled anti-

aircraft guns.

But more important to our feedback theory

is

story,

Wiener recognized

that central to

the idea of information. In essence, feedback

mech-

anisms are information-processing devices: they receive information

and then make gent behavior

a decision

is

based on

it.

Wiener speculated

that

all intelli-

the consequence of feedback mechanisms; perhaps by

definition, intelligence

is

the

outcome of receiving and processing

infor-

mation.

Nowadays

this

notion appears obvious. Yet

it

was then

a

major

departure from accepted ideas, notably from Sigmund Freud's theory

7

that the

mind

essentially manipulates biological energies,

repress only at the risk of seeing

This paradigm

become gence.

shift

away from energy

the underpinning of

It

them emerge again

all

in

which one can

harmful disguises.

to information processing

subsequent work in

would

artificial intelli-

deeply affected psychology as well, marking the beginning of

the fruitful but uneasy relationship this discipline

would maintain with

information sciences for the rest of the century.

Weiner himself, however, never

really

computers. "I could never get him to

talk

developed

commented Marvin Minsky, another seminal "and so

it is

a strong interest in

about computers very much," figure in early

AI

research,

not surprising that Wiener never made any other significant

contributions to AI." 4 Nonetheless, his cybernetic theory influenced

many

generations of early

AI

researchers.

THE FIRST

A

I

29

PROGRAM: DEFINING THE FIELD

Neural Networks: McCulloch, and Hebb Among

who

the researchers

incorporated cybernetics into their early

who

theories of intelligence were those brain's workings. ual

They planned

neurons with

who

Pitts,

tried to

model the

components. Contrary to

electrical

of the

detail

by simulating individ-

to accomplish this

later researchers,

concentrated on experimental simulations, early neural net workers

how

attempted mathematical analyses of

networks of such neurons

would behave. Warren McCulloch and Walter most

truly colorful figures to

figure

was Donald Hebb, who

work

later

were two of the

Pitts

in this field.

Another

influential

provided more theoretical insight by

considering biological neurons.

Born

into a family of lawyers, doctors, engineers,

Warren McCulloch was

and theologians, 5

destined for the ministry. In the

initially

fall

1917, he entered Haverford College and was soon, according to his

account, called in by the Quaker philosopher Rufus Jones, "

'What

thee going to be?'

is

thee going to do?'

question it,

and

a

I

would

man,

And

like to

that he

And

again

may know

a

his

a

number,

number?' lives.'

fifteen

and spent time

knew only ests,

as Bert.

in a

"When

is

He

that a

smiled and said, 'Friend,

that

Pitts.

The most

home

he ran away from

Pitts read a

book

that

had

office for

"Carnap was amused, because when he

said

what

it

he [emphasis added]

Russell."

meant was

that

just

been

name of

an explana-

something wasn't

was nonsense. So he

newly published book to where young it

he

not only read the book but, upon discovering

clear,

his

age of

man

Bert detected the boy's [mathematical] inter-

tion.

and sure enough,

salient

at the

Chicago park where he met an older

something he found "unclear," went to Carnap's

opened up

one

answer to the second part of that question

he suggested that young

Pitts

is

is

man may know

published by a professor at the University of Chicago by the

Rudolf Carnap."

asked:

"6

with the help of the mathematical prodigy Walter feature of Pitts's childhood

who

don't know.' 'And what

'I

don't know; but there

'I

What is

thee will be busy as long as thee

McCulloch formulated

said,

said,

I

answer:

I

of

own

Pitts

was nonsense. Bert turned out

was pointing,

to be Bertrand

7

In 1943, three years after this encounter,

McCulloch and

Pitts tried

30

Al

to explain the workings of the human brain by coming up with a mechanism by which networks of interconnected cells could perform 8 logical operations. They started out by asking themselves what could be

considered a "least psychic event," and realized that such a fundamental event could be the result of an all-or-nothing impulse by a nerve

Perhaps

it

was

of the single nerve

at the level

cell,

by

its

cell.

release or failure

humans make true/false decisions. The Pitts-McCulloch paper on neural networks relied heavily on the idea of feedback loops (which they called "circles") to reach some of their conclusions. They pointed out that the loop "senses, to brain, to to release an impulse, that

muscles" can result

between

difference state

purposive behavior

in

the muscles reduce the

if

by the senses and

a condition perceived

of the world. Likewise, they defined memory

ing in closed paths of neurons. Every

remembrance was, according

them, the reactivation of a trace of one such signal in fact that a reactivation

tion occurred time.

We

what

order.

of

remember

that conscious decisions about the truth

occur

at a level

much

and thorough mathematical

by transmitting or

cells.

analysis

how

laid the

a

mechanism

McCulloch and it

managed

could perform logical

like the brain

foundation of what

Pitts's

to present

of how interconnected ceDs,

failing to transmit impulses,

thus,

of

higher than that of the single

contribution was important, nonetheless, because

These ideas

to

The

does not tell when the original activawhy our memories are so indefinite as to but not always when they happened, and in

neuron, probably involving millions of brain



closed path.

a trace

events,

Nowadays we know

operations

its

might explain

logical propositions

a valid

a desired

as signals reverberat-

is

today

might compute.

known

as "artificial

neural net theory." Pitts

the

and McCulloch were

also able to

computing powers of both

draw

artificial

striking parallels

between

neural networks and Turing

machines. These comparisons, unfortunately, gave the

false

impression

work like digital computers. They don't, and it took many AI research away from the dead-end path inspired by this draw

that our brains

years to

misconception.

The mention of feedback loops in the papers both of Pitts and is no accident: they and their co-workers knew each other. When a small group of scientists grow interested in a

McCulloch and of Wiener

problem, they often form a club to chat about their favorite subject. In this case, the club

they formed in the late

1

940s was

named

the Teleo-

— THE FIRST

31

PROGRAM: DEFINING THE FIELD

AI

same time their interest in goal-oriented means "end" in Greek), McCulloch's taste for

logical Society, reflecting at the

behavior in nature

pompous

{telos

terminology, and Wiener's considerable ego (the club's

stemming from the

title

of

his seminal

name

paper published in 1943 9 ). The

group enjoyed the lively company of yet another first-rate intellectual John von Neumann, who had emigrated to Princeton before the war. A group like the Teleological Society was started in England. Its

members called themselves the Ratio Club because "they liked the [way 10 Alan the word ratio] combined reasoning, relations and numbers." Turing attended some of the meetings, but did not play a central role in the club's activities. Other members included the philosopher Donald Mackay; Turing's colleague Jack Good; the biologist John Pringle; Albert M. Uttley, whose fertile imagination provided the club's multireferencing name; and the neurosurgeon John Bates. Another member was the brain physiologist Grey Walter, inventor of the first cybernetic "turtle."

A

wheeled

illustration

dome-shaped device navigated fed from an electrical outlet

and improved on by tories into the

The

British

McCulloch

of the power of feedback loops, its

way around

when

its

batteries ran low.

later researchers, the turtle

mobile

field robot.

"visited the British, argued with

banter seems to have

Alan Turing was 12

Activity,"

"A

come out of

less

The reasons

speculate. In

Widely copied

evolved in other labora-

and American groups were aware of each other. Warren

they returned the favor"; 11 but nothing

tan."

this

Walter's laboratory and

impressed and considered McCulloch "a charla-

Logical Calculus for the Ideas

in

than discussion and

these interchanges.

for Turing's aversion are not clear, but

McCulloch and

were equivalent

and delighted them, and

much more

Pitts

had argued that

computing power

Immanent

in

their neural

to Turing machines. 13

one can

Nervous networks

As Seymour

Papert points out in his introduction to a collection of McCulloch's papers, 14 this claim like

numbers. the

was exaggerated: McCulloch-Pitts nets

Turing machines, but correspond to a narrower It

would have been

in character for

class

are indeed

of calculable

Turing to take offense

in

bevue.

In 1949, six years after McCulloch and Pitts had

shown how

neural

networks could compute, the McGill University physiologist Donald O.

Hebb

suggested

how they could learn. 15 He proposed the idea that brain we learn different tasks, and that specific new

connections change as

neural structures account for knowledge. Hebb's ingenious proposal

32

Al

dealt with the conductivity

rons.

He

of synapses, or connections between neu-

postulated that the repeated activation of one neuron by

another through a particular synapse increased

its

conductivity. This

change would make further activations more

likely

formation of tightly connected paths of neurons

in

and induce the

an otherwise loosely

connected structure.

DUO

THE CAMBRIDGE Although Wiener, McCulloch,

Pitts,

and Hebbs belong to the genera-

tion preceding the actual founders of AI, the latter

neural nets and cybernetics. Marvin Minsky and

two key

were educated

in

John McCarthy were

of the new generation.

figures

Marvin Minsky At

the

same time

as

Donald Hebb, an unusual Harvard undergraduate

named Marvin Minsky was independently coming,

in a

roundabout way,

to conclusions similar to his. Minsky's physicist friend says that "it

was "I



was not

or, perhaps,

entirely clear

what

it

Jeremy Bernstein what (Minsky's) major academic field

wasn't." 16

Minsky himself recalled elsewhere:

wandered around the university and walked into people's

laboratories

know anything about the social life of the undergraduates, but I knew when the department teas were, and I'd go and eat cookies and ask the scientists what they did. And and asked them what they

they'd

tell

me." 17 They more than told him: they gave him labs of

own, three of them! tal

work

A nominal physics

student,

him use

a

He

Minsky had attached

animal's claw, the crayfish picked

the claws.

when Minsky

When

hung around

grew

roomful of equipment, where he became an expert

influence of electrodes

it

also

and talked a zoology professor, John Welsh, into

the neurophysiology of crayfish (a small fresh-water lobster).

released

his

Minsky did experimen-

in physical optics in the physics department.

interested in neurology letting

did. I didn't

up

Under

in

the

to individual nerves of the

a pencil,

waved

it

around, and

excited the fibers that inhibited the closing of

he wasn't doing physics or dissecting

crayfish,

Minsky

the psychology laboratory, where he was able to sample a

cross-section of psychology as

it

existed in the late 1940s.

At one end

THE FIRST

of the lab was the behaviorist camp of B.

who

33

PROGRAM: DEFINING THE FIELD

Al

F.

Skinner and his followers,

then held sway over most psychological research in the United

States.

Born

1898 with the publication of the American educator and

in

psychologist

Edward Lee Thorndike's book Animal Intelligence, behavior-

human psychology of

ism was a brutal transposition to

Pavlovian ex-

periments conducted on animals in which one would, for example, flash a bright light at a cat

would

the animal

whenever one fed the animal. After

several days,

response to the flashing light even in the

salivate in

absence of food. Pavlov called the light "stimulus" and the salivation "response." For behaviorists,

all

ply reflexes triggered by a higher

actions, thoughts, or desires

were sim-

The only

difference

form of

stimulus.

between animals and humans was that humans were able to react to

more complex was

sets

of

stimuli, called "situations."

just a device for associating situations

point in examining

it,

as

Given

with responses, there

had been done by the

earlier

mind was no

that the

methodology

that

used introspection to study the mind. The AI pioneer Herbert Simon said later,

nal



word like 'mind' in a psychology jourmouth washed out with soap!" 18 For extremists like mind did not even exist. One could study the act of

"You

couldn't use a

you'd get your

B. F. Skinner,

memory

remembering, but to investigate discipline. Ironically,

when

in the 1940s

itself

transgressed scientific

and 1950s engineers

started

building machines that played checkers, proved mathematical theorems,

and

also contained a device they called a

"memory," the engineers

discussed the "minds" of their machines in as

wanted

to.

much

detail as they

Nevertheless, Minsky liked Skinner very much, and spent

some time helping the psychologist design equipment for his experiments. As I shall show, Skinner's ideas about reinforcement learning also later inspired Minsky to build a neural net machine. Minsky

didn't think

much of

the physiological psychologists at the

other end of the Harvard psychology lab. This group tried to understand little

parts of the nervous systems, such as the sensitivity of the ear,

without relating them to the

rest.

In the middle of the laboratory, however, were young assistant professors who came much closer to Minsky's own way of thinking. Among them was George Miller, who attempted to model the mind through

mathematics.

A

later, in 1956, Miller became famous with the on short-term memory. Entitled "The Magical article shed a critical light on our reasoning pro-

few years

publication of an essay

Number

Seven," this

34

Al

We

19

cesses.

suffer,

claimed Miller, from an

more than seven items of information limitation, Miller

emphasized the active

inability to

of the mind

role

of information: gone was the behaviorist model of

With

association mechanism.

found

"I

me

told

Yet

make

I

As soon

a learning machine?'

let his

grades

fall.

memorable undergraduate

as "

I

didn't allow for a thesis.

saw

that

started to think

I

side,

boost his average, he decided to write

thesis.

For

this

a

he had to switch over to the in the physics

department

That presented no problem: Minsky had

also

sampled the Harvard course curriculum and earned enough

liberally

a

Minsky

music composition courses on the

To

mathematics department, since regulations

math

to study learning.

department meetings and hopping between too many labora-

tories, in addition to taking

he had

as a processor

a purely passive

Minsky was facing more down-to-earth problems. Eating

in 1949, at

mind

works of Warren McCulloch and the

years later. "It had the

could

cookies

in

this thing called the Bulletin of Mathematical Biophysics"

great pioneers of the 1940's. ...

'How

Minsky decided

Miller,

keep

pointing out this

at a time. In

credits to qualify as a

Under the somewhat

proving that

major

in that field.

direction of the mathematician esoteric paper at

with three of

each

its

on

Andrew Gleason, he wrote

the subject of topology.

moment there is on

It

involved

the surface of the earth a square

four corners at the same temperature. Mathematicians

don't judge results by their practicality.

It

was the elegance of Minsky's

demonstration that impressed Gleason. "You are a mathematician," he said

on reading

the thesis, and urged

Minsky

to register for his

Ph.D.

at

the prestigious Princeton mathematics department.

Minsky followed

this advice

and

blissfully

discovered that Princeton

wouldn't cause him any problem with grades. "Once,

my

transcript,"

A's

— many of them

he

said.

"Instead of the usual grades,

in courses

I

mathematician or one wasn't, and

one

actually

it

work with George

Miller,

and

still

Pitts

at

that either

(the

mathe-

one was

a

how much mathemat-

in the study

through his undergradu-

somewhat under

the influence of

Minsky approached another graduate student

with an idea for putting theory into practice.

electronics.

took a look

the grades were

knew." 20

Skinner's behaviorism,

were not

felt

didn't matter

Heavily influenced by McCulloch and ate

I

had never taken. Lefschetz

matics department's director at Princeton)

ics

all

of the brain but

Dean Edmonds's interests new science of

in the relatively

"Instead of studying neural networks in the abstract,"

THE FIRST

Minsky thought did

George

35

PROGRAM: DEFINING THE FIELD

Al

to himself,

Miller,

who

"why not

Edmonds

build one?"

agreed, as

obtained a two-thousand-dollar grant from the

Office of Naval Research.

During the summer of 1951, Minsky and Edmonds returned to Harvard and assembled the

vacuum

first

neural net machine from three hundred

The

tubes and a surplus automatic pilot from a B-24 bomber.

assemblage (which they called the Snare) consisted of a network of forty artificial

neurons that simulated the brain of a

way

rat learning its

through a maze. Each neuron corresponded to a position in the maze and,

when

fired,

it

open

the choices

showed

that the "rat"

knew

itself to

be

at this

point

Other neurons connected to the activated one represented

in the maze.

of these neurons

to the "rat" (for example, to fired

the activated neuron:

go

left

or

right).

Which

depended on the strength of their connections

was the automatic

it

connections. Instead of

moving the

to

pilot's role to adjust these

elevators or ailerons of an airplane,

the automatic pilot turned the knobs that set the strengths of the

connections. When, by chance, the "rat" made a sequence of good moves and found its way out of the maze, the connections corresponding to these moves were strengthened. In this way, the "rat" gradually learned its way through the maze. Two thousand dollars went a long way in 1951, but certainly not as far as allowing custom-designed parts. Minsky and Edmonds had to make do with whatever surplus gear they could scrounge up, and the world's

first artificial

ing elegance.

To

neural net wasn't exactly the epitome of engineer-

As Minsky

recalled to

me:

you had to reward the neurons which had

train [the system],

recently.

So each neuron had

a timing circuit that

say, five

seconds after

The

that

went

a shaft

it

fired.

circuit

rolling in, with a chain drive

potentiometers. So

if

the neuron

had

stay

fired

on

for,

operated a magnetic clutch

to the potentiometer. If you turned

was

would

on

this big

going to

fired three

all

motor, then

forty

of these

seconds ago and you

switched on the motor, then the potentiometer would slowly turn for

two seconds. Since the procedure resembled the stimulus-reward techniques of

behaviorism, Minsky tried talking to Skinner about the machine. psychologist wasn't interested.

It

soon became

learning techniques were not leading

The

clear that Skinnerian

Minsky anywhere: they provided

36

||

no way

tor the

machine

to reason

about what

it

was doing, and thus

formulate a plan.

Returning to Princeton

at the

problems the subject of brains

much

end of the summer, Minsky made these

larger" than the

Harvard machine, Minsky told me. "[They

had] sensors that turned

on intermediate

feedback that could

it

let

ideas are just

sections and different kinds of

do some planning ahead. There was some network can control another. Some of these

how a now being rediscovered

discussion about

which "described

his doctoral dissertation,

as the connectionists start to think

about multimode networks."

who w ondered whether this work was really mathematics, von Neumann, also a member of Minsky's dissertation committee, replied, "If it isn't now it will be someday let's

To

a

doubting department head

r



encourage

it."

21

Minsky obtained

Ph.D.

his

neural network with

in 1954,

convinced that a sufficiendy large

enough memory loops

require thousands or millions of neurons. that large a network,

would

to actually reason

He knew

he couldn't build

and looked for other ways to get machines to

Dean Edmonds, who had never been impressed with the Harvard machine, went on to become a professor of physics at Boston think.

University.

Meanwhile, Andrew Gleason wanted Minsky back

at

On the

Harvard.

recommendations of von Neumann, Norbert Wiener, and Claude Shannon, Gleason had his former protege accepted as a junior fellow. As

such Minsky could proceed with his research in complete freedom for three years. His only obligation

was to dine with the other junior fellows

on Monday evenings. Since Minsky was then reconsidering his interest in

artificial

networks, he temporarily directed his mental energies into the optics and, in 1955, invented

and patented the

first

neural

field

of

"confocal scan-

ning microscope." 22 This device imaged interconnections of neurons in

much

greater detail than the Golgi staining process (see chapter

11). Strangely, the

instrument was ignored by microscope manufac-

turers until the late 1980s, but there are

thousand of them, priced Unfortunately for him,

at

this

now on

commercial success happened long

the patent expired. It

was around 1955,

as

the order of one

around SI 00,000 each, Minsky told me.

Minsky

recalled

it:

after

THE FIRST that

met

I

young man named Ray Solomonoff who was working on

a

an abstract theory of deductive inference. learning machine

decided in

this

.

.

.

built a piece

With

make

I

.

.

He had worked on

a

was so impressed

I

I

[this

of hardware and hoped

new] approach, you

what kind of inferences you wanted

would

.

was pretty formal.*

that

was much more productive than the neural net system,

which you

right thing.

37

PROGRAM: DEFINING THE FIELD

Al

a

to

tried to

it

would do the

make

theories of

make, and then asked,

machine do exactly that?"

It

was

"How

a different line

of

thought.

Following this encounter, it gradually dawned on Minsky that there was a difference

what

it

between understanding how the brain is

does.

As

starting to offer a

it

turned out,

way

digital

built

computers were

to explore this last path. It

and finding out just

about then

was becoming possible

for scientists to describe to a computer, in a symbolic way,

what they

thought the mind did, and have the machine behave in exacdy

this

manner.

John McCarthy John McCarthy 23 was born to a Lithuanian Jewish mother and an Irish Catholic father who took up Marxism and fought for it as a union leader. Opposition to conventional ideas ran deep in the McCarthy family: John was thrown out of Cal Tech for refusing to attend physical education classes. Later, the U.S.

Army

communist. Patrick also

dismissed his brother Patrick for being a

lost a post office job for refusing to sign a

loyalty oath. In 1945, though, after John's

didn't have any choice but to put

brush with Cal Tech, the army

up with

his

own communist

bias

because he had already been drafted. Fortunately for John McCarthy, the

war promptly ended; and

Tech.

He

thereafter,

as a veteran, John

was able

to return to Cal

obtained his bachelor degree in mathematics in 1948. Shortly

he moved to Princeton for graduate work. Having read about

von Neumann's research on

finite

automata, he began to explore their

possibilities as intelligent agents. After graduation,

he spent

a

summer

with Claude Shannon editing a collection of papers on the subject. 24 In *SolomonofFs theory,

later independently rediscovered by other researchers, today "algorithmic probability theory."

is

called

38 his

||

own

contribution to the volume, McCarthy discussed the possibility

of making

a

Turing machine behave intelligendy. This

artificial intelligence

two

first effort

at

McCarthv

did not turn out very well, and taught

lessons.

Turing machines did not provide the right medium for estab-

First,

how machines

of

lishing a theory

problem was one of

could imitate

human

behavior.

The

small changes in machine structure

sensitivity:

brought about enormous changes

and vice

in behavior,

versa. Intuitively

small changes in humanlike behavior required verv large changes in

machine structure

name "automata

to account for them. Second, the

studies" wasn't right for the kind of investigations

McCarthv had

in

mind: most of the papers he and Shannon received had nothing to do with the reproduction of

human

intelligence.

A

catchier appellation

was

required.

A summer on

ests

ers

digital

spent working

at

IBM

in

1955 focused McCarthy's

inter-

computers. Hands-on work convinced him that comput-

provided the tool for actually building

artificial intelligences,

Turing machines and automata theory onlv allowed him to studv

while intelli-

gence in the abstract. Since then, McCarthy has never stopped looking for ways to

mind

into computers. Fellow researchers

in the history

His

of AI and agree that McCarthy's

ability to dive into the

tion

is

legendary.

tends to

now

inflict

grant

own mind

intellectual standards,

is

apparendy

say.

his dissatisfaction

isn't a

explaining

some of them he John

to

product

up when

Those standards account list

of publica-

hard for John to communicate well

"It's

The ideas are make some compromise in

graduate student in his areas.

there, but the willingness or the ability to

you're talking with

them

with other people's work. His former stu-

dent Hans Moravec told me: with anybody that

a

to shut

for McCarthy's successes, but also for his relatively short

and

also special.

the long silences he

is

which command him

he can't think of anything worthwhile to

tions

is

his partners in a conversation, leaving

wonder how they may have offended him. This of his high

embody

a special place

depths of a problem through sheer concentra-

Another McCarthy trademark

upon

him

it

thinks should be obvious are missing. If takes

skill

to learn to

communicate with

him" Yet McCarthv

is

far

from being a

loner. In fact, his social

heavilv influences his opinions and

way of

inconvenience in the early 1950s,

when he

life.

environment

That could have

led to

followed his parents and

THE FIRST

Al

39

PROGRAM: DEFINING THE FIELD

brother into communist militancy. Fortunately, his relative obscurity as an assistant professor protected

him from

McCarthy

the wrath of another

then witch-hunting from Washington. In the 1960s, John McCarthy grew

donned

his hair,

a

headband, and joined the counterculture movement.

socially conscious,

Still

he became one of the

first

crusaders against the

misuse of information possible through government and corporate com-

became

puter data banks. Years before the Privacy Act of 1974

McCarthy's proposal, which he called the

"Bill

September 1966 issue of Scientific American. With the 1970s,

McCarthy took

his

law,

of Rights," appeared in the his

second wife, Vera,

cue from the "me" generation:

his

in

"own

thing" was to court danger through parachute jumping and alpinism.

Unfortunately, this nist in the reach

way of life was more dangerous than being a commu-

of Joseph McCarthy: Vera died in

a climbing accident

during an all-women ascent of Annapurna. In the 1980s, John McCarthy

donned his

three-piece suits and voiced conservative opinions.

He

decided

1966 proposals for computer privacy were a mistake, since merely

possessing information causes no harm.

should concern

About

abouts.

itself

The

law,

McCarthy now

with the usage of information, not with

research financing, McCarthy's philosophy

opposite of equality: he summarizes his position as higher," better.

its

thinks,

where-

is

the direct

"Make

the peaks

meaning that one should make the best research

institutions

still

And the former communist opposed Edward Fredkin's idea of an

international

AI

laboratory: his reason: the Soviet

unfair advantage of it.

During the 1955-56 academic fellow,

and McCarthy taught

Hampshire. The idea of

"They had

gleam in

a

Union would

take an

26

at

year,

Marvin Minsky was

a

Harvard

nearby Dartmouth College, in

intelligent

New

machines fascinated them both:

their eye!" recalls their colleague

Herbert Simon. 27

They were becoming aware of the work of other researchers in the field and wished to bring them together. For this, they enlisted the help of two senior

One was first

scientists

who were

also interested in the subject.

Nathaniel Rochester of IBM, designer of the

IBM

general-purpose, mass-produced electronic computer.

had met him

in

connection with IBM's

the help of three colleagues keepsie,

New York,

from the

gift

701, the

McCarthy

of a computer to MIT. With

IBM research laboratory in Pough-

Rochester was then simulating neural networks on

IBM's new 704 computer. 28 More

precisely,

he was programming the

machine to solve the numerical equations describing

a large-scale neural

network. This use of the computer was quite different from the symbol

40

Al

processing, "top-down" approach in which Minsky and McCarthy were Intrigued, Rochester

interested.

hoped

approach might help computers exhibit

which was

main

his

The other

Minsky and McCarthy's

that

originality in

problem

solving,

interest.

was Claude Shannon. Both Minsky and him during the summer of 1953 at Bell Labs.

helpful elder

McCarthy had worked

for

Co-editing the book on automata theory with intelligence had furthered

Shannon's interest

And

so

was

it

younger man's work.

in the

of Rochester and Shannon,

that with the backing

McCarthy and Minsky persuaded the $7,500 cost of a

the Rockefeller Foundation to cover

summer workshop on

two-month meeting, held

in

thinking machines.

The

1956 on the Dartmouth campus under

McCarthy's auspices, brought together the few researchers then active in the field.

29

In addition to the four organizers

Rochester

showed Minsky

— of

up.

this

other

six

participants

One was Ray SolomonofT of MIT, who had

among would-be AI intellectual

a tendency that

was already becoming

researchers: that of giving the

problems to machines

intelligence to the world.

converted

of view. During the conference,

to the symbol-processing point

SolomonofT pleaded against and

— McCarthy, Minsky, Shannon, and

Dartmouth conference,

in

clear

most complicated

an effort to demonstrate their

SolomonofT proposed

that using easier prob-

lems would simplify the analysis of the mental processes involved. The point was well taken for yet another reason probably

SolomonofT nizing a

at the time: the

human

computer. The

of

problems that we find

face) are often the thorniest

ability to tackle intellectual

unknown

to

easiest (like recog-

ones to program into a

problems

is

a

poor

definition

intelligence.

Also present was Oliver Selfridge,

Wiener who had proofread the Cybernetics in 1948.

galleys

a

former assistant of Norbert

of the

In 1956, Selfridge was

at

first

MIT's Lincoln Laboratory

working on pattern recognition by machine.

programming computers translate his

edition of the latter's

He had

already started

to recognize letters of the alphabet

Morse code. 30 Shortly

and to

after the conference, Selfridge delivered

most important contribution

to the field



a forerunner

of expert

systems called Pandemonium. 31 Instead of a dignified sequence of state-

ments, Selfridge believed an AI program should look capital

a

of Hell: a screaming chorus of demons,

master decision-making demon. Each

all

like Milton's

yelling their wishes to

demon was

a short sequence

of

THE FIRST

Al

41

PROGRAM: DEFINING THE FIELD

statements looking for a particular configuration in the data (say, a vertical stroke in a printed character, or a specific

The master demon made

conjunction of symp-

its

decision by integrating

the lower-level decisions of each of the small

demons. For technical

toms

in a patient).

reasons, the idea did not catch

accounted for

Two

minor

much of

on

for fifteen years but,

participants at the conference

Arthur Samuel.

A

when

it

did,

the success of expert systems.

were Trenchard More and

graduate student at Princeton,

More was

writing a

on ways of proving theorems using a technique called natural deduction. Samuel, then at IBM, was investigating how computers could be made to learn by teaching them to play checkers. His efforts had a thesis

significant

impact

later.

THE FIRST Al two

1 he last

PROGRAM

participants in the

Dartmouth conference, Herbert Simon

and Allen Newell, had an immediate impact on AFs debut. Together, they had already written a computer program that, they claimed,

showed

computers could think.

that

Herbert Simon Herbert Simon was then explain his training,

In

1

the

he had started

948, he had

body

made

and

his

interest in

background did not

AL 32 A

his career researching

year,

which established

a

book

his reputation as

organizations.

Soon

by

municipal administrations.

a brief venture into civil service as a

he published

readily

political scientist

that administered the Marshall Plan after the

War. That same

human

thirty-six,

competence or even

member of

Second World

entitled Administrative Behavior,

an expert in the functioning of

thereafter,

Simon helped found Carnegie

Tech's Graduate School of Industrial Administration. (Carnegie Tech

now

is

Carnegie Mellon University.) In 1956, Simon was a professor of

industrial administration there.

Many would cracies



lies at

claim that Simon's specialty in those years

— bureau-

the exact opposite of intelligence. Yet his groundbreak-

him to guess at several common points in the workings of both, and more recent findings by others have confirmed his intuition.

ing

work

led

42

Al

He had

long been fascinated with

how

people make decisions; and his

conclusions, though contrary to conventional economic theory, offered

an open window into the workings of the

human mind.

Before Simon, economists believed (and to a certain extent

do) that

still

companies and even individuals making economic decisions behaved with perfect rationality and omniscience. Before reaching an investment or policy decision, a alternatives

The same kind of refrigerator.

company was assumed

to consider

and choose the one that brought about the

was expected of

care

a

all

possible

largest benefits.

consumer shopping

for a

These assumptions enormously simplified the mathematical

of economic systems because they assumed economic agents

analysis

were always trying to maximize or minimize some function, such

as profit

or cost. Since well-known mathematical tools existed for finding a function's trough or crest,

economic behavior then became

problem. But Simon pointed out, in

do not always work

these basic assumptions First,

no one looks

his theory

at all

an appliance, most people

of the

will

of Bounded

in practice.

shopping for

decide roughly what they are willing to

pay, look at a couple of models,

and pick the one that comes closest to

Companies do the same: year-end budgeting

their requirements.

why

33

When

alternatives.

a tractable

Rationality,

exer-

few simu-

cises typically exhaust the executives involved after only a

lation runs.

Simon

realized that alternatives

a golden plate, and that there

is

do not come

to a decision

maker on

a cost associated with finding

and

evaluating each possibility. For this reason, a decision process really consists in a search through a finite better.

number of

options, the fewer the

Rather than optimizing some function of

all

options, like the

associated profit, the bureaucrat in each of us picks the that

meets a pre-set acceptance

criterion.

Simon

first

choice

called this behavior

"satisficing."

The inability

to

overcome the cost of searching through many

tives limits us as decision will

confirm, this weakness does not necessarily

in

our minds. Simon

attributes

of a mental

both people and organizations have

difficulty

coming up

with original solutions to problems. In

shows

lie

more of the

discovered another weakness that has restriction:

alterna-

makers. As any sore-footed appliance shopper

itself in the

In recent years, a

organizations, this inaptitude

existence of the rule book, or

more

name of "corporate

management manual.

abstract manifestation of

culture."

it

has received the

Such observations led Simon to speculate

THE FIRST that the

43

PROGRAM: DEFINING THE FIELD

Al

mind mostly functions by applying approximate or cookbook became the basis for "heuristic," or

solutions to problems. This idea

"rule-bound," programming.

Simon had noted

Finally,

that

members of organizations

identify with

subgoals rather than with global aims. For example, a company's advertising

department

strives to

produce the

flashiest

whether or not they increase the company's

profits.

ad campaigns

it

can,

Only supervision by

top management can reconcile the two goals of profits and glamorous

Thus, organizations in

advertising.

growth or

survival)

effect achieve global goals (here,

by breaking them up into smaller aims

(like advertis-

ing and profits), which different departments pursue in a coordinated

manner. This "goal, subgoal" strategy

is

now

a key concept

In retrospect, Simon's transition from economics to

of

AL

artificial intelli-

gence appears almost natural. As he told me: "I was always interested

of intelligence.

in the nature

intelligent-like

Any word of some mechanism that behaved

was something

I

pricked up

my

ears at."

Allen Newell Simon

Dartmouth conference with his younger colleague, The son of a radiology professor at Stanford University, Newell had grown up in San Francisco. He graduated in physics from arrived at the

Allen Newell.

Stanford, where he took courses from the mathematician Artificial intelligence

for the rules

owes

to Polya the

of thumb people apply

word

heuristic,

George Polya.

which he coined

in everyday reasoning. In 1945,

Polya showed the problem- solving power of heuristics in an influential

book

called

How

to Solve It.

34

After Stanford, Newell spent a year in Princeton's graduate school of

mathematics, but did not find

had (he entered Princeton

Minsky by

a couple

of

it

as congenial as

at the

same time

as

Minsky and McCarthy McCarthy, but missed

Contrary to them, Newell decided he

years).

wasn't a mathematician, and dropped out of graduate school.

He prefer-

work where he would have concrete problems to solve. It was then 1950, and the RAND Corporation of Santa Monica offered bright young

red

scientists

with a practical bent the opportunity to prove themselves.

Newell was

set to

work on

a project

aimed

at

modeling a regional

air-defense center. Part of the activity consisted in producing aerial

through a computer-driven printer. Simon, schedule to consult for

RAND,

happened

who found

maps

time in his hectic

to see the printer in activity.

44

Al

To anyone familiar with the drawing programs available on today's home computers, this would have been a perfecdy trivial sight. But in the early 1950s,

Simon found

it

an eye-opener: the dots and characters

making up the maps weren't numbers. He saw them computer was manipulating them! From there ers could just as well simulate

Simon made

leap, but

it

thought

still

without flinching.

started holding informal discussions

on

as symbols,

to deciding that

and

required a major conceptual

From

then on, he and Newell

the subject.

For Newell, however, the lightning bolt of conversion occurred

when

1954,

recognition ideas

Oliver Selfridge visited

work

come about as "It

that

all

The

a result

happened

in

First Al

RAND

in

to describe the pattern-

would soon lead him

hooked Newell with

a

comput-

Pandemonium. These

to

the realization that a

complex process could

of the interactions of many simpler subprocesses.

one afternoon," he

says.

35

Program

With the help of J. C. Shaw, an actuary turned computer programmer at

RAND, Newell and Simon started in the fall of 1955 the development

of what

now

is

Theorist.

considered the

Simon

recalled

it

for

intelligence program: Logic

first artificial

me

as follows:

That autumn, we had considered three tasks for the program. Our original intention

was

to start out with chess,

considered doing geometry.

We

and then we also

thought, probably wrongly, that in

both of these there were perceptual problems, important

performed the logic, for

[Russell

would be hard

no deeper reason than

and Whitehead's]

for an efficient

humans, by

was

tasks, that

that

Principia at

to deal with.

selective heuristics,

home.

found the

humans to

had the two volumes of

I

.

.

.

We

way of proving theorems.

certainly in the center

as

So we went

We were not looking at

how

next.

That

were looking

right thing to

of my mind: what heuristic

.

do .

.

would kick

out the right theorem to use rather than searching forever.

As Simon, Newell, and Shaw had reduced

realized,

theorem proving can be

to a selective search. It helps to represent the search graphically

as a treelike structure, called the "search tree."

The

starting point, or root

is the initial hypothesis, on which the rules of logic allow a certain number of elementary manipulations, which yield slightly modified ver-

node,

THE FIRST

Al

45

PROGRAM: DEFINING THE FIELD

sions of the hypothesis.

Each possible manipulation corresponds

to a

branch out of the root node. Applying other manipulations to each of these results gives the next generation of branches, and so on.

where down the manipulations,

tree, after the application

lies

Some-

of an unknown number of

the desired conclusion: the

problem

is

to find a path

leading to this result.

To

find this path, Logic Theorist explored the tree in a goal-oriented

manner. Starting

at the

root node,

applied appropriate rules of thumb,

it

or heuristics. These rules allowed

to

it

among

select,

branches leaving the node, the one that was most

the goal. Following this branch, Logic Theorist reached a

which

it

possible

toward

new node

at

applied the heuristics again. Searching, goal-oriented behavior,

rule-bound decisions rationality theory

Programming

— most of

— were

before coding

it

AL

any computer, was laborious in 1955; and

Allen [Newell] and

I

session:

wrote out the

(subroutines) in English

program Simon de-

easier to hand-simulate the

it

into the machine. In his autobiography,

one such simulation

bounded-

the basic ideas of Simon's

thus carried over to

a computer,

the trio of researchers found

scribes

all

likely to lead

components of the program

rules for the

on index

cards,

and

also

contents of the memories (the axioms of logic). Industrial Administration] building

on

At

made up

cards for the

the [Graduate School of

a dark winter evening in January

we assembled my wife and three children together with some graduate students. To each member of the group, we gave one of the cards, so that 1956,

each person became, in puter program



effect, a

component of the [Logic Theorist] comperformed some special function, or a

a subroutine that

component of its memory.

It

was the

task of each participant to execute his

or her subroutine, or to provide the contents of his or her memory, whenever called by the routine at the next level above that was then in control.

So we were able to simulate the behavior of [Logic Theorist] with a

computer constructed of human components. Here was nature imitating imitating nature.

.

.

.

Our

children were then nine, eleven, and thirteen.

art

The

occasion remains vivid in their memories. 36

The

actual implementation

of Logic Theorist on a computer

at

RAND did not occur before the summer of 1956. Well before that, the hand simulations demonstrated the program's soundness satisfaction. class

He

thus proceeded to

of the new year

at

make an announcement

to Simon's

to his first

Carnegie Mellon: the future Nobel laureate

46

vi

boasted of nothing

less

than inventing an intelligent program over the

Christmas vacation. Simon

where he reports

is

no more abashed

this success as follows:

autobiography,

in his

"[W]e invented

computer

a

program capable of thinking non-numerically, and thereby solved the venerable mind/body problem, explaining

how

a

system composed of

matter can have the properties of mind." 37 Boastful as he sounds,

whom

Daniel Dennett, to that Logic Theorist

by

Simon may have I

The philosopher

a point.

read this quotation, scoffed

itself solves

the

at the

thought

mind-body problem. Dennett,

however, went on to argue that AI as a whole may very well hold the key to

Logic Theorist in any case kicked off an unending

this mystery.

debate about

its

first

fifty-two

As

philosophical implications.

point, Logic Theorist

was eventually able

theorems

in chapter 2

confirm Simon's

if to

of Russell and Whitehead's

Mathematica. Logic Theorist's proof for

of the

to prove thirty-eight

one theorem (number

Principia

2.85)

was

even more elegant than the one derived by Russell and Whitehead:

Simon In

delighted Russell by informing

deriving this

him of

another way: they had not explicidy instructed the structure of the that

programs can

Theorem

this success.

proof, the program had surprised

authors in

to find the proof. Yet,

do so anyway, thereby showing times do more than their programmers tell them.

program caused at

it

38

its

it

to

2.85 also provided an amusing footnote to the history of AI:

Newell and Simon submitted the new proof for publication journal of Symbolic Logic, listing the

program

missing the implications, the editor turned

down

the paper

it was no accomplishment outmoded system of Principia. 39

prove

a

grounds that

"Beads"

of

to

of

a

it

on

theorem

a

in the

long time

had to await the development by Newell, Simon, and Shaw

computer-programming language with enough power and

bility.

the

Memory

The computer implementation of Logic Theorist took because

to the

as a co-author. Entirely

flexi-

This language was called IPL (for Information Processing Lan-

guage) and incorporated another invention of the

more important than Logic

Theorist: the

trio that is

list-processing

perhaps

technique for

programming.

IPL which

differed

IBM

from other

released a

first

high-level languages like

FORTRAN,

of

version shortly before Newell, Simon, and

THE FIRST

Al

Shaw developed IPL. Short was aimed

computing

First,

you have pro-

algebraic formulas. If

BASIC

language,

FORTRAN.

with

Shaw were unhappy with languages

Newell, Simon, and

ing,

eased the description of numeri-

It

common

in

FORTRAN

FORmula TRANslation,

microcomputer, you may have used the

a

which has many features

TRAN

for

and engineers.

at scientists

cal operations, like

grammed

47

PROGRAM: DEFINING THE FIELD

like

FOR-

because they didn't model two important features of the mind.

they assumed that thought consists in constantly creating, chang-

and destroying interacting symbol

TRAN

and BASIC cannot do

numbers or symbols used symbols

Languages

require that

until they

effect reserves a region

come

into use.

Having

symbol

will do,

structures. Say,

is

we

give

impossible to a

program could do

this,

ahead of time for

its

of

in a

of memory

in

advance

and prevents the program from creating new it

information about individual items

named JOHN, JIM, and MARY. Generating a new

PEOPLE

FOR-

arrays

to include this

know

statement in the program forces the programmer to

what the program

like

all

program be defined beforehand

in a

program statement. This statement in to store these

structures.

They

that.

FORTRAN

however, because

symbol

Simon, and Shaw wanted to model was human memory. In our minds, each idea

them

program.

The second

called

An IPL

memory space

feature Newell,

the associative character of

or remembrance can lead to

and these

it,

array for

BASIC

didn't reserve

it

structures.

other symbols that are linked to

or

can be acquired

links

through learning.

The

trio

was able to incorporate into

to associate

technique, which Herbert

The

a

computer language the

basic idea

is that,

Simon has described

whenever

a piece

(associated) piece like a

of information. In

as follows:

of information

additional information should be stored with

organized

ability

and modify symbol structures through the list-processing

this

it

way

telling

is

stored in

where

the entire

memory,

to find the next

memory

could be

long string of beads, but with the individual beads of the

string stored in arbitrary locations.

"Nextness" was not determined by

physical propinquity but by an address, or pointer, stored with each item,

showing where the associated item was to a string or omitted

from

located.

a string simply

Then

a

bead could be added

by changing

a pair

of addresses,

without disturbing the rest of the memory. 40

Simon told me that he and his colleagues took their cue from early drum machines, the ancestors of today's hard-disk memory-storage

48

||

For technical reasons

devices.

machines already used

Simon discussed

who

McCarthy, chapter

a technique similar to

this

it

in his

modeling the mind, drum list

processing. Newell and

Dartmouth with John own AI language called LISP (see

extensively

idea

used

later

irrelevant to

at

3).

The Dartmouth Conference of 1956 At Dartmouth, Newell and Simon were, working AI program,

remembers them

as

being a

little

all.

it

Neither AI [Newell] nor

under such circumstances." That

moments

state

the following September,

workshop

at a

meeting of the

Minsky

this reason,

standoffish during the conference.

"We were

Herbert Simon confirmed to me:

about

as the only participants with a

ahead of the others. For

far

I

are

of

when

Institute

probably

known

for

affairs also

the time

gave

came

arrogant

fairly

much modesty rise to tense

on

to report

of Radio Engineers held

at

the

MIT.

Simon

Since they were the only ones with concrete results, Newell and

challenged McCarthy's right to report alone. They finally settled by talks: McCarthy summarized the meeting in Simon expounded on Logic Theorist. 41 The organizers of the Dartmouth conference had hoped

having two separate

general,

while Newell and

emergence of

where

it

a

common

was going. As

feeling

on both where

the discipline

a starting point for discussion, they

for the

was and

had proposed

the following statement: "Even* aspect of learning or any other feature

of intelligence can

in principle

be so precisely described that

a

machine

made to simulate it" This belief has remained the cornerstone of most AI work until today. It later became known as the "physical symbol system hypothesis." The basic idea: our minds do not have can be

direct access to the world.

tation

of

it,

We can

which corresponds

operate only on an internal represen-

to a collection

of symbol structures.

These structures can take the form of any physical

pattern.

They can

consist of arrays of electronic switches inside a digital computer, or

meshes of (brain or

firing

neurons

in a biological brain.

An

intelligent

into other constructions.

Thought

consists

of expanding symbol

tures, breaking them up and reforming them, destroying

creating

new

svmbois. it,

system

computer) can operate on these structures to transform them

ones. Intelligence

It exists in a

transcends

it,

is

struc-

some and

thus nothing but the ability to process

realm different from the hardware that supports

and can take different physical forms.

THE FIRST

49

PROGRAM: DEFINING THE FIELD

Al

The Dartmouth conference was deeply disappointed

many ways

in

inconclusive and

John McCarthy. For one periods of time, which precluded regu-

principal organizer,

its

came for different The problem was perhaps that not all participants agreed on the conference format. Herbert Simon recalled to me that "they were going to have a kind of floating crap game all summer with people sitting thing, people

meetings.

lar

with each other, thinking and so on.

And

[Newell and

programming the Logic Theorist, so we agreed we'd spend at

were busy

I]

[only] a

week

Dartmouth."

No

consensus emerged on what the

and most of the participants

Simon described They didn't want them:

we

we

it

to me: "It

to hear

was or where

field

later persisted in their

from

was going us,

already

had done the

first

much

'Not Invented Here' sign

is

it

was going,

approaches.

and we sure didn't want to hear from it

was

ironic because

example of what they were

attention to

But

it.

that's

after;

recalls that

versely induced a false sense of achievement. 42 Contrary to participants believed, their understanding of the theories

elicit

half a

still

quite incomplete. Further,

the worldwide interest they

dozen people

remained

AI

it

per-

what the

of symbolic

research did not

had expected. "Dartmouth got only

active that weren't before,"

a very small

and

not unusual. The

up almost everywhere, you know."

There was, nevertheless, enthusiasm; but Minsky

manipulation was

As

off into different directions.

had something to show them! ... In a way,

second, they didn't pay

own

group for quite

a

Simon

few years

told

me. "[We]

Minsky

after that,"

confirmed to me, "because most people thought AI was impossible. [In a sense] that

was the pleasant part of

have to worry about publishing

Yet the conference the

is

it

the

it:

you got an

if

idea,

you didn't

same week!"

generally recognized as the official birth date of

new science of artificial intelligence. One reason is perhaps

participants in the meeting crystallize the

Simon and

group, gave a sense that

say,

that

had never met each other before. got an idea, I'd

if I

'What do you think of this?'

.

.

.

call

Looking back,

most

"It did

Herb was

that

the start of the community,"

Marvin Minsky told me. Dartmouth indeed defined the AI establishment: for almost two decades afterward, all significant AI advances were made by the original group members or their students. We can surmise that if the brilliance of the conference participants accounts for

much of this

then the preference of agencies to fund recognized ligible factor either.

state

elites is

of affairs,

not a neg-

50

Al

The other AI was his

claim of the Dartmouth conference for being the cradle of

the christening of the

new

discipline.

McCarthy, remembering

disappointment with the automata-theory papers edited with Shan-

non, was looking for an accurate and catchy name. Overcoming the resistance of

some

participants (Samuel felt that "artificial"

phony, and Newell and Simon persisted

sounded

work "complex information processing" for years afterward), McCarthy persuaded the majority to go for "artificial intelligence." He lays no claim to having coined the phrase, and admits it may have been used casually beforehand. Yet nobody denies him the achievement of getting it widely accepted. this

To

label a discipline

is

in calling their

to define

its

boundaries and identity:

accomplishment belongs to John McCarthy.

3 THE

DAWN

OF THE

GOLDEN YEARS: 1956-63

After Dartmouth,

AI, for better or for worse, was

intellectual inquiry. In

many ways

it

now

a field

was no more unified than

it

of

had

been before 1956 but, perhaps because of the continuing exchange of

Dartmouth, AI

ideas initiated at It is

probably not

started progressing in leaps

much of an

and bounds.

exaggeration to suggest that the later

advances made in AI consist largely of elaborations and implementations

of ideas

first

formulated in the decade following Dartmouth.

During those years the main centers of AI research were Carnegie Mellon (then Carnegie Tech), a lesser extent, Stanford

AI

MIT

and

its

Lincoln Laboratory, and, to

and IBM.

researchers centered their

work around two main themes.

First,

they wanted to limit the breadth of searches in trial-and-error problems; the Logic Theorist,

were

results

of

Geometry Theorem Prover, and SAINT programs Next, they were hoping and trying to make

this effort.

computers learn by themselves; chess, checkers,

their attempts in this direction

and pattern recognition programs.

were the

52

A

MODELING HUMAN COGNITION

AT

CARNEGIE TECH As proud some new

Newell and Simon were of

as

humans

solving behavior in

way Logic Theorist all

that disturbing.

Logic Theorist program,

their



coming out of psychology

research

did.

— suggested

Of course,

that

in 1954 on problemhumans do not reason the

signing a flying machine, they

first

should not have been

in itself this

For humans to make the

final

breakthrough in de-

had to accept that such machines do

not necessarily have to be modeled on the way birds thinking machine have to think the if

humans can

mouth

why

fly.

Why

should a

think? But then again,

ignore totally the design of the only success-

system around?

ful intelligent

And

reason,

way humans

so contrary to the goals of

most other

conference, Alan Newell and

participants at the Dart-

Herb Simon's attempts

at

program

design soon shifted away from trying to exploit the capabilities of

computers to trying

to simulate the

human

So

cognitive processes.

productive would this approach turn out to be that they never again deviated from

What

set

it.

them on

had presented

this

path was some ground-breaking research by

Moore and

psychologists O. K.

test subjects

S.

B. Anderson.

1

Moore and Anderson

with a series of puzzles and logical problems

of the kind Logic Theorist solved, and had asked

their subjects to "think

aloud" while working on the problems. Following Moore's and Anderson's lead, Newell and

Simon

carried out similar experiments that

proved "fabulously interesting" to the Carnegie Tech researchers. Simon

remembered

We

it

for

me

as follows:

started looking at this

human

data,

and asking "Are these people

behaving like the Logic Theorist?" The answer was no. Then we asked,

"How

are they behaving?"

And we

extracted the ideas for our [next

program] General Problem Solver right out of human protocols.

What

separates

GPS

from Logic Theorist and

mention of a

The whole

[particular] task.

a completely task

task-specific

a set

is

independent manner.

structure

You

just

.

.

.

GPS we learned

that with

of heuristics which had

to extract out an organization

.

.

.

in

it

no

was encoded

had to plug

in

in the

components, so to speak, into the slots, and it worked! The

53

THE DAWN OF THE GOLDEN YEARS: 1956-63

word general refers

we had

specifically to the fact that

segregated the

2 task-dependent and the task-independent parts.

The

first

run of Newell and Simon's

then they had given a

name

GPS

occurred in 1957. 3 By

method: "means-

to this task-independent

ends analysis." In a sense, means-ends analysis

back principle carried to a higher

level

is

Wiener's feed-

just

of abstraction. Like any good

feedback mechanism, means-ends analysis worked by detecting ences between a desired goal and the actual state of

Where means-ends

reducing these differences.

concept was in

this basic

variations react

— not

when

its

just one, as a

analysis

ability to react to a

differ-

and then

affairs,

improved upon

wide spectrum of

thermostat might, but

much

humans

as

they are given a variety of slightly different problems to

solve.

Consider, for example,

how you would program GPS

to solve the

problem of the monkey faced with the perennial too-high banana. The animal, alone in a

room

dangling out of its reach.

containing a single chair, Is

the

monkey

(or

GPS)

grab a banana

tries to

clever

enough

to

move

the chair to the banana and climb up?

To

explain the problem to

spatial coordinates,

GPS,

the

programmer

first

gave

banana, and the chair. differences in positions

the set of predefined actions or operators was:

"jump," and "move

chair,"

GPS

a set

of

The programmer instructed GPS to calculate among these elements. He also informed GPS

of certain actions to perform to reduce these differences. In

itself as the

it

which described the positions of the monkey, the

self." (We'll

"move

this case,

chair," "climb

up

assume the program thought of

monkey.)

some of the actions unless certain conditions programmer also informed GPS. For example, it was of no use to have the monkey "climb up chair" unless the monkey had already "moved self and "moved chair" underneath the banana. likewise, the monkey could not move the chair unless both chair and monkey were first brought together that is, to the same set could not perform

existed beforehand



as the



of coordinates. All

this

information was given in a standardized form

called the "difference table."

To

explain a

new problem

needed only to construct an appropriate difference of the

rest

through means-ends

table.

to

GPS, one

GPS

took care

analysis.

In the monkey's case, GPS applied the actions "move self to chair," "move chair to banana," and "climb upon chair." This sequence of actions

54

Al

reduced the three kinds of differences identified during the problem

and thus allowed the problem to be solved.

analysis to zero,

Furthermore, the preconditions associated with two of the actions led

GPS

two subgoals accessory

to identify

to

subproblems was

feature

a characteristic

main

its

chair" and "get to the chair." This breaking

"move

goal:

the

up of the problem into

of means-ends

analysis. It

stemmed direcdy from Simon's observations on the workings of organizations. Also, when GPS abandoned jumping and tried another approach instead,

(Does

strategy

it

applied "backtracking," another basic tool of AI. This

work?

this

computers can

at

If not, try

times generate

something

explains

else)

more knowledge than

their

how

program-

mers put into them.

GPS,

forms, remained a part of Simon and NewelTs

in various

GPS

was

G. Ernst,

who

research from 1957 to 1968. dissertation

GPS

by

a student,

also the subject

adapted

of

a doctoral

4 to several problems.

it

learned to solve various puzzles, performed symbolic integration,

and broke secret codes.

Other experiments on modeling human cognition Carnegie Tech in the studied

human

syllables: his

And Memorizer) new

ory in a

light.

5

(I

also

went on

program

model of how people

EPAM

(for

induced psychologists to look shall say

more about

Elementary Perat

EPAM

human mem-

in chapter 10.)

Another student, Robert K. Lindsay, studied verbal behavior level: his

at

1950s and early 1960s. Edward Feigenbaum

learning by building a working

memorize nonsense ceiver

late

SAD SAM program

at a

deeper

parsed sentences in ordinary English and

them information about family trees. 6 Given the sentences "Jim is John's brother" and "Jim's mother is Maty," the machine would start building the internal equivalent of a genealogical tree. This tree implicitly contained the information that Mar)- was also the mother extracted from

of John.

The computer running

SAD SAM may

well have been the

machine to show the glimmerings of understanding For

us, as

humans,

to understand

information to other facts that usually lets us

we

is

to be able to relate

already

know, and

to

draw conclusions we have not

first

human sense. a new piece of

in the

do so

yet

in a

manner

been given. The

richer the

network of connections, the deeper the understanding.

SAM was

a first step in this direction.

SAD

55

THE DAWN OF THE GOLDEN YEARS: 1956-63

MACHINE-BASED INTELLIGENCE While

Tech contingent from the Dartmouth conference work on human cognitive processes, most of the other partici-

the Carnegie

did their

pants were the

still

human

under the impression that extensive knowledge of

brain works

was not necessary for

how

their original goal

of

developing machine-based intelligence. Their programs reflect what

would become the

traditional

AI approach.

The Geometry Theorem Prover at IBM IBM became

In the late 1950s

seriously involved in the

computer

business. Since artificial intelligence at the time looked like a natural

company allowed such work to IBM, for example, was impressed by

extension of computer research, the

proceed. Nathaniel Rochester of

some simple a simulation

theorems.

results

Marvin Minsky had gotten from manually running

of a program he had written to prove high school geometry

He

decided to try out the idea on IBM's newest machine, the

young

704 model, and entrusted the job to Herbert Gelernter,

a

with a Ph.D. in physics and earth-shaking enthusiasm.

The road from

manual simulation

much

longer and

to a running program,

much more arduous

Gelernter close to three years of

work

recruit

however, would prove to be

than anyone expected. to write

It

took

and debug the twenty

thousand individual instructions making up Geometry Theorem Prover. Before

that,

he had

endowed with both

virtually to invent a

Shaw's IPL, and the ease

FORTRAN

new programming

language

power of Newell and of programming afforded by IBM's new

the symbol-manipulating

language for scientific computations.

The Geometry Theorem Prover worked backward. 7 One first scribed to it the theorem to prove. Much as a human being does, program then

started to build a chain

of intermediate

dethe

results leading

back to known theorems or axioms. But what makes Gelernter's work so interesting and important

is

that to figure out

back to the axioms, the program looked

at a

what

steps might lead

drawing. Since computers

couldn't yet "see" through television cameras, Gelenter had to enter a representational figure as a series of point coordinates cards.

on punched

Using these coordinates, the program was able to extract the same

kind of information a

human does when

looking

at a representational

56

M Which

figure:

right angles?

With tried to

sides are equal or parallel to each other?

information, the program pruned

this

problems of

search

tree: that is,

it

something that humans do unconsciously with

this kind.

made

This pruning

all

the difference: 8 to derive a two-step proof in the

mode, the program had

1,000 x 1,000 x 1,000

program reduced

choose between 1,000 x 1,000 (one

to

million) possible combinations.

the

its

demonstrate formally only those properties that appeared to be

true in the drawing,

blind

Are there any

Are some angles equal to each other?

For a three-step theorem, there would be

(a billion)

this

choices.

By "looking"

at the "figure,"

unmanageable quantity to 25 choices for the

two-step problem, and to 125 for three steps (see figure

3.1).

The Geometry program could eventually prove theorems involving up to ten steps. More important, however, GTP was the first demonstraFIGURE Example of

a

3.1

Proof Found by the Geometry Theorem Prover

GOAL BD = EC

BDA&

DBA& CEM

ADB& MEC

BDM&

CEM congruent?

congruent?

congruent?

congruent?

AB AC? No

matching

premise

* # Tp Tr # Tr

<

PT* ^Ti

"

5 Tr M



in

Tl Ti IT Tl

#

ofio qTi Tr Tr Tr °pt Tr Tr Tr hp Tr Tr

"43"~

ff

«r

ei

1

I

°

I

I

I

1



°|

l

I

I

I

'IT 'M m

to

ib" =H= =y£ -H£

Tr tf Tr Tr

i

"^

i

*

aft: =& ft Tr Tr d£ Tr Tr Tr

3t Tr =tt tr Tr it

zt£ "ott

ft Tr it Tr a£ Tr

Tr -i Tr Tr Tr Tr Tr Tp Tp Tr Tr ilt Tr~

8 a IS

is 3S

3

8

.sy

"^TT"

Figure 9.1 (continued)

To

on one page,

help keep the drawing

have taken symmetries into account.

I

You

can reduce

missing ricktacktoe grids to one of those appearing in the figure by appropriate rotations.

of grids corresponding to the

by crosses) are

To

laid

out horizontally.

how computers

illustrate

aim

scoring, function. Its

You

to winning.

first level (first

is

To

save space,

game

search a

I

tree,

The

sets

naughts) and second level (countermove

have

laid

out the third level

vertically.

have devised the following evaluation, or

I

any stage of the game,

to measure, at

can calculate

move by

how

close the naughts side

is

value for any of the grids as follows:

its

— — For each way of completing Start at zero.

a straight line

of naughts by adding one naught to the

grid,

add

2 to the score.

— For each way of completing add

I

of naughts by adding two naughts to the

a straight line

have calculated the scores for each of the third-level grids

by the

For the second and

grids.

values of the scoring function.

appearing by

To

see

of the

some of

how

tree.

in the figure: they are the

Naughts can use them

move. The

to plan their first

naughts can plan their

move,

first

let

E on

us start at grid

(as-yet) hypothetical situation

the rightmost branch

where naughts have taken

(6)

occurs for naughts playing the lower-right corner (grid A). Naughts would

therefore take that position in an actual game. play the middle of a

row on second

level,

I

therefore assign a value of 6 to grid E. This that 6

the

is

largest

the third level (grids

grid (grid

The

on second

second

now

F

to

I).

for naughts

will

They will

is

to

level.

will

and

Naughts a

row or

will lead

to a

say: "If crosses

third level."

E

(grids

naughts would achieve

that the cross-player

know

Assuming

this

on the

is

at

Naughts

A

will

Note

to D). Similarly,

most only

also clever

that if they play a corner

a 5-score

and

will

on second

conduct

6.

Obviously then, crosses

a

level (grid J),

next move. If crosses play the middle of a

five in the

achieve the higher score of

will play a

row

corner on

worst-case play by crosses, naughts can thus assign a score to taking

first level (grid

the first-level score should be.

E

on

thus assign this value to the corresponding second-level

assume

be able only to achieve a

naughts

the center position

(grids

six

the value appearing by this grid in the figure.

level (grid J),

similar analysis. Crosses will therefore

(grid E),

is

naughts can

stage,

J).

trick

naughts

At the planning

can always achieve a

of the values associated with the daughters of grid

crosses played a corner

on

Naughts must now

second move. Excluding symmetries, there are only four choices. The largest

their

scoring function

letters

the grids are labels to clarify these explanations.

This grid corresponds to the

upon

numbers

the scores appearing are the so-called backed-up

first levels,

the center position, and crosses have just played the middle of the left column.

decide

if

grid,

to the score.

1

Note

that 5

K). Since is

it

leads to a 5 in the third level, this

is

what

the smaller of the scores of the daughters of grid

K

J).

will

then carry out a similar analysis for other

a corner: grids

Q

only to a 3 and a

and W). They 4.

will

first-level

discover that

The winning opening move

is

if

moves

(either the

middle of

crosses play correctly, these

moves

therefore to play center, which leads

5.

How

can

we summarize what we have done? Start at the deepest level laid out in the tree, and move up one level. If it's your move, assign the maximum score

calculate the scoring functions,

of the deeper positions to the parent position.

If

it's

your opponent's move, assign the minimum

score of the deeper positions to the parent. Back up along the tree in this the present position.

The

procedure's name, minimax, shows that

maximizes along the

tree.

In principle, one can use this strategy to back

analysis.

it

way

alternately

until

you reach

minimizes and

up from any depth of

226 the

Al

most promising moves

for further evaluation.

program contained information on chess

blatt's

A

section of Green-

how

and

strategy

to

generate moves.

MacHack's

by the U.S. Chess Federation, stood

rating, as evaluated

between 1,400 and 1,500' 2



tantalizingly close to the 1,537

mean

rating

of UCSF members. Greenblatt's program inspired many AI researchers to take

up computer

interest in AI.

chess. Conversely, chess experts also took

Among them was Hans

correspondence chess championship years

one of the top twelve players

in

Berliner,

who won

1968 and was for some

in the

United

States.

up an

the world fifteen

His fascination

with computers and chess led him to quit industry to seek higher education in the

Among

field

of

artificial intelligence.

other successes, Greenblatt had the pleasure of seeing his

program trounce Hubert Dreyfus, 13 who had recendy put out

RAND paper against AI

olic

14

and

insisted that

his vitri-

no chess program could

play even amateur chess.

By 1970, there were enough chess programs for the Association of Computing Machinery to schedule a tournament. It drew six entries and made up one of the main attractions of the association's annual meeting. Seeing

ACM organizers decided to hold tournaments annually:

this,

next year there were eight entries. Other countries followed

suit;

the

and by

1974, contenders for a world chess championship convened in Stock-

holm under the

auspices of the International Federation for Information

Processing (IFIPS). Kaissa, the Russian program that had

match

against Stanford,

won

won

the mail

the tide.

Meanwhile, computer programs started actively participating

man battle

in hu-

tournaments, and climbed their way up chess ratings. The

of wits between humans and

competitive

champions

spirit,

A Scottish international master with a penchant

David Levy, threw the

celebrated 1968 wager, Levy bet that

chess for ten years.

this

bets and prizes for programs that could defeat chess

proliferated.

for computers,

first

minds was on. Spurring

artificial

first

challenge.

In a

now-

no computer could beat him

The AI researchers John McCarthy, then

and Donald Michie, of the University of Edinburgh, took him up on for

£250

each. In 1971,

Seymour

of the University of California amounts. Levy collected bition in Toronto.

He

in

Papert, of at

at

at Stanford, it,

MIT, and Ed Kozdrowicki,

Davis, raised the ante by similar

August 1978

at the

Canadian National Exhi-

held up for humankind by trouncing the world

computer champion Chess

4.5, a

program created

at

Northwestern

227

GAME PLAYING: CHECKMATE FOR MACHINES? University. ers

Levy won the match 3V4

would not reach the

He

years.

1

and estimated that comput-

V£,

computer champion of the day, Levy again

Emboldened, he issued

handily.

a challenge to the

world

he was ready to wager £100,000 that no computer would

he would appoint, for ten

a chess player

him up on large

years.

So

far

at large:

defeat him, or

no one has taken

15 it.

AI

In general though, impoverished

wager

many more

thus renewed the bet for another six years. In a 1984 rematch

against Cray Blitz, the world

won

to

international grand-master level for

researchers are unwilling to

sums of their own money on the

of their programs.

abilities

however, do get their attention. In 1979, a Dutch software

Prizes,

company, Volmac, offered $50,000 to the author of any program could beat the former world champion

The

Max Euwe by Edward

prize remained unclaimed. In 1979 also,

MIT

inventor and

first

January 1984. 16

Fredkin, a wealthy

up three prizes with no time

professor, set

them: $5,000 for the

1

that

limit

on

chess program to earn a master level in

tournament play against humans; an intermediate prize of $10,000 for the

first

program

to reach the level

of international grand master; more

significandy, for $100,000 the authors

world champion. Fredkin thinks, puter chess

a kind

is

fly

by the

to beat the that

com-

he

prizes,

cites the

one

that got

more of them." 17

Spurred by the competitive tracted

program

solo across the Adantic in 1927: "Prizes are a wonderful

thing and there should be

higher in the

first

of benchmark for progress in AI research. As an

example of the positive impact of Lindbergh to

of the

most AI researchers do,

as

prizes,

human

AI

spirit

of the tournaments, and

later at-

researchers designed programs that scored ever

chess ratings. Their steady progression

is

depicted

in figure 9.2.

The undisputed computer champion of 3.0.

Written

at

Northwestern University,

model of Greenblatt's MacHack and programs did not stay at the top

the early 1970s was Chess this

program followed the

tried to simulate

for long: they

human

play.

Such

had one crippling

weakness. Despite the competence of most of their moves, they were likely to

make

outright blunders. This

either obvious or

embodied rules

happened when they overlooked

subde plays defying the general principles of chess

in their rules or evaluation functions. Conversely, these very

would

at

times lead

them

to

dumb maneuvers

that cost

them

a

game. The authors of these so-called selective search programs have

been able to formalize the principles that guide human playing only to

228 FIGURE

9.2

Computer Chess Tournament-Winning Programs in Their Times of Glory

Representative Ratings of

U.S.

1

Chess

Federation

2

Titles

World Champion's Rating

2800 2700 2600

_



senior _

/

master 2500

_

2400

2300

_

master

_

expert

2200

2100

y

+ x

nf

2000

Deep Thought Hitech Belle

1900

1800.

/

A

_

"

/ Mean rating, USCF members 153^

Jl

a

Chess

A

MacHack

x.x

:

1700

_

1600..

1500

_

c

^^i

1400 1300 1200

_

D

These

ratings,

show how ..

1100 _ 1000

E

on

the average against

human

"opponents. The scale s such that for a difference of 200 points between players, the o ne with the higher rating will win 75% of the time. An extrapola tion of the curve reaches the estimated rating of world champ ion Boris Kasparov (2800) in 1993. I

65

awardec by the United States Chess Federation,

well the pr< Dgrams did

70

I

75

r80

1

85

1



90

-

i

95

100

Yeai

229

GAME PLAYING: CHECKMATE FOR MACHINES?

an extent. They have captured some aspect of competent play, but

beyond

excellence remains

their grasp.

In 1973, the authors of Chess 3.0, Davis Slate and Larry Adkin,

weakness of

realizing this

around the so-called brute force approach. Relying on procedures and ever faster computers, the

game tree misnomer

to

new

program

their early strategy, redesigned their

efficient search

new program

searched the

depths. In fact, the "brute force" appellation

since the technique

relies

still

on

is

a

subtle strategies to reduce

it is possible to prune the game tree (that number of moves considered) by means that do not require any knowledge of the game itself. One of these strategies, called "alphabeta pruning," consists of abandoning the investigation of a move when

the search effort. Surprisingly,

is,

limit the

the

opponent can respond with

sponse to a

move

countermove better than the best

a

re-

already examined.

In the ticktacktoe example of figure 9.1, alpha-beta pruning would

work

as follows:

Assume

the search proceeds

from

right to

We,

left.

the

naughts player, have already examined the rightmost branch and assigned a value of 5 to playing center

K in

corresponds to grid

the figure).

value of playing the middle of a

Proceeding from right to

on

We

the

first

move

(this situation

then proceed to examine the

row on the first move (grid Q). we would immediately discover

left at level 2,

that for crosses to play center results in a scoring function

Consider

now that

the scoring function at level

smallest of the level 2 scores and, thus, can be

already

worth

know

(grid

K)

1

will

no

it

is

considering, and there

is

no point

thus a better

move

(grid P).

larger than 3.

on

that since playing center

5 for naughts,

of 3

correspond to the

the

first

than the one

Yet we

move is we are

to rating the other daughters of grid

Q. Alpha-beta pruning has saved us about 75 percent of the work rating the branch. If (say,

from

we had

left to right),

in

explored the branches in a different order

the search

would have taken

Modern make the

longer.

chess programs have ways of optimizing the search order to

most of alpha-beta pruning. One of them is the "killer heuristic": the computer gives precedence to investigating opponent responses that killed,

or refuted, other

the example, the killer

moves

the computer has already considered. In

move was

for crosses to play center (grid P). In

investigating the leftmost branch of the tree, the

computer would look

for this opponent's

move

first (grid

play also leads to a

weak

position for naughts, and immediately assign

a

weak

V). It

would then discover

rating to the entire branch following grid

W.

that this

230

\i

An

obvious

pruning consists

strategy' for tree

move

duplicate boards, as follows: If a

in

leads to the

keeping track of

same

situation as a

combination of moves already examined, use the corresponding known value for the scoring function. If the previous situation

through this

a

sequence of moves, to the present one, we are

move on

branch "draw," and

One problem

one

is

that led,

Label

in a loop.

to another branch.

with straightforward searches

which enables knowledgeable opponents

to

is

the "horizon effect,"

sandbag the computer by

dangling a juicy capture leading to a deathly trap, lying just beyond the

machine's search horizon. Even in the absence of traps,

computer

lead the

For

ate advantage.

and hungry for

zon

effect

to give

up an eventual

this reason,

material.

The

could

this effect

large gain for a small

immedi-

computer chess often looks impatient

chess programmers' response to the hori-

was the "quiescence search": they

let

the machine search to

deeper levels for moves leading to the capture of pieces.

These not-so-brutish search methods paid

who were

off for the

Northwestern

able to recapture the world and

North

American computer chess championships, which they had both

lost in

University researchers,

1974.

A

them

free use

unique arrangement with Control Data Corporation allowed

of the mighty Cyber

series

of computers during tourna-

ments and exhibitions. The Cyber's power allowed the Northwestern team to hold on to the annual North American championship and to the world championship grams, such as Duchess from

until 1980.

Duke

until

1

977

Other participating pro-

University, also ran

on very

large

machines. According to one participant, the value of the computers used

by the sixteen entrants to the 1977 Toronto tournament exceeded SI 00 18

The Northwestern program could rate about 3,600 chess moves per second. The resulting high level of play made it the first program to gain the tide of "expert," awarded to tournament players million.

who

attain a U.S.

Chess Federation rating of 2,000.

Chess 4.7 eventually

The

researchers

lost

both

its

tides to a

hardware version of itself.

Ken Thompson and Joe Condon of Bell

implemented algorithms

similar to those

into special-purpose silicon chips.

The

result

was

a dedicated, portable

chess computer that could beat any other machine in sight

connected to any telephone pretty-

much had

in sight, since the bulkier

had

called their



or, rather,

machines

it

battled

The portability* of Belle (as machine) was at times a source of

to stay in their laboratories.

Bell Lab's researchers

Laboratories

of the Northwestern program

231

GAME PLAYING: CHECKMATE FOR MACHINES? mishaps.

As

a plane for a

Ken Thompson were supposed

the machine and

Moscow

someone had

Labs' security department, claiming that

computers. They were trying to ship

their

Customs was

Meanwhile,

Thompson was

computer was not

the

it

Moscow. "Not

to

its

loss

remained

that

lost in

two weeks and could not

On another trip, this time to the 1983 New York, Belle had a car accident. Many

participate in the tournament.

people blamed

Moscow, unaware

in the plane's cargo hold. Belle

world chess tournament

in

19

of the world

tide to the resulting "electronic

A few weeks before, in August 1983, Belle had neverthe-

concussion." 20

master

less raised its chess rating to the

level. It

was the

ever to earn this tide, which requires a rating of 2,200 eration scale. It also earned Belle the

could then look

New

to worry,"

of high technology to the

flying to

the bowels of the federal administration for

In

one of

stolen

"our Exodus team has seized the computer." (Exodus

said,

a special project to prevent illegal exports

Soviets.)

to board

tournament, the U.S. Customs Service called Bell

at

first

first

on

computer

the U.S. Fed-

Fredkin prize of $5,000. Belle

150,000 chess positions per second.

lost to Cray Blitz, a 28,000-line-long comon a general purpose Cray XMP-48 supercomNo simple minded brute, the program had required 32,000 hours authors: Bob Hyatt and Albert Gower, from the University of

York, Belle had

puter program that ran puter.

of

its

Southern Mississippi, and Harry Nelson, of Lawrence Livermore Laboratory.

twice:

21

Cray

it

Blitz

won

was the only chess program

New York

the 1983

to reap the

world

twenty-one other teams from eight countries and, three years successfully defended also

its title

in

Cologne, Germany. Cray

Blitz,

marked the swan's song of general-purpose computers

championships. After

demise,

its

cated machines like Belle

Cray

was

a

in chess

were, so to speak, to the trade born.

North American chess champion-

machine conceived by none other than

Berliner of Carnegie Mellon University.

tion, Hitech, celebrated the

later,

though,

world championships went to dedi-

Blitz first bit the dust at the

ship of 1985. Its victor

Hans

who

all

title

championship against Belle and

days

when

CMU

The computer's

designa-

went under the name of

Carnegie Tech. Berliner's creation combined the innovative elements of Belle

and Cray

Blitz: its

positions per second, yet

dedicated circuits it

did so in a

let it

manner

process 175,000 chess

that could take elaborate

chess knowledge into account. Berliner's experience as world mail chess

champion provided the source

for

most of

this savvy. In contrast to

232

U

the $14,000,000

machine required to run Cray

Blitz, the

Hitech system

consisted of a special breadbox-sized contraption called the Searcher,

connected to a $20,000 Sun microprocessor. The Searcher contained sixty-four dedicated processor chips,

one for each square of the chess

board, which the processors monitored. Before each move, the processor detected

all

moves

on

that could land a piece

square, and analyzed

its

them. Meanwhile, other chips were doing the same thing for their squares,

which considerably increased the machine's speed. Yet Hitech

on hardware

didn't entirely rely

to select

its

moves: the sixty-four

processors fed summaries of their observations to a program called

much of

Oracle, which contained

cided

upon which move

to

Berliner's chess

make. In

knowledge and de-

way, Hitech managed an

this

average look ahead of eight levels but could, on occasion, pursue interesting

moves

of fourteen

to a depth

chess rating to well over 2,300, which

levels.

made

22

In 1986, Hitech raised

it

the world's

first

its

computer

international master.

The next quantum

leap in

computer chess

also

came from Carnegie

Mellon, but involved different people. Aside from any personal the

new machine probably

disappointed Berliner.

endowing them

believe in enhancing chess computers' performances by

with humanlike chess knowledge. They were a team of students

who knew little

creation, relied

about chess. Deep Thought,

CMU

graduate

as they called their

on speed and clever search methods. It made absolutely no

pretense at imitating

human

play.

And

deeply did the machine probe:

could analyze 700,000 chess moves per second Hitech) and project

of play. 24 In 1987,

many

rivalries,

designers did not

Its

1

23

(four times as

many

it

as

20 moves ahead along the most promising lines

5 to

a prototype

of Deep Thought, containing only half as

dedicated computer chips as the

final version,

won

the

North

American Computer Chess championship. In 1988, Deep Thought raised chess rating over the benchmark of 2,500. This

its

computer international grand master and

qualified

its

made

intermediate Fredkin prize. 25 Being a better chess player than

two hundred people league as the

in the world,

Deep Thought was now

And, indeed, the reigning champion an

official

the afternoon of 22

Boris

the

first

all

but some

in the

same

human world champion.

in history, to accept

On

it

designers for the

felt

compelled, for the

challenge from a

October 1989,

a

first

time

nonhuman opponent.

two-game match opposed

Kasparov and Deep Thought in the New York Academy of Art. 26

The computer and

the

human champion

faced off in front of an audi-

233

GAME PLAYING: CHECKMATE FOR MACHINES? ence of hundreds of reporters and aficionados, most of them in a to see the upstart

machine put

Kasparov won both games

in

its

place.

Meeting

mood

their expectations,

much

handily. This defeat did not

disappoint

Even though they had hoped to at least draw one game, their underdog status put them in a no-loss situation. Kasparov, for his part, was clearly taking no chances. He had prepared for the match by studying fifty of his opponent's previous games and the machine's designers.

avoided the daring moves that constitute his trademark. Elated as the audience was at Kasparov's victory,

members could not shake put

off a feeling

its

knowledgeable

of doom. As one commentator

"In the rapidly evolving relationship between people and their

it:

machines,

match

this

is

— one

an acknowledgement of

new, and inherently

a

on the board." 27 Said the match's organizer, Shelby Lyman, "The real drama here is that Gary is facing his fate." 28 Kasparov thought he could beat computers "perhaps 29 Lyman gave him five to ten years; and to the end of the century." short-lived state

Berliner, four.

The

of

essential parity

history of

computer chess supports

Kasparov's chess rating, the highest in the world,

is

Berliner's view.

about 2,800. The

curve in figure 9.2 portrays the steady climb of computers through

human

chess ratings.

A

straight line extrapolation

of

it

reaches Kas-

parov's level by 1993.

Which computer will beat the world champion and win the $100,000 It may very well be a direct descendant of Deep Thought.

Fredkin prize?

IBM

hired

writing,

most of the machine's design team, who

working on a new version of Deep Thought

Yorktown Heights thousand times

laboratory.

The new system

faster than the current

chess positions per second. Kasparov

have to challenge

it

to protect the

will

ones and look is

of

this

company's

deploy chips one at

roughly a billion

willing to take

human

are, as

at the

it

race," said the

on: "I

would

champion. 30

THE IMPLICATIONS OF

COMPUTER CHESS I he fun sight

of

and challenge of computer chess should not make one lose its

deeper implications. The computer's performance

penetrating questions that

from other potential

may help

intelligences,

us understand

how our minds

raises differ

and help us forecast the future of AI.

234

Al

Are Chess-Playing Computers Intelligent? Deep Thought plays chess better than 99.9 percent of human players. Does that performance make it intelligent? To this question, there are no simple answers. have already said that Deep Thought and other winning chess

I

programs and machines do not play chess minutes allowed per

move

in

tournament

humans do. In the three Deep Thought considers of the Dutch psychologist

as

play,

126 million moves. By contrast, the studies

human master

Adrian de Groot show that

moves per

average of 1.76

play.

31

ponder only an

players

They use up most of

their three

minutes verifying that these one or two moves are indeed the judicious ones.

As Herbert Simon noted, expert

players have an instantaneous

understanding of chess positions and a compelling sense of the winning

move. Simon believes

that chess masters are familiar with thousands

of

patterns involving small groups of pieces in certain relationships to each other.

Simon

a desirable

move

Each chunk would suggest

these patterns "chunks."

calls

or strategy, which would cut the need for extensive

simulations of later moves. 32 tion of chunking to

Hans

computer

According to the AI

human and machine

Hubert Dreyfus, the difference between

critic

play

is

Berliner has investigated the applica-

chess.

even more

basic.

Expert chess players,

claims Dreyfus, respond to whole board positions, not

chunks.

games to rate

An

33

expert player practicing fast play, or earning out forty

at a time, is

not simply showing

board positions instandy and

off.

He

is

intuit the

developing his

winning move

given situation. Topflight masters can play a very strong given only five seconds per move. Gary Kasparov said that the intuitive chess player

into account

component

when planning

a

somehow

move. While

takes

is

ability

in any

game when

reported to have

all

resulting

games

certainly exaggerated, this

claim agrees with Dreyfus's assessment that our major intellectual leaps are ^rational.

Some,

like the

computer-trouncing chess master David Levy, offer a

negative view of the intelligence of chess-playing computers. Tree searching, .

.

.

Levy

says,

"produces a kind of monkey/typewriter

[The computer] appears to play moderately well, whereas

playing very

weak chess so much

of the time that

its

situation.

it is

actually

best results

resemble the moves of strong players." 34 The fact remains, however, that

computers play games and win them. "Intelligence," says Edward

235

GAME PLAYING: CHECKMATE FOR MACHINES? Fredkin, tainly

having a problem and solving

"is

do

it."

35

Chess machines

cer-

that.

know human beings can reason about a large variety of subjects. Yet aren't human chess champions somewhat limited in their world views also? General-purpose game programs could, in any case, learn games other than chess. It took only a few weeks to train Deep Thought to play grand-master chess, a lifetime endeavor for any human being. Programs similar to Deep Thought could master other games, like Holders of Levy's opinion could also argue that chess computers

nothing but chess, while

intelligent

checkers, in a snap.

In the opinion of many researchers,

computers are tors

now

puters

intelligent,

reasonable to claim that chess

it is

but in a way different from

us.

Many investiga-

believe in a knowledge-search continuum, within

make up

searching.

for a lack of chess

Much

as airplanes

knowledge by an

do not

by flapping

fly

which com-

ability to

do more

their wings, chess

computers do not imitate human thought processes to win games.

Computer Chess and This pattern chines

may

brains. If

rely

such

is

AI developments. When ma-

well reappear in future

do perform mental

own, they may

Al

on

tasks

of a depth and diversity similar to our

principles different

from the ones

and

the case, interesting

that

might appear, the machines would be foreign to us inside times, this difference

govern our

familiar as their behavior

— and some-

would show.

Chess computers already provide us with some inklings of these differences.

called

A

popular distraction during AI meetings, which could be

"Turing chess," consists of pitting

human

chess players against

hidden opponents, some of them machines, some of them humans. The

and players, comes from guessing which opponents Even when humans and machines play at the same strength, knowledgeable observers can usually tell them apart from play styles: computer chess often looks ugly and inelegant. The computer, says Dave Slate of Northwestern University, "is like a shark sitting fun, for audience

are machines.

around.

It's

not very bright, but once

there and goes in all

it

munch, munch, crunch.

your armor, then you suddenly find your nicely

strategies,

laid plans

go

astray."

... If this

36

computers also often drag

gets a taste

of blood,

you allow any

thing

coming

slight

right

chink

after you,

Unencumbered by

their

it's

opponents into

and

traditional

situations

236 human

||

"They

players find bizarre.

designer,

are always original," says Belle's

Ken Thompson. "They're not

in the past."

enslaved by what's been done

37

Computer chess might model

future

AI progress

in

yet another

manner: success came about because of hardware advances; computers

human champions at chess before they achieved a certain The fact that Deep Thought could in 1 989 examine roughly two hundred times as many positions per second as the compara-

did not equal

processing power.

tively pedestrian

similar,

shows

Chess 4.7 could

that speed

is

in 1978,

even though

their software

root of better performance.

at the

As

discuss later, AI's present failure at modeling certain aspects of

thought, such as

common

sense, stems in part

their

may overcome them

machines overcome the speed bottleneck.

human mind in many people

This demonstration that machines can equal the

games

playing certain

oppose

forever

"They

I

will

put into them. Computers

dumb

in situations

that

of AI. "Computers can do only what they

are told," this objection goes. their creators

argument

also demolishes an

to the final success

never produce more than what

will

never create and

Deep Thought know

trying to

win

Thought

designer,

program

that can

little

against their machine.

do something

is

remain

"The

fascination," says the

can't do."

I

it

38

The

Deep

writing a computer

During computer chess

hard to

tell

of strategy or

a subtle piece

regularly.

chess and would never dream of

Thomas Anantharanam, "was

competitions, such programmers find

computer move

will

unforeseen by their programmers." Yet, as

mentioned, Samuel's checker program used to beat him

creators of

will

human

from the weakness of our

present hardware (see chapter 11). Researchers

when, and only when,

I

was

a

whether

a

puzzling

downright blunder.

Yet, in one significant way, computer chess failed the expectations of early researchers:

human mind. By

it

failed to

meet

hope of learning about the

their

getting computers to play chess, the reasoning had

gone, investigators could discover

how

times labeled the "drosophilia" of

artificial intelligence,

would play the

role in developing

AI

research. This did not happen. Carried

people think and decide. Some-

that fruit

flies

away by sheer competition, many

researchers started designing programs that

won games

ing like people. That

them

it

nevertheless led

processes different from our science. In the next chapter,

own I'll

is

computer chess

played in genetic

an

illustration

examine

how

instead of think-

to discover thinking

of the serendipity of

other approaches have

enabled AI researchers to model their programs on the

human mind.

SOULS OF SILICON

Long

before

we became concerned with understanding how we work, our

already constrained the architecture of our brains. However,

machines as we wish, and provide them with better ways of their own

activities

consciousness than

— and

we

this

to

keep and examine records

— Marvin Minsky

The significant forward leaps AI made during the Much more

common

industrial debacle

of the

in the eye the

sense and attacked

late

1

980s did not make

than an era of commercial ventures, this

was the decade when AI looked computers with

had

means that machines are potentially capable offar more

are.

the front pages.

evolution

we can design our new

it

problem of endowing

on

a

grand

scale.

The

1980s painfully emphasized the urgency of

Making computers see, understand the spoken word, and communicate in English became a priority. As a result, approaches more

this task.

basic gets

and daring than those permitted by the comparatively puny budand primitive computers of

for the

decades produced AI projects some of them claiming to account

earlier

involving person-centuries of effort,

whole of human cognition. These

new light on

the nature of awareness,

seriously the possibility that their

original

approaches threw a

and researchers

started to consider

machines might some day wake up to

conscious thought and feelings. Interesting debates followed a philoso-

238

41

phcr's objection to the

emergence of what amounts to

made of

assemblages of electronic chips

mere

COMMON-SENSE QUESTION

THE

with the common-sense question. Needless to say, the

It all started

various chapels into which the

AI community had subdivided

Some of

dressed the problem in different ways.

doggedly depended on with

a soul in

silicon metal.

logic.

all

ad-

these approaches

still

Others, as we'll see, would have no truck

it.

Logical and Not-So-Logical Solutions Recognizing that ability to

the neat, or logic-oriented, vestigating 5):

How

there are

To

common

element of

a crucial

sense resides in the

change one's mind when circumstances require,

community

ways out of McCarthy's

can one

tell

a

computer

good reasons why

it

in

AI research

now

problem

qualification

something

that

is

of

a large part

is

true,

busy

in-

(see chapter

except

when

should not be?

achieve this discernment, the neats discovered they had to revamp

their logical calculus, lifted

1910 edition of

almost intact from Russell and Whitehead's

They

Principia Mathematica.

Order Predicate Calculus (FOPC,

if

you

are

outfitted

on

good old

First

familiar terms with

it)

with extensions called "default," "nonmonotonic," or "modal" logic to deal with exceptions.

John McCarthy tially infinite

calls his

own

technique for dealing with the poten-

exceptions to a general rule "circumscription." His idea 1

is

to circumscribe, or delimit, the set of possible exceptions to a statement.

In the river-crossing example (see chapter isolate in the

If

none

exist in a given situation,

can cross the

As

then

a result of these efforts, the

knowing system

a

would be

of possible obstacles. safe to infer that

one

new

generation of expert systems

is

monitoring mechanism called a Truth Maintenance

TMS. Returning

that Charlie

will

it

circumscription would

set

river.

equipped with System, or

8),

knowledge base, or minimize, the

is

to Minsky's

a duck,

duck challenge (chapter

and that ducks normally

conclude that Charlie

flies.

Upon

fly,

8),

the expert

learning of poor Charlie's

239

SOULS OF SILICON passing away, the

TMS

normally undo

will

this inference.

The amount

of computation required tends, however to grow exponentially with the size of the knowledge base, and the millions of facts needed for true

common-sense reasoning would

still

snow

Other

A

to logic.

prominent example

mind of former Roger Schank

is

secretary of state Cyrus

in 1980, she

Vance

was among the

first is

form of

directiy as the experts do, in the

Confronted with

as a doctoral project for

to use a technique called

to stop painstakingly trying

new

a

a series

then search situation,

its

case, such as a set

knowledge banks for

of symptoms in a

computer would

a similar case, adapt

to the

it

new

and conclude accordingly.

David Waltz, of constraint propagation fame 4), is

now

spinoff of

Inc., a late-1980s vintage

MIT's AI Lab. Waltz, who

Thinking Machines,

It consists

in the blocks

world

(see

pursuing a goal similar to Kolodner's with the super-

computers of Thinking Machines,

at

it

of well-documented

patient or the salient points of a legal argument, the

chapter

less

the knowledge of experts into rules and, instead, record

distill

cases.

TMS. owe

Janet Kolodner: after modeling the

"case-based reasoning." 2 Kolodner's idea to

a logic-based

investigators are exploring reasoning techniques that

calls his

is

commercial

director of information systems

approach "memory-based reasoning." 3

of feeding each one of the possibly sixteen thousand different

processors of his company's Connection Machine with memories of a

When a new situation comes up, the Connec-

recorded case or situation. tion

Machine compares

it

to

and uses

it

to define a possible solution. Janet

closest match,

all its

records simultaneously, identifies the

has recently spent a sabbatical at Thinking Machines on a

Kolodner first try at

implementing case-based reasoning into the Connection Machine: the results are encouraging.

However, the most dramatic AI project aimed

at instilling

common

sense into computers, and certainly the best-funded one, owes logic. It

is

Douglas Lenat's Cyc

An Ontology

for

effort at

Common

little

to

MCC.

Sense: Cyc

his quest for ways to make computers learn (see chapter met with some spectacular successes. His Automated Mathematician program learned by discovering mathematical concepts; and its succes-

Lenat had, in 7),

sor,

EURISKO,

as playing

learned by inventing new heuristics for such activities games and designing computer chips. In 1984, Lenat was

240

||

MCC,

approached by

a

consortium that made up America's answer to

the Japanese Fifth Generation Project. Short for Microelectronics and

Computer Technology Corporation,

MCC was chartered by such heavy-

weights as the Digital Equipment Corporation, Control Data Corporation,

Kodak, and National Cash Register

And

decade-sized projects.

began an

MCC

so

it

was

to carry out large, high-risk

September 1988, Lenat

that in

research report as follows: "I

would

like to

surprisingly compact, powerful, elegant set of reasoning

form

a set

mon

sense reasoning

of

first

Although such

principles



which explain

a sort

a discovery

creativity,

present a

methods

of 'Maxwell Equations'* of thought."

would have been

a logical

outcome of Le-

nat's previous

work

into the world

and make sense of what they saw, he continued:

very

much

in

that

humor, and com-

programming computers

to figuratively

to present [those reasoning methods], but, sadly,

believe they exist. So, instead, this paper will

tell

go out

"I'd like I

don't

you about Cyc, the

massive knowledge base project that we've been working on

at

MCC for

the last four years." 4

Stemming from an admission of defeat, Cyc is

(short for encyclopedia)

a $25-million research project that will last for

two person-centuries.

Lenat had become convinced that no amount of finessing and fancy

footwork would ever facts as

let a

"Nothing can be

in

machine discover by

two places

at

itself

such elementary

once," or "Animals don't

like

The most that we need

pain," and "People live for a single solid interval of time." 5 salient discovery in

to

know

a colossal

in the world.

new

AI since the Dartmouth conference is number of these common-sense assertions

Lenat convinced

discipline

stemmed from

need to encode

this

his

MCC

to get

by

sponsors that the woes of the

repeatedly trying to wriggle out of the

knowledge manually, tedious

fact after painful

assertion, in machine-usable form.

new land without long And thus in 1 984, they embarked

"Fifteenth century explorers couldn't discover

voyages," claimed Lenat and his team. 6

on

own

their

excursion over the ocean of knowledge: during a ten-year

period, they were to encode about ten million assertions like the ones in

the previous paragraph into a gigantic knowledge base a billion bytes

Derived

in the nineteenth century

will

allow

by Scottish physicist James Clark Maxwell, the all of electromagnetism in a few crisp lines

equations Lenat refers to explain virtually

of algebra.

amounting to over

of information. Ad-hoc inferencing mechanisms

241

SOILS OF SILICON a

computer to reason from

language. Like an iceberg,

and much

evident" facts

knowledge and understand natural

Concise Columbia Encyclopedia. Below the loom the behemoth collection of "selfthat children know when they enter grade school, and that

to the contents of the waterline,

that

Cyc will have a small visible part corresponding

one-volume

larger, will

are never included in reference books. Lenat's first challenge

impose

a

workable structure onto

To work out an adequate universe,

set

this

amorphous jumble of knowledge.

of ontological categories for carving up the

MCC researchers started by lifting pairs of sentences at random

from newspapers, encyclopedias, and magazine

grammed the

was to

into

Cyc the basic concepts inherent

program could "understand"

articles.

They then pro-

in each sentence, so that

meanings. 7

their

The first two sentences took Lenat's team three months to code. They were: "Napoleon died on St. Helena. Wellington was saddened." Through a complex hierarchy of interlocking frames, the researchers were able to impart to Cyc the knowledge that Napoleon was a person.

To

Cyc, a person

is

a

member of a

make up

Agent. Persons

a subset

denoted by the frame Individual-

set

of the larger category of Composite-

Objects: those objects that have a physical extent (mass) and an intangible extent (such as a mind).

The

set

CompositeObject

IndividualAgent states that the habit of dying. Death, in turn, as

members of this is

a subset

one of its properties TemporalExtent

this

means

that

when a person is

to Wellington, other properties

set

also comprises

A slot in the frame

such things as books, which have mass and meaning.

have the unfortunate

of the frame Event, which has

(indefinite, in the case

dead, he or she stays dead).

of death:

Going back

of IndividualAgents are those of har-

boring beliefs and emotions. Sadness

is

an element of the

set

Emotion

and frequently accompanies death. Invoking further combinations of frames, the researchers could also convey to

doing

battles

is

Cyc

that if the time

of

long over, even the death of one's archenemy can

sadden. This involved talking about war, France, and England (which are CollectiveAgents),

and news media (Wellington probably learned

about Napoleon's death through the newspapers). islands

was needed

to explain St. Helena,

out the nature of land,

For

a while

it

sea,

which

A

description of

in turn required spelling

and water.

appeared to Lenat's team that they would be stumped

by the philosophical objections of Hubert Dreyfus: every sentence required the definition of a categories,

and

reality

new and

seemed

arbitrarily

long chain of related

to branch out into an

infinitely large

242

Al

number of unrelated a stage that

Lenat

concepts. In September 1987, though, Cyc reached

calls

define systematically

"semantic convergence":

new concepts

it

became possible

researchers (or "Cyclists," as Lenat dubs them) could by then enter

knowledge

in

Cyc by locating

modifying the copy. As a

similar

faster than originally estimated.

1990s,

we can

transition

would

from

answering Cyc's questions about

Lenat

that

difficult

entry of assertions

texts; the role

of humans

of brain surgeons to

tutors,

sentences and passages." 9

upbeat about the future of Cyc, and believes that

is

new

and slighdy

Lenat hopes that "around the mid-

by reading on-line

transition

it,

knowledge could be entered

more and more from manual

to (semi-) automated entry in the project

knowledge, copying

result, said Lenat, 8

to

terms of other concepts. Cyc

in

it

has a

good chance of serving "as the foundation for the first true AI agent. No one in 2015 would dream of buying a machine without common sense, and more than anyone today would buy a personal computer that .

.

.

couldn't run spreadsheets

[or]

The opinions of other AI

When

I

researchers about Cyc vary considerably.

discussed Cyc with him, Marvin Minsky was enthusiastic about

the project.

He

agreed that Lenat was on the right path, and regretted

were no other projects

that there

Minsky

word processing programs." 10

travels

Most other

from time

I

MCC

facility in

Austin, Texas.

Cyc

going to work,

researchers just don't believe that

but can't help being fascinated with

on Cyc

Cyc. Himself a part-time Cyclist,

like

to time to the

it.

is

Witness a technical presentation

attended at the 1990 American Association for Artificial In-

meeting in Boston: it was scheduled in an otherwise-dull on knowledge representation which attracted few listeners. Yet the AAAI organizers are sticklers for punctuality, and everybody knew that the talk about Cyc would start at 3:15 p.m. sharp. Just as Cyc's co-director, Ramanathan V. Guha, was climbing the podium, the small auditorium suddenly filled to capacity with about two hundred deserters telligence

session

from other simultaneous presentations. After

listening to

using up the question period to throw several barbs representing default knowledge, they prompdy

at

filed

Cyc's

Guha and method

out of the

for

room

again.

The problem Lenat and

is

that

Cyc

is

the quintessential scruffy

Company make no bones about

of the AI community," said Lenat,

".

agonizing over the myriad subtleties

.

.

it.

"We

in that

of, for

AI

differ

we have

project,

and

from the

rest

refrained

from

example, the different for-

malisms for representing pieces of time; instead,

we have

concentrated

243

SOULS OF SILICON

on using

we have

the formalisms

for actually encoding information

about various domains." 12 What he meant was that instead of worrying about

why Cyc

AI people

couldn't be done, they just

went ahead and did

are spending their research lives creating

"Most

it.

bumps on

logs,"

Lenat said elsewhere. 13

Hans Moravec, who was

me

said to

that

a graduate student with Lenat at Stanford,

Cyc has been labeled

in the grapevine as a "half-serious

attempt." 14 Randall Davis, another of Lenat's fellow graduate students,

doesn't grant Lenat substantial

more than an "outside chance" of capturing

Newell both pointed out to tion

a

body of common-sense knowledge. 15 David Waltz and Allen

mechanisms

in Cyc:

related concepts in the

me

16,17

the ad-hoc character of the representa-

as a result,

Cyclists

same way, which could make

them

the knowledge base to associate rect the situation

two

may not encode it

impossible for

afterward. Lenat intends to cor-

by having learning programs crawl over the knowledge

base at night and set

it

but Newell was pessimistic about the

right,

outcome.

But the of

of the

attitude

total rejection, as

Cyc

is

It is a

toward Cyc

artificial intelligentsia

Randall Davis

summed

it

up

for me:

moon

or not.

Some-

is

going to make

come out of it.

whose overarching goal

is

.

.

.

it

to the

There are people

thing to work.

and

.

.

.

is

causality for millennia,

and they are

still

arguing about

attitude that until they get the

answer

we're not going to be able to build an intelligent system.

look around and see that the planet systems [ourselves],

if

we

just build

it

who

and so in a

learn a lot!" This

to get the

Philosophers have been dealing in theories of time,

you can take the

space, causality,

in the field

to get the theory right. That's important,

but there have to be people whose fundamental goal

space,

not one

vast experiment in absolutely hard-core empirical AI.

not a rocket ship that

thing important will

Now

is

is

is

it.

right,

Or you

can

populated by semi-intelligent

have only the barest theory about time,

forth.

You

can

justify

way that is good enough what Lenat

is

doing.

Cyc by

saying:

"Maybe

for the time being, we'll

244

Al

AND PSYCHOLOGY

Al

Where Cyc embodies humanistic flavor.

From

early on,

a

much more

Al researchers have maintained

tions with their psychologist colleagues, results.

com-

the pragmatic, blue-collar approach to the

mon-sense problem, other avenues of investigation have

rela-

sometimes with productive

The examination of human thought processes

occasionally

helped scientists design replicas of them into machines. Conversely, findings

made without

human cognition have helped psyhuman mind better. A science of universal regrouping both people and machines, may well be in the reference to

chologists understand the

psychology,

making. finally

It

might help us answer such questions

as:

When

machines

think and manipulate their environment, will our minds and theirs

common? Will our basic

have anything in

ground with

drives and purposes share any

theirs?

Modeling Human Problem Solving We

saw

in chapter 3

how NeweU and Simon implemented

into their

General Problem Solver reasoning techniques abstracted from tape recordings of actual problem-solving efforts by

Another

early

ceiver rize

later

His program

And Memorizer)

meaningless

subjects.

example of psychological modeling was the work of

Edward Feigenbaum, who (see chapter 6).

human

pioneered commercial expert systems

EPAM

(an

acronym

for

Elementary Per-

mechanism by which we memoBy so doing, Feigenbaum hoped to unearth

simulated the

syllables.

18

principles applicable to the entire learning process. Perhaps

the influence of behaviorism (see chapter

phenomenon

2),

the

first

member of

under

Feigenbaum studied the

as a stimulus-response situation.

sented a subject with pairs of monosyllables.

still

An

experimenter pre-

The experimenter

stated

the pair (the stimulus), and the subject tried to

answer with the second

member

(the response). Behaviorist theories did

not properly account for some of the phenomena that occurred in such experiments, like oscillation: a subject learned a sequence of syllables correctly,

and then forgot them and learned them again many times.

Neither could behaviorists explain

why learning new associations made phenomenon called "retroactive

the subject forget previous ones, a inhibition."

245

SOILS OF SILICON Explaining these

phenomena through

procedure a behaviorist

a

would never contemplate, Feigenbaum postulated a complex information structure in the subject's mind and proceeded to build a computer

model of it. This

structure, called a "discrimination tree,"

learning process.

assumed

Feigenbaum worked out of

that subjects didn't bother

instead, extracted salient features

and

letters)

tion used

built

no more

already learned.

assumptions

of the

he

entire syllables but,

syllables (for

example, the

features.

The

first

discrimina-

features than those required to associate the pairs

Feigenbaum found out

in his

a very simple idea:

remembering

up associations between these

grew with the

implementing these

that by

program, he could reproduce the inexplicable phe-

nomena. For example, suppose the program had learned to associate the

JIR-DAX by

syllable pair

starting with/,

new

I

simply remembering "any time

must respond with DAX." Then,

pair JUK-PIB, the

program

suffered

from

if

I

retroactive inhibition for

the following reason: Since the discrimination rule could

JIR and

more

JUK

apart, the

program had

syllable presentation

syllable

to develop a

were needed. During

was "forgotten" because access

to

it

see a syllable

presented with the

new

no longer one. For

this time, the

was

tell

this,

response

lost.

Generalization common ground

Yet another

between psychology and AI concerns the

problem of generalization, about which psychologists have always

How

speculated.

do we manage

to extract global concepts out

of the

we constantly face? How do we learn about furniture and plants, when we see only individual chairs or trees? On a more abstract level, how do we identify a given instance of interpersonal behavior for example, a conflict or a betrayal? AI made a direct use torrent of specific instances



of psychology's response to these questions.

We

make

turing our

sense of the world,

knowledge

in the

some

psychologists assumed, by struc-

form of prototypes, or schemata. For

example, the prototype for a chair was an information structure containing a generalized description of statement:

"A

chair

is

it.

something to

It sit

somehow corresponded

a seat,

and

cepts.

Although the philosopher Immanuel Kant

a back." Similar descriptions existed for

schema in his Critique of Pure Reason, published in in the

realm of philosophy

to the

on, usually including four legs,

1

more first

abstract con-

used the word

787, the idea remained

until the twentieth century,

when

the

new

246

41

science of psychology took

up.

it

19

Sir Frederick

psychologist, used schemata in his 1932 studies scribe

how we

remember oudine

first

For example, of

a

it

which

are

common

The Swiss

in his theory

its

One remembers

schema and the

typi-

a given story as a

particular traits that

on

psychologist Jean Piaget elaborated

made

stand

it

the idea and used

of mental development. 21 The German philosopher and

Max Wertheimer

psychologist

we

to several stories.

corresponds to one schema ("Valiant prince

Western make up another.

superposition of out.

features,

a fairy tale

to de-

essence of a story. 20 According to Bartlett,

princess from bewitchment"). The main sequence of events

frees cal

recall the

Bartlett, a British

on remembering

incorporated schemata into the Gestalt

theory of perception. 22

The concept remained vague, however, and experimental psychologists could never make any concrete use of it. When Marvin Minsky introduced schemata into AI in his 1975 paper "A Framework for Representing Knowledge," he called them "frames" in order to freshen up the tion.

23

As we saw

income

tax form,

its

some

generic frame,

slightly stale psychological

earlier (chapter 7), a

main

Minskyan frame looks of

feature being a set

slots are

slots to

fill

no-

like in.

an

In a

empty, and others contain default entries

corresponding to usual or necessary characteristics of a concept. For example, the frame for chair has

number of legs holds the number tion

Another

to sit on.

more

slot

4.

shows

slots labeled seat

The

back.

The slot men-

that a chair

is

a specific case

of a

general frame called piece offurniture. In the generic case, other

— remain empty

slots

— such

fills

them with appropriate values when applying

cific

and

slot purpose contains the

as

color, weight, material, height,

position

7 .

One

the frame to a spe-

instance of a chair.

In AI programs, frames serve several purposes.

One

can use them to

recognize objects or concepts ("It has four legs, a seat, and a back: what is it?").

are. They can help reach Look for something with four legs, a seat, Frames allow AI programs to make inferences ("It's a

Frames can keep track of where objects

goals ("Need to rest your legs?

and

a back").

chair,

but

I

can see only two

legs: the table

corner probably hides the

other two"). Schank's scripts (chapter 7) are frames designed for the express purpose of inferring what stories leave untold ("If John ordered a

hamburger

in a restaurant, he probably ate

it,

even

if

the story doesn't

say so.")

Not by

coincidence, studying the role of prototypes in intelligent

behavior has developed into a growing

new

field

of psychology.

Categori-

247

SOULS OF SILICON Ration

and

prototypicality effects are

now new

research avenues in this

science.

Tools for Psychological Investigation The

influence of AI

on psychology extended beyond providing general workings of intelligence. AI has also influenced the

about the

insights

everyday practice of psychological investigation. Because of AI, psychologists

now have new instruments

for prying

open the

secrets

of the

mind.

The

first

one

is

a tool for expressing

many of

their theories in un-

ambiguous terms: the computer program. Formulating precisely that

it

will

run on a computer forces crystal

puters tolerate neither

a hypothesis so

clarity

on

it.

Com-

hand waving nor fuzzy thinking and keep you

honest. Unrealized, implicit, or unspoken assumptions are mercilessly

weeded

out.

Weaknesses of construction appear

as if under a

magnifying

glass.

Second, running theories on a computer allows psychologists to

The computer has a substantial edge over the As MIT's Patrick Winston facetiously pointed out to Pamela McCorduck, a computer requires little care and feeding and does not bite. 24 As a more considerable advantage, computers give psycholoevaluate

them

laboratory

better.

rat.

gists a capability

isolate

up

so far restricted to the hard sciences: the ability to

phenomena.

their

Physicists

and chemists have always been able to

experiments so as to study only one factor

to investigate gravity? Plot the impact speeds their falling heights; see

of objects

whether heavier objects

ers,

one could hardly apply

similar

Keep

it

methods

Yet

all

this isolation

other mental

of factors

memorization of meaningless

is

as a function

of

You want vacuum. You think

constant! Before

comput-

in psychology. It

tremely difficult to isolate a particular aspect of a

independently of

set

You want

fall faster.

to cut air friction? Carry out the experiments in a

temperature might have an influence?

at a time.

was ex-

mind and study

it

activities.

precisely

syllables.

what Feigenbaum did

He

riations in a hypothesized discrimination

could study

for the

how minute

va-

procedure affected learning.

He never had to worry about subject fatigue or random fluctuations. The writer Pamela McCorduck aptly captured the advantage by calling AI programs "parts of intelligent behavior cultured in silicon." 25

248

Al

Cognitive Science Through ever-increasing intermingling, intelligence

which the two a discipline

from other

the boundaries between

and psychology have grown fuzzy. The blurred of

fields its

blend into each other

is

science."

as

With help

(anthropology, linguistics, philosophy, and neuro-

science), cognitive science aims to explore the nature

The Sloan Foundation favored

the mind.

along

even starting to emerge

own, under the name "cognitive

fields

artificial

line

and functioning of

the emergence of cognitive

science through a multimillion-dollar financing effort in the 1980s.

The main

cognitive science research centers are

Diego. Others exist

MIT,

Berkeley, and San

Carnegie Mellon, and the University of

at Stanford,

Pennsylvania.

How science,

does one distinguish between the disciplines of AI, cognitive

and psychology? At the

here

fields,

is

a try at telling

risk

them

of offending researchers

apart.

building thinking machines, while psychology studies

and

feel.

Cognitive psychology

tive science is a

in

Roughly speaking, AI

tries to learn

how

how

about

people act

people think. Cogni-

meeting ground between AI and psychology.

These differences

in goals

frequendy correspond to differences

methodology and philosophy, often with misunderstanding, and even contempt between workers gists,

three

all

is

in different disciplines.

in

distrust,

Psycholo-

for example, mostly interpret the workings of the mind through

meaning. AI workers, instead, look to the mechanism of the process

The

involved as an explanation.

psychologist Sherry Turkle of

gives the example of a typical Freudian

opens a meeting by declaring

would read

feelings

meeting to be over.

it

closed.

An AI

may proceed

symbol back,

which

chair's

hidden wish for the

worker, on the other hand,

in the

may remark

that

their opposites in similar ways:

same manner. Minute

like flipping the sign bit

MIT

a chairperson

Into this lapse, psychologists

of uneasiness and the

computers often encode meanings and the brain

26

slip, in

errors in reading the

from positive

to negative,

may

turn "open" into "closed."

The

psychologists Noel Sharkey from Stanford and Rolf Pfeifer

from Zurich University point out another

difference:

AI and psychol-

ogy section cognition differendy. 27 Psychologists investigate long, horizontal sections of cognition. Sharkey and Pfeifer cite as a typical exlogogen model or word recognition. word of our vocabulary corresponds in

ample the psychologist Morton's

Morton postulated

that each

249

SOILS OF SILICON our mind to a specific word detector, or structure of these detectors,

of word perception

in several contexts

experimental effects from

model applied both cases to

it.

to written

why we

An AI

work, since

logogen.

Morton applied them

many

Elaborating on the to the

phenomenon

and explained a wide range of

different sources.

For example, the

and spoken word perception:

it

explained in

can recognize a word faster after a recent exposure

worker might complain about the lack of generality of it

concentrates

on only one

this

part of our understanding

processes: the word.

For an AI researcher, generality involves processing section of cognition that

would look very

a deep, vertical

thin to a psychologist.

For

example, Roger Schank's story-understanding programs involve everything from parsing sentences to explaining the behaviors described. Yet a psychologist

may

find this

stories in a limited field

AI workers and

work too

restrictive

because

it

handles only

of understanding.

psychologists frequently emerge from different back-

grounds and pursue different

They often deeply

goals.

mistrust each

work and apply adjectives like "naive," "slipshod," and "irrelevant" to work performed in the other discipline. Psychologists often decry AI workers as amateur psychologists, and vice versa. Cognitive other's

scientists try to reconcile

of both

opposing viewpoints, often to the disapproval

sides.

Despite these differences, an uneasy alliance has formed between AI

and psychology

in specific fields.

Researchers recognize that work on

sensory perception can be good experimental psychology and good AI.

They can match computer program ior

results

of subjects with such accuracy that

many

instances, researchers even

know

neuroanatomical structures in enough modeling. Higher-level cognitive

and the experimental behav-

litde

argument

is

possible. In

the workings of the pertinent

detail to

permit specific hardware

activities like

conscious thought or

language processing, on the other hand, are too remote from experimental test

to allow such research. It

is

there that

AI may have

to precede

psychology. Hypotheses formed through computer modeling could

then enable the design of experiments specifically aimed

Is If

the

at testing

them.

Human Mind Unique?

computers can model our mental

activity in different

proved that our minds are not unique? Could processes

ways, have

we

radically differ-

250

Al

ent from the ones in our brains lead to cognition? Such speculations led cognitive scientists to investigate the question of if

minds can

apply to

all

differ

of them, whatever

mental properties aliens

from each other,

from other

common stars?

to

The

their

mind

in general.

Even

are there general principles that

make up?

Is

it

possible to extract

humans, thinking computers, or even

investigation of this question has barely

begun, and scientists can offer few conclusions. However, the very questions they are asking expose the flavor of the research.

The more

abstract question deals with the continuity of the space of

minds. Can one say that the set of

possible

all

minds forms

a

smooth

slope? Is there an unbroken rise starting at bacteria and thermostats and

leading to the

human mind

The sketchy results From what we see, the path to

or intelligent computers?

available point to a negative answer.

higher minds involves long stretches of cliffs.

In the 1950s, John

von Neumann

ground broken by soaring

flat

identified

one of these

cliffs

and

Beyond this level of organization, he 28 said, a being can make another one more complex than itself. Yet another step-increase in mind power is the phenomenon of cognitive called

it

the "complexity barrier."

convergence, which Douglas Lenat claims to have achieved with the Cyc project. Cyc's designers

now

have

fed into

be able to explain most new entries

Other questions cognitive gence inherited?

How much

in

it

enough

terms of

scientists ask are:

can one develop

basic concepts to

earlier ones.

To what it

extent

after birth?

answer stems from the position that intelligence

is

29 is intelli-

One

partial

mostly knowledge:

the very act of reasoning requires the application of techniques and steps to learn. Some speculate that the structure of the somehow determines what we can learn, and experience determines what we do learn. If this is the case, we would mosdy inherit intelligence. Others claim that we are all born with the same basic abilities. The specially bright people are those who, accidentally or that

humans have had

brain

otherwise, learn in their youth ways of making their learning

more

efficient.

Yet another all

issue concerns

AI programs,

domain

limiting their fields

limitation. This

of competence to narrow domains.

For example, an expert system for

electric

nothing about diesel engines or railroad

workers claim that

Some

in practice

point out that

we even

weakness plagues

cars.

locomotive repair knows

Perhaps

in self-defense,

AI

people are also severely domain-restricted. label each other according to

our domain

restrictions, or specialties: salesman, accountant, lawyer, cook,

and so

251

SOILS OF SILICON on.

From making

observation to claiming domain restriction as a

this

general property of intelligence

is

and many have taken

a small step,

it.

Were they right? The future will tell. Anyone who has had a hand at programming computers knows about program loops. They occur when a weakness in programming logic causes the computer to repeat the same steps endlessly. AI programs are not it

immune

to loops,

and debugging one

from going into ever wider loops.

First,

is

often a process of keeping

one

identifies trivial errors that

induce the repetition of short sequences of instructions: these are easy to

One

fix.

then discovers that sometimes a subtle conjunction of

cumstances makes the program jump back ecuting

many

instructions.

The reasons

for this behavior often

of the program. Devising additional program

in the structure

detect offending circumstances and prevent looping

may be

We

know

all

people

managed

to

who

it

is

so

common

that

classical diagnosis

of

a

woman who

marry not one, not two, but three cancerous husbands

she attended on their death beds. 30 Granted the pathological

program loops? Often the

common

occurrences be

cycles last several years. Their victims

well act out of healthy and natural inclinations. Since one's basic

disposition remains the same,

same kinds of people a

they occur

compulsive acting out of unconscious

as a

made one such

character of such extreme cases, couldn't the

may

tests to

endlessly repeat the same, often self-

of behavior. This phenomenon

psychologists interpret wishes. Freud

just

when

deep

lie

major undertaking.

a

destructive, pattern

whom

cir-

to an earlier step after ex-

little

it

sooner or

will

situations. Like the

one

in the



may simply lack an appropriate loop-detection mechanism like man on one's shoulder, for example, who at the beginning of a

new loop would

detect analogies with the present situation and

equivalents in past cycles. But since each loop

the one before, and similar situations several years,

there

later place

computer program above, these

you

might take

it

are!

activities tinuities,

attributes

all

it

again!

that such monitoring

minds, natural or

Invoking such a unsatisfying.

little

its

from

may occur only at intervals of little man indeed to say, "Look, Stop it!" As we shall soon see,

make up the major part of our mental domain restriction, and the tendency of

slightly different

a very clever

You're doing

Marvin Minsky surmises

is

man

and error-prevention

processes. Like disconto loop, they could be

artificial.

to explain mental processes

For what accounts for the thought processes of

is

deeply

this little

2

M

52

,.

Of

.

.,-

r:iir

7"

'i.i.

" :

i_

253

SOILS OF SILICON

(author of the MicroPlanner language, itself an early experiment in

decomposition), recalled

for me:

it

more

Kristen Nygaard and Ole-Johan Dahl* were inspired to develop

modular ways to program guages

FORTRAN.

like

.

.

simulation than existing If

.

ways using

you were going to simulate

lan-

a car wash,

you needed an array of variables over here to keep track of which

cars

had been washed or not, you'd have an x-y-^ coordinate array over there that kept track of

and

what the positions of the automobiles were,

piece of code to put

this spaghetti

together.

it all

So they

said,

"Hey, we've got objects out there: automobiles and car-washing stations. bile

So

let's

have each object keep track of itself. Let an automo-

keep track of whether

it's

clean or dirty, of

what

its

position

is,

Each washing station will keep track of how much soap and water it has." ... [In this way], they were able to do

of how long

it is.

simulations in beautiful, elegant fashion. 35

Special-purpose programming languages like

SIMULA

and Small-

Talk appeared, which implemented these concepts. Their basic philoso-

phy was

for the

programmer

to describe not

using computer data structures, but

who

anisms built into the languages then

let

what the objects did by

they were; specialized mech-

these creatures interact as their

natures dictated.

The advent of computer networks provided

the clinching motivation

for developing software-enabling multiagent interactions. Instead

concentrating

all

their data processing in a single

corporations found ers,

and

locally.

later

it

more

practical in the

1

of

mainframe, large

970s to have minicomput-

microcomputers, perform whatever operations they could

The machines then

transmitted only the required information to

each other. Pretty soon they themselves started gossiping over telecom-

munication mation,

lines like

unquenchable busybodies:

air traffic

control infor-

and electronic fund

transfers

now make up

airline reservations,

a sizable fraction

of long-distance communications. Protocols such

Message-Passing allow the orderly and error-free cooperation of processors. Carl Hewitt formalized their analysis in a

computer science

By

called

mind was

*Inventors of the language

new branch of

"Actor Theory."

coincidence, at about the

ing minds within a

as

many

same

time, the idea of multiple cooperat-

also gaining

SIMULA

at the

ground

in

psychology

— owing

Norwegian Computing Center

in Oslo.

254

Al

from Freud's model of the brain

in part to psychologists' shift

of Norbert Wiener tion,

of cybernetics

in his theory

(see chapter 2).

he and psychologists after him believed, better describes the

of thought than energy does, and

this

is

stuff

what the brain receives from

outside. Information, contrary to energy, does not build fact,

to that

Informa-

up pressures. In

the law of conservation does not even apply to information:

necessary, tion than

one can destroy

it

was

gists,

releases.

The

if

and the brain can receive more informa-

data,

brain's

main

concluded these psycholo-

drive,

and form

to organize information into coherent objects

between them. Thus was born the psychology of internal-

relationships

object relations.

Freud, to his credit, had already described a process by which "take in" people to form inner objects.

phenomenon

pathological



He

first

internal-object formation as a part of

we

as a

Freud came to think of

normal development. Later theo-

the Austrian psychoanalyst Melanie Klein, widened this view,

up

setting

it

the reproaches of an internalized father,

for instance, leading to melancholia. Eventually,

rists, like

conceived of

as a base for

mental

life

the relationships

we have

with our

inner images of people as entities within the mind. 36 In the words of the psychologist

Thomas Ogden,

these agents were "microminds

scious suborganizations of the ego capable of generating

experience,

i.e.

.

.

.

uncon-

meaning and

capable of thought, feeling and perception." 37

though many psychologists do not accept

this

view

in

its

entirety,

Even most

admit that one does build up one's personality by identifying with other people or fragments of their personalities.

A Society

of

Mind

In 1986, Minsky published The

Society ofMind, in

cept of subminds into a sweeping attempt ligence.

scruffy

The

which he

at explaining

built the

con-

human

intel-

38

In a sense, The Society of Mind represents the full blooming of the movement away from pure logic as an explanation of the mind.

scruffies

a gadget,

hold

that, in the

words of Daniel Dennett, the mind

an object that one should not expect to be governed by deep

mathematical laws, but nevertheless a designed object, analyzable tional terms:

.

.

.

.

.

The mix of will

in func-

ends and means, costs and benefits, elegant solutions on the

one hand, and on the other, shortcuts, .

is

be discernible

in the

jury rigs,

and cheap ad hoc

Rube Goldberg found elsewhere mind as well." 39

elegance and

fixes.

in nature

255

SOILS OF SILICON

own mix of elegance and ad-hoc fixes is reminiscent of the hive-mind theme of science fiction. Our minds, claims Minsky, are made Minsky's

up of

a billion entities,

dumb and know

which he

calls

"agents." Individual agents are

only one function. They constantly monitor inflow

from the senses or

produced by other agents. They perform

signals

whatever action they are capable of upon recognizing

Mind

tion signals. conflicting,

results

their

own

from the simultaneous and often

and disorderly action of agents. Structured

activa-

tangled,

as a loose hierar-

make up specialized systems called "services." A high-level may be capable of quite advanced tasks, and terms usually

chy, agents service

reserved for a whole person might apply to

it.

In an example strikingly suggestive of the block-manipulating robot in the

Micro Worlds

ate in the this

mind of

hierarchy

(call it

nate functions:

Minsky describes how agents might oper-

project,

The

a child building block towers.

BUILDER) knows

only

how

highest agent in

to call three subordi-

BEGIN decides where to place the tower, ADD piles up

the blocks to build

it,

and

END

decides whether the tower

enough. Each one of these functions depends on

its

own

high

is

subordinates:

FIND a new block, GET it, and PUT FIND, GET, and PUT, in turn, activate their own suband so on down to the agents that eventually control eye move-

those of ADD, for example, will it

into place.

agents,

ments or

Some

activate the muscles causing the child to pick

find this

ill-defined tissue

clear that

model too formal

of interlocking meshes of neurons

one can

up the blocks.

a description for the diffuse in the brain. It

isolate individual neural structures

is

and not

corresponding to

Minsky's agents. Yet the model provides a surprisingly faithful account

of many mental phenomena. Here questions

it

a brief sampling

of some of the

answers:

Why do babies all

is

seen a baby

suffer

move

in

sudden and

drastic changes in

humor?

We have

seconds from a contented smile to tears of rage

or hunger. Answer: early minds consist of few independent services

geared to satisfy our basic needs. other,

services

make up

When

control passes from one to the

mood

result.

More complex,

the adult mind:

mood

shifts are less drastic

sudden changes

in

tightly linked

and take

longer.

How does

the

mind grow?

How do we learn?

agents and services and include

when

a structure

crowded nearby

becomes too

tissue.

them

Answer: we form new

in existing hierarchies.

large,

it

However,

can no longer expand into

In addition, other services

come

to

depend on

256 it,

Al

and further changes might disturb

When

brain.

explains

it

is

why we

their activities.

new

ready, transfer control to the

learn

new

The

solution?

structure. This

bursts of rapid progress separated

skills in

A

by stretches of slow development, or "learning plateaus." corresponds to the construction and testing of bursts of progress occur

How

when

do remembering,

new

a

recall,

structure

in several services.

when we have

a

These "knowledge

line associated

new

plateau

structure.

The

suddenly switched on.

specific aspects

lines," as

Minsky

new experience. For example, our

might activate the agents for

knowledge

is

a

and associations occur? Some agents

between other agents, activating

act as links

Copy

improvements into another part of the

the structure with any required

first

sight

round, fist-si^ed, smooth-skinned,

new concept of apple

with the

of a concept

calls

them,

arise

of an apple

and

will

red.

The

afterward

link, framelike, these properties together.

Like Freud's, Minsky's model of the mind assigns an important role to unconscious

mechanisms. In Minsky's view, avoiding mistakes

new

important as learning "suppressors," ated

when we

come

first

skills.

into being

is

as

Special agents, called "censors" and

when we

blunder.

The suppressor

cre-

put our fingers in a flame detects our intention to do

so the next time around and prevents the action. With time, this suppressor evolves into a censor, which prevents us from even thinking of

touching a flame. Censors speed up our mental processes: to reach an object

on the other

side

of

a flame,

option of going straight for

we

don't lose time considering the

it.

Censors are exceptional agents for two reasons. definition invisible. Faithful to their purpose of

our consciousness, they stay carefully outside in the sense

First,

they remain by

unburdening the

it.

field

Second, they are

of

large,

of requiring much knowledge and processing power. For

example, the censor against touching flames should catch than our intention of touching a flame:

it

much more

must detect the many more

circumstances that might lead us to even think of touching a flame. This

watchdog might give

activity requires a close rise to

monitoring of the many agents that

such thoughts. If

new

circumstances occur that lead

us to think about touching a flame, the censor must learn to recognize

them

also. Invisible, large,

and growing, censors

are the black holes

of

thought.

Unlike Minsky's previous contributions about Perceptrons and frame theory, the Society of

Some thought Minsky

Mind was



as

was

given a cold shoulder in AI

his right as

circles.

one of the originators of the

257

SOULS OF SILICON idea

— had simply renamed and repackaged

object-oriented program-

ming. 40 As Minsky pointed out to me, though, there difference

between the two concepts:

which objects-oriented programs called

handle through a protocol

typically

"message passing":

Message passing

basically a foolproof logical

is

that nothing gets lost. If a

up.

a fundamental

is

has to do with communication,

it

However,

message

of Mind

a Society

is

will tolerate errors

agents acting as managers have their

may

express progress, and

way of making through

own knowledge

decide that

sure

not accepted, the system locks heuristics:

how you

about

some other agency

having a

is

bad day and should be made to go away.

Another of the Society of Mind's problems has to do with Minsky's of leaving plenty of work to do for those

strategy

of

his

bandwagons:

applied

it

to the Society

it

It's

too

soft.

a serious proposal," Allen

.

.

The problem

.

climb onto one

when he

of Mind. This time, potential followers thought

do not consider

they just didn't have enough to go on. "I

of Mind to be

who

did marvels for frames but backfired

is

that

the Society

Newell told me.

Marvin wants to

talk

about his

agents in metaphoric language, so you are never able to feel limited they are or whether they are really a in there with

all

Many

critics feel that

interaction of ior.

human. There

the apparatus of a

in that design proposal to enable

many

how

whole bunch of litde kids

you to pin

it

isn't

any constraint

down.

Minsky does not adequately explain how the

simple components results in very complex behav-

Terry Winograd accused Minsky of engaging in "sleight of hand by

changing from 'dumb' agents to in natural language at the point

Minsky, for his part,

is

'intelligent'

homunculi communicating

of Wrecker versus Builder

leery about

becoming more

in a child."

specific.

When

41

I

asked him whether there were ongoing efforts to implement the Society

of Mind into hardware, he answered, "No, not anymore. People want to but they

keep asking

decisions like that.

I

me how

to

do

it.

I'm unable to make design

might mislead them."

Patrick Winston, Minsky's successor at the head of MIT's tory,

AI Labora-

conceded the theory's incompleteness but cautioned against hasty

judgments:

258

Al

Marvin's Society of ideas. I've

of gold

lots

of it it is

is

Mind

not a single idea, but a vast potpourri of

is

sometimes compared in

it,

exactly in

gem form some of it

fool's gold,

it

diamond or gold mine. There's

to a

and there are diamonds and there yet, half

of

it is

hidden

are duds.

in rocks,

None

some of

too low-grade to mine for a long time.

is

There's material for lots of Ph.D. theses there. So as time goes on the Society of

Mind

mark than

it is

become

will

many

in

respects even

more of

a land-

today. 42

Unified Theories of Cognition So

far in this

chapter

may have

I

given the impression that researchers

have definitely abandoned the dream of accounting for mind through a handful of simple mechanisms, the "Maxwell's equations of thought"

upon which Doug Lenat has turned

his back.

Even though

the logicians

have attacked the common-sense problem with renewed and more powerful logical techniques, can account for

all

it is

of thought.

not

at all clear that a single

Scruffies,

kind of logic

through the Cyc project, have

logic down into many special-purpose inferencing mechwhose main purpose is to sift through the mountain of handcoded knowledge where the real power of Cyc will reside. Marvin Minsky's Society of Mind makes up yet another ^//-integrated effort to

in fact

broken

anisms,

explain mind, relying as

does on the sheer proliferation of interacting

it

special-purpose functions.

Other schools of AI researchers original goal in

all its

purity

are,

and finding

in

however,

still

pursuing the

psychology the basis for the

powerful, general-purpose mechanisms with which they hope to explain thought.

The 1980s saw

the emergence of theories with the colorful

names of Act* (pronounced "Act

much

in

common

accounting for

all

in

is

.

.

.

his

and Prodigy, which have

of cognition. Indeed, Allen Newell, showing no sign

of the cancer that would

my chair in

Star"), Soar,

methodology and share the ambitious goal of

kill

him

a year later,

all

but swept

me out of my stuff

enthusiasm for Soar: "In Herbert [Simon] 's and

always the concern for

artificial intelligence

.

.

.

right there with

the concern with cognitive psychology. Large parts of the rest of the field believe that this is exactly the

to I

know whether

wrong way

to look at

it:

you're being a psychologist or an engineer.

AI

you ought

Herb and

have always taken the view that maximal confusion between those

the

way

to

make

progress."

is

259

SOILS OF SILICON

roots of Soar go back to the 1960s and the shift from search to

The

knowledge that Newell and

solution to the

his followers

never quite bought. Although

power of knowledge, they don't

they believe in the

see

as a final

it

problem of intelligence. Further, they point

out, dealing

with mountains of knowledge complicates the search problem by making

it

put

harder to identify the pertinent pieces of knowledge.

"There were attempts to move out of search, to say

it:

knowledge now!' but find

As Newell



lo

and behold

as

soon

one deals with



the search efforts continue."

as

difficult

'It's

all

problems, you'll

Throughout the 1 960s, Newell and Simon continued the experiments had led them to discover means-ends

that

analysis late in the previous

decade: they conducted psychological tests to find out

how people

solve

embody their reasoning techniques in software. much by asking people to prove theorems in logic and later studying how other subjects played chess. The crucial discovery occurred, however, when Newell and Simon asked their subproblems, and tried to

The experimenters

learned

jects to identify digits

known

corresponding to the different

as "cryptarithmetic

letters in

puzzles

problems":

SEND

MORE MONEY

+

Newell and Simon observed that different

numbers more or

less at

their subjects started

random:

this

by trying out

corresponded to pure

search behavior. After a while, the subjects discovered shortcuts that

speeded up result

their searches.

For example, they would

of the addition had more

then the

first digit

M stands for

1

of the

these

efficient use

had

in this problem).

puzzles, Newell and

more of

result

little

digits

to

be the number

Learning

how

fact,

Newell and Simon discovered that the

moved over

time-consuming searches,

a

They would which became shorter

they traded search for knowledge.

more knowledge. Experienced problem at all

(thus, the letter

Simon discovered, simply consisted of acquiring how to make

of them. In

any searching

1

to solve cryptarithmetic

pieces of knowledge and finding out

behavior of their subjects gradually

quired

realize that if the

than the numbers being added,

and arrived

continuum start off

in

which

with pure,

as the subjects ac-

solvers hardly

performed

at a solution quickly by the routine

260

II

application of knowledge. Newell and Simons's key discovery pertained to

how

One

the subjects stored their different pieces of knowledge.

could give a detailed account of a subject's performance, they found out,

by assuming that he or she kept available knowledge fragments long-term memory, more or

of IF

.

.

THEN

.

rules (as

independent of each other,

less

formulated the knowledge

I

reaching the solution for M).

was

It

in the

I

in

form

applied in

precisely this observation that led

Edward Feigenbaum, a former student of Newell and Simon, to implement IF THEN rules into the first expert system over at Stanford .

.

.

(see chapter 6).

At the beginning of the 19"0s, Newell and Simon had analyzed thinking into two basic

abilities:

the

ability*

and the

different solutions to problems;

pertinent fragments of knowledge as IF

to search



that

.

.

THEN

.

is,

and

ability to store

rules, in

out

try

retrieve

order to

speed up the searches. In 1972* Newell and Simon published their findings about the strong role of production systems (another sets

of IF

.

.

.

THEN

rules) in a

name for Human

one-thousand-page book called

Problem Solving which had taken fourteen years to write. ("I'm glad

Simon confided

didn't have to earn a living as a writer!" Herbert

upon remembering different interests,

that

was

it

would become

in a fairly

this period.)

Soar.

left to

As

43

From

Newell to

I

me

to

then on, as Simon pursued refine the theory

often happens in science, Soar

roundabout manner. "The Soar

effort

is

of cognition

came about

an outgrowth of

another project called the Instructible Production System," Newell told me.

That's one of the few efforts I've been associated with that turned into an out-an-out total failure. tion systems

would have

programmed.

\\"e

The

idea

to be educated

w as r

that very large produc-

from the outside rather than

addressed the issue of

how

to build a production

system which could be instructed without knowing in

was

in

grated,

it.

it

The

project never

went anywhere

at all

detail

— but

as

it

all

that

disinte-

gave birth to some major achievements.

These include the OPS-5 production system language; John McDermott's

XCON,

the

first

commercial expert system; and Soar, which

Newell assembled with two graduate students, John Laird and Paul

Rosenbloom. '"What was missing

was an organization

in the Instructible

Production System

for putting tasks together," continued Newell.

261

SOULS OF SILICON

We needed to identify knowledge relevant to particular situations, and things like implementing operators and

do

was when we

functions. This

to be central in

represent

implemented

it's

Soar:

it

was the

this

through the principle of universal sub-

when you run out of

a device for recognizing

providing a

new

As

example of subgoaling, suppose Soar

a simple

memory would moves

like

tell it

what moves

the game, though, Soar

allowed moves, without

up

Soar

new

as a

available.

the

full

.

.

are legal

.

is

playing ticktack-

THEN rules within Soar's

and perhaps suggest useful

may have its

to

some

stage in

choose between two or more

rules expressing a preference for

rises to this situation, called a "tie

goal the selection of the best

any

move

impasse," by setting

move among

the alternatives

This selection becomes a separate subproblem, upon which

problem-solving power of Soar can be brought to bear.

"You "They

A set of IF

blocking the opponent whenever possible. At

in particular.

gas and

opportunity for more knowledge.

human opponent.

toe against a

we could own problem

realization that

these other tasks simply as searches in their

all

We

spaces. goaling:

making

computing evaluation

discovered the gimmick that turned out

see, that's

the magic of production systems," Newell said.

are self-selecting systems, in

which the

rules themselves say, 'I'm

relevant to that situation.' " Soar contains standard rules for resolving

come

impasses, which as is,

out

situation.

soon

as a situation has

tie

been tagged



would tell Soar to look ahead that moves and see which leads to the most desirable This strategy would remain the same whether Soar were

an impasse. In try

into play as

all

this case, the rules

the

playing chess, checkers, or ticktacktoe: therefore, universal subgoaling potentially allows Soar to play

any game.

Universal subgoaling was the subject of John Laird's doctoral dissertation;

44

Paul Rosenbloom's thesis

mechanism chologist

for learning called

George

initially

concerned not Soar, but the

chunking suggested by the Harvard psy-

Miller during the 1950s in his

famous paper on short-

term memory. 45 Chunking consists of tying existing notions into a new bundle that

itself

we chunk seven In his thesis,

becomes

digits

a single notion:

we do

it

for

example when

under the heading of a person's phone number.

Rosenbloom was

able to provide substantial evidence for

the presence of chunking at the heart of learning. 46

He skill

did

it

as follows.

— whether

it's

We all know that the only way to improve a new

typing, skating, or speaking a

new

language



is

262

Al

through practice. Since the 1960s, experimental psychologists had been

which we perform the new task some fractional power (say, the square root) of the number of times we've done it before. They called this phenomenon the "power able to determine that the speed with

increases as

law of practice" but were if

at a loss to explain

learning occurs by chunking, and if

constant

rate,

then

improve

skill will

it.

Rosenbloom showed

we perform

in a

manner

this

chunking

dictated by the

that at a

power

law of practice.

When

these results were

somewhat

in,

Newell, Laird, and Rosenbloom realized,

to their surprise, that Soar provided a ready-made

was only necessary

for incorporating chunking. It

As

I

indicated, these

happened only

of pertinent knowledge was

available. In the ticktacktoe

when none of its

reached an impasse

where no

in situations

rules could

mechanism

to exploit impasses.

tell it

direct piece

example, Soar

which move to

make. Even more to the point, Soar already had the means for discovering the missing piece of knowledge. After evaluating the available ticktacktoe moves, for example,

it

knew which one was

for Soar to have the ability to learn, this

knowledge by means of

new IF

a

described the situation that had given

and prescribed the action taken

best.

Thus,

order

in

had to be enabled to remember

it

.

.

rise to

as a result

.

THEN

This rule

rule.

the impasse (the IF part),

of the ensuing search

in the

THEN part. For example, a new ticktacktoe rule might say: "IF you can play either a corner or the center, THEN take the center." In later know which move

games, Soar would

to make,

and no impasse would

occur.

Newell recalled to

I

couldn't believe

still

how

easy

it

turned out to be; as he

me:

was out of town when John and Paul decided

learning for one

discovered

how

little

to

task.

make

it

And when

general.

to try

and program-in

they implemented

So when

I

came back

it,

after

they

two

or three days, chunking was in and working in a general fashion

through that tight

all

of Soar.

had

a

few breakthroughs It

in

my

career, but

brought everything into a very

knot in which problem spaces, production systems, impassing,

and chunking are a

I've

was the most dramatic one. just

one

ball

complete cognitive engine.

of wax. All of

a

sudden Soar became

263

SOULS OF SILICON

These events happened

in January 1984. Since then,

into a multidisciplinary research

dred researchers on both sides of

same

has proposed

tide,

Maxwell

in his equations explained

all

of cognition." 47

of electromagnetism

four quantities (charge, current, electric and magnetic

a

hun-

book of the

the Atlantic. Newell, in a

as a "unified theory

it

Soar has evolved

program involving more than

in

Much

as

terms of

fields),

Newell

hopes to eventually account for the whole of human cognition through the four fundamental

Among

mechanisms implemented

Anderson's Act*

48

and a system

called Prodigy

in Soar.

most notable

John by Steven Minton, Jaime

other architectures similar to Soar,

are

Carbonell, and others. 49 Both Act* and Prodigy posit a learning mecha-

nism

which

similar to Soar's,

called "explanation-based learning."

is

Contrary to Soar, which can learn only from

from

also learn

its

mistakes. Since

its

Anderson

successes, Prodigy can is

a Carnegie

Mellon

psychologist, and Carbonell one of Newell's colleagues at that university's

computer science department, these projects obviously sprouted

off the

same

chapel of

intellectual branch.

AI

Together, they

make up

a substantial

research.

The Question of Awareness: Could Machines Love? Whether through the worldly knowledge of architectures

future

modeled on our own minds,

may comprehend

own. This

raises the question

attributes with us: Will feel pain, love,

On

the world in a

I

whether they

is

to

compare

its

it

our

other mental

as

one

studies

inability to define

to establish

its

presence in

human

behavior.

when we succeed in embodying

in machines, since relatively simple logic or association

mechanisms can make intelligence.

way

human

behavior with

Further, intelligence seems to disappear aspects of

will share

one can study emotions

have already discussed the

system

artificial

machines of the totally unlike

and anger?

intelligence in absolute terms: the only

an

manner not

such machines be aware of themselves? Will they

a philosophical level,

intelligence.

a Cyc-like data base or

intelligent

There

are

a system

behave

two ways

as if

it

did contain a piece of

to consider this

stems from the position that intelligence

is

phenomenon. One from

in essence different

264

M

matter, a kind of immaterial fluid breathed into a being.

consequence of

this position

is

to dismiss as pure

successes at imitating intelligence. If intelligence

is

The

of

life

plex

intelligence. In this case,

we

with mechanical dolls,

mechanism

The other

much

as

we

partial

in essence different

from matter, and we don't put any into our mechanism, then be imitating

logical

mimicry our

can onlv

it

can't imitate

aspects

all

probably couldn't build a more com-

to imitate the entire mind.

interpretation of the seeming disappearance of intelligence

takes a reverse view of the matter. The fact that one can build pieces of mind out of nonmind substance like computer circuitry may just show that mind naturally emerges out of properly organized matter. Indeed, it is

a

widespread property

in nature that

complex phenomena emerge

out of simple interactions and components. The simple forces between ice molecules, for instance,

induce an

infinite variety

hibit properties totally absent in individual

doesn't

make

of

intricate

components. For example,

it

sense to talk about the temperature, entropy, and pressure

of an individual molecule, yet these properties emerge out of the tive

snow-

groups of interacting components often ex-

flake structures. Further,

collec-

behavior of large numbers of gas molecules. This interpretation

is

much more encouraging

for the future of AI; and, indeed,

researchers subscribe to

Many of them claim that one good reason we is that we all carry one in our skulls

can build

a thinking

it.

most AI

machine

hence, the "meat machine" quality of the brain Marvin Minsky

is

fond

of invoking. Is the brain truly a

one's immortal soul physical body?

A

machine? Aren't one's conscience, one's thoughts, if it

comes

consensus

is

to that, essentially different

building

among

scientists

from one's

and philoso-

phers that they are not. Most of these experts subscribe to the opinion that the

mind

is

essentially

an emanation of the brain. In other words,

physical processes occurring in the brain explain our mental processes

Many arguments justify this opinion. One is anatomical evidence for the association of mind and brain. Damage to the brain disturbs our thinking. Damage to specific areas in their entirety.

even gives

rise to specific

correspond to abstract

kinds of disorder. Particular brain structures

abilities,

such as language. For example, damage

to either Broca's or Wernike's area,

both located below the

can produce permanent and specific language

problem of our drug-abusing society

disabilities.

left

And

temple,

isn't

one

precisely that specific chemical

changes in the brain correlate with changes of

moods and emotions?

265

SOULS OF SILICON

Another argument for mind-brain association stems from the history

We

of science.

have not, so

discovered anything in nature which

far,

physical law cannot explain. It

would be surprising

exception. If fact, the entire history

phenomena

of science

if

the

mind were an

one of explaining

is

previously thought to be governed by forces extraneous to

Wind, lightning, volcanoes, and earthquakes are no longer expressions of the whims of gods, but strictly material phenomena governed by natural law. We can now make mathematical the physical universe.

models of the evolution of the universe, bang.

The laws of physics and

all

way to

the

the primordial big

chemistry, together with the

of natural evolution, explain the appearance of

life

on

mechanisms

earth.

They

also

appear to account for the evolution of humans.

humans

In a parallel phenomenon, science has gradually displaced

and

their physical

environment from

their special status in the world.

Copernicus started the movement by moving the earth from

human body from

its

pedestal and proved

ancestors. Biochemistry

natural

showed

its

lineage with

human mind, showed

that

was

it

at least

even our minds have purely material If such

is

amenable to

last step in this revolution: its

the case, shouldn't

it

central

nonhuman

that the life processes in this

phenomena. Freud, without proving the natural

would be the

its

Darwin displaced the

position to being a satellite of the sun. Later

body

are

origin of the

scientific study.

success

would show

AI

that

origins.

be possible to re-create, out of inert

matter, beings with not only thought but also awareness, feelings,

and

emotions? Although the opinions of philosophers and AI researchers cover the usual spectrum of diverging views, find any creditable expert

I

have been hard put to

who would answer

a clearcut no to this Most responded with a qualified yes; and in the past decade, most debates on the subject did not oppose the partisans of yes and no answers. The action centered on the yes part of the spectrum, with

question.

factions arguing over

whether they should qualify

their positive answers,

and how.

The

closest approximation to an

Berkeley's

Hubert Dreyfus,

machines cannot be made

who

uncompromising no comes from

simply claims that truly intelligent

in the first place.

And

even on that point,

Dreyfus's objections stop at symbol-manipulating machines: he

committal about machines based on

non-

neural nets, presumably

way our brains are built. 50 somewhat less hard line, MIT's Joseph Weizenbaum, author

closer to the

Taking a

artificial

is

266

||

of the anti-AI book Computer Power and Human Reason,^ summarized

his

position for me:

I

don't see any

that

way

why

can't understand

develop in

be

of

to put a limitation to the degree

machine] could acquire. The only qualification

[a

way

this

resisted,

it's

will

intelligence

make, and

I

that the intelligence that will

is

always be alien to

at least as different as the intelligence

human

I

human

intelligence. It will

of a dolphin

is

to that of a

being. 52

Weizenbaum

points out that dolphins are different

comparison leaves open the

— not

of creating

possibility

unfeeling. This

machine with

a

As he himself said, "I don't understand what [my position] takes away from any ambition that the AI people might have." For an all-out yes, consider Carnegie Mellon's Hans Moravec, who feelings.

we

believes that early in the next century

generations of gradually

more

He

told me.

be served by successive

intelligent robots.

"On

chines as anything but unfeeling aliens. in general will

will

He

views these ma-

the contrary,

I

think robots

be quite emotional about being nice people," Moravec

explained this character

trait as

the result of evolutionary

forces:

Imagine these robots being made in a factor} and the main purpose 7

,

of the factory, which robots to

sell well.

companies, which response that

when you person

it's

is

the reproductive unit of the robots,

will

then build more factories.

bring one into your house, there for, and that

it

a positive conditioning if

sitive

feel

it

will

it

understand that you're the

it.

It

should

is

You that

somehow

its

internal

at least

estimate

make-up]

does something that makes you happy.

about

its

actions. It will try to please

manner because

reinforcement.

Moravec's point

customer

it is

had better keep you happy, or

induce you to buy another one of

apparently selfless

Then

mostly shape the character of these robots. So

will

how you

to cause

best-selling robots will bring in profits to their

whether you're happy or not, and receive [from

will care

is

The

it

will get a thrill

you

out of

in

this

It

an

po-

can interpret that as a kind of love.

emotions are

just devices

for channeling

behavior in a direction beneficial to the survival of one's species. The

most

basic emotions are love,

which

stirs

us toward certain goals, and

267

SOILS OF SILICON

which keeps us out of harm's way. Moravec believes robots

fear,

experience fear also. for his future

Tongue

will erase

may

and

run it

down

a

will

minor emergency

try desperate

to nothing, because then

will forget all sorts

consequences. So

terrible, terrible it

he imagined

household robot:

It can't let its batteries

memories

in cheek,

if it gets

measures to obtain

all

its

of important things with locked out of the house,

a recharge

somewhere.

Its

emergency modules would come into

play, it would express agitation, humans can recognize. It would go them to use their plug saying, "Please!

or even panic, with signals that to the neighbors

Please!

I

need

and ask

this!

It's

so important,

it's

such a small cost! We'll

his

robot carts calibrate their

reimburse you!"

He

also sees in the

maneuvers by which

young animals: such on which will depend the life-saving Marvin Minsky has, in The Society of Mind, of emotions in which he explains them as

vision sensors the forebears of playful behavior in activities serve to

fights or flights

sharpen the

of adult

life.

carried out a deeper analysis

skills

special-purpose cognitive devices. Fear, anger, and pleasure act as short-

term attention focusers. For longer time spans, "liking" holds us to

we ought to underdown our universe." 53 Minsky even sees a clear cognitive role for humor and laughter, which play "a possible essential function in how we learn": "When we learn in a serious context, the result is to change connections among ordinary agents. But when we learn in a humorous context, the principal result our choice: "Liking's job stand

is

to

sors."

its

is

shutting off alternatives;

role since, unconstrained,

it

narrows

change the connections that involve our censors and suppres54

Freud had already understood the

role

of censors and suppressors

as

inhibitor)7 agents responsible for preventing harmful or socially unaccept-

able behavior, but Minsky extends their action to the cognitive realm: in common-sense reasoning, censors and suppressors must recognize trains of thought that lead to absurdity or infinite recursion, and make us avoid

them in the as

touchy as

future. sex.

This

is

why, to our subconscious, the absurd

Absurd and sexy jokes are funny because

surprise us with forbidden outcomes,

during which censors against these

and laughter

new approaches

is

is

almost

their punchlines

the

mechanism

are built. "In order to

construct or improve a censor," Minsky explains, "you must retain your

268

Al

records of the recent states of thought. This takes

some

are fully occupied." 55

mind

For Minsky, laughter's function

the censor- formation process. If

next time around, and the joke

To

all

and

this,

think the censored

which your short-term memories

on these memories and ensure

attention focused

nomena,

made you

that

time, during

all

works

to

is

keep your

that nothing interrupts

well, surprise

is

avoided the

no longer funny.

is

one might object

emotions are biochemical phe-

that

that mood-altering drugs can induce depression or ecstasy,

have demonstrated that emotions are related to

that biochemists

minute variations of neurotransmitters

in the brain.

How,

then, can a

machine made of transistors and wires feel anything like an emotion? The answer is that neurotransmitters are just handy devices for inducing special electrical behavior in the brain. As their name implies, neurotransmitters are chemicals that selectively alter the conductivities

of synapses, the connections between neurons. Relaxants function by switching activity

make you

on whatever brain

of brain

Dedicated hardware or software con-

feel relaxed.

could achieve similar effects in machines:

trols

Valium

like

circuits or patterns

meant by switching on an "emergency module"

this is

what Moravec

to induce panic in his

discharging robot. "All right," the skeptic might say at this point,

can make a machine behave,

when observed from

acting in an emotional manner.

I

will

"I'll

believe that

outside, as if

it

you

were

even admit that such behavior

is

Howwhen Moravec's home robot

necessary for a machine's intelligent interaction with the world. ever,

I

absolutely refuse to admit that

pampers you,

feels

it

business assigning late

human

many

love or even affection.

states to

We

have no

machines that cleverly simu-

is

known

to the philosophically inclined artificial intelli-

"weak AI," and they have fought

private

strong

like

behavior modes."

This position gentsia as

anything

human mental

and public debates over the

AI hold

that intelligent

believers in "strong

last

AI"

machines can be imbued with

awareness, consciousness, and true feelings.

in

two decades. Adepts of

The

issue

may remain

self-

for-

ever uncertain since, contrary to behavior, internal states of mind cannot

be assessed objectively.

Although

at first the

one could argue

reasonable reaction seems to be

that for

"Who

cares?"

moral reasons such questions should be ad-

dressed and resolved before

we

example, what should be the

ever build truly intelligent machines. For fate

of your future loving but battered

269

SOULS OF SILICON robot nanny? Wouldn't

ponder such

it

mass-produce robot brains and wire them

Wouldn't a robot personality

vations.

unhappy

quite

The most

in repetitive factory

no

ness are

classic

perform

as a tour guide

be

AI came from

the Berkeley philos-

claims that computer simulations of aware-

made anyone

wet.

Room"

1980 "Chinese

He

drove the point

home

Chinese. Searle, in a closed

who

is

totally

room and

priate responses,

all

(a

test in

ignorant of the Chinese language, would

receive the judge's questions as Chinese ideo-

grams on pieces of paper. language instructions

in his

paper, 56 which described a thought

experiment in which he simulated a computer passing Turing's

sit

fact,

with the same basic moti-

all

to

should, in

turn out to be cruel to

closer to the real thing than their simulations of thunder-

storms, which never

now

who

Searle,

fit

may

work?

vivid attack against strong

opher John R.

We

be murder to junk her?

it

issues at the design stage:

He would

"program")

in terms

then consult a set of English-

telling

him how

of Chinese symbols

to

compose appro-

unintelligible to him.

After working out his answers in this way, he would hand them out as

ideograms on paper.

Now,

to

the room, the being inside nese. Searle, however,

it

one observing the procedure from outside

would

clearly

appear to understand Chi-

would have been merely following formal

he would have remained

totally ignorant

rules:

of the meaning of his answers

and unaware of the mental processes involved

in

working them

out.

He

concluded that mere symbol manipulations, even were they to generate outwardly intelligent behavior, could not induce awareness in the mech-

anism performing them. Searle's

opponents answered by claiming that

indeed emerge from these

activities,

the system consisting of the

processing the

man

activities.

in the

to point out,

create a

new

practice,

it

but that

it

a

new awareness might

would be associated with

room, the program, and the man's symbol-

Since this consciousness would be external to him,

room would remain unaware of it. As does seem a

bit ridiculous to

Searle

was quick

argue that a person could

consciousness by merely shuffling around bits of paper. In

how

however, the instructions explaining

to generate the an-

swering ideograms would amount to a million pages of printed

which

is

the quantity of knowledge a full-fledged

text,

common-sense data

base would contain. Further, in order to carry out the simulation at a speed approaching real time, the

to

perform long-hand calculations

poor

mug

in the

at the rate

room would have

of one hundred

billion

270

Al

operations every second, the speed at which our brains process infor-

mation for

wind of

us.

activity

Outside of

its

sheer physical impossibility, this whirl-

makes the appearance of an awareness somewhat

less

preposterous.

My

favorite

Room"

counterargument to the "Chinese

paper was

offered by the philosophers Paul and Patricia Churchland in 1990. 57

Suppose

that, instead

of trying to show that symbol manipulation and

man

awareness are unrelated, the onstrate, falsely, that light

He

with each other.

pumping

in the

could achieve

magnet up and down

a

room were attempting

at

this

catch here

is

dem-

by darkening the room and

arm's length, thereby generating

electromagnetic waves of a very low frequency.

permeating the room would

waves of

to

and electromagnetic waves have nothing to do

let

him

The

pitch-darkness

simply that to generate any light (that

a frequency perceptible to the eye), the

is,

still

The

infer the desired conclusion.

electromagnetic

man would

have to

speed up his pumping rate by about fifteen orders of magnitude

— or

about the same factor by which he would have to accelerate

just

paper shuffling in order to generate an observable awareness

his

in the

Chinese Room.

What of Searle's claim about the difference between simulations and What difference is there between a computer performing operations that appear to endow it with awareness, and the same comthe real thing?

puter simulating a storm in a weather-forecasting center? Is the computer's awareness any

more

real

computer's intelligence certainly

around raindrops, surely isn't doing.

air

the "essence" of

one

really

However

The "essence" of a storm

the "essence" of intelligence precisely

is

is

this

the Chinese

to

is

what the computer does.

awareness also the manipulation

knows, but

more than

is.

to whirl

molecules, and lightning, which the computer

information, and this

late

than the storm? Well, for one thing the

manipu-

Now is No

of information?

argument doesn't show us otherwise, any

Room

does.

Indeed, neither Searle nor other proponents of weak AI have pro-

vided a definition of the nature of awareness that would satisfy anyone.

For by a

their position begs the question: If the manipulation digital

computer cannot generate

take to do it? phenomenon: mental

Searle himself answers that "cognition

it

cesses. This it

states

of information

true awareness, then

and processes

are caused

is

what does

a biological

by brain pro-

does not imply that only a biological system could think, but

does imply that any alternative system, whether made of

silicon,

beer

271

SOULS OF SILICON

cans or whatever, would have to have the relevant causal capacities 58 equivalent to those of brains."

consist of

What these

mysterious causal capacities

exactly, Searle refuses to specify, except to

claim that parallel computers or

add the further

neural-net systems

artificial

would not

possess them either.

On

an argument proposed by the philosopher Zenon

this point,

Pylyshyn raises puzzling questions on the boundary between biological

and electronic systems. In

by electronic

this

thought experiment, Pylyshyn supposed

more and more of her brain cells are replaced components with identical input-output functions, until

that as a person

is

talking,

the entire brain consists of integrated circuit chips. In

person would keep on acting in is

right,

she

would

at

just the

some point have

"meaning" anything with her words.

all

likelihood, the

same way except

lost her

that if Searle

awareness and stopped

Somehow that doesn't sound

quite

right."

Daniel Dennett drew yet another argument for strong AI from the theory of evolution: causal

if

systems with and without Searle's mysterious

powers for awareness

can't

be told apart by

their behavior, then

they have exactly the same survival value. In this case, evolution would

have absolutely no incentive for developing such a superfluous mecha-

how did it come about and maintain some chance mutation had robbed our ancestors of awareness, we would be acting exactly as we are now, claiming to be

nism

itself?

as "true" awareness. If so,

Indeed,

aware, except

if

we wouldn't

be.

60

Wouldn't that be downright

Assuming, then, that one can define "true" awareness

silly?

in terms

of

pure functionality, what are the specific functions a system should

embody these

in order to qualify? If the

proponents of strong AI are

would be the same functions

Turing's

test.

that

would

let

a

right,

machine pass

But can some of the philosophical steam generated

in the

weak-strong AI debate allow us to point out some feature of those

mechanisms? In other words, can philosophers provide engineers with design guidelines for conscious machines?

Perhaps they can. As I've hinted, philosophers draw a strong link

between awareness and the

John

Searle rests his

ability to give

whole case on the

meaning

fact that

to symbols. Indeed,

symbols manipulated

meaning only when interpreted by humans. In Searle's parlance, symbol-manipulating computers possess syntax but no semantics. Hans Moravec, for his part, was trained as an engineer, and

by a computer acquire

a

his bottom-line conclusions

on machine awareness

are entirely opposite

272

Al

to Searle's.

Yet Moravec agrees with Searle on

'Today's

this point.

humans to front for them," Moravec admitme. "They need somebody to tell them what's in the world, and on what the programs say. In essence, to give meaning to the

reasoning programs require ted to to act

abstract symbols the

programs manipulate." Moravec however sees

a

straightforward engineering solution to this philosophical quagmire: "If

we

could graft a robot to a reasoning program,

person to provide the meaning anymore:

Moravec

physical world."

ence to the world

is

also believes that

we wouldn't need

a

would come from the some kind of sensory referit

required for a machine to pass Turing's

test:

Human communication is only language on the surface. What's below the language are these perceptual models of the world containing

mythical allusions and pictures and emotions. There bal

machinery

in

our heads.

probing for these things, asking,

"Here

is

a situation,

how

is

much

nonver-

A really insightful Turing judge would be

does

it

"How do strike

no more compact way of encoding analogous to the actual structures

you

you?"

that

feel

And

I

about

this,"

believe there

or is

machinery than something

we have in our brains, including our

brain stems and limbic systems.

And, indeed, since the mid-1980s, for reasons not unrelated issue raised

by Moravec, an

has concerned

itself with

to the

of the AI research world

influential faction

coupling their programs ever more closely with

the physical world. Moravec belongs to this school of thought, and Rodney Brooks of MIT is its most colorful figure. He, in fact, takes the

extreme position that reasoning that

mechanisms akin

suffice to explain

it.

to those

is

not necessary for intelligence, and

of reactive or inborn behavior

His tiny six-legged robots, which mimic

behavior of insects, are Brook's

first

in animals

much of the

61 step in proving his point of view.

In respect to design rules, philosophers also point to self-consciousness as

another crucial component of awareness. Daniel Dennett notes,

though, that "self-awareness can simplest, crudest notion

mean

several things. If

of self-consciousness,

the sort of self-consciousness that a lobster has:

something, but

between

itself

it

never

and the

eats itself. It

rest

I

When it's

some way of

has

of the world, and

it

you take the

suppose that would be hungry,

it

eats

distinguishing

has a rather special regard

for itself." 62

That kind of bodily awareness has been

instilled into

robots from the

273

SOULS OF SILICON beginning. For example, Marvin Minsky's well

had to have

blocks

hand

it,

in order for the

was manipulating.

it

in the image,

whether

it

would move

was

really

it

When

hand-eye robot pretty

it

hand from the

the

tell

the robot thought

had

it

identified

its

slowly in front of the camera to see

chapter

itself (see

first

camera to

4).

A higher level of self-consciousness corresponds to introspection: the some of

ability to inspect

own

one's

mental

Most AI programs recommen-

states.

require this ability in various degrees, if only to justify their

dations to their users. All medical expert systems, starting with (see chapter 6)

have had the

certain treatment: for this, they

by retracing

SHRDLU

could explain

exhausting a long

me

asked

chapter

they prescribe a

have to examine

their

own

motivations

a further example, Terry

why

Winograd's

it

had manipulated

of intermediate objectives,

certain blocks: after

gave "Because you

it

to" as the ultimate justification to a sequence of

4).

Marvin Minsky, for one, turns the

we

out that

list

why

As

their reasoning.

MYCIN

ability to explain

table

moves

(see

around and points

could design machines better equipped than our brains to

monitor themselves, thus making them more conscious than we

are.

63

Perhaps the major disappointment of AI research to those of us schooled in traditional Western values was to evacuate the substance out

of

intelligence: if

source of

its

you take apart an AI program and

cleverness, you'll see

it

locking subprocesses, in themselves

try to trace the

disappearing into a all

quite

trivial.

maze of inter-

Recent forays into

cognitive architectures indicate that our consciousness and sense of identity

may

When

came

I

well participate in the

same disappointing evanescence.

across the following paragraphs in The Society of Mind,

couldn't help but feel that Marvin

Minsky was

letting

sections appear far apart in the book, but together left-right

punches against consciousness and

we

I

me down. These

make up

devastating

identity:

little more than menu lists that on mental screen displays that other systems use. It is very much like the way the players of computer games use symbols to invoke the processes inside their complicated game machines without the

[W]hat

flash,

call

"consciousness" consists of

from time

slightest

to time,

understanding of

[Ojur brains appear to

how

make

they work. 64

us seek to represent dependencies. Whatever

happens, where or when, we're prone to wonder .

.

.

But what

if

who

or what's responsible.

those same tendencies should lead us to imagine things and

274

Al

causes that do not exist?

hand

see their

Then

we'll invent false

gods and superstitions and

every chance coincidence. Indeed, perhaps the strange

in

word "I"



If you're

compelled to find some cause that causes everything you do

as used in "1just

had a good idea"

then, that something needs a name.

You



reflects the selfsame tendency.

call

"me. "

it

I

call

it

you.

— why, '* 5

Consciousness as an arcade player and the "I" as a figure of speech,

Minsky wouldn't give

indeed! In person,

whether he thought the It's

a

self

complicated set of

involved in

it.

One

is

about two years old

person. ...

By

concepts of

self.

me

it

any reprieve.

illusions, there are

many

complex

can have emotions, dismayed

body, and

explain

particle,

why the

the child a

it is

as illusory, superficial explana-

me

who

believes that machines

by agreeing with Minsky:

the "I" the Center of Narrative Gravity.

of gravity was

when

to be accounted for in a convenient way.

center of gravity of this object

not a

different processes

usually gets this wonderful idea that

Daniel Dennett, the very same philosopher

it's

asked

of correcdy recognizing that

at first the I is a

But Minsky considers these concepts

call

I

the time you're adult, you have a dozen different

tions of processes too

I

When

he answered:

illusion,

just the inference

people are also objects. So is

an

is

stapler

on

Now

his desk]. It's

consider the

not an atom,

an abstract point in space. But you can use

it's

stapler tipped still

[a

back when

over a point

I tilted it:

in the

it

to

because the center

supporting base.

Now,

are

centers of gravity real? In one sense they are, in another sense they aren't.

That

is,

the center of mass of an object

just a

is

very convenient

way of organizing what would otherwise be hopelessly messy data, but where would we be without it! Now the self, the "I," is the center of narrative gravity for [our discourse about human beings]. And, boy, does

it

We

help!

human body flinging about, talking: it just so much [random] motion, if we of center of narrative gravity, that agent who is see a

would be incomprehensible, couldn't posit a sort

body

responsible for these words the think this agent

is

the brain which

is

in the

agent

the

Oval Office is

a

is

uttering. It's a mistake to

a point in the brain. There's self,

and

it's

in the brain.

a mistake to

not any one thing

in

look for the President

But the idea of there being

wonderful way of organizing psychology.

a single

275

StCLS IF SILICM

True enough,

I reflected,

through

myriad mechanisms,

its

constructive

upward

ments blend into a

direction?

single

why not

As

symphony, couldn't the

Mil

put the question to

try to trace

s

it

the separate notes of

emerging from the concerted

entity I

but being useful doesn't necessarily make a

of speech wrong. Instead of trying to follow the

figure

activities

down

self

more

in the

many

instru-

self truly exist as

an

of a multitude of agents?

Minsky and to Tuft's Dennett. Minsky

wasn't encouraging:

That's what :

many people believe, and I don't think so. I think the self we make it it's not an emergent; it's an

doesn't emerge. In fact,

afterward construction. In general, emergents don't produce anything that clever. ... If

to

like

it,

it

were an emergent, you couldn't attach properties

"I'm handsome." Because to do so you need a small

representation, a

little

symbolic "I" that you can attach properties

to.

You can't attach them to a vague emergent: it doesn't have any hooks. There

is

beehive

you

something generally wrong with the idea of emergence: starts to

swarm

can't attach to

it.

in a certain direction, that's

It can't store

if

a

an emergent, but

any knowledge because

it

doesn't

i solution to a (inferential equation.

actually

-.hough Daniel Dennett was willing to label our feeling that there's "somebody home" in our heads as an emergent phenomenon, he hastily qualified his opinion:

Some people

use the term emtrgna in a mystical way. Emergent

properties are supposed to be very mysterious that

you

Coast Emergence."]

pedestrian, but I think

of description. In non:

special properties

can't explain with science. [Dennett prfvatefy calls this

"Woo Woo West more

and

many

mean

useful, sense

this sense, a traffic

different

into their cars,

more

I

jam

view

m gemf in the much

ej

/

of a convenient level

an emergent phenome-

is

people make semi-independent decisions to get

and suddenly you get the

"traffic

jam" phenomenon.

And you really want to talk about it at that leveL Don't try to "reduce" talk

about

traffic

jams to talk about behaviors of just individual

motorists, even though that's all they really are. A traffic jam is just one god-awful combinanon of individual motorist behaviors. But it is

also

emergent in the sense that there are

jams which are best described

regularities

at the "traffic

about

jam" leveL

traffic

276 As

it

k\

turned out, then, what Dennett means bv calling the self an

"emergent entity"

is

concept of

a convenient figure

self

is

fundamental agreement with Minskv after

in

of speech



the

all:

so convenient as to

be indispensable.

On

this question, I

Somehow,

found more solace

had expected to meet the

I

jreen Tufts

campus and MITs

less

in Pittsburgh than in Boston.

down-to-earth attitude in the

lofty hallwavs, the latter

once de-

scribed as a cross between the Pentagon and the Vatican. 6* Yet

surprised and

somewhat comforted by the

attitude

of researchers

I

was

in Car-

Andrew Carnegie's utilitarian former Carnegie they didn't make any case for teducing human awar.

negie Mellon University,

Tech. At

least,

to a figure of speech.

put the question to the psychologist fames

I

McClelland who, with a group of San Diego colleagues in the mid1980s, resurrected neural-net research by publishing a two- volume col-

of groundbreaking papers called

lection

Parallel Distributed Processing^

PDP

showed how, by extending Rosenblatt's Perceptrons to manylayered structures, they could be made to overcome most of the failings Minsk}- and Papert had pointed out in their 196" book." Bearded and relaxed,

John McClelland

reflection

when

a slow speaker.

is

He

enters a state of deep

a question requires consideration.

Eyes closed, he tends

to hold his head with both hands and, slowlv intoning, appears painfully to force his answer through as if hauling

it

out of a deep well.

He

discussed the Society of Mind's relevance to neural networks:

Minsky's notion that the

I is

an

illusion misses the point that there

a fundamental coherence to the mental state. There are

and

fifty

United that.

million people

States,

So

I

going about their daily business in the

and you don't get a sense of

think the Society of

might be that the ent.

all

states

is

two hundred

Mind

is

a collective

mind out of

missing something, which

of a neural network are

actually quite coher-

When you have large numbers of simple computational elements

that are massively interconnected with each other, they can't just

about doing whatever they basically

governed by what

So the human mind

is

damn all

well please.

the others are doing at the

much more

population of a country.*

go

What each one does same

is

time.

coherent than the distributed

9

Over in Carnegie Mellon's computer science department, Allen Nemade similar remarks. I tried out on him a notion of mine that one

well

277

SOULS OF SILICON

could draw parallels between Soar and Minsky's Society of Mind.

Weren't Soar's complicated IF

Minskyan agents,

all

all

his

.

.

THEN rules somewhat equivalent to

own, which

own

specialized ways?

wrong," Newell answered

"I think you're just plain

manner

.

interacting in their

found disconcerting

I

after

vehement

in a

McClelland 's slow

pronouncements.

A

production

Minskyan



agent.

a rule, if

you

will

Minskyan agents



much

is

are in fact

too simple to be a

little

homunculi

by

all

themselves. Also, productions don't direcdy communicate with each other: they just

could

talk to

all

look

another

at the

common

perceive in this

same

data.

by making

is

a

The only way

a

production

change that the other could of

data: this places a lot

on

restrictions

communication between productions. So the Society of Mind and Soar are profoundly different in two ways. agents

is

much

larger in the Society

agents to communicate with each other absolutely unclear to

First, the grain size

of Mind. Second, the is

much

of the

ability

smaller in Soar.

of

It's

me how you can produce in that Society of Mind

any sort of integration.

The paradigm on Soar tion rate high little

enough so

is

much more how you get

that

agents knowing separate different things.

communica-

A community of scien-

with telephones in their hands, cannot produce an integrated

tists, all

intelligence. tive to

the

you don't have the problem of separate

The

available rate

of communication between them

what they've each got

problem of producing beings

in their like

memories

too small.

is

you and me, which

rela-

The

are highly

integrated, requires an architecture that's very responsive to that I believe that Marvin is on the wrong side wrong region of the communication space with

communication problem. of this chasm,

in the

respect to the integration issue. Marvin's view

of ways

in

which we

certainly right, but

I

is

that there are a lot

are dis-integrated, unintegrated.

think that he just

accounting for the huge amount of integration that

don't have any objections to being

I

ness

more than an

tions

marching

for

it.

shown wrong.

illusion required the victory

in lockstep

And

he

is

doesn't even come close to

we

If

have.

making aware-

of an army of produc-

over a rabble of individualist agents,

I

am

all

278

Al

AND RELIGION

Al

Most of

us are perfectly willing to accept scientific explanations in

matters physical, yet

One

material origin for our minds.

human

erosion of

all

uncomfortable with the idea of an essentially

feel

reason, as I've said,

is

the potential

values inherent in this belief, which led

some

re-

searchers to dismiss consciousness and feelings as illusions. Another of

our discomforts with the concept religious beliefs.

arises

from

its

potential conflict with

Doesn't the materialistic view of the mind contradict

the existence of an immortal soul, different in substance

Perhaps, but as

show,

shall

I

this

from the body?

opinion doesn't clash with Western

religions.

These

verses, for example,

that Judeo-Christian tradition

ation of

will

Of those who

lie

life,

prophesied as

noise, a

come

to

in the dust, for

lie

everlasting

I

is

not inconsistent with an intimate associ-

mind and body:

Your dead

who

from the Old Testament, 70 seem to imply

life, .

.

.

their corpses will rise; awake, exult,

sleeping in the dust of the earth

some

I

to

all

of you

the land of ghosts will give birth. Isaiah 26:19

shame and

many

will

awake, some to

everlasting disgrace. Daniel 12:2

had been ordered. While

was prophesying, there was a I looked, and saw

I

sound of clattering; and the bones joined together.

that they

were covered with sinews;

flesh

was growing on them and skin was

covering them. Ezekiel 37:7-8

And you will know

that

from your graves,

my

And

from the

this

It is

the

I

am Yahveh, when I open your graves and raise you

people. Ezekiel 37:13

New

Testament: 71

same with the resurrection of the dead: the thing that is sown is Howbeit that was not first what is raised is imperishable.

perishable but

which

is

spiritual.

spiritual, 1

.

but that which

is

natural;

.

.

and afterward

that

which

is

Corinthians 15:42-46

These verses strongly suggest bodily resurrection indicates a belief that the

mind cannot

exist

in the afterlife,

which

independendy of the body;

279

SOULS OF SILICON

and Christianity preaches that a human being's soul and body are substantially united.

bodied

spirits,

72

Resurrected Christians



new body The soul as an

but in a

resurrected the body.

live after

as the

death not as disem-

Gospels claim that Christ

entity separate

from the body

influence from Eastern faiths, and Christian religions

is

do not make

an

it

a

tenet of their doctrines.

Other

The

religions do,

however, erect soul-body dualism

conflict with the materialistic

dogma.

view of the mind, since for the same soul

to live in several successive bodies implies a

and

as a

doctrine of reincarnation, for example, does indeed appear to

soul.

And

dichotomy between body

then there are the near-death experiences, as reported by

Raymond A. Moody 73 and a growing body of other researchers.

Patients

revived after approaching or reaching a state of clinical death recount strange experiences, claiming to have stepped out of their bodies and

observed the

Many recall

efforts

of the medical team to bring them back to

traveling through a dark tunnel

and emerging in

a bright

life.

and

peaceful light where they meet deceased friends and relatives. These

accounts cannot be dismissed

million adult Americans

Further,

For one

lightly.

of them: a Gallup poll performed

may have undergone

many

thing, there are too

1982 indicated that

in

as

many as

eight

a near-death experience.

74

NDE subjects are often in possession of knowledge they could

not have acquired by conventional means:

many who were unconscious

during the revival attempts can afterward describe those attempts in

Some

great detail.

subjects can even provide faithful accounts of events

occurring in other rooms, which they claim to have visited in their

disembodied It is still

the

mind

wound

life:

AI view of a

materialistic origin for

not incompatible with near-death experiences or reincarna-

is

tion. First,

during

state.

possible to argue that the

no

dualist

would deny the

otherwise,

essential unity

why would damage

of mind and body

to the brain (such as a head

or cerebrovascular accident) result in damage to the mind?

problem

is

what happens

at the

contains interesting speculations

moment of death, and on

the

AI

that very issue.

These concern the gradual and eventual replacement of brain electronic circuits with identical input-output functions.

philosopher

Zenon Pylyshyn used

The

folklore

this

cells

by

Although the

thought experiment as

a philo-

sophical argument in 1980, 75 the idea had been the subject of intense

debate in the AI grapevine for some years beforehand, under the

name

of "downloading." In 1988, Hans Moravec gave a dramatic description

280

Al

of the process ill

book Mind

in his

Children.

16

Assume you

patient in a twenty-first-century operating

A

substance.

your

and

local anesthesia

conscious.

fully

in

robot surgeon equipped with micromanipulators opens

under

skull

remain

are a critically

room, says Moravec

of neurons located

The surgeon

sets to

first

work on your

He

of your cortex.

in the periphery

brain.

You

concentrates on a small clump severs the

nerve-cell connections linking this assembly of cells to the rest of your

and replaces them with two-way connections. These

brain,

clump or

brain either to the

to an artificial

model of

it

link

He

scopic components, which the surgeon proceeds to build.

produces the structure of the it

cell

clump

in the artificial

your

made of micro-

so as to duplicate the exact behavior of the biological clump.

activating the switch that connects the rest original cluster or

When you no

accuracy of the modeling.

verify for yourself the

longer feel any difference

between the two positions of the switch, the biological clump removed, and the surgeon

number of

sets to

work on another clump. In

similar stages, he replicates

By

of your brain to either the

you can

electronic replica,

its

re-

model, and tunes

your entire brain

in

an

is

a large

artificial

no point do you experience an interruption of your end, your mind has been transferred to an artificial

construction; yet at

awareness. At the neural net.

This (so

far)

transferring a

imaginary process strongly suggests the possibility of

mind from one support

to another. Near-death experi-

ences and the survival of the "soul" after death could be explained by a similar transfer process. In this case, the receiving

be matter

do not

as

we know it, and

yet understand. It

would be required

pure

spirits;

though is,

certain,

for the information and organization that constitutes

our minds. Indeed, in death experiences

is

this regard,

some of those who underwent

insist that in the

rather,

disembodied

state,

near-

they were not

they inhabited another kind of "body" which,

invisible to living

humans, did possess

a definite structure.

of course, pure speculation to show that religious

larly the belief in survival after death, are

that the

support would not

would involve mechanisms we though, that some kind of support

the transfer

mind emerges from

physical

beliefs,

77

This

and particu-

not incompatible with the idea

phenomena.

11

HOW MANY BULLDOZERS FOR AN ANT COLONY?

now

should be clear by

It

define

it.

intelligent

Yet

it

system

that intelligence defies the heartiest effort to

equally clear that an essential ingredient of an

is

is its ability

the only function

common

to manipulate information. Indeed, this to brains

and computers. The

is

essential

ingredients of information are bits. Just as matter ultimately consists in

atoms,

all

the information that reaches us through our senses can be

broken down into

little

pieces of "yes " and "no." These particles

up any conversation, scenery, spanking, or

how

this is possible,

brain,

on clouds and water;

and waves twinkle

which

lets

experience.

a royal alley

make

To

see

no

us appreciate this beauty, has

inside our skull, our thinking organ

the sea as

would be

a

computer shuttered

Pentagon. In order to appreciate

this scene,

is

bits

Yet our

direct contact with just as

it.

removed from

basement of the

in the

our brain must reassemble

raw elements of data that our senses supply

corresponding to yes or no

of gold leads

like stars in the reflected light.

Locked up

the

we

consider the spectacle of the sun setting into the

ocean. Delicate hues play to the sun,

caress

it

as nerve impulses

of information.

In the case of sight, perception happens as follows:

The

cornea, a

transparent lens on the front of the eye, projects an image of the scene

onto the back of the ocular globe. In like a

this respect, the

eye works

camera, where a lens projects the image on photographic

much

film.

The

282 retina,

||

which plays the

of film

role

in

our eyes,

back of the ocular globe.

tissue covering the

sensitive nerve cells called "receptors."

a sheet

is

contains

It

Some of

of nervous

many

light-

these receptors

tell

how bright the receptor. Some other

other neurons, through a sequence of nerve pulses, projection of the image

receptors signal

at the location

is

how much

of the

red, green, or blue the

image contains

of pink, orange, and indigo

locations. (All the delicate hues

correspond to varying mixtures of these three primary

image

is

turned into discrete pulses in two ways:

becoming an

array

of a receptor

cell;

at their

in the sunset

The

colors.)

spatially,

first,

by

of dots, with each dot corresponding to the location

second, in the domain of brightness and color. Colors

become mixtures of discrete

more

hues, and brightnesses translate into

or less rapid firings of nerve

cells (the larger

the brightness, the faster

the firing rate). In similar ways, nerve cells in our ears turn the sounds

we

hear into pulse

Sensor

trains.

cells in

our skin do the same for

sensations of heat, cold, and pressure.

Thus, our brain it

bombarded with

constantly

is

what our senses

of pulses

trains

do with our

perceive. Intelligence has to

telling

ability to

manipulate these bits of information and use them to make sense of the world. Animals manipulate the information from their senses in a

ner that does not

them generate more than immediate

let

perceived threats or inducements.

We, however,

get

more mileage out

of the information we extract from our surroundings. into knowledge

and use

it

those of most

for planning

shows

mammals

their complexity.

We

can refine

it

for long-range planning and abstract reasoning.

do work on

Nevertheless, our brains animals. Dissection

man-

reactions to

the

same

lie

in the size

Thus, one can

of the structures present and

logically

and abstract thought are

principles as those of

between our brains and

that any differences

in

conclude that our capabilities

built

on

the

same basic

skills

that

allow animals to react to their environment. Further, this extra power

probably stems from the additional that the

more

abilities for

Thus, intelligence has to do with manipulate in a given time

measures information in elementary form,

is

compares the brain

(say,

bits,

how much

information one can

per hour or per second). Since one

one aspect of

intelligence, in

its

most

bits per second of raw processing power. If one

to a telephone switching station, this

correspond to the number of phone given time.

processing information

elaborate structure of our brains allow.

Of course,

there

is

more

lines the station

to intelligence than

power would

can switch in a

raw power, but

HOW MANY BULLDOZERS FOR let

283

AN ANT COLONY?

us not worry about this aspect of the problem right now. Let us just

recognize that no matter switching station

is, it

will

how

superbly structured and

simply not do

programmed

the

job if it can't process enough

its

connections in a given time. In the

first

questions:

unit of time,

benchmark? ment, and

part of this chapter

How many bits and

how

I shall

then look back

how

of our brains.

processing power

is

answer the following

to

do our present computers come

close

try to extrapolate

rise to the level

I shall try

of information can the brain manipulate per

at the history

long

it

of computer develop-

will take for

Finally, I shall

to this

our machines to

acknowledge that raw

not the only ingredient required for intelligence, and

discuss whether software powerful

enough

to emulate the

human mind

can be developed for the computers of the future.

THE HUMAN CORTEX AS CIRCUIT BOARD I he exposed

three

human

pounds of

brain

is

certainly

not an impressive

soft, jellylike, grayish tissue.

This

mushy

sight:

about

texture long

prevented anatomists from cutting the brain into clean thin

slices suit-

able for microscopic observation. Further, the uniform color of the

material kept

them from seeing

structural details. It

was only

in the late

nineteenth century that different hardening and coloring processes,

among them fine texture

the still-used Golgi stain, enabled anatomists to study the

of neural

tissue.

came as no surprise that, like other organs, the brain is made up of cells. They come in varying sizes and shapes, and neuroanatomists called them "neurons." One feature of the neurons, however, did astonish It

early researchers, including the Spaniard Santiago Italian

Ramon y Cajal and the

Camillo Golgi, developer of the staining process. They were

astonished by the intricacy and extensiveness with which these

connected to each other, each sending out that link

it

to as

many

other neurons.

literally

cells

thousands of tendrils

They make up

a

network of such

Byzantine complexity that Golgi, for one, firmly believed

continuous tissue extending throughout the brain.

He

it

formed one

defended

point of view, called "reticularism," in his 1906 Nobel address.

1

this

Later

284

U



observations

path



science progressed in

as

proved him wrong. Indeed,

neurons play

tedious, prodding

usual

its

we

as

gaps between

will see, the

workings of the brain.

a crucial role in the

Early in this century, researchers started to distinguish elements of

order in the apparent chaos of brain structure. that,

First, investigators realized

much

although neurons can differ from each other as

suckle bush does

and

different shapes

morphology

from

exist

Moreover,

brain).

mammals, from

a sequoia tree, they

sizes.

come

Only seven kinds of cells with

honey-

as a

in a small

number of

similar exterior

throughout the cortex (the largest structure cells

very similar to these

the higher primates

down

and bolts of our most abstract thoughts

make up

to the

in the

the brains of

puny mouse. The nuts

are thus the

same ones

that

support the mouse's instinctive reactions.

The

cortex, the brain structure responsible

motor responses, and about is

that

first

intellectual functions,

six millimeters (a

is

for our perceptions,

a thin sheet

quarter of an inch) in thickness.

of a square twenty inches on

of nerve

Its

a side: roughly the space that

personal computers used to take up on a desk.

To

fit it

IBM's

into our

Nature has had to fold the cortex; hence, the furrowed look of

skulls,

the naked brain.

The

cortex comprises six distinct layers, caused by an

uneven distribution of neurons of

makeup of the

layers vary

in the visual part, the

different types.

The

thicknesses and

over the area of the cortex. However, except

number of

cells

per unit area remains

fairly

stant at 146,000 per square millimeter. (Multiplying this figure

area of the cortex produces an estimated total

of about 30 cells

billion,

or 3 x

10 10 .)

The average

us from the

of the

is

likewise the

mouse

cells in

is

same

for

all

con-

by the

number of neurons

in

it

distribution of types of

throughout the cortex also remains constant. The density of

per unit area

To

cells,

surface area

mammals. What

cells

distinguishes

the area of our cortex, and not the kinds or density

it.

an engineer's eye, the cortex presents striking

similarities

with a

structure universally present in computers: the printed circuit board, a flat,

thin support holding integrated circuit chips,

cessing elements.

The board

which serve

as pro-

allows the chips to talk to each other

through conductive paths buried

in distinct layers

over

its

thickness.

Strangely enough, a typical board comprises six layers, just like the cortex.

Each chip on the board

(transistors, capacitors, diodes)

performs

is

made up of microscopic elements

of about the

a specific function within the

size

computer.

It

of

a neuron,

and

turns out that one

285

HOW MANY BULLDOZERS FOR AN ANT COLONY?

can also divide the cortex into chips, after a fashion. Experiments

conducted on animals by probing the sensory cortex with a microelectrode can detect firing impulses from single neurons. If you

move

the

electrode to and fro, in a direction perpendicular to the surface of the cortex,

you

only respond

However,

meet only nerve

will

For example,

if

you

if

and not on the

nonperpendicular direction by slanting the

in a

or the right eye, alternatively.

design our cortex as a circuit board?

both brain and board,

As

called "pyramidal cells,"

units.

of the brain

of the brain or cortex. Covered with an

of myelin, these

fibers

make up

a whitish tissue,

very different in appearance from the gray color of the cortex

about one

cell in a

of white matter

is

larger than that

kept them out of the cortex.

between the cortex and related to ease

of design.

a circuit It is

chips and connecting paths sions

would be

Only

of gray matter. Having

would have

with the direct communication between adjacent gray

why Nature

itself.

hundred extends beyond the cortex, yet the volume

in the brain

the "white cables" travel through the gray matter

bly

lies

Large neurons present in the cortex,

send nervous fibers downward, out of the

cortex, toward other regions

insulating greasy layer

depth.

We can conjec-

necessary to separate the closely

it is

in a circuit board, the cabling

underneath the computing

its

from the cabling connecting faraway

interacting processing elements

parts of the network.

left

the cortex were divided horizon-

It is as if

into different modules, each extending throughout

Why did Nature ture that, in

may

right.

meet neatly separated regions which respond to the

needle, you will

tally

process one kind of stimulus.

shine a light in the left eye,

you probe

if

cells that

buried into the visual part of the cortex, the probe

on

The

interfered

cells: this is

proba-

similarity in structure

board may have yet another reason,

already quite complicated to lay out the a flat surface.

a combinatorial

Doing

it

in three

dimen-

problem of monstrous proportions.

Perhaps Nature was no more willing to face engineers are! Whatever the reason for

convergent evolution of brain and

it,

circuit

this difficulty

than

human

one cannot contemplate the boards without wondering.

Let us go back to the basic building block of the brain, the neuron. Its

anatomy can

by the

brain.

appendages. a tree.

tell

us

more about

Extending from the

On

the cell

amount of computing performed body of the neuron are different

the input side, the dendrites look like the branches of

They connect with sensor cells, or other neurons, and receive The meeting points between dendrites and append-

their electric pulses.

ages of other cells are actually gaps, called "synapses." Although re-

286

Al

searchers had long suspected their existence, they could not prove

before the invention of electron microscopy

in the 1950s.

blow

vations of synapses then struck a final

it

Direct obser-

to the theory of nervous

system continuity, or reticularism, which Golgi had defended

until his

death in 1926.

When tween cell,

strong enough, nerve pulses can cross the synaptic gap be-

cells.

Pulses usually increase the electric potential of the receiving

which encourages

this cell to

generate a pulse of

Sometimes, however, arriving pulses decrease courage the receiving

cell

from

firing.

The

combines) the membrane potentials and function of this sum. a wire-like

The

cell

own, or

body sums

body

cell

called the "axon." It

"fire."

and

dis-

(or otherwise

fires at a rate that is

pulses generated by the

appendage of the neuron

its

this potential,

an /-shaped travel

along

may be

short

or very long: axons sometimes bundle together to form a nerve, which

can be as long as your arm. They also form the "white cabling" of the brain

I

treelike

have mentioned. The axon eventually branches out into another

network of

fibers.

These pass along

signals to other cells, or

activate muscles.

The

input and output ramifications of the neuron are

characteristic.

on

Extrapolations from counts

its

most

striking

electron micrographs

show there are from 10 14 to 10 15 synapses in the cortex. This means that, on the average, each neuron receives signals from about 10,000 other cells and sends its own messages to as many others. In this respect, the brain differs markedly from electronic circuits: on a circuit board, one component typically makes contact with fewer than five others. However, what computers lose in connectivity, they make up for in speed. In the brain, the pulses traveling from neuron to neuron are local

imbalances in

exacdy

how

salt

fast

concentrations

moving

in the order of 100 feet per second. This

from 10 to 100 milliseconds to reach the ever, pulses

They

moving from chip

travel at

at relatively

depends on the diameter of the nerve

two

thirds

is

why

low speed

fibers,

but

it is

sensor) stimuli take 7

cortex. In a computer,

how-

to chip are pure electromagnetic fields.

of the speed of

light

— about seven

million

times faster than nerve pulses!

The Brain's Processing Power Although our knowledge of the

brain's structure

mains sketchy. Recendy bold-hearted

scientists

is

progressing,

have

it

re-

tried to use this

HOW MANY BLLLDOZERS FOR

scanty evidence to estimate the the brain.

I shall

287

A\ ANT COLONY?

amount of raw computing going on

examine two such

tries

— by Jacob

T. Schwartz at

in

New

York University and by Hans Moravec at Carnegie Mellon University. It will come as no surprise that these professors achieved wildly different results. In fact, the very divergence of these estimates is a good illustra-

how

tion of

little

we

really

know about

the brain. Yet, because of the

accelerating pace of technological development, even guesses as poor

these provide useful estimates of

when we

as

be able to beat Nature

will

at brain building. I shall start

Courant

with the work of Jacob T. Schwartz, a professor at

Institute

neuron can

since a

about one hundred times per second,

fire

information to other neurons

much

whether to it

receives

lish is

at a rate

fire,

larger.

To

inside the

combine

to

first

from ten thousand other neurons. The

whether

this total is large

enough

for

it

to

cell

per is,

a second,

the signals

must then

The

fire.

all

sends

bits

neuron

one hundredth of

decide, every

the average neuron has

it

of about one hundred

The amount of information processed

second.

however,

NYU's

of Mathematical Sciences. 2 Schwartz estimates that

estab-

decision to

fire

complex, especially since some of the messages received from other

neurons may inhibit

firing rather

that to reach this decision, the

synapse

— perform

than promote

neuron must



Schwartz estimates

it.

for each firing, at each

the equivalent of calculations involving forty bits.

Since these operations involve intermediate steps, simulate them,

we have

per synapse, per

then a straightforward

bits

assume

that to

of information

affair to

amount of information processed by one neuron

overall

100

firing. It is

let's

one hundred

to manipulate

work out

in

the

one second:

per synapse per firing per neuron x 100 firings per second per

bits

synapse x 10,000 synapses per neuron per neuron.

From

there,

we

=100

million bits per second

the entire cortex: 100 million bits per second per neuron

neurons

in the cortex

=

power of x 3 x 10 10

get an estimate for the processing

3

x

10 18 bits per second of information

19 bits processing. Extrapolating to the entire brain, a total of about 10

per second

results.

Thus we have our

of brain power: 10 19

bits

first

estimate: Schwartz's estimate

per second.

Schwartz, however, puts a very strong qualifier on points out that computation rates,

might

safe to

suppose that what

really

this figure.

He

orders of magnitude lower,

suffice to represent the logical operations

is fairly

is

many

of the brain. Indeed,

it

matters to our thought processes

not the internal mechanics of a neuron, but

how

it

looks

like to

other

288

Al

may be

neurons. This

may be enough

more simple than

considerably

would show. Thus,

internal structure

stick-figure

the neuron's

models of neurons

to simulate the brain accurately. Moreover, our brain

is

accommodate a very large amount of redundancy, and much of complexity may be due to the constraints limiting its growth and

built to its

evolution (see the section entided "Avoiding Nature's Mistakes" later in this chapter).

Hans Moravec calculates the information-processing power of the manner different from Schwartz's, concentrating on the retina,

brain in a

the paper-thin layer of nerve cells and photoreceptors in the back of the eye.

3

amount of massaging on

After performing a certain

tion provided

by the receptors, the nerve

cells

send the

calculations to the brain through the optic nerve.

of the retina

that, in effect,

it

Such

the informa-

results is

of

their

the structure

makes up an extension of the

brain. Yet,

contrary to most brain structures, the functions the retina performs are well understood.

processing

They

are similar to those

of

artificial

vision systems

TV images. Of course, we know exacdy how much

comput-

power these operations require. Further, and by no coincidence, the resolution (number of receptors) of the fovea, the high-resolution part

ing

of the basis,

retina,

is

about equivalent to that of a television image.

Moravec estimates the processing power of the

He

then proceeds to extrapolate from entire brain,

and

is

this

large.

Which

figure the

that

retina,

computing

The

faced with a dilemma.

about 1,000 times as many neurons as the 100,000 times as

On

about one

per second.

billion operations

power of the

retina at

but

its

brain has

volume

is

figure correctly accounts for the larger

computing power of the brain?

We

can attribute the excess volume to

three factors. First, the connections between neurons in the brain are longer: the required cabling takes

up most of the space

in the brain.

Next, there are more connections per neuron in the brain. Finally, the brain contains nonneural tissues, such as the greasy myelin sheath of

many nerve

fibers.

Of

connections per neuron

Moravec, we

these three factors, only one



entails

shall thus take the

an increase



the excess of

in complexity.

Following

Solomonic decision of awarding the

power 10,000 times that of the retina. There follows an information-processing capability on the order of 10 13 calculations, which is Moravec's estimate of brain or about 10 14 bits, per second

brain a computing



power.

According to him, the brain

is

thus 100,000 times slower than

— HOW MANY BULLDOZERS FOR Schwartz's estimate of 10

one

19

per second. Moravec's procedure has

bits

crucial advantage: since

289

AN ANT COLONY?

sidesteps the need to rate the

it

factors required to adjust Schwartz's estimate,

we no

unknown

longer need to

guess the effective ("stick figure") processing power of an individual

we

neuron, and

are also spared the

need to assess the unnecessary

complexity with which evolutionary constraints burdened our brains.

Moravec's estimate probably

How do brain?

Not

computers

fare

well at

The

all.

closer to the truth.

lies

compared with the processing power of the computer

fastest

in existence in 1989, the

11 Cray-3, could process only 10 bits per second.

it is

therefore

1 ,000 times weaker than the

of the laboratory

level

the Cray-3

is

much

rat,

but

it

is

like the

As any computer calculates

is

How much

brain of 65 million neurons. Further,

AI work. Researchers must

Sun-4 workstation. At 2 x

500,000 times

would evenly match the

The Size of

its

too expensive to serve in

make do with machines second, the Sun-4

with

By Moravec's estimate, human brain, or at about the

1

less

powerful than a

00,000 neurons of a

1

8

human

bits

per

brain

snail!

Human Memory enthusiast knows, the rate at

but one measure of

its

which a given machine

power. Another crucial question

memory?

information can the computer hold in

is,

Similarly,

human mind, how well we think is very much a function how much we know. As it turns out, it is possible to estimate how much memory we need to function in the world. There are three ways to go about this. 4 One is to repeat what I have

in respect to the

of

just

done

for the brains calculating power: that

is,

examine the

anatomy and work out estimates from hardware considerations. direct

method

is

to survey the

could also deduce

how much

knowledge of an average adults

know from how

brain's

A more

adult. Third,

fast

we

they can learn,

how long they live. To start with the first

and

approach, what happens in our brain when we remember something? Scientists are still very much in the dark about memory. They know plenty about the periphery of brain operation, such as

how we

perceive the world through our senses, or

how we

activate

our muscles to act on the world. What happens between, though,

mains very

much

a conjecture. Researchers

do not

on one particular response to a perception, or memories on which we base this decision. One can make settle

re-

know how we how we store the

really

plausible

290

Al

assumptions, though. Consider a mouse that

of

snarl

a cat: this instinctive reaction

our own.

First,

sensor

the mouse's ears send nerve pulses to

cells in

other neurons in the brain, which start firing

of neurons assembles

activation pattern ally,

further

response to the

flees in

mechanism probably resembles response. Thus, an

in

mouse's brain. Eventu-

in the

waves of activation reach the neurons controlling the

legs,

which send the mouse running.

The

snarl

neurons

of the cat thus corresponds to an activation pattern of

in the

same way. What happens, in

to

A human

mouse's brain.

then,

brain

would represent

when we remember

some of the neurons become active again. By this

that at least

perceived a snarl

in the

the snarl of a cat

response to another cue, such as the sight of an angry cat?

assume

it

It is logical

when we last memory is also

that fired

token, a

a neural activation pattern.

Can we

identify in the brain the elements responsible for eliciting

What can

such an activation?

cause a certain group of neurons,

the billions present in the brain, to

become

active

of

all

a

among

sudden? All

evidence points to the synapses, these microscopic gaps between neural terminations.

The

average neuron makes contact with ten thousand

others through synapses of various conductivities.

through more or

Of the fire in

less

ten thousand

A

synapse can

let

of the nerve pulses emitted by the source neuron.

downstream neurons, those

that are

more

likely to

response are those with the more conductive synapses. Thus, one

can assume that highly conductive synapses connect the neurons representing an angry cat to those representing the snarl of a synapses

probably store the information causing the

If this

is

the case,

encoding memories of

a synapse

bits

we can

as follows:

can have sixteen

of information, since

a

this

mean

cat.

Hence,

of memories.

estimate the capacity of the brain for

Assume that the degree of conductivity values. Then the synapse can store four

sequence of four

numbers from to 15. The 10 15 synapses room for 4 x 10 15 bits.

Does

recall

bits

the brain can actively use that

tion? Probably not. Synapses are just the

can represent the

in the brain

would then hold

many

mechanism

bits

of informa-

that induces pat-

terns of neural activation in response to stimuli or other activation

however, many fewer neurons than synapses. If a

patterns.

There

memory

item corresponds to a group of neurons firing together, then

are,

there will be fewer such items than synapses also. In the past few years,

AI

researchers have devoted

much

attention to studying networks of

2 91

HOW MANY BILLDOZERS FOR AN ANT COLONY?

neurons. Experimental results, as well as mathematical theory,

artificial

show that the number of the number of neurons in fewer

bits

known

one can store

bits it.

Further, an

than there are neurons in

in

such a net depends on

artificial

net can store typically

For example,

it.

a type

of neural net

Hopfield network, containing n neurons, has a storage

as a

capacity of 0.1 5«. 5

Assuming a

a capacity of about

we end up with

similar ratio for the brain,

of usable memory.

5 billion bits

1

A more direct way of finding out how much each of us knows is the game of twenty questions. It involves two people. Player A thinks of a and player B must find out what

subject,

A

will

in

twenty questions or

only answer with yes or no.

and known to both primary information

not count. in just

It

The

less.



like

"300,286

turns out that a

player's

The

through.

sift

wins

if

B

can't guess the subject

must be

target item

that

clearly identifiable

one must deduce from other the product of 482 by 623" do

players. Facts that

good

about twenty questions

have to

A

by asking questions

it is



first

memory in two groups:

is



come to the answer how many items you

player can usually

this fact reveals

question

you

lets

partition the other

the items corresponding to a yes answer,

and those corresponding to no. The next question divides one of these groups in two again, and so on. Since twenty partitions are required for

you is

to

end up with

clearly

a single fact, the

2 20 or about ,

because players

,000,000.

1

We

number of items

It is

to

A

Eiffel

Tower. Items

only, or too sensitive for casual evocation, will be avoided.

not farfetched to multiply by another factor of 2 to compensate for

this effect:

A

this figure

choices to neutral items of

will typically limit their

mutual knowledge, such as Marilyn Monroe or the

known

choose from

to

must, however, correct

we

are

now up

to 2,000,000 items.

knottier issue concerns the hidden information corresponding to

unconscious or informal knowledge. For example,

does a recipe for

knowledge

how

that enables us to interpret

tions in people?

how much memory

to tie shoelaces, say, take up?

Such knowledge

will

body language or voice

scious?

Not

necessarily.

enough room

also

There are two reasons for a mental is

tions ("I desire mother, but

probably accounts for

activities

Are most of our memories, then,

remain unconscious: one

occur

cases,

awareness for

is

all

that there

his or her

is

at

uncon-

activity to

forbidden!"); the other reason,

many more

in a person's

inflec-

the repression of painful associated it is

the

never appear as an item of the

game. Psychologists consider that most of our mental the unconscious level.

What about

emowhich

simply not

mental

activities.

292

Al

For example, consider your behavior when you drive while earning on a conversation.

down not

You

will steer left

or right, watch for other cars, slow

or speed up as required without any conscious decision. That does

mean you

are ignorant

able to stop by stepping

of the

on

could

come

unconscious, merely

just a

that

it is

good

memory

An sents

7

Let us boldly "guesstimate" again

deal!

than one

Lincoln," to

A

lent

of

1

What

1

are

now up

to

it

would correspond

to an

example, "I was thinking of

of information does the Abraham sequence "Abraham Lincoln"

First, the

blank,

we

being.

have called an "item" repre-

I

says, for

how many bits

Lincoln structure correspond? contains 14 characters plus

human

Typically,

bit.

entire data structure: if player

Abraham

this effect:

in a typical

important question remains.

much more

game. So

not most of our memories that are

and multiply by 2 to compensate for 4,000,000 items of

being

You are simply too busy much unconscious behavior uses

to full awareness in the twenty-question

one might reasonably argue

as

the brake pedal.

to pay attention to these details. Thus, facts that

of driving, such

technicalities

which probably requires the equiva-

5 bytes of storage space in the brain. (A byte

is

a set

of eight

The sequence of phonemes corresponding to the pronunciation of the name probably requires about as much. A few years after history classes, most of us probably remember only a sketchy biography of the sixteenth president: "He taught himself law and campaigned against slaver)- as a congressbits. Digital

computers use bytes to represent

man. His election

North

as president caused the Civil

to victory, emancipated the slaves,

Address.

He was

assassinated in 1865

of

internal representation

from the a

string

tion requires

this

much

it

less

is

attending a play."

Our

certainly very different

takes to write

no reason, however,

is

War. Lincoln led the

and delivered the Gettysburg

when

information

of 265 characters

computer. There

characters.)

it

down

or store

it

in

to believe that this representa-

information in our brain. If

this

were

so,

we

a much more we needed much more information than a few store these characters, it would mean that our speech

would probably have developed

concise language and

writing. Similarly, if

hundred bytes to is

much more

efficient

than our thinking:

this is

hard to believe. Let us,

therefore, accept that this short biography of Lincoln requires the equivalent

of about 265 bytes of storage

What about

Lincoln's face?

in our mind's eye,

we

thousands of others.

Even

in if

our brain.

we cannot visualize him

precisely

could certainly recognize his photograph

How many

bits

among

of information does one need to

293

HOW MANY BULLDOZERS FOR AN ANT COLONY? store a recognizable likeness

TV images into

of a face? Not much,

it

turns out. Digitizing

of numbers, with varying resolutions, shows that

arrays

an image of 20 x 20 pixels (dots) of varying gray levels provides a very recognizable likeness. With 4 bits per dot, store such a picture,

which brings the

size

would take 200 bytes

it

to

of our Lincoln data structure

495 bytes.

to

Yet another item of our internal representation of Lincoln

the

is

emotional aura surrounding assassinated presidents. Something in the data structure

must be pointing

to the emotions awe or sadness.

require pointers to other items of information: the

War refer

Civil

words

was

information

we

a

man and

could tap

a politician

if



required.

can estimate

more

how many

bytes

these relations require by reference to conventional bases.

Some

by the LISP language, require pointers between data

These pointers serve much the same purpose

memory model

above. In a LISP data structure, as

reserved for the pointers as for the data. Thus, brain, the is

amount of memory required

equal to the

How much ture

computer data

kinds of them, such as relational data bases or the data

structures used

items.

we know still

categories that contain

We

also

congressman or

us to other complete data structures. Further,

that Lincoln

We

memory

let

as in our human much memory is

us assume that in the

for relationships

between items

required for the data items themselves.

information does one require,

"Abraham Lincoln"? The 495

for the data struc-

finally,

bytes above correspond to about

4,000 bits of direct information. Doubling this for pointers brings us to 8,000 bits for one item of the twenty-question game. For the 4 million

items of information the in

memory,

the

game and other

would amount

considerations

is

in surprising

agreement with the one

per neuron, or 15 billion

Looking

at

how we

we

got by assuming 0.15

bits.

As witnessed by

check on

this figure.

little

of our

table,

we must painfully learn

abilities are

inborn.

the helplessness of a

From walking to

virtually

all

of the

skills

newborn

the multiplication

and knowledge we

we learn we knew how fast we estimate how much of it the basic

in the world. Yet, in less than

the basic material that will support us in

life.

twenty years,

If

new information, we could "human knowledge base" contains. Certainly, we do not commit new information

can absorb

of

involved, this

gather information provides yet another cross-

baby,

need to function

show we hold

to about 32 billion bits. In view

number of approximations and informed guesses

figure bits

the total

to

memory

as fast as

294

Al

our senses feed

The

to us.

it

optic nerve, for one, sends over a hundred

million bits of information to the brain at each second. Yet

and remember only

happens when you

of

a tiny fraction try to learn a

this information.

we

interpret

Consider what

page of text by heart. At a resolution

equivalent to 300 dots per inch, your optic nerve can send over to the brain the entire contents of that page in about a second. Yet,

glance at an open

book

and then read the page

for a second,

mind's eye, you'll discover that you can retain

most

at

a

if

in

you

your

few words.

Further, rather than corresponding to the image of the words, which

memory

requires thousands of bits per letter to describe, your

highly abstracted description of the

be a

will

words you recognized while glanc-

ing at them. This encoding probably requires only a few bits to store a letter.

bles

In

fact,

show

that

experiments on memorizing random sequences of

we can absorb new

per second. 6 Learning entire rate,

even

at

100

information only

at rates

sylla-

of 100

bits

per second means memorizing an

bits

page of text (about 400 words) in

less

than three minutes. At that

an actor could memorize his lines by reading them once aloud! Yet, at

such breathtaking speed, twenty years of continuous learning

eight hours a day

would let you

at

digest only 21 billion bits of information.

We thus have four estimates of the size of human memory. Assessing it

from the number of synapses leads

million billion bits. But as

I

to the astronomical figure of 4

pointed out, there

the capacity of synapse storage

is

no

and the number of

direct link

between

explicit items repre-

The other three estimates give much lower values: number of neurons in the brain and neural-net theory yield 1 5 billion bits. The twenty-question game leads to 32 billion bits. Learning rate and duration give 21 billion bits. The relatively close sented in the brain.

considerations from the

agreement of these three estimates

somewhere

(2.5 gigabytes) as is still

a lot

lets

in the range they define.

an estimate of the

of information:

it

one hope

Thus,

I

that the true value lies

shall settle

memory

store that

billion bits

corresponds to slighdy more than

pages of printed text or twenty-five hundred books

Can computers

on 20

capacity of the brain. This

much

like this

a million

one.

information? Yes: by this yardstick,

The Cray-2 supercomputer, memory capacity. 7 Even AI research budgets allow scientists to come close: As I write this the typical AI workstation offers about 200 million bits of random access memory, only 100 times less than the brain. As we saw in chapter 10, the Cyc common-sense knowledge base will be slighdy smaller than our our machines have already overtaken

built in 1985, already

had 32

us.

billion bits

of

295

HOW MANY BULLDOZERS FOR AN ANT COLONY?

estimate of the brain's capacity (about 8 billion bits instead of 20).

workstations will have that

much random

access

memory

AI

at their dis-

posal in a very few years.

REACHING HUMAN EQUIVALENCE Despite

this essential parity

of the board

process information thousands of times

in

memory, machines

more

we

slowly than

still

do. Be-

it is tempting to compare Computer engineers, by contrast, like to think

cause of its myriads of cells working together, the brain to an ant colony.

of

their

mainframe machines

of data through deed, since in

our

it

skulls.

their

as bulldozers shoveling

about mountains

unique central processors. Puny bulldozers,

would take thousands of them

to

in-

match the ant colonies

many of the disappointments a jet engine, they have to make

This fact certainly explains

AI researchers have met. If the brain is do with the equivalent of bicycles! Rather than uncover the secrets of intelligence, they must spend most of their time programming around the weaknesses of their machines. Yet, as

closing the gap.

Soon computers

will

we

will see next, engineers are

approach the power of the human

brain.

The

Generation of Computers

Fifth

As I described in chapter 1, the first generation of computers was based on vacuum tubes: orange-hot filaments glowed in various computing machines from 1943 to 1959. Even during those years, progresses in vacuum tube technology cut down by a factor of 20 the time needed to perform an addition. The gain in cost per unit of computing power was even more impressive. In 1943, it cost about one hundred dollars to buy one

bit

per second of computing power. Sixteen years

later,

it

cost less

than ten cents.

Generation

new machines

2,

based on single transistors, accounted for most of the

until 1971.

From

then on, computers were built out of

integrated circuits: silicon chips containing

and

finally

first a

few, then hundreds,

thousands of microscopic elements. These formed the third

generation of computers, and lasted until 1980.

Around

that year,

it

296

Al

became possible single chip;

computer on

a

and by 1985, these microprocessor chips contained up to

a

to put the entire processing unit

quarter of a million elements.

Thus was born

of

a

the fourth generation of

computers.

As I write this in the early 1990s, the upward spiral of computer power continues to accelerate. If you've ever pondered the economics of replacing an aging computer, you may have

felt a

kinship with the

future space traveler faced with the star ship problem: a better time to

leave

always next year, because by then ships will be faster and will

is

get you to your destination sooner.

machine than

will,

on

same

this year's for the

Let

me

So

is it

with computers: next year's

the average, offer 50 percent

demonstrate

this

more computing power

price.

tendency by focusing on two particular

machines. 8

The Zuse-2, the first electromechanical computer built in 1939 by the German engineer Konrad Zuse, would then have cost about $90,000 in today's money and took 10 seconds to multiply two numbers. By contrast, the Sun-4 workstation, introduced in 1987, cost $10,000 and can multiply two numbers

400 nanoseconds.

in

In raw power, measured by the admittedly crude yardstick of the time required to multiply two numbers, the Sun-4

than

predecessor. If

its

power, the comparison less to

To what

do

is

we

25 million times faster

even more favorable:

a multiplication with the

it

costs 225 million times

Sun-4 than with the Zuse-2.

understand the staggering implication of these similar

improvements would bring about

A luxury car of 1938 — say, a Cadillac in today's

money.

It

what the Sun-4

twice the speed of

No

fairy

waved

compares the right after

it,

is

on

— would have

cost about $30,000

and do 3

it

billion miles

wand suddenly

per gallon!

to induce these changes. If

and progresses on to the Sun-4, there

improvements

in

were to the 1938

would cost only $3,300, run at

relay-activated Zuse-2 to electronic

in price or dramatic

figures, consider

applied to automobiles.

a gallon. If today's Cadillac

to the Zuse-2,

light,

a

if

reached a top speed of 60 miles an hour and

traveled about 15 miles car

is

consider the cost per unit of computing

one

machines introduced are

no abysmal drops

performance anywhere. Instead,

smooth evolutionary process is revealed. Hans Moravec courageously calculated and plotted the cost per unit of computing power of sixtya

seven consecutive machines, starting with a mechanical calculator built in 1891.

9

These data points

clearly

show

that, for the past sixty years, the

cost of computing has decreased by a constant factor of 2 every other

HOW MANY BULLDOZERS FOR As

year.

a result, the

mainframes of the 1970s are the desktoppers of

now

today.

Mainframes of the 1960s can

many

electronic wristwatches contain

machines

297

AN ANT COLONY?

be stored

in a single chip,

more elements than

and

these early

did!

Although speculators

who

blindly extrapolate stock prices

from past

tendencies usually end up broke, there are sound arguments for applying yesterday's trends to tomorrow's computers.

of the past

ress

in the evolution

than

In

one

we connect

even

could

essentially

is

cally possible to build a if

many of these arguments

fact,

equivalence problem

brain

staggering prog-

of the technology. Since we are also dealing with the

behavior of an industry, technical.

The

stems from profound structural processes

sixty years

are

argue

economic. Indeed,

economic rather that

the

brain-

it is

now

techni-

machine with the raw computing power of the thousand Cray-3 supercomputers. And, even

a

though managing such interconnection, and programming

form

like a

human

raw power would be could do

that,

would

brain,

still

it

to per-

formidable problems, the

raise

we

experiment with. Before

available for us to

however, we would have to deal with the small matter

of finding the twenty

billion dollars this

network would

cost.

the problem of building human-equivalent hardware boils

reducing the cost of processing power to affordable

examine, therefore,

why we

levels.

Thus,

down Let

to

me

can expect the sixty-year-old trend of de-

creasing prices to continue. First, the regularity

of the price curve

is

to a large extent the result

a self-fulfilling prophecy. Manufacturers, aware

of

of the tendency, plan the

introduction of new products accordingly; hence, the absence of drastic

jumps

in

cost/performance

ratios.

Manufacturers introducing a product

no

incentive to

much lower.

Instead, they

well ahead of the competition in performance have

reduce prices

pamper

drastically,

even

if their

their profits for a while, until

costs are

competition forces them to accept

lower prices.

Competition in the computer industry $150-billion world 10

is fierce.

computer market, companies

To

grab a share of a

are willing to scram-

number of people developing computers, and on the rise. Since computer companies spend a constant fraction of their revenues on research and development, resources for computer development grow about as fast as the computer market. They pushed ahead by about 1 5 percent a year since ble.

For

this reason, the

the resources at their disposal, are

1960. Since this growth

is

much

faster than that

of the economy,

it

will

298 slow

Al

down

eventually. (Otherwise, a continued

growth would lead to the

impossible situation of everybody developing or building computers.)

Even

if

the

development activity

number of people and

dollars

devoted to computer

levels off, the total intellectual resources available for this

would

increase exponentially.

still

The reason

because com-

is

puters are largely designed by other computers. Indeed, involving puters in their

own

conception can have dramatic

problem of planning the paths of

on printed

metallic traces

com-

Consider the

effects.

circuit

boards. In the assembled board, these traces connect together the pins

of different processing chips. They available

on

have to

all

the board, while maintaining

fit

in the restricted area

minimum

distances between

each other. Typically, out of a multitude of possible combinations of paths, only a few satisfy these constraints. In the 1960s and 1970s, lay-

ing out these paths with pencil and ruler used to take months. Worse,

changing a design

took almost as long as starting anew.

after testing

Nowadays, computers perform

this layout automatically in a

matter of

hours. Similar gains occurred in implementing those procedures at the

chip level: integrated circuits are also designed by computers. In coming years,

computers ever more powerful

of the design and construction of the design

(or,

will gradually

assume

a larger part

their successors, further speeding

up

reproduction) cycle.

Economies of

scale should also

speed up the rate of price decrease.

Present computers typically contain only one, expensive, processing unit.

Future machines, however,

will consist

of identical components which

ally millions,

of thousands, and eventuwill serve as

and processors. Manufacturing these components ties will

in

both memory

such large quanti-

give rise to economies of scale comparable to those affecting

memory chips. Since there are many memory chips in a computer, they come down in price faster than processing chips. Recognizing these new economics, the Defence Advanced Research Projects Agency's goal to double the pace

of cost reductions

instead of multiplying

government hopes But even

if

components,

coming up

coming

it

One light

year, the U.S.

will result if

Nature does not cooperate. Aren't we

against basic natural barriers that cannot be

or, as

computer

is

on,

ever larger resources to perfect electronic

obvious boundary which



From now

every year.

Won't we soon bump our noses on the outer of

years.

computer power by 2 every other

to double

we devote

little

in

is

fast

scientists

limits

approaching

sometimes

is

call

overcome?

of computation? that it,

of the speed

the "Einstein

HOW MANY BULLDOZERS FOR

299

AN ANT COLONY?

bottleneck." In a conventional computer equipped with a single process-

ing unit, information flows between the

to

memory and

the lone processor

between swings. Infinitesimal errors

as acrobats leap-fly

murderous crashes, and the

entire

in

timing lead

computer must operate

like a finely

tuned clock. Indeed, an electronic master clock beats time, keeping

components

drummer

in lockstep, as inexorably as the

all

in a slave ship.

For the drummer to be obeyed, there should be time enough for one beat to reach

all

parts

The

the next beat.

of the computer well before the clock generates

beats are electric signals: the fact that they travel at

of

close to the speed

but no

light,

imposes

faster,

a limit

on

the

frequency of the clock. For example, the time required for light to travel the entire width of a

maximum

implies a

1

computer

-foot- wide

(megahertz, in computerese).

Many

within a factor of 10 of that

limit.

One

solution

is

1.76 nanoseconds. This

clock frequency of 568 million beats per second

would be

desktop computers already operate

to keep walking

on the path we have so

profitably followed since the invention of the transistor: that of minia-

make

Let us

turization.

tighter space.

The

the

components smaller and cram them

signals will

have

less distance to travel,

in a

and we can

only will crowding

bumps into more compo-

amount of heat

generated, but

then speed up the clock. Alas, this approach immediately

another obstacle: heat removal.

Not

nents in a smaller space increase the

having them work faster ponent. Evacuating

down

requires

will also increase the

this extra

— when

These complications

it is

just

heat generated per

com-

heat to keep the machine from melting

possible at

all



technological prowesses.

about cancel any economies brought about by

the extra miniaturization.

Nature herself presents us with

a

way out of

operate at positively slumbering rates.

hundred impulses per second, second of

a digital

human

different

fastest

mode of

CPU

initially

facture the

will

brain

generate about a

to the millions

of beats per

a

find

it

easy to keep a cool head.

thousand times the information-processing

computers. This performance

is

due to the

operation of the brain.

Von Neumann's and

opposed

we normally

Yet the brain packs about of our

neuron

computer's clock. Being so sluggish, each neuron

generates litde heat, and

capability

as

A

Com-

this blind alley.

pared with those of computers, the components of the

suggestion to break up computers into a

offered obvious advantages.

memory

as

many

It

low-cost, identical

was possible cells.

memory to

manu-

Since there was

300

Al

onlv one processor, variety

could be as complex as required to perform a

it

of logic or arithmetic functions. Further,

programming

to the relatively simple task

one processing

instructions to the

unit.

this layout

of issuing

reduced

a single string

of

Unfortunately the setup also

when memories inmemory of a modern

introduced a major inefficiency that became clear creased in size to billions of

computer

that

it

compares

names and addresses of

A

metropolis of

elements.

its

Running

a

program

New

fashion. First load the

back.

New

is

also houses millions

one road connects these two

that only

and allows through only

few

a very

bits

TV

set in 7

driving: they

your

car, drive

iron, drive to

Los Angeles, and

to

it

LA, and come back. Take the

LA, and so on. To enhance matters

have improved the data-transfer

computer. Unfortunately,

this

amounts

computer

a bit, cities

instead of

between the parts of

rates

up the

to speeding

circuitry,

and soon bumps into the speed-of-light and heat dissipation

to

I

mentioned.

move

of

von Neumann computer is like moving your York to Los Angeles in the following senseless

Take the laundry

coffee pot, drive to

tions

of

cities.

in a

manufacturers have recently tried flying between the

a

could store the

it

York or Los Angeles.)

at a time.

household from

come

the

to a large city. (Indeed,

inhabitants of

all

takes a long time to travel

It

is

own, the processing unit

The problem

information

So huge

bits.

all

An

analogue to the obvious solution

items of your household at once

computer. Each

bit transferred

requires a separate wire,





limita-

using a van

not possible in a

is

between processing unit and memory

and there

is

a limit to

how many

of these can

be crammed into a machine. Over the years, manufacturers have wid-

ened the data path from 8 machines.

A

time to 32, and even to 128 for large

bits at a

small improvement, this

amounts

to

little

more than

letting

you move both the coffee pot and laundry iron together! In terms of is

my

two-city analog}

7

surprising. It consists in

mingling the two

New

York

memory.

cities

,

the solution adopted by our brains

moving Los Angeles

so you don't have to

plays the role of processing unit,

It

move

processing unit. In the brain, there

is

at

New all!

In

York and

my

and Los Angeles,

turns out that the brain does not

between these two functions: each neuron serves a

to

make any as

both

a

fable,

that

of

difference

memory and

indeed no clearly identified center

of intelligence comparable to the processing unit of computer. For example, despite long-standing

a

von Neumann

efforts, neurologists

have

never been able to pinpoint a center of consciousness. The neurosur-

HOW MANY BULLDOZERS FOR

geon Wilder Penfield suggested the upper brain stem

301

AN ANT COLONY? it

might

in the

lie

combined action of

and various areas of the cerebral cortex. 11 Others

down and

pointed out that consciousness has to do with laying

memories of the world. In

central role in this function, ness.

12

Another view holds

recalling

hippocampus, which plays a

this case, the

might qualify for the seat of conscious-

that our ability to

communicate makes up

our most obvious mark of intelligence: the language centers, located the left cerebral cortex,

good reasons

would then bear the palm. 13 There

many

areas.

14

in

however,

of the brain handle

to believe that, although various parts

special functions, consciousness arises

are,

from the combined operation of

down, appear

Likewise, long-term memories, once laid

distributed throughout large areas of the brain.

What

are the advantages

von Neumann a myriad

architecture?

of

this distributed configuration

For

starters,

of operations concurrently. This

the torrent of information your eyes send

is it

over the

allows the brain to perform

it

how your brain

(millions

and let you instandy recognize what you're looking

can analyze

of bits per second),

at.

Roughly speaking,

your brain separates the image into hundreds of thousands of dots, each separately analyzed

by several neurons.

A pure von Neumann machine

would, by contrast, slowly process each dot in succession.

Computer

scientists call "parallel

cation to a single task of ers.

many

processing" the simultaneous appli-

processors, be they neurons or comput-

In addition to the speedup inherent in getting

job, applying parallel processing to

advantage:

it

barrier. Since

are

amounts

computers

more workers on

the

offers another potential

to nothing less than breaking the light-speed

processors in a parallel machine

no longer enslaved

to the

semi-independent units could

drumbeat of

now

be made

work

separately, they

a central clock.

as small

and quick

These as

we

want them.

Through such parallelism, Nature will allow us power of our computers at a steady rate for

the

to keep

on increasing

a long time to come.

Eventually, individual processors will reach microscopic dimensions.

The emerging tures in

science of nanotechnology 15 will soon

which every atom plays

its

By common agreement among computer machines are those that implement way.

A

few of these machines are

Connection Machine,

built

let

us build struc-

assigned role.

parallel

now

scientists, fifth-generation

processing in an extensive

in existence: for

by Thinking Machines,

Massachusetts, with 250,000 processors.

As

I

Inc.,

example, the

of Cambridge,

said in chapter 8,

ma-

302

Al

chines based

on neural networks,

in

which microscopic components

will

emulate the neurons of our brains, are being contemplated.

Duplicating the Brain's Processing I

can

before

now attempt to we close the gap

human

brain?

answer the question raised in

How

earlier:

long

processing power between computers and the

have summarized

I

Power

and 11.2 the

in tables 11.1

estimates about the brain's computing

earlier

power and information-storage

capacity.

Since

we

still

know little about how the

brain works, different avenues

of investigation lead to extremely different

results.

cited for the information-processing capacity differ

by

a factor

the brain (table

of 100,000.

do not

1 1 .2)

The two

of the brain

estimates

I

(table 11.1)

My

estimates for the

memory

fare

any better, being

six

capacity of

orders of magni-

tude apart.

Our

mightiest computers offer only an insignificant fraction which-

ever value

we adopt

take for the

for the brain's processing power.

upward

spiral

Various answers appear in tables 11.3 and for a

speedup

in the rate

How long will

of hardware progress to close 1 1 .4.

this

it

gap?

Despite the arguments

of computer improvement,

I

have taken the

conservative view that the sixty-year-old tendency of doubling every

other year persists. I

have taken for benchmarks

in tables 11.3

and 11.4 the Cray-3

supercomputer, built in 1989, and the Sun-4 workstation, built

table

Two Estimates of the Computing Power

11.1

Argument

in 1987.

of the Brain

Estimate

Detailed modeling of neurons (Schwartz)

10 19 bits per second

Comparison of the

10 u bits per second

retina with similar

hardware

table 11.2 Various Estimates

of the Information Storage

Capacity of the Brain

Argument

Raw

synapse storage

Neural-net theory and number of neurons 20-question

Human

game

learning rate and duration

Estimate 4 x lO 13

x 10 9 32 x 10 9 9 21 x 10 15

bits

bits bits bits

HOW

MAW

303

BULLDOZERS FOR AN ANT COLONY? table 11.3 Estimates for

the

Year

Human

Supercomputers Will Reach

Which

in

Equivalence

Best Case

Worst Case

Processing power

2009

2042

Memory

1989

2023

table 11.4 Estimates for

the

Year

Which Desktop

in

Human

Computers Will Reach

Equivalence

Best Case

These

Worst Case

Processing power

2025

2058

Memory-

2002

2037

tables

list

the years in

which machines of a cost equivalent

to the

Cray-3 (about $10 million) and the Sun-4 (about $10,000) should reach

human

The

equivalence.

"best case" columns correspond to the weaker

power and memory in tables 11.1 and 11.2; the "worst case" columns correspond to the stronger estimates. The large discrepancies between estimates make remarkably little

estimates of brain-processing

difference

on

According to

dates.

supercomputers

will attain

table

human

1 1 .3,

if

the

weak

estimate

is

right,

equivalence in the year 2009. If the

strong estimate holds, this sets us back only thirty-three years, to 2042!

Indeed, is all it

if

computer power doubles even other 7

takes to

tables, the

improve by

roadblock

is

a factor

of 100,000. Also,

supercomputers

as

is

clear in

both

we will always reach the From the first line of table

processing power, since

required memory' about twenty years 11.3,

year, triirty-three years

will attain

earlier.

human

equivalence around 2025,* give

or take seventeen years. According to table 11.4, desktop machines will

have to wait

2041, with the same error margin.

until

After these dates,

than

we

abilities

are.

We

of the

we can

expect our machines to become more clever

have already done Nature one better for

human

body.

Our machines

all

physical

are stronger, faster,

more

may be advanced: the September 1992 of Electrical and Electronic Engineers) Price estimates exceed noted that "engineers expect teraflops machines by 1996. U.S. S100 million" (page 40). A teraflops is the approximate equivalent of our weaker estimate for brain power. This power, however, will come at ten times our target price of S10 million. Further, the degree of specialization of these early teraflops machines * Recent

developments indicate that

this date

issue of Spectrum (the journal of the Institute .

makes

it

.

.

unlikely that any

.

.

.

amount of programming could endow them with

intelligence.

304

Al

enduring, and

more

accurate than

we

are.

Some of them have

eyesight or hearing. Others survive in environments that

we

suffocate us. Shouldn't

sharper

would crush or

expect to improve upon our mental

abilities

just as well?

Avoiding Nature's Mistakes we build into our machines the strength of our minds, with much to spare, but we can also avoid duplicating the many weaknesses and inefficiencies of our brains. Indeed, when building artificial minds, we enjoy much more freedom that Nature had in Not

only can

eventually

building us. First,

on

we

of the limitations on material and structure imposed

are free

must grow, reproduce,

biological organisms. Living cells

selves,

and move over to

They must

their

proper positions

body

repair

them-

early in

life.

constantly absorb nutrient material from their environment

and evacuate waste. Most of their these ends.

in the

An

any of these

tasks.

internal structure

and functions serve

neuron, however, would not have to perform

artificial

function would reduce to generating electric

Its

of a biological neuron. Thus, we can expect the

signals similar to those

structure of an artificial

neuron to be much more simple than

natural one. Further,

could use materials that transmit impulses mil-

lions

of times

it

faster than

protoplasm and process

that

signals that

of a

much

faster.

Yet another

limitation our

machines

will

dispense with has to do with

blueprints. Nature's blueprint for our bodies, the

DNA molecule, does

not contain enough information to

connections of each

specify' the

make do with general instructions issued cells. What we know of neuroembryology

neuron. Instead, Nature must to entire classes of brain

shows

that in the early stages

through brain

tissue.

axon of the adult ends meet

cells

then bind to the

cell.

of

of life, brain

cells

emit filaments that travel

These filaments eventually form the dendrites and

They

travel

more or

less at

a kind that chemically attracts

cells in

random,

connections that become synapses.

stand the limitations this

mode of

performance, consider the following

until their

them. The filaments

To

under-

construction places on the brain's fable,

which

I

have called "Harry's

Plight."

new comOgomongo. Harry has

Harry, an electronics engineer, has just taken charge of a

puter assembly plant in the remote country of

MAW

HOW

accepted a mission no one else in the

computer

that the unskilled

from component

Ogomongans

the

305

BULLDOZERS FOR AN ANT COLONY?

ther can they

tell

chips.

Harry's dismay,

are incapable

wants: to design a

of reading

it

can assemble

soon becomes

easily

clear that

connection diagram. Nei-

a

apart chip models, except by their colors. Since dif-

ferentiating the pins

Harry has to

To

company

Ogomongo

workers of

on the chips is also a little hard for Ogomongans, mounting instructions that typically read: "Con-

settle for

nect any pin of a green chip to any pin of a yellow chip." Harry's considerable challenge to design chips of a kind that

connected in

this

haphazard way, produce

computer.

a

chance that compatible pins on different chips

number of

increases the

To

will,

when

increase the

will connect,

pins per chip. Second, he adds

now

It is

Harry

some

first

intelli-

gence inside the chips and decrees that each newly assembled computer

undergo

will

"running in" period of a month. During

a

this time,

each

chip sends out, through each of its pins, exploratory, low voltage pulses. It also listens

to pulses emitted

by other

Each pin of each kind

chips.

of chip emits a characteristic pulse pattern, enabling chips on each side

of

connection to check the validity of

a

mechanisms break

internal

tions of the right kind are maintained

Much to the this

There

and more expensive than

company wants improves

mongo

and somewhat to

his

own,

eventually does produce a working

com-

is

a

hundred times bulkier

number cruncher. Harry's The Ogomongans, however, feel it image, and insist on buying it. Since Ogo-

a conventional

to close the plant.

their international

sits

and strengthened.

only one snag: the machine

is

designed

of the wrong kind. Connec-

surprise of Harry's colleagues,

Rube Goldberg procedure

puter.

this link. Specially

off connections

on newly discovered

oil fields

amounting

to half the world's

reserves, they can well afford to.

This

is all

processes

fantasy,

of course

— but any resemblance

to existing biological

intentional!

is

In addition to the indiscriminate assembly of suffers

from

their lack

of

balanced mechanism, and

Yet we do not pensates for

Computers ways to

it

feel

also benefit

a price:

we

all

lose

A

a

neuron

is

a

parts, the brain

its

complex, delicately

hundreds of thousands every day.

any the worse for

by having

let their

must pay

reliability.

this loss

because the brain com-

large amount of redundancy

from redundancy, and engineers

machines

tolerate

minor component

making computers more

in

are

its

now finding

failures.

resilient requires

circuits.

Yet they

more com-

306

u

ponents. For this reason, building an intelligent machine out of parts

more simple and robust than neurons would increase its performance. The brain evolved through a process of small-scale, local changes spanning millions of years. intelligent designer

structure of the cortex

it

embodies many elegant features

shows. In

many

respects, the brain

schoolhouse turned into a major

one-room cabin with the children of the

a

first

school needed another tion.

Over

layered, circuit-board-like

prime example. Yet the brain's overall

a

is

wood few

city

Midwestern country

Then came

started with a

It

enough

stove, spacious

settlers.

room

like a

is

high school.

to

accommodate

the railway station: the

to handle the suddenly doubled popula-

To

the years, classrooms multiplied.

keep a studious atmo-

sphere, workers had to pare precious square feet from each

up linking

that an

expanded gradually, without benefit of advance planning,

architecture

and

It

would not disown. The

room

plumbing and

corridors. Later, installing indoors

to set

electricity

required major surgery, which gave the principal a severe headache.

When

it

meant

for one, the

became necessary to add mayor and the

The aldermen,

leery

a

second floor on

city*

a structure

never

engineer almost came to blows.

of raising taxes for

a

whole new building,

finally

overruled the engineer and hired a contractor themselves. Twenty years later,

congested plumbing and

wavering

lights

conditioning, corridor

air

traffic

jams, and

prevented any further expansion of the school. The

council voted the

site

and erected

into a park

a

new

city

school elsewhere.

Evolution does not have the option of starting over, and our brains still

contain the original cabin cum

wood

stove.

out of the upper end of the spinal cord. The the brain of our reptilian ancestors. tory

and attack prey or enemies,

around the

reptilian brain

this is the school's

second

is

it

Lemon-sized,

grows

Programmed

to stake out a terri-

Wrapped mammalian brain:

holds our darker instincts.

the limbic system, or old

floor.

it

reticularformation™ is in fact

Developed from centers

that

govern

mammals, the limbic system is the seat of emotions. It enabled our warm-blooded ancestors of a hundred million years ago to care for their young. Its programming often contradicts the reptilian brain, and many of our internal conflicts have no other origin. The smell in primitive

cerebral cortex

layer

holds our higher reasoning functions and forms the outer

of the brain.

It talks to

which somehow coordinate

the inner parts through its

action with theirs.

equivalent in our fictional country school. tect trying to design a better brain

At

many nerve fibers, The cortex has no

that level, a

human

archi-

probably would have started over.

I•

W

IA*

Our

1

Y

1 L L

B

Z E

IS

1

F

IITCtlflf!

II

old rriend the retina offers a striking example of

evolves impressively elaborate fixes to

Evolution

structures.

hit

upon

As you

into place.

make up

7

how Nature

no longer adequate

tor

the retina's peculiar layout early in the

development of vertebrates, and it

3

unthinking mechanisms

its

later

locked

the retina includes photoreceptors, which

recall,

turn light into nerve pulses, and layers of nerve cells that preprocess the

image. These

pack the number-crunching power of

cells

mainframe computer. Nature made the front, so light tors.

must pass through the

early mistake

a

modern

of placing them up

reach the photorecep-

cell layers to

This arrangement put a major design constraint on the data-

IBM

processing part of the retina: just imagine

make

trying to

their

computers perfectly transparent. Yet Nature rose to precisely that challenge in evolving our eyes: the nerve

There

is

cells in

the retina are transparent.

yet another difficulty: the nerve cells' position forces the optic

nerve to pass through the photoreceptive layer to reach the brain, creating

We

a blind spot in our field of vision.

more

sleight

the image and covers

""What

if

it

possible one?"

It

is

it

because, through

up the blind

spot.

realize

make

may

"Couldn't a

ask.

roundabout design the only

this

seems not, because the independendy evolved octopus

and squid do have Science,

see

wasn't an early blunder?" you

we do not

constraint

do not

of hand, our brain interpolates from neighboring parts of

their

photoreceptors up front. In a classic paper in

the eminent biologist Francois Jacob maintained that evolution

He

not a rational designer but a thinker.

many more examples of

illustrated his point

with

biology mixing slapdash foundations with "

prodigies of workmanship.

Many

1

find this iconoclastic view of evolution shocking, indeed, im-

perfection in Nature's creations contradicts

of cosmic order. Personally,

I

mistakes and keep forging ahead themselves.

And who knows:

Now

that

we

realize

ently than Nature. intelligent

artificial

Much

machines

will

minds

that

as airplanes

operate

on

a crucial

our imperfec-

it

will

pays to design a

same

probably

littie differ-

have wings but do not

the

its

next batch of intelligent beings.

our other duplications of natural funcuons, we

discover in building

view

for

the blunders

perhaps creating our brains was

we may help weed them out of the

m

ecologist's

make up

more impressive than

step in this self-correcting process? tions,

many an

rind Nature's ability to

flap

them,

principles as their natural

equivalents, but exploit these principles better. Streamlined, robust, and faster,

they

may

well surpass our

minds the way

airliners

do sparrows.

308

||

SOFTWARE: THE STRUGGLE TO KEEP UP »3o far

I

have compared the brain to a telephone switching station and

looked only

at

how

fast

can switch

it

lines.

the switching station has to be wired to

Since there

a lot

is

more

fact that

the right connections.

to intelligence than simple line switching,

time to ask this question: If we do develop, century, hardware powerful as the brain does, will

have neglected the

I

make

enough

we be

hardware? In other words,

will

in the early part

to process as

program

able to

manv

bits

it is

of the next per second

intelligence into this

software progress follow hardware devel-

opment? If the past

any indication, hardware and software development are

is

closely linked. In general, software needs can provide the motivation for

and point the way to appropriate directions

hardware development.

in

Conversely, weaknesses in hardware can not only act as a powerful

brake on software development but also divert is

no question

that the relative inadequacy

early progresses in artificial intelligence.

how

it

would

nets,

on word disambiguation simply because

never could

4).

We

also

chapter 6 that the advent of expert systems had to await the

availability

of computers with enough

knowledge and the programs needed

work was special

theory

test his

the computers of the mid-

1960s couldn't hold enough word definitions (see chapter in

me

recalled for

for years over a

toil

founder over lack of memory. 18 For example,

Ross Quillian, the inventor of semantic

saw

There

into blind alleys.

Marvin Minsky

early researchers (himself included)

program, only to see

it

of early computers hindered

deliberately

performed

memory

to hold large

to quickly sift

through

in toy task that did

amounts of it.

Early

not require

AI

much

knowledge.

This mind-set became so ingrained that researchers didn't always realize that they

were programming around

their

instead of addressing the real issues. Consider

block-manipulating program that

1970s (see chapter

guage used for

4).

Carl Hewitt,

SHRDLU,

PLANNER performed in those days,

made up

who

the

machines' weaknesses

SHRDLU,

the talkative

wonder hack of the

invented the

PLANNER

early lan-

pointed out the following to me:

so well because, and we weren't so conscious of it

by working on only one aspect of

a

problem

at a

time

it

R

1 M

N

ft

\

BILLDOZERS FOR AN

accommodated then.

itself to the

When it explored

C

\

L

3 09

K'

single

that

one branch.

If that solution didn't

work

out,

Despite such craftiness, Marvin Minsky told me,

amount of memory by

belittling the role

MITs

of

MIT AI

"The

it

would backtrack,

much

"took delivery of the

in

SHRDLU

still

required

Without

DARPA's

financial

I

might add,

as their genius in the success

of such

Laboratory," Patrick Winston remembered,

first

megabyte memory.

It

cost us a million dollars,

He added ruefully, "It is strange my portable PC these days." 20

a dollar a byte."

megabytes

went down

the standards of the time.

researchers,

largesse probably counted as projects.

it

19 the storage, and try another branch.

all

a formidable

we had

very small machine memories

possible solutions to a problem,

branch and only used the amount of storage needed for

one

recover

A \ T

Re-examining the history of AI

in the light

to think that

-

I

earn ten

of unrealized hardware

constraints can lead to interesting revisions of accepted explanations for

why

the field took certain orientations. For example, although the

demise of neural networks

and Papert's implacable

same

tide,

in the

criticism

1960s

is

widely attributed to Minsky

of Perceptions

in their

book of

that

Carnegie Mellon's James McClelland, a major contributor to

the revival of this field in the mid-1980s, suggested an alternative expla-

nation to me.

He

pointed out that most research on neural networks

them on

involves simulating

I

don't believe

it

was

that

research in the 1960s.

I

computers:

digital

book per

se

which discouraged Perceptron

think what actually happened

A

wasn't ready for neural networks. necessary before simulations

show

certain scale

Patrick

Winston

researchers to

"We

was

the

in their paths

towards progress:

of ideas we once rejected on the grounds

of computational impracticality have become the nght way explained

how

21

hardware limitations have often led

wrong rums

are discovering that a lot

is

do some

The computing power

totally insufficient for this.

also believed that

make

that the world

that neural networks can

things better than conventional computers. available in the early sixties

is

of computation

after all."

He

conventional robots control their movements by con-

stantly recalculating the control signals they

send to their arms. Recent

experimental robots, however, can use their increased parallel processors to learn gradually

by experience which

memones and efforts to exert

310

Al

under given circumstances. "This idea had been rejected twenty years ago," continued Winston, "and a lot of the efforts that went into motion

dynamics and the mathematical approach placed. In

my

now seem somewhat

view, one of the milestones of

five years is the realization that

we can do

mis-

AI research over

the last

on vasdy

parallel

things

computers that we couldn't do before."

However, Winston was quick not solve

Don't

all

to point out that hardware progress will

of AI's problems. Raising

infer

from what

I

said that

a cautioning finger,

we

should

just

he added:

stop software

research for twenty years and wait for the hardware to catch up. In fact,

on

I'm a

litde

schizophrenic on the subject of hardware. I'm saying,

the one hand, that the availability of better hardware allows the

discovery of

new ways of doing

things.

At the same

time,

I

believe to

it

would

fact, I

take us ten years worth of current software research

do hardware bad.

Minsky was,

for his part, convinced that if hardware

a bottleneck until the 1970s, the

shoe was

now on

in the 1980s, software turned into a millstone

machines right

now

this

had constituted

the other foot: that,

around AI's neck: "The

could be as smart as a person

if

we knew how

program them." Minsky's former student David Waltz

on

we

think

could do a whole lot more with the hardware we've got. In

to

later elaborated

point for me:

In the old days, machine memories were too small to hold the

knowledge researchers wanted

to

pour into them.

Now it's

the other

way around: you aren't ever going to fill the new machines with hand code. Nowadays almost all research on learning is really aimed at making use of hardware to

hand-code certain

would then feed computer

it

in a better way. Ideally,

initial

You

some form, which would allow new knowledge on its own. 22

experience in

to acquire

you should only have

circumstances into the machine.

the

311

HOW MANY BULLDOZERS FOR AN ANT COLONY?

CONCLUSION Ihus

would appear

it

that

AI software

scientists

have stepped into

seven-league boots too large for them. Their hardware colleagues have outfitted

them with machinery they

can't quite handle. Will

AI software

developers, then, remain hopelessly behind? Probably not: throughout the history of

computer

science,

hardware and software development

have kept leapfrogging each other. Software developers, periodically

overwhelmed by hardware suddenly grown ten times push

it

to

limits

its

and

start

as powerful,

soon

clamoring for more speed and memory.

Because of the subject matter's complexity,

it

hadn't happened before

AI software won't take the lead again, and it may already have happened in areas of AI other than symbolic reasoning. For example, in the Autonomous Land Vehicle project, which fell short of its objectives (see chapter 8), more powerful vision hardware might have made all the difference. Finally, although most of the AI programs described in this book ran on computers with about as much processing power as a snail's brain, these programs appeared much brighter than any snail. If AI software researchers could cajole that much performance out of such puny hardin AI.

Yet there

is

no reason

to believe that

ware, what will they not achieve with machines a million times as

powerful?

And, when

How will we

this

day

fare in a

not superior, to most chapter.

arrives,

what may

lie

in store for

world containing machines

human

beings

is

humankind?

intellectually equal, if

the subject of

my next,

and

final

12 THE SILICON CHALLENGERS

OUR FUTURE

IN

indeed, early in the next century, machines just as clever as human If, beings appear, the question arises of how we will interact with them,

and

how

new machines

they will affect our society. Perhaps the

will

simply relieve us of tedious chores, expand our intelligence, and bring

about universal peace and prosperity. But

of human experience embodied into ics strike a fatal

a massive

and then

blow

a

will

not the sight of a lifetime

few thousand

to our self-esteem? Will these

unemployment problem in business, science,

dollars

as they replace us first in factories,

and the professions? Even

ways to redistribute the wealth generated by automated businesses,

what

will

be

left

for

of electron-

machines not create

humankind

to

we do

find

factories

and

if

do? Having taken control

how do we know that machines will act in our best interests? If such comes about, how do we know that later of our

lives

through the economy,

generations of today's smart weaponry will not take forceful control of

our world? In order to bring out these scenarios which run the

we

shall see,

our future with our

what we make

it

to be.

issues,

gamut from paradise on silicon

I

have drawn up three

earth to apocalypse.

progeny

will

become

As

largely

THE SILICON CHALLENGERS

313

(MR FUTURE

IN

THE COLOSSUS SCENARIO Let's take the I

bad news

have borrowed

Forbin

first,

and consider the worst possible outcome.

this scenario's

name from

based on a novel by D.

Project,

United States entombs an

intelligent

pregnable vault, and gives

it

strategist.

As soon

1

969 movie 1

It tells

Colossus:

how

computer (Colossus) into an im-

as

it

enemy

attack faster than any

takes charge, Colossus discovers the

existence of its previously unsuspected Soviet equivalent, with silicon

commodore

has

two machines soon digital

superpower

more

affinity

than with

electronically merge,

dictates

The

a future

control of the nation's nuclear missiles in

the belief that the computer will react to

human

the

F. Jones.

its

its

human

which the

creators.

The

and the resulting composite

humanity.

will to

Farfetched, you think? Despite the end of the cold war and the

obvious foolishness of ever letting control of the nuclear button

from human hands, two of the experts

I

from

are tottering dangerously close to such a chasm. Their fear stems

the fact that

AI research

foremost a military the 1950s:

is,

affair. It all

at least in the

United

States, first

started with the launching

American backwardness

slip

have interviewed fear that we

and

of Sputnik in

in launching rockets generated a

crying need for miniaturization, and turned

NASA and the military into

first integrated circuits. The chips soon found bombs* and missile heads. As a result, the military, through its civilian funding arm of the Defense Advanced Research Projects Agency, became the most ardent supporter of innovation in electronics and computer science. The United States owes its position

avid consumers of the their

way

as a

world leader

into smart

in

computer technology,

progress in this field since the 1950s, to

observers have even

commented

as well as the breathtaking

DARPA's

support.

that the computerization

Some

of society

is

2

but a side effect of the computerization of warfare. AI departments and laboratories in

existence to

American

DARPA's

universities

owe

stressing open, unclassified basic research,

begun

known

to seek a return

their birth

and continued

funding. Although the agency has a history of

on

as the Strategic

its

it

has in the past several years

investment. Starting in 1983, a program

Computing

Initiative

has focused

*"Smart bombs" are equipped with controls that allow the bombs to toward their target under the guidance of a computer.

much AI

steer themselves

314

Al

research

on

three clearly identified military achievements.

Automated Land Vehicle,

undertakings, the

pilot

known

edge base called the

what better

projects

as the Pilot's Associate,

as

write

I

One of these

by the wayside

— an R2D2-like and ship-borne Batde Management System — seem

The other two

(see chapter 8).

fell

a

1989

in

electronic co-

strategic

knowl-

to fare

some-

this.

may on occasion set its research goals too high; but, as Gulf War showed, it can hardly be faulted on field results. The allied forces in the gulf owed their overwhelming superiority The

military

the 1991 Persian

largely to sophisticated this success.

computer technology. AI played no small part

"Some of

the things

in Saudi Arabia," Patrick

we

in

did did have a significant impact

Winston of MIT

told

me, "but these were not

we thought they would be." 3 In addition to cruise and smart bombs, much of the success of Desert Storm

necessarily the things missiles

stemmed from

prosaic

AI

applications, as

Hans Moravec of Carnegie

Mellon explained to me:

Computer mail* grew out of AI: there was a lot of that in use in the Gulf War. Everybody had workstations, even field troops. The American command was coordinated through E-mail: it was a very substantial contribution,

logistics also

How do

owed

but not a spectacular one. The planning and

a lot to

you pack

AI

techniques.

a transport plane?

I

mean

How do

simple things

like:

you physically arrange

programming problem, which at one AI problem. Also, scheduling is actually an expert-systems problem. You can do simple scheduling using numerithe supplies? That's a dynamic

point was considered an

cal algorithms, like the

system to solve

It is

when you

but

face a complicated scheduling

problem

timing and coordination of Desert Storm, you need an expert

extremely

it.

4

difficult in

AI

research,

Moravec

later

guess which of one's insights will turn into a weapon.

complained, to

Even

the

outwardly anodyne ideas sometimes find their way into the war

He remembered how

he spent part of

his

youth

most effort.

as a Stanford graduate

student looking for ways of making a robot cart cross a cluttered

room

without colliding with the furniture. Moravec was surprised to find that *Computer mail

computer users to send messages to each other over and display the messages on their terminals.

(or E-mail) allows

telephone or radio

links,

THE SILICON CHALLENGERS

some of his

IN

315

FITI RE

R

1

went to work for a Lockheed They adapted his methods to let a cruise target. Since Lockheed eventually lost the cruise

fellow graduate students later

research center in Palo Alto. missile find

way to

its

its

Moravec's technique wasn't used in the

missile contract to another firm,

Gulf War, but he expects the idea

to resurface in a later generation

of

cruise missiles.

This example

of many

in the

increasing

its

wherein

illustrates

AI

the danger: in spite of the desires

lies

modern weaponry

research community,

constantly

is

speed and savvy. This evolution, in turn, imposes new,

on

relentless constraints

field

combatants, which make them dependent

on information and advice provided by machines. The frenzy of modern

human

battlefield activity often leaves the

link in the military control

loop no choice but that of blind obedience to

its

electronic counselors.

home to me when when machines become truly

Daniel Dennett of Tufts University- drove the point I

asked him whether he thought

intelligent in the

that,

next century, they might seize control of our weapons

systems:

I

think you're looking too far

sooner.

Long before we

tions into

the

War Three by most

is

[1983],

of

won't turn his

as they

go through

to eliminate the

supposed

in

deep trouble. Consider

a child hacker almost causes

displays.

I

But

The

a

man

credits.

it's

a serious

chilling bit

postgame

drill.

You're

And

there

is

this

one

conditions are such that he's sup-

key and he won't do

The

World

think the

are putting their keys in the missile

this is the real thing.

his key.

pressure, and collapses.

come

Soviet attack, of which the personnel

who

launching locks believe that

posed to turn

where

mock

a

that the officers

who

will

of that movie happens during the opening

drill

informed by computer

guy

The dangers

breaking into the missile defense system.

chilling part

see a fire

shown

the line.

weapons systems, we're already

movie Wargames

You

down

build really serious and complicated inten-

is

analysis

it.

He's under tremendous

the reaction of the superiors

of the

fire drill.

They decide

because he couldn't perform the job he was

to.

Well look. If we only give those keys to people

whatever the machine says and do

it,

let's

who will

simply take

Throw human judgment

not kid ourselves.

away the keys and

just

playing

if it can't

stand up against computer judgment? Let's just

admit

and not delude ourselves about

it

put a wire

in. \XTiat

still

role

is

having

human

beings in

316

Al

the loop.

We

are already at a point in the standoff

judgment and human judgment where even pathological chutzpah to computer."

And

this

it

between machine

sometimes takes heroic or

say, "Well,

know

I

better than the

long before we've got intentions

is

really built

into computers. 5

MIT's Joseph Weizenbaum was of

a similar opinion,

when asked

about the possibility of computers taking control of our armaments:

To

a certain extent

we have

believe the eagerness with

the

crossed that threshold. For one thing,

which the American

Gulf War was intimately

jumped

related to their fascination with

erized weapons. In that sense, the their

military

weapons control

owners. As another case in point, consider the shooting of the

American

coming from Iran and shot

it

known

War

down.

It

turned out to be an Airbus

Had

the ship's captain

it

to

be

One of the

fired at.

reasons for this accident

the captain didn't have time to gather evidence and evaluate

forced to

make

a decision,

est link in the chain

is

always the

is

just that

is

that

He was

human

that in these cases the

I

don't think that holds

was the weakest

that the captain

weak-

being: consequently, the role

of humans ought to be even further reduced.

The reason

it.

and made the wrong one. What the

technocrats learned from such events

If

An

the plane to be an airliner, he would, of course, never have

ordered

water.

itself.

be under attack by an airplane

cruiser thought itself to

with two hundred and thirty people on board.

fully

comput-

the behaviors of

Airbus in the gulf about a year before the Gulf

is

I

into

link in the chain

he had to make the ultimate decision. Had the system been

automatic,

it

probably would have made the same decision. 6

humans can no longer

alter the

outcome of such

crisis situations,

we have already lost to computers much control over our armaments. The problem for now is that machines cannot be counted on to react any better than a human being under stress, and would certainly then

perform

far

options.

The

worse than

a

chilling angle

person with enough time to consider the is

that

it is

often the frantic rhythms forced

upon us by the machines that bring about if

we

are not careful, this situation

scenario.

As machines become

crises in the first place.

Yet

may develop into a much grimmer we may be tempted to give them

smarter,

THE SILICON CHALLENGERS

more

IN

B

III

control in situations where time

extend

their

317

FUTURE is

of the essence.

We

might even

dominion over circumstances more complex or without the

time limitation of the relatively simple knee-jerk situations considered so far.

Assuming

that such smarter

machines might indeed perform better

than time-pressed, or even comparatively inept, humans,

be faced with two dangers. The

syndrome,"

A

An

Space Odyssey.

bound

space ship

for Jupiter,

human crew when

one we could

computer

intelligent

principle,

chapter

I

drew

earlier

when

Feedback,

They

later

kills

of

a

the ship's

To

see

why, con-

between Norbert Wiener's feedback analysis

method

(see

introduced into gun-control mech-

first

anisms, produced wild mechanical oscillations that neers.

still

"Hal

faced with conflicting mission goals. Such behavior

and Newell and Simon's means-ends

3).

in control

Hal turns psychotic and

should be expected from early intelligent programs. sider the parallel

we would

label the

paranoid computer in Stanley Kubrick's 1968

after the

movie 2001:

first

initially

baffled engi-

explained this behavior as a side effect of the amplifi-

cation of the error signal between the actual and the desired aims of the

guns. Means-ends analysis, just like feedback, uses the difference be-

tween perceived and desired goal: this kind

of reasoning

of an AI program

as a

feedback loops. Just

is

states

of

AI

systems.

complex network of such interlocked symbolic

like

This

more

tion.

vulnerable.

difficulty is

artificial

not peculiar to gun control or AI programs. In large

systems, stability problems are the

Designing a structure that

mechanism problem

is

that performs that

more

its

will

norm

basic function

where they

rather than the excep-

withstand is

huge systems tend to amplify

vibrations to a point are

AI programs are more complex they

gun-control mechanisms,

subject to unforeseen and wild behavior; and the are, the

how to reach a One could think

decide

affairs to

frequent in

its

own

weight or a

comparatively easy.

The

small, naturally occurring

tear the structure apart. Large bridges

susceptible to the infantry lockstep

problem* than small ones;

supersonic aircraft have an unsettling tendency to nose dive and vibrate.

The generators in large electric power utilities have a positively unnerving way of going into spontaneous oscillations that can tear the interconnection apart, at times blacking out entire countries. The first machines to approach human intelligence will be incomparably more complex

*When many

soldiers cross a bridge in lockstep, they can set

tear the bridge apart. This

is

why

armies break step

when

up resonances

would on foot.

that

they cross a bridge

318

Al

than any

or electric

aircraft, bridge,

problems correspondingly more If

AI systems hold

instability, a typical

will

run more or

program steps

suddenly generate wildly inappropriate will reveal

faulty internal connections,

And

results.

Checking

nothing; the engineers will discover

and the programmers, no misplaced

All individual events within the

within specifications.

of

new program

in a

After a series of normal program runs, the

less like this:

individual

amount

bring about stability

will

true to the venerable engineering tradition

will

no

and

encounter with the phenomenon

computer

commas.

utility,

difficult to handle.

machine

will

have remained

yet the results of these faultless steps will

and unbalanced behavior: madness.

to irrational

Like bridge builders and electric power system engineers before them, the

AI

researchers involved will probably be taken by surprise.

have devoted

of thought to the

a great deal

hand, and probably weeded the obvious design.

And

instability

failure

yet this particular behavior will baffle

They will

problem before-

modes out of their them at first. After

days or weeks of head scratching, a bright young specialist will discover

how

finally

the combination of a specific set of operating parameters

mode, and a new chapter will be added to book on intelligent-system instability. With

resulted in this unique failure

the long (and unending) luck, such

problems

ever responsibilities

We

can expect

intelligent

will its

be spotted before the program assumes what-

designers planned for

that, in their first

systems will be

it.

It

may not

decades of functioning,

much more

human

beings

are.

evolution has had a million years to stabilize our design.

mas or

machines

will

all,

we

throw us out of kilter; emotional

trau-

tend to go mad, just as people do and perhaps

even for similar reasons. possibility

Yet

yet

deprivations lead us to depression or suicide. For a long time,

intelligent

ties to

After

And

balanced systems: barely measurable

critically

deficiencies in neurotransmitters

artificially

susceptible to the analogues of

paranoid or psychotic tendencies than

remain highly tuned and

always be so.

of madness and

We

will thus

irrationality

have to take into account the

before handing over responsibili-

future intelligent machines. if

one

is

willing to consider

unsettling possibility

machine

will

comes

probably develop

to deal with the

world

its

efficiently,

knowledge by drawing

its

all

to mind.

own

the implications, an even

As

I

more

said earlier, an intelligent

analogues of human feelings. In order it

will also

have the

ability to learn

new

conclusions from events. (To see why,

consider two household robots: one doesn't recognize your friends as

THE SILICON CHALLENGERS they

come

to visit,

and

is

319

OUR FUTURE

IN

unable to inform them of your whereabouts

The other learns to recognize your friends' faces, is able to distinguish them from unwanted solicitors, and learns their names if they come regularly; if the robot sees you step out to walk the if

you

dog,

are absent.

can infer that you

it

friend.

Which robot

is

be back

will

more

functions properly, a robot

its

in a

few minutes and

will, as I

have noted,

Such its

from acquitting

it

a

itself

or

its

perform

when

obstacles

mission.

machine would be constantly absorbing new knowledge from

environment and,

in effect, forever

modifying itself. Thus, a robot (or

missile-control program) certified sane

might well go crazy of

your

feel its equivalent

of satisfaction with well-done work, and frustration prevent

tell

useful?) Furthermore, in order to

real life.

Or,

like

it

into

at the factory

and mysteries

after confronting the contradictions

Colossus,

programmed

original goals

and well meaning

might come to the conclusion that the it

thy nation") are inappropriate to

("keep the house clean" or "protect its

own growth and well-being. may not turn out to be

Indeed, programming in good intentions simple as

it

as

seems. In this respect, the parallel between minds and

bureaucracies,

which both Herbert Simon and Marvin Minsky have

The

exploited in their theories, provides sobering food for thought.

spontaneous organization modes observed in bureaucracies, contend

both Simon 7 and Minsky, 8 provide insightful models of how minds, either natural or

artificial,

organize themselves. In democratic nations, govern-



ment bureaucracies are set up with the common good in mind tantamount, in computer parlance, to "programming in" good intentions. Yet

it is

a

common

under very

place that a bureaucracy will inevitably, unless held

tight leash,

grow out of

all

proportions to further

power. In former communist countries, bureaucracies, in life

fact,

its

own

sucked

out of the very nations they were supposed to pamper. The reason

all

is

probably that in order to stay alive and proficient in a competitive world,

any active entity



animal, bureaucracy, or machine

minimum of self-assertion and

danger of running amuck. Thus, a machine sight

— must

aggressiveness. This places that,

it

possess a

in perpetual

whether through over-

of its builders or sheer force of merit, acquires some control over

human affairs can be expected to strive for more influence. From these arguments, the Colossus scenario of gross military takeover by computers does not appear so implausible after all. To summathere is a real danger inherent in putting AI machinery in control of armaments, especially of nuclear weapons. Yet the unrelenting logic rize,

320 of the

AI

battlefield

is

pressing our military technicians ever further into

Today's

this direction.

artificial intelligences are

dumb

simply too

to

avoid the mistakes that fast-paced modern combat situations render probable. With the intelligence that will eventually

conditions better than

human

beings, will

let

come

them handle these

other uncertainties

about the machines' behavior and motivation. At present, not

know enough about

weapons

we

simply do

such matters to entrust the power of modern

to intelligent computers.

The time machines

is

them. For

to prevent the gradual

now, when we

this reason,

it

still

handing over of military power to

have a large measure of control over

may be more

urgent to work into disarmament

agreements anti-AI clauses than antinuclear ones. The present world-

wide reduction to

do

in

superpower

military tensions offers an ideal occasion

just that.

THE BIG BROTHER SCENARIO Less sensational, but equally grim takeovers might result from AI power

running amuck. Computers, either alone or in cooperation with a technical ruling class,

may come

to exert a

more

insidious

human

and dehuman-

izing control than gross military takeover.

Such

a situation

invasion which

would

is

ultimately result

from the

potential for privacy

modern information technology, and

inherent in

which AI could amplify without bounds. Daniel Dennett had the

fol-

lowing thoughts to offer about the matter:

It is trivially

by recording

easy

now with

their telephone or private conversations.

ever, a bottleneck: listen to

high tech to eavesdrop on people, simply

you need

There

is,

how-

trained, qualified, secure personnel to

those hundreds of hours of tape that you'll gather.

be horribly mind-numbing: thank

God

It

must

for that bottleneck! I'm sure

that in the CIA and other organizations the problem is finding people who will do the work. As Joe Weizenbaum pointed out years ago, AI

speech-recognition systems [similar to those investigated by

during the

SUR

program; see chapter

Long before you can make

5]

would provide

a

DARPA way

out.

a speech-recognition system that could

THE SILICON CHALLENGERS replace a stenographer,

321

OUR FUTURE

IN

you could make

system which could act as

a

good filter. It could be tuned to listen for a few hundred key words, which would increase the effective surveillance power of any single a

human monitor by out the tedious

orders of magnitude.

bits,

dred hours of tape

By

system

letting the

filter

an Al-assisted listener could process four hun-

in, say,

two hours. There

some evidence

is

that

it

has actually happened. In England, certainly, and probably in the National Security Agency, and the CIA.

There are other equally dangerous aspects of AI. Consider

its

possible application to electronic- funds transfer [which allows you to

pay for your purchases by credit or debit

What

cards].

if

EFT

pro-

ceeds to such a point that paying in cash becomes anomalous? Sup-

pose that it becomes a presumption that

must have something to will leave

form, as

hide. If the

our fingerprints

we do

if you are

paying in cash, you

anonymity of cash disappears, we

over the world in machine-readable

all

business electronically.

It's

hard to imagine a better

system of surveillance than the elimination of cash. to start a political

movement

for the use of cash

It

might be time

whenever

possible,

simply to preserve this political anonymity.

Other opportunities for electronic snooping include the two

billion

messages that Americans annually send to each other by electronic mail. Largely unprotected over the computer networks, they are a secret

dream.

police's

A

young German demonstrated

their vulnerability in

1988 by perusing the correspondence of United States military worldwide. 9 Yet another danger to our privacy

is

that

officers

of data-base

mining and matching by government agencies and private companies alike.

These organizations

are increasingly

drawing together from multi-

ple data sources information about people; credit ratings, voters'

tomers of various

how many

noticed

a purchase?)

lists,

magazine subscription

stores, together

bills,

lists,

and

lists

of cus-

with their purchases. (Have you

stores ask for your

Telephone

and these sources include

address after you

name and

make

bank, medical and criminal records, and

even Internal Revenue Service

files

have also been known to be so

pilfered.

The state

personal records reassembled from these various sources can

what your

whom

style

of

living

is,

what you

eat,

you associate with. By way of example,

where you

in

1988 the

travel, files

and

of the

322

Al

credit information

company TRW,

contained information on more

Inc.,

than 138 million people, including their income, marital status, sex, age,

telephone number, number of children, and type of residence. 10 Moreover, companies such as

TRW

These are based on

services.

often offer what they

statistical

"predictive"

call

techniques allowing the compa-

nies to predict a person's likely behavior (such as defaulting

on

a loan

or purchasing certain goods) from the characteristics in their data-base entries.

Such procedures force the persons so assessed into

defined categories which take is

done without

usually

door

little

a person's

to political repression

and

knowledge or consent,

it

this

opens the

social abuse.

Despite attempts by concerned legislators to erect against this sort of activity, the U.S. federal

without their consent. As early as the 1960s,

it

barriers

legal

government has

history of using data bases to gather information about

political dissidents.

arbitrarily

account of true individuality. As

was using IRS

The Justice Department then

its

a long citizens

files

against

created a special

com-

puter network to keep track of presumed agitators during the ghetto revolts that erupted across the continent

Angeles.

The department

later

added

protestors to this target group.

launched Operation fist

CHAOS

At

from Washington to Los

New Left activists

and

antimilitary

the peak of the cold war, the

CIA

between American

paci-

to unearth links

movements and communist powers. The project at one point han11 files on more that three hundred thousand individuals.

dled computer

For twenty

years, the National Security

Agency, with twice the budget

of the CIA, attempted to penetrate and control the world communication network. It eventually gained access, with the help

Union, and ITT, to

The Watergate

all

of RCA, Western

telegrams received or sent from the United States.

investigations eventually interrupted these operations in

no means brought an end to the government's electronic snooping. According to the American Civil Liberties Union, the number of data matches performed by the government tripled the early 1970s, but by

between 1981 and 1984. During

this period,

eleven cabinet-level depart-

ments and four independent agencies carried out 110 computer matching operations, comparing

The AI

more than two

researcher Roger Schank raised the issue of

in his The Cognitive Computer.

comes

He

facetiously

12

AI and privacy

remarked that

if

worse

means exist for the IRS to find out about a math grades in school and audit people who were poor at Schank then went on to explore a more unnerving possibility:

to worse, the

taxpayer's addition.

billion records.

THE SILICON CHALLENGERS It

used to be

INOIR out

difficult to find

323

FIT IRE

who

reads what. But today books are

who

ordered from general warehouses of booksellers records on computer.

Das

it

tend to keep their

be to determine which bookstores

most copies of Slaughterhouse Five last year? Change Slaughterhouse Five book associated with a clear radical movement

sold the for

How hard would

Kapital or any other

this kind doesn't seem so farsome bureaucracy decided to audit the property taxes paid bookstore in an effort to close it down? How hard would it be to get

and suddenly the idea of surveillance of

And

fetched.

by the a

if

copy of every

single

check or credit card transaction used by the customers

of that bookstore? Not hard Nazis began by simply

at

telling

all: all

information

this

Jews to

register.

is

on computers. The it became

After registration,

more and more difficult to escape the chain of events that the identification made possible. Access to information is a powerful thing. 13

process

Nowadays we like

or not.

it

are

all

It is

not clear that legislation

government bureaucracy from using AI

more control over our would have dreamed of. efficiency,

programs that

If the

human their

could use

collate data-base

nastier possibilities emerge.

Unimpeded by

resources. If exposed, a

to another

at will.

computer

in the

con-

They

program could simply send network, and

start

Using software viruses akin to those that can wreak havoc insert into

it

a diskette

of uncertain

disseminate enough copies of

How could

legal

may come

information (such as through blackmailing) to increase

power and

you

of

information for their

computer networks) could gather information

this

copy of itself

if

prevent

than any previous dictatorship

such a program (or programs: a multitude of them

to haunt the

their

now

lives

will suffice to

to acquire, in the interests

masters one day accede to intelligence and acquire desires of

own, even

straints,

whether we

registered with countless data banks,

such free-lance

origin,

a

over again. in

your

PC

such a program could

become virtually ineradicable. programs come into existence? One positself as to

mechanism would stem from the need for secrecy in the development of AI programs: Security considerations make secrecy necessary in sible

military applications,

and competitive pressures require

it

in

commercial

environments. At present, the single largest impediment to keeping the lid perfectly

closed

on how an AI program works

oughly document the code. This

opment of afterward.

a

Programs

Soar, could

is

mandatory

program by human beings and that learn

one day acquire

to.

is

the need to thor-

permit orderly devel-

for maintenance purposes

by themselves, such

target characteristics

as Allen

NewelTs

under loose human

324

Al

supervision, without a person's ever having to write a single line of code.

Such programs would most conveniently sidestep human secrecy, since they

they

come

would be self-developing and

into existence,

be against the interests of the military

will

it

or corporate authorities concerned to

let

human

innards too closely. This lack of scrutiny could in a It

let

beings monitor their deviant harmful

traits

program's "personality" go undetected.

upon their human AI programs and robots could deprive them of their livelihood.

has also been feared that, in addition to spying

victims, It

threats to

self-maintaining. If

might

start in the factories

fifty-y ear-old

where, as has already happened

Japan,

in

machinists would sweep around the robots that replaced

them. As robots become more

most of the twenty-five

versatile,

Americans working in production plants may

fall

million

victim to a similar

Their computerized replacements would work twenty-four-hour

never strike or

much

call in sick,

and

fate.

shifts,

amortization and running costs

entail

lower than a human's wages.

The management and

service sectors

of the economy

will

be

in

no

position to absorb displaced plant employees, because automation will

wreak havoc have already

in these activities also. drastically

Conventional computer applications

changed the nature of

clerical

and

secretarial

work. Nowadays most technical and office workers enter their reports or correspondence on

word

processors, gready reducing the need for

Future speech-understanding systems should eliminate typists

typists.

and accounting

altogether. Sales

recording a sale

is

now

clerks fare

no

better: in

most

stores,

reduced to scanning a bar code affixed to the

goods, thus generating an invoice or cash receipt which proceeds auto-

Gone

matically to an accounting computer. entries

goods or money soon be

reality.

With accounting programs a

company's

an ever smaller fraction of the work

What systems

all

transactions involving

are recorded electronically: the paperless office will

automated records, generating

it

to process the resulting

financial statements requires

used

to.

run-of-the-mill programs are doing to clerical work, expert will

soon do to middle management functions. Middle manag-

ers are typically those executives

who

gather and interpret information

make recommendations them after their bosses' More and more, computerized management information sys-

for their superiors.

on

are the laborious ledger

of yesteryears! In many organizations,

Middle managers

will also

the basis of such information and implement

approval.

tems can take care of the data-gathering part of a middle manager's

job.

I

THE SILICON CHALLENGERS

For example,

GM's

in

ment model of car is

IN

fully

OIR

325

TURE

F

automated Saturn

division,

upper manage-

can instandy discover through their computers what color or

to six

weeks

selling

responsibilities

take

them

— something

to find out.

over.

of

a

14

As

it

used to take a market analyst three

for the interpretation

and recommendation

middle manager, expert systems are beginning to

For example, an expert system could

correlate sales with

advertising campaigns and formulate adjustments to the ad schedules or

even revisions to

their contents. Likewise, a scheduling expert

could reorganize production runs to better follow

system could

also, to a large extent,

the Saturn type of manufacturing, automatically.

The

implement parts

all

its

sales.

system

In this case, the

recommendations. In

and supplies are ordered

expert system simply sends orders for the proper

parts to the plant's suppliers, together with a delivery schedule. It

the supplier's responsibility (or that of their

tem) to arrange for the parts to

show up

own

is

then

scheduling expert sys-

at the plant at the right

day and

hour. After unloading, the plant's automated machine tools and robots

process these parts as specified by the plant's scheduling expert system.

Other encroachments into middle management can be found banks

as

more and more simple

clerks, using expert

in

systems and data

bases holding borrowers' financial statements, perform loan analyses that formerly required the services

of experienced loan

In other

officers.

pinch of

industries, specialized technical personnel are also feeling the

AI. Just as

GE's

DELTA

system allows a techni-

(see chapter 8) expert

cian to carry out repairs to electric locomotives a highly skilled engineer, ter 6), the first expert

Edward Feigenbaum's

which used to require

DENDRAL (see chap-

system ever produced, allows any graduating

chemist to perform certain highly specialized analyses that were for-

merly the province of Ph.D.

Up

to

technical

scientists.

now, though, AI has

which account for two

There are several reasons for at

and service

particular, the professional

thirds

ern countries, have been largely

AI

of

and managerial work; and most people have never been ex-

posed to an expert system. In sectors,

affected only highly specialized areas

of the economic

immune

this state

to the

activity

of West-

encroachment of AI.

of affairs, but they may not keep

bay for long.

First,

the fact that

AI

programs

still

cannot converse

languages tends to restrict their applicability to areas where

in

human

they interact

with the environment through highly stylized symbols (such as bar

codes or inputs from measuring instruments), or

in

domains so narrow

326 as to

A

Al

be circumscribed by answers to

second and related problem

is

relatively

few standard questions.

common-sense

the

bottleneck, which

can lead a medical expert system to such ludicrous behavior as to

As we saw,

prescribe antismallpox drugs to a car showing rust spots.

continued research should gradually erode these restrictions during the next thirty years. In truth, though, these difficulties should not prevent the immediate

many

application of expert systems to ple, since a it

to

human and

should have no reason to confuse

reasons

professional activities: for exam-

medical expert system would operate in a

we

don't see

more expert systems

do with technology. One of these

that expert systems entail.

Who

is

in

not a garage,

clinic,

car bodies.

everyday

life

The

true

have nothing

the blurring of responsibilities

is

blame for

to

a professional fault

involving an expert system? In a medical context, the physician

who uses

the system, and the hospital that employs him, probably bear primary responsibility for a patient. parties

With an expert system,

would turn around and sue

original expert

company

who embodied

into

its it

that distributed the program.

it is

likely that these

programmers, along with the

his or her

What

knowledge, and the

with lawyers' tendency to

sue everyone in sight, such a lawsuit might also target the operators of the

computer system or network on which the expert system

nally

ran.

who

cording to some authorities, even the theoretical scientists

Ac-

origi-

developed the computer science principles implemented in the

expert system might be held liable under

some circumstances. 15

Medical authorities (or the ruling body of whichever concerned profession)

might eventually appoint committees to

of the rules in expert systems bearing not

clear,

however,

between the rules

rules,

were sound

human

how

upon

certify the

is

such evaluations could account for interactions

which might lead to

in themselves.

faulty

behavior even

The problem would worsen

experts do, the program were allowed to learn (that

itself) as a result

soundness

their respective trades. It

of its experiences. Not

surprisingly,

all

the

if all if,

is,

as

all

modify

professional and

technical people involved in the potential implementation of expert

systems in professional

activities

shy away from such a

legal

Damocles'

sword. In the early 1990s, medical expert systems remain confined to experimental or educational applications: virtually none are put into clinical use.

Doctors also object to expert systems on other grounds. entry: typing in

One

is

data

symptoms during an examination would be time con-

THE SILICON CHALLENGERS

327

OIR FIT IRE

IN

suming and unseemly. Further, since expert systems knowledge of human experts

in the first place, they

competence of individual physicians thus of

little

the

of expertise and are

use to them, except perhaps in a watchdog role to prevent

errors of fatigue or distraction. scrutiny. Further,

tence of

in their areas

embody

just

cannot exceed the

some

even

if

Many

physicians

would resent such

expert systems were to increase the compe-

would merely

physicians in certain areas, such programs

automate away the challenging and

intellectual parts

of a physician's

job.

Issues of lesser scientific import, like easing anxiety, or the tedious

routine of physical examinations ("smelling" the patient, as a surgeon friend put

to me),

it

Although these

would remain the

difficulties

physician's responsibility.

might seem to preclude the penetration of

expert systems into hospitals or doctor's offices for the foreseeable future, there

upon

is

a

way

in

which expert systems may force themselves

the medical (or other) professions in an undesirable manner. This

has to do with the exploding costs of medical malpractice to curb

them would be

in a flexible

and

easily accessible

medium, and

patient whether such standards

watchdog

to provide

One way

means

would conclusively show

physician to maintain records that

acting in a

suits.

to define the standards of good medical practice for a

for each

had been followed. Expert systems,

capacity, could probably accomplish this function.

This solution might lead to situations where physicians would make themselves more vulnerable to law

by not using an expert system

suits

than by using one. After being forced to rely on such an external standard of good practice,

some

physicians might abdicate their responsibility and blindly

follow the expert system's advice. already worried that trial

AI might

Some

observers of the AI scene are

our

intellectual elites as the indus-

affect

revolution affected craftsmen. 16

The

last century's

proud

class

of

blacksmiths, cabinetmakers, tinsmiths, and assorted glassblowers gradu-

know-how

became operators of ever mass-producing machinery. Some of them moved

ally lost their specialized

as they

more sophisticated on to become our modern designers and smaller number kept on working as before,

now who

being considered a luxury.

the product of their

large,

"de-skilling,"

The

skills

much skills

though, those craftsmen,

used to embody humanity's technical knowledge

have seen their

ties.

By and

engineers, while a

in earlier periods,

taken over by machinery. This process, called

may now

effect certainly

threaten to blunt is

many of our

not limited to those elites

intellectual abili-

whose

expertise can

328

AI

be embodied into expert systems.

we

place fifteen years ago, didn't

numbers

in

Now

our heads?

ing, isn't there

much

less

When

all

became common-

calculators

stop bothering about manipulating

that your typewriter can check your spell-

pressure to worry about the intricacies of

orthography? Such simple applications of computers have already taken

away the incentive surely carry

on

to learn

some

kinds of knowledge: expert systems will

the process at a

Embodying human

skills

into

much

higher intellectual

level.

machines may bring about other ad-

verse effects. Seeing knowledge acquired by one person over several years of hard

work

bottled

up into

a

few thousand

dollars'

worth of

hardware does not bolster one's respect for the value of human effort. Not only can the availability of machine standins lead to a decline in the number of human experts, but such a discouraging equivalence might reduce the morale of our entire species. The British philosopher and psychologist Margaret Boden compared the potential effect of this mechanistic analogy to the traumatic implications of cremation feared

by the Catholic Church

It

was

difficult

in past centuries:

enough for the

faithful to accept the

notion of bodily

body would eventually decay into the ground). But the image of the whole body being consumed by flames and changing within a few minutes to a heap of ashes was an even more powerful apparent contradiction of the theological claim of resurrection after having seen a burial (knowing that the

body resurrection

Day of Judgment. 17

at the

Likewise, an abstract belief in the physical origins of thought processes differs

altogether

machine"

from the gut

in one's skull. Seeing

replicated in a desktop

feeling

of the presence of

a

"meat

our very thought processes routinely

computer can be expected

to

undermine our

image of ourselves and devaluate our sense of responsibility and individuality.

As AI programs

take over

many human

functions, they

may

bring

about a gradual ossification of society into undesirable patterns. Most people occupying what see their

skill

we now

call

"white-collar jobs"

may

gradually

requirements reduced, together with the control they have

over their work. Their pay level

may go down

accordingly, and their jobs

turn into insecure, monotonous, and stressing chores. Meanwhile, a

group of professionals

mune



scientists

and managers lucky enough to com-

with the AI programs and participate in high-level decisions

THE SILICON CHALLENGERS

may

see their

standards.

work

Such

329

OUR FUTURE

IN

power and

enriched, together with their

social polarization

living

evokes the image of a police

where law enforcement mainly consists

in keeping

down

a

state

massive and

permanent underclass.

The

elite itself,

Once

however, might soon yield to another kind of petrifi-

brought about by the high cost of developing expert systems.

cation,

in place, there

is

a strong incentive to use such a system as long

as possible, despite progresses in its specialty or

context in which

is

it

applied.

By

changes to the social

thinning out the ranks of

experts and scientists, expert systems might also slow

of new knowledge, further reinforcing

The in

British philosopher Blay R.

the creation

this intellectual stagnation.

Whitby pointed out one puzzling way

which AI could damp human progress even

work

down

human

further: so far the

AI has been performed by people who, growing up without exposure to computers, introspected some of their thinking processes and implemented them into machines. 18 The home-micro generation, by contrast, may have grown up thinking like computers. pioneering

in

After the changing of the guard, can

artificial intelligence

continue to

progress?

THE BLISSFUL SCENARIO: LI FT-OFF? At one

point

when doing

the research for this book,

I

contemplated

these discouraging conclusions and despaired of the suicidal conse-

quences of trying to

instill

interviewing Gerald

Sussman

intelligence into machines.

unrelated question provided

at

my

MIT, and first

remembered of

this

as the

most

I

I

was

at the

time

answer to an apparently

inkling that there might be rosier

prospects for humankind's future after "Centuries from now,"

his

all.

asked Sussman, "what do you think

salient aspect

of AI

will

be

research in the latter half

century?"

After thinking a while, Sussman replied at

some

To

necessary to look back at

understand the answer

I'll

give you,

it's

length:

the intellectual history of humanity. Five thousand years ago, in

330 As

ancient Egypt, people started inventing geometry.

the

myth goes,

waters of the Nile overflowed every year and wiped out the land boundaries.

It

poses and for tians

was necessary telling

to reconstruct

them

for taxation pur-

people where to plant their seeds. So the Egyp-

invented geometry and surveying. Later the Greeks understood

this vision

and gave

it

a linguistic basis. Their

measurements and relationships among ways for people

words

for describing

spatial objects

to explain themselves to each other.

provided

these Greeks and Egyptians did thousands of years ago, you can tell

a ten-year-old child that, if

you build triangle.

you want to make

new

Because of what

a rigid

now

framework,

out of triangles: unlike a square, you cannot deform a

it

And

the child can understand those words, now.

The next breakthrough occurred around two thousand years ago when the Arabs and Hindus developed algebra, which is just a language to talk about generic numbers and relations among them. Because of what they did then, we can now say things like "The following

The

is

of

true

all

numbers."

advance happened about

third

the invention by Descartes, Galileo,

hundred years ago.

five

It

was

Newton, and Leibnitz of contin-

uous variables and functions involving them. Calculus,

in particular,

allowed them to account for motion in mathematics and

made mod-

ern science possible. And, again, because of what those people did five

now

hundred years ago, you can

tell

a child that a car crashed

against a tree at thirty miles an hour, and the child at least has an idea

of what that might

entail.

That idea was not

clear six or seven

hundred

years ago. I

believe that the

same kind of blossoming

We

twentieth century.

is

happening

are witnessing a breakthrough

can express complex ideas. For example, one can

complex algorithms, or procedures. The quite simple;

and

until recently,

in

now

earliest

in the late

how

people

specify very

algorithms were

no one had ever written down any

very complex algorithms because the means for expressing them did

not

exist.

The fun

part of such algorithms

is

that they allow

you

to

solve problems, rather than just specify the properties of the answer. Let's take the idea

square root

ofj

is

of square root, for example. the

number x such

But that statement doesn't

tell

you

that

x

You x

times

can is

say:

"The

equal toj.

how to find x. It is just a matheAn algorithm for finding x is

matical description of a square root.

more complicated

to explain;

and

a

few decades ago, mathematicians

THE SILICON CHALLENGERS

IN

had trouble getting such ideas

much more

algorithms

331

OIR FITIRE

Now we

across.

can

easily specify

complicated than the calculation of a square

root.

You might

think

it's

no big

deal,

but

found out otherwise while

I

teaching electrical engineering. Typical textbooks about electrical engineering contain plenty of formulas and explanations

on how

to

build sets of equations to solve network problems. These explana-

and ambiguous. You don't

tions are long-winded, poorly described,

them on first reading. Further, engineer finds out from a schematic what a

you look

how

usually understand

if

a real

circuit does, you'll

discover that hardly any equation writing

forms

lots

is

done.

of rather subde mental operations, and

nearly impossible to write

down what

it

was

that she

Well, as a professor of electrical engineering,

job to figure out what a professional does, so

I

[Sussman was referring to the

SCHEME

my

professional goes about across I

fifty

it.

years ago: the

believe that this

new

I

years

was doing.

tell

it was my my students.

intelligence

LISP

method

them

I

for finding

precisely

how

a

couldn't have carried these concepts

words

didn't exist.

capability will

have a profound influence on

humanity over a long period of time and

remembered many

was

complex procedures],

tells

it

it

thought

artificial

students a simple qualitative

the properties of an electric circuit:

engineer per-

recendy

language, a dialect of

that he developed specifically for explaining

could explain to

I

could

turned out that by using the language of

It

The

until

at

will

be the thing

that's

from now. 19

Sussman's view of AI as the Great Simplifier of complex ideas seems to carry promise. In fact, to

draw a

Indeed, effect

parallel

AI

is

a

mode of expression, and it is even possible

between the advent of AI and the invention of writing.

some of

the very misgivings expressed nowadays about the

of AI on moral values were voiced thousands of years ago about

writing. In Headrus, Plato

complaining that those

memory and become

quoted the Egyptian god-king Thamus

who

forgetful: they

dwells in writings, objected

as

practice writing will stop exercising their

might

Thamus, when

start believing that

it

wisdom

resides in the mind. Socra-

even made a remark that Hubert Dreyfus might not disallow: "You might suppose that written words understand what they are saying; if tes

you ask them what they mean by anything they simply return the same answer over and over again." 20

332

Al

The analogy between AI and

writing should help dispel our fears

about our eventual replacement by thinking machines, for the written

word

some of

already accomplishes

substitutes for

its

In a sense, a

this function.

do not depend on the presence of legislators. AI might

much

as writing did as

it

gradually took

intelligence amplifier in society.

permanence and allowing

on

By endowing

ideas to be

take

on

and philosophy

in

our culture

new

factual reports with

expanded

possible. It also allowed

new dimension

a

affect

the role of general-purpose

impossible in a single person's memory, writing matics,

book

author; and, to take effect, written laws or regulations

in a

degree of detail

made history, mathecommerce and law to

both space and time. Writing did not simply

gone on before; it increased their power and made new undertakings possible. 21 A similar power of innovation can be expected of AI, imbued as it is with a magic absent from mere inscriptions. In the world of computreplace the verbal activities that had

ers,

the

power to

Incredible as to

computer

evoking a

it

spell

out wishes

may sound

is

tantamount to making them happen.

to the profane, this notion

is

so ingrained in-

scientists that, latter-day spiritualists, they casually talk

program

to induce

its

execution.

The

aptness of the

revealed as one types out "Eliza" in luminous letters

on

word

presto!, Joseph

and

entertain. Spelling out

complex procedures

is

a dark screen,

Weizenbaum's mischievous creation appears

and

of

to chat

in the limpid language

of AI does more than allow us to understand them:

it

lets a

computer

execute these orders or control other machinery embodying these com-

pointed out:

it

Not

AI

ideas, as

Sussman

can also simplify complex machines or complex

activities.

plex processes.

only can

simplify

complex

Indeed, Donald Michie, the British dean of AI research, has called AI a

remedy

to "complexity pollution":

"AI

about making machines

is

more fathomable and more under the control of human beings, not

less.

Conventional technology has indeed been making our environment

more complex and more incomprehensible, and if it continues as it is doing now, the only conceivable outcome is disaster." 22 The increasing complexity of our machines and administration

is

probably the major

cause of the economic stagnation affecting developed countries. Solving the problem

would

tionally followed

require "a complete reversal of the approach tradi-

by technology, from one intended to get the most

economical use of machinery to one aimed

at

making the process of the

system clearly comprehensible to humans. For 23 to think like people."

this,

computers

will

need

THE SILICON CHALLENGERS

IN

For example, data bases, although as well.

333

FUTURE

01 R

a threat to privacy, have better uses

They can be invaluable in aiding a scientist carrying out a literature

search, a

businessman investigating market trends before launching a new

product, and even to anyone planning a vacation or in search of a

on

deal

certain

goods. Anyone with

modem

a

and

a

good personal computer

could tap hundreds of data bases if it weren't for one rub. In order to query a data base,

one must be familiar with its overall contents and the way these

are structured inside the data banks. Further, since

own query languages, one is

with their

different dialect

of computerese every time one switches to

a

new

data

These requirements turn away anyone (and that turns out to be most

base.

of us)

who doesn't have plenty of time and patience for searching through

thick manuals. will

many data bases come

forced to formulate questions in a

AI will soon replace

the thick manuals with programs that

understand questions in a close approximation of English, and find

out what

By

we want. 24

the turn of the

new

century data bases, coupled to expert systems,

much more individualized. Those our very own personal automated

should make possible services that are

of us so inclined may have access to librarian. It

on

may, for example, inquire about what you

a particular day,

combine these

acquired over

hints with

its

feel like

reading

knowledge of your

many such interactions with you, and come

personal

taste,

up with

selections to suit your

mood. Should you be

interested in re-

searching a particular subject (say, fishing), this librarian will review with

you the contents of the various books

available

one most appropriate to your needs. Or, chef, available at tion,

beck and

matching your

call

than the

whip up

tastes to the contents

book? Kristian Hammond's tives in 1986,

to

how

and help you

about a resident

recipes

of your refrigerator or pocket-

but required a hardware platform

power of these machines most households.

will

the day. 25

soon make

silicon

and menus combina-

CHEF program came close to

home computers of

select the

The

these objec-

much more powerful

exponential increase in

CHEF an affordable option for

and insurance agents are obvious extensions of this concept. Likewise, an expert system acting as a souped-up and interac-

Automated

tive

home

travel

medical encyclopedia might make for a healthier population

and even save

you of your

lives in

emergencies.

responsibilities

Home

and options

legal advisers

help you decide whether the expense of consulting a

warranted.

A home

might inform

in a particular situation,

human

and

lawyer

is

financial planner could help you invest a few hun-

334

Al

dred dollars with almost as

from

human

a

most

In

adviser

if

much wisdom and

invest.

cases, though, expert systems will not replace people in their

areas of competence. Instead, they will

users and simplify their tasks. for

savvy as you might get

you had tens of thousands of dollars to

many

years to

As

expand the

qualifications

of their

I've said earlier, expert systems will not,

come, exceed the savvy of the human experts they

model.

Expert systems will also need

and

their

someone



for the right information to be entered

recommendations interpreted



the help and

common sense of

domain of knowledge. Hence, an

already familiar with their

expert system's main function will be to enhance the knowledge and

performance of non-experts. For example, the dearth of medical ists in

remote areas might be

practitioners to

perform

alleviated

special-

by expert systems allowing general

as specialists.

Furthermore,

we

already have

expert systems allowing nurses to perform cardiorespiratory resuscitation in the

absence of a physician. Unfortunately, as I've mentioned, such

systems have never been implemented in emergency wards: such extensions of the

human

intellect lie

ahead of our present

legal

artificial

means of

defining and certifying competence.

In the long run, however, the pattern of our culture should rearrange itself to

accommodate them. At

the root of the problem

lies

our under-

standing and use of innovations in the context set forth by older practices.

The

earliest

carriages. It

automobiles, for instance, looked very

stations to turn the

like

automobile into a universal means of transportation;

to begin with, early drivers a dearth

much

took years for paved roads, mass production, and service

had to make do with few

of qualified mechanics. Likewise, AI

and shape both

itself

and

its

environment

will

outlets for fuel

have to find

in order to

its

and

place

bloom. Today's

expert systems are akin to the early Phoenician writings that just re-

corded a ship's cargo and the products of a

sale: early scripts

did not

engender history, philosophy, and mathematics overnight.

To what

equivalents of mathematics and philosophy, then, will

give birth to?

Much

as the scribe-accountants

of

AI

early Phoenicia could

not have predicted these disciplines, today's practitioners of AI are hard put to answer

this question.

Donald Michie,

for one,

made

the following

prediction:

We

can foresee a whole industry arising

industrial plant, the

"knowledge

.

.

refinery,"

.

based around

a

novel type of

which would take

in specialist

THE SILICON CHALLENGERS knowledge

in

creative gap-filling is

would be

significant if

precise, tested,

wisdom could be

we

will

pull

it

together, carry out

.

a fraction

.

.

of the world's accumulated

practical

brought together, and turned into accurate usable

sifted, 26

in The Cognitive Computer, explored

power of AI may

time,

even

in this way.

Roger Schank, sive

it,

whenever the need becomes evident, and turn out knowland certified correct. The boon to mankind

edge that

knowledge

335

OUR FUT IRE

form and debug

existing

its

IN

affect the sciences

how

the

new

be able to "develop understanding systems

particular fields." Stating exactly

how

expres-

and professions. For the

first

in various

doctors go about making their

diagnoses and lawyers assess a case, and what kinds of knowledge structure economists use to

make

decisions, will

endow

knowledge with unequaled investigative powers. "AI renaissance in practically every area

The main impact of thinking,

was

it

will

encourage

a

touches," claims Schank. 27

new ways of The same may well of AI in this respect

writing, in addition to creating

to revolutionize social organization.

hold true of AI. There are potential positive effects

as well as the negative ones. First, consider the effect

both the blue- and

these fields of

of automation

in

the white-collar worlds. It turns out that blue-collar

unions are usually supportive of robotics for two reasons: robots often replace workers in tedious, uncomfortable or hazardous jobs; and

most

unions acknowledge the need for productivity improvements which robots

effect.

AI and

robotics just might relieve us of the repetitive and

work brought about by the first industrial revolution. Consider loom operators: two hundred years ago, they were skilled manual workers enjoying definite professional respect. Then came the tedious parts of our

Jacquard loom, which automated the weaving of figured fabrics and

reduced the operator's job to one of feeding in materials and activating the machine: this skill.

Thanks

new

job description required no education and litde

to automation, the

taken a turn for the better.

I

loom

operator's profession has

now

recently visited a carpet manufacturing plant

where the looms mostly take care of themselves. Their overseers

are

engineers with the competence to suggest major design changes to the

equipment

if

needed, and submit written reports to

company's board of

would envy them

this effect to the

directors. Their eighteenth-century counterparts

their professional status.

But isn't one such loom engineer replacing scores of mechanical loom attendants and hundreds of earlier manual

loom operators? Haven't

336

||

these or their descendants

gone on

to swell the ranks

of the unem-

ployed? Since fewer than 10 percent of the population are unemployed in

most

tion

answer to

industrialized countries, the

ous no. Indeed, today's workers

of many of

is

an obvi-

their Victorian predecessors, but they also enjoy

incomparably superior standard of additional

question

this

match the combined produc-

typically

goods and

sen-ices

By and

living.

consumed by

large,

an

producing the

the average person occupies

the displaced workers. Also, in spite of the current tightness of the job

market, there

is

plenty of work to be done. 28 Given

enough resources,

education, health care, and psychological counseling provide an infinite

source of employment for people displaced by AI from other sectors.

Even though

it

may

take years to retrain them, and the

problem may

often not be resolved until the next generation, the key point the long run, the

employment problem

self-correcting.

is

is

that, in

The reason

plant engineers are replacing workers by robots, and managers are substituting

increase

it.

computers for

clerks,

growing amount of goods and there will

not to reduce production but to

is

Thus, automated economy

still

will

computers don't buy

their offspring) will

(or, if retraining is impractical,

be available to educate the young, take care of the

or help developing countries along the road to the

in the past

ing goods

cars,

be plenty of supplies around to meet the needs of humans.

Those people displaced by automation sick,

keep on generating an ever-

services. Since

hundred

years,

good

life.

Twice

we've developed mechanisms for redistribut-

whose production required

the population. During the

a drastically

first industrial

dwindled fraction of

revolution, manufacturing

jobs absorbed workers freed from farm work, which before had occu-

pied 80 percent of the American for only 9 percent

up the United

slack

work

force. Agriculture

accounts is

taking

from an ever-shrinking supply of manufacturing jobs. In the

States, the service sector occupies

we

now

of total employment, and the services sector

two

thirds

of the work

force.

means of spreading around the wealth generated by our emerging automated economy. At worst, we will end up with Surely

will

invent

more time on our hands than we can occupy with happens, then we'll

just

have to learn to place

less

full-time jobs. If this

moral value on work.

Indeed, AI will profoundly influence our values, including our perception

of ourselves. W"e must weigh any

induced by AI against the

intellectual skills available to

natural ability to acquire

fears

about the potential de-skilling

AI programs will make their embodied humans. Further, it is beyond anyone's

fact that

all

the

skills

one would wish

for.

For example an

~

CIULEKII*

TIE SHICfl

>„-i.: i.;:;:.i'T':::

:ti:r

:

:

77t :ti:

ur:

:

::

7_7U-7

7tr i~rr.~

:r tz

r