Jared Diamond
PREFACE TO THE PAPERBACK EDITION
WHY IS WORLD HISTORY LIKE AN ONION?
THIS BOOK ATTEMPTS TO PROVIDE A SHORT HISTORY OF everybody for the last
13,000 years. The question motivating the book is: Why did history unfold
differently on different continents? In case this question immediately makes
you shudder at the thought that you are about to read a racist treatise, you
aren’t: as you will see, the answers to the question don’t involve human racial
differences at all. The book’s emphasis is on the search for ultimate
explanations, and on pushing back the chain of historical causation as far as
possible.
Most books that set out to recount world history concentrate on histories
of literate Eurasian and North African societies. Native societies of other parts
of the world—sub-Saharan Africa, the Americas, Island Southeast Asia,
Australia, New Guinea, the Pacific Islands—receive only brief treatment,
mainly as concerns what happened to them very late in their history, after they
were discovered and subjugated by western Europeans. Even within Eurasia,
much more space gets devoted to the history of western Eurasia than of
China, India, Japan, tropical Southeast Asia, and other eastern Eurasian
societies. History before the emergence of writing around 3,000 B.C. also
receives brief treatment, although it constitutes 99.9% of the five-million-year
history of the human species.
Such narrowly focused accounts of world history suffer from three
disadvantages. First, increasing numbers of people today are, quite
understandably, interested in other societies besides those of western Eurasia.
After all, those “other” societies encompass most of the world’s population
and the vast majority of the world’s ethnic, cultural, and linguistic groups.
Some of them already are, and others are becoming, among the world’s most
powerful economies and political forces.
Second, even for people specifically interested in the shaping of the
modern world, a history limited to developments since the emergence of
writing cannot provide deep understanding. It is not the case that societies on
the different continents were comparable to each other until 3,000 B.C.,
whereupon western Eurasian societies suddenly developed writing and began
for the first time to pull ahead in other respects as well. Instead, already by
3,000 B.C., there were Eurasian and North African societies not only with
incipient writing but also with centralized state governments, cities,
widespread use of metal tools and weapons, use of domesticated animals for
transport and traction and mechanical power, and reliance on agriculture and
domestic animals for food. Throughout most or all parts of other continents,
none of those things existed at that time; some but not all of them emerged
later in parts of the Native Americas and sub-Saharan Africa, but only over
the course of the next five millennia; and none of them emerged in Aboriginal
Australia. That should already warn us that the roots of western Eurasian
dominance in the modern world lie in the preliterate past before 3,000 B.C.
(By western Eurasian dominance, I mean the dominance of western Eurasian
societies themselves and of the societies that they spawned on other
continents.)
Third, a history focused on western Eurasian societies completely
bypasses the obvious big question. Why were those societies the ones that
became disproportionately powerful and innovative? The usual answers to
that question invoke proximate forces, such as the rise of capitalism,
mercantilism, scientific inquiry, technology, and nasty germs that killed
peoples of other continents when they came into contact with western
Eurasians. But why did all those ingredients of conquest arise in western
Eurasia, and arise elsewhere only to a lesser degree or not at all?
All those ingredients are just proximate factors, not ultimate explanations.
Why didn’t capitalism flourish in Native Mexico, mercantilism in sub-
Saharan Africa, scientific inquiry in China, advanced technology in Native
North America, and nasty germs in Aboriginal Australia? If one responds by
invoking idiosyncratic cultural factors—e.g., scientific inquiry supposedly
stifled in China by Confucianism but stimulated in western Eurasia by Greek
or Judaeo-Christian traditions—then one is continuing to ignore the need for
ultimate explanations: why didn’t traditions like Confucianism and the
Judaeo-Christian ethic instead develop in western Eurasia and China,
respectively? In addition, one is ignoring the fact that Confucian China was
technologically more advanced than western Eurasia until about A.D. 1400.
It is impossible to understand even just western Eurasian societies
themselves, if one focuses on them. The interesting questions concern the
distinctions between them and other societies. Answering those questions
requires us to understand all those other societies as well, so that western
Eurasian societies can be fitted into the broader context.
Some readers may feel that I am going to the opposite extreme from
conventional histories, by devoting too little space to western Eurasia at the
expense of other parts of the world. I would answer that some other parts of
the world are very instructive, because they encompass so many societies and
such diverse societies within a small geographical area. Other readers may
find themselves agreeing with one reviewer of this book. With mildly critical
tongue in cheek, the reviewer wrote that I seem to view world history as an
onion, of which the modern world constitutes only the surface, and whose
layers are to be peeled back in the search for historical understanding. Yes,
world history is indeed such an onion! But that peeling back of the onion’s
layers is fascinating, challenging—and of overwhelming importance to us
today, as we seek to grasp our past’s lessons for our future.
J. D.
PROLOGUE
YALI’S QUESTION
WE ALL KNOW THAT HISTORY HAS PROCEEDED VERY DIFFERENTLY for peoples
from different parts of the globe. In the 13,000 years since the end of the last
Ice Age, some parts of the world developed literate industrial societies with
metal tools, other parts developed only nonliterate farming societies, and still
others retained societies of hunter-gatherers with stone tools. Those historical
inequalities have cast long shadows on the modern world, because the literate
societies with metal tools have conquered or exterminated the other societies.
While those differences constitute the most basic fact of world history, the
reasons for them remain uncertain and controversial. This puzzling question
of their origins was posed to me 25 years ago in a simple, personal form.
In July 1972 I was walking along a beach on the tropical island of New
Guinea, where as a biologist I study bird evolution. I had already heard about
a remarkable local politician named Yali, who was touring the district then.
By chance, Yali and I were walking in the same direction on that day, and he
overtook me. We walked together for an hour, talking during the whole time.
Yali radiated charisma and energy. His eyes flashed in a mesmerizing
way. He talked confidently about himself, but he also asked lots of probing
questions and listened intently. Our conversation began with a subject then on
every New Guinean’s mind—the rapid pace of political developments. Papua
New Guinea, as Yali’s nation is now called, was at that time still administered
by Australia as a mandate of the United Nations, but independence was in the
air. Yali explained to me his role in getting local people to prepare for self-
government.
After a while, Yali turned the conversation and began to quiz me. He had
never been outside New Guinea and had not been educated beyond high
school, but his curiosity was insatiable. First, he wanted to know about my
work on New Guinea birds (including how much I got paid for it). I explained
to him how different groups of birds had colonized New Guinea over the
course of millions of years. He then asked how the ancestors of his own
people had reached New Guinea over the last tens of thousands of years, and
how white Europeans had colonized New Guinea within the last 200 years.
The conversation remained friendly, even though the tension between the
two societies that Yali and I represented was familiar to both of us. Two
centuries ago, all New Guineans were still “living in the Stone Age.” That is,
they still used stone tools similar to those superseded in Europe by metal tools
thousands of years ago, and they dwelt in villages not organized under any
centralized political authority. Whites had arrived, imposed centralized
government, and brought material goods whose value New Guineans instantly
recognized, ranging from steel axes, matches, and medicines to clothing, soft
drinks, and umbrellas. In New Guinea all these goods were referred to
collectively as “cargo.”
Many of the white colonialists openly despised New Guineans as
“primitive.” Even the least able of New Guinea’s white “masters,” as they
were still called in 1972, enjoyed a far higher standard of living than New
Guineans, higher even than charismatic politicians like Yali. Yet Yali had
quizzed lots of whites as he was then quizzing me, and I had quizzed lots of
New Guineans. He and I both knew perfectly well that New Guineans are on
the average at least as smart as Europeans. All those things must have been on
Yali’s mind when, with yet another penetrating glance of his flashing eyes, he
asked me, “Why is it that you white people developed so much cargo and
brought it to New Guinea, but we black people had little cargo of our own?”
It was a simple question that went to the heart of life as Yali experienced
it. Yes, there still is a huge difference between the lifestyle of the average
New Guinean and that of the average European or American. Comparable
differences separate the lifestyles of other peoples of the world as well. Those
huge disparities must have potent causes that one might think would be
obvious.
Yet Yali’s apparently simple question is a difficult one to answer. I didn’t
have an answer then. Professional historians still disagree about the solution;
most are no longer even asking the question. In the years since Yali and I had
that conversation, I have studied and written about other aspects of human
evolution, history, and language. This book, written twenty-five years later,
attempts to answer Yali.
ALTHOUGH YALI’S QUESTION concerned only the contrasting lifestyles of New
Guineans and of European whites, it can be extended to a larger set of
contrasts within the modern world. Peoples of Eurasian origin, especially
those still living in Europe and eastern Asia, plus those transplanted to North
America, dominate the modern world in wealth and power. Other peoples,
including most Africans, have thrown off European colonial domination but
remain far behind in wealth and power. Still other peoples, such as the
aboriginal inhabitants of Australia, the Americas, and southernmost Africa,
are no longer even masters of their own lands but have been decimated,
subjugated, and in some cases even exterminated by European colonialists.
Thus, questions about inequality in the modern world can be reformulated
as follows. Why did wealth and power become distributed as they now are,
rather than in some other way? For instance, why weren’t Native Americans,
Africans, and Aboriginal Australians the ones who decimated, subjugated, or
exterminated Europeans and Asians?
We can easily push this question back one step. As of the year A.D. 1500,
when Europe’s worldwide colonial expansion was just beginning, peoples on
different continents already differed greatly in technology and political
organization. Much of Europe, Asia, and North Africa was the site of metal-
equipped states or empires, some of them on the threshold of industrialization.
Two Native American peoples, the Aztecs and the Incas, ruled over empires
with stone tools. Parts of sub-Saharan Africa were divided among small states
or chiefdoms with iron tools. Most other peoples—including all those of
Australia and New Guinea, many Pacific islands, much of the Americas, and
small parts of sub-Saharan Africa—lived as farming tribes or even still as
hunter-gatherer bands using stone tools.
Of course, those technological and political differences as of A.D. 1500
were the immediate cause of the modern world’s inequalities. Empires with
steel weapons were able to conquer or exterminate tribes with weapons of
stone and wood. How, though, did the world get to be the way it was in A.D.
1500?
Once again, we can easily push this question back one step further, by
drawing on written histories and archaeological discoveries. Until the end of
the last Ice Age, around 11,000 B.C., all peoples on all continents were still
hunter-gatherers. Different rates of development on different continents, from
11,000 B.C. to A.D. 1500, were what led to the technological and political
inequalities of A.D. 1500. While Aboriginal Australians and many Native
Americans remained hunter-gatherers, most of Eurasia and much of the
Americas and sub-Saharan Africa gradually developed agriculture, herding,
metallurgy, and complex political organization. Parts of Eurasia, and one area
of the Americas, independently developed writing as well. However, each of
these new developments appeared earlier in Eurasia than elsewhere. For
instance, the mass production of bronze tools, which was just beginning in the
South American Andes in the centuries before A.D. 1500, was already
established in parts of Eurasia over 4,000 years earlier. The stone technology
of the Tasmanians, when first encountered by European explorers in A.D.
1642, was simpler than that prevalent in parts of Upper Paleolithic Europe
tens of thousands of years earlier.
Thus, we can finally rephrase the question about the modern world’s
inequalities as follows: why did human development proceed at such different
rates on different continents? Those disparate rates constitute history’s
broadest pattern and my book’s subject.
While this book is thus ultimately about history and prehistory, its subject
is not of just academic interest but also of overwhelming practical and
political importance. The history of interactions among disparate peoples is
what shaped the modern world through conquest, epidemics, and genocide.
Those collisions created reverberations that have still not died down after
many centuries, and that are actively continuing in some of the world’s most
troubled areas today.
For example, much of Africa is still struggling with its legacies from
recent colonialism. In other regions—including much of Central America,
Mexico, Peru, New Caledonia, the former Soviet Union, and parts of
Indonesia—civil unrest or guerrilla warfare pits still-numerous indigenous
populations against governments dominated by descendants of invading
conquerors. Many other indigenous populations—such as native Hawaiians,
Aboriginal Australians, native Siberians, and Indians in the United States,
Canada, Brazil, Argentina, and Chile—became so reduced in numbers by
genocide and disease that they are now greatly outnumbered by the
descendants of invaders. Although thus incapable of mounting a civil war,
they are nevertheless increasingly asserting their rights.
In addition to these current political and economic reverberations of past
collisions among peoples, there are current linguistic reverberations—
especially the impending disappearance of most of the modern world’s 6,000
surviving languages, becoming replaced by English, Chinese, Russian, and a
few other languages whose numbers of speakers have increased enormously
in recent centuries. All these problems of the modern world result from the
different historical trajectories implicit in Yali’s question.
BEFORE SEEKING ANSWERS to Yali’s question, we should pause to consider
some objections to discussing it at all. Some people take offense at the mere
posing of the question, for several reasons.
One objection goes as follows. If we succeed in explaining how some
people came to dominate other people, may this not seem to justify the
domination? Doesn’t it seem to say that the outcome was inevitable, and that
it would therefore be futile to try to change the outcome today? This objection
rests on a common tendency to confuse an explanation of causes with a
justification or acceptance of results. What use one makes of a historical
explanation is a question separate from the explanation itself. Understanding
is more often used to try to alter an outcome than to repeat or perpetuate it.
That’s why psychologists try to understand the minds of murderers and
rapists, why social historians try to understand genocide, and why physicians
try to understand the causes of human disease. Those investigators do not
seek to justify murder, rape, genocide, and illness. Instead, they seek to use
their understanding of a chain of causes to interrupt the chain.
Second, doesn’t addressing Yali’s question automatically involve a
Eurocentric approach to history, a glorification of western Europeans, and an
obsession with the prominence of western Europe and Europeanized America
in the modern world? Isn’t that prominence just an ephemeral phenomenon of
the last few centuries, now fading behind the prominence of Japan and
Southeast Asia? In fact, most of this book will deal with peoples other than
Europeans. Rather than focus solely on interactions between Europeans and
non-Europeans, we shall also examine interactions between different non-
European peoples—especially those that took place within sub-Saharan
Africa, Southeast Asia, Indonesia, and New Guinea, among peoples native to
those areas. Far from glorifying peoples of western European origin, we shall
see that most basic elements of their civilization were developed by other
peoples living elsewhere and were then imported to western Europe.
Third, don’t words such as “civilization,” and phrases such as “rise of
civilization,” convey the false impression that civilization is good, tribal
hunter-gatherers are miserable, and history for the past 13,000 years has
involved progress toward greater human happiness? In fact, I do not assume
that industrialized states are “better” than hunter-gatherer tribes, or that the
abandonment of the hunter-gatherer lifestyle for iron-based statehood
represents “progress,” or that it has led to an increase in human happiness. My
own impression, from having divided my life between United States cities
and New Guinea villages, is that the so-called blessings of civilization are
mixed. For example, compared with hunter-gatherers, citizens of modern
industrialized states enjoy better medical care, lower risk of death by
homicide, and a longer life span, but receive much less social support from
friendships and extended families. My motive for investigating these
geographic differences in human societies is not to celebrate one type of
society over another but simply to understand what happened in history.
DOES YALI’S QUESTION really need another book to answer it? Don’t we
already know the answer? If so, what is it?
Probably the commonest explanation involves implicitly or explicitly
assuming biological differences among peoples. In the centuries after A.D.
1500, as European explorers became aware of the wide differences among the
world’s peoples in technology and political organization, they assumed that
those differences arose from differences in innate ability. With the rise of
Darwinian theory, explanations were recast in terms of natural selection and
of evolutionary descent. Technologically primitive peoples were considered
evolutionary vestiges of human descent from apelike ancestors. The
displacement of such peoples by colonists from industrialized societies
exemplified the survival of the fittest. With the later rise of genetics, the
explanations were recast once again, in genetic terms. Europeans became
considered genetically more intelligent than Africans, and especially more so
than Aboriginal Australians.
Today, segments of Western society publicly repudiate racism. Yet many
(perhaps most!) Westerners continue to accept racist explanations privately or
subconsciously. In Japan and many other countries, such explanations are still
advanced publicly and without apology. Even educated white Americans,
Europeans, and Australians, when the subject of Australian Aborigines comes
up, assume that there is something primitive about the Aborigines themselves.
They certainly look different from whites. Many of the living descendants of
those Aborigines who survived the era of European colonization are now
finding it difficult to succeed economically in white Australian society.
A seemingly compelling argument goes as follows. White immigrants to
Australia built a literate, industrialized, politically centralized, democratic
state based on metal tools and on food production, all within a century of
colonizing a continent where the Aborigines had been living as tribal hunter-
gatherers without metal for at least 40,000 years. Here were two successive
experiments in human development, in which the environment was identical
and the sole variable was the people occupying that environment. What
further proof could be wanted to establish that the differences between
Aboriginal Australian and European societies arose from differences between
the peoples themselves?
The objection to such racist explanations is not just that they are
loathsome, but also that they are wrong. Sound evidence for the existence of
human differences in intelligence that parallel human differences in
technology is lacking. In fact, as I shall explain in a moment, modern “Stone
Age” peoples are on the average probably more intelligent, not less
intelligent, than industrialized peoples. Paradoxical as it may sound, we shall
see in Chapter 15 that white immigrants to Australia do not deserve the credit
usually accorded to them for building a literate industrialized society with the
other virtues mentioned above. In addition, peoples who until recently were
technologically primitive—such as Aboriginal Australians and New Guineans
—routinely master industrial technologies when given opportunities to do so.
An enormous effort by cognitive psychologists has gone into the search
for differences in IQ between peoples of different geographic origins now
living in the same country. In particular, numerous white American
psychologists have been trying for decades to demonstrate that black
Americans of African origins are innately less intelligent than white
Americans of European origins. However, as is well known, the peoples
compared differ greatly in their social environment and educational
opportunities. This fact creates double difficulties for efforts to test the
hypothesis that intellectual differences underlie technological differences.
First, even our cognitive abilities as adults are heavily influenced by the social
environment that we experienced during childhood, making it hard to discern
any influence of preexisting genetic differences. Second, tests of cognitive
ability (like IQ tests) tend to measure cultural learning and not pure innate
intelligence, whatever that is. Because of those undoubted effects of
childhood environment and learned knowledge on IQ test results, the
psychologists’ efforts to date have not succeeded in convincingly establishing
the postulated genetic deficiency in IQs of nonwhite peoples.
My perspective on this controversy comes from 33 years of working with
New Guineans in their own intact societies. From the very beginning of my
work with New Guineans, they impressed me as being on the average more
intelligent, more alert, more expressive, and more interested in things and
people around them than the average European or American is. At some tasks
that one might reasonably suppose to reflect aspects of brain function, such as
the ability to form a mental map of unfamiliar surroundings, they appear
considerably more adept than Westerners. Of course, New Guineans tend to
perform poorly at tasks that Westerners have been trained to perform since
childhood and that New Guineans have not. Hence when unschooled New
Guineans from remote villages visit towns, they look stupid to Westerners.
Conversely, I am constantly aware of how stupid I look to New Guineans
when I’m with them in the jungle, displaying my incompetence at simple
tasks (such as following a jungle trail or erecting a shelter) at which New
Guineans have been trained since childhood and I have not.
It’s easy to recognize two reasons why my impression that New Guineans
are smarter than Westerners may be correct. First, Europeans have for
thousands of years been living in densely populated societies with central
governments, police, and judiciaries. In those societies, infectious epidemic
diseases of dense populations (such as smallpox) were historically the major
cause of death, while murders were relatively uncommon and a state of war
was the exception rather than the rule. Most Europeans who escaped fatal
infections also escaped other potential causes of death and proceeded to pass
on their genes. Today, most live-born Western infants survive fatal infections
as well and reproduce themselves, regardless of their intelligence and the
genes they bear. In contrast, New Guineans have been living in societies
where human numbers were too low for epidemic diseases of dense
populations to evolve. Instead, traditional New Guineans suffered high
mortality from murder, chronic tribal warfare, accidents, and problems in
procuring food.
Intelligent people are likelier than less intelligent ones to escape those
causes of high mortality in traditional New Guinea societies. However, the
differential mortality from epidemic diseases in traditional European societies
had little to do with intelligence, and instead involved genetic resistance
dependent on details of body chemistry. For example, people with blood
group B or O have a greater resistance to smallpox than do people with blood
group A. That is, natural selection promoting genes for intelligence has
probably been far more ruthless in New Guinea than in more densely
populated, politically complex societies, where natural selection for body
chemistry was instead more potent.
Besides this genetic reason, there is also a second reason why New
Guineans may have come to be smarter than Westerners. Modern European
and American children spend much of their time being passively entertained
by television, radio, and movies. In the average American household, the TV
set is on for seven hours per day. In contrast, traditional New Guinea children
have virtually no such opportunities for passive entertainment and instead
spend almost all of their waking hours actively doing something, such as
talking or playing with other children or adults. Almost all studies of child
development emphasize the role of childhood stimulation and activity in
promoting mental development, and stress the irreversible mental stunting
associated with reduced childhood stimulation. This effect surely contributes
a non-genetic component to the superior average mental function displayed
by New Guineans.
That is, in mental ability New Guineans are probably genetically superior
to Westerners, and they surely are superior in escaping the devastating
developmental disadvantages under which most children in industrialized
societies now grow up. Certainly, there is no hint at all of any intellectual
dis advantage of New Guineans that could serve to answer Yali’s question.
The same two genetic and childhood developmental factors are likely to
distinguish not only New Guineans from Westerners, but also hunter-
gatherers and other members of technologically primitive societies from
members of technologically advanced societies in general. Thus, the usual
racist assumption has to be turned on its head. Why is it that Europeans,
despite their likely genetic disadvantage and (in modern times) their
undoubted developmental disadvantage, ended up with much more of the
cargo? Why did New Guineans wind up technologically primitive, despite
what I believe to be their superior intelligence?
A GENETIC EXPLANATION isn’t the only possible answer to Yali’s question.
Another one, popular with inhabitants of northern Europe, invokes the
supposed stimulatory effects of their homeland’s cold climate and the
inhibitory effects of hot, humid, tropical climates on human creativity and
energy. Perhaps the seasonally variable climate at high latitudes poses more
diverse challenges than does a seasonally constant tropical climate. Perhaps
cold climates require one to be more technologically inventive to survive,
because one must build a warm home and make warm clothing, whereas one
can survive in the tropics with simpler housing and no clothing. Or the
argument can be reversed to reach the same conclusion: the long winters at
high latitudes leave people with much time in which to sit indoors and invent.
Although formerly popular, this type of explanation, too, fails to survive
scrutiny. As we shall see, the peoples of northern Europe contributed nothing
of fundamental importance to Eurasian civilization until the last thousand
years; they simply had the good luck to live at a geographic location where
they were likely to receive advances (such as agriculture, wheels, writing, and
metallurgy) developed in warmer parts of Eurasia. In the New World the cold
regions at high latitude were even more of a human backwater. The sole
Native American societies to develop writing arose in Mexico south of the
Tropic of Cancer; the oldest New World pottery comes from near the equator
in tropical South America; and the New World society generally considered
the most advanced in art, astronomy, and other respects was the Classic Maya
society of the tropical Yucatán and Guatemala in the first millennium A.D.
Still a third type of answer to Yali invokes the supposed importance of
lowland river valleys in dry climates, where highly productive agriculture
depended on large-scale irrigation systems that in turn required centralized
bureaucracies. This explanation was suggested by the undoubted fact that the
earliest known empires and writing systems arose in the Tigris and Euphrates
Valleys of the Fertile Crescent and in the Nile Valley of Egypt. Water control
systems also appear to have been associated with centralized political
organization in some other areas of the world, including the Indus Valley of
the Indian subcontinent, the Yellow and Yangtze Valleys of China, the Maya
lowlands of Mesoamerica, and the coastal desert of Peru.
However, detailed archaeological studies have shown that complex
irrigation systems did not accompany the rise of centralized bureaucracies but
followed after a considerable lag. That is, political centralization arose for
some other reason and then permitted construction of complex irrigation
systems. None of the crucial developments preceding political centralization
in those same parts of the world were associated with river valleys or with
complex irrigation systems. For example, in the Fertile Crescent food
production and village life originated in hills and mountains, not in lowland
river valleys. The Nile Valley remained a cultural backwater for about 3,000
years after village food production began to flourish in the hills of the Fertile
Crescent. River valleys of the southwestern United States eventually came to
support irrigation agriculture and complex societies, but only after many of
the developments on which those societies rested had been imported from
Mexico. The river valleys of southeastern Australia remained occupied by
tribal societies without agriculture.
Yet another type of explanation lists the immediate factors that enabled
Europeans to kill or conquer other peoples—especially European guns,
infectious diseases, steel tools, and manufactured products. Such an
explanation is on the right track, as those factors demonstrably were directly
responsible for European conquests. However, this hypothesis is incomplete,
because it still offers only a proximate (first-stage) explanation identifying
immediate causes. It invites a search for ultimate causes: why were
Europeans, rather than Africans or Native Americans, the ones to end up with
guns, the nastiest germs, and steel?
While some progress has been made in identifying those ultimate causes
in the case of Europe’s conquest of the New World, Africa remains a big
puzzle. Africa is the continent where protohumans evolved for the longest
time, where anatomically modern humans may also have arisen, and where
native diseases like malaria and yellow fever killed European explorers. If a
long head start counts for anything, why didn’t guns and steel arise first in
Africa, permitting Africans and their germs to conquer Europe? And what
accounts for the failure of Aboriginal Australians to pass beyond the stage of
hunter-gatherers with stone tools?
Questions that emerge from worldwide comparisons of human societies
formerly attracted much attention from historians and geographers. The best-
known modern example of such an effort was Arnold Toynbee’s 12-volume
Study of History. Toynbee was especially interested in the internal dynamics
of 23 advanced civilizations, of which 22 were literate and 19 were Eurasian.
He was less interested in prehistory and in simpler, nonliterate societies. Yet
the roots of inequality in the modern world lie far back in prehistory. Hence
Toynbee did not pose Yali’s question, nor did he come to grips with what I see
as history’s broadest pattern. Other available books on world history similarly
tend to focus on advanced literate Eurasian civilizations of the last 5,000
years; they have a very brief treatment of pre-Columbian Native American
civilizations, and an even briefer discussion of the rest of the world except for
its recent interactions with Eurasian civilizations. Since Toynbee’s attempt,
worldwide syntheses of historical causation have fallen into disfavor among
most historians, as posing an apparently intractable problem.
Specialists from several disciplines have provided global syntheses of
their subjects. Especially useful contributions have been made by ecological
geographers, cultural anthropologists, biologists studying plant and animal
domestication, and scholars concerned with the impact of infectious diseases
on history. These studies have called attention to parts of the puzzle, but they
provide only pieces of the needed broad synthesis that has been missing.
Thus, there is no generally accepted answer to Yali’s question. On the one
hand, the proximate explanations are clear: some peoples developed guns,
germs, steel, and other factors conferring political and economic power before
others did; and some peoples never developed these power factors at all. On
the other hand, the ultimate explanations—for example, why bronze tools
appeared early in parts of Eurasia, late and only locally in the New World, and
never in Aboriginal Australia—remain unclear.
Our present lack of such ultimate explanations leaves a big intellectual
gap, since the broadest pattern of history thus remains unexplained. Much
more serious, though, is the moral gap left unfilled. It is perfectly obvious to
everyone, whether an overt racist or not, that different peoples have fared
differently in history. The modern United States is a European-molded
society, occupying lands conquered from Native Americans and incorporating
the descendants of millions of sub-Saharan black Africans brought to
America as slaves. Modern Europe is not a society molded by sub-Saharan
black Africans who brought millions of Native Americans as slaves.
These results are completely lopsided: it was not the case that 51 percent
of the Americas, Australia, and Africa was conquered by Europeans, while 49
percent of Europe was conquered by Native Americans, Aboriginal
Australians, or Africans. The whole modern world has been shaped by
lopsided outcomes. Hence they must have inexorable explanations, ones more
basic than mere details concerning who happened to win some battle or
develop some invention on one occasion a few thousand years ago.
It seems logical to suppose that history’s pattern reflects innate differences
among people themselves. Of course, we’re taught that it’s not polite to say so
in public. We read of technical studies claiming to demonstrate inborn
differences, and we also read rebuttals claiming that those studies suffer from
technical flaws. We see in our daily lives that some of the conquered peoples
continue to form an underclass, centuries after the conquests or slave imports
took place. We’re told that this too is to be attributed not to any biological
shortcomings but to social disadvantages and limited opportunities.
Nevertheless, we have to wonder. We keep seeing all those glaring,
persistent differences in peoples’ status. We’re assured that the seemingly
transparent biological explanation for the world’s inequalities as of A.D. 1500
is wrong, but we’re not told what the correct explanation is. Until we have
some convincing, detailed, agreed-upon explanation for the broad pattern of
history, most people will continue to suspect that the racist biological
explanation is correct after all. That seems to me the strongest argument for
writing this book.
AUTHORS ARE REGULARLY asked by journalists to summarize a long book in
one sentence. For this book, here is such a sentence: “History followed
different courses for different peoples because of differences among peoples’
environments, not because of biological differences among peoples
themselves.”
Naturally, the notion that environmental geography and biogeography
influenced societal development is an old idea. Nowadays, though, the view is
not held in esteem by historians; it is considered wrong or simplistic, or it is
caricatured as environmental determinism and dismissed, or else the whole
subject of trying to understand worldwide differences is shelved as too
difficult. Yet geography obviously has some effect on history; the open
question concerns how much effect, and whether geography can account for
history’s broad pattern.
The time is now ripe for a fresh look at these questions, because of new
information from scientific disciplines seemingly remote from human history.
Those disciplines include, above all, genetics, molecular biology, and
biogeography as applied to crops and their wild ancestors; the same
disciplines plus behavioral ecology, as applied to domestic animals and their
wild ancestors; molecular biology of human germs and related germs of
animals; epidemiology of human diseases; human genetics; linguistics;
archaeological studies on all continents and major islands; and studies of the
histories of technology, writing, and political organization.
This diversity of disciplines poses problems for would-be authors of a
book aimed at answering Yali’s question. The author must possess a range of
expertise spanning the above disciplines, so that relevant advances can be
synthesized. The history and prehistory of each continent must be similarly
synthesized. The book’s subject matter is history, but the approach is that of
science—in particular, that of historical sciences such as evolutionary biology
and geology. The author must understand from firsthand experience a range
of human societies, from hunter-gatherer societies to modern space-age
civilizations.
These requirements seem at first to demand a multi-author work. Yet that
approach would be doomed from the outset, because the essence of the
problem is to develop a unified synthesis. That consideration dictates single
authorship, despite all the difficulties that it poses. Inevitably, that single
author will have to sweat copiously in order to assimilate material from many
disciplines, and will require guidance from many colleagues.
My background had led me to several of these disciplines even before
Yali put his question to me in 1972. My mother is a teacher and linguist; my
father, a physician specializing in the genetics of childhood diseases. Because
of my father’s example, I went through school expecting to become a
physician. I had also become a fanatical bird-watcher by the age of seven. It
was thus an easy step, in my last undergraduate year at university, to shift
from my initial goal of medicine to the goal of biological research. However,
throughout my school and undergraduate years, my training was mainly in
languages, history, and writing. Even after deciding to obtain a Ph.D. in
physiology, I nearly dropped out of science during my first year of graduate
school to become a linguist.
Since completing my Ph.D. in 1961, I have divided my scientific research
efforts between two fields: molecular physiology on the one hand,
evolutionary biology and biogeography on the other hand. As an unforeseen
bonus for the purposes of this book, evolutionary biology is a historical
science forced to use methods different from those of the laboratory sciences.
That experience has made the difficulties in devising a scientific approach to
human history familiar to me. Living in Europe from 1958 to 1962, among
European friends whose lives had been brutally traumatized by 20th-century
European history, made me start to think more seriously about how chains of
causes operate in history’s unfolding.
For the last 33 years my fieldwork as an evolutionary biologist has
brought me into close contact with a wide range of human societies. My
specialty is bird evolution, which I have studied in South America, southern
Africa, Indonesia, Australia, and especially New Guinea. Through living with
native peoples of these areas, I have become familiar with many
technologically primitive human societies, from those of hunter-gatherers to
those of tribal farmers and fishing peoples who depended until recently on
stone tools. Thus, what most literate people would consider strange lifestyles
of remote prehistory are for me the most vivid part of my life. New Guinea,
though it accounts for only a small fraction of the world’s land area,
encompasses a disproportionate fraction of its human diversity. Of the modern
world’s 6,000 languages, 1,000 are confined to New Guinea. In the course of
my work on New Guinea birds, my interests in language were rekindled, by
the need to elicit lists of local names of bird species in nearly 100 of those
New Guinea languages.
Out of all those interests grew my most recent book, a nontechnical
account of human evolution entitled The Third Chimpanzee. Its Chapter 14,
called “Accidental Conquerors,” sought to understand the outcome of the
encounter between Europeans and Native Americans. After I had completed
that book, I realized that other modern, as well as prehistoric, encounters
between peoples raised similar questions. I saw that the question with which I
had wrestled in that Chapter 14 was in essence the question Yali had asked me
in 1972, merely transferred to a different part of the world. And so at last,
with the help of many friends, I shall attempt to satisfy Yali’s curiosity—and
my own.
THIS BOOK’S CHAPTERS are divided into four parts. Part 1, entitled “From
Eden to Cajamarca,” consists of three chapters. Chapter 1 provides a
whirlwind tour of human evolution and history, extending from our
divergence from apes, around 7 million years ago, until the end of the last Ice
Age, around 13,000 years ago. We shall trace the spread of ancestral humans,
from our origins in Africa to the other continents, in order to understand the
state of the world just before the events often lumped into the term “rise of
civilization” began. It turns out that human development on some continents
got a head start in time over developments on others.
Chapter 2 prepares us for exploring effects of continental environments
on history over the past 13,000 years, by briefly examining effects of island
environments on history over smaller time scales and areas. When ancestral
Polynesians spread into the Pacific around 3,200 years ago, they encountered
islands differing greatly in their environments. Within a few millennia that
single ancestral Polynesian society had spawned on those diverse islands a
range of diverse daughter societies, from hunter-gatherer tribes to proto-
empires. That radiation can serve as a model for the longer, larger-scale, and
less understood radiation of societies on different continents since the end of
the last Ice Age, to become variously hunter-gatherer tribes and empires.
The third chapter introduces us to collisions between peoples from
different continents, by retelling through contemporary eyewitness accounts
the most dramatic such encounter in history: the capture of the last
independent Inca emperor, Atahuallpa, in the presence of his whole army, by
Francisco Pizarro and his tiny band of conquistadores, at the Peruvian city of
Cajamarca. We can identify the chain of proximate factors that enabled
Pizarro to capture Atahuallpa, and that operated in European conquests of
other Native American societies as well. Those factors included Spanish
germs, horses, literacy, political organization, and technology (especially
ships and weapons). That analysis of proximate causes is the easy part of this
book; the hard part is to identify the ultimate causes leading to them and to
the actual outcome, rather than to the opposite possible outcome of
Atahuallpa’s coming to Madrid and capturing King Charles I of Spain.
Part 2, entitled “The Rise and Spread of Food Production” and consisting
of Chapters 4–10, is devoted to what I believe to be the most important
constellation of ultimate causes. Chapter 4 sketches how food production—
that is, the growing of food by agriculture or herding, instead of the hunting
and gathering of wild foods—ultimately led to the immediate factors
permitting Pizarro’s triumph. But the rise of food production varied around
the globe. As we shall see in Chapter 5, peoples in some parts of the world
developed food production by themselves; some other peoples acquired it in
prehistoric times from those independent centers; and still others neither
developed nor acquired food production prehistorically but remained hunter-
gatherers until modern times. Chapter 6 explores the numerous factors driving
the shift from the hunter-gatherer lifestyle toward food production, in some
areas but not in others.
Chapters 7, 8, and 9 then show how crops and livestock came in
prehistoric times to be domesticated from ancestral wild plants and animals,
by incipient farmers and herders who could have had no vision of the
outcome. Geographic differences in the local suites of wild plants and animals
available for domestication go a long way toward explaining why only a few
areas became independent centers of food production, and why it arose earlier
in some of those areas than in others. From those few centers of origin, food
production spread much more rapidly to some areas than to others. A major
factor contributing to those differing rates of spread turns out to have been the
orientation of the continents’ axes: predominantly west-east for Eurasia,
predominantly north-south for the Americas and Africa (Chapter 10).
Thus, Chapter 3 sketched the immediate factors behind Europe’s conquest
of Native Americans, and Chapter 4 the development of those factors from
the ultimate cause of food production. In Part 3 (“From Food to Guns, Germs,
and Steel,” Chapters 11–14), the connections from ultimate to proximate
causes are traced in detail, beginning with the evolution of germs
characteristic of dense human populations (Chapter 11). Far more Native
Americans and other non-Eurasian peoples were killed by Eurasian germs
than by Eurasian guns or steel weapons. Conversely, few or no distinctive
lethal germs awaited would-be European conquerors in the New World. Why
was the germ exchange so unequal? Here, the results of recent molecular
biological studies are illuminating in linking germs to the rise of food
production, in Eurasia much more than in the Americas.
Another chain of causation led from food production to writing, possibly
the most important single invention of the last few thousand years (Chapter
12). Writing has evolved de novo only a few times in human history, in areas
that had been the earliest sites of the rise of food production in their
respective regions. All other societies that have become literate did so by the
diffusion of writing systems or of the idea of writing from one of those few
primary centers. Hence, for the student of world history, the phenomenon of
writing is particularly useful for exploring another important constellation of
causes: geography’s effect on the ease with which ideas and inventions
spread.
What holds for writing also holds for technology (Chapter 13). A crucial
question is whether technological innovation is so dependent on rare inventor-
geniuses, and on many idiosyncratic cultural factors, as to defy an
understanding of world patterns. In fact, we shall see that, paradoxically, this
large number of cultural factors makes it easier, not harder, to understand
world patterns of technology. By enabling farmers to generate food surpluses,
food production permitted farming societies to support full-time craft
specialists who did not grow their own food and who developed technologies.
Besides sustaining scribes and inventors, food production also enabled
farmers to support politicians (Chapter 14). Mobile bands of hunter-gatherers
are relatively egalitarian, and their political sphere is confined to the band’s
own territory and to shifting alliances with neighboring bands. With the rise
of dense, sedentary, food-producing populations came the rise of chiefs,
kings, and bureaucrats. Such bureaucracies were essential not only to
governing large and populous domains but also to maintaining standing
armies, sending out fleets of exploration, and organizing wars of conquest.
Part 4 (“Around the World in Five Chapters,” Chapters 15–19) applies the
lessons of Parts 2 and 3 to each of the continents and some important islands.
Chapter 15 examines the history of Australia itself, and of the large island of
New Guinea, formerly joined to Australia in a single continent. The case of
Australia, home to the recent human societies with the simplest technologies,
and the sole continent where food production did not develop indigenously,
poses a critical test of theories about intercontinental differences in human
societies. We shall see why Aboriginal Australians remained hunter-gatherers,
even while most peoples of neighboring New Guinea became food producers.
Chapters 16 and 17 integrate developments in Australia and New Guinea
into the perspective of the whole region encompassing the East Asian
mainland and Pacific islands. The rise of food production in China spawned
several great prehistoric movements of human populations, or of cultural
traits, or of both. One of those movements, within China itself, created the
political and cultural phenomenon of China as we know it today. Another
resulted in a replacement, throughout almost the whole of tropical Southeast
Asia, of indigenous hunter-gatherers by farmers of ultimately South Chinese
origin. Still another, the Austronesian expansion, similarly replaced the
indigenous hunter-gatherers of the Philippines and Indonesia and spread out
to the most remote islands of Polynesia, but was unable to colonize Australia
and most of New Guinea. To the student of world history, all those collisions
among East Asian and Pacific peoples are doubly important: they formed the
countries where one-third of the modern world’s population lives, and in
which economic power is increasingly becoming concentrated; and they
furnish especially clear models for understanding the histories of peoples
elsewhere in the world.
Chapter 18 returns to the problem introduced in Chapter 3, the collision
between European and Native American peoples. A summary of the last
13,000 years of New World and western Eurasian history makes clear how
Europe’s conquest of the Americas was merely the culmination of two long
and mostly separate historical trajectories. The differences between those
trajectories were stamped by continental differences in domesticable plants
and animals, germs, times of settlement, orientation of continental axes, and
ecological barriers.
Finally, the history of sub-Saharan Africa (Chapter 19) offers striking
similarities as well as contrasts with New World history. The same factors that
molded Europeans’ encounters with Africans molded their encounters with
Native Americans as well. But Africa also differed from the Americas in all
these factors. As a result, European conquest did not create widespread or
lasting European settlement of sub-Saharan Africa, except in the far south. Of
more lasting significance was a large-scale population shift within Africa
itself, the Bantu expansion. It proves to have been triggered by many of the
same causes that played themselves out at Cajamarca, in East Asia, on Pacific
islands, and in Australia and New Guinea.
I harbor no illusions that these chapters have succeeded in explaining the
histories of all the continents for the past 13,000 years. Obviously, that would
be impossible to accomplish in a single book even if we did understand all the
answers, which we don’t. At best, this book identifies several constellations
of environmental factors that I believe provide a large part of the answer to
Yali’s question. Recognition of those factors emphasizes the unexplained
residue, whose understanding will be a task for the future.
The Epilogue, entitled “The Future of Human History as a Science,” lays
out some pieces of the residue, including the problem of the differences
between different parts of Eurasia, the role of cultural factors unrelated to
environment, and the role of individuals. Perhaps the biggest of these
unsolved problems is to establish human history as a historical science, on a
par with recognized historical sciences such as evolutionary biology, geology,
and climatology. The study of human history does pose real difficulties, but
those recognized historical sciences encounter some of the same challenges.
Hence the methods developed in some of these other fields may also prove
useful in the field of human history.
Already, though, I hope to have convinced you, the reader, that history is
not “just one damn fact after another,” as a cynic put it. There really are broad
patterns to history, and the search for their explanation is as productive as it is
fascinating.
PART ONE
FROM EDEN TO CAJAMARCA
CHAPTER 1
UP TO THE STARTING LINE
A SUITABLE STARTING POINT FROM WHICH TO COMPARE historical developments
on the different continents is around 11,000
*
B.C. This date corresponds
approximately to the beginnings of village life in a few parts of the world, the
first undisputed peopling of the Americas, the end of the Pleistocene Era and
last Ice Age, and the start of what geologists term the Recent Era. Plant and
animal domestication began in at least one part of the world within a few
thousand years of that date. As of then, did the people of some continents
already have a head start or a clear advantage over peoples of other
continents?
If so, perhaps that head start, amplified over the last 13,000 years,
provides the answer to Yali’s question. Hence this chapter will offer a
whirlwind tour of human history on all the continents, for millions of years,
from our origins as a species until 13,000 years ago. All that will now be
summarized in less than 20 pages. Naturally, I shall gloss over details and
mention only what seem to me the trends most relevant to this book.
Our closest living relatives are three surviving species of great ape: the
gorilla, the common chimpanzee, and the pygmy chimpanzee (also known as
bonobo). Their confinement to Africa, along with abundant fossil evidence,
indicates that the earliest stages of human evolution were also played out in
Africa. Human history, as something separate from the history of animals,
began there about 7 million years ago (estimates range from 5 to 9 million
years ago). Around that time, a population of African apes broke up into
several populations, of which one proceeded to evolve into modern gorillas, a
second into the two modern chimps, and the third into humans. The gorilla
line apparently split off slightly before the split between the chimp and the
human lines.
Fossils indicate that the evolutionary line leading to us had achieved a
substantially upright posture by around 4 million years ago, then began to
increase in body size and in relative brain size around 2.5 million years ago.
Those protohumans are generally known as Australopithecus africanus,
Homo habilis, and Homo erectus, which apparently evolved into each other in
that sequence. Although Homo erectus, the stage reached around 1.7 million
years ago, was close to us modern humans in body size, its brain size was still
barely half of ours. Stone tools became common around 2.5 million years ago,
but they were merely the crudest of flaked or battered stones. In zoological
significance and distinctiveness, Homo erectus was more than an ape, but still
much less than a modern human.
All of that human history, for the first 5 or 6 million years after our
origins about 7 million years ago, remained confined to Africa. The first
human ancestor to spread beyond Africa was Homo erectus, as is attested by
fossils discovered on the Southeast Asian island of Java and conventionally
known as Java man (see Figure 1.1). The oldest Java “man” fossils—of
course, they may actually have belonged to a Java woman—have usually
been assumed to date from about a million years ago. However, it has recently
been argued that they actually date from 1.8 million years ago. (Strictly
speaking, the name Homo erectus belongs to these Javan fossils, and the
African fossils classified as Homo erectus may warrant a different name.) At
present, the earliest unquestioned evidence for humans in Europe stems from
around half a million years ago, but there are claims of an earlier presence.
One would certainly assume that the colonization of Asia also permitted the
simultaneous colonization of Europe, since Eurasia is a single landmass not
bisected by major barriers.
That illustrates an issue that will recur throughout this book. Whenever
some scientist claims to have discovered “the earliest X”—whether X is the
earliest human fossil in Europe, the earliest evidence of domesticated corn in
Mexico, or the earliest anything anywhere—that announcement challenges
other scientists to beat the claim by finding something still earlier. In reality,
there must be some truly “earliest X,” with all claims of earlier X’s being
false. However, as we shall see, for virtually any X, every year brings forth
new discoveries and claims of a purported still earlier X, along with
refutations of some or all of previous years’ claims of earlier X. It often takes
decades of searching before archaeologists reach a consensus on such
questions.
By about half a million years ago, human fossils had diverged from older
Homo erectus skeletons in their enlarged, rounder, and less angular skulls.
African and European skulls of half a million years ago were sufficiently
similar to skulls of us moderns that they are classified in our species, Homo
sapiens, instead of in Homo erectus. This distinction is arbitrary, since Homo erectus evolved into Homo sapiens. However, these early Homo sapiens still
differed from us in skeletal details, had brains significantly smaller than ours,
and were grossly different from us in their artifacts and behavior. Modern
stone-tool-making peoples, such as Yali’s great-grandparents, would have
scorned the stone tools of half a million years ago as very crude. The only
other significant addition to our ancestors’ cultural repertoire that can be
documented with confidence around that time was the use of fire.
No art, bone tool, or anything else has come down to us from early Homo
sapiens except for their skeletal remains, plus those crude stone tools. There
were still no humans in Australia, for the obvious reason that it would have
taken boats to get there from Southeast Asia. There were also no humans
anywhere in the Americas, because that would have required the occupation
of the nearest part of the Eurasian continent (Siberia), and possibly boat-
building skills as well. (The present, shallow Bering Strait, separating Siberia
from Alaska, alternated between a strait and a broad intercontinental bridge of
dry land, as sea level repeatedly rose and fell during the Ice Ages.) However,
boat building and survival in cold Siberia were both still far beyond the
capabilities of early Homo sapiens.
After half a million years ago, the human populations of Africa and
western Eurasia proceeded to diverge from each other and from East Asian
populations in skeletal details. The population of Europe and western Asia
between 130,000 and 40,000 years ago is represented by especially many
skeletons, known as Neanderthals and sometimes classified as a separate
species, Homo neanderthalensis. Despite being depicted in innumerable
cartoons as apelike brutes living in caves, Neanderthals had brains slightly
larger than our own. They were also the first humans to leave behind strong
evidence of burying their dead and caring for their sick. Yet their stone tools
were still crude by comparison with modern New Guineans’ polished stone
axes and were usually not yet made in standardized diverse shapes, each with
a clearly recognizable function.
The few preserved African skeletal fragments contemporary with the
Neanderthals are more similar to our modern skeletons than to Neanderthal
skeletons. Even fewer preserved East Asian skeletal fragments are known, but
they appear different again from both Africans and Neanderthals. As for the
lifestyle at that time, the best-preserved evidence comes from stone artifacts
and prey bones accumulated at southern African sites. Although those
Africans of 100,000 years ago had more modern skeletons than did their
Neanderthal contemporaries, they made essentially the same crude stone tools
as Neanderthals, still lacking standardized shapes. They had no preserved art.
To judge from the bone evidence of the animal species on which they preyed,
their hunting skills were unimpressive and mainly directed at easy-to-kill, not-
at-all-dangerous animals. They were not yet in the business of slaughtering
buffalo, pigs, and other dangerous prey. They couldn’t even catch fish: their
sites immediately on the seacoast lack fish bones and fishhooks. They and
their Neanderthal contemporaries still rank as less than fully human.
Human history at last took off around 50,000 years ago, at the time of
what I have termed our Great Leap Forward. The earliest definite signs of that
leap come from East African sites with standardized stone tools and the first
preserved jewelry (ostrich-shell beads). Similar developments soon appear in
the Near East and in southeastern Europe, then (some 40,000 years ago) in
southwestern Europe, where abundant artifacts are associated with fully
modern skeletons of people termed Cro-Magnons. Thereafter, the garbage
preserved at archaeological sites rapidly becomes more and more interesting
and leaves no doubt that we are dealing with biologically and behaviorally
modern humans.
Cro-Magnon garbage heaps yield not only stone tools but also tools of
bone, whose suitability for shaping (for instance, into fishhooks) had
apparently gone unrecognized by previous humans. Tools were produced in
diverse and distinctive shapes so modern that their functions as needles, awls,
engraving tools, and so on are obvious to us. Instead of only single-piece tools
such as hand-held scrapers, multipiece tools made their appearance.
Recognizable multipiece weapons at Cro-Magnon sites include harpoons,
spear-throwers, and eventually bows and arrows, the precursors of rifles and
other multipiece modern weapons. Those efficient means of killing at a safe
distance permitted the hunting of such dangerous prey as rhinos and
elephants, while the invention of rope for nets, lines, and snares allowed the
addition of fish and birds to our diet. Remains of houses and sewn clothing
testify to a greatly improved ability to survive in cold climates, and remains
of jewelry and carefully buried skeletons indicate revolutionary aesthetic and
spiritual developments.
Of the Cro-Magnons’ products that have been preserved, the best known
are their artworks: their magnificent cave paintings, statues, and musical
instruments, which we still appreciate as art today. Anyone who has
experienced firsthand the overwhelming power of the life-sized painted bulls
and horses in the Lascaux Cave of southwestern France will understand at
once that their creators must have been as modern in their minds as they were
in their skeletons.
Obviously, some momentous change took place in our ancestors’
capabilities between about 100,000 and 50,000 years ago. That Great Leap
Forward poses two major unresolved questions, regarding its triggering cause
and its geographic location. As for its cause, I argued in my book The Third
Chimpanzee for the perfection of the voice box and hence for the anatomical
basis of modern language, on which the exercise of human creativity is so
dependent. Others have suggested instead that a change in brain organization
around that time, without a change in brain size, made modern language
possible.
As for the site of the Great Leap Forward, did it take place primarily in
one geographic area, in one group of humans, who were thereby enabled to
expand and replace the former human populations of other parts of the world?
Or did it occur in parallel in different regions, in each of which the human
populations living there today would be descendants of the populations living
there before the leap? The rather modern-looking human skulls from Africa
around 100,000 years ago have been taken to support the former view, with
the leap occurring specifically in Africa. Molecular studies (of so-called
mitochondrial DNA) were initially also interpreted in terms of an African
origin of modern humans, though the meaning of those molecular findings is
currently in doubt. On the other hand, skulls of humans living in China and
Indonesia hundreds of thousands of years ago are considered by some
physical anthropologists to exhibit features still found in modern Chinese and
in Aboriginal Australians, respectively. If true, that finding would suggest
parallel evolution and multiregional origins of modern humans, rather than
origins in a single Garden of Eden. The issue remains unresolved.
The evidence for a localized origin of modern humans, followed by their
spread and then their replacement of other types of humans elsewhere, seems
strongest for Europe. Some 40,000 years ago, into Europe came the Cro-
Magnons, with their modern skeletons, superior weapons, and other advanced
cultural traits. Within a few thousand years there were no more Neanderthals,
who had been evolving as the sole occupants of Europe for hundreds of
thousands of years. That sequence strongly suggests that the modern Cro-
Magnons somehow used their far superior technology, and their language
skills or brains, to infect, kill, or displace the Neanderthals, leaving behind
little or no evidence of hybridization between Neanderthals and Cro-
Magnons.
THE GREAT LEAP FORWARD coincides with the first proven major extension of
human geographic range since our ancestors’ colonization of Eurasia. That
extension consisted of the occupation of Australia and New Guinea, joined at
that time into a single continent. Many radiocarbon-dated sites attest to
human presence in Australia / New Guinea between 40,000 and 30,000 years
ago (plus the inevitable somewhat older claims of contested validity). Within
a short time of that initial peopling, humans had expanded over the whole
continent and adapted to its diverse habitats, from the tropical rain forests and
high mountains of New Guinea to the dry interior and wet southeastern corner
of Australia.
During the Ice Ages, so much of the oceans’ water was locked up in
glaciers that worldwide sea levels dropped hundreds of feet below their
present stand. As a result, what are now the shallow seas between Asia and
the Indonesian islands of Sumatra, Borneo, Java, and Bali became dry land.
(So did other shallow straits, such as the Bering Strait and the English
Channel.) The edge of the Southeast Asian mainland then lay 700 miles east
of its present location. Nevertheless, central Indonesian islands between Bali
and Australia remained surrounded and separated by deepwater channels. To
reach Australia / New Guinea from the Asian mainland at that time still
required crossing a minimum of eight channels, the broadest of which was at
least 50 miles wide. Most of those channels divided islands visible from each
other, but Australia itself was always invisible from even the nearest
Indonesian islands, Timor and Tanimbar. Thus, the occupation of Australia /
New Guinea is momentous in that it demanded watercraft and provides by far
the earliest evidence of their use in history. Not until about 30,000 years later
(13,000 years ago) is there strong evidence of watercraft anywhere else in the
world, from the Mediterranean.
Initially, archaeologists considered the possibility that the colonization of
Australia / New Guinea was achieved accidentally by just a few people swept
to sea while fishing on a raft near an Indonesian island. In an extreme
scenario the first settlers are pictured as having consisted of a single pregnant
young woman carrying a male fetus. But believers in the fluke-colonization
theory have been surprised by recent discoveries that still other islands, lying
to the east of New Guinea, were colonized soon after New Guinea itself, by
around 35,000 years ago. Those islands were New Britain and New Ireland, in
the Bismarck Archipelago, and Buka, in the Solomon Archipelago. Buka lies
out of sight of the closest island to the west and could have been reached only
by crossing a water gap of about 100 miles. Thus, early Australians and New
Guineans were probably capable of intentionally traveling over water to
visible islands, and were using watercraft sufficiently often that the
colonization of even invisible distant islands was repeatedly achieved
unintentionally.
The settlement of Australia / New Guinea was perhaps associated with
still another big first, besides humans’ first use of watercraft and first range
extension since reaching Eurasia: the first mass extermination of large animal
species by humans. Today, we regard Africa as the continent of big mammals.
Modern Eurasia also has many species of big mammals (though not in the
manifest abundance of Africa’s Serengeti Plains), such as Asia’s rhinos and
elephants and tigers, and Europe’s moose and bears and (until classical times)
lions. Australia / New Guinea today has no equally large mammals, in fact no
mammal larger than 100-pound kangaroos. But Australia / New Guinea
formerly had its own suite of diverse big mammals, including giant
kangaroos, rhinolike marsupials called diprotodonts and reaching the size of a
cow, and a marsupial “leopard.” It also formerly had a 400-pound ostrichlike
flightless bird, plus some impressively big reptiles, including a one-ton lizard,
a giant python, and land-dwelling crocodiles.
All of those Australian / New Guinean giants (the so-called megafauna)
disappeared after the arrival of humans. While there has been controversy
about the exact timing of their demise, several Australian archaeological sites,
with dates extending over tens of thousands of years, and with prodigiously
abundant deposits of animal bones, have been carefully excavated and found
to contain not a trace of the now extinct giants over the last 35,000 years.
Hence the megafauna probably became extinct soon after humans reached
Australia.
The near-simultaneous disappearance of so many large species raises an
obvious question: what caused it? An obvious possible answer is that they
were killed off or else eliminated indirectly by the first arriving humans.
Recall that Australian / New Guinean animals had evolved for millions of
years in the absence of human hunters. We know that Galápagos and
Antarctic birds and mammals, which similarly evolved in the absence of
humans and did not see humans until modern times, are still incurably tame
today. They would have been exterminated if conservationists had not
imposed protective measures quickly. On other recently discovered islands
where protective measures did not go into effect quickly, exterminations did
indeed result: one such victim, the dodo of Mauritius, has become virtually a
symbol for extinction. We also know now that, on every one of the well-
studied oceanic islands colonized in the prehistoric era, human colonization
led to an extinction spasm whose victims included the moas of New Zealand,
the giant lemurs of Madagascar, and the big flightless geese of Hawaii. Just as
modern humans walked up to unafraid dodos and island seals and killed them,
prehistoric humans presumably walked up to unafraid moas and giant lemurs
and killed them too.
Hence one hypothesis for the demise of Australia’s and New Guinea’s
giants is that they met the same fate around 40,000 years ago. In contrast,
most big mammals of Africa and Eurasia survived into modern times, because
they had coevolved with protohumans for hundreds of thousands or millions
of years. They thereby enjoyed ample time to evolve a fear of humans, as our
ancestors’ initially poor hunting skills slowly improved. The dodo, moas, and
perhaps the giants of Australia / New Guinea had the misfortune suddenly to
be confronted, without any evolutionary preparation, by invading modern
humans possessing fully developed hunting skills.
However, the overkill hypothesis, as it is termed, has not gone
unchallenged for Australia / New Guinea. Critics emphasize that, as yet, no
one has documented the bones of an extinct Australian / New Guinean giant
with compelling evidence of its having been killed by humans, or even of its
having lived in association with humans. Defenders of the overkill hypothesis
reply: you would hardly expect to find kill sites if the extermination was
completed very quickly and long ago, such as within a few millennia some
40,000 years ago. The critics respond with a countertheory: perhaps the giants
succumbed instead to a change in climate, such as a severe drought on the
already chronically dry Australian continent. The debate goes on.
Personally, I can’t fathom why Australia’s giants should have survived
innumerable droughts in their tens of millions of years of Australian history,
and then have chosen to drop dead almost simultaneously (at least on a time
scale of millions of years) precisely and just coincidentally when the first
humans arrived. The giants became extinct not only in dry central Australia
but also in drenching wet New Guinea and southeastern Australia. They
became extinct in every habitat without exception, from deserts to cold rain
forest and tropical rain forest. Hence it seems to me most likely that the giants
were indeed exterminated by humans, both directly (by being killed for food)
and indirectly (as the result of fires and habitat modification caused by
humans). But regardless of whether the overkill hypothesis or the climate
hypothesis proves correct, the disappearance of all of the big animals of
Australia / New Guinea had, as we shall see, heavy consequences for
subsequent human history. Those extinctions eliminated all the large wild
animals that might otherwise have been candidates for domestication, and left
native Australians and New Guineans with not a single native domestic
animal.
THUS, THE COLONIZATION of Australia/New Guinea was not achieved until
around the time of the Great Leap Forward. Another extension of human
range that soon followed was the one into the coldest parts of Eurasia. While
Neanderthals lived in glacial times and were adapted to the cold, they
penetrated no farther north than northern Germany and Kiev. That’s not
surprising, since Neanderthals apparently lacked needles, sewn clothing,
warm houses, and other technology essential to survival in the coldest
climates. Anatomically modern peoples who did possess such technology had
expanded into Siberia by around 20,000 years ago (there are the usual much
older disputed claims). That expansion may have been responsible for the
extinction of Eurasia’s woolly mammoth and woolly rhinoceros.
With the settlement of Australia / New Guinea, humans now occupied
three of the five habitable continents. (Throughout this book, I count Eurasia
as a single continent, and I omit Antarctica because it was not reached by
humans until the 19th century and has never had any self-supporting human
population.) That left only two continents, North America and South
America. They were surely the last ones settled, for the obvious reason that
reaching the Americas from the Old World required either boats (for which
there is no evidence even in Indonesia until 40,000 years ago and none in
Europe until much later) in order to cross by sea, or else it required the
occupation of Siberia (unoccupied until about 20,000 years ago) in order to
cross the Bering land bridge.
However, it is uncertain when, between about 14,000 and 35,000 years
ago, the Americas were first colonized. The oldest unquestioned human
remains in the Americas are at sites in Alaska dated around 12,000 B.C.,
followed by a profusion of sites in the United States south of the Canadian
border and in Mexico in the centuries just before 11,000 B.C. The latter sites
are called Clovis sites, named after the type site near the town of Clovis, New
Mexico, where their characteristic large stone spearpoints were first
recognized. Hundreds of Clovis sites are now known, blanketing all 48 of the
lower U.S. states south into Mexico. Unquestioned evidence of human
presence appears soon thereafter in Amazonia and in Patagonia. These facts
suggest the interpretation that Clovis sites document the Americas’ first
colonization by people, who quickly multiplied, expanded, and filled the two
continents.
One might at first be surprised that Clovis descendants could reach
Patagonia, lying 8,000 miles south of the U.S.-Canada border, in less than a
thousand years. However, that translates into an average expansion of only 8
miles per year, a trivial feat for a hunter-gatherer likely to cover that distance
even within a single day’s normal foraging.
One might also at first be surprised that the Americas evidently filled up
with humans so quickly that people were motivated to keep spreading south
toward Patagonia. That population growth also proves unsurprising when one
stops to consider the actual numbers. If the Americas eventually came to hold
hunter-gatherers at an average population density of somewhat under one
person per square mile (a high value for modern hunter-gatherers), then the
whole area of the Americas would eventually have held about 10 million
hunter-gatherers. But even if the initial colonists had consisted of only 100
people and their numbers had increased at a rate of only 1.1 percent per year,
the colonists’ descendants would have reached that population ceiling of 10
million people within a thousand years. A population growth rate of 1.1
percent per year is again trivial: rates as high as 3.4 percent per year have
been observed in modern times when people colonized virgin lands, such as
when the HMS Bounty mutineers and their Tahitian wives colonized Pitcairn
Island.
The profusion of Clovis hunters’ sites within the first few centuries after
their arrival resembles the site profusion documented archaeologically for the
more recent discovery of New Zealand by ancestral Maori. A profusion of
early sites is also documented for the much older colonization of Europe by
anatomically modern humans, and for the occupation of Australia / New
Guinea. That is, everything about the Clovis phenomenon and its spread
through the Americas corresponds to findings for other, unquestioned virgin-
land colonizations in history.
What might be the significance of Clovis sites’ bursting forth in the
centuries just before 11,000 B.C., rather than in those before 16,000 or 21,000
B.C.? Recall that Siberia has always been cold, and that a continuous ice sheet
stretched as an impassable barrier across the whole width of Canada during
much of the Pleistocene Ice Ages. We have already seen that the technology
required for coping with extreme cold did not emerge until after anatomically
modern humans invaded Europe around 40,000 years ago, and that people did
not colonize Siberia until 20,000 years later. Eventually, those early Siberians
crossed to Alaska, either by sea across the Bering Strait (only 50 miles wide
even today) or else on foot at glacial times when Bering Strait was dry land.
The Bering land bridge, during its millennia of intermittent existence, would
have been up to a thousand miles wide, covered by open tundra, and easily
traversable by people adapted to cold conditions. The land bridge was flooded
and became a strait again most recently when sea level rose after around
14,000 B.C. Whether those early Siberians walked or paddled to Alaska, the
earliest secure evidence of human presence in Alaska dates from around
12,000 B.C.
Soon thereafter, a north–south ice-free corridor opened in the Canadian
ice sheet, permitting the first Alaskans to pass through and come out into the
Great Plains around the site of the modern Canadian city of Edmonton. That
removed the last serious barrier between Alaska and Patagonia for modern
humans. The Edmonton pioneers would have found the Great Plains teeming
with game. They would have thrived, increased in numbers, and gradually
spread south to occupy the whole hemisphere.
One other feature of the Clovis phenomenon fits our expectations for the
first human presence south of the Canadian ice sheet. Like Australia / New
Guinea, the Americas had originally been full of big mammals. About 15,000
years ago, the American West looked much as Africa’s Serengeti Plains do
today, with herds of elephants and horses pursued by lions and cheetahs, and
joined by members of such exotic species as camels and giant ground sloths.
Just as in Australia / New Guinea, in the Americas most of those large
mammals became extinct. Whereas the extinctions took place probably before
30,000 years ago in Australia, they occurred around 17,000 to 12,000 years
ago in the Americas. For those extinct American mammals whose bones are
available in greatest abundance and have been dated especially accurately,
one can pinpoint the extinctions as having occurred around 11,000 B.C.
Perhaps the two most accurately dated extinctions are those of the Shasta
ground sloth and Harrington’s mountain goat in the Grand Canyon area; both
of those populations disappeared within a century or two of 11,100 B.C.
Whether coincidentally or not, that date is identical, within experimental
error, to the date of Clovis hunters’ arrival in the Grand Canyon area.
The discovery of numerous skeletons of mammoths with Clovis
spearpoints between their ribs suggests that this agreement of dates is not a
coincidence. Hunters expanding southward through the Americas,
encountering big animals that had never seen humans before, may have found
those American animals easy to kill and may have exterminated them. A
countertheory is that America’s big mammals instead became extinct because
of climate changes at the end of the last Ice Age, which (to confuse the
interpretation for modern paleontologists) also happened around 11,000 B.C.
Personally, I have the same problem with a climatic theory of megafaunal
extinction in the Americas as with such a theory in Australia / New Guinea.
The Americas’ big animals had already survived the ends of 22 previous Ice
Ages. Why did most of them pick the 23rd to expire in concert, in the
presence of all those supposedly harmless humans? Why did they disappear in
all habitats, not only in habitats that contracted but also in ones that greatly
expanded at the end of the last Ice Age? Hence I suspect that Clovis hunters
did it, but the debate remains unresolved. Whichever theory proves correct,
most large wild mammal species that might otherwise have later been
domesticated by Native Americans were thereby removed.
Also unresolved is the question whether Clovis hunters really were the
first Americans. As always happens whenever anyone claims the first
anything, claims of discoveries of pre-Clovis human sites in the Americas are
constantly being advanced. Every year, a few of those new claims really do
appear convincing and exciting when initially announced. Then the inevitable
problems of interpretation arise. Were the reported tools at the site really tools
made by humans, or just natural rock shapes? Are the reported radiocarbon
dates really correct, and not invalidated by any of the numerous difficulties
that can plague radiocarbon dating? If the dates are correct, are they really
associated with human products, rather than just being a 15,000-year-old
lump of charcoal lying next to a stone tool actually made 9,000 years ago?
To illustrate these problems, consider the following typical example of an
often quoted pre-Clovis claim. At a Brazilian rock shelter named Pedra
Furada, archaeologists found cave paintings undoubtedly made by humans.
They also discovered, among the piles of stones at the base of a cliff, some
stones whose shapes suggested the possibility of their being crude tools. In
addition, they came upon supposed hearths, whose burnt charcoal yielded
radiocarbon dates of around 35,000 years ago. Articles on Pedra Furada were
accepted for publication in the prestigious and highly selective international
scientific journal Nature.
But none of those rocks at the base of the cliff is an obviously
humanmade tool, as are Clovis points and Cro-Magnon tools. If hundreds of
thousands of rocks fall from a high cliff over the course of tens of thousands
of years, many of them will become chipped and broken when they hit the
rocks below, and some will come to resemble crude tools chipped and broken
by humans. In western Europe and elsewhere in Amazonia, archaeologists
have radiocarbon-dated the actual pigments used in cave paintings, but that
was not done at Pedra Furada. Forest fires occur frequently in the vicinity and
produce charcoal that is regularly swept into caves by wind and streams. No
evidence links the 35,000-year-old charcoal to the undoubted cave paintings
at Pedra Furada. Although the original excavators remain convinced, a team
of archaeologists who were not involved in the excavation but receptive to
pre-Clovis claims recently visited the site and came away unconvinced.
The North American site that currently enjoys the strongest credentials as
a possible pre-Clovis site is Meadowcroft rock shelter, in Pennsylvania,
yielding reported human-associated radiocarbon dates of about 16,000 years
ago. At Meadowcroft no archaeologist denies that many human artifacts do
occur in many carefully excavated layers. But the oldest radiocarbon dates
don’t make sense, because the plant and animal species associated with them
are species living in Pennsylvania in recent times of mild climates, rather than
species expected for the glacial times of 16,000 years ago. Hence one has to
suspect that the charcoal samples dated from the oldest human occupation
levels consist of post-Clovis charcoal infiltrated with older carbon. The
strongest pre-Clovis candidate in South America is the Monte Verde site, in
southern Chile, dated to at least 15,000 years ago. It too now seems
convincing to many archaeologists, but caution is warranted in view of all the
previous disillusionments.
If there really were pre-Clovis people in the Americas, why is it still so
hard to prove that they existed? Archaeologists have excavated hundreds of
American sites unequivocally dating to between 2000 and 11,000 B.C.,
including dozens of Clovis sites in the North American West, rock shelters in
the Appalachians, and sites in coastal California. Below all the archaeological
layers with undoubted human presence, at many of those same sites, deeper
older layers have been excavated and still yield undoubted remains of animals
—but with no further evidence of humans. The weaknesses in pre-Clovis
evidence in the Americas contrast with the strength of the evidence in Europe,
where hundreds of sites attest to the presence of modern humans long before
Clovis hunters appeared in the Americas around 11,000 B.C. Even more
striking is the evidence from Australia / New Guinea, where there are barely
one-tenth as many archaeologists as in the United States alone, but where
those few archaeologists have nevertheless discovered over a hundred
unequivocal pre-Clovis sites scattered over the whole continent.
Early humans certainly didn’t fly by helicopter from Alaska to
Meadowcroft and Monte Verde, skipping all the landscape in between.
Advocates of pre-Clovis settlement suggest that, for thousands or even tens of
thousands of years, pre-Clovis humans remained at low population densities
or poorly visible archaeologically, for unknown reasons unprecedented
elsewhere in the world. I find that suggestion infinitely more implausible than
the suggestion that Monte Verde and Meadowcroft will eventually be
reinterpreted, as have other claimed pre-Clovis sites. My feeling is that, if
there really had been pre-Clovis settlement in the Americas, it would have
become obvious at many locations by now, and we would not still be arguing.
However, archaeologists remain divided on these questions.
The consequences for our understanding of later American prehistory
remain the same, whichever interpretation proves correct. Either: the
Americas were first settled around 11,000 B.C. and quickly filled up with
people. Or else: the first settlement occurred somewhat earlier (most
advocates of pre-Clovis settlement would suggest by 15,000 or 20,000 years
ago, possibly 30,000 years ago, and few would seriously claim earlier); but
those pre-Clovis settlers remained few in numbers, or inconspicuous, or had
little impact, until around 11,000 B.C. In either case, of the five habitable
continents, North America and South America are the ones with the shortest
human prehistories.
WITH THE OCCUPATION of the Americas, most habitable areas of the
continents and continental islands, plus oceanic islands from Indonesia to east
of New Guinea, supported humans. The settlement of the world’s remaining
islands was not completed until modern times: Mediterranean islands such as
Crete, Cyprus, Corsica, and Sardinia between about 8500 and 4000 B.C.;
Caribbean islands beginning around 4000 B.C.; Polynesian and Micronesian
islands between 1200 B.C. and A.D. 1000; Madagascar sometime between A.D.
300 and 800; and Iceland in the ninth century A.D. Native Americans, possibly
ancestral to the modern Inuit, spread throughout the High Arctic around 2000
B.C. That left, as the sole uninhabited areas awaiting European explorers over
the last 700 years, only the most remote islands of the Atlantic and Indian
Oceans (such as the Azores and Seychelles), plus Antarctica.
What significance, if any, do the continents’ differing dates of settlement
have for subsequent history? Suppose that a time machine could have
transported an archaeologist back in time, for a world tour at around 11,000
B.C. Given the state of the world then, could the archaeologist have predicted
the sequence in which human societies on the various continents would
develop guns, germs, and steel, and thus predicted the state of the world
today?
Our archaeologist might have considered the possible advantages of a
head start. If that counted for anything, then Africa enjoyed an enormous
advantage: at least 5 million more years of separate protohuman existence
than on any other continent. In addition, if it is true that modern humans arose
in Africa around 100,000 years ago and spread to other continents, that would
have wiped out any advantages accumulated elsewhere in the meantime and
given Africans a new head start. Furthermore, human genetic diversity is
highest in Africa; perhaps more-diverse humans would collectively produce
more-diverse inventions.
But our archaeologist might then reflect: what, really, does a “head start”
mean for the purposes of this book? We cannot take the metaphor of a
footrace literally. If by head start you mean the time required to populate a
continent after the arrival of the first few pioneering colonists, that time is
relatively brief: for example, less than 1,000 years to fill up even the whole
New World. If by head start you instead mean the time required to adapt to
local conditions, I grant that some extreme environments did take time: for
instance, 9,000 years to occupy the High Arctic after the occupation of the
rest of North America. But people would have explored and adapted to most
other areas quickly, once modern human inventiveness had developed. For
example, after the ancestors of the Maori reached New Zealand, it apparently
took them barely a century to discover all worthwhile stone sources; only a
few more centuries to kill every last moa in some of the world’s most rugged
terrain; and only a few centuries to differentiate into a range of diverse
societies, from that of coastal hunter-gatherers to that of farmers practicing
new types of food storage.
Our archaeologist might therefore look at the Americas and conclude that
Africans, despite their apparently enormous head start, would have been
overtaken by the earliest Americans within at most a millennium. Thereafter,
the Americas’ greater area (50 percent greater than Africa’s) and much greater
environmental diversity would have given the advantage to Native Americans
over Africans.
The archaeologist might then turn to Eurasia and reason as follows.
Eurasia is the world’s largest continent. It has been occupied for longer than
any other continent except Africa. Africa’s long occupation before the
colonization of Eurasia a million years ago might have counted for nothing
anyway, because protohumans were at such a primitive stage then. Our
archaeologist might look at the Upper Paleolithic flowering of southwestern
Europe between 20,000 and 12,000 years ago, with all those famous artworks
and complex tools, and wonder whether Eurasia was already getting a head
start then, at least locally.
Finally, the archaeologist would turn to Australia / New Guinea, noting
first its small area (it’s the smallest continent), the large fraction of it covered
by desert capable of supporting few humans, the continent’s isolation, and its
later occupation than that of Africa and Eurasia. All that might lead the
archaeologist to predict slow development in Australia / New Guinea.
But remember that Australians and New Guineans had by far the earliest
watercraft in the world. They were creating cave paintings apparently at least
as early as the Cro-Magnons in Europe. Jonathan Kingdon and Tim Flannery
have noted that the colonization of Australia / New Guinea from the islands of
the Asian continental shelf required humans to learn to deal with the new
environments they encountered on the islands of central Indonesia—a maze
of coastlines offering the richest marine resources, coral reefs, and mangroves
in the world. As the colonists crossed the straits separating each Indonesian
island from the next one to the east, they adapted anew, filled up that next
island, and went on to colonize the next island again. It was a hitherto
unprecedented golden age of successive human population explosions.
Perhaps those cycles of colonization, adaptation, and population explosion
were what selected for the Great Leap Forward, which then diffused back
westward to Eurasia and Africa. If this scenario is correct, then Australia /
New Guinea gained a massive head start that might have continued to propel
human development there long after the Great Leap Forward.
Thus, an observer transported back in time to 11,000 B.C. could not have
predicted on which continent human societies would develop most quickly,
but could have made a strong case for any of the continents. With hindsight,
of course, we know that Eurasia was the one. But it turns out that the actual
reasons behind the more rapid development of Eurasian societies were not at
all the straightforward ones that our imaginary archaeologist of 11,000 B.C.
guessed. The remainder of this book consists of a quest to discover those real
reasons.
CHAPTER 2
A NATURAL EXPERIMENT OF HISTORY
ON THE CHATHAM ISLANDS, 500 MILES EAST OF NEW ZEALAND, centuries of
independence came to a brutal end for the Moriori people in December 1835.
On November 19 of that year, a ship carrying 500 Maori armed with guns,
clubs, and axes arrived, followed on December 5 by a shipload of 400 more
Maori. Groups of Maori began to walk through Moriori settlements,
announcing that the Moriori were now their slaves, and killing those who
objected. An organized resistance by the Moriori could still then have
defeated the Maori, who were outnumbered two to one. However, the Moriori
had a tradition of resolving disputes peacefully. They decided in a council
meeting not to fight back but to offer peace, friendship, and a division of
resources.
Before the Moriori could deliver that offer, the Maori attacked en masse.
Over the course of the next few days, they killed hundreds of Moriori, cooked
and ate many of the bodies, and enslaved all the others, killing most of them
too over the next few years as it suited their whim. A Moriori survivor
recalled, “[The Maori] commenced to kill us like sheep…. [We] were
terrified, fled to the bush, concealed ourselves in holes underground, and in
any place to escape our enemies. It was of no avail; we were discovered and
killed—men, women, and children indiscriminately.” A Maori conqueror
explained, “We took possession…in accordance with our customs and we
caught all the people. Not one escaped. Some ran away from us, these we
killed, and others we killed—but what of that? It was in accordance with our
custom.”
The brutal outcome of this collision between the Moriori and the Maori
could have been easily predicted. The Moriori were a small, isolated
population of hunter-gatherers, equipped with only the simplest technology
and weapons, entirely inexperienced at war, and lacking strong leadership or
organization. The Maori invaders (from New Zealand’s North Island) came
from a dense population of farmers chronically engaged in ferocious wars,
equipped with more-advanced technology and weapons, and operating under
strong leadership. Of course, when the two groups finally came into contact, it
was the Maori who slaughtered the Moriori, not vice versa.
The tragedy of the Moriori resembles many other such tragedies in both
the modern and the ancient world, pitting numerous well-equipped people
against few ill-equipped opponents. What makes the Maori-Moriori collision
grimly illuminating is that both groups had diverged from a common origin
less than a millennium earlier. Both were Polynesian peoples. The modern
Maori are descendants of Polynesian farmers who colonized New Zealand
around A.D. 1000. Soon thereafter, a group of those Maori in turn colonized
the Chatham Islands and became the Moriori. In the centuries after the two
groups separated, they evolved in opposite directions, the North Island Maori
developing more-complex and the Moriori less-complex technology and
political organization. The Moriori reverted to being hunter-gatherers, while
the North Island Maori turned to more intensive farming.
Those opposite evolutionary courses sealed the outcome of their eventual
collision. If we could understand the reasons for the disparate development of
those two island societies, we might have a model for understanding the
broader question of differing developments on the continents.
MORIORI AND MAORI history constitutes a brief, small-scale natural
experiment that tests how environments affect human societies. Before you
read a whole book examining environmental effects on a very large scale—
effects on human societies around the world for the last 13,000 years—you
might reasonably want assurance, from smaller tests, that such effects really
are significant. If you were a laboratory scientist studying rats, you might
perform such a test by taking one rat colony, distributing groups of those
ancestral rats among many cages with differing environments, and coming
back many rat generations later to see what had happened. Of course, such
purposeful experiments cannot be carried out on human societies. Instead,
scientists must look for “natural experiments,” in which something similar
befell humans in the past.
Such an experiment unfolded during the settlement of Polynesia.
Scattered over the Pacific Ocean beyond New Guinea and Melanesia are
thousands of islands differing greatly in area, isolation, elevation, climate,
productivity, and geological and biological resources (Figure 2.1). For most of
human history those islands lay far beyond the reach of watercraft. Around
1200 B.C. a group of farming, fishing, seafaring people from the Bismarck
Archipelago north of New Guinea finally succeeded in reaching some of
those islands. Over the following centuries their descendants colonized
virtually every habitable scrap of land in the Pacific. The process was mostly
complete by A.D. 500, with the last few islands settled around or soon after
A.D. 1000.
Thus, within a modest time span, enormously diverse island environments
were settled by colonists all of whom stemmed from the same founding
population. The ultimate ancestors of all modern Polynesian populations
shared essentially the same culture, language, technology, and set of
domesticated plants and animals. Hence Polynesian history constitutes a
natural experiment allowing us to study human adaptation, devoid of the
usual complications of multiple waves of disparate colonists that often
frustrate our attempts to understand adaptation elsewhere in the world.
Within that medium-sized test, the fate of the Moriori forms a smaller test.
It is easy to trace how the differing environments of the Chatham Islands and
of New Zealand molded the Moriori and the Maori differently. While those
ancestral Maori who first colonized the Chathams may have been farmers,
Maori tropical crops could not grow in the Chathams’ cold climate, and the
colonists had no alternative except to revert to being hunter-gatherers. Since
as hunter-gatherers they did not produce crop surpluses available for
redistribution or storage, they could not support and feed nonhunting craft
specialists, armies, bureaucrats, and chiefs. Their prey were seals, shellfish,
nesting seabirds, and fish that could be captured by hand or with clubs and
required no more elaborate technology. In addition, the Chathams are
relatively small and remote islands, capable of supporting a total population
of only about 2,000 hunter-gatherers. With no other accessible islands to
colonize, the Moriori had to remain in the Chathams, and to learn how to get
along with each other. They did so by renouncing war, and they reduced
potential conflicts from overpopulation by castrating some male infants. The
result was a small, unwarlike population with simple technology and
weapons, and without strong leadership or organization.
In contrast, the northern (warmer) part of New Zealand, by far the largest
island group in Polynesia, was suitable for Polynesian agriculture. Those
Maori who remained in New Zealand increased in numbers until there were
more than 100,000 of them. They developed locally dense populations
chronically engaged in ferocious wars with neighboring populations. With the
crop surpluses that they could grow and store, they fed craft specialists,
chiefs, and part-time soldiers. They needed and developed varied tools for
growing their crops, fighting, and making art. They erected elaborate
ceremonial buildings and prodigious numbers of forts.
Thus, Moriori and Maori societies developed from the same ancestral
society, but along very different lines. The resulting two societies lost
awareness even of each other’s existence and did not come into contact again
for many centuries, perhaps for as long as 500 years. Finally, an Australian
seal-hunting ship visiting the Chathams en route to New Zealand brought the
news to New Zealand of islands where “there is an abundance of sea and
shellfish; the lakes swarm with eels; and it is a land of the karaka berry….
The inhabitants are very numerous, but they do not understand how to fight,
and have no weapons.” That news was enough to induce 900 Maori to sail to
the Chathams. The outcome clearly illustrates how environments can affect
economy, technology, political organization, and fighting skills within a short
time.
AS I ALREADY mentioned, the Maori-Moriori collision represents a small test
within a medium-sized test. What can we learn from all of Polynesia about
environmental influences on human societies? What differences among
societies on different Polynesian islands need to be explained?
Polynesia as a whole presented a much wider range of environmental
conditions than did just New Zealand and the Chathams, although the latter
define one extreme (the simple end) of Polynesian organization. In their
subsistence modes, Polynesians ranged from the hunter-gatherers of the
Chathams, through slash-and-burn farmers, to practitioners of intensive food
production living at some of the highest population densities of any human
societies. Polynesian food producers variously intensified production of pigs,
dogs, and chickens. They organized work forces to construct large irrigation
systems for agriculture and to enclose large ponds for fish production. The
economic basis of Polynesian societies consisted of more or less self-
sufficient households, but some islands also supported guilds of hereditary
part-time craft specialists. In social organization, Polynesian societies ran the
gamut from fairly egalitarian village societies to some of the most stratified
societies in the world, with many hierarchically ranked lineages and with
chief and commoner classes whose members married within their own class.
In political organization, Polynesian islands ranged from landscapes divided
into independent tribal or village units, up to multi-island proto-empires that
devoted standing military establishments to invasions of other islands and
wars of conquest. Finally, Polynesian material culture varied from the
production of no more than personal utensils to the construction of
monumental stone architecture. How can all that variation be explained?
Contributing to these differences among Polynesian societies were at least
six sets of environmental variables among Polynesian islands: island climate,
geological type, marine resources, area, terrain fragmentation, and isolation.
Let’s examine the ranges of these factors, before considering their specific
consequences for Polynesian societies.
The climate in Polynesia varies from warm tropical or subtropical on
most islands, which lie near the equator, to temperate on most of New
Zealand, and cold subantarctic on the Chathams and the southern part of New
Zealand’s South Island. Hawaii’s Big Island, though lying well within the
Tropic of Cancer, has mountains high enough to support alpine habitats and
receive occasional snowfalls. Rainfall varies from the highest recorded on
Earth (in New Zealand’s Fjordland and Hawaii’s Alakai Swamp on Kauai) to
only one-tenth as much on islands so dry that they are marginal for
agriculture.
Island geological types include coral atolls, raised limestone, volcanic
islands, pieces of continents, and mixtures of those types. At one extreme,
innumerable islets, such as those of the Tuamotu Archipelago, are flat, low
atolls barely rising above sea level. Other former atolls, such as Henderson
and Rennell, have been lifted far above sea level to constitute raised limestone
islands. Both of those atoll types present problems to human settlers, because
they consist entirely of limestone without other stones, have only very thin
soil, and lack permanent fresh water. At the opposite extreme, the largest
Polynesian island, New Zealand, is an old, geologically diverse, continental
fragment of Gondwanaland, offering a range of mineral resources, including
commercially exploitable iron, coal, gold, and jade. Most other large
Polynesian islands are volcanoes that rose from the sea, have never formed
parts of a continent, and may or may not include areas of raised limestone.
While lacking New Zealand’s geological richness, the oceanic volcanic
islands at least are an improvement over atolls (from the Polynesians’
perspective) in that they offer diverse types of volcanic stones, some of which
are highly suitable for making stone tools.
The volcanic islands differ among themselves. The elevations of the
higher ones generate rain in the mountains, so the islands are heavily
weathered and have deep soils and permanent streams. That is true, for
instance, of the Societies, Samoa, the Marquesas, and especially Hawaii, the
Polynesian archipelago with the highest mountains. Among the lower islands,
Tonga and (to a lesser extent) Easter also have rich soil because of volcanic
ashfalls, but they lack Hawaii’s large streams.
As for marine resources, most Polynesian islands are surrounded by
shallow water and reefs, and many also encompass lagoons. Those
environments teem with fish and shellfish. However, the rocky coasts of
Easter, Pitcairn, and the Marquesas, and the steeply dropping ocean bottom
and absence of coral reefs around those islands, are much less productive of
seafood.
Area is another obvious variable, ranging from the 100 acres of Anuta, the
smallest permanently inhabited isolated Polynesian island, up to the 103,000
square miles of the minicontinent of New Zealand. The habitable terrain of
some islands, notably the Marquesas, is fragmented into steep-walled valleys
by ridges, while other islands, such as Tonga and Easter, consist of gently
rolling terrain presenting no obstacles to travel and communication.
The last environmental variable to consider is isolation. Easter Island and
the Chathams are small and so remote from other islands that, once they were
initially colonized, the societies thus founded developed in total isolation
from the rest of the world. New Zealand, Hawaii, and the Marquesas are also
very remote, but at least the latter two apparently did have some further
contact with other archipelagoes after the first colonization, and all three
consist of many islands close enough to each other for regular contact
between islands of the same archipelago. Most other Polynesian islands were
in more or less regular contact with other islands. In particular, the Tongan
Archipelago lies close enough to the Fijian, Samoan, and Wallis
Archipelagoes to have permitted regular voyaging between archipelagoes, and
eventually to permit Tongans to undertake the conquest of Fiji.
AFTER THAT BRIEF look at Polynesia’s varying environments, let’s now see
how that variation influenced Polynesian societies. Subsistence is a
convenient facet of society with which to start, since it in turn affected other
facets.
Polynesian subsistence depended on varying mixes of fishing, gathering
wild plants and marine shellfish and Crustacea, hunting terrestrial birds and
breeding seabirds, and food production. Most Polynesian islands originally
supported big flightless birds that had evolved in the absence of predators,
New Zealand’s moas and Hawaii’s flightless geese being the best-known
examples. While those birds were important food sources for the initial
colonists, especially on New Zealand’s South Island, most of them were soon
exterminated on all islands, because they were easy to hunt down. Breeding
seabirds were also quickly reduced in number but continued to be important
food sources on some islands. Marine resources were significant on most
islands but least so on Easter, Pitcairn, and the Marquesas, where people as a
result were especially dependent on food that they themselves produced.
Ancestral Polynesians brought with them three domesticated animals (the
pig, chicken, and dog) and domesticated no other animals within Polynesia.
Many islands retained all three of those species, but the more isolated
Polynesian islands lacked one or more of them, either because livestock
brought in canoes failed to survive the colonists’ long overwater journey or
because livestock that died out could not be readily obtained again from the
outside. For instance, isolated New Zealand ended up with only dogs; Easter
and Tikopia, with only chickens. Without access to coral reefs or productive
shallow waters, and with their terrestrial birds quickly exterminated, Easter
Islanders turned to constructing chicken houses for intensive poultry farming.
At best, however, these three domesticated animal species provided only
occasional meals. Polynesian food production depended mainly on
agriculture, which was impossible at subantarctic latitudes because all
Polynesian crops were tropical ones initially domesticated outside Polynesia
and brought in by colonists. The settlers of the Chathams and the cold
southern part of New Zealand’s South Island were thus forced to abandon the
farming legacy developed by their ancestors over the previous thousands of
years, and to become hunter-gatherers again.
People on the remaining Polynesian islands did practice agriculture based
on dryland crops (especially taro, yams, and sweet potatoes), irrigated crops
(mainly taro), and tree crops (such as breadfruit, bananas, and coconuts). The
productivity and relative importance of those crop types varied considerably
on different islands, depending on their environments. Human population
densities were lowest on Henderson, Rennell, and the atolls because of their
poor soil and limited fresh water. Densities were also low on temperate New
Zealand, which was too cool for some Polynesian crops. Polynesians on these
and some other islands practiced a nonintensive type of shifting, slash-and-
burn agriculture.
Other islands had rich soils but were not high enough to have large
permanent streams and hence irrigation. Inhabitants of those islands
developed intensive dryland agriculture requiring a heavy input of labor to
build terraces, carry out mulching, rotate crops, reduce or eliminate fallow
periods, and maintain tree plantations. Dryland agriculture became especially
productive on Easter, tiny Anuta, and flat and low Tonga, where Polynesians
devoted most of the land area to the growing of food.
The most productive Polynesian agriculture was taro cultivation in
irrigated fields. Among the more populous tropical islands, that option was
ruled out for Tonga by its low elevation and hence its lack of rivers. Irrigation
agriculture reached its peak on the westernmost Hawaiian islands of Kauai,
Oahu, and Molokai, which were big and wet enough to support not only large
permanent streams but also large human populations available for
construction projects. Hawaiian labor corvées built elaborate irrigation
systems for taro fields yielding up to 24 tons per acre, the highest crop yields
in all of Polynesia. Those yields in turn supported intensive pig production.
Hawaii was also unique within Polynesia in using mass labor for aquaculture,
by constructing large fishponds in which milkfish and mullet were grown.
AS A RESULT of all this environmentally related variation in subsistence,
human population densities (measured in people per square mile of arable
land) varied greatly over Polynesia. At the lower end were the hunter-
gatherers of the Chathams (only 5 people per square mile) and of New
Zealand’s South Island, and the farmers of the rest of New Zealand (28 people
per square mile). In contrast, many islands with intensive agriculture attained
population densities exceeding 120 per square mile. Tonga, Samoa, and the
Societies achieved 210–250 people per square mile and Hawaii 300. The
upper extreme of 1,100 people per square mile was reached on the high island
of Anuta, whose population converted essentially all the land to intensive
food production, thereby crammed 160 people into the island’s 100 acres, and
joined the ranks of the densest self-sufficient populations in the world.
Anuta’s population density exceeded that of modern Holland and even rivaled
that of Bangladesh.
Population size is the product of population density (people per square
mile) and area (square miles). The relevant area is not the area of an island
but that of a political unit, which could be either larger or smaller than a
single island. On the one hand, islands near one another might become
combined into a single political unit. On the other hand, single large rugged
islands were divided into many independent political units. Hence the area of
the political unit varied not only with an island’s area but also with its
fragmentation and isolation.
For small isolated islands without strong barriers to internal
communication, the entire island constituted the political unit—as in the case
of Anuta, with its 160 people. Many larger islands never did become unified
politically, whether because the population consisted of dispersed bands of
only a few dozen hunter-gatherers each (the Chathams and New Zealand’s
southern South Island), or of farmers scattered over large distances (the rest of
New Zealand), or of farmers living in dense populations but in rugged terrain
precluding political unification. For example, people in neighboring steep-
sided valleys of the Marquesas communicated with each other mainly by sea;
each valley formed an independent political entity of a few thousand
inhabitants, and most individual large Marquesan islands remained divided
into many such entities.
The terrains of the Tongan, Samoan, Society, and Hawaiian islands did
permit political unification within islands, yielding political units of 10,000
people or more (over 30,000 on the large Hawaiian islands). The distances
between islands of the Tongan archipelago, as well as the distances between
Tonga and neighboring archipelagoes, were sufficiently modest that a multi-
island empire encompassing 40,000 people was eventually established. Thus,
Polynesian political units ranged in size from a few dozen to 40,000 people.
A political unit’s population size interacted with its population density to
influence Polynesian technology and economic, social, and political
organization. In general, the larger the size and the higher the density, the
more complex and specialized were the technology and organization, for
reasons that we shall examine in detail in later chapters. Briefly, at high
population densities only a portion of the people came to be farmers, but they
were mobilized to devote themselves to intensive food production, thereby
yielding surpluses to feed nonproducers. The nonproducers mobilizing them
included chiefs, priests, bureaucrats, and warriors. The biggest political units
could assemble large labor forces to construct irrigation systems and
fishponds that intensified food production even further. These developments
were especially apparent on Tonga, Samoa, and the Societies, all of which
were fertile, densely populated, and moderately large by Polynesian
standards. The trends reached their zenith on the Hawaiian Archipelago,
consisting of the largest tropical Polynesian islands, where high population
densities and large land areas meant that very large labor forces were
potentially available to individual chiefs.
The variations among Polynesian societies associated with different
population densities and sizes were as follows. Economies remained simplest
on islands with low population densities (such as the hunter-gatherers of the
Chathams), low population numbers (small atolls), or both low densities and
low numbers. In those societies each household made what it needed; there
was little or no economic specialization. Specialization increased on larger,
more densely populated islands, reaching a peak on Samoa, the Societies, and
especially Tonga and Hawaii. The latter two islands supported hereditary part-
time craft specialists, including canoe builders, navigators, stone masons, bird
catchers, and tattooers.
Social complexity was similarly varied. Again, the Chathams and the
atolls had the simplest, most egalitarian societies. While those islands retained
the original Polynesian tradition of having chiefs, their chiefs wore little or no
visible signs of distinction, lived in ordinary huts like those of commoners,
and grew or caught their food like everyone else. Social distinctions and
chiefly powers increased on high-density islands with large political units,
being especially marked on Tonga and the Societies.
Social complexity again reached its peak in the Hawaiian Archipelago,
where people of chiefly descent were divided into eight hierarchically ranked
lineages. Members of those chiefly lineages did not intermarry with
commoners but only with each other, sometimes even with siblings or half-
siblings. Commoners had to prostrate themselves before high-ranking chiefs.
All the members of chiefly lineages, bureaucrats, and some craft specialists
were freed from the work of food production.
Political organization followed the same trends. On the Chathams and
atolls, the chiefs had few resources to command, decisions were reached by
general discussion, and landownership rested with the community as a whole
rather than with the chiefs. Larger, more densely populated political units
concentrated more authority with the chiefs. Political complexity was greatest
on Tonga and Hawaii, where the powers of hereditary chiefs approximated
those of kings elsewhere in the world, and where land was controlled by the
chiefs, not by the commoners. Using appointed bureaucrats as agents, chiefs
requisitioned food from the commoners and also conscripted them to work on
large construction projects, whose form varied from island to island: irrigation
projects and fishponds on Hawaii, dance and feast centers on the Marquesas,
chiefs’ tombs on Tonga, and temples on Hawaii, the Societies, and Easter.
At the time of Europeans’ arrival in the 18th century, the Tongan
chiefdom or state had already become an inter-archipelagal empire. Because
the Tongan Archipelago itself was geographically close-knit and included
several large islands with unfragmented terrain, each island became unified
under a single chief; then the hereditary chiefs of the largest Tongan island
(Tongatapu) united the whole archipelago, and eventually they conquered
islands outside the archipelago up to 500 miles distant. They engaged in
regular long-distance trade with Fiji and Samoa, established Tongan
settlements in Fiji, and began to raid and conquer parts of Fiji. The conquest
and administration of this maritime proto-empire were achieved by navies of
large canoes, each holding up to 150 men.
Like Tonga, Hawaii became a political entity encompassing several
populous islands, but one confined to a single archipelago because of its
extreme isolation. At the time of Hawaii’s “discovery” by Europeans in 1778,
political unification had already taken place within each Hawaiian island, and
some political fusion between islands had begun. The four largest islands—
Big Island (Hawaii in the narrow sense), Maui, Oahu, and Kauai—remained
independent, controlling (or jockeying with each other for control of) the
smaller islands (Lanai, Molokai, Kahoolawe, and Niihau). After the arrival of
Europeans, the Big Island’s King Kamehameha I rapidly proceeded with the
consolidation of the largest islands by purchasing European guns and ships to
invade and conquer first Maui and then Oahu. Kamehameha thereupon
prepared invasions of the last independent Hawaiian island, Kauai, whose
chief finally reached a negotiated settlement with him, completing the
archipelago’s unification.
The remaining type of variation among Polynesian societies to be
considered involves tools and other aspects of material culture. The differing
availability of raw materials imposed an obvious constraint on material
culture. At the one extreme was Henderson Island, an old coral reef raised
above sea level and devoid of stone other than limestone. Its inhabitants were
reduced to fabricating adzes out of giant clamshells. At the opposite extreme,
the Maori on the minicontinent of New Zealand had access to a wide range of
raw materials and became especially noted for their use of jade. Between
those two extremes fell Polynesia’s oceanic volcanic islands, which lacked
granite, flint, and other continental rocks but did at least have volcanic rocks,
which Polynesians worked into ground or polished stone adzes used to clear
land for farming.
As for the types of artifacts made, the Chatham Islanders required little
more than hand-held clubs and sticks to kill seals, birds, and lobsters. Most
other islanders produced a diverse array of fishhooks, adzes, jewelry, and
other objects. On the atolls, as on the Chathams, those artifacts were small,
relatively simple, and individually produced and owned, while architecture
consisted of nothing more than simple huts. Large and densely populated
islands supported craft specialists who produced a wide range of prestige
goods for chiefs—such as the feather capes reserved for Hawaiian chiefs and
made of tens of thousands of bird feathers.
The largest products of Polynesia were the immense stone structures of a
few islands—the famous giant statues of Easter Island, the tombs of Tongan
chiefs, the ceremonial platforms of the Marquesas, and the temples of Hawaii
and the Societies. This monumental Polynesian architecture was obviously
evolving in the same direction as the pyramids of Egypt, Mesopotamia,
Mexico, and Peru. Naturally, Polynesia’s structures are not on the scale of
those pyramids, but that merely reflects the fact that Egyptian pharaohs could
draw conscript labor from a much larger human population than could the
chief of any Polynesian island. Even so, the Easter Islanders managed to erect
30-ton stone statues—no mean feat for an island with only 7,000 people, who
had no power source other than their own muscles.
THUS, POLYNESIAN ISLAND societies differed greatly in their economic
specialization, social complexity, political organization, and material
products, related to differences in population size and density, related in turn
to differences in island area, fragmentation, and isolation and in opportunities
for subsistence and for intensifying food production. All those differences
among Polynesian societies developed, within a relatively short time and
modest fraction of the Earth’s surface, as environmentally related variations
on a single ancestral society. Those categories of cultural differences within
Polynesia are essentially the same categories that emerged everywhere else in
the world.
Of course, the range of variation over the rest of the globe is much greater
than that within Polynesia. While modern continental peoples included ones
dependent on stone tools, as were Polynesians, South America also spawned
societies expert in using precious metals, and Eurasians and Africans went on
to utilize iron. Those developments were precluded in Polynesia, because no
Polynesian island except New Zealand had significant metal deposits. Eurasia
had full-fledged empires before Polynesia was even settled, and South
America and Mesoamerica developed empires later, whereas Polynesia
produced just two proto-empires, one of which (Hawaii) coalesced only after
the arrival of Europeans. Eurasia and Mesoamerica developed indigenous
writing, which failed to emerge in Polynesia, except perhaps on Easter Island,
whose mysterious script may however have postdated the islanders’ contact
with Europeans.
That is, Polynesia offers us a small slice, not the full spectrum, of the
world’s human social diversity. That shouldn’t surprise us, since Polynesia
provides only a small slice of the world’s geographic diversity. In addition,
since Polynesia was colonized so late in human history, even the oldest
Polynesian societies had only 3,200 years in which to develop, as opposed to
at least 13,000 years for societies on even the last-colonized continents (the
Americas). Given a few more millennia, perhaps Tonga and Hawaii would
have reached the level of full-fledged empires battling each other for control
of the Pacific, with indigenously developed writing to administer those
empires, while New Zealand’s Maori might have added copper and iron tools
to their repertoire of jade and other materials.
In short, Polynesia furnishes us with a convincing example of
environmentally related diversification of human societies in operation. But
we thereby learn only that it can happen, because it happened in Polynesia.
Did it also happen on the continents? If so, what were the environmental
differences responsible for diversification on the continents, and what were
their consequences?
CHAPTER 3
COLLISION AT CAJAMARCA
THE BIGGEST POPULATION SHIFT OF MODERN TIMES HAS been the colonization of
the New World by Europeans, and the resulting conquest, numerical
reduction, or complete disappearance of most groups of Native Americans
(American Indians). As I explained in Chapter 1, the New World was initially
colonized around or before 11,000 B.C. by way of Alaska, the Bering Strait,
and Siberia. Complex agricultural societies gradually arose in the Americas
far to the south of that entry route, developing in complete isolation from the
emerging complex societies of the Old World. After that initial colonization
from Asia, the sole well-attested further contacts between the New World and
Asia involved only hunter-gatherers living on opposite sides of the Bering
Strait, plus an inferred transpacific voyage that introduced the sweet potato
from South America to Polynesia.
As for contacts of New World peoples with Europe, the sole early ones
involved the Norse who occupied Greenland in very small numbers between
A.D. 986 and about 1500. But those Norse visits had no discernible impact on
Native American societies. Instead, for practical purposes the collision of
advanced Old World and New World societies began abruptly in A.D. 1492,
with Christopher Columbus’s “discovery” of Caribbean islands densely
populated by Native Americans.
The most dramatic moment in subsequent European–Native American
relations was the first encounter between the Inca emperor Atahuallpa and the
Spanish conquistador Francisco Pizarro at the Peruvian highland town of
Cajamarca on November 16, 1532. Atahuallpa was absolute monarch of the
largest and most advanced state in the New World, while Pizarro represented
the Holy Roman Emperor Charles V (also known as King Charles I of Spain),
monarch of the most powerful state in Europe. Pizarro, leading a ragtag group
of 168 Spanish soldiers, was in unfamiliar terrain, ignorant of the local
inhabitants, completely out of touch with the nearest Spaniards (1,000 miles
to the north in Panama) and far beyond the reach of timely reinforcements.
Atahuallpa was in the middle of his own empire of millions of subjects and
immediately surrounded by his army of 80,000 soldiers, recently victorious in
a war with other Indians. Nevertheless, Pizarro captured Atahuallpa within a
few minutes after the two leaders first set eyes on each other. Pizarro
proceeded to hold his prisoner for eight months, while extracting history’s
largest ransom in return for a promise to free him. After the ransom—enough
gold to fill a room 22 feet long by 17 feet wide to a height of over 8 feet—
was delivered, Pizarro reneged on his promise and executed Atahuallpa.
Atahuallpa’s capture was decisive for the European conquest of the Inca
Empire. Although the Spaniards’ superior weapons would have assured an
ultimate Spanish victory in any case, the capture made the conquest quicker
and infinitely easier. Atahuallpa was revered by the Incas as a sun-god and
exercised absolute authority over his subjects, who obeyed even the orders he
issued from captivity. The months until his death gave Pizarro time to
dispatch exploring parties unmolested to other parts of the Inca Empire, and
to send for reinforcements from Panama. When fighting between Spaniards
and Incas finally did commence after Atahuallpa’s execution, the Spanish
forces were more formidable.
Thus, Atahuallpa’s capture interests us specifically as marking the
decisive moment in the greatest collision of modern history. But it is also of
more general interest, because the factors that resulted in Pizarro’s seizing
Atahuallpa were essentially the same ones that determined the outcome of
many similar collisions between colonizers and native peoples elsewhere in
the modern world. Hence Atahuallpa’s capture offers us a broad window onto
world history.
WHAT UNFOLDED THAT day at Cajamarca is well known, because it was
recorded in writing by many of the Spanish participants. To get a flavor of
those events, let us relive them by weaving together excerpts from eyewitness
accounts by six of Pizarro’s companions, including his brothers Hernando and
Pedro:
“The prudence, fortitude, military discipline, labors, perilous navigations,
and battles of the Spaniards—vassals of the most invincible Emperor of the
Roman Catholic Empire, our natural King and Lord—will cause joy to the
faithful and terror to the infidels. For this reason, and for the glory of God our
Lord and for the service of the Catholic Imperial Majesty, it has seemed good
to me to write this narrative, and to send it to Your Majesty, that all may have
a knowledge of what is here related. It will be to the glory of God, because
they have conquered and brought to our holy Catholic Faith so vast a number
of heathens, aided by His holy guidance. It will be to the honor of our
Emperor because, by reason of his great power and good fortune, such events
happened in his time. It will give joy to the faithful that such battles have
been won, such provinces discovered and conquered, such riches brought
home for the King and for themselves; and that such terror has been spread
among the infidels, such admiration excited in all mankind.
“For when, either in ancient or modern times, have such great exploits
been achieved by so few against so many, over so many climes, across so
many seas, over such distances by land, to subdue the unseen and unknown?
Whose deeds can be compared with those of Spain? Our Spaniards, being few
in number, never having more than 200 or 300 men together, and sometimes
only 100 and even fewer, have, in our times, conquered more territory than
has ever been known before, or than all the faithful and infidel princes
possess. I will only write, at present, of what befell in the conquest, and I will
not write much, in order to avoid prolixity.
“Governor Pizarro wished to obtain intelligence from some Indians who
had come from Cajamarca, so he had them tortured. They confessed that they
had heard that Atahuallpa was waiting for the Governor at Cajamarca. The
Governor then ordered us to advance. On reaching the entrance to Cajamarca,
we saw the camp of Atahuallpa at a distance of a league, in the skirts of the
mountains. The Indians’ camp looked like a very beautiful city. They had so
many tents that we were all filled with great apprehension. Until then, we had
never seen anything like this in the Indies. It filled all our Spaniards with fear
and confusion. But we could not show any fear or turn back, for if the Indians
had sensed any weakness in us, even the Indians that we were bringing with
us as guides would have killed us. So we made a show of good spirits, and
after carefully observing the town and the tents, we descended into the valley
and entered Cajamarca.
“We talked a lot among ourselves about what to do. All of us were full of
fear, because we were so few in number and we had penetrated so far into a
land where we could not hope to receive reinforcements. We all met with the
Governor to debate what we should undertake the next day. Few of us slept
that night, and we kept watch in the square of Cajamarca, looking at the
campfires of the Indian army. It was a frightening sight. Most of the campfires
were on a hillside and so close to each other that it looked like the sky
brightly studded with stars. There was no distinction that night between the
mighty and the lowly, or between foot soldiers and horsemen. Everyone
carried out sentry duty fully armed. So too did the good old Governor, who
went about encouraging his men. The Governor’s brother Hernando Pizarro
estimated the number of Indian soldiers there at 40,000, but he was telling a
lie just to encourage us, for there were actually more than 80,000 Indians.
“On the next morning a messenger from Atahuallpa arrived, and the
Governor said to him, ‘Tell your lord to come when and how he pleases, and
that, in what way soever he may come I will receive him as a friend and
brother. I pray that he may come quickly, for I desire to see him. No harm or
insult will befall him.’
“The Governor concealed his troops around the square at Cajamarca,
dividing the cavalry into two portions of which he gave the command of one
to his brother Hernando Pizarro and the command of the other to Hernando de
Soto. In like manner he divided the infantry, he himself taking one part and
giving the other to his brother Juan Pizarro. At the same time, he ordered
Pedro de Candia with two or three infantrymen to go with trumpets to a small
fort in the plaza and to station themselves there with a small piece of artillery.
When all the Indians, and Atahuallpa with them, had entered the Plaza, the
Governor would give a signal to Candia and his men, after which they should
start firing the gun, and the trumpets should sound, and at the sound of the
trumpets the cavalry should dash out of the large court where they were
waiting hidden in readiness.
“At noon Atahuallpa began to draw up his men and to approach. Soon we
saw the entire plain full of Indians, halting periodically to wait for more
Indians who kept filing out of the camp behind them. They kept filling out in
separate detachments into the afternoon. The front detachments were now
close to our camp, and still more troops kept issuing from the camp of the
Indians. In front of Atahuallpa went 2,000 Indians who swept the road ahead
of him, and these were followed by the warriors, half of whom were marching
in the fields on one side of him and half on the other side.
“First came a squadron of Indians dressed in clothes of different colors,
like a chessboard. They advanced, removing the straws from the ground and
sweeping the road. Next came three squadrons in different dresses, dancing
and singing. Then came a number of men with armor, large metal plates, and
crowns of gold and silver. So great was the amount of furniture of gold and
silver which they bore, that it was a marvel to observe how the sun glinted
upon it. Among them came the figure of Atahuallpa in a very fine litter with
the ends of its timbers covered in silver. Eighty lords carried him on their
shoulders, all wearing a very rich blue livery. Atahuallpa himself was very
richly dressed, with his crown on his head and a collar of large emeralds
around his neck. He sat on a small stool with a rich saddle cushion resting on
his litter. The litter was lined with parrot feathers of many colors and
decorated with plates of gold and silver.
“Behind Atahuallpa came two other litters and two hammocks, in which
were some high chiefs, then several squadrons of Indians with crowns of gold
and silver. These Indian squadrons began to enter the plaza to the
accompaniment of great songs, and thus entering they occupied every part of
the plaza. In the meantime all of us Spaniards were waiting ready, hidden in a
courtyard, full of fear. Many of us urinated without noticing it, out of sheer
terror. On reaching the center of the plaza, Atahuallpa remained in his litter on
high, while his troops continued to file in behind him.
“Governor Pizarro now sent Friar Vicente de Valverde to go speak to
Atahuallpa, and to require Atahuallpa in the name of God and of the King of
Spain that Atahuallpa subject himself to the law of our Lord Jesus Christ and
to the service of His Majesty the King of Spain. Advancing with a cross in
one hand and the Bible in the other hand, and going among the Indian troops
up to the place where Atahuallpa was, the Friar thus addressed him: ‘I am a
Priest of God, and I teach Christians the things of God, and in like manner I
come to teach you. What I teach is that which God says to us in this Book.
Therefore, on the part of God and of the Christians, I beseech you to be their
friend, for such is God’s will, and it will be for your good.’
“Atahuallpa asked for the Book, that he might look at it, and the Friar
gave it to him closed. Atahuallpa did not know how to open the Book, and the
Friar was extending his arm to do so, when Atahuallpa, in great anger, gave
him a blow on the arm, not wishing that it should be opened. Then he opened
it himself, and, without any astonishment at the letters and paper he threw it
away from him five or six paces, his face a deep crimson.
“The Friar returned to Pizarro, shouting, ‘Come out! Come out,
Christians! Come at these enemy dogs who reject the things of God. That
tyrant has thrown my book of holy law to the ground! Did you not see what
happened? Why remain polite and servile toward this over-proud dog when
the plains are full of Indians? March out against him, for I absolve you!’
“The governor then gave the signal to Candia, who began to fire off the
guns. At the same time the trumpets were sounded, and the armored Spanish
troops, both cavalry and infantry, sallied forth out of their hiding places
straight into the mass of unarmed Indians crowding the square, giving the
Spanish battle cry, ‘Santiago!’ We had placed rattles on the horses to terrify
the Indians. The booming of the guns, the blowing of the trumpets, and the
rattles on the horses threw the Indians into panicked confusion. The Spaniards
fell upon them and began to cut them to pieces. The Indians were so filled
with fear that they climbed on top of one another, formed mounds, and
suffocated each other. Since they were unarmed, they were attacked without
danger to any Christian. The cavalry rode them down, killing and wounding,
and following in pursuit. The infantry made so good an assault on those that
remained that in a short time most of them were put to the sword.
“The Governor himself took his sword and dagger, entered the thick of
the Indians with the Spaniards who were with him, and with great bravery
reached Atahuallpa’s litter. He fearlessly grabbed Atahuallpa’s left arm and
shouted ‘Santiago!,’ but he could not pull Atahuallpa out of his litter because
it was held up high. Although we killed the Indians who held the litter, others
at once took their places and held it aloft, and in this manner we spent a long
time in overcoming and killing Indians. Finally seven or eight Spaniards on
horseback spurred on their horses, rushed upon the litter from one side, and
with great effort they heaved it over on its side. In that way Atahuallpa was
captured, and the Governor took Atahuallpa to his lodging. The Indians
carrying the litter, and those escorting Atahuallpa, never abandoned him: all
died around him.
“The panic-stricken Indians remaining in the square, terrified at the firing
of the guns and at the horses—something they had never seen—tried to flee
from the square by knocking down a stretch of wall and running out onto the
plain outside. Our cavalry jumped the broken wall and charged into the plain,
shouting, ‘Chase those with the fancy clothes! Don’t let any escape! Spear
them!’ All of the other Indian soldiers whom Atahuallpa had brought were a
mile from Cajamarca ready for battle, but not one made a move, and during
all this not one Indian raised a weapon against a Spaniard. When the
squadrons of Indians who had remained in the plain outside the town saw the
other Indians fleeing and shouting, most of them too panicked and fled. It was
an astonishing sight, for the whole valley for 15 or 20 miles was completely
filled with Indians. Night had already fallen, and our cavalry were continuing
to spear Indians in the fields, when we heard a trumpet calling for us to
reassemble at camp.
“If night had not come on, few out of the more than 40,000 Indian troops
would have been left alive. Six or seven thousand Indians lay dead, and many
more had their arms cut off and other wounds. Atahuallpa himself admitted
that we had killed 7,000 of his men in that battle. The man killed in one of the
litters was his minister, the lord of Chincha, of whom he was very fond. All
those Indians who bore Atahuallpa’s litter appeared to be high chiefs and
councillors. They were all killed, as well as those Indians who were carried in
the other litters and hammocks. The lord of Cajamarca was also killed, and
others, but their numbers were so great that they could not be counted, for all
who came in attendance on Atahuallpa were great lords. It was extraordinary
to see so powerful a ruler captured in so short a time, when he had come with
such a mighty army. Truly, it was not accomplished by our own forces, for
there were so few of us. It was by the grace of God, which is great.
“Atahuallpa’s robes had been torn off when the Spaniards pulled him out
of his litter. The Governor ordered clothes to be brought to him, and when
Atahuallpa was dressed, the Governor ordered Atahuallpa to sit near him and
soothed his rage and agitation at finding himself so quickly fallen from his
high estate. The Governor said to Atahuallpa, ‘Do not take it as an insult that
you have been defeated and taken prisoner, for with the Christians who come
with me, though so few in number, I have conquered greater kingdoms than
yours, and have defeated other more powerful lords than you, imposing upon
them the dominion of the Emperor, whose vassal I am, and who is King of
Spain and of the universal world. We come to conquer this land by his
command, that all may come to a knowledge of God and of His Holy Catholic
Faith; and by reason of our good mission, God, the Creator of heaven and
earth and of all things in them, permits this, in order that you may know Him
and come out from the bestial and diabolical life that you lead. It is for this
reason that we, being so few in number, subjugate that vast host. When you
have seen the errors in which you live, you will understand the good that we
have done you by coming to your land by order of his Majesty the King of
Spain. Our Lord permitted that your pride should be brought low and that no
Indian should be able to offend a Christian.’”
LET US NOW trace the chain of causation in this extraordinary confrontation,
beginning with the immediate events. When Pizarro and Atahuallpa met at
Cajamarca, why did Pizarro capture Atahuallpa and kill so many of his
followers, instead of Atahuallpa’s vastly more numerous forces capturing and
killing Pizarro? After all, Pizarro had only 62 soldiers mounted on horses,
along with 106 foot soldiers, while Atahuallpa commanded an army of about
80,000. As for the antecedents of those events, how did Atahuallpa come to
be at Cajamarca at all? How did Pizarro come to be there to capture him,
instead of Atahuallpa’s coming to Spain to capture King Charles I? Why did
Atahuallpa walk into what seems to us, with the gift of hindsight, to have
been such a transparent trap? Did the factors acting in the encounter of
Atahuallpa and Pizarro also play a broader role in encounters between Old
World and New World peoples and between other peoples?
Why did Pizarro capture Atahuallpa? Pizarro’s military advantages lay in
the Spaniards’ steel swords and other weapons, steel armor, guns, and horses.
To those weapons, Atahuallpa’s troops, without animals on which to ride into
battle, could oppose only stone, bronze, or wooden clubs, maces, and hand
axes, plus slingshots and quilted armor. Such imbalances of equipment were
decisive in innumerable other confrontations of Europeans with Native
Americans and other peoples.
The sole Native Americans able to resist European conquest for many
centuries were those tribes that reduced the military disparity by acquiring
and mastering both horses and guns. To the average white American, the
word “Indian” conjures up an image of a mounted Plains Indian brandishing a
rifle, like the Sioux warriors who annihilated General George Custer’s U.S.
Army battalion at the famous battle of the Little Big Horn in 1876. We easily
forget that horses and rifles were originally unknown to Native Americans.
They were brought by Europeans and proceeded to transform the societies of
Indian tribes that acquired them. Thanks to their mastery of horses and rifles,
the Plains Indians of North America, the Araucanian Indians of southern
Chile, and the Pampas Indians of Argentina fought off invading whites longer
than did any other Native Americans, succumbing only to massive army
operations by white governments in the 1870s and 1880s.
Today, it is hard for us to grasp the enormous numerical odds against
which the Spaniards’ military equipment prevailed. At the battle of Cajamarca
recounted above, 168 Spaniards crushed a Native American army 500 times
more numerous, killing thousands of natives while not losing a single
Spaniard. Time and again, accounts of Pizarro’s subsequent battles with the
Incas, Cortés’s conquest of the Aztecs, and other early European campaigns
against Native Americans describe encounters in which a few dozen
European horsemen routed thousands of Indians with great slaughter. During
Pizarro’s march from Cajamarca to the Inca capital of Cuzco after
Atahuallpa’s death, there were four such battles: at Jauja, Vilcashuaman,
Vilcaconga, and Cuzco. Those four battles involved a mere 80, 30, 110, and
40 Spanish horsemen, respectively, in each case ranged against thousands or
tens of thousands of Indians.
These Spanish victories cannot be written off as due merely to the help of
Native American allies, to the psychological novelty of Spanish weapons and
horses, or (as is often claimed) to the Incas’ mistaking Spaniards for their
returning god Viracocha. The initial successes of both Pizarro and Cortés did
attract native allies. However, many of them would not have become allies if
they had not already been persuaded, by earlier devastating successes of
unassisted Spaniards, that resistance was futile and that they should side with
the likely winners. The novelty of horses, steel weapons, and guns
undoubtedly paralyzed the Incas at Cajamarca, but the battles after Cajamarca
were fought against determined resistance by Inca armies that had already
seen Spanish weapons and horses. Within half a dozen years of the initial
conquest, Incas mounted two desperate, large-scale, well-prepared rebellions
against the Spaniards. All those efforts failed because of the Spaniards’ far
superior armament.
By the 1700s, guns had replaced swords as the main weapon favoring
European invaders over Native Americans and other native peoples. For
example, in 1808 a British sailor named Charlie Savage equipped with
muskets and excellent aim arrived in the Fiji Islands. The aptly named Savage
proceeded single-handedly to upset Fiji’s balance of power. Among his many
exploits, he paddled his canoe up a river to the Fijian village of Kasavu,
halted less than a pistol shot’s length from the village fence, and fired away at
the undefended inhabitants. His victims were so numerous that surviving
villagers piled up the bodies to take shelter behind them, and the stream
beside the village was red with blood. Such examples of the power of guns
against native peoples lacking guns could be multiplied indefinitely.
In the Spanish conquest of the Incas, guns played only a minor role. The
guns of those times (so-called harquebuses) were difficult to load and fire,
and Pizarro had only a dozen of them. They did produce a big psychological
effect on those occasions when they managed to fire. Far more important
were the Spaniards’ steel swords, lances, and daggers, strong sharp weapons
that slaughtered thinly armored Indians. In contrast, Indian blunt clubs, while
capable of battering and wounding Spaniards and their horses, rarely
succeeded in killing them. The Spaniards’ steel or chain mail armor and,
above all, their steel helmets usually provided an effective defense against
club blows, while the Indians’ quilted armor offered no protection against
steel weapons.
The tremendous advantage that the Spaniards gained from their horses
leaps out of the eyewitness accounts. Horsemen could easily outride Indian
sentries before the sentries had time to warn Indian troops behind them, and
could ride down and kill Indians on foot. The shock of a horse’s charge, its
maneuverability, the speed of attack that it permitted, and the raised and
protected fighting platform that it provided left foot soldiers nearly helpless in
the open. Nor was the effect of horses due only to the terror that they inspired
in soldiers fighting against them for the first time. By the time of the great
Inca rebellion of 1536, the Incas had learned how best to defend themselves
against cavalry, by ambushing and annihilating Spanish horsemen in narrow
passes. But the Incas, like all other foot soldiers, were never able to defeat
cavalry in the open. When Quizo Yupanqui, the best general of the Inca
emperor Manco, who succeeded Atahuallpa, besieged the Spaniards in Lima
in 1536 and tried to storm the city, two squadrons of Spanish cavalry charged
a much larger Indian force on flat ground, killed Quizo and all of his
commanders in the first charge, and routed his army. A similar cavalry charge
of 26 horsemen routed the best troops of Emperor Manco himself, as he was
besieging the Spaniards in Cuzco.
The transformation of warfare by horses began with their domestication
around 4000 B.C., in the steppes north of the Black Sea. Horses permitted
people possessing them to cover far greater distances than was possible on
foot, to attack by surprise, and to flee before a superior defending force could
be gathered. Their role at Cajamarca thus exemplifies a military weapon that
remained potent for 6,000 years, until the early 20th century, and that was
eventually applied on all the continents. Not until the First World War did the
military dominance of cavalry finally end. When we consider the advantages
that Spaniards derived from horses, steel weapons, and armor against foot
soldiers without metal, it should no longer surprise us that Spaniards
consistently won battles against enormous odds.
How did Atahuallpa come to be at Cajamarca? Atahuallpa and his army
came to be at Cajamarca because they had just won decisive battles in a civil
war that left the Incas divided and vulnerable. Pizarro quickly appreciated
those divisions and exploited them. The reason for the civil war was that an
epidemic of smallpox, spreading overland among South American Indians
after its arrival with Spanish settlers in Panama and Colombia, had killed the
Inca emperor Huayna Capac and most of his court around 1526, and then
immediately killed his designated heir, Ninan Cuyuchi. Those deaths
precipitated a contest for the throne between Atahuallpa and his half brother
Huascar. If it had not been for the epidemic, the Spaniards would have faced a
united empire.
Atahuallpa’s presence at Cajamarca thus highlights one of the key factors
in world history: diseases transmitted to peoples lacking immunity by
invading peoples with considerable immunity. Smallpox, measles, influenza,
typhus, bubonic plague, and other infectious diseases endemic in Europe
played a decisive role in European conquests, by decimating many peoples on
other continents. For example, a smallpox epidemic devastated the Aztecs
after the failure of the first Spanish attack in 1520 and killed Cuitláhuac, the
Aztec emperor who briefly succeeded Montezuma. Throughout the Americas,
diseases introduced with Europeans spread from tribe to tribe far in advance
of the Europeans themselves, killing an estimated 95 percent of the pre-
Columbian Native American population. The most populous and highly
organized native societies of North America, the Mississippian chiefdoms,
disappeared in that way between 1492 and the late 1600s, even before
Europeans themselves made their first settlement on the Mississippi River. A
smallpox epidemic in 1713 was the biggest single step in the destruction of
South Africa’s native San people by European settlers. Soon after the British
settlement of Sydney in 1788, the first of the epidemics that decimated
Aboriginal Australians began. A well-documented example from Pacific
islands is the epidemic that swept over Fiji in 1806, brought by a few
European sailors who struggled ashore from the wreck of the ship Argo.
Similar epidemics marked the histories of Tonga, Hawaii, and other Pacific
islands.
I do not mean to imply, however, that the role of disease in history was
confined to paving the way for European expansion. Malaria, yellow fever,
and other diseases of tropical Africa, India, Southeast Asia, and New Guinea
furnished the most important obstacle to European colonization of those
tropical areas.
How did Pizarro come to be at Cajamarca? Why didn’t Atahuallpa
instead try to conquer Spain? Pizarro came to Cajamarca by means of
European maritime technology, which built the ships that took him across the
Atlantic from Spain to Panama, and then in the Pacific from Panama to Peru.
Lacking such technology, Atahuallpa did not expand overseas out of South
America.
In addition to the ships themselves, Pizarro’s presence depended on the
centralized political organization that enabled Spain to finance, build, staff,
and equip the ships. The Inca Empire also had a centralized political
organization, but that actually worked to its disadvantage, because Pizarro
seized the Inca chain of command intact by capturing Atahuallpa. Since the
Inca bureaucracy was so strongly identified with its godlike absolute
monarch, it disintegrated after Atahuallpa’s death. Maritime technology
coupled with political organization was similarly essential for European
expansions to other continents, as well as for expansions of many other
peoples.
A related factor bringing Spaniards to Peru was the existence of writing.
Spain possessed it, while the Inca Empire did not. Information could be
spread far more widely, more accurately, and in more detail by writing than it
could be transmitted by mouth. That information, coming back to Spain from
Columbus’s voyages and from Cortés’s conquest of Mexico, sent Spaniards
pouring into the New World. Letters and pamphlets supplied both the
motivation and the necessary detailed sailing directions. The first published
report of Pizarro’s exploits, by his companion Captain Cristóbal de Mena, was
printed in Seville in April 1534, a mere nine months after Atahuallpa’s
execution. It became a best-seller, was rapidly translated into other European
languages, and sent a further stream of Spanish colonists to tighten Pizarro’s
grip on Peru.
Why did Atahuallpa walk into the trap? In hindsight, we find it
astonishing that Atahuallpa marched into Pizarro’s obvious trap at Cajamarca.
The Spaniards who captured him were equally surprised at their success. The
consequences of literacy are prominent in the ultimate explanation.
The immediate explanation is that Atahuallpa had very little information
about the Spaniards, their military power, and their intent. He derived that
scant information by word of mouth, mainly from an envoy who had visited
Pizarro’s force for two days while it was en route inland from the coast. That
envoy saw the Spaniards at their most disorganized, told Atahuallpa that they
were not fighting men, and that he could tie them all up if given 200 Indians.
Understandably, it never occurred to Atahuallpa that the Spaniards were
formidable and would attack him without provocation.
In the New World the ability to write was confined to small elites among
some peoples of modern Mexico and neighboring areas far to the north of the
Inca Empire. Although the Spanish conquest of Panama, a mere 600 miles
from the Incas’ northern boundary, began already in 1510, no knowledge even
of the Spaniards’ existence appears to have reached the Incas until Pizarro’s
first landing on the Peruvian coast in 1527. Atahuallpa remained entirely
ignorant about Spain’s conquests of Central America’s most powerful and
populous Indian societies.
As surprising to us today as Atahuallpa’s behavior leading to his capture
is his behavior thereafter. He offered his famous ransom in the naive belief
that, once paid off, the Spaniards would release him and depart. He had no
way of understanding that Pizarro’s men formed the spearhead of a force bent
on permanent conquest, rather than an isolated raid.
Atahuallpa was not alone in these fatal miscalculations. Even after
Atahuallpa had been captured, Francisco Pizarro’s brother Hernando Pizarro
deceived Atahuallpa’s leading general, Chalcuchima, commanding a large
army, into delivering himself to the Spaniards. Chalcuchima’s miscalculation
marked a turning point in the collapse of Inca resistance, a moment almost as
significant as the capture of Atahuallpa himself. The Aztec emperor
Montezuma miscalculated even more grossly when he took Cortés for a
returning god and admitted him and his tiny army into the Aztec capital of
Tenochtitlán. The result was that Cortés captured Montezuma, then went on to
conquer Tenochtitlán and the Aztec Empire.
On a mundane level, the miscalculations by Atahuallpa, Chalcuchima,
Montezuma, and countless other Native American leaders deceived by
Europeans were due to the fact that no living inhabitants of the New World
had been to the Old World, so of course they could have had no specific
information about the Spaniards. Even so, we find it hard to avoid the
conclusion that Atahuallpa “should” have been more suspicious, if only his
society had experienced a broader range of human behavior. Pizarro too
arrived at Cajamarca with no information about the Incas other than what he
had learned by interrogating the Inca subjects he encountered in 1527 and
1531. However, while Pizarro himself happened to be illiterate, he belonged
to a literate tradition. From books, the Spaniards knew of many contemporary
civilizations remote from Europe, and about several thousand years of
European history. Pizarro explicitly modeled his ambush of Atahuallpa on the
successful strategy of Cortés.
In short, literacy made the Spaniards heirs to a huge body of knowledge
about human behavior and history. By contrast, not only did Atahuallpa have
no conception of the Spaniards themselves, and no personal experience of any
other invaders from overseas, but he also had not even heard (or read) of
similar threats to anyone else, anywhere else, anytime previously in history.
That gulf of experience encouraged Pizarro to set his trap and Atahuallpa to
walk into it.
THUS, PIZARRO’S CAPTURE of Atahuallpa illustrates the set of proximate
factors that resulted in Europeans’ colonizing the New World instead of
Native Americans’ colonizing Europe. Immediate reasons for Pizarro’s
success included military technology based on guns, steel weapons, and
horses; infectious diseases endemic in Eurasia; European maritime
technology; the centralized political organization of European states; and
writing. The title of this book will serve as shorthand for those proximate
factors, which also enabled modern Europeans to conquer peoples of other
continents. Long before anyone began manufacturing guns and steel, others of
those same factors had led to the expansions of some non-European peoples,
as we shall see in later chapters.
But we are still left with the fundamental question why all those
immediate advantages came to lie more with Europe than with the New
World. Why weren’t the Incas the ones to invent guns and steel swords, to be
mounted on animals as fearsome as horses, to bear diseases to which
Europeans lacked resistance, to develop oceangoing ships and advanced
political organization, and to be able to draw on the experience of thousands
of years of written history? Those are no longer the questions of proximate
causation that this chapter has been discussing, but questions of ultimate
causation that will take up the next two parts of this book.
PART TWO
THE RISE AND SPREAD OF FOOD
PRODUCTION
CHAPTER 4
FARMER POWER
AS A TEENAGER, I SPENT THE SUMMER OF 1956 IN MONTANA, working for an
elderly farmer named Fred Hirschy. Born in Switzerland, Fred had come to
southwestern Montana as a teenager in the 1890s and proceeded to develop
one of the first farms in the area. At the time of his arrival, much of the
original Native American population of hunter-gatherers was still living there.
My fellow farmhands were, for the most part, tough whites whose normal
speech featured strings of curses, and who spent their weekdays working so
that they could devote their weekends to squandering their week’s wages in
the local saloon. Among the farmhands, though, was a member of the
Blackfoot Indian tribe named Levi, who behaved very differently from the
coarse miners—being polite, gentle, responsible, sober, and well spoken. He
was the first Indian with whom I had spent much time, and I came to admire
him.
It was therefore a shocking disappointment to me when, one Sunday
morning, Levi too staggered in drunk and cursing after a Saturday-night
binge. Among his curses, one has stood out in my memory: “Damn you, Fred
Hirschy, and damn the ship that brought you from Switzerland!” It poignantly
brought home to me the Indians’ perspective on what I, like other white
schoolchildren, had been taught to view as the heroic conquest of the
American West. Fred Hirschy’s family was proud of him, as a pioneer farmer
who had succeeded under difficult conditions. But Levi’s tribe of hunters and
famous warriors had been robbed of its lands by the immigrant white farmers.
How did the farmers win out over the famous warriors?
For most of the time since the ancestors of modern humans diverged from
the ancestors of the living great apes, around 7 million years ago, all humans
on Earth fed themselves exclusively by hunting wild animals and gathering
wild plants, as the Blackfeet still did in the 19th century. It was only within
the last 11,000 years that some peoples turned to what is termed food
production: that is, domesticating wild animals and plants and eating the
resulting livestock and crops. Today, most people on Earth consume food that
they produced themselves or that someone else produced for them. At current
rates of change, within the next decade the few remaining bands of hunter-
gatherers will abandon their ways, disintegrate, or die out, thereby ending our
millions of years of commitment to the hunter-gatherer lifestyle.
Different peoples acquired food production at different times in
prehistory. Some, such as Aboriginal Australians, never acquired it at all. Of
those who did, some (for example, the ancient Chinese) developed it
independently by themselves, while others (including ancient Egyptians)
acquired it from neighbors. But, as we’ll see, food production was indirectly a
prerequisite for the development of guns, germs, and steel. Hence geographic
variation in whether, or when, the peoples of different continents became
farmers and herders explains to a large extent their subsequent contrasting
fates. Before we devote the next six chapters to understanding how
geographic differences in food production arose, this chapter will trace the
main connections through which food production led to all the advantages
that enabled Pizarro to capture Atahuallpa, and Fred Hirschy’s people to
dispossess Levi’s (Figure 4.1).
The first connection is the most direct one: availability of more
consumable calories means more people. Among wild plant and animal
species, only a small minority are edible to humans or worth hunting or
gathering. Most species are useless to us as food, for one or more of the
following reasons: they are indigestible (like bark), poisonous (monarch
butterflies and death-cap mushrooms), low in nutritional value (jellyfish),
tedious to prepare (very small nuts), difficult to gather (larvae of most
insects), or dangerous to hunt (rhinoceroses). Most biomass (living biological
matter) on land is in the form of wood and leaves, most of which we cannot
digest.
By selecting and growing those few species of plants and animals that we
can eat, so that they constitute 90 percent rather than 0.1 percent of the
biomass on an acre of land, we obtain far more edible calories per acre. As a
result, one acre can feed many more herders and farmers—typically, 10 to
100 times more—than hunter-gatherers. That strength of brute numbers was
the first of many military advantages that food-producing tribes gained over
hunter-gatherer tribes.
In human societies possessing domestic animals, livestock fed more
people in four distinct ways: by furnishing meat, milk, and fertilizer and by
pulling plows. First and most directly, domestic animals became the societies’
major source of animal protein, replacing wild game. Today, for instance,
Americans tend to get most of their animal protein from cows, pigs, sheep,
and chickens, with game such as venison just a rare delicacy. In addition,
some big domestic mammals served as sources of milk and of milk products
such as butter, cheese, and yogurt. Milked mammals include the cow, sheep,
goat, horse, reindeer, water buffalo, yak, and Arabian and Bactrian camels.
Those mammals thereby yield several times more calories over their lifetime
than if they were just slaughtered and consumed as meat.
Big domestic mammals also interacted with domestic plants in two ways
to increase crop production. First, as any modern gardener or farmer still
knows by experience, crop yields can be greatly increased by manure applied
as fertilizer. Even with the modern availability of synthetic fertilizers
produced by chemical factories, the major source of crop fertilizer today in
most societies is still animal manure—especially of cows, but also of yaks
and sheep. Manure has been valuable, too, as a source of fuel for fires in
traditional societies.
In addition, the largest domestic mammals interacted with domestic plants
to increase food production by pulling plows and thereby making it possible
for people to till land that had previously been uneconomical for farming.
Those plow animals were the cow, horse, water buffalo, Bali cattle, and yak /
cow hybrids. Here is one example of their value: the first prehistoric farmers
of central Europe, the so-called Linearbandkeramik culture that arose slightly
before 5000 B.C., were initially confined to soils light enough to be tilled by
means of hand-held digging sticks. Only over a thousand years later, with the
introduction of the ox-drawn plow, were those farmers able to extend
cultivation to a much wider range of heavy soils and tough sods. Similarly,
Native American farmers of the North American Great Plains grew crops in
the river valleys, but farming of the tough sods on the extensive uplands had
to await 19th-century Europeans and their animal-drawn plows.
All those are direct ways in which plant and animal domestication led to
denser human populations by yielding more food than did the hunter-gatherer
lifestyle. A more indirect way involved the consequences of the sedentary
lifestyle enforced by food production. People of many hunter-gatherer
societies move frequently in search of wild foods, but farmers must remain
near their fields and orchards. The resulting fixed abode contributes to denser
human populations by permitting a shortened birth interval. A hunter-gatherer
mother who is shifting camp can carry only one child, along with her few
possessions. She cannot afford to bear her next child until the previous toddler
can walk fast enough to keep up with the tribe and not hold it back. In
practice, nomadic hunter-gatherers space their children about four years apart
by means of lactational amenorrhea, sexual abstinence, infanticide, and
abortion. By contrast, sedentary people, unconstrained by problems of
carrying young children on treks, can bear and raise as many children as they
can feed. The birth interval for many farm peoples is around two years, half
that of hunter-gatherers. That higher birthrate of food producers, together with
their ability to feed more people per acre, lets them achieve much higher
population densities than hunter-gatherers.
A separate consequence of a settled existence is that it permits one to
store food surpluses, since storage would be pointless if one didn’t remain
nearby to guard the stored food. While some nomadic hunter-gatherers may
occasionally bag more food than they can consume in a few days, such a
bonanza is of little use to them because they cannot protect it. But stored food
is essential for feeding non-food-producing specialists, and certainly for
supporting whole towns of them. Hence nomadic hunter-gatherer societies
have few or no such full-time specialists, who instead first appear in sedentary
societies.
Two types of such specialists are kings and bureaucrats. Hunter-gatherer
societies tend to be relatively egalitarian, to lack full-time bureaucrats and
hereditary chiefs, and to have small-scale political organization at the level of
the band or tribe. That’s because all able-bodied hunter-gatherers are obliged
to devote much of their time to acquiring food. In contrast, once food can be
stockpiled, a political elite can gain control of food produced by others, assert
the right of taxation, escape the need to feed itself, and engage full-time in
political activities. Hence moderate-sized agricultural societies are often
organized in chiefdoms, and kingdoms are confined to large agricultural
societies. Those complex political units are much better able to mount a
sustained war of conquest than is an egalitarian band of hunters. Some hunter-
gatherers in especially rich environments, such as the Pacific Northwest coast
of North America and the coast of Ecuador, also developed sedentary
societies, food storage, and nascent chiefdoms, but they did not go farther on
the road to kingdoms.
A stored food surplus built up by taxation can support other full-time
specialists besides kings and bureaucrats. Of most direct relevance to wars of
conquest, it can be used to feed professional soldiers. That was the decisive
factor in the British Empire’s eventual defeat of New Zealand’s well-armed
indigenous Maori population. While the Maori achieved some stunning
temporary victories, they could not maintain an army constantly in the field
and were in the end worn down by 18,000 full-time British troops. Stored
food can also feed priests, who provide religious justification for wars of
conquest; artisans such as metalworkers, who develop swords, guns, and other
technologies; and scribes, who preserve far more information than can be
remembered accurately.
So far, I’ve emphasized direct and indirect values of crops and livestock
as food. However, they have other uses, such as keeping us warm and
providing us with valuable materials. Crops and livestock yield natural fibers
for making clothing, blankets, nets, and rope. Most of the major centers of
plant domestication evolved not only food crops but also fiber crops—notably
cotton, flax (the source of linen), and hemp. Several domestic animals yielded
animal fibers—especially wool from sheep, goats, llamas, and alpacas, and
silk from silkworms. Bones of domestic animals were important raw materials
for artifacts of Neolithic peoples before the development of metallurgy. Cow
hides were used to make leather. One of the earliest cultivated plants in many
parts of the Americas was grown for nonfood purposes: the bottle gourd, used
as a container.
Big domestic mammals further revolutionized human society by
becoming our main means of land transport until the development of railroads
in the 19th century. Before animal domestication, the sole means of
transporting goods and people by land was on the backs of humans. Large
mammals changed that: for the first time in human history, it became possible
to move heavy goods in large quantities, as well as people, rapidly overland
for long distances. The domestic animals that were ridden were the horse,
donkey, yak, reindeer, and Arabian and Bactrian camels. Animals of those
same five species, as well as the llama, were used to bear packs. Cows and
horses were hitched to wagons, while reindeer and dogs pulled sleds in the
Arctic. The horse became the chief means of long-distance transport over
most of Eurasia. The three domestic camel species (Arabian camel, Bactrian
camel, and llama) played a similar role in areas of North Africa and Arabia,
Central Asia, and the Andes, respectively.
The most direct contribution of plant and animal domestication to wars of
conquest was from Eurasia’s horses, whose military role made them the jeeps
and Sherman tanks of ancient warfare on that continent. As I mentioned in
Chapter 3, they enabled Cortés and Pizarro, leading only small bands of
adventurers, to overthrow the Aztec and Inca Empires. Even much earlier
(around 4000 B.C.), at a time when horses were still ridden bareback, they may
have been the essential military ingredient behind the westward expansion of
speakers of Indo-European languages from the Ukraine. Those languages
eventually replaced all earlier western European languages except Basque.
When horses later were yoked to wagons and other vehicles, horse-drawn
battle chariots (invented around 1800 B.C.) proceeded to revolutionize warfare
in the Near East, the Mediterranean region, and China. For example, in 1674
B.C., horses even enabled a foreign people, the Hyksos, to conquer then
horseless Egypt and to establish themselves temporarily as pharaohs.
Still later, after the invention of saddles and stirrups, horses allowed the
Huns and successive waves of other peoples from the Asian steppes to
terrorize the Roman Empire and its successor states, culminating in the
Mongol conquests of much of Asia and Russia in the 13th and 14th centuries
A.D. Only with the introduction of trucks and tanks in World War I did horses
finally become supplanted as the main assault vehicle and means of fast
transport in war. Arabian and Bactrian camels played a similar military role
within their geographic range. In all these examples, peoples with domestic
horses (or camels), or with improved means of using them, enjoyed an
enormous military advantage over those without them.
Of equal importance in wars of conquest were the germs that evolved in
human societies with domestic animals. Infectious diseases like smallpox,
measles, and flu arose as specialized germs of humans, derived by mutations
of very similar ancestral germs that had infected animals (Chapter 11). The
humans who domesticated animals were the first to fall victim to the newly
evolved germs, but those humans then evolved substantial resistance to the
new diseases. When such partly immune people came into contact with others
who had had no previous exposure to the germs, epidemics resulted in which
up to 99 percent of the previously unexposed population was killed. Germs
thus acquired ultimately from domestic animals played decisive roles in the
European conquests of Native Americans, Australians, South Africans, and
Pacific islanders.
In short, plant and animal domestication meant much more food and
hence much denser human populations. The resulting food surpluses, and (in
some areas) the animal-based means of transporting those surpluses, were a
prerequisite for the development of settled, politically centralized, socially
stratified, economically complex, technologically innovative societies. Hence
the availability of domestic plants and animals ultimately explains why
empires, literacy, and steel weapons developed earliest in Eurasia and later, or
not at all, on other continents. The military uses of horses and camels, and the
killing power of animal-derived germs, complete the list of major links
between food production and conquest that we shall be exploring.
CHAPTER 5
HISTORY’S HAVES AND HAVE-NOTS
MUCH OF HUMAN HISTORY HAS CONSISTED OF UNEQUAL conflicts between the
haves and the have-nots: between peoples with farmer power and those
without it, or between those who acquired it at different times. It should come
as no surprise that food production never arose in large areas of the globe, for
ecological reasons that still make it difficult or impossible there today. For
instance, neither farming nor herding developed in prehistoric times in North
America’s Arctic, while the sole element of food production to arise in
Eurasia’s Arctic was reindeer herding. Nor could food production spring up
spontaneously in deserts remote from sources of water for irrigation, such as
central Australia and parts of the western United States.
Instead, what cries out for explanation is the failure of food production to
appear, until modern times, in some ecologically very suitable areas that are
among the world’s richest centers of agriculture and herding today. Foremost
among these puzzling areas, where indigenous peoples were still hunter-
gatherers when European colonists arrived, were California and the other
Pacific states of the United States, the Argentine pampas, southwestern and
southeastern Australia, and much of the Cape region of South Africa. Had we
surveyed the world in 4000 B.C., thousands of years after the rise of food
production in its oldest sites of origin, we would have been surprised too at
several other modern breadbaskets that were still then without it—including
all the rest of the United States, England and much of France, Indonesia, and
all of subequatorial Africa. When we trace food production back to its
beginnings, the earliest sites provide another surprise. Far from being modern
breadbaskets, they include areas ranking today as somewhat dry or
ecologically degraded: Iraq and Iran, Mexico, the Andes, parts of China, and
Africa’s Sahel zone. Why did food production develop first in these
seemingly rather marginal lands, and only later in today’s most fertile
farmlands and pastures?
Geographic differences in the means by which food production arose are
also puzzling. In a few places it developed independently, as a result of local
people domesticating local plants and animals. In most other places it was
instead imported, in the form of crops and livestock that had been
domesticated elsewhere. Since those areas of nonindependent origins were
suitable for prehistoric food production as soon as domesticates had arrived,
why did the peoples of those areas not become farmers and herders without
outside assistance, by domesticating local plants and animals?
Among those regions where food production did spring up independently,
why did the times at which it appeared vary so greatly—for example,
thousands of years earlier in eastern Asia than in the eastern United States and
never in eastern Australia? Among those regions into which it was imported
in the prehistoric era, why did the date of arrival also vary so greatly—for
example, thousands of years earlier in southwestern Europe than in the
southwestern United States? Again among those regions where it was
imported, why in some areas (such as the southwestern United States) did
local hunter-gatherers themselves adopt crops and livestock from neighbors
and survive as farmers, while in other areas (such as Indonesia and much of
subequatorial Africa) the importation of food production involved a
cataclysmic replacement of the region’s original hunter-gatherers by invading
food producers? All these questions involve developments that determined
which peoples became history’s have-nots, and which became its haves.
BEFORE WE CAN hope to answer these questions, we need to figure out how
to identify areas where food production originated, when it arose there, and
where and when a given crop or animal was first domesticated. The most
unequivocal evidence comes from identification of plant and animal remains
at archaeological sites. Most domesticated plant and animal species differ
morphologically from their wild ancestors: for example, in the smaller size of
domestic cattle and sheep, the larger size of domestic chickens and apples, the
thinner and smoother seed coats of domestic peas, and the corkscrew-twisted
rather than scimitar-shaped horns of domestic goats. Hence remains of
domesticated plants and animals at a dated archaeological site can be
recognized and provide strong evidence of food production at that place and
time, whereas finding the remains only of wild species at a site fails to
provide evidence of food production and is compatible with hunting-
gathering. Naturally, food producers, especially early ones, continued to
gather some wild plants and hunt wild animals, so the food remains at their
sites often include wild species as well as domesticated ones.
Archaeologists date food production by radiocarbon dating of carbon-
containing materials at the site. This method is based on the slow decay of
radioactive carbon 14, a very minor component of carbon, the ubiquitous
building block of life, into the nonradioactive isotope nitrogen 14. Carbon 14
is continually being generated in the atmosphere by cosmic rays. Plants take
up atmospheric carbon, which has a known and approximately constant ratio
of carbon 14 to the prevalent isotope carbon 12 (a ratio of about one to a
million). That plant carbon goes on to form the body of the herbivorous
animals that eat the plants, and of the carnivorous animals that eat those
herbivorous animals. Once the plant or animal dies, though, half of its carbon
14 content decays into nitrogen 14 every 5,700 years, until after about 40,000
years the carbon 14 content is very low and difficult to measure or to
distinguish from contamination with small amounts of modern materials
containing carbon 14. Hence the age of material from an archaeological site
can be calculated from the material’s carbon 14 / carbon 12 ratio.
Radiocarbon is plagued by numerous technical problems, of which two
deserve mention here. One is that radiocarbon dating until the 1980s required
relatively large amounts of carbon (a few grams), much more than the amount
in small seeds or bones. Hence scientists instead often had to resort to dating
material recovered nearby at the same site and believed to be “associated
with” the food remains—that is, to have been deposited simultaneously by the
people who left the food. A typical choice of “associated” material is charcoal
from fires.
But archaeological sites are not always neatly sealed time capsules of
materials all deposited on the same day. Materials deposited at different times
can get mixed together, as worms and rodents and other agents churn up the
ground. Charcoal residues from a fire can thereby end up close to the remains
of a plant or animal that died and was eaten thousands of years earlier or later.
Increasingly today, archaeologists are circumventing this problem by a new
technique termed accelerator mass spectrometry, which permits radiocarbon
dating of tiny samples and thus lets one directly date a single small seed,
small bone, or other food residue. In some cases big differences have been
found between recent radiocarbon dates based on the direct new methods
(which have their own problems) and those based on the indirect older ones.
Among the resulting controversies remaining unresolved, perhaps the most
important for the purposes of this book concerns the date when food
production originated in the Americas: indirect methods of the 1960s and
1970s yielded dates as early as 7000 B.C., but more recent direct dating has
been yielding dates no earlier than 3500 B.C.
A second problem in radiocarbon dating is that the carbon 14 / carbon 12
ratio of the atmosphere is in fact not rigidly constant but fluctuates slightly
with time, so calculations of radiocarbon dates based on the assumption of a
constant ratio are subject to small systematic errors. The magnitude of this
error for each past date can in principle be determined with the help of long-
lived trees laying down annual growth rings, since the rings can be counted
up to obtain an absolute calendar date in the past for each ring, and a carbon
sample of wood dated in this manner can then be analyzed for its carbon 14 /
carbon 12 ratio. In this way, measured radiocarbon dates can be “calibrated”
to take account of fluctuations in the atmospheric carbon ratio. The effect of
this correction is that, for materials with apparent (that is, uncalibrated) dates
between about 1000 and 6000 B.C., the true (calibrated) date is between a few
centuries and a thousand years earlier. Somewhat older samples have more
recently begun to be calibrated by an alternative method based on another
radioactive decay process and yielding the conclusion that samples apparently
dating to about 9000 B.C. actually date to around 11,000 B.C.
Archaeologists often distinguish calibrated from uncalibrated dates by
writing the former in upper-case letters and the latter in lower-case letters (for
example, 3000 B.C. vs. 3000 B.C., respectively). However, the archaeological
literature can be confusing in this respect, because many books and papers
report un calibrated dates as B.C. and fail to mention that they are actually
uncalibrated. The dates that I report in this book for events within the last
15,000 years are calibrated dates. That accounts for some of the discrepancies
that readers may note between this book’s dates and those quoted in some
standard reference books on early food production.
Once one has recognized and dated ancient remains of domestic plants or
animals, how does one decide whether the plant or animal was actually
domesticated in the vicinity of that site itself, rather than domesticated
elsewhere and then spread to the site? One method is to examine a map of the
geographic distribution of the crop’s or animal’s wild ancestor, and to reason
that domestication must have taken place in the area where the wild ancestor
occurs. For example, chickpeas are widely grown by traditional farmers from
the Mediterranean and Ethiopia east to India, with the latter country
accounting for 80 percent of the world’s chickpea production today. One
might therefore have been deceived into supposing that chickpeas were
domesticated in India. But it turns out that ancestral wild chickpeas occur
only in southeastern Turkey. The interpretation that chickpeas were actually
domesticated there is supported by the fact that the oldest finds of possibly
domesticated chickpeas in Neolithic archaeological sites come from
southeastern Turkey and nearby northern Syria that date to around 8000 B.C.;
not until over 5,000 years later does archaeological evidence of chickpeas
appear on the Indian subcontinent.
A second method for identifying a crop’s or animal’s site of domestication
is to plot on a map the dates of the domesticated form’s first appearance at
each locality. The site where it appeared earliest may be its site of initial
domestication—especially if the wild ancestor also occurred there, and if the
dates of first appearance at other sites become progressively later with
increasing distance from the putative site of initial domestication, suggesting
spread to those other sites. For instance, the earliest known cultivated emmer
wheat comes from the Fertile Crescent around 8500 B.C. Soon thereafter, the
crop appears progressively farther west, reaching Greece around 6500 B.C. and
Germany around 5000 B.C. Those dates suggest domestication of emmer
wheat in the Fertile Crescent, a conclusion supported by the fact that ancestral
wild emmer wheat is confined to the area extending from Israel to western
Iran and Turkey.
However, as we shall see, complications arise in many cases where the
same plant or animal was domesticated independently at several different
sites. Such cases can often be detected by analyzing the resulting
morphological, genetic, or chromosomal differences between specimens of
the same crop or domestic animal in different areas. For instance, India’s zebu
breeds of domestic cattle possess humps lacking in western Eurasian cattle
breeds, and genetic analyses show that the ancestors of modern Indian and
western Eurasian cattle breeds diverged from each other hundreds of
thousands of years ago, long before any animals were domesticated
anywhere. That is, cattle were domesticated independently in India and
western Eurasia, within the last 10,000 years, starting with wild Indian and
western Eurasian cattle subspecies that had diverged hundreds of thousands of
years earlier.
LET’S NOW RETURN to our earlier questions about the rise of food production.
Where, when, and how did food production develop in different parts of the
globe?
At one extreme are areas in which food production arose altogether
independently, with the domestication of many indigenous crops (and, in
some cases, animals) before the arrival of any crops or animals from other
areas. There are only five such areas for which the evidence is at present
detailed and compelling: Southwest Asia, also known as the Near East or
Fertile Crescent; China; Mesoamerica (the term applied to central and
southern Mexico and adjacent areas of Central America); the Andes of South
America, and possibly the adjacent Amazon Basin as well; and the eastern
United States (Figure 5.1). Some or all of these centers may actually comprise
several nearby centers where food production arose more or less
independently, such as North China’s Yellow River valley and South China’s
Yangtze River valley.
In addition to these five areas where food production definitely arose de
novo, four others—Africa’s Sahel zone, tropical West Africa, Ethiopia, and
New Guinea—are candidates for that distinction. However, there is some
uncertainty in each case. Although indigenous wild plants were undoubtedly
domesticated in Africa’s Sahel zone just south of the Sahara, cattle herding
may have preceded agriculture there, and it is not yet certain whether those
were independently domesticated Sahel cattle or, instead, domestic cattle of
Fertile Crescent origin whose arrival triggered local plant domestication. It
remains similarly uncertain whether the arrival of those Sahel crops then
triggered the undoubted local domestication of indigenous wild plants in
tropical West Africa, and whether the arrival of Southwest Asian crops is
what triggered the local domestication of indigenous wild plants in Ethiopia.
As for New Guinea, archaeological studies there have provided evidence of
early agriculture well before food production in any adjacent areas, but the
crops grown have not been definitely identified.
Table 5.1 summarizes, for these and other areas of local domestication,
some of the best-known crops and animals and the earliest known dates of
domestication. Among these nine candidate areas for the independent
evolution of food production, Southwest Asia has the earliest definite dates
for both plant domestication (around 8500 B.C.) and animal domestication
(around 8000 B.C.); it also has by far the largest number of accurate
radiocarbon dates for early food production. Dates for China are nearly as
early, while dates for the eastern United States are clearly about 6,000 years
later. For the other six candidate areas, the earliest well-established dates do
not rival those for Southwest Asia, but too few early sites have been securely
dated in those six other areas for us to be certain that they really lagged
behind Southwest Asia and (if so) by how much.
The next group of areas consists of ones that did domesticate at least a
couple of local plants or animals, but where food production depended mainly
on crops and animals that were domesticated elsewhere. Those imported
domesticates may be thought of as “founder” crops and animals, because they
founded local food production. The arrival of founder domesticates enabled
local people to become sedentary, and thereby increased the likelihood of
local crops’ evolving from wild plants that were gathered, brought home and
planted accidentally, and later planted intentionally.
TABLE 5.1 Examples of Species Domesticated in Each Area
Area
Domesticated
Earliest Attested Date of
Domestication
Plants
Animals
Independent Origins of Domestication
1. Southwest Asia
wheat, pea, olive
sheep, goat
8500 B.C.
2. China
rice, millet
pig, silkworm
by 7500 B.C.
3. Mesoamerica
corn, beans, squash
turkey
by 3500 B.C.
4. Andes and
llama, guinea
potato, manioc
by 3500 B.C.
Amazonia
pig
5. Eastern United
sunflower, goosefoot none
2500 B.C.
States
? 6. Sahel
sorghum, African rice guinea fowl
by 5000 B.C.
? 7. Tropical West
African yams, oil
none
by 3000 B.C.
Africa
palm
? 8. Ethiopia
coffee, teff
none
?
? 9. New Guinea
sugar cane, banana
none
7000 B.C.?
Local Domestication Following Arrival of Founder Crops from Elsewhere
10. Western Europe
poppy, oat
none
6000–3500 B.C.
11. Indus Valley
sesame, eggplant
humped cattle
7000 B.C.
12. Egypt
sycamore fig, chufa
donkey, cat
6000 B.C.
In three or four such areas, the arriving founder package came from
Southwest Asia. One of them is western and central Europe, where food
production arose with the arrival of Southwest Asian crops and animals
between 6000 and 3500 B.C., but at least one plant (the poppy, and probably
oats and some others) was then domesticated locally. Wild poppies are
confined to coastal areas of the western Mediterranean. Poppy seeds are
absent from excavated sites of the earliest farming communities in eastern
Europe and Southwest Asia; they first appear in early farming sites in western
Europe. In contrast, the wild ancestors of most Southwest Asian crops and
animals were absent from western Europe. Thus, it seems clear that food
production did not evolve independently in western Europe. Instead, it was
triggered there by the arrival of Southwest Asian domesticates. The resulting
western European farming societies domesticated the poppy, which
subsequently spread eastward as a crop.
Another area where local domestication appears to have followed the
arrival of Southwest Asian founder crops is the Indus Valley region of the
Indian subcontinent. The earliest farming communities there in the seventh
millennium B.C. utilized wheat, barley, and other crops that had been
previously domesticated in the Fertile Crescent and that evidently spread to
the Indus Valley through Iran. Only later did domesticates derived from
indigenous species of the Indian subcontinent, such as humped cattle and
sesame, appear in Indus Valley farming communities. In Egypt as well, food
production began in the sixth millennium B.C. with the arrival of Southwest
Asian crops. Egyptians then domesticated the sycamore fig and a local
vegetable called chufa.
The same pattern perhaps applies to Ethiopia, where wheat, barley, and
other Southwest Asian crops have been cultivated for a long time. Ethiopians
also domesticated many locally available wild species to obtain crops most of
which are still confined to Ethiopia, but one of them (the coffee bean) has
now spread around the world. However, it is not yet known whether
Ethiopians were cultivating these local plants before or only after the arrival
of the Southwest Asian package.
In these and other areas where food production depended on the arrival of
founder crops from elsewhere, did local hunter-gatherers themselves adopt
those founder crops from neighboring farming peoples and thereby become
farmers themselves? Or was the founder package instead brought by invading
farmers, who were thereby enabled to outbreed the local hunters and to kill,
displace, or outnumber them?
In Egypt it seems likely that the former happened: local hunter-gatherers
simply added Southwest Asian domesticates and farming and herding
techniques to their own diet of wild plants and animals, then gradually phased
out the wild foods. That is, what arrived to launch food production in Egypt
was foreign crops and animals, not foreign peoples. The same may have been
true on the Atlantic coast of Europe, where local hunter-gatherers apparently
adopted Southwest Asian sheep and cereals over the course of many
centuries. In the Cape of South Africa the local Khoi hunter-gatherers became
herders (but not farmers) by acquiring sheep and cows from farther north in
Africa (and ultimately from Southwest Asia). Similarly, Native American
hunter-gatherers of the U.S. Southwest gradually became farmers by
acquiring Mexican crops. In these four areas the onset of food production
provides little or no evidence for the domestication of local plant or animal
species, but also little or no evidence for the replacement of human
population.
At the opposite extreme are regions in which food production certainly
began with an abrupt arrival of foreign people as well as of foreign crops and
animals. The reason why we can be certain is that the arrivals took place in
modern times and involved literate Europeans, who described in innumerable
books what happened. Those areas include California, the Pacific Northwest
of North America, the Argentine pampas, Australia, and Siberia. Until recent
centuries, these areas were still occupied by hunter-gatherers—Native
Americans in the first three cases and Aboriginal Australians or Native
Siberians in the last two. Those hunter-gatherers were killed, infected, driven
out, or largely replaced by arriving European farmers and herders who
brought their own crops and did not domesticate any local wild species after
their arrival (except for macadamia nuts in Australia). In the Cape of South
Africa the arriving Europeans found not only Khoi hunter-gatherers but also
Khoi herders who already possessed only domestic animals, not crops. The
result was again the start of farming dependent on crops from elsewhere, a
failure to domesticate local species, and a massive modern replacement of
human population.
Finally, the same pattern of an abrupt start of food production dependent
on domesticates from elsewhere, and an abrupt and massive population
replacement, seems to have repeated itself in many areas in the prehistoric
era. In the absence of written records, the evidence of those prehistoric
replacements must be sought in the archaeological record or inferred from
linguistic evidence. The best-attested cases are ones in which there can be no
doubt about population replacement because the newly arriving food
producers differed markedly in their skeletons from the hunter-gatherers
whom they replaced, and because the food producers introduced not only
crops and animals but also pottery. Later chapters will describe the two
clearest such examples: the Austronesian expansion from South China into
the Philippines and Indonesia (Chapter 17), and the Bantu expansion over
subequatorial Africa (Chapter 19).
Southeastern Europe and central Europe present a similar picture of an
abrupt onset of food production (dependent on Southwest Asian crops and
animals) and of pottery making. This onset too probably involved
replacement of old Greeks and Germans by new Greeks and Germans, just as
old gave way to new in the Philippines, Indonesia, and subequatorial Africa.
However, the skeletal differences between the earlier hunter-gatherers and the
farmers who replaced them are less marked in Europe than in the Philippines,
Indonesia, and subequatorial Africa. Hence the case for population
replacement in Europe is less strong or less direct.
IN SHORT, ONLY a few areas of the world developed food production
independently, and they did so at widely differing times. From those nuclear
areas, hunter-gatherers of some neighboring areas learned food production,
and peoples of other neighboring areas were replaced by invading food
producers from the nuclear areas—again at widely differing times. Finally,
peoples of some areas ecologically suitable for food production neither
evolved nor acquired agriculture in prehistoric times at all; they persisted as
hunter-gatherers until the modern world finally swept upon them. The peoples
of areas with a head start on food production thereby gained a head start on
the path leading toward guns, germs, and steel. The result was a long series of
collisions between the haves and the have-nots of history.
How can we explain these geographic differences in the times and modes
of onset of food production? That question, one of the most important
problems of prehistory, will be the subject of the next five chapters.
CHAPTER 6
TO FARM OR NOT TO FARM
FORMERLY, ALL PEOPLE ON EARTH WERE HUNTER-GATHERERS. Why did any of
them adopt food production at all? Given that they must have had some
reason, why did they do so around 8500 B.C. in Mediterranean habitats of the
Fertile Crescent, only 3,000 years later in the climatically and structurally
similar Mediterranean habitats of southwestern Europe, and never
indigenously in the similar Mediterranean habitats of California, southwestern
Australia, and the Cape of South Africa? Why did people of the Fertile
Crescent wait until 8500 B.C., instead of becoming food producers around
18,500 or 28,500 B.C.?
From our modern perspective, all these questions at first seem silly,
because the drawbacks of being a hunter-gatherer appear so obvious.
Scientists used to quote a phrase of Thomas Hobbes’s in order to characterize
the lifestyle of hunter-gatherers as “nasty, brutish, and short.” They seemed to
have to work hard, to be driven by the daily quest for food, often to be close
to starvation, to lack such elementary material comforts as soft beds and
adequate clothing, and to die young.
In reality, only for today’s affluent First World citizens, who don’t
actually do the work of raising food themselves, does food production (by
remote agribusinesses) mean less physical work, more comfort, freedom from
starvation, and a longer expected lifetime. Most peasant farmers and herders,
who constitute the great majority of the world’s actual food producers, aren’t
necessarily better off than hunter-gatherers. Time budget studies show that
they may spend more rather than fewer hours per day at work than hunter-
gatherers do. Archaeologists have demonstrated that the first farmers in many
areas were smaller and less well nourished, suffered from more serious
diseases, and died on the average at a younger age than the hunter-gatherers
they replaced. If those first farmers could have foreseen the consequences of
adopting food production, they might not have opted to do so. Why, unable to
foresee the result, did they nevertheless make that choice?
There exist many actual cases of hunter-gatherers who did see food
production practiced by their neighbors, and who nevertheless refused to
accept its supposed blessings and instead remained hunter-gatherers. For
instance, Aboriginal hunter-gatherers of northeastern Australia traded for
thousands of years with farmers of the Torres Strait Islands, between
Australia and New Guinea. California Native American hunter-gatherers
traded with Native American farmers in the Colorado River valley. In
addition, Khoi herders west of the Fish River of South Africa traded with
Bantu farmers east of the Fish River, and continued to dispense with farming
themselves. Why?
Still other hunter-gatherers in contact with farmers did eventually become
farmers, but only after what may seem to us like an inordinately long delay.
For example, the coastal peoples of northern Germany did not adopt food
production until 1,300 years after peoples of the Linearbandkeramik culture
introduced it to inland parts of Germany only 125 miles to the south. Why did
those coastal Germans wait so long, and what led them finally to change their
minds?
BEFORE WE CAN answer these questions, we must dispel some
misconceptions about the origins of food production and then reformulate the
question. What actually happened was not a discovery of food production, nor
an invention, as we might first assume. There was often not even a conscious
choice between food production and hunting-gathering. Specifically, in each
area of the globe the first people who adopted food production could
obviously not have been making a conscious choice or consciously striving
toward farming as a goal, because they had never seen farming and had no
way of knowing what it would be like. Instead, as we shall see, food
production evolved as a by-product of decisions made without awareness of
their consequences. Hence the question that we have to ask is why food
production did evolve, why it evolved in some places but not others, why at
different times in different places, and why not instead at some earlier or later
date.
Another misconception is that there is necessarily a sharp divide between
nomadic hunter-gatherers and sedentary food producers. In reality, although
we frequently draw such a contrast, hunter-gatherers in some productive
areas, including North America’s Pacific Northwest coast and possibly
southeastern Australia, became sedentary but never became food producers.
Other hunter-gatherers, in Palestine, coastal Peru, and Japan, became
sedentary first and adopted food production much later. Sedentary groups
probably made up a much higher fraction of hunter-gatherers 15,000 years
ago, when all inhabited parts of the world (including the most productive
areas) were still occupied by hunter-gatherers, than they do today, when the
few remaining hunter-gatherers survive only in unproductive areas where
nomadism is the sole option.
Conversely, there are mobile groups of food producers. Some modern
nomads of New Guinea’s Lakes Plains make clearings in the jungle, plant
bananas and papayas, go off for a few months to live again as hunter-
gatherers, return to check on their crops, weed the garden if they find the
crops growing, set off again to hunt, return months later to check again, and
settle down for a while to harvest and eat if their garden has produced.
Apache Indians of the southwestern United States settled down to farm in the
summer at higher elevations and toward the north, then withdrew to the south
and to lower elevations to wander in search of wild foods during the winter.
Many herding peoples of Africa and Asia shift camp along regular seasonal
routes to take advantage of predictable seasonal changes in pasturage. Thus,
the shift from hunting-gathering to food production did not always coincide
with a shift from nomadism to sedentary living.
Another supposed dichotomy that becomes blurred in reality is a
distinction between food producers as active managers of their land and
hunter-gatherers as mere collectors of the land’s wild produce. In reality,
some hunter-gatherers intensively manage their land. For example, New
Guinea peoples who never domesticated sago palms or mountain pandanus
nevertheless increase production of these wild edible plants by clearing away
encroaching competing trees, keeping channels in sago swamps clear, and
promoting growth of new sago shoots by cutting down mature sago trees.
Aboriginal Australians who never reached the stage of farming yams and seed
plants nonetheless anticipated several elements of farming. They managed the
landscape by burning it, to encourage the growth of edible seed plants that
sprout after fires. In gathering wild yams, they cut off most of the edible tuber
but replaced the stems and tops of the tubers in the ground so that the tubers
would regrow. Their digging to extract the tuber loosened and aerated the soil
and fostered regrowth. All that they would have had to do to meet the
definition of farmers was to carry the stems and remaining attached tubers
home and similarly replace them in soil at their camp.
FROM THOSE PRECURSORS of food production already practiced by hunter-
gatherers, it developed stepwise. Not all the necessary techniques were
developed within a short time, and not all the wild plants and animals that
were eventually domesticated in a given area were domesticated
simultaneously. Even in the cases of the most rapid independent development
of food production from a hunting-gathering lifestyle, it took thousands of
years to shift from complete dependence on wild foods to a diet with very few
wild foods. In early stages of food production, people simultaneously
collected wild foods and raised cultivated ones, and diverse types of
collecting activities diminished in importance at different times as reliance on
crops increased.
The underlying reason why this transition was piecemeal is that food
production systems evolved as a result of the accumulation of many separate
decisions about allocating time and effort. Foraging humans, like foraging
animals, have only finite time and energy, which they can spend in various
ways. We can picture an incipient farmer waking up and asking: Shall I spend
today hoeing my garden (predictably yielding a lot of vegetables several
months from now), gathering shellfish (predictably yielding a little meat
today), or hunting deer (yielding possibly a lot of meat today, but more likely
nothing)? Human and animal foragers are constantly prioritizing and making
effort-allocation decisions, even if only unconsciously. They concentrate first
on favorite foods, or ones that yield the highest payoff. If these are
unavailable, they shift to less and less preferred foods.
Many considerations enter into these decisions. People seek food in order
to satisfy their hunger and fill their bellies. They also crave specific foods,
such as protein-rich foods, fat, salt, sweet fruits, and foods that simply taste
good. All other things being equal, people seek to maximize their return of
calories, protein, or other specific food categories by foraging in a way that
yields the most return with the greatest certainty in the least time for the least
effort. Simultaneously, they seek to minimize their risk of starving: moderate
but reliable returns are preferable to a fluctuating lifestyle with a high time-
averaged rate of return but a substantial likelihood of starving to death. One
suggested function of the first gardens of nearly 11,000 years ago was to
provide a reliable reserve larder as insurance in case wild food supplies failed.
Conversely, men hunters tend to guide themselves by considerations of
prestige: for example, they might rather go giraffe hunting every day, bag a
giraffe once a month, and thereby gain the status of great hunter, than bring
home twice a giraffe’s weight of food in a month by humbling themselves and
reliably gathering nuts every day. People are also guided by seemingly
arbitrary cultural preferences, such as considering fish either delicacies or
taboo. Finally, their priorities are heavily influenced by the relative values
they attach to different lifestyles—just as we can see today. For instance, in
the 19th-century U.S. West, the cattlemen, sheepmen, and farmers all
despised each other. Similarly, throughout human history farmers have tended
to despise hunter-gatherers as primitive, hunter-gatherers have despised
farmers as ignorant, and herders have despised both. All these elements come
into play in people’s separate decisions about how to obtain their food.
AS WE ALREADY noted, the first farmers on each continent could not have
chosen farming consciously, because there were no other nearby farmers for
them to observe. However, once food production had arisen in one part of a
continent, neighboring hunter-gatherers could see the result and make
conscious decisions. In some cases the hunter-gatherers adopted the
neighboring system of food production virtually as a complete package; in
others they chose only certain elements of it; and in still others they rejected
food production entirely and remained hunter-gatherers.
For example, hunter-gatherers in parts of southeastern Europe had quickly
adopted Southwest Asian cereal crops, pulse crops, and livestock
simultaneously as a complete package by around 6000 B.C. All three of these
elements also spread rapidly through central Europe in the centuries before
5000 B.C. Adoption of food production may have been rapid and wholesale in
southeastern and central Europe because the hunter-gatherer lifestyle there
was less productive and less competitive. In contrast, food production was
adopted piecemeal in southwestern Europe (southern France, Spain, and
Italy), where sheep arrived first and cereals later. The adoption of intensive
food production from the Asian mainland was also very slow and piecemeal
in Japan, probably because the hunter-gatherer lifestyle based on seafood and
local plants was so productive there.
Just as a hunting-gathering lifestyle can be traded piecemeal for a food-
producing lifestyle, one system of food production can also be traded
piecemeal for another. For example, Indians of the eastern United States were
domesticating local plants by about 2500 B.C. but had trade connections with
Mexican Indians who developed a more productive crop system based on the
trinity of corn, squash, and beans. Eastern U.S. Indians adopted Mexican
crops, and many of them discarded many of their local domesticates,
piecemeal; squash was domesticated independently, corn arrived from Mexico
around A.D. 200 but remained a minor crop until around A.D. 900, and beans
arrived a century or two later. It even happened that food-production systems
were abandoned in favor of hunting-gathering. For instance, around 3000 B.C.
the hunter-gatherers of southern Sweden adopted farming based on Southwest
Asian crops, but abandoned it around 2700 B.C. and reverted to hunting-
gathering for 400 years before resuming farming.
ALL THESE CONSIDERATIONS make it clear that we should not suppose that the
decision to adopt farming was made in a vacuum, as if the people had
previously had no means to feed themselves. Instead, we must consider food
production and hunting-gathering as alternative strategies competing with
each other. Mixed economies that added certain crops or livestock to hunting-
gathering also competed against both types of “pure” economies, and against
mixed economies with higher or lower proportions of food production.
Nevertheless, over the last 10,000 years, the predominant result has been a
shift from hunting-gathering to food production. Hence we must ask: What
were the factors that tipped the competitive advantage away from the former
and toward the latter?
That question continues to be debated by archaeologists and
anthropologists. One reason for its remaining unsettled is that different factors
may have been decisive in different parts of the world. Another has been the
problem of disentangling cause and effect in the rise of food production.
However, five main contributing factors can still be identified; the
controversies revolve mainly around their relative importance.
One factor is the decline in the availability of wild foods. The lifestyle of
hunter-gatherers has become increasingly less rewarding over the past 13,000
years, as resources on which they depended (especially animal resources)
have become less abundant or even disappeared. As we saw in Chapter 1,
most large mammal species became extinct in North and South America at the
end of the Pleistocene, and some became extinct in Eurasia and Africa, either
because of climate changes or because of the rise in skill and numbers of
human hunters. While the role of animal extinctions in eventually (after a
long lag) nudging ancient Native Americans, Eurasians, and Africans toward
food production can be debated, there are numerous incontrovertible cases on
islands in more recent times. Only after the first Polynesian settlers had
exterminated moas and decimated seal populations on New Zealand, and
exterminated or decimated seabirds and land birds on other Polynesian
islands, did they intensify their food production. For instance, although the
Polynesians who colonized Easter Island around A.D. 500 brought chickens
with them, chicken did not become a major food until wild birds and
porpoises were no longer readily available as food. Similarly, a suggested
contributing factor to the rise of animal domestication in the Fertile Crescent
was the decline in abundance of the wild gazelles that had previously been a
major source of meat for hunter-gatherers in that area.
A second factor is that, just as the depletion of wild game tended to make
hunting-gathering less rewarding, an increased availability of domesticable
wild plants made steps leading to plant domestication more rewarding. For
instance, climate changes at the end of the Pleistocene in the Fertile Crescent
greatly expanded the area of habitats with wild cereals, of which huge crops
could be harvested in a short time. Those wild cereal harvests were precursors
to the domestication of the earliest crops, the cereals wheat and barley, in the
Fertile Crescent.
Still another factor tipping the balance away from hunting-gathering was
the cumulative development of technologies on which food production would
eventually depend—technologies for collecting, processing, and storing wild
foods. What use can would-be farmers make of a ton of wheat grains on the
stalk, if they have not first figured out how to harvest, husk, and store them?
The necessary methods, implements, and facilities appeared rapidly in the
Fertile Crescent after 11,000 B.C., having been invented for dealing with the
newly available abundance of wild cereals.
Those inventions included sickles of flint blades cemented into wooden or
bone handles, for harvesting wild grains; baskets in which to carry the grains
home from the hillsides where they grew; mortars and pestles, or grinding
slabs, to remove the husks; the technique of roasting grains so that they could
be stored without sprouting; and underground storage pits, some of them
plastered to make them waterproof. Evidence for all of these techniques
becomes abundant at sites of hunter-gatherers in the Fertile Crescent after
11,000 B.C. All these techniques, though developed for the exploitation of wild
cereals, were prerequisites to the planting of cereals as crops. These
cumulative developments constituted the unconscious first steps of plant
domestication.
A fourth factor was the two-way link between the rise in human
population density and the rise in food production. In all parts of the world
where adequate evidence is available, archaeologists find evidence of rising
densities associated with the appearance of food production. Which was the
cause and which the result? This is a long-debated chicken-or-egg problem:
did a rise in human population density force people to turn to food
production, or did food production permit a rise in human population density?
In principle, one expects the chain of causation to operate in both
directions. As I’ve already discussed, food production tends to lead to
increased population densities because it yields more edible calories per acre
than does hunting-gathering. On the other hand, human population densities
were gradually rising throughout the late Pleistocene anyway, thanks to
improvements in human technology for collecting and processing wild foods.
As population densities rose, food production became increasingly favored
because it provided the increased food outputs needed to feed all those
people.
That is, the adoption of food production exemplifies what is termed an
autocatalytic process—one that catalyzes itself in a positive feedback cycle,
going faster and faster once it has started. A gradual rise in population
densities impelled people to obtain more food, by rewarding those who
unconsciously took steps toward producing it. Once people began to produce
food and become sedentary, they could shorten the birth spacing and produce
still more people, requiring still more food. This bidirectional link between
food production and population density explains the paradox that food
production, while increasing the quantity of edible calories per acre, left the
food producers less well nourished than the hunter-gatherers whom they
succeeded. That paradox developed because human population densities rose
slightly more steeply than did the availability of food.
Taken together, these four factors help us understand why the transition to
food production in the Fertile Crescent began around 8500 B.C., not around
18,500 or 28,500 B.C. At the latter two dates hunting-gathering was still much
more rewarding than incipient food production, because wild mammals were
still abundant; wild cereals were not yet abundant; people had not yet
developed the inventions necessary for collecting, processing, and storing
cereals efficiently; and human population densities were not yet high enough
for a large premium to be placed on extracting more calories per acre.
A final factor in the transition became decisive at geographic boundaries
between hunter-gatherers and food producers. The much denser populations
of food producers enabled them to displace or kill hunter-gatherers by their
sheer numbers, not to mention the other advantages associated with food
production (including technology, germs, and professional soldiers). In areas
where there were only hunter-gatherers to begin with, those groups of hunter-
gatherers who adopted food production outbred those who didn’t.
As a result, in most areas of the globe suitable for food production,
hunter-gatherers met one of two fates: either they were displaced by
neighboring food producers, or else they survived only by adopting food
production themselves. In places where they were already numerous or where
geography retarded immigration by food producers, local hunter-gatherers did
have time to adopt farming in prehistoric times and thus to survive as farmers.
This may have happened in the U.S. Southwest, in the western Mediterranean,
on the Atlantic coast of Europe, and in parts of Japan. However, in Indonesia,
tropical Southeast Asia, most of subequatorial Africa, and probably in parts of
Europe, the hunter-gatherers were replaced by farmers in the prehistoric era,
whereas a similar replacement took place in modern times in Australia and
much of the western United States.
Only where especially potent geographic or ecological barriers made
immigration of food producers or diffusion of locally appropriate food-
producing techniques very difficult were hunter-gatherers able to persist until
modern times in areas suitable for food production. The three outstanding
examples are the persistence of Native American hunter-gatherers in
California, separated by deserts from the Native American farmers of
Arizona; that of Khoisan hunter-gatherers at the Cape of South Africa, in a
Mediterranean climate zone unsuitable for the equatorial crops of nearby
Bantu farmers; and that of hunter-gatherers throughout the Australian
continent, separated by narrow seas from the food producers of Indonesia and
New Guinea. Those few peoples who remained hunter-gatherers into the 20th
century escaped replacement by food producers because they were confined
to areas not fit for food production, especially deserts and Arctic regions.
Within the present decade, even they will have been seduced by the
attractions of civilization, settled down under pressure from bureaucrats or
missionaries, or succumbed to germs.
CHAPTER 7
HOW TO MAKE AN ALMOND
IF YOU’RE A HIKER WHOSE APPETITE IS JADED BY FARM-grown foods, it’s fun to
try eating wild foods. You know that some wild plants, such as wild
strawberries and blueberries, are both tasty and safe to eat. They’re
sufficiently similar to familiar crops that you can easily recognize the wild
berries, even though they’re much smaller than those we grow. Adventurous
hikers cautiously eat mushrooms, aware that many species can kill us. But not
even ardent nut lovers eat wild almonds, of which a few dozen contain
enough cyanide (the poison used in Nazi gas chambers) to kill us. The forest
is full of many other plants deemed inedible.
Yet all crops arose from wild plant species. How did certain wild plants
get turned into crops? That question is especially puzzling in regard to the
many crops (like almonds) whose wild progenitors are lethal or bad-tasting,
and to other crops (like corn) that look drastically different from their wild
ancestors. What cavewoman or caveman ever got the idea of “domesticating”
a plant, and how was it accomplished?
Plant domestication may be defined as growing a plant and thereby,
consciously or unconsciously, causing it to change genetically from its wild
ancestor in ways making it more useful to human consumers. Crop
development is today a conscious, highly specialized effort carried out by
professional scientists. They already know about the hundreds of existing
crops and set out to develop yet another one. To achieve that goal, they plant
many different seeds or roots, select the best progeny and plant their seeds,
apply knowledge of genetics to develop good varieties that breed true, and
perhaps even use the latest techniques of genetic engineering to transfer
specific useful genes. At the Davis campus of the University of California, an
entire department (the Department of Pomology) is devoted to apples and
another (the Department of Viticulture and Enology) to grapes and wine.
But plant domestication goes back over 10,000 years. Early farmers
surely didn’t use molecular genetic techniques to arrive at their results. The
first farmers didn’t even have any existing crop as a model to inspire them to
develop new ones. Hence they couldn’t have known that, whatever they were
doing, they would enjoy a tasty treat as a result.
How, then, did early farmers domesticate plants unwittingly? For
example, how did they turn poisonous almonds into safe ones without
knowing what they were doing? What changes did they actually make in wild
plants, besides rendering some of them bigger or less poisonous? Even for
valuable crops, the times of domestication vary greatly: for instance, peas
were domesticated by 8000 B.C., olives around 4000 B.C., strawberries not
until the Middle Ages, and pecans not until 1846. Many valuable wild plants
yielding food prized by millions of people, such as oaks sought for their
edible acorns in many parts of the world, remain untamed even today. What
made some plants so much easier or more inviting to domesticate than others?
Why did olive trees yield to Stone Age farmers, whereas oak trees continue to
defeat our brightest agronomists?
LET’S BEGIN BY looking at domestication from the plant’s point of view. As
far as plants are concerned, we’re just one of thousands of animal species that
unconsciously “domesticate” plants.
Like all animal species (including humans), plants must spread their
offspring to areas where they can thrive and pass on their parents’ genes.
Young animals disperse by walking or flying, but plants don’t have that
option, so they must somehow hitchhike. While some plant species have
seeds adapted for being carried by the wind or for floating on water, many
others trick an animal into carrying their seeds, by wrapping the seed in a
tasty fruit and advertising the fruit’s ripeness by its color or smell. The hungry
animal plucks and swallows the fruit, walks or flies off, and then spits out or
defecates the seed somewhere far from its parent tree. Seeds can in this
manner be carried for thousands of miles.
It may come as a surprise to learn that plant seeds can resist digestion by
your gut and nonetheless germinate out of your feces. But any adventurous
readers who are not too squeamish can make the test and prove it for
themselves. The seeds of many wild plant species actually must pass through
an animal’s gut before they can germinate. For instance, one African melon
species is so well adapted to being eaten by a hyena-like animal called the
aardvark that most melons of that species grow on the latrine sites of
aardvarks.
As an example of how would-be plant hitchhikers attract animals,
consider wild strawberries. When strawberry seeds are still young and not yet
ready to be planted, the surrounding fruit is green, sour, and hard. When the
seeds finally mature, the berries turn red, sweet, and tender. The change in the
berries’ color serves as a signal attracting birds like thrushes to pluck the
berries and fly off, eventually to spit out or defecate the seeds.
Naturally, strawberry plants didn’t set out with a conscious intent of
attracting birds when, and only when, their seeds were ready to be dispersed.
Neither did thrushes set out with the intent of domesticating strawberries.
Instead, strawberry plants evolved through natural selection. The greener and
more sour the young strawberry, the fewer the birds that destroyed the seeds
by eating berries before the seeds were ready; the sweeter and redder the final
strawberry, the more numerous the birds that dispersed its ripe seeds.
Countless other plants have fruits adapted to being eaten and dispersed by
particular species of animals. Just as strawberries are adapted to birds, so
acorns are adapted to squirrels, mangos to bats, and some sedges to ants. That
fulfills part of our definition of plant domestication, as the genetic
modification of an ancestral plant in ways that make it more useful to
consumers. But no one would seriously describe this evolutionary process as
domestication, because birds and bats and other animal consumers don’t
fulfill the other part of the definition: they don’t consciously grow plants. In
the same way, the early unconscious stages of crop evolution from wild plants
consisted of plants evolving in ways that attracted humans to eat and disperse
their fruit without yet intentionally growing them. Human latrines, like those
of aardvarks, may have been a testing ground of the first unconscious crop
breeders.
LATRINES ARE MERELY one of the many places where we accidentally sow the
seeds of wild plants that we eat. When we gather edible wild plants and bring
them home, some spill en route or at our houses. Some fruit rots while still
containing perfectly good seeds, and gets thrown out uneaten into the
garbage. As parts of the fruit that we actually take into our mouths, strawberry
seeds are tiny and inevitably swallowed and defecated, but other seeds are
large enough to be spat out. Thus, our spittoons and garbage dumps joined our
latrines to form the first agricultural research laboratories.
At whichever such “lab” the seeds ended up, they tended to come from
only certain individuals of edible plants—namely, those that we preferred to
eat for one reason or another. From your berry-picking days, you know that
you select particular berries or berry bushes. Eventually, when the first
farmers began to sow seeds deliberately, they would inevitably sow those
from the plants they had chosen to gather, even though they didn’t understand
the genetic principle that big berries have seeds likely to grow into bushes
yielding more big berries.
So, when you wade into a thorny thicket amid the mosquitoes on a hot,
humid day, you don’t do it for just any strawberry bush. Even if
unconsciously, you decide which bush looks most promising, and whether it’s
worth it at all. What are your unconscious criteria?
One criterion, of course, is size. You prefer large berries, because it’s not
worth your while to get sunburned and mosquito bitten for some lousy little
berries. That provides part of the explanation why many crop plants have
much bigger fruits than their wild ancestors do. It’s especially familiar to us
that supermarket strawberries and blueberries are gigantic compared with
wild ones; those differences arose only in recent centuries.
Such size differences in other plants go back to the very beginnings of
agriculture, when cultivated peas evolved through human selection to be 10
times heavier than wild peas. The little wild peas had been collected by
hunter-gatherers for thousands of years, just as we collect little wild
blueberries today, before the preferential harvesting and planting of the most
appealing largest wild peas—that is, what we call farming—began
automatically to contribute to increases in average pea size from generation to
generation. Similarly, supermarket apples are typically around three inches in
diameter, wild apples only one inch. The oldest corn cobs are barely more
than half an inch long, but Mexican Indian farmers of A.D. 1500 already had
developed six-inch cobs, and some modern cobs are one and a half feet long.
Another obvious difference between seeds that we grow and many of their
wild ancestors is in bitterness. Many wild seeds evolved to be bitter, bad-
tasting, or actually poisonous, in order to deter animals from eating them.
Thus, natural selection acts oppositely on seeds and on fruits. Plants whose
fruits are tasty get their seeds dispersed by animals, but the seed itself within
the fruit has to be bad-tasting. Otherwise, the animal would also chew up the
seed, and it couldn’t sprout.
Almonds provide a striking example of bitter seeds and their change
under domestication. Most wild almond seeds contain an intensely bitter
chemical called amygdalin, which (as was already mentioned) breaks down to
yield the poison cyanide. A snack of wild almonds can kill a person foolish
enough to ignore the warning of the bitter taste. Since the first stage in
unconscious domestication involves gathering seeds to eat, how on earth did
domestication of wild almonds ever reach that first stage?
The explanation is that occasional individual almond trees have a
mutation in a single gene that prevents them from synthesizing the bitter-
tasting amygdalin. Such trees die out in the wild without leaving any progeny,
because birds discover and eat all their seeds. But curious or hungry children
of early farmers, nibbling wild plants around them, would eventually have
sampled and noticed those nonbitter almond trees. (In the same way,
European peasants today still recognize and appreciate occasional individual
oak trees whose acorns are sweet rather than bitter.) Those nonbitter almond
seeds are the only ones that ancient farmers would have planted, at first
unintentionally in their garbage heaps and later intentionally in their orchards.
Already by 8000 B.C. wild almonds show up in excavated archaeological
sites in Greece. By 3000 B.C. they were being domesticated in lands of the
eastern Mediterranean. When the Egyptian king Tutankhamen died, around
1325 B.C., almonds were one of the foods left in his famous tomb to nourish
him in the afterlife. Lima beans, watermelons, potatoes, eggplants, and
cabbages are among the many other familiar crops whose wild ancestors were
bitter or poisonous, and of which occasional sweet individuals must have
sprouted around the latrines of ancient hikers.
While size and tastiness are the most obvious criteria by which human
hunter-gatherers select wild plants, other criteria include fleshy or seedless
fruits, oily seeds, and long fibers. Wild squashes and pumpkins have little or
no fruit around their seeds, but the preferences of early farmers selected for
squashes and pumpkins consisting of far more flesh than seeds. Cultivated
bananas were selected long ago to be all flesh and no seed, thereby inspiring
modern agricultural scientists to develop seedless oranges, grapes, and
watermelons as well. Seedlessness provides a good example of how human
selection can completely reverse the original evolved function of a wild fruit,
which in nature serves as a vehicle for dispersing seeds.
In ancient times many plants were similarly selected for oily fruits or
seeds. Among the earliest fruit trees domesticated in the Mediterranean world
were olives, cultivated since around 4000 B.C. for their oil. Crop olives are not
only bigger but also oilier than wild ones. Ancient farmers selected sesame,
mustard, poppies, and flax as well for oily seeds, while modern plant
scientists have done the same for sunflower, safflower, and cotton.
Before that recent development of cotton for oil, it was of course selected
for its fibers, used to weave textiles. The fibers (termed lint) are hairs on the
cotton seeds, and early farmers of both the Americas and the Old World
independently selected different species of cotton for long lint. In flax and
hemp, two other plants grown to supply the textiles of antiquity, the fibers
come instead from the stem, and plants were selected for long, straight stems.
While we think of most crops as being grown for food, flax is one of our
oldest crops (domesticated by around 7000 B.C.). It furnished linen, which
remained the chief textile of Europe until it became supplanted by cotton and
synthetics after the Industrial Revolution.
SO FAR, ALL the changes that I’ve described in the evolution of wild plants
into crops involve characters that early farmers could actually notice—such as
fruit size, bitterness, fleshiness, and oiliness, and fiber length. By harvesting
those individual wild plants possessing these desirable qualities to an
exceptional degree, ancient peoples unconsciously dispersed the plants and
set them on the road to domestication.
In addition, though, there were at least four other major types of change
that did not involve berry pickers making visible choices. In these cases the
berry pickers caused changes either by harvesting available plants while other
plants remained unavailable for invisible reasons, or by changing the selective
conditions acting on plants.
The first such change affected wild mechanisms for the dispersal of seeds.
Many plants have specialized mechanisms that scatter seeds (and thereby
prevent humans from gathering them efficiently). Only mutant seeds lacking
those mechanisms would have been harvested and would thus have become
the progenitors of crops.
A clear example involves peas, whose seeds (the peas we eat) come
enclosed in a pod. Wild peas have to get out of the pod if they are to
germinate. To achieve that result, pea plants evolved a gene that makes the
pod explode, shooting out the peas onto the ground. Pods of occasional
mutant peas don’t explode. In the wild the mutant peas would die entombed
in their pod on their parent plants, and only the popping pods would pass on
their genes. But, conversely, the only pods available to humans to harvest
would be the nonpopping ones left on the plant. Thus, once humans began
bringing wild peas home to eat, there was immediate selection for that single-
gene mutant. Similar nonpopping mutants were selected in lentils, flax, and
poppies.
Instead of being enclosed in a poppable pod, wild wheat and barley seeds
grow at the top of a stalk that spontaneously shatters, dropping the seeds to
the ground where they can germinate. A single-gene mutation prevents the
stalks from shattering. In the wild that mutation would be lethal to the plant,
since the seeds would remain suspended in the air, unable to germinate and
take root. But those mutant seeds would have been the ones waiting
conveniently on the stalk to be harvested and brought home by humans. When
humans then planted those harvested mutant seeds, any mutant seeds among
the progeny again became available to the farmers to harvest and sow, while
normal seeds among the progeny fell to the ground and became unavailable.
Thus, human farmers reversed the direction of natural selection by 180
degrees: the formerly successful gene suddenly became lethal, and the lethal
mutant became successful. Over 10,000 years ago, that unconscious selection
for nonshattering wheat and barley stalks was apparently the first major
human “improvement” in any plant. That change marked the beginning of
agriculture in the Fertile Crescent.
The second type of change was even less visible to ancient hikers. For
annual plants growing in an area with a very unpredictable climate, it could
be lethal if all the seeds sprouted quickly and simultaneously. Were that to
happen, the seedlings might all be killed by a single drought or frost, leaving
no seeds to propagate the species. Hence many annual plants have evolved to
hedge their bets by means of germination inhibitors, which make seeds
initially dormant and spread out their germination over several years. In that
way, even if most seedlings are killed by a bout of bad weather, some seeds
will be left to germinate later.
A common bet-hedging adaptation by which wild plants achieve that
result is to enclose their seeds in a thick coat or armor. The many wild plants
with such adaptations include wheat, barley, peas, flax, and sunflowers. While
such late-sprouting seeds still have the opportunity to germinate in the wild,
consider what must have happened as farming developed. Early farmers
would have discovered by trial and error that they could obtain higher yields
by tilling and watering the soil and then sowing seeds. When that happened,
seeds that immediately sprouted grew into plants whose seeds were harvested
and planted in the next year. But many of the wild seeds did not immediately
sprout, and they yielded no harvest.
Occasional mutant individuals among wild plants lacked thick seed coats
or other inhibitors of germination. All such mutants promptly sprouted and
yielded harvested mutant seeds. Early farmers wouldn’t have noticed the
difference, in the way that they did notice and selectively harvest big berries.
But the cycle of sow / grow / harvest / sow would have selected immediately
and unconsciously for the mutants. Like the changes in seed dispersal, these
changes in germination inhibition characterize wheat, barley, peas, and many
other crops compared with their wild ancestors.
The remaining major type of change invisible to early farmers involved
plant reproduction. A general problem in crop development is that occasional
mutant plant individuals are more useful to humans (for example, because of
bigger or less bitter seeds) than are normal individuals. If those desirable
mutants proceeded to interbreed with normal plants, the mutation would
immediately be diluted or lost. Under what circumstances would it remain
preserved for early farmers?
For plants that reproduce themselves, the mutant would automatically be
preserved. That’s true of plants that reproduce vegetatively (from a tuber or
root of the parent plant), or that are hermaphrodites capable of fertilizing
themselves. But the vast majority of wild plants don’t reproduce that way.
They’re either hermaphrodites incapable of fertilizing themselves and forced
to interbreed with other hermaphrodite individuals (my male part fertilizes
your female part, your male part fertilizes my female part), or else they occur
as separate male and female individuals, like all normal mammals. The
former plants are termed self-incompatible hermaphrodites; the latter,
dioecious species. Both were bad news for ancient farmers, who would
thereby have promptly lost any favorable mutants without understanding why.
The solution involved another type of invisible change. Numerous plant
mutations affect the reproductive system itself. Some mutant individuals
developed fruit without even having to be pollinated, resulting in our seedless
bananas, grapes, oranges, and pineapples. Some mutant hermaphrodites lost
their self-incompatibility and became able to fertilize themselves—a process
exemplified by many fruit trees such as plums, peaches, apples, apricots, and
cherries. Some mutant grapes that normally would have had separate male
and female individuals also became self-fertilizing hermaphrodites. By all
these means, ancient farmers, who didn’t understand plant reproductive
biology, still ended up with useful crops that bred true and were worth
replanting, instead of initially promising mutants whose worthless progeny
were consigned to oblivion.
Thus, farmers selected from among individual plants on the basis not only
of perceptible qualities like size and taste, but also of invisible features like
seed dispersal mechanisms, germination inhibition, and reproductive biology.
As a result, different plants became selected for quite different or even
opposite features. Some plants (like sunflowers) were selected for much
bigger seeds, while others (like bananas) were selected for tiny or even
nonexistent seeds. Lettuce was selected for luxuriant leaves at the expense of
seeds or fruit; wheat and sunflowers, for seeds at the expense of leaves; and
squash, for fruit at the expense of leaves. Especially instructive are cases in
which a single wild plant species was variously selected for different purposes
and thereby gave rise to quite different-looking crops. Beets, grown already in
Babylonian times for their leaves (like the modern beet varieties called
chards), were then developed for their edible roots and finally (in the 18th
century) for their sugar content (sugar beets). Ancestral cabbage plants,
possibly grown originally for their oily seeds, underwent even greater
diversification as they became variously selected for leaves (modern cabbage
and kale), stems (kohlrabi), buds (brussels sprouts), or flower shoots
(cauliflower and broccoli).
So far, we have been discussing transformations of wild plants into crops
as a result of selection by farmers, consciously or unconsciously. That is,
farmers initially selected seeds of certain wild plant individuals to bring into
their gardens and then chose certain progeny seeds each year to grow in the
next year’s garden. But much of the transformation was also effected as a
result of plants’ selecting themselves. Darwin’s phrase “natural selection”
refers to certain individuals of a species surviving better, and / or reproducing
more successfully, than competing individuals of the same species under
natural conditions. In effect, the natural processes of differential survival and
reproduction do the selecting. If the conditions change, different types of
individuals may now survive or reproduce better and become “naturally
selected,” with the result that the population undergoes evolutionary change.
A classic example is the development of industrial melanism in British moths:
darker moth individuals became relatively commoner than paler individuals
as the environment became dirtier during the 19th century, because dark
moths resting on a dark, dirty tree were more likely than contrasting pale
moths to escape the attention of predators.
Much as the Industrial Revolution changed the environment for moths,
farming changed the environment for plants. A tilled, fertilized, watered,
weeded garden provides growing conditions very different from those on a
dry, unfertilized hillside. Many changes of plants under domestication
resulted from such changes in conditions and hence in the favored types of
individuals. For example, when a farmer sows seeds densely in a garden,
there is intense competition among the seeds. Big seeds that can take
advantage of the good conditions to grow quickly will now be favored over
small seeds that were previously favored on dry, unfertilized hillsides where
seeds were sparser and competition less intense. Such increased competition
among plants themselves made a major contribution to larger seed size and to
many other changes developing during the transformation of wild plants into
ancient crops.
WHAT ACCOUNTS FOR the great differences among plants in ease of
domestication, such that some species were domesticated long ago and others
not until the Middle Ages, whereas still other wild plants have proved
immune to all our activities? We can deduce many of the answers by
examining the well-established sequence in which various crops developed in
Southwest Asia’s Fertile Crescent.
It turns out that the earliest Fertile Crescent crops, such as the wheat and
barley and peas domesticated around 10,000 years ago, arose from wild
ancestors offering many advantages. They were already edible and gave high
yields in the wild. They were easily grown, merely by being sown or planted.
They grew quickly and could be harvested within a few months of sowing, a
big advantage for incipient farmers still on the borderline between nomadic
hunters and settled villagers. They could be readily stored, unlike many later
crops such as strawberries and lettuce. They were mostly self-pollinating: that
is, the crop varieties could pollinate themselves and pass on their own
desirable genes unchanged, instead of having to hybridize with other varieties
less useful to humans. Finally, their wild ancestors required very little genetic
change to be converted into crops—for instance, in wheat, just the mutations
for nonshattering stalks and uniform quick germination.
A next stage of crop development included the first fruit and nut trees,
domesticated around 4000 B.C. They comprised olives, figs, dates,
pomegranates, and grapes. Compared with cereals and legumes, they had the
drawback of not starting to yield food until at least three years after planting,
and not reaching full production until after as much as a decade. Thus,
growing these crops was possible only for people already fully committed to
the settled village life. However, these early fruit and nut trees were still the
easiest such crops to cultivate. Unlike later tree domesticates, they could be
grown directly by being planted as cuttings or even seeds. Cuttings have the
advantage that, once ancient farmers had found or developed a productive
tree, they could be sure that all its descendants would remain identical to it.
A third stage involved fruit trees that proved much harder to cultivate,
including apples, pears, plums, and cherries. These trees cannot be grown
from cuttings. It’s also a waste of effort to grow them from seed, since the
offspring even of an outstanding individual tree of those species are highly
variable and mostly yield worthless fruit. Instead, those trees must be grown
by the difficult technique of grafting, developed in China long after the
beginnings of agriculture. Not only is grafting hard work even once you know
the principle, but the principle itself could have been discovered only through
conscious experimentation. The invention of grafting was hardly just a matter
of some nomad relieving herself at a latrine and returning later to be
pleasantly surprised by the resulting crop of fine fruit.
Many of these late-stage fruit trees posed a further problem in that their
wild progenitors were the opposite of self-pollinating. They had to be cross-
pollinated by another plant belonging to a genetically different variety of their
species. Hence early farmers either had to find mutant trees not requiring
cross-pollination, or had consciously to plant genetically different varieties or
else male and female individuals nearby in the same orchard. All those
problems delayed the domestication of apples, pears, plums, and cherries until
around classical times. At about the same time, though, another group of late
domesticates arose with much less effort, as wild plants that established
themselves initially as weeds in fields of intentionally cultivated crops. Crops
starting out as weeds included rye and oats, turnips and radishes, beets and
leeks, and lettuce.
ALTHOUGH THE DETAILED sequence that I’ve just described applies to the
Fertile Crescent, partly similar sequences also appeared elsewhere in the
world. In particular, the Fertile Crescent’s wheat and barley exemplify the
class of crops termed cereals or grains (members of the grass family), while
Fertile Crescent peas and lentils exemplify pulses (members of the legume
family, which includes beans). Cereal crops have the virtues of being fast
growing, high in carbohydrates, and yielding up to a ton of edible food per
hectare cultivated. As a result, cereals today account for over half of all
calories consumed by humans and include five of the modern world’s 12
leading crops (wheat, corn, rice, barley, and sorghum). Many cereal crops are
low in protein, but that deficit is made up by pulses, which are often 25
percent protein (38 percent in the case of soybeans). Cereals and pulses
together thus provide many of the ingredients of a balanced diet.
As Table 7.1 summarizes, the domestication of local cereal / pulse
combinations launched food production in many areas. The most familiar
examples are the combination of wheat and barley with peas and lentils in the
Fertile Crescent, the combination of corn with several bean species in
Mesoamerica, and the combination of rice and millets with soybeans and
other beans in China. Less well known are Africa’s combination of sorghum,
African rice, and pearl millet with cowpeas and groundnuts, and the Andes’
combination of the noncereal grain quinoa with several bean species.
Table 7.1 also shows that the Fertile Crescent’s early domestication of
flax for fiber was paralleled elsewhere. Hemp, four cotton species, yucca, and
agave variously furnished fiber for rope and woven clothing in China,
Mesoamerica, India, Ethiopia, sub-Saharan Africa, and South America,
supplemented in several of those areas by wool from domestic animals. Of the
centers of early food production, only the eastern United States and New
Guinea remained without a fiber crop.
TABLE 7.1 Examples of Early Major Crop Types around the Ancient
World
Crop Type
Area
Cereals, Other
Pulses
Fiber
Roots, Tubers Melons
Grasses
Fertile
emmer wheat,
pea, lentil, chickpea flax
—
muskmelon
Crescent
einkorn wheat,
barley
China
foxtail millet,
soybean, adzuki
hemp
—
[muskmelon]
broomcorn millet,
bean, mung bean
rice
Mesoamerica corn
common bean,
cotton (G.
jicama
squashes (C.
tepary bean, scarlet
hirsutum),
pepo, etc.)
runner bean
yucca, agave
Andes,
quinoa, [corn]
lima bean, common
cotton (G.
manioc, sweet squashes (C.
Amazonia
bean, peanut
barbadense)
potato,
maxima,
potato, oca
etc.)
West Africa sorghum, pearl
cowpea, groundnut
cotton (G.
African yams
watermelon,
and Sahel
millet, African
herbaceum)
bottle gourd
rice
India
[wheat, barley, rice, hyacinth bean, black cotton (G
—
cucumber
sorghum, millets]
gram, green gram
arboreum),
flax
Ethiopia
teff, finger millet,
[pea, lentil]
[flax]
—
—
[wheat, barley]
Eastern
maygrass, little
—
—
Jerusalem
squash (C.
United
barley, knotweed,
artichoke
pepo)
States
goosefoot
New Guinea sugar cane
—
—
yams, taro
—
The table gives major crops, of five crop classes, from early agricultural
sites in various parts of the world. Square brackets enclose names of crops
first domesticated elsewhere; names not enclosed in brackets refer to local
domesticates. Omitted are crops that arrived or became important only later,
such as bananas in Africa, corn and beans in the eastern United States, and
sweet potato in New Guinea. Cottons are four species of the genus
Gossypium, each species being native to a particular part of the world;
squashes are five species of the genus Cucurbita. Note that cereals, pulses,
and fiber crops launched agriculture in most areas, but that root and tuber
crops and melons were of early importance in only some areas.
Alongside these parallels, there were also some major differences in food
production systems around the world. One is that agriculture in much of the
Old World came to involve broadcast seeding and monoculture fields, and
eventually plowing. That is, seeds were sown by being thrown in handfuls,
resulting in a whole field devoted to a single crop. Once cows, horses, and
other large mammals were domesticated, they were hitched to plows, and
fields were tilled by animal power. In the New World, however, no animal
was ever domesticated that could be hitched to a plow. Instead, fields were
always tilled by hand-held sticks or hoes, and seeds were planted individually
by hand and not scattered as whole handfuls. Most New World fields thus
came to be mixed gardens of many crops planted together, rather than
monoculture.
Another major difference among agricultural systems involved the main
sources of calories and carbohydrates. As we have seen, these were cereals in
many areas. In other areas, though, that role of cereals was taken over or
shared by roots and tubers, which were of negligible importance in the ancient
Fertile Crescent and China. Manioc (alias cassava) and sweet potato became
staples in tropical South America, potato and oca in the Andes, African yams
in Africa, and Indo-Pacific yams and taro in Southeast Asia and New Guinea.
Tree crops, notably bananas and breadfruit, also furnished carbohydrate-rich
staples in Southeast Asia and New Guinea.
THUS, BY ROMAN times, almost all of today’s leading crops were being
cultivated somewhere in the world. Just as we shall see for domestic animals
too (Chapter 9), ancient hunter-gatherers were intimately familiar with local
wild plants, and ancient farmers evidently discovered and domesticated
almost all of those worth domesticating. Of course, medieval monks did begin
to cultivate strawberries and raspberries, and modern plant breeders are still
improving ancient crops and have added new minor crops, notably some
berries (like blueberries, cranberries, and kiwifruit) and nuts (macadamias,
pecans, and cashews). But these few modern additions have remained of
modest importance compared with ancient staples like wheat, corn, and rice.
Still, our list of triumphs lacks many wild plants that, despite their value
as food, we never succeeded in domesticating. Notable among these failures
of ours are oak trees, whose acorns were a staple food of Native Americans in
California and the eastern United States as well as a fallback food for
European peasants in famine times of crop failure. Acorns are nutritionally
valuable, being rich in starch and oil. Like many otherwise edible wild foods,
most acorns do contain bitter tannins, but acorn lovers learned to deal with
tannins in the same way that they dealt with bitter chemicals in almonds and
other wild plants: either by grinding and leaching the acorns to remove the
tannins, or by harvesting acorns from the occasional mutant individual oak
tree low in tannins.
Why have we failed to domesticate such a prized food source as acorns?
Why did we take so long to domesticate strawberries and raspberries? What is
it about those plants that kept their domestication beyond the reach of ancient
farmers capable of mastering such difficult techniques as grafting?
It turns out that oak trees have three strikes against them. First, their slow
growth would exhaust the patience of most farmers. Sown wheat yields a crop
within a few months; a planted almond grows into a nut-bearing tree in three
or four years; but a planted acorn may not become productive for a decade or
more. Second, oak trees evolved to make nuts of a size and taste suitable for
squirrels, which we’ve all seen burying, digging up, and eating acorns. Oaks
grow from the occasional acorn that a squirrel forgets to dig up. With billions
of squirrels each spreading hundreds of acorns every year to virtually any spot
suitable for oak trees to grow, we humans didn’t stand a chance of selecting
oaks for the acorns that we wanted. Those same problems of slow growth and
fast squirrels probably also explain why beech and hickory trees, heavily
exploited as wild trees for their nuts by Europeans and Native Americans,
respectively, were also not domesticated.
Finally, perhaps the most important difference between almonds and
acorns is that bitterness is controlled by a single dominant gene in almonds
but appears to be controlled by many genes in oaks. If ancient farmers planted
almonds or acorns from the occasional nonbitter mutant tree, the laws of
genetics dictate that half of the nuts from the resulting tree growing up would
also be nonbitter in the case of almonds, but almost all would still be bitter in
the case of oaks. That alone would kill the enthusiasm of any would-be acorn
farmer who had defeated the squirrels and remained patient.
As for strawberries and raspberries, we had similar trouble competing
with thrushes and other berry-loving birds. Yes, the Romans did tend wild
strawberries in their gardens. But with billions of European thrushes
defecating wild strawberry seeds in every possible place (including Roman
gardens), strawberries remained the little berries that thrushes wanted, not the
big berries that humans wanted. Only with the recent development of
protective nets and greenhouses were we finally able to defeat the thrushes,
and to redesign strawberries and raspberries according to our own standards.
WE’VE THUS SEEN that the difference between gigantic supermarket
strawberries and tiny wild ones is just one example of the various features
distinguishing cultivated plants from their wild ancestors. Those differences
arose initially from natural variation among the wild plants themselves. Some
of it, such as the variation in berry size or in nut bitterness, would have been
readily noticed by ancient farmers. Other variation, such as that in seed
dispersal mechanisms or seed dormancy, would have gone unrecognized by
humans before the rise of modern botany. But whether or not the selection of
wild edible plants by ancient hikers relied on conscious or unconscious
criteria, the resulting evolution of wild plants into crops was at first an
unconscious process. It followed inevitably from our selecting among wild
plant individuals, and from competition among plant individuals in gardens
favoring individuals different from those favored in the wild.
That’s why Darwin, in his great book On the Origin of Species, didn’t
start with an account of natural selection. His first chapter is instead a lengthy
account of how our domesticated plants and animals arose through artificial
selection by humans. Rather than discussing the Galápagos Island birds that
we usually associate with him, Darwin began by discussing—how farmers
develop varieties of gooseberries! He wrote, “I have seen great surprise
expressed in horticultural works at the wonderful skill of gardeners, in having
produced such splendid results from such poor materials; but the art has been
simple, and as far as the final result is concerned, has been followed almost
unconsciously. It has consisted in always cultivating the best-known variety,
sowing its seeds, and, when a slightly better variety chanced to appear,
selecting it, and so onwards.” Those principles of crop development by
artificial selection still serve as our most understandable model of the origin
of species by natural selection.
CHAPTER 8
APPLES OR INDIANS
WE HAVE JUST SEEN HOW PEOPLES OF SOME REGIONS began to cultivate wild
plant species, a step with momentous unforeseen consequences for their
lifestyle and their descendants’ place in history. Let us now return to our
questions: Why did agriculture never arise independently in some fertile and
highly suitable areas, such as California, Europe, temperate Australia, and
subequatorial Africa? Why, among the areas where agriculture did arise
independently, did it develop much earlier in some than in others?
Two contrasting explanations suggest themselves: problems with the local
people, or problems with the locally available wild plants. On the one hand,
perhaps almost any well-watered temperate or tropical area of the globe offers
enough species of wild plants suitable for domestication. In that case, the
explanation for agriculture’s failure to develop in some of those areas would
lie with cultural characteristics of their peoples. On the other hand, perhaps at
least some humans in any large area of the globe would have been receptive
to the experimentation that led to domestication. Only the lack of suitable
wild plants might then explain why food production did not evolve in some
areas.
As we shall see in the next chapter, the corresponding problem for
domestication of big wild mammals proves easier to solve, because there are
many fewer species of them than of plants. The world holds only about 148
species of large wild mammalian terrestrial herbivores or omnivores, the large
mammals that could be considered candidates for domestication. Only a
modest number of factors determines whether a mammal is suitable for
domestication. It’s thus straightforward to review a region’s big mammals and
to test whether the lack of mammal domestication in some regions was due to
the unavailability of suitable wild species, rather than to local peoples.
That approach would be much more difficult to apply to plants because of
the sheer number—200,000—of species of wild flowering plants, the plants
that dominate vegetation on the land and that have furnished almost all of our
crops. We can’t possibly hope to examine all the wild plant species of even a
circumscribed area like California, and to assess how many of them would
have been domesticable. But we shall now see how to get around that
problem.
WHEN ONE HEARS that there are so many species of flowering plants, one’s
first reaction might be as follows: surely, with all those wild plant species on
Earth, any area with a sufficiently benign climate must have had more than
enough species to provide plenty of candidates for crop development.
But then reflect that the vast majority of wild plants are unsuitable for
obvious reasons: they are woody, they produce no edible fruit, and their
leaves and roots are also inedible. Of the 200,000 wild plant species, only a
few thousand are eaten by humans, and just a few hundred of these have been
more or less domesticated. Even of these several hundred crops, most provide
minor supplements to our diet and would not by themselves have sufficed to
support the rise of civilizations. A mere dozen species account for over 80
percent of the modern world’s annual tonnage of all crops. Those dozen
blockbusters are the cereals wheat, corn, rice, barley, and sorghum; the pulse
soybean; the roots or tubers potato, manioc, and sweet potato; the sugar
sources sugarcane and sugar beet; and the fruit banana. Cereal crops alone
now account for more than half of the calories consumed by the world’s
human populations. With so few major crops in the world, all of them
domesticated thousands of years ago, it’s less surprising that many areas of
the world had no wild native plants at all of outstanding potential. Our failure
to domesticate even a single major new food plant in modern times suggests
that ancient peoples really may have explored virtually all useful wild plants
and domesticated all the ones worth domesticating.
Yet some of the world’s failures to domesticate wild plants remain hard to
explain. The most flagrant cases concern plants that were domesticated in one
area but not in another. We can thus be sure that it was indeed possible to
develop the wild plant into a useful crop, and we have to ask why that wild
species was not domesticated in certain areas.
A typical puzzling example comes from Africa. The important cereal
sorghum was domesticated in Africa’s Sahel zone, just south of the Sahara. It
also occurs as a wild plant as far south as southern Africa, yet neither it nor
any other plant was cultivated in southern Africa until the arrival of the whole
crop package that Bantu farmers brought from Africa north of the equator
2,000 years ago. Why did the native peoples of southern Africa not
domesticate sorghum for themselves?
Equally puzzling is the failure of people to domesticate flax in its wild
range in western Europe and North Africa, or einkorn wheat in its wild range
in the southern Balkans. Since these two plants were among the first eight
crops of the Fertile Crescent, they were presumably among the most readily
domesticated of all wild plants. They were adopted for cultivation in those
areas of their wild range outside the Fertile Crescent as soon as they arrived
with the whole package of food production from the Fertile Crescent. Why,
then, had peoples of those outlying areas not already begun to grow them of
their own accord?
Similarly, the four earliest domesticated fruits of the Fertile Crescent all
had wild ranges stretching far beyond the eastern Mediterranean, where they
appear to have been first domesticated: the olive, grape, and fig occurred west
to Italy and Spain and Northwest Africa, while the date palm extended to all
of North Africa and Arabia. These four were evidently among the easiest to
domesticate of all wild fruits. Why did peoples outside the Fertile Crescent
fail to domesticate them, and begin to grow them only when they had already
been domesticated in the eastern Mediterranean and arrived thence as crops?
Other striking examples involve wild species that were not domesticated
in areas where food production never arose spontaneously, even though those
wild species had close relatives domesticated elsewhere. For example, the
olive Olea europea was domesticated in the eastern Mediterranean. There are
about 40 other species of olives in tropical and southern Africa, southern
Asia, and eastern Australia, some of them closely related to Olea europea, but
none of them was ever domesticated. Similarly, while a wild apple species
and a wild grape species were domesticated in Eurasia, there are many related
wild apple and grape species in North America, some of which have in
modern times been hybridized with the crops derived from their wild Eurasian
counterparts in order to improve those crops. Why, then, didn’t Native
Americans domesticate those apparently useful apples and grapes
themselves?
One can go on and on with such examples. But there is a fatal flaw in this
reasoning: plant domestication is not a matter of hunter-gatherers’
domesticating a single plant and otherwise carrying on unchanged with their
nomadic lifestyle. Suppose that North American wild apples really would
have evolved into a terrific crop if only Indian hunter-gatherers had settled
down and cultivated them. But nomadic hunter-gatherers would not throw
over their traditional way of life, settle in villages, and start tending apple
orchards unless many other domesticable wild plants and animals were
available to make a sedentary food-producing existence competitive with a
hunting-gathering existence.
How, in short, do we assess the potential of an entire local flora for
domestication? For those Native Americans who failed to domesticate North
American apples, did the problem really lie with the Indians or with the
apples?
In order to answer this question, we shall now compare three regions that
lie at opposite extremes among centers of independent domestication. As we
have seen, one of them, the Fertile Crescent, was perhaps the earliest center of
food production in the world, and the site of origin of several of the modern
world’s major crops and almost all of its major domesticated animals. The
other two regions, New Guinea and the eastern United States, did domesticate
local crops, but these crops were very few in variety, only one of them gained
worldwide importance, and the resulting food package failed to support
extensive development of human technology and political organization as in
the Fertile Crescent. In the light of this comparison, we shall ask: Did the
flora and environment of the Fertile Crescent have clear advantages over
those of New Guinea and the eastern United States?
ONE OF THE central facts of human history is the early importance of the part
of Southwest Asia known as the Fertile Crescent (because of the crescent-like
shape of its uplands on a map: see Figure 8.1). That area appears to have been
the earliest site for a whole string of developments, including cities, writing,
empires, and what we term (for better or worse) civilization. All those
developments sprang, in turn, from the dense human populations, stored food
surpluses, and feeding of nonfarming specialists made possible by the rise of
food production in the form of crop cultivation and animal husbandry. Food
production was the first of those major innovations to appear in the Fertile
Crescent. Hence any attempt to understand the origins of the modern world
must come to grips with the question why the Fertile Crescent’s domesticated
plants and animals gave it such a potent head start.
Fortunately, the Fertile Crescent is by far the most intensively studied and
best understood part of the globe as regards the rise of agriculture. For most
crops domesticated in or near the Fertile Crescent, the wild plant ancestor has
been identified; its close relationship to the crop has been proven by genetic
and chromosomal studies; its wild geographic range is known; its changes
under domestication have been identified and are often understood at the level
of single genes; those changes can be observed in successive layers of the
archaeological record; and the approximate place and time of domestication
are known. I don’t deny that other areas, notably China, also had advantages
as early sites of domestication, but those advantages and the resulting
development of crops can be specified in much more detail for the Fertile
Crescent.
One advantage of the Fertile Crescent is that it lies within a zone of so-
called Mediterranean climate, a climate characterized by mild, wet winters
and long, hot, dry summers. That climate selects for plant species able to
survive the long dry season and to resume growth rapidly upon the return of
the rains. Many Fertile Crescent plants, especially species of cereals and
pulses, have adapted in a way that renders them useful to humans: they are
annuals, meaning that the plant itself dries up and dies in the dry season.
Within their mere one year of life, annual plants inevitably remain small
herbs. Many of them instead put much of their energy into producing big
seeds, which remain dormant during the dry season and are then ready to
sprout when the rains come. Annual plants therefore waste little energy on
making inedible wood or fibrous stems, like the body of trees and bushes. But
many of the big seeds, notably those of the annual cereals and pulses, are
edible by humans. They constitute 6 of the modern world’s 12 major crops. In
contrast, if you live near a forest and look out your window, the plant species
that you see will tend to be trees and shrubs, most of whose body you cannot
eat and which put much less of their energy into edible seeds. Of course,
some forest trees in areas of wet climate do produce big edible seeds, but
these seeds are not adapted to surviving a long dry season and hence to long
storage by humans.
A second advantage of the Fertile Crescent flora is that the wild ancestors
of many Fertile Crescent crops were already abundant and highly productive,
occurring in large stands whose value must have been obvious to hunter-
gatherers. Experimental studies in which botanists have collected seeds from
such natural stands of wild cereals, much as hunter-gatherers must have been
doing over 10,000 years ago, show that annual harvests of up to nearly a ton
of seeds per hectare can be obtained, yielding 50 kilocalories of food energy
for only one kilocalorie of work expended. By collecting huge quantities of
wild cereals in a short time when the seeds were ripe, and storing them for use
as food through the rest of the year, some hunting-gathering peoples of the
Fertile Crescent had already settled down in permanent villages even before
they began to cultivate plants.
Since Fertile Crescent cereals were so productive in the wild, few
additional changes had to be made in them under cultivation. As we discussed
in the preceding chapter, the principal changes—the breakdown of the natural
systems of seed dispersal and of germination inhibition—evolved
automatically and quickly as soon as humans began to cultivate the seeds in
fields. The wild ancestors of our wheat and barley crops look so similar to the
crops themselves that the identity of the ancestor has never been in doubt.
Because of this ease of domestication, big-seeded annuals were the first, or
among the first, crops developed not only in the Fertile Crescent but also in
China and the Sahel.
Contrast this quick evolution of wheat and barley with the story of corn,
the leading cereal crop of the New World. Corn’s probable ancestor, a wild
plant known as teosinte, looks so different from corn in its seed and flower
structures that even its role as ancestor has been hotly debated by botanists for
a long time. Teosinte’s value as food would not have impressed hunter-
gatherers: it was less productive in the wild than wild wheat, it produced
much less seed than did the corn eventually developed from it, and it enclosed
its seeds in inedible hard coverings. For teosinte to become a useful crop, it
had to undergo drastic changes in its reproductive biology, to increase greatly
its investment in seeds, and to lose those rock-like coverings of its seeds.
Archaeologists are still vigorously debating how many centuries or millennia
of crop development in the Americas were required for ancient corn cobs to
progress from a tiny size up to the size of a human thumb, but it seems clear
that several thousand more years were then required for them to reach modern
sizes. That contrast between the immediate virtues of wheat and barley and
the difficulties posed by teosinte may have been a significant factor in the
differing developments of New World and Eurasian human societies.
A third advantage of the Fertile Crescent flora is that it includes a high
percentage of hermaphroditic “selfers”—that is, plants that usually pollinate
themselves but that are occasionally cross-pollinated. Recall that most wild
plants either are regularly cross-pollinated hermaphrodites or consist of
separate male and female individuals that inevitably depend on another
individual for pollination. Those facts of reproductive biology vexed early
farmers, because, as soon as they had located a productive mutant plant, its
offspring would cross-breed with other plant individuals and thereby lose
their inherited advantage. As a result, most crops belong to the small
percentage of wild plants that either are hermaphrodites usually pollinating
themselves or else reproduce without sex by propagating vegetatively (for
example, by a root that genetically duplicates the parent plant). Thus, the high
percentage of hermaphroditic selfers in the Fertile Crescent flora aided early
farmers, because it meant that a high percentage of the wild flora had a
reproductive biology convenient for humans.
Selfers were also convenient for early farmers in that they occasionally
did become cross-pollinated, thereby generating new varieties among which
to select. That occasional cross-pollination occurred not only between
individuals of the same species, but also between related species to produce
interspecific hybrids. One such hybrid among Fertile Crescent selfers, bread
wheat, became the most valuable crop in the modern world.
Of the first eight significant crops to have been domesticated in the Fertile
Crescent, all were selfers. Of the three selfer cereals among them—einkorn
wheat, emmer wheat, and barley—the wheats offered the additional
advantage of a high protein content, 8–14 percent. In contrast, the most
important cereal crops of eastern Asia and of the New World—rice and corn,
respectively—had a lower protein content that posed significant nutritional
problems.
THOSE WERE SOME of the advantages that the Fertile Crescent’s flora
afforded the first farmers: it included an unusually high percentage of wild
plants suitable for domestication. However, the Mediterranean climate zone
of the Fertile Crescent extends westward through much of southern Europe
and northwestern Africa. There are also zones of similar Mediterranean
climates in four other parts of the world: California, Chile, southwestern
Australia, and South Africa (Figure 8.2). Yet those other Mediterranean zones
not only failed to rival the Fertile Crescent as early sites of food production;
they never gave rise to indigenous agriculture at all. What advantage did that
particular Mediterranean zone of western Eurasia enjoy?
It turns out that it, and especially its Fertile Crescent portion, possessed at
least five advantages over other Mediterranean zones. First, western Eurasia
has by far the world’s largest zone of Mediterranean climate. As a result, it
has a high diversity of wild plant and animal species, higher than in the
comparatively tiny Mediterranean zones of southwestern Australia and Chile.
Second, among Mediterranean zones, western Eurasia’s experiences the
greatest climatic variation from season to season and year to year. That
variation favored the evolution, among the flora, of an especially high
percentage of annual plants. The combination of these two factors—a high
diversity of species and a high percentage of annuals—means that western
Eurasia’s Mediterranean zone is the one with by far the highest diversity of
annuals.
The significance of that botanical wealth for humans is illustrated by the
geographer Mark Blumler’s studies of wild grass distributions. Among the
world’s thousands of wild grass species, Blumler tabulated the 56 with the
largest seeds, the cream of nature’s crop: the grass species with seeds at least
10 times heavier than the median grass species (see Table 8.1). Virtually all of
them are native to Mediterranean zones or other seasonally dry environments.
Furthermore, they are overwhelmingly concentrated in the Fertile Crescent or
other parts of western Eurasia’s Mediterranean zone, which offered a huge
selection to incipient farmers: about 32 of the world’s 56 prize wild grasses!
Specifically, barley and emmer wheat, the two earliest important crops of the
Fertile Crescent, rank respectively 3rd and 13th in seed size among those top
56. In contrast, the Mediterranean zone of Chile offered only two of those
species, California and southern Africa just one each, and southwestern
Australia none at all. That fact alone goes a long way toward explaining the
course of human history.
A third advantage of the Fertile Crescent’s Mediterranean zone is that it
provides a wide range of altitudes and topographies within a short distance.
Its range of elevations, from the lowest spot on Earth (the Dead Sea) to
mountains of 18,000 feet (near Teheran), ensures a corresponding variety of
environments, hence a high diversity of the wild plants serving as potential
ancestors of crops. Those mountains are in proximity to gentle lowlands with
rivers, flood plains, and deserts suitable for irrigation agriculture. In contrast,
the Mediterranean zones of southwestern Australia and, to a lesser degree, of
South Africa and western Europe offer a narrower range of altitudes, habitats,
and topographies.
TABLE 8.1 World Distribution of Large-Seeded Grass Species
Area
Number of Species
West Asia, Europe, North Africa
33
Mediterranean zone
32
England
1
East Asia
6
Sub-Saharan Africa
4
Americas
11
North America
4
Mesoamerica
5
South America
2
Northern Australia
2
Total:
56
Table 12.1 of Mark Blumler’s Ph.D. dissertation, “Seed Weight and
Environment in Mediterranean-type Grasslands in California and Israel”
(University of California, Berkeley, 1992), listed the world’s 56 heaviest-
seeded wild grass species (excluding bamboos) for which data were available.
Grain weight in those species ranged from 10 milligrams to over 40
milligrams, about 10 times greater than the median value for all of the world’s
grass species. Those 56 species make up less than 1 percent of the world’s
grass species. This table shows that these prize grasses are overwhelmingly
concentrated in the Mediterranean zone of western Eurasia.
The range of altitudes in the Fertile Crescent meant staggered harvest
seasons: plants at higher elevations produced seeds somewhat later than
plants at lower elevations. As a result, hunter-gatherers could move up a
mountainside harvesting grain seeds as they matured, instead of being
overwhelmed by a concentrated harvest season at a single altitude, where all
grains matured simultaneously. When cultivation began, it was a simple
matter for the first farmers to take the seeds of wild cereals growing on
hillsides and dependent on unpredictable rains, and to plant those seeds in the
damp valley bottoms, where they would grow reliably and be less dependent
on rain.
The Fertile Crescent’s biological diversity over small distances
contributed to a fourth advantage—its wealth in ancestors not only of
valuable crops but also of domesticated big mammals. As we shall see, there
were few or no wild mammal species suitable for domestication in the other
Mediterranean zones of California, Chile, southwestern Australia, and South
Africa. In contrast, four species of big mammals—the goat, sheep, pig, and
cow—were domesticated very early in the Fertile Crescent, possibly earlier
than any other animal except the dog anywhere else in the world. Those
species remain today four of the world’s five most important domesticated
mammals (Chapter 9). But their wild ancestors were commonest in slightly
different parts of the Fertile Crescent, with the result that the four species
were domesticated in different places: sheep possibly in the central part, goats
either in the eastern part at higher elevations (the Zagros Mountains of Iran)
or in the southwestern part (the Levant), pigs in the north-central part, and
cows in the western part, including Anatolia. Nevertheless, even though the
areas of abundance of these four wild progenitors thus differed, all four lived
in sufficiently close proximity that they were readily transferred after
domestication from one part of the Fertile Crescent to another, and the whole
region ended up with all four species.
Agriculture was launched in the Fertile Crescent by the early
domestication of eight crops, termed “founder crops” (because they founded
agriculture in the region and possibly in the world). Those eight founders
were the cereals emmer wheat, einkorn wheat, and barley; the pulses lentil,
pea, chickpea, and bitter vetch; and the fiber crop flax. Of these eight, only
two, flax and barley, range in the wild at all widely outside the Fertile
Crescent and Anatolia. Two of the founders had very small ranges in the wild,
chickpea being confined to southeastern Turkey and emmer wheat to the
Fertile Crescent itself. Thus, agriculture could arise in the Fertile Crescent
from domestication of locally available wild plants, without having to wait for
the arrival of crops derived from wild plants domesticated elsewhere.
Conversely, two of the eight founder crops could not have been domesticated
anywhere in the world except in the Fertile Crescent, since they did not occur
wild elsewhere.
Thanks to this availability of suitable wild mammals and plants, early
peoples of the Fertile Crescent could quickly assemble a potent and balanced
biological package for intensive food production. That package comprised
three cereals, as the main carbohydrate sources; four pulses, with 20–25
percent protein, and four domestic animals, as the main protein sources,
supplemented by the generous protein content of wheat; and flax as a source
of fiber and oil (termed linseed oil: flax seeds are about 40 percent oil).
Eventually, thousands of years after the beginnings of animal domestication
and food production, the animals also began to be used for milk, wool,
plowing, and transport. Thus, the crops and animals of the Fertile Crescent’s
first farmers came to meet humanity’s basic economic needs: carbohydrate,
protein, fat, clothing, traction, and transport.
A final advantage of early food production in the Fertile Crescent is that it
may have faced less competition from the hunter-gatherer lifestyle than that in
some other areas, including the western Mediterranean. Southwest Asia has
few large rivers and only a short coastline, providing relatively meager
aquatic resources (in the form of river and coastal fish and shellfish). One of
the important mammal species hunted for meat, the gazelle, originally lived in
huge herds but was overexploited by the growing human population and
reduced to low numbers. Thus, the food production package quickly became
superior to the hunter-gatherer package. Sedentary villages based on cereals
were already in existence before the rise of food production and predisposed
those hunter-gatherers to agriculture and herding. In the Fertile Crescent the
transition from hunting-gathering to food production took place relatively
fast: as late as 9000 B.C. people still had no crops and domestic animals and
were entirely dependent on wild foods, but by 6000 B.C. some societies were
almost completely dependent on crops and domestic animals.
The situation in Mesoamerica contrasts strongly: that area provided only
two domesticable animals (the turkey and the dog), whose meat yield was far
lower than that of cows, sheep, goats, and pigs; and corn, Mesoamerica’s
staple grain, was, as I’ve already explained, difficult to domesticate and
perhaps slow to develop. As a result, domestication may not have begun in
Mesoamerica until around 3500 B.C. (the date remains very uncertain); those
first developments were undertaken by people who were still nomadic hunter-
gatherers; and settled villages did not arise there until around 1500 B.C.
IN ALL THIS discussion of the Fertile Crescent’s advantages for the early rise
of food production, we have not had to invoke any supposed advantages of
Fertile Crescent peoples themselves. Indeed, I am unaware of anyone’s even
seriously suggesting any supposed distinctive biological features of the
region’s peoples that might have contributed to the potency of its food
production package. Instead, we have seen that the many distinctive features
of the Fertile Crescent’s climate, environment, wild plants, and animals
together provide a convincing explanation.
Since the food production packages arising indigenously in New Guinea
and in the eastern United States were considerably less potent, might the
explanation there lie with the peoples of those areas? Before turning to those
regions, however, we must consider two related questions arising in regard to
any area of the world where food production never developed independently
or else resulted in a less potent package. First, do hunter-gatherers and
incipient farmers really know well all locally available wild species and their
uses, or might they have overlooked potential ancestors of valuable crops?
Second, if they do know their local plants and animals, do they exploit that
knowledge to domesticate the most useful available species, or do cultural
factors keep them from doing so?
As regards the first question, an entire field of science, termed
ethnobiology, studies peoples’ knowledge of the wild plants and animals in
their environment. Such studies have concentrated especially on the world’s
few surviving hunting-gathering peoples, and on farming peoples who still
depend heavily on wild foods and natural products. The studies generally
show that such peoples are walking encyclopedias of natural history, with
individual names (in their local language) for as many as a thousand or more
plant and animal species, and with detailed knowledge of those species’
biological characteristics, distribution, and potential uses. As people become
increasingly dependent on domesticated plants and animals, this traditional
knowledge gradually loses its value and becomes lost, until one arrives at
modern supermarket shoppers who could not distinguish a wild grass from a
wild pulse.
Here’s a typical example. For the last 33 years, while conducting
biological exploration in New Guinea, I have been spending my field time
there constantly in the company of New Guineans who still use wild plants
and animals extensively. One day, when my companions of the Foré tribe and
I were starving in the jungle because another tribe was blocking our return to
our supply base, a Foré man returned to camp with a large rucksack full of
mushrooms he had found, and started to roast them. Dinner at last! But then I
had an unsettling thought: what if the mushrooms were poisonous?
I patiently explained to my Foré companions that I had read about some
mushrooms’ being poisonous, that I had heard of even expert American
mushroom collectors’ dying because of the difficulty of distinguishing safe
from dangerous mushrooms, and that although we were all hungry, it just
wasn’t worth the risk. At that point my companions got angry and told me to
shut up and listen while they explained some things to me. After I had been
quizzing them for years about names of hundreds of trees and birds, how
could I insult them by assuming they didn’t have names for different
mushrooms? Only Americans could be so stupid as to confuse poisonous
mushrooms with safe ones. They went on to lecture me about 29 types of
edible mushroom species, each species’ name in the Foré language, and where
in the forest one should look for it. This one, the tánti, grew on trees, and it
was delicious and perfectly edible.
Whenever I have taken New Guineans with me to other parts of their
island, they regularly talk about local plants and animals with other New
Guineans whom they meet, and they gather potentially useful plants and bring
them back to their home villages to try planting them. My experiences with
New Guineans are paralleled by those of ethnobiologists studying traditional
peoples elsewhere. However, all such peoples either practice at least some
food production or are the partly acculturated last remnants of the world’s
former hunter-gatherer societies. Knowledge of wild species was presumably
even more detailed before the rise of food production, when everyone on
Earth still depended entirely on wild species for food. The first farmers were
heirs to that knowledge, accumulated through tens of thousands of years of
nature observation by biologically modern humans living in intimate
dependence on the natural world. It therefore seems extremely unlikely that
wild species of potential value would have escaped the notice of the first
farmers.
The other, related question is whether ancient hunter-gatherers and
farmers similarly put their ethnobiological knowledge to good use in selecting
wild plants to gather and eventually to cultivate. One test comes from an
archaeological site at the edge of the Euphrates Valley in Syria, called Tell
Abu Hureyra. Between 10,000 and 9000 B.C. the people living there may
already have been residing year-round in villages, but they were still hunter-
gatherers; crop cultivation began only in the succeeding millennium. The
archaeologists Gordon Hillman, Susan Colledge, and David Harris retrieved
large quantities of charred plant remains from the site, probably representing
discarded garbage of wild plants gathered elsewhere and brought to the site
by its residents. The scientists analyzed over 700 samples, each containing an
average of over 500 identifiable seeds belonging to over 70 plant species. It
turned out that the villagers were collecting a prodigious variety (157
species!) of plants identified by their charred seeds, not to mention other
plants that cannot now be identified.
Were those naive villagers collecting every type of seed plant that they
found, bringing it home, poisoning themselves on most of the species, and
nourishing themselves from only a few species? No, they were not so silly.
While 157 species sounds like indiscriminate collecting, many more species
growing wild in the vicinity were absent from the charred remains. The 157
selected species fall into three categories. Many of them have seeds that are
nonpoisonous and immediately edible. Others, such as pulses and members of
the mustard family, have toxic seeds, but the toxins are easily removed,
leaving the seeds edible. A few seeds belong to species traditionally used as
sources of dyes or medicine. The many wild species not represented among
the 157 selected are ones that would have been useless or harmful to people,
including all of the most toxic weed species in the environment.
Thus, the hunter-gatherers of Tell Abu Hureyra were not wasting time and
endangering themselves by collecting wild plants indiscriminately. Instead,
they evidently knew the local wild plants as intimately as do modern New
Guineans, and they used that knowledge to select and bring home only the
most useful available seed plants. But those gathered seeds would have
constituted the material for the unconscious first steps of plant domestication.
My other example of how ancient peoples apparently used their
ethnobiological knowledge to good effect comes from the Jordan Valley in the
ninth millennium B.C., the period of the earliest crop cultivation there. The
valley’s first domesticated cereals were barley and emmer wheat, which are
still among the world’s most productive crops today. But, as at Tell Abu
Hureyra, hundreds of other seed-bearing wild plant species must have grown
in the vicinity, and a hundred or more of them would have been edible and
gathered before the rise of plant domestication. What was it about barley and
emmer wheat that caused them to be the first crops? Were those first Jordan
Valley farmers botanical ignoramuses who didn’t know what they were
doing? Or were barley and emmer wheat actually the best of the local wild
cereals that they could have selected?
Two Israeli scientists, Ofer Bar-Yosef and Mordechai Kislev, tackled this
question by examining the wild grass species still growing wild in the valley
today. Leaving aside species with small or unpalatable seeds, they picked out
23 of the most palatable and largest-seeded wild grasses. Not surprisingly,
barley and emmer wheat were on that list.
But it wasn’t true that the 21 other candidates would have been equally
useful. Among those 23, barley and emmer wheat proved to be the best by
many criteria. Emmer wheat has the biggest seeds and barley the second
biggest. In the wild, barley is one of the 4 most abundant of the 23 species,
while emmer wheat is of medium abundance. Barley has the further
advantage that its genetics and morphology permit it to evolve quickly the
useful changes in seed dispersal and germination inhibition that we discussed
in the preceding chapter. Emmer wheat, however, has compensating virtues: it
can be gathered more efficiently than barley, and it is unusual among cereals
in that its seeds do not adhere to husks. As for the other 21 species, their
drawbacks include smaller seeds, in many cases lower abundance, and in
some cases their being perennial rather than annual plants, with the
consequence that they would have evolved only slowly under domestication.
Thus, the first farmers in the Jordan Valley selected the 2 very best of the
23 best wild grass species available to them. Of course, the evolutionary
changes (following cultivation) in seed dispersal and germination inhibition
would have been unforeseen consequences of what those first farmers were
doing. But their initial selection of barley and emmer wheat rather than other
cereals to collect, bring home, and cultivate would have been conscious and
based on the easily detected criteria of seed size, palatability, and abundance.
This example from the Jordan Valley, like that from Tell Abu Hureyra,
illustrates that the first farmers used their detailed knowledge of local species
to their own benefit. Knowing far more about local plants than all but a
handful of modern professional botanists, they would hardly have failed to
cultivate any useful wild plant species that was comparably suitable for
domestication.
WE CAN NOW examine what local farmers, in two parts of the world (New
Guinea and the eastern United States) with indigenous but apparently
deficient food production systems compared to that of the Fertile Crescent,
actually did when more-productive crops arrived from elsewhere. If it turned
out that such crops did not become adopted for cultural or other reasons, we
would be left with a nagging doubt. Despite all our reasoning so far, we
would still have to suspect that the local wild flora harbored some ancestor of
a potential valuable crop that local farmers failed to exploit because of similar
cultural factors. These two examples will also demonstrate in detail a fact
critical to history: that indigenous crops from different parts of the globe were
not equally productive.
New Guinea, the largest island in the world after Greenland, lies just
north of Australia and near the equator. Because of its tropical location and
great diversity in topography and habitats, New Guinea is rich in both plant
and animal species, though less so than continental tropical areas because it is
an island. People have been living in New Guinea for at least 40,000 years—
much longer than in the Americas, and slightly longer than anatomically
modern peoples have been living in western Europe. Thus, New Guineans
have had ample opportunity to get to know their local flora and fauna. Were
they motivated to apply this knowledge to developing food production?
I mentioned already that the adoption of food production involved a
competition between the food producing and the hunting-gathering lifestyles.
Hunting-gathering is not so rewarding in New Guinea as to remove the
motivation to develop food production. In particular, modern New Guinea
hunters suffer from the crippling disadvantage of a dearth of wild game: there
is no native land animal larger than a 100-pound flightless bird (the
cassowary) and a 50-pound kangaroo. Lowland New Guineans on the coast
do obtain much fish and shellfish, and some lowlanders in the interior still
live today as hunter-gatherers, subsisting especially on wild sago palms. But
no peoples still live as hunter-gatherers in the New Guinea highlands; all
modern highlanders are instead farmers who use wild foods only to
supplement their diets. When highlanders go into the forest on hunting trips,
they take along garden-grown vegetables to feed themselves. If they have the
misfortune to run out of those provisions, even they starve to death despite
their detailed knowledge of locally available wild foods. Since the hunting-
gathering lifestyle is thus nonviable in much of modern New Guinea, it comes
as no surprise that all New Guinea highlanders and most lowlanders today are
settled farmers with sophisticated systems of food production. Extensive,
formerly forested areas of the highlands were converted by traditional New
Guinea farmers to fenced, drained, intensively managed field systems
supporting dense human populations.
Archaeological evidence shows that the origins of New Guinea
agriculture are ancient, dating to around 7000 B.C. At those early dates all the
landmasses surrounding New Guinea were still occupied exclusively by
hunter-gatherers, so this ancient agriculture must have developed
independently in New Guinea. While unequivocal remains of crops have not
been recovered from those early fields, they are likely to have included some
of the same crops that were being grown in New Guinea at the time of
European colonization and that are now known to have been domesticated
locally from wild New Guinea ancestors. Foremost among these local
domesticates is the modern world’s leading crop, sugarcane, of which the
annual tonnage produced today nearly equals that of the number two and
number three crops combined (wheat and corn). Other crops of undoubted
New Guinea origin are a group of bananas known as Australimusa bananas,
the nut tree Canarium indicum, and giant swamp taro, as well as various
edible grass stems, roots, and green vegetables. The breadfruit tree and the
root crops yams and (ordinary) taro may also be New Guinean domesticates,
although that conclusion remains uncertain because their wild ancestors are
not confined to New Guinea but are distributed from New Guinea to
Southeast Asia. At present we lack evidence that could resolve the question
whether they were domesticated in Southeast Asia, as traditionally assumed,
or independently or even only in New Guinea.
However, it turns out that New Guinea’s biota suffered from three severe
limitations. First, no cereal crops were domesticated in New Guinea, whereas
several vitally important ones were domesticated in the Fertile Crescent,
Sahel, and China. In its emphasis instead on root and tree crops, New Guinea
carries to an extreme a trend seen in agricultural systems in other wet tropical
areas (the Amazon, tropical West Africa, and Southeast Asia), whose farmers
also emphasized root crops but did manage to come up with at least two
cereals (Asian rice and a giant-seeded Asian cereal called Job’s tears). A
likely reason for the failure of cereal agriculture to arise in New Guinea is a
glaring deficiency of the wild starting material: not one of the world’s 56
largest-seeded wild grasses is native there.
Second, the New Guinea fauna included no domesticable large mammal
species whatsoever. The sole domestic animals of modern New Guinea, the
pig and chicken and dog, arrived from Southeast Asia by way of Indonesia
within the last several thousand years. As a result, while New Guinea
lowlanders obtain protein from the fish they catch, New Guinea highland
farmer populations suffer from severe protein limitation, because the staple
crops that provide most of their calories (taro and sweet potato) are low in
protein. Taro, for example, consists of barely 1 percent protein, much worse
than even white rice, and far below the levels of the Fertile Crescent’s wheats
and pulses (8–14 percent and 20–25 percent protein, respectively).
Children in the New Guinea highlands have the swollen bellies
characteristic of a high-bulk but protein-deficient diet. New Guineans old and
young routinely eat mice, spiders, frogs, and other small animals that peoples
elsewhere with access to large domestic mammals or large wild game species
do not bother to eat. Protein starvation is probably also the ultimate reason
why cannibalism was widespread in traditional New Guinea highland
societies.
Finally, in former times New Guinea’s available root crops were limiting
for calories as well as for protein, because they do not grow well at the high
elevations where many New Guineans live today. Many centuries ago,
however, a new root crop of ultimately South American origin, the sweet
potato, reached New Guinea, probably by way of the Philippines, where it had
been introduced by Spaniards. Compared with taro and other presumably
older New Guinea root crops, the sweet potato can be grown up to higher
elevations, grows more quickly, and gives higher yields per acre cultivated
and per hour of labor. The result of the sweet potato’s arrival was a highland
population explosion. That is, even though people had been farming in the
New Guinea highlands for many thousands of years before sweet potatoes
were introduced, the available local crops had limited them in the population
densities they could attain, and in the elevations they could occupy.
In short, New Guinea offers an instructive contrast to the Fertile Crescent.
Like hunter-gatherers of the Fertile Crescent, those of New Guinea did evolve
food production independently. However, their indigenous food production
was restricted by the local absence of domesticable cereals, pulses, and
animals, by the resulting protein deficiency in the highlands, and by
limitations of the locally available root crops at high elevations. Yet New
Guineans themselves know as much about the wild plants and animals
available to them as any peoples on Earth today. They can be expected to
have discovered and tested any wild plant species worth domesticating. They
are perfectly capable of recognizing useful additions to their crop larder, as is
shown by their exuberant adoption of the sweet potato when it arrived. That
same lesson is being driven home again in New Guinea today, as those tribes
with preferential access to introduced new crops and livestock (or with the
cultural willingness to adopt them) expand at the expense of tribes without
that access or willingness. Thus, the limits on indigenous food production in
New Guinea had nothing to do with New Guinea peoples, and everything
with the New Guinea biota and environment.
OUR OTHER EXAMPLE of indigenous agriculture apparently constrained by
the local flora comes from the eastern United States. Like New Guinea, that
area supported independent domestication of local wild plants. However,
early developments are much better understood for the eastern United States
than for New Guinea: the crops grown by the earliest farmers have been
identified, and the dates and crop sequences of local domestication are
known. Well before other crops began to arrive from elsewhere, Native
Americans settled in eastern U.S. river valleys and developed intensified food
production based on local crops. Hence they were in a position to take
advantage of the most promising wild plants. Which ones did they actually
cultivate, and how did the resulting local crop package compare with the
Fertile Crescent’s founder package?
It turns out that the eastern U.S. founder crops were four plants
domesticated in the period 2500–1500 B.C., a full 6,000 years after wheat and
barley domestication in the Fertile Crescent. A local species of squash
provided small containers, as well as yielding edible seeds. The remaining
three founders were grown solely for their edible seeds (sunflower, a daisy
relative called sumpweed, and a distant relative of spinach called goosefoot).
But four seed crops and a container fall far short of a complete food
production package. For 2,000 years those founder crops served only as minor
dietary supplements while eastern U.S. Native Americans continued to
depend mainly on wild foods, especially wild mammals and waterbirds, fish,
shellfish, and nuts. Farming did not supply a major part of their diet until the
period 500–200 B.C., after three more seed crops (knotweed, maygrass, and
little barley) had been brought into cultivation.
A modern nutritionist would have applauded those seven eastern
U.S.crops. All of them were high in protein—17–32 percent, compared with
8–14 percent for wheat, 9 percent for corn, and even lower for barley and
white rice. Two of them, sunflower and sumpweed, were also high in oil (45–
47 percent). Sumpweed, in particular, would have been a nutritionist’s
ultimate dream, being 32 percent protein and 45 percent oil. Why aren’t we
still eating those dream foods today?
Alas, despite their nutritional advantage, most of these eastern U.S. crops
suffered from serious disadvantages in other respects. Goosefoot, knotweed,
little barley, and maygrass had tiny seeds, with volumes only one-tenth that of
wheat and barley seeds. Worse yet, sumpweed is a wind-pollinated relative of
ragweed, the notorious hayfever-causing plant. Like ragweed’s, sumpweed’s
pollen can cause hayfever where the plant occurs in abundant stands. If that
doesn’t kill your enthusiasm for becoming a sumpweed farmer, be aware that
it has a strong odor objectionable to some people and that handling it can
cause skin irritation.
Mexican crops finally began to reach the eastern United States by trade
routes after A.D. 1. Corn arrived around A.D. 200, but its role remained very
minor for many centuries. Finally, around A.D. 900 a new variety of corn
adapted to North America’s short summers appeared, and the arrival of beans
around A.D. 1100 completed Mexico’s crop trinity of corn, beans, and squash.
Eastern U.S. farming became greatly intensified, and densely populated
chiefdoms developed along the Mississippi River and its tributaries. In some
areas the original local domesticates were retained alongside the far more
productive Mexican trinity, but in other areas the trinity replaced them
completely. No European ever saw sumpweed growing in Indian gardens,
because it had disappeared as a crop by the time that European colonization of
the Americas began, in A.D. 1492. Among all those ancient eastern U.S. crop
specialties, only two (sunflower and eastern squash) have been able to
compete with crops domesticated elsewhere and are still grown today. Our
modern acorn squashes and summer squashes are derived from those
American squashes domesticated thousands of years ago.
Thus, like the case of New Guinea, that of the eastern United States is
instructive. A priori, the region might have seemed a likely one to support
productive indigenous agriculture. It has rich soils, reliable moderate rainfall,
and a suitable climate that sustains bountiful agriculture today. The flora is a
species-rich one that includes productive wild nut trees (oak and hickory).
Local Native Americans did develop an agriculture based on local
domesticates, did thereby support themselves in villages, and even developed
a cultural florescence (the Hopewell culture centered on what is today Ohio)
around 200 B.C.–A.D. 400. They were thus in a position for several thousand
years to exploit as potential crops the most useful available wild plants,
whatever those should be.
Nevertheless, the Hopewell florescence sprang up nearly 9,000 years after
the rise of village living in the Fertile Crescent. Still, it was not until after A.D.
900 that the assembly of the Mexican crop trinity triggered a larger population
boom, the so-called Mississippian florescence, which produced the largest
towns and most complex societies achieved by Native Americans north of
Mexico. But that boom came much too late to prepare Native Americans of
the United States for the impending disaster of European colonization. Food
production based on eastern U.S. crops alone had been insufficient to trigger
the boom, for reasons that are easy to specify. The area’s available wild
cereals were not nearly as useful as wheat and barley. Native Americans of
the eastern United States domesticated no locally available wild pulse, no
fiber crop, no fruit or nut tree. They had no domesticated animals at all except
for dogs, which were probably domesticated elsewhere in the Americas.
It’s also clear that Native Americans of the eastern United States were not
overlooking potential major crops among the wild species around them. Even
20th-century plant breeders, armed with all the power of modern science,
have had little success in exploiting North American wild plants. Yes, we
have now domesticated pecans as a nut tree and blueberries as a fruit, and we
have improved some Eurasian fruit crops (apples, plums, grapes, raspberries,
blackberries, strawberries) by hybridizing them with North American wild
relatives. However, those few successes have changed our food habits far less
than Mexican corn changed food habits of Native Americans in the eastern
United States after A.D. 900.
The farmers most knowledgeable about eastern U.S. domesticates, the
region’s Native Americans themselves, passed judgment on them by
discarding or deemphasizing them when the Mexican trinity arrived. That
outcome also demonstrates that Native Americans were not constrained by
cultural conservativism and were quite able to appreciate a good plant when
they saw it. Thus, as in New Guinea, the limitations on indigenous food
production in the eastern United States were not due to Native American
peoples themselves, but instead depended entirely on the American biota and
environment.
WE HAVE NOW considered examples of three contrasting areas, in all of
which food production did arise indigenously. The Fertile Crescent lies at one
extreme; New Guinea and the eastern United States lie at the opposite
extreme. Peoples of the Fertile Crescent domesticated local plants much
earlier. They domesticated far more species, domesticated far more productive
or valuable species, domesticated a much wider range of types of crops,
developed intensified food production and dense human populations more
rapidly, and as a result entered the modern world with more advanced
technology, more complex political organization, and more epidemic diseases
with which to infect other peoples.
We found that these differences between the Fertile Crescent, New
Guinea, and the eastern United States followed straightforwardly from the
differing suites of wild plant and animal species available for domestication,
not from limitations of the peoples themselves. When more-productive crops
arrived from elsewhere (the sweet potato in New Guinea, the Mexican trinity
in the eastern United States), local peoples promptly took advantage of them,
intensified food production, and increased greatly in population. By
extension, I suggest that areas of the globe where food production never
developed indigenously at all—California, Australia, the Argentine pampas,
western Europe, and so on—may have offered even less in the way of wild
plants and animals suitable for domestication than did New Guinea and the
eastern United States, where at least a limited food production did arise.
Indeed, Mark Blumler’s worldwide survey of locally available large-seeded
wild grasses mentioned in this chapter, and the worldwide survey of locally
available big mammals to be presented in the next chapter, agree in showing
that all those areas of nonexistent or limited indigenous food production were
deficient in wild ancestors of domesticable livestock and cereals.
Recall that the rise of food production involved a competition between
food production and hunting-gathering. One might therefore wonder whether
all these cases of slow or nonexistent rise of food production might instead
have been due to an exceptional local richness of resources to be hunted and
gathered, rather than to an exceptional availability of species suitable for
domestication. In fact, most areas where indigenous food production arose
late or not at all offered exceptionally poor rather than rich resources to
hunter-gatherers, because most large mammals of Australia and the Americas
(but not of Eurasia and Africa) had become extinct toward the end of the Ice
Ages. Food production would have faced even less competition from hunting-
gathering in these areas than it did in the Fertile Crescent. Hence these local
failures or limitations of food production cannot be attributed to competition
from bountiful hunting opportunities.
LEST THESE CONCLUSIONS be misinterpreted, we should end this chapter with
caveats against exaggerating two points: peoples’ readiness to accept better
crops and livestock, and the constraints imposed by locally available wild
plants and animals. Neither that readiness nor those constraints are absolute.
We’ve already discussed many examples of local peoples’ adopting more-
productive crops domesticated elsewhere. Our broad conclusion is that people
can recognize useful plants, would therefore have probably recognized better
local ones suitable for domestication if any had existed, and aren’t barred
from doing so by cultural conservatism or taboos. But a big qualifier must be
added to this sentence: “in the long run and over large areas.” Anyone
knowledgeable about human societies can cite innumerable examples of
societies that refused crops, livestock, and other innovations that would have
been productive.
Naturally, I don’t subscribe to the obvious fallacy that every society
promptly adopts every innovation that would be useful for it. The fact is that,
over entire continents and other large areas containing hundreds of competing
societies, some societies will be more open to innovation, and some will be
more resistant. The ones that do adopt new crops, livestock, or technology
may thereby be enabled to nourish themselves better and to outbreed,
displace, conquer, or kill off societies resisting innovation. That’s an
important phenomenon whose manifestations extend far beyond the adoption
of new crops, and to which we shall return in Chapter 13.
Our other caveat concerns the limits that locally available wild species set
on the rise of food production. I’m not saying that food production could
never, in any amount of time, have arisen in all those areas where it actually
had not arisen indigenously by modern times. Europeans today who note that
Aboriginal Australians entered the modern world as Stone Age hunter-
gatherers often assume that the Aborigines would have gone on that way
forever.
To appreciate the fallacy, consider a visitor from Outer Space who
dropped in on Earth in the year 3000 B.C. The spaceling would have observed
no food production in the eastern United States, because food production did
not begin there until around 2500 B.C. Had the visitor of 3000 B.C. drawn the
conclusion that limitations posed by the wild plants and animals of the eastern
United States foreclosed food production there forever, events of the
subsequent millennium would have proved the visitor wrong. Even a visitor
to the Fertile Crescent in 9500 B.C. rather than in 8500 B.C. could have been
misled into supposing the Fertile Crescent permanently unsuitable for food
production.
That is, my thesis is not that California, Australia, western Europe, and all
the other areas without indigenous food production were devoid of
domesticable species and would have continued to be occupied just by hunter-
gatherers indefinitely if foreign domesticates or peoples had not arrived.
Instead, I note that regions differed greatly in their available pool of
domesticable species, that they varied correspondingly in the date when local
food production arose, and that food production had not yet arisen
independently in some fertile regions as of modern times.
Australia, supposedly the most “backward” continent, illustrates this point
well. In southeastern Australia, the well-watered part of the continent most
suitable for food production, Aboriginal societies in recent millennia appear
to have been evolving on a trajectory that would eventually have led to
indigenous food production. They had already built winter villages. They had
begun to manage their environment intensively for fish production by
building fish traps, nets, and even long canals. Had Europeans not colonized
Australia in 1788 and aborted that independent trajectory, Aboriginal
Australians might within a few thousand years have become food producers,
tending ponds of domesticated fish and growing domesticated Australian
yams and small-seeded grasses.
In that light, we can now answer the question implicit in the title of this
chapter. I asked whether the reason for the failure of North American Indians
to domesticate North American apples lay with the Indians or with the apples.
I’m not thereby implying that apples could never have been domesticated
in North America. Recall that apples were historically among the most
difficult fruit trees to cultivate and among the last major ones to be
domesticated in Eurasia, because their propagation requires the difficult
technique of grafting. There is no evidence for large-scale cultivation of
apples even in the Fertile Crescent and in Europe until classical Greek times,
8,000 years after the rise of Eurasian food production began. If Native
Americans had proceeded at the same rate in inventing or acquiring grafting
techniques, they too would eventually have domesticated apples—around the
year A.D. 5500, some 8,000 years after the rise of domestication in North
America around 2500 B.C.
Thus, the reason for the failure of Native Americans to domesticate North
American apples by the time Europeans arrived lay neither with the people
nor with the apples. As far as biological prerequisites for apple domestication
were concerned, North American Indian farmers were like Eurasian farmers,
and North American wild apples were like Eurasian wild apples. Indeed,
some of the supermarket apple varieties now being munched by readers of
this chapter have been developed recently by crossing Eurasian apples with
wild North American apples. Instead, the reason Native Americans did not
domesticate apples lay with the entire suite of wild plant and animal species
available to Native Americans. That suite’s modest potential for
domestication was responsible for the late start of food production in North
America.
CHAPTER 9
ZEBRAS, UNHAPPY MARRIAGES, AND THE ANNA KARENINA
PRINCIPLE
DOMESTICABLE ANIMALS ARE ALL ALIKE; EVERY UNDOMESTICABLE animal is
undomesticable in its own way.
If you think you’ve already read something like that before, you’re right.
Just make a few changes, and you have the famous first sentence of Tolstoy’s
great novel Anna Karenina: “Happy families are all alike; every unhappy
family is unhappy in its own way.” By that sentence, Tolstoy meant that, in
order to be happy, a marriage must succeed in many different respects: sexual
attraction, agreement about money, child discipline, religion, in-laws, and
other vital issues. Failure in any one of those essential respects can doom a
marriage even if it has all the other ingredients needed for happiness.
This principle can be extended to understanding much else about life
besides marriage. We tend to seek easy, single-factor explanations of success.
For most important things, though, success actually requires avoiding many
separate possible causes of failure. The Anna Karenina principle explains a
feature of animal domestication that had heavy consequences for human
history—namely, that so many seemingly suitable big wild mammal species,
such as zebras and peccaries, have never been domesticated and that the
successful domesticates were almost exclusively Eurasian. Having in the
preceding two chapters discussed why so many wild plant species seemingly
suitable for domestication were never domesticated, we shall now tackle the
corresponding question for domestic mammals. Our former question about
apples or Indians becomes a question of zebras or Africans.
IN CHAPTER 4 we reminded ourselves of the many ways in which big
domestic mammals were crucial to those human societies possessing them.
Most notably, they provided meat, milk products, fertilizer, land transport,
leather, military assault vehicles, plow traction, and wool, as well as germs
that killed previously unexposed peoples.
In addition, of course, small domestic mammals and domestic birds and
insects have also been useful to humans. Many birds were domesticated for
meat, eggs, and feathers: the chicken in China, various duck and goose
species in parts of Eurasia, turkeys in Mesoamerica, guinea fowl in Africa,
and the Muscovy duck in South America. Wolves were domesticated in
Eurasia and North America to become our dogs used as hunting companions,
sentinels, pets, and, in some societies, food. Rodents and other small
mammals domesticated for food included the rabbit in Europe, the guinea pig
in the Andes, a giant rat in West Africa, and possibly a rodent called the hutia
on Caribbean islands. Ferrets were domesticated in Europe to hunt rabbits,
and cats were domesticated in North Africa and Southwest Asia to hunt
rodent pests. Small mammals domesticated as recently as the 19th and 20th
centuries include foxes, mink, and chinchillas grown for fur and hamsters
kept as pets. Even some insects have been domesticated, notably Eurasia’s
honeybee and China’s silkworm moth, kept for honey and silk, respectively.
Many of these small animals thus yielded food, clothing, or warmth. But
none of them pulled plows or wagons, none bore riders, none except dogs
pulled sleds or became war machines, and none of them have been as
important for food as have big domestic mammals. Hence the rest of this
chapter will confine itself to the big mammals.
THE IMPORTANCE OF domesticated mammals rests on surprisingly few species
of big terrestrial herbivores. (Only terrestrial mammals have been
domesticated, for the obvious reason that aquatic mammals were difficult to
maintain and breed until the development of modern Sea World facilities.) If
one defines “big” as “weighing over 100 pounds,” then only 14 such species
were domesticated before the twentieth century (see Table 9.1 for a list). Of
those Ancient Fourteen, 9 (the “Minor Nine” of Table 9.1) became important
livestock for people in only limited areas of the globe: the Arabian camel,
Bactrian camel, llama / alpaca (distinct breeds of the same ancestral species),
donkey, reindeer, water buffalo, yak, banteng, and gaur. Only 5 species
became widespread and important around the world. Those Major Five of
mammal domestication are the cow, sheep, goat, pig, and horse.
This list may at first seem to have glaring omissions. What about the
African elephants with which Hannibal’s armies crossed the Alps? What
about the Asian elephants still used as work animals in Southeast Asia today?
No, I didn’t forget them, and that raises an important distinction. Elephants
have been tamed, but never domesticated. Hannibal’s elephants were, and
Asian work elephants are, just wild elephants that were captured and tamed;
they were not bred in captivity. In contrast, a domesticated animal is defined
as an animal selectively bred in captivity and thereby modified from its wild
ancestors, for use by humans who control the animal’s breeding and food
supply.
That is, domestication involves wild animals’ being transformed into
something more useful to humans. Truly domesticated animals differ in
various ways from their wild ancestors. These differences result from two
processes: human selection of those individual animals more useful to
humans than other individuals of the same species, and automatic
evolutionary responses of animals to the altered forces of natural selection
operating in human environments as compared with wild environments. We
already saw in Chapter 7 that all of these statements also apply to plant
domestication.
The ways in which domesticated animals have diverged from their wild
ancestors include the following. Many species changed in size: cows, pigs,
and sheep became smaller under domestication, while guinea pigs became
larger. Sheep and alpacas were selected for retention of wool and reduction or
loss of hair, while cows have been selected for high milk yields. Several
species of domestic animals have smaller brains and less developed sense
organs than their wild ancestors, because they no longer need the bigger
brains and more developed sense organs on which their ancestors depended to
escape from wild predators.
TABLE 9.1 The Ancient Fourteen Species of Big Herbivorous Domestic
Mammals
The Major Five
1. Sheep. Wild ancestor: the Asiatic mouflon sheep of West and
Central Asia. Now worldwide.
2. Goat. Wild ancestor: the bezoar goat of West Asia. Now
worldwide.
3. Cow, alias ox or cattle. Wild ancestor: the now extinct aurochs,
formerly distributed over Eurasia and North Africa. Now worldwide.
4. Pig. Wild ancestor: the wild boar, distributed over Eurasia and
North Africa. Now worldwide. Actually an omnivore (regularly eats
both animal and plant food), whereas the other 13 of the Ancient
Fourteen are more strictly herbivores.
5. Horse. Wild ancestor: now extinct wild horses of southern
Russia; a different subspecies of the same species survived in the wild
to modern times as Przewalski’s horse of Mongolia. Now worldwide.
The Minor Nine
6. Arabian (one-humped) camel. Wild ancestor: now extinct,
formerly lived in Arabia and adjacent areas. Still largely restricted to
Arabia and northern Africa, though feral in Australia.
7. Bactrian (two-humped) camel: Wild ancestor: now extinct,
lived in Central Asia. Still largely confined to Central Asia.
8. Llama and alpaca. These appear to be well-differentiated
breeds of the same species, rather than different species. Wild
ancestor: the guanaco of the Andes. Still largely confined to the
Andes, although some are bred as pack animals in North America.
9. Donkey. Wild ancestor: the African wild ass of North Africa and
formerly perhaps the adjacent area of Southwest Asia. Originally
confined as a domestic animal to North Africa and western Eurasia,
more recently also used elsewhere.
10. Reindeer. Wild ancestor: the reindeer of northern Eurasia. Still
largely confined as a domestic animal to that area, though now some
are also used in Alaska.
11. Water buffalo. Wild ancestor lives in Southeast Asia. Still used
as a domestic animal mainly in that area, though many are also used
in Brazil and others have escaped to the wild in Australia and other
places.
12. Yak. Wild ancestor: the wild yak of the Himalayas and Tibetan
plateau. Still confined as a domestic animal to that area.
13. Bali cattle. Wild ancestor: the banteng (a relative of the
aurochs) of Southeast Asia. Still confined as a domestic animal to that
area.
14. Mithan. Wild ancestor: the gaur (another relative of the
aurochs) of India and Burma. Still confined as a domestic animal to
that area.
To appreciate the changes that developed under domestication, just
compare wolves, the wild ancestors of domestic dogs, with the many breeds
of dogs. Some dogs are much bigger than wolves (Great Danes), while others
are much smaller (Pekingese). Some are slimmer and built for racing
(greyhounds), while others are short-legged and useless for racing
(dachshunds). They vary enormously in hair form and color, and some are
even hairless. Polynesians and Aztecs developed dog breeds specifically
raised for food. Comparing a dachshund with a wolf, you wouldn’t even
suspect that the former had been derived from the latter if you didn’t already
know it.
THE WILD ANCESTORS of the Ancient Fourteen were spread unevenly over the
globe. South America had only one such ancestor, which gave rise to the
llama and alpaca. North America, Australia, and sub-Saharan Africa had none
at all. The lack of domestic mammals indigenous to sub-Saharan Africa is
especially astonishing, since a main reason why tourists visit Africa today is
to see its abundant and diverse wild mammals. In contrast, the wild ancestors
of 13 of the Ancient Fourteen (including all of the Major Five) were confined
to Eurasia. (As elsewhere in this book, my use of the term “Eurasia” includes
in several cases North Africa, which biogeographically and in many aspects
of human culture is more closely related to Eurasia than to sub-Saharan
Africa.)
Of course, not all 13 of these wild ancestral species occurred together
throughout Eurasia. No area had all 13, and some of the wild ancestors were
quite local, such as the yak, confined in the wild to Tibet and adjacent
highland areas. However, many parts of Eurasia did have quite a few of these
13 species living together in the same area: for example, seven of the wild
ancestors occurred in Southwest Asia.
This very unequal distribution of wild ancestral species among the
continents became an important reason why Eurasians, rather than peoples of
other continents, were the ones to end up with guns, germs, and steel. How
can we explain the concentration of the Ancient Fourteen in Eurasia?
TABLE 9.2 Mammalian Candidates for Domestication
Continent
Eurasia Sub-Saharan Africa The Americas Australia
Candidates
72
51
24
1
Domesticated species
13
0
1
0
Percentage of Candidates domesticated
18%
0%
4%
0%
A “candidate” is defined as a species of terrestrial, herbivorous or omnivorous, wild mammal
weighing on the average over 100 pounds.
One reason is simple. Eurasia has the largest number of big terrestrial
wild mammal species, whether or not ancestral to a domesticated species.
Let’s define a “candidate for domestication” as any terrestrial herbivorous or
omnivorous mammal species (one not predominantly a carnivore) weighing
on the average over 100 pounds (45 kilograms). Table 9.2 shows that Eurasia
has the most candidates, 72 species, just as it has the most species in many
other plant and animal groups. That’s because Eurasia is the world’s largest
landmass, and it’s also very diverse ecologically, with habitats ranging from
extensive tropical rain forests, through temperate forests, deserts, and
marshes, to equally extensive tundras. Sub-Saharan Africa has fewer
candidates, 51 species, just as it has fewer species in most other plant and
animal groups—because it’s smaller and ecologically less diverse than
Eurasia. Africa has smaller areas of tropical rain forest than does Southeast
Asia, and no temperate habitats at all beyond latitude 37 degrees. As I
discussed in Chapter 1, the Americas may formerly have had almost as many
candidates as Africa, but most of America’s big wild mammals (including its
horses, most of its camels, and other species likely to have been domesticated
had they survived) became extinct about 13,000 years ago. Australia, the
smallest and most isolated continent, has always had far fewer species of big
wild mammals than has Eurasia, Africa, or the Americas. Just as in the
Americas, in Australia all of those few candidates except the red kangaroo
became extinct around the time of the continent’s first colonization by
humans.
Thus, part of the explanation for Eurasia’s having been the main site of
big mammal domestication is that it was the continent with the most
candidate species of wild mammals to start out with, and lost the fewest
candidates to extinction in the last 40,000 years. But the numbers in Table 9.2
warn us that that’s not the whole explanation. It’s also true that the percentage
of candidates actually domesticated is highest in Eurasia (18 percent), and is
especially low in sub-Saharan Africa (no species domesticated out of 51
candidates!). Particularly surprising is the large number of species of African
and American mammals that were never domesticated, despite their having
Eurasian close relatives or counterparts that were domesticated. Why were
Eurasia’s horses domesticated, but not Africa’s zebras? Why Eurasia’s pigs,
but not American peccaries or Africa’s three species of true wild pigs? Why
Eurasia’s five species of wild cattle (aurochs, water buffalo, yak, gaur,
banteng), but not the African buffalo or American bison? Why the Asian
mouflon sheep (ancestor of our domestic sheep), but not North American
bighorn sheep?
DID ALL THOSE peoples of Africa, the Americas, and Australia, despite their
enormous diversity, nonetheless share some cultural obstacles to
domestication not shared with Eurasian peoples? For example, did Africa’s
abundance of big wild mammals, available to kill by hunting, make it
superfluous for Africans to go to the trouble of tending domestic stock?
The answer to that question is unequivocal: No! The interpretation is
refuted by five types of evidence: rapid acceptance of Eurasian domesticates
by non-Eurasian peoples, the universal human penchant for keeping pets, the
rapid domestication of the Ancient Fourteen, the repeated independent
domestications of some of them, and the limited successes of modern efforts
at further domestications.
First, when Eurasia’s Major Five domestic mammals reached sub-Saharan
Africa, they were adopted by the most diverse African peoples wherever
conditions permitted. Those African herders thereby achieved a huge
advantage over African hunter-gatherers and quickly displaced them. In
particular, Bantu farmers who acquired cows and sheep spread out of their
homeland in West Africa and within a short time overran the former hunter-
gatherers in most of the rest of sub-Saharan Africa. Even without acquiring
crops, Khoisan peoples who acquired cows and sheep around 2,000 years ago
displaced Khoisan hunter-gatherers over much of southern Africa. The arrival
of the domestic horse in West Africa transformed warfare there and turned the
area into a set of kingdoms dependent on cavalry. The only factor that
prevented horses from spreading beyond West Africa was trypanosome
diseases borne by tsetse flies.
The same pattern repeated itself elsewhere in the world, whenever peoples
lacking native wild mammal species suitable for domestication finally had the
opportunity to acquire Eurasian domestic animals. European horses were
eagerly adopted by Native Americans in both North and South America,
within a generation of the escape of horses from European settlements. For
example, by the 19th century North America’s Great Plains Indians were
famous as expert horse-mounted warriors and bison hunters, but they did not
even obtain horses until the late 17th century. Sheep acquired from Spaniards
similarly transformed Navajo Indian society and led to, among other things,
the weaving of the beautiful woolen blankets for which the Navajo have
become renowned. Within a decade of Tasmania’s settlement by Europeans
with dogs, Aboriginal Tasmanians, who had never before seen dogs, began to
breed them in large numbers for use in hunting. Thus, among the thousands of
culturally diverse native peoples of Australia, the Americas, and Africa, no
universal cultural taboo stood in the way of animal domestication.
Surely, if some local wild mammal species of those continents had been
domesticable, some Australian, American, and African peoples would have
domesticated them and gained great advantage from them, just as they
benefited from the Eurasian domestic animals that they immediately adopted
when those became available. For instance, consider all the peoples of sub-
Saharan Africa living within the range of wild zebras and buffalo. Why
wasn’t there at least one African hunter-gatherer tribe that domesticated those
zebras and buffalo and that thereby gained sway over other Africans, without
having to await the arrival of Eurasian horses and cattle? All these facts
indicate that the explanation for the lack of native mammal domestication
outside Eurasia lay with the locally available wild mammals themselves, not
with the local peoples.
A SECOND TYPE of evidence for the same interpretation comes from pets.
Keeping wild animals as pets, and taming them, constitute an initial stage in
domestication. But pets have been reported from virtually all traditional
human societies on all continents. The variety of wild animals thus tamed is
far greater than the variety eventually domesticated, and includes some
species that we would scarcely have imagined as pets.
For example, in the New Guinea villages where I work, I often see people
with pet kangaroos, possums, and birds ranging from flycatchers to ospreys.
Most of these captives are eventually eaten, though some are kept just as pets.
New Guineans even regularly capture chicks of wild cassowaries (an ostrich-
like large, flightless bird) and raise them to eat as a delicacy—even though
captive adult cassowaries are extremely dangerous and now and then
disembowel village people. Some Asian peoples tame eagles for use in
hunting, although those powerful pets have also been known on occasion to
kill their human handlers. Ancient Egyptians and Assyrians, and modern
Indians, tamed cheetahs for use in hunting. Paintings made by ancient
Egyptians show that they further tamed (not surprisingly) hoofed mammals
such as gazelles and hartebeests, birds such as cranes, more surprisingly
giraffes (which can be dangerous), and most astonishingly hyenas. African
elephants were tamed in Roman times despite the obvious danger, and Asian
elephants are still being tamed today. Perhaps the most unlikely pet is the
European brown bear (the same species as the American grizzly bear), which
the Ainu people of Japan regularly captured as young animals, tamed, and
reared to kill and eat in a ritual ceremony.
Thus, many wild animal species reached the first stage in the sequence of
animal-human relations leading to domestication, but only a few emerged at
the other end of that sequence as domestic animals. Over a century ago, the
British scientist Francis Galton summarized this discrepancy succinctly: “It
would appear that every wild animal has had its chance of being
domesticated, that [a] few…were domesticated long ago, but that the large
remainder, who failed sometimes in only one small particular, are destined to
perpetual wildness.”
DATES OF DOMESTICATION provide a third line of evidence confirming
Galton’s view that early herding peoples quickly domesticated all big
mammal species suitable for being domesticated. All species for whose dates
of domestication we have archaeological evidence were domesticated
between about 8000 and 2500 B.C.—that is, within the first few thousand
years of the sedentary farming-herding societies that arose after the end of the
last Ice Age. As summarized in Table 9.3, the era of big mammal
domestication began with the sheep, goat, and pig and ended with camels.
Since 2500 B.C. there have been no significant additions.
It’s true, of course, that some small mammals were first domesticated
long after 2500 B.C. For example, rabbits were not domesticated for food until
the Middle Ages, mice and rats for laboratory research not until the 20th
century, and hamsters for pets not until the 1930s. The continuing
development of domesticated small mammals isn’t surprising, because there
are literally thousands of wild species as candidates, and because they were of
too little value to traditional societies to warrant the effort of raising them. But
big mammal domestication virtually ended 4,500 years ago. By then, all of
the world’s 148 candidate big species must have been tested innumerable
times, with the result that only a few passed the test and no other suitable ones
remained.
STILL A FOURTH line of evidence that some mammal species are much more
suitable than others is provided by the repeated independent domestications of
the same species. Genetic evidence based on the portions of our genetic
material known as mitochondrial DNA recently confirmed, as had long been
suspected, that humped cattle of India and humpless European cattle were
derived from two separate populations of wild ancestral cattle that had
diverged hundreds of thousands of years ago. That is, Indian peoples
domesticated the local Indian subspecies of wild aurochs, Southwest Asians
independently domesticated their own Southwest Asian subspecies of
aurochs, and North Africans may have independently domesticated the North
African aurochs.
Similarly, wolves were independently domesticated to become dogs in the
Americas and probably in several different parts of Eurasia, including China
and Southwest Asia. Modern pigs are derived from independent sequences of
domestication in China, western Eurasia, and possibly other areas as well.
These examples reemphasize that the same few suitable wild species attracted
the attention of many different human societies.
THE FAILURES OF modern efforts provide a final type of evidence that past
failures to domesticate the large residue of wild candidate species arose from
shortcomings of those species, rather than from shortcomings of ancient
humans. Europeans today are heirs to one of the longest traditions of animal
domestication on Earth—that which began in Southwest Asia around 10,000
years ago. Since the fifteenth century, Europeans have spread around the
globe and encountered wild mammal species not found in Europe. European
settlers, such as those that I encounter in New Guinea with pet kangaroos and
possums, have tamed or made pets of many local mammals, just as have
indigenous peoples. European herders and farmers emigrating to other
continents have also made serious efforts to domesticate some local species.
TABLE 9.3 Approximate Dates of First Attested Evidence for
Domestication of Large Mammal Species
Species
Date (B.C.)
Place
Dog
10,000
Southwest Asia, China, North America
Sheep
8,000
Southwest Asia
Goat
8,000
Southwest Asia
Pig
8,000
China, Southwest Asia
Cow
6,000
Southwest Asia, India, (?)North Africa
Horse
4,000
Ukraine
Donkey
4,000
Egypt
Water buffalo
4,000
China?
Llama / alpaca
3,500
Andes
Bactrian camel
2,500
Central Asia
Arabian camel
2,500
Arabia
For the other four domesticated large mammal species—reindeer, yak, gaur, and banteng—there
is as yet little evidence concerning the date of domestication. Dates and places shown are merely the earliest ones attested to date; domestication may actually have begun earlier and at a different location.
In the 19th and 20th centuries at least six large mammals—the eland, elk,
moose, musk ox, zebra, and American bison—have been the subjects of
especially well-organized projects aimed at domestication, carried out by
modern scientific animal breeders and geneticists. For example, eland, the
largest African antelope, have been undergoing selection for meat quality and
milk quantity in the Askaniya-Nova Zoological Park in the Ukraine, as well
as in England, Kenya, Zimbabwe, and South Africa; an experimental farm for
elk (red deer, in British terminology) has been operated by the Rowett
Research Institute at Aberdeen, Scotland; and an experimental farm for moose
has operated in the Pechero-Ilych National Park in Russia. Yet these modern
efforts have achieved only very limited successes. While bison meat
occasionally appears in some U.S. supermarkets, and while moose have been
ridden, milked, and used to pull sleds in Sweden and Russia, none of these
efforts has yielded a result of sufficient economic value to attract many
ranchers. It is especially striking that recent attempts to domesticate eland
within Africa itself, where its disease resistance and climate tolerance would
give it a big advantage over introduced Eurasian wild stock susceptible to
African diseases, have not caught on.
Thus, neither indigenous herders with access to candidate species over
thousands of years, nor modern geneticists, have succeeded in making useful
domesticates of large mammals beyond the Ancient Fourteen, which were
domesticated by at least 4,500 years ago. Yet scientists today could
undoubtedly, if they wished, fulfill for many species that part of the definition
of domestication that specifies the control of breeding and food supply. For
example, the San Diego and Los Angeles zoos are now subjecting the last
surviving California condors to a more draconian control of breeding than that
imposed upon any domesticated species. All individual condors have been
genetically identified, and a computer program determines which male shall
mate with which female in order to achieve human goals (in this case, to
maximize genetic diversity and thereby preserve this endangered bird). Zoos
are conducting similar breeding programs for many other threatened species,
including gorillas and rhinos. But the zoos’ rigorous selection of California
condors shows no prospects of yielding an economically useful product. Nor
do zoos’ efforts with rhinos, although rhinos offer up to over three tons of
meat on the hoof. As we shall now see, rhinos (and most other big mammals)
present insuperable obstacles to domestication.
IN ALL, OF the world’s 148 big wild terrestrial herbivorous mammals—the
candidates for domestication—only 14 passed the test. Why did the other 134
species fail? To which conditions was Francis Galton referring, when he
spoke of those other species as “destined to perpetual wildness”?
The answer follows from the Anna Karenina principle. To be
domesticated, a candidate wild species must possess many different
characteristics. Lack of any single required characteristic dooms efforts at
domestication, just as it dooms efforts at building a happy marriage. Playing
marriage counselor to the zebra / human couple and other ill-sorted pairs, we
can recognize at least six groups of reasons for failed domestication.
Diet. Every time that an animal eats a plant or another animal, the
conversion of food biomass into the consumer’s biomass involves an
efficiency of much less than 100 percent: typically around 10 percent. That is,
it takes around 10,000 pounds of corn to grow a 1,000-pound cow. If instead
you want to grow 1,000 pounds of carnivore, you have to feed it 10,000
pounds of herbivore grown on 100,000 pounds of corn. Even among
herbivores and omnivores, many species, like koalas, are too finicky in their
plant preferences to recommend themselves as farm animals.
As a result of this fundamental inefficiency, no mammalian carnivore has
ever been domesticated for food. (No, it’s not because its meat would be
tough or tasteless: we eat carnivorous wild fish all the time, and I can
personally attest to the delicious flavor of lion burger.) The nearest thing to an
exception is the dog, originally domesticated as a sentinel and hunting
companion, but breeds of dogs were developed and raised for food in Aztec
Mexico, Polynesia, and ancient China. However, regular dog eating has been
a last resort of meat-deprived human societies: the Aztecs had no other
domestic mammal, and the Polynesians and ancient Chinese had only pigs
and dogs. Human societies blessed with domestic herbivorous mammals have
not bothered to eat dogs, except as an uncommon delicacy (as in parts of
Southeast Asia today). In addition, dogs are not strict carnivores but
omnivores: if you are so naive as to think that your beloved pet dog is really a
meat eater, just read the list of ingredients on your bag of dog food. The dogs
that the Aztecs and Polynesians reared for food were efficiently fattened on
vegetables and garbage.
Growth Rate. To be worth keeping, domesticates must also grow quickly.
That eliminates gorillas and elephants, even though they are vegetarians with
admirably nonfinicky food preferences and represent a lot of meat. What
would-be gorilla or elephant rancher would wait 15 years for his herd to reach
adult size? Modern Asians who want work elephants find it much cheaper to
capture them in the wild and tame them.
Problems of Captive Breeding. We humans don’t like to have sex under
the watchful eyes of others; some potentially valuable animal species don’t
like to, either. That’s what derailed attempts to domesticate cheetahs, the
swiftest of all land animals, despite our strong motivation to do so for
thousands of years.
As I already mentioned, tame cheetahs were prized by ancient Egyptians
and Assyrians and modern Indians as hunting animals infinitely superior to
dogs. One Mogul emperor of India kept a stable of a thousand cheetahs. But
despite those large investments that many wealthy princes made, all of their
cheetahs were tamed ones caught in the wild. The princes’ efforts to breed
cheetahs in captivity failed, and not until 1960 did even biologists in modern
zoos achieve their first successful cheetah birth. In the wild, several cheetah
brothers chase a female for several days, and that rough courtship over large
distances seems to be required to get the female to ovulate or to become
sexually receptive. Cheetahs usually refuse to carry out that elaborate
courtship ritual inside a cage.
A similar problem has frustrated schemes to breed the vicuña, an Andean
wild camel whose wool is prized as the finest and lightest of any animal’s.
The ancient Incas obtained vicuña wool by driving wild vicuñas into corrals,
shearing them, and then releasing them alive. Modern merchants wanting this
luxury wool have had to resort either to this same method or simply to killing
wild vicuñas. Despite strong incentives of money and prestige, all attempts to
breed vicuñas for wool production in captivity have failed, for reasons that
include vicuñas’ long and elaborate courtship ritual before mating, a ritual
inhibited in captivity; male vicuñas’ fierce intolerance of each other; and their
requirement for both a year-round feeding territory and a separate year-round
sleeping territory.
Nasty Disposition. Naturally, almost any mammal species that is
sufficiently large is capable of killing a human. People have been killed by
pigs, horses, camels, and cattle. Nevertheless, some large animals have much
nastier dispositions and are more incurably dangerous than are others.
Tendencies to kill humans have disqualified many otherwise seemingly ideal
candidates for domestication.
One obvious example is the grizzly bear. Bear meat is an expensive
delicacy, grizzlies weigh up to 1,700 pounds, they are mainly vegetarians
(though also formidable hunters), their vegetable diet is very broad, they
thrive on human garbage (thereby creating big problems in Yellowstone and
Glacier National Parks), and they grow relatively fast. If they would behave
themselves in captivity, grizzlies would be a fabulous meat production
animal. The Ainu people of Japan made the experiment by routinely rearing
grizzly cubs as part of a ritual. For understandable reasons, though, the Ainu
found it prudent to kill and eat the cubs at the age of one year. Keeping
grizzly bears for longer would be suicidal; I am not aware of any adult that
has been tamed.
Another otherwise suitable candidate that disqualifies itself for equally
obvious reasons is the African buffalo. It grows quickly up to a weight of a
ton and lives in herds that have a well-developed dominance hierarchy, a trait
whose virtues will be discussed below. But the African buffalo is considered
the most dangerous and unpredictable large mammal of Africa. Anyone
insane enough to try to domesticate it either died in the effort or was forced to
kill the buffalo before it got too big and nasty. Similarly, hippos, as four-ton
vegetarians, would be great barnyard animals if they weren’t so dangerous.
They kill more people each year than do any other African mammals,
including even lions.
Few people would be surprised at the disqualification of those notoriously
ferocious candidates. But there are other candidates whose dangers are not so
well known. For instance, the eight species of wild equids (horses and their
relatives) vary greatly in disposition, even though all eight are genetically so
close to each other that they will interbreed and produce healthy (though
usually sterile) offspring. Two of them, the horse and the North African ass
(ancestor of the donkey), were successfully domesticated. Closely related to
the North African ass is the Asiatic ass, also known as the onager. Since its
homeland includes the Fertile Crescent, the cradle of Western civilization and
animal domestication, ancient peoples must have experimented extensively
with onagers. We know from Sumerian and later depictions that onagers were
regularly hunted, as well as captured and hybridized with donkeys and horses.
Some ancient depictions of horselike animals used for riding or for pulling
carts may refer to onagers. However, all writers about them, from Romans to
modern zookeepers, decry their irascible temper and their nasty habit of biting
people. As a result, although similar in other respects to ancestral donkeys,
onagers have never been domesticated.
Africa’s four species of zebras are even worse. Efforts at domestication
went as far as hitching them to carts: they were tried out as draft animals in
19th-century South Africa, and the eccentric Lord Walter Rothschild drove
through the streets of London in a carriage pulled by zebras. Alas, zebras
become impossibly dangerous as they grow older. (That’s not to deny that
many individual horses are also nasty, but zebras and onagers are much more
uniformly so.) Zebras have the unpleasant habit of biting a person and not
letting go. They thereby injure even more American zookeepers each year
than do tigers! Zebras are also virtually impossible to lasso with a rope—even
for cowboys who win rodeo championships by lassoing horses—because of
their unfailing ability to watch the rope noose fly toward them and then to
duck their head out of the way.
Hence it has rarely (if ever) been possible to saddle or ride a zebra, and
South Africans’ enthusiasm for their domestication waned. Unpredictably
aggressive behavior on the part of a large and potentially dangerous mammal
is also part of the reason why the initially so promising modern experiments
in domesticating elk and eland have not been more successful.
Tendency to Panic. Big mammalian herbivore species react to danger
from predators or humans in different ways. Some species are nervous, fast,
and programmed for instant flight when they perceive a threat. Other species
are slower, less nervous, seek protection in herds, stand their ground when
threatened, and don’t run until necessary. Most species of deer and antelope
(with the conspicuous exception of reindeer) are of the former type, while
sheep and goats are of the latter.
Naturally, the nervous species are difficult to keep in captivity. If put into
an enclosure, they are likely to panic, and either die of shock or batter
themselves to death against the fence in their attempts to escape. That’s true,
for example, of gazelles, which for thousands of years were the most
frequently hunted game species in some parts of the Fertile Crescent. There is
no mammal species that the first settled peoples of that area had more
opportunity to domesticate than gazelles. But no gazelle species has ever been
domesticated. Just imagine trying to herd an animal that bolts, blindly bashes
itself against walls, can leap up to nearly 30 feet, and can run at a speed of 50
miles per hour!
Social Structure. Almost all species of domesticated large mammals prove
to be ones whose wild ancestors share three social characteristics: they live in
herds; they maintain a well-developed dominance hierarchy among herd
members; and the herds occupy overlapping home ranges rather than
mutually exclusive territories. For example, herds of wild horses consist of
one stallion, up to half a dozen mares, and their foals. Mare A is dominant
over mares B, C, D, and E; mare B is submissive to A but dominant over C,
D, and E; C is submissive to B and A but dominant over D and E; and so on.
When the herd is on the move, its members maintain a stereotyped order: in
the rear, the stallion; in the front, the top-ranking female, followed by her
foals in order of age, with the youngest first; and behind her, the other mares
in order of rank, each followed by her foals in order of age. In that way, many
adults can coexist in the herd without constant fighting and with each
knowing its rank.
That social structure is ideal for domestication, because humans in effect
take over the dominance hierarchy. Domestic horses of a pack line follow the
human leader as they would normally follow the top-ranking female. Herds or
packs of sheep, goats, cows, and ancestral dogs (wolves) have a similar
hierarchy. As young animals grow up in such a herd, they imprint on the
animals that they regularly see nearby. Under wild conditions those are
members of their own species, but captive young herd animals also see
humans nearby and imprint on humans as well.
Such social animals lend themselves to herding. Since they are tolerant of
each other, they can be bunched up. Since they instinctively follow a
dominant leader and will imprint on humans as that leader, they can readily be
driven by a shepherd or sheepdog. Herd animals do well when penned in
crowded conditions, because they are accustomed to living in densely packed
groups in the wild.
In contrast, members of most solitary territorial animal species cannot be
herded. They do not tolerate each other, they do not imprint on humans, and
they are not instinctively submissive. Who ever saw a line of cats (solitary
and territorial in the wild) following a human or allowing themselves to be
herded by a human? Every cat lover knows that cats are not submissive to
humans in the way dogs instinctively are. Cats and ferrets are the sole
territorial mammal species that were domesticated, because our motive for
doing so was not to herd them in large groups raised for food but to keep
them as solitary hunters or pets.
While most solitary territorial species thus haven’t been domesticated, it’s
not conversely the case that most herd species can be domesticated. Most
can’t, for one of several additional reasons.
First, herds of many species don’t have overlapping home ranges but
instead maintain exclusive territories against other herds. It’s no more
possible to pen two such herds together than to pen two males of a solitary
species.
Second, many species that live in herds for part of the year are territorial
in the breeding season, when they fight and do not tolerate each other’s
presence. That’s true of most deer and antelope species (again with the
exception of reindeer), and it’s one of the main factors that has disqualified all
the social antelope species for which Africa is famous from being
domesticated. While one’s first association to African antelope is “vast dense
herds spreading across the horizon,” in fact the males of those herds space
themselves into territories and fight fiercely with each other when breeding.
Hence those antelope cannot be maintained in crowded enclosures in
captivity, as can sheep or goats or cattle. Territorial behavior similarly
combines with a fierce disposition and a slow growth rate to banish rhinos
from the farmyard.
Finally, many herd species, including again most deer and antelope, do
not have a well-defined dominance hierarchy and are not instinctively
prepared to become imprinted on a dominant leader (hence to become
misimprinted on humans). As a result, though many deer and antelope species
have been tamed (think of all those true Bambi stories), one never sees such
tame deer and antelope driven in herds like sheep. That problem also derailed
domestication of North American bighorn sheep, which belong to the same
genus as Asiatic mouflon sheep, ancestor of our domestic sheep. Bighorn
sheep are suitable to us and similar to mouflons in most respects except a
crucial one: they lack the mouflon’s stereotypical behavior whereby some
individuals behave submissively toward other individuals whose dominance
they acknowledge.
LET’S NOW RETURN to the problem I posed at the outset of this chapter.
Initially, one of the most puzzling features of animal domestication is the
seeming arbitrariness with which some species have been domesticated while
their close relatives have not. It turns out that all but a few candidates for
domestication have been eliminated by the Anna Karenina principle. Humans
and most animal species make an unhappy marriage, for one or more of many
possible reasons: the animal’s diet, growth rate, mating habits, disposition,
tendency to panic, and several distinct features of social organization. Only a
small percentage of wild mammal species ended up in happy marriages with
humans, by virtue of compatibility on all those separate counts.
Eurasian peoples happened to inherit many more species of domesticable
large wild mammalian herbivores than did peoples of the other continents.
That outcome, with all of its momentous advantages for Eurasian societies,
stemmed from three basic facts of mammalian geography, history, and
biology. First, Eurasia, befitting its large area and ecological diversity, started
out with the most candidates. Second, Australia and the Americas, but not
Eurasia or Africa, lost most of their candidates in a massive wave of late-
Pleistocene extinctions—possibly because the mammals of the former
continents had the misfortune to be first exposed to humans suddenly and late
in our evolutionary history, when our hunting skills were already highly
developed. Finally, a higher percentage of the surviving candidates proved
suitable for domestication on Eurasia than on the other continents. An
examination of the candidates that were never domesticated, such as Africa’s
big herd-forming mammals, reveals particular reasons that disqualified each
of them. Thus, Tolstoy would have approved of the insight offered in another
context by an earlier author, Saint Matthew: “Many are called, but few are
chosen.”
CHAPTER 10
SPACIOUS SKIES AND TILTED AXES
ON THE MAP OF THE WORLD ON CHAPTER 10 (FIGURE 10.1), compare the shapes
and orientations of the continents. You’ll be struck by an obvious difference.
The Americas span a much greater distance north–south (9,000 miles) than
east–west: only 3,000 miles at the widest, narrowing to a mere 40 miles at the
Isthmus of Panama. That is, the major axis of the Americas is north–south.
The same is also true, though to a less extreme degree, for Africa. In contrast,
the major axis of Eurasia is east–west. What effect, if any, did those
differences in the orientation of the continents’ axes have on human history?
This chapter will be about what I see as their enormous, sometimes tragic,
consequences. Axis orientations affected the rate of spread of crops and
livestock, and possibly also of writing, wheels, and other inventions. That
basic feature of geography thereby contributed heavily to the very different
experiences of Native Americans, Africans, and Eurasians in the last 500
years.
FOOD PRODUCTION’S SPREAD proves as crucial to understanding geographic
differences in the rise of guns, germs, and steel as did its origins, which we
considered in the preceding chapters. That’s because, as we saw in Chapter 5,
there were no more than nine areas of the globe, perhaps as few as five, where
food production arose independently. Yet, already in prehistoric times, food
production became established in many other regions besides those few areas
of origins. All those other areas became food producing as a result of the
spread of crops, livestock, and knowledge of how to grow them and, in some
cases, as a result of migrations of farmers and herders themselves.
The main such spreads of food production were from Southwest Asia to
Europe, Egypt and North Africa, Ethiopia, Central Asia, and the Indus Valley;
from the Sahel and West Africa to East and South Africa; from China to
tropical Southeast Asia, the Philippines, Indonesia, Korea, and Japan; and
from Mesoamerica to North America. Moreover, food production even in its
areas of origin became enriched by the addition of crops, livestock, and
techniques from other areas of origin.
Just as some regions proved much more suitable than others for the
origins of food production, the ease of its spread also varied greatly around
the world. Some areas that are ecologically very suitable for food production
never acquired it in prehistoric times at all, even though areas of prehistoric
food production existed nearby. The most conspicuous such examples are the
failure of both farming and herding to reach Native American California from
the U.S. Southwest or to reach Australia from New Guinea and Indonesia, and
the failure of farming to spread from South Africa’s Natal Province to South
Africa’s Cape. Even among all those areas where food production did spread
in the prehistoric era, the rates and dates of spread varied considerably. At the
one extreme was its rapid spread along east–west axes: from Southwest Asia
both west to Europe and Egypt and east to the Indus Valley (at an average rate
of about 0.7 miles per year); and from the Philippines east to Polynesia (at 3.2
miles per year). At the opposite extreme was its slow spread along north–
south axes: at less than 0.5 miles per year, from Mexico northward to the U.S.
Southwest; at less than 0.3 miles per year, for corn and beans from Mexico
northward to become productive in the eastern United States around A.D. 900;
and at 0.2 miles per year, for the llama from Peru north to Ecuador. These
differences could be even greater if corn was not domesticated in Mexico as
late as 3500 B.C., as I assumed conservatively for these calculations, and as
some archaeologists now assume, but if it was instead domesticated
considerably earlier, as most archaeologists used to assume (and many still
do).
There were also great differences in the completeness with which suites of
crops and livestock spread, again implying stronger or weaker barriers to their
spreading. For instance, while most of Southwest Asia’s founder crops and
livestock did spread west to Europe and east to the Indus Valley, neither of the
Andes’ domestic mammals (the llama / alpaca and the guinea pig) ever
reached Mesoamerica in pre-Columbian times. That astonishing failure cries
out for explanation. After all, Mesoamerica did develop dense farming
populations and complex societies, so there can be no doubt that Andean
domestic animals (if they had been available) would have been valuable for
food, transport, and wool. Except for dogs, Mesoamerica was utterly without
indigenous mammals to fill those needs. Some South American crops
nevertheless did succeed in reaching Mesoamerica, such as manioc, sweet
potatoes, and peanuts. What selective barrier let those crops through but
screened out llamas and guinea pigs?
A subtler expression of this geographically varying ease of spread is the
phenomenon termed preemptive domestication. Most of the wild plant species
from which our crops were derived vary genetically from area to area,
because alternative mutations had become established among the wild
ancestral populations of different areas. Similarly, the changes required to
transform wild plants into crops can in principle be brought about by
alternative new mutations or alternative courses of selection to yield
equivalent results. In this light, one can examine a crop widespread in
prehistoric times and ask whether all of its varieties show the same wild
mutation or same transforming mutation. The purpose of this examination is
to try to figure out whether the crop was developed in just one area or else
independently in several areas.
If one carries out such a genetic analysis for major ancient New World
crops, many of them prove to include two or more of those alternative wild
variants, or two or more of those alternative transforming mutations. This
suggests that the crop was domesticated independently in at least two
different areas, and that some varieties of the crop inherited the particular
mutation of one area while other varieties of the same crop inherited the
mutation of another area. On this basis, botanists conclude that lima beans
( Phaseolus lunatus), common beans ( Phaseolus vulgaris), and chili peppers
of the Capsicum annuum / chinense group were all domesticated on at least
two separate occasions, once in Mesoamerica and once in South America; and
that the squash Cucurbita pepo and the seed plant goosefoot were also
domesticated independently at least twice, once in Mesoamerica and once in
the eastern United States. In contrast, most ancient Southwest Asian crops
exhibit just one of the alternative wild variants or alternative transforming
mutations, suggesting that all modern varieties of that particular crop stem
from only a single domestication.
What does it imply if the same crop has been repeatedly and
independently domesticated in several different parts of its wild range, and
not just once and in a single area? We have already seen that plant
domestication involves the modification of wild plants so that they become
more useful to humans by virtue of larger seeds, a less bitter taste, or other
qualities. Hence if a productive crop is already available, incipient farmers
will surely proceed to grow it rather than start all over again by gathering its
not yet so useful wild relative and redomesticating it. Evidence for just a
single domestication thus suggests that, once a wild plant had been
domesticated, the crop spread quickly to other areas throughout the wild
plant’s range, preempting the need for other independent domestications of
the same plant. However, when we find evidence that the same wild ancestor
was domesticated independently in different areas, we infer that the crop
spread too slowly to preempt its domestication elsewhere. The evidence for
predominantly single domestications in Southwest Asia, but frequent multiple
domestications in the Americas, might thus provide more subtle evidence that
crops spread more easily out of Southwest Asia than in the Americas.
Rapid spread of a crop may preempt domestication not only of the same
wild ancestral species somewhere else but also of related wild species. If
you’re already growing good peas, it’s of course pointless to start from
scratch to domesticate the same wild ancestral pea again, but it’s also
pointless to domesticate closely related wild pea species that for farmers are
virtually equivalent to the already domesticated pea species. All of Southwest
Asia’s founder crops preempted domestication of any of their close relatives
throughout the whole expanse of western Eurasia. In contrast, the New World
presents many cases of equivalent and closely related, but nevertheless
distinct, species having been domesticated in Mesoamerica and South
America. For instance, 95 percent of the cotton grown in the world today
belongs to the cotton species Gossypium hirsutum, which was domesticated in
prehistoric times in Mesoamerica. However, prehistoric South American
farmers instead grew the related cotton Gossypium barbadense. Evidently,
Mesoamerican cotton had such difficulty reaching South America that it
failed in the prehistoric era to preempt the domestication of a different cotton
species there (and vice versa). Chili peppers, squashes, amaranths, and
chenopods are other crops of which different but related species were
domesticated in Mesoamerica and South America, since no species was able
to spread fast enough to preempt the others.
We thus have many different phenomena converging on the same
conclusion: that food production spread more readily out of Southwest Asia
than in the Americas, and possibly also than in sub-Saharan Africa. Those
phenomena include food production’s complete failure to reach some
ecologically suitable areas; the differences in its rate and selectivity of spread;
and the differences in whether the earliest domesticated crops preempted
redomestications of the same species or domestications of close relatives.
What was it about the Americas and Africa that made the spread of food
production more difficult there than in Eurasia?
TO ANSWER THIS question, let’s begin by examining the rapid spread of food
production out of Southwest Asia (the Fertile Crescent). Soon after food
production arose there, somewhat before 8000 B.C., a centrifugal wave of it
appeared in other parts of western Eurasia and North Africa farther and
farther removed from the Fertile Crescent, to the west and east. On this page I
have redrawn the striking map (Figure 10.2) assembled by the geneticist
Daniel Zohary and botanist Maria Hopf, in which they illustrate how the wave
had reached Greece and Cyprus and the Indian subcontinent by 6500 B.C.,
Egypt soon after 6000 B.C., central Europe by 5400 B.C., southern Spain by
5200 B.C., and Britain around 3500 B.C. In each of those areas, food production
was initiated by some of the same suite of domestic plants and animals that
launched it in the Fertile Crescent. In addition, the Fertile Crescent package
penetrated Africa southward to Ethiopia at some still-uncertain date.
However, Ethiopia also developed many indigenous crops, and we do not yet
know whether it was these crops or the arriving Fertile Crescent crops that
launched Ethiopian food production.
Of course, not all pieces of the package spread to all those outlying areas:
for example, Egypt was too warm for einkorn wheat to become established. In
some outlying areas, elements of the package arrived at different times: for
instance, sheep preceded cereals in southwestern Europe. Some outlying areas
went on to domesticate a few local crops of their own, such as poppies in
western Europe and watermelons possibly in Egypt. But most food production
in outlying areas depended initially on Fertile Crescent domesticates. Their
spread was soon followed by that of other innovations originating in or near
the Fertile Crescent, including the wheel, writing, metalworking techniques,
milking, fruit trees, and beer and wine production.
Why did the same plant package launch food production throughout
western Eurasia? Was it because the same set of plants occurred in the wild in
many areas, were found useful there just as in the Fertile Crescent, and were
independently domesticated? No, that’s not the reason. First, many of the
Fertile Crescent’s founder crops don’t even occur in the wild outside
Southwest Asia. For instance, none of the eight main founder crops except
barley grows wild in Egypt. Egypt’s Nile Valley provides an environment
similar to the Fertile Crescent’s Tigris and Euphrates Valleys. Hence the
package that worked well in the latter valleys also worked well enough in the
Nile Valley to trigger the spectacular rise of indigenous Egyptian civilization.
But the foods to fuel that spectacular rise were originally absent in Egypt. The
sphinx and pyramids were built by people fed on crops originally native to the
Fertile Crescent, not to Egypt.
Second, even for those crops whose wild ancestor does occur outside of
Southwest Asia, we can be confident that the crops of Europe and India were
mostly obtained from Southwest Asia and were not local domesticates. For
example, wild flax occurs west to Britain and Algeria and east to the Caspian
Sea, while wild barley occurs east even to Tibet. However, for most of the
Fertile Crescent’s founding crops, all cultivated varieties in the world today
share only one arrangement of chromosomes out of the multiple arrangements
found in the wild ancestor; or else they share only a single mutation (out of
many possible mutations) by which the cultivated varieties differ from the
wild ancestor in characteristics desirable to humans. For instance, all
cultivated peas share the same recessive gene that prevents ripe pods of
cultivated peas from spontaneously popping open and spilling their peas, as
wild pea pods do.
Evidently, most of the Fertile Crescent’s founder crops were never
domesticated again elsewhere after their initial domestication in the Fertile
Crescent. Had they been repeatedly domesticated independently, they would
exhibit legacies of those multiple origins in the form of varied chromosomal
arrangements or varied mutations. Hence these are typical examples of the
phenomenon of preemptive domestication that we discussed above. The quick
spread of the Fertile Crescent package preempted any possible other attempts,
within the Fertile Crescent or elsewhere, to domesticate the same wild
ancestors. Once the crop had become available, there was no further need to
gather it from the wild and thereby set it on the path to domestication again.
The ancestors of most of the founder crops have wild relatives, in the
Fertile Crescent and elsewhere, that would also have been suitable for
domestication. For example, peas belong to the genus Pisum, which consists
of two wild species: Pisum sativum, the one that became domesticated to
yield our garden peas, and Pisum fulvum, which was never domesticated. Yet
wild peas of Pisum fulvum taste good, either fresh or dried, and are common
in the wild. Similarly, wheats, barley, lentil, chickpea, beans, and flax all have
numerous wild relatives besides the ones that became domesticated. Some of
those related beans and barleys were indeed domesticated independently in
the Americas or China, far from the early site of domestication in the Fertile
Crescent. But in western Eurasia only one of several potentially useful wild
species was domesticated—probably because that one spread so quickly that
people soon stopped gathering the other wild relatives and ate only the crop.
Again as we discussed above, the crop’s rapid spread preempted any possible
further attempts to domesticate its relatives, as well as to redomesticate its
ancestor.
WHY WAS THE spread of crops from the Fertile Crescent so rapid? The
answer depends partly on that east–west axis of Eurasia with which I opened
this chapter. Localities distributed east and west of each other at the same
latitude share exactly the same day length and its seasonal variations. To a
lesser degree, they also tend to share similar diseases, regimes of temperature
and rainfall, and habitats or biomes (types of vegetation). For example,
Portugal, northern Iran, and Japan, all located at about the same latitude but
lying successively 4,000 miles east or west of each other, are more similar to
each other in climate than each is to a location lying even a mere 1,000 miles
due south. On all the continents the habitat type known as tropical rain forest
is confined to within about 10 degrees latitude of the equator, while
Mediterranean scrub habitats (such as California’s chaparral and Europe’s
maquis) lie between about 30 and 40 degrees of latitude.
But the germination, growth, and disease resistance of plants are adapted
to precisely those features of climate. Seasonal changes of day length,
temperature, and rainfall constitute signals that stimulate seeds to germinate,
seedlings to grow, and mature plants to develop flowers, seeds, and fruit.
Each plant population becomes genetically programmed, through natural
selection, to respond appropriately to signals of the seasonal regime under
which it has evolved. Those regimes vary greatly with latitude. For example,
day length is constant throughout the year at the equator, but at temperate
latitudes it increases as the months advance from the winter solstice to the
summer solstice, and it then declines again through the next half of the year.
The growing season—that is, the months with temperatures and day lengths
suitable for plant growth—is shortest at high latitudes and longest toward the
equator. Plants are also adapted to the diseases prevalent at their latitude.
Woe betide the plant whose genetic program is mismatched to the latitude
of the field in which it is planted! Imagine a Canadian farmer foolish enough
to plant a race of corn adapted to growing farther south, in Mexico. The
unfortunate corn plant, following its Mexico-adapted genetic program, would
prepare to thrust up its shoots in March, only to find itself still buried under
10 feet of snow. Should the plant become genetically reprogrammed so as to
germinate at a time more appropriate to Canada—say, late June—the plant
would still be in trouble for other reasons. Its genes would be telling it to
grow at a leisurely rate, sufficient only to bring it to maturity in five months.
That’s a perfectly safe strategy in Mexico’s mild climate, but in Canada a
disastrous one that would guarantee the plant’s being killed by autumn frosts
before it had produced any mature corn cobs. The plant would also lack genes
for resistance to diseases of northern climates, while uselessly carrying genes
for resistance to diseases of southern climates. All those features make low-
latitude plants poorly adapted to high-latitude conditions, and vice versa. As a
consequence, most Fertile Crescent crops grow well in France and Japan but
poorly at the equator.
Animals too are adapted to latitude-related features of climate. In that
respect we are typical animals, as we know by introspection. Some of us can’t
stand cold northern winters with their short days and characteristic germs,
while others of us can’t stand hot tropical climates with their own
characteristic diseases. In recent centuries overseas colonists from cool
northern Europe have preferred to emigrate to the similarly cool climates of
North America, Australia, and South Africa, and to settle in the cool
highlands within equatorial Kenya and New Guinea. Northern Europeans who
were sent out to hot tropical lowland areas used to die in droves of diseases
such as malaria, to which tropical peoples had evolved some genetic
resistance.
That’s part of the reason why Fertile Crescent domesticates spread west
and east so rapidly: they were already well adapted to the climates of the
regions to which they were spreading. For instance, once farming crossed
from the plains of Hungary into central Europe around 5400 B.C., it spread so
quickly that the sites of the first farmers in the vast area from Poland west to
Holland (marked by their characteristic pottery with linear decorations) were
nearly contemporaneous. By the time of Christ, cereals of Fertile Crescent
origin were growing over the 8,000-mile expanse from the Atlantic coast of
Ireland to the Pacific coast of Japan. That west–east expanse of Eurasia is the
largest land distance on Earth.
Thus, Eurasia’s west–east axis allowed Fertile Crescent crops quickly to
launch agriculture over the band of temperate latitudes from Ireland to the
Indus Valley, and to enrich the agriculture that arose independently in eastern
Asia. Conversely, Eurasian crops that were first domesticated far from the
Fertile Crescent but at the same latitudes were able to diffuse back to the
Fertile Crescent. Today, when seeds are transported over the whole globe by
ship and plane, we take it for granted that our meals are a geographic
mishmash. A typical American fast-food restaurant meal would include
chicken (first domesticated in China) and potatoes (from the Andes) or corn
(from Mexico), seasoned with black pepper (from India) and washed down
with a cup of coffee (of Ethiopian origin). Already, though, by 2,000 years
ago, Romans were also nourishing themselves with their own hodgepodge of
foods that mostly originated elsewhere. Of Roman crops, only oats and
poppies were native to Italy. Roman staples were the Fertile Crescent founder
package, supplemented by quince (originating in the Caucasus); millet and
cumin (domesticated in Central Asia); cucumber, sesame, and citrus fruit
(from India); and chicken, rice, apricots, peaches, and foxtail millet
(originally from China). Even though Rome’s apples were at least native to
western Eurasia, they were grown by means of grafting techniques that had
developed in China and spread westward from there.
While Eurasia provides the world’s widest band of land at the same
latitude, and hence the most dramatic example of rapid spread of
domesticates, there are other examples as well. Rivaling in speed the spread
of the Fertile Crescent package was the eastward spread of a subtropical
package that was initially assembled in South China and that received
additions on reaching tropical Southeast Asia, the Philippines, Indonesia, and
New Guinea. Within 1,600 years that resulting package of crops (including
bananas, taro, and yams) and domestic animals (chickens, pigs, and dogs) had
spread more than 5,000 miles eastward into the tropical Pacific to reach the
islands of Polynesia. A further likely example is the east–west spread of crops
within Africa’s wide Sahel zone, but paleobotanists have yet to work out the
details.
CONTRAST THE EASE of east–west diffusion in Eurasia with the difficulties of
diffusion along Africa’s north–south axis. Most of the Fertile Crescent
founder crops reached Egypt very quickly and then spread as far south as the
cool highlands of Ethiopia, beyond which they didn’t spread. South Africa’s
Mediterranean climate would have been ideal for them, but the 2,000 miles of
tropical conditions between Ethiopia and South Africa posed an insuperable
barrier. Instead, African agriculture south of the Sahara was launched by the
domestication of wild plants (such as sorghum and African yams) indigenous
to the Sahel zone and to tropical West Africa, and adapted to the warm
temperatures, summer rains, and relatively constant day lengths of those low
latitudes.
Similarly, the spread southward of Fertile Crescent domestic animals
through Africa was stopped or slowed by climate and disease, especially by
trypanosome diseases carried by tsetse flies. The horse never became
established farther south than West Africa’s kingdoms north of the equator.
The advance of cattle, sheep, and goats halted for 2,000 years at the northern
edge of the Serengeti Plains, while new types of human economies and
livestock breeds were being developed. Not until the period A.D. 1–200, some
8,000 years after livestock were domesticated in the Fertile Crescent, did
cattle, sheep, and goats finally reach South Africa. Tropical African crops had
their own difficulties spreading south in Africa, arriving in South Africa with
black African farmers (the Bantu) just after those Fertile Crescent livestock
did. However, those tropical African crops could never be transmitted across
South Africa’s Fish River, beyond which they were stopped by Mediterranean
conditions to which they were not adapted.
The result was the all-too-familiar course of the last two millennia of
South African history. Some of South Africa’s indigenous Khoisan peoples
(otherwise known as Hottentots and Bushmen) acquired livestock but
remained without agriculture. They became outnumbered and were replaced
northeast of the Fish River by black African farmers, whose southward spread
halted at that river. Only when European settlers arrived by sea in 1652,
bringing with them their Fertile Crescent crop package, could agriculture
thrive in South Africa’s Mediterranean zone. The collisions of all those
peoples produced the tragedies of modern South Africa: the quick decimation
of the Khoisan by European germs and guns; a century of wars between
Europeans and blacks; another century of racial oppression; and now, efforts
by Europeans and blacks to seek a new mode of coexistence in the former
Khoisan lands.
CONTRAST ALSO THE ease of diffusion in Eurasia with its difficulties along
the Americas’ north–south axis. The distance between Mesoamerica and
South America—say, between Mexico’s highlands and Ecuador’s—is only
1,200 miles, approximately the same as the distance in Eurasia separating the
Balkans from Mesopotamia. The Balkans provided ideal growing conditions
for most Mesopotamian crops and livestock, and received those domesticates
as a package within 2,000 years of its assembly in the Fertile Crescent. That
rapid spread preempted opportunities for domesticating those and related
species in the Balkans. Highland Mexico and the Andes would similarly have
been suitable for many of each other’s crops and domestic animals. A few
crops, notably Mexican corn, did indeed spread to the other region in the pre-
Columbian era.
But other crops and domestic animals failed to spread between
Mesoamerica and South America. The cool highlands of Mexico would have
provided ideal conditions for raising llamas, guinea pigs, and potatoes, all
domesticated in the cool highlands of the South American Andes. Yet the
northward spread of those Andean specialties was stopped completely by the
hot intervening lowlands of Central America. Five thousand years after llamas
had been domesticated in the Andes, the Olmecs, Maya, Aztecs, and all other
native societies of Mexico remained without pack animals and without any
edible domestic mammals except for dogs.
Conversely, domestic turkeys of Mexico and domestic sunflowers of the
eastern United States might have thrived in the Andes, but their southward
spread was stopped by the intervening tropical climates. The mere 700 miles
of north–south distance prevented Mexican corn, squash, and beans from
reaching the U.S. Southwest for several thousand years after their
domestication in Mexico, and Mexican chili peppers and chenopods never did
reach it in prehistoric times. For thousands of years after corn was
domesticated in Mexico, it failed to spread northward into eastern North
America, because of the cooler climates and shorter growing season
prevailing there. At some time between A.D. 1 and A.D. 200, corn finally
appeared in the eastern United States but only as a very minor crop. Not until
around A.D. 900, after hardy varieties of corn adapted to northern climates had
been developed, could corn-based agriculture contribute to the flowering of
the most complex Native American society of North America, the
Mississippian culture—a brief flowering ended by European-introduced
germs arriving with and after Columbus.
Recall that most Fertile Crescent crops prove, upon genetic study, to
derive from only a single domestication process, whose resulting crop spread
so quickly that it preempted any other incipient domestications of the same or
related species. In contrast, many apparently widespread Native American
crops prove to consist of related species or even of genetically distinct
varieties of the same species, independently domesticated in Mesoamerica,
South America, and the eastern United States. Closely related species replace
each other geographically among the amaranths, beans, chenopods, chili
peppers, cottons, squashes, and tobaccos. Different varieties of the same
species replace each other among the kidney beans, lima beans, the chili
pepper Capsicum annuum / chinense, and the squash Cucurbita pepo. Those
legacies of multiple independent domestications may provide further
testimony to the slow diffusion of crops along the Americas’ north–south axis.
Africa and the Americas are thus the two largest landmasses with a
predominantly north–south axis and resulting slow diffusion. In certain other
parts of the world, slow north–south diffusion was important on a smaller
scale. These other examples include the snail’s pace of crop exchange
between Pakistan’s Indus Valley and South India, the slow spread of South
Chinese food production into Peninsular Malaysia, and the failure of tropical
Indonesian and New Guinean food production to arrive in prehistoric times in
the modern farmlands of southwestern and southeastern Australia,
respectively. Those two corners of Australia are now the continent’s
breadbaskets, but they lie more than 2,000 miles south of the equator.
Farming there had to await the arrival from faraway Europe, on European
ships, of crops adapted to Europe’s cool climate and short growing season.
I HAVE BEEN dwelling on latitude, readily assessed by a glance at a map,
because it is a major determinant of climate, growing conditions, and ease of
spread of food production. However, latitude is of course not the only such
determinant, and it is not always true that adjacent places at the same latitude
have the same climate (though they do necessarily have the same day length).
Topographic and ecological barriers, much more pronounced on some
continents than on others, were locally important obstacles to diffusion.
For instance, crop diffusion between the U.S. Southeast and Southwest
was very slow and selective although these two regions are at the same
latitude. That’s because much of the intervening area of Texas and the
southern Great Plains was dry and unsuitable for agriculture. A corresponding
example within Eurasia involved the eastern limit of Fertile Crescent crops,
which spread rapidly westward to the Atlantic Ocean and eastward to the
Indus Valley without encountering a major barrier. However, farther eastward
in India the shift from predominantly winter rainfall to predominantly
summer rainfall contributed to a much more delayed extension of agriculture,
involving different crops and farming techniques, into the Ganges plain of
northeastern India. Still farther east, temperate areas of China were isolated
from western Eurasian areas with similar climates by the combination of the
Central Asian desert, Tibetan plateau, and Himalayas. The initial development
of food production in China was therefore independent of that at the same
latitude in the Fertile Crescent, and gave rise to entirely different crops.
However, even those barriers between China and western Eurasia were at
least partly overcome during the second millennium B.C., when West Asian
wheat, barley, and horses reached China.
By the same token, the potency of a 2,000-mile north–south shift as a
barrier also varies with local conditions. Fertile Crescent food production
spread southward over that distance to Ethiopia, and Bantu food production
spread quickly from Africa’s Great Lakes region south to Natal, because in
both cases the intervening areas had similar rainfall regimes and were suitable
for agriculture. In contrast, crop diffusion from Indonesia south to
southwestern Australia was completely impossible, and diffusion over the
much shorter distance from Mexico to the U.S. Southwest and Southeast was
slow, because the intervening areas were deserts hostile to agriculture. The
lack of a high-elevation plateau in Mesoamerica south of Guatemala, and
Mesoamerica’s extreme narrowness south of Mexico and especially in
Panama, were at least as important as the latitudinal gradient in throttling crop
and livestock exchanges between the highlands of Mexico and the Andes.
Continental differences in axis orientation affected the diffusion not only
of food production but also of other technologies and inventions. For
example, around 3,000 B.C. the invention of the wheel in or near Southwest
Asia spread rapidly west and east across much of Eurasia within a few
centuries, whereas the wheels invented independently in prehistoric Mexico
never spread south to the Andes. Similarly, the principle of alphabetic writing,
developed in the western part of the Fertile Crescent by 1500 B.C., spread west
to Carthage and east to the Indian subcontinent within about a thousand years,
but the Mesoamerican writing systems that flourished in prehistoric times for
at least 2,000 years never reached the Andes.
Naturally, wheels and writing aren’t directly linked to latitude and day
length in the way crops are. Instead, the links are indirect, especially via food
production systems and their consequences. The earliest wheels were parts of
ox-drawn carts used to transport agricultural produce. Early writing was
restricted to elites supported by food-producing peasants, and it served
purposes of economically and socially complex food-producing societies
(such as royal propaganda, goods inventories, and bureaucratic record
keeping). In general, societies that engaged in intense exchanges of crops,
livestock, and technologies related to food production were more likely to
become involved in other exchanges as well.
America’s patriotic song “America the Beautiful” invokes our spacious
skies, our amber waves of grain, from sea to shining sea. Actually, that song
reverses geographic realities. As in Africa, in the Americas the spread of
native crops and domestic animals was slowed by constricted skies and
environmental barriers. No waves of native grain ever stretched from the
Atlantic to the Pacific coast of North America, from Canada to Patagonia, or
from Egypt to South Africa, while amber waves of wheat and barley came to
stretch from the Atlantic to the Pacific across the spacious skies of Eurasia.
That faster spread of Eurasian agriculture, compared with that of Native
American and sub-Saharan African agriculture, played a role (as the next part
of this book will show) in the more rapid diffusion of Eurasian writing,
metallurgy, technology, and empires.
To bring up all those differences isn’t to claim that widely distributed
crops are admirable, or that they testify to the superior ingenuity of early
Eurasian farmers. They reflect, instead, the orientation of Eurasia’s axis
compared with that of the Americas or Africa. Around those axes turned the
fortunes of history.
PART THREE
FROM FOOD TO GUNS, GERMS, AND STEEL
CHAPTER 11
LETHAL GIFT OF LIVESTOCK
WE HAVE NOW TRACED HOW FOOD PRODUCTION AROSE in a few centers, and
how it spread at unequal rates from there to other areas. Those geographic
differences constitute important ultimate answers to Yali’s question about why
different peoples ended up with disparate degrees of power and affluence.
However, food production itself is not a proximate cause. In a one-on-one
fight, a naked farmer would have no advantage over a naked hunter-gatherer.
Instead, one part of the explanation for farmer power lies in the much
denser populations that food production could support: ten naked farmers
certainly would have an advantage over one naked hunter-gatherer in a fight.
The other part is that neither farmers nor hunter-gatherers are naked, at least
not figuratively. Farmers tend to breathe out nastier germs, to own better
weapons and armor, to own more-powerful technology in general, and to live
under centralized governments with literate elites better able to wage wars of
conquest. Hence the next four chapters will explore how the ultimate cause of
food production led to the proximate causes of germs, literacy, technology,
and centralized government.
The links connecting livestock and crops to germs were unforgettably
illustrated for me by a hospital case about which I learned through a physician
friend. When my friend was an inexperienced young doctor, he was called
into a hospital room to deal with a married couple stressed-out by a
mysterious illness. It did not help that the couple was also having difficulty
communicating with each other, and with my friend. The husband was a
small, timid man, sick with pneumonia caused by an unidentified microbe,
and with only limited command of the English language. Acting as translator
was his beautiful wife, worried about her husband’s condition and frightened
by the unfamiliar hospital environment. My friend was also stressed-out from
a long week of hospital work, and from trying to figure out what unusual risk
factors might have brought on the strange illness. The stress caused my friend
to forget everything he had been taught about patient confidentiality: he
committed the awful blunder of requesting the woman to ask her husband
whether he’d had any sexual experiences that could have caused the infection.
As the doctor watched, the husband turned red, pulled himself together so
that he seemed even smaller, tried to disappear under his bedsheets, and
stammered out words in a barely audible voice. His wife suddenly screamed
in rage and drew herself up to tower over him. Before the doctor could stop
her, she grabbed a heavy metal bottle, slammed it with full force onto her
husband’s head, and stormed out of the room. It took a while for the doctor to
revive her husband and even longer to elicit, through the man’s broken
English, what he’d said that so enraged his wife. The answer slowly emerged:
he had confessed to repeated intercourse with sheep on a recent visit to the
family farm; perhaps that was how he had contracted the mysterious microbe.
This incident sounds bizarrely one-of-a-kind and of no possible broader
significance. In fact, it illustrates an enormous subject of great importance:
human diseases of animal origins. Very few of us love sheep in the carnal
sense that this patient did. But most of us platonically love our pet animals,
such as our dogs and cats. As a society, we certainly appear to have an
inordinate fondness for sheep and other livestock, to judge from the vast
numbers of them that we keep. For example, at the time of a recent census,
Australia’s 17,085,400 people thought so highly of sheep that they kept
161,600,000 of them.
Some of us adults, and even more of our children, pick up infectious
diseases from our pets. Usually they remain no more than a nuisance, but a
few have evolved into something far more serious. The major killers of
humanity throughout our recent history—smallpox, flu, tuberculosis, malaria,
plague, measles, and cholera—are infectious diseases that evolved from
diseases of animals, even though most of the microbes responsible for our
own epidemic illnesses are paradoxically now almost confined to humans.
Because diseases have been the biggest killers of people, they have also been
decisive shapers of history. Until World War II, more victims of war died of
war-borne microbes than of battle wounds. All those military histories
glorifying great generals oversimplify the ego-deflating truth: the winners of
past wars were not always the armies with the best generals and weapons, but
were often merely those bearing the nastiest germs to transmit to their
enemies.
The grimmest examples of germs’ role in history come from the European
conquest of the Americas that began with Columbus’s voyage of 1492.
Numerous as were the Native American victims of the murderous Spanish
conquistadores, they were far outnumbered by the victims of murderous
Spanish microbes. Why was the exchange of nasty germs between the
Americas and Europe so unequal? Why didn’t Native American diseases
instead decimate the Spanish invaders, spread back to Europe, and wipe out
95 percent of Europe’s population? Similar questions arise for the decimation
of many other native peoples by Eurasian germs, as well as for the decimation
of would-be European conquistadores in the tropics of Africa and Asia.
Thus, questions of the animal origins of human disease lie behind the
broadest pattern of human history, and behind some of the most important
issues in human health today. (Think of AIDS, an explosively spreading
human disease that appears to have evolved from a virus resident in wild
African monkeys.) This chapter will begin by considering what a “disease” is,
and why some microbes have evolved so as to “make us sick,” whereas most
other species of living things don’t make us sick. We’ll examine why many of
our most familiar infectious diseases run in epidemics, such as our current
AIDS epidemic and the Black Death (bubonic plague) epidemics of the
Middle Ages. We’ll then consider how the ancestors of microbes now
confined to us transferred themselves from their original animal hosts.
Finally, we’ll see how insight into the animal origins of our infectious
diseases helps explain the momentous, almost one-way exchange of germs
between Europeans and Native Americans.
NATURALLY, WE’RE DISPOSED to think about diseases just from our own point
of view: what can we do to save ourselves and to kill the microbes? Let’s
stamp out the scoundrels, and never mind what their motives are! In life in
general, though, one has to understand the enemy in order to beat him, and
that’s especially true in medicine.
Hence let’s begin by temporarily setting aside our human bias and
considering disease from the microbes’ point of view. After all, microbes are
as much a product of natural selection as we are. What evolutionary benefit
does a microbe derive from making us sick in bizarre ways, like giving us
genital sores or diarrhea? And why should microbes evolve so as to kill us?
That seems especially puzzling and self-defeating, since a microbe that kills
its host kills itself.
Basically, microbes evolve like other species. Evolution selects for those
individuals most effective at producing babies and at helping them spread to
suitable places to live. For a microbe, spread may be defined mathematically
as the number of new victims infected per each original patient. That number
depends on how long each victim remains capable of infecting new victims,
and how efficiently the microbe is transferred from one victim to the next.
Microbes have evolved diverse ways of spreading from one person to
another, and from animals to people. The germ that spreads better leaves more
babies and ends up favored by natural selection. Many of our “symptoms” of
disease actually represent ways in which some damned clever microbe
modifies our bodies or our behavior such that we become enlisted to spread
microbes.
The most effortless way a germ could spread is by just waiting to be
transmitted passively to the next victim. That’s the strategy practiced by
microbes that wait for one host to be eaten by the next host: for instance,
salmonella bacteria, which we contract by eating already infected eggs or
meat; the worm responsible for trichinosis, which gets from pigs to us by
waiting for us to kill the pig and eat it without proper cooking; and the worm
causing anisakiasis, with which sushi-loving Japanese and Americans
occasionally infect themselves by consuming raw fish. Those parasites pass to
a person from an eaten animal, but the virus causing laughing sickness (kuru)
in the New Guinea highlands used to pass to a person from another person
who was eaten. It was transmitted by cannibalism, when highland babies
made the fatal mistake of licking their fingers after playing with raw brains
that their mothers had just cut out of dead kuru victims awaiting cooking.
Some microbes don’t wait for the old host to die and get eaten, but instead
hitchhike in the saliva of an insect that bites the old host and flies off to find a
new host. The free ride may be provided by mosquitoes, fleas, lice, or tsetse
flies that spread malaria, plague, typhus, or sleeping sickness, respectively.
The dirtiest of all tricks for passive carriage is perpetrated by microbes that
pass from a woman to her fetus and thereby infect babies already at birth. By
playing that trick, the microbes responsible for syphilis, rubella, and now
AIDS pose ethical dilemmas with which believers in a fundamentally just
universe have had to struggle desperately.
Other germs take matters into their own hands, figuratively speaking.
They modify the anatomy or habits of their host in such a way as to accelerate
their transmission. From our perspective, the open genital sores caused by
venereal diseases like syphilis are a vile indignity. From the microbes’ point
of view, however, they’re just a useful device to enlist a host’s help in
inoculating microbes into a body cavity of a new host. The skin lesions
caused by smallpox similarly spread microbes by direct or indirect body
contact (occasionally very indirect, as when U.S. whites bent on wiping out
“belligerent” Native Americans sent them gifts of blankets previously used by
smallpox patients).
More vigorous yet is the strategy practiced by the influenza, common
cold, and pertussis (whooping cough) microbes, which induce the victim to
cough or sneeze, thereby launching a cloud of microbes toward prospective
new hosts. Similarly, the cholera bacterium induces in its victim a massive
diarrhea that delivers bacteria into the water supplies of potential new victims,
while the virus responsible for Korean hemorrhagic fever broadcasts itself in
the urine of mice. For modification of a host’s behavior, nothing matches
rabies virus, which not only gets into the saliva of an infected dog but drives
the dog into a frenzy of biting and thus infecting many new victims. But for
physical effort on the bug’s own part, the prize still goes to worms such as
hookworms and schistosomes, which actively burrow through a host’s skin
from the water or soil into which their larvae had been excreted in a previous
victim’s feces.
Thus, from our point of view, genital sores, diarrhea, and coughing are
“symptoms of disease.” From a germ’s point of view, they’re clever
evolutionary strategies to broadcast the germ. That’s why it’s in the germ’s
interests to “make us sick.” But why should a germ evolve the apparently
self-defeating strategy of killing its host?
From the germ’s perspective, that’s just an unintended by-product (fat
consolation to us!) of host symptoms promoting efficient transmission of
microbes. Yes, an untreated cholera patient may eventually die from
producing diarrheal fluid at a rate of several gallons per day. At least for a
while, though, as long as the patient is still alive, the cholera bacterium profits
from being massively broadcast into the water supplies of its next victims.
Provided that each victim thereby infects on the average more than one new
victim, the bacterium will spread, even though the first host happens to die.
SO MUCH FOR our dispassionate examination of the germ’s interests. Now
let’s get back to considering our own selfish interests: to stay alive and
healthy, best done by killing the damned germs. One common response of
ours to infection is to develop a fever. Again, we’re used to considering fever
as a “symptom of disease,” as if it developed inevitably without serving any
function. But regulation of body temperature is under our genetic control and
doesn’t just happen by accident. A few microbes are more sensitive to heat
than our own bodies are. By raising our body temperature, we in effect try to
bake the germs to death before we get baked ourselves.
Another common response of ours is to mobilize our immune system.
White blood cells and other cells of ours actively seek out and kill foreign
microbes. The specific antibodies that we gradually build up against a
particular microbe infecting us make us less likely to get reinfected once we
become cured. As we all know from experience, there are some illnesses,
such as flu and the common cold, to which our resistance is only temporary;
we can eventually contract the illness again. Against other illnesses, though—
including measles, mumps, rubella, pertussis, and the now defeated smallpox
—our antibodies stimulated by one infection confer lifelong immunity. That’s
the principle of vaccination: to stimulate our antibody production without our
having to go through the actual experience of the disease, by inoculating us
with a dead or weakened strain of microbe.
Alas, some clever microbes don’t just cave in to our immune defenses.
Some have learned to trick us by changing those molecular pieces of the
microbe (its so-called antigens) that our antibodies recognize. The constant
evolution or recycling of new strains of flu, with differing antigens, explains
why your having gotten flu two years ago didn’t protect you against the
different strain that arrived this year. Malaria and sleeping sickness are even
more slippery customers in their ability rapidly to change their antigens.
Among the slipperiest of all is AIDS, which evolves new antigens even as it
sits within an individual patient, thereby eventually overwhelming his or her
immune system.
Our slowest defensive response is through natural selection, which
changes our gene frequencies from generation to generation. For almost any
disease, some people prove to be genetically more resistant than are others. In
an epidemic those people with genes for resistance to that particular microbe
are more likely to survive than are people lacking such genes. As a result,
over the course of history, human populations repeatedly exposed to a
particular pathogen have come to consist of a higher proportion of individuals
with those genes for resistance—just because unfortunate individuals without
the genes were less likely to survive to pass their genes on to babies.
Fat consolation, you may be thinking again. This evolutionary response is
not one that does the genetically susceptible dying individual any good. It
does mean, though, that a human population as a whole becomes better
protected against the pathogen. Examples of those genetic defenses include
the protections (at a price) that the sickle-cell gene, Tay-Sachs gene, and
cystic fibrosis gene may confer on African blacks, Ashkenazi Jews, and
northern Europeans against malaria, tuberculosis, and bacterial diarrheas,
respectively.
In short, our interaction with most species, as exemplified by
hummingbirds, doesn’t make us or the hummingbird “sick.” Neither we nor
hummingbirds have had to evolve defenses against each other. That peaceful
relationship was able to persist because hummingbirds don’t count on us to
spread their babies or to offer our bodies for food. Hummingbirds evolved
instead to feed on nectar and insects, which they find by using their own
wings.
But microbes evolved to feed on the nutrients within our own bodies, and
they don’t have wings to let them reach a new victim’s body once the original
victim is dead or resistant. Hence many germs have had to evolve tricks to let
them spread between potential victims, and many of those tricks are what we
experience as “symptoms of disease.” We’ve evolved countertricks of our
own, to which the germs have responded by evolving counter-countertricks.
We and our pathogens are now locked in an escalating evolutionary contest,
with the death of one contestant the price of defeat, and with natural selection
playing the role of umpire. Now let’s consider the form of the contest:
blitzkrieg or guerrilla war?
SUPPOSE THAT ONE counts the number of cases of some particular infectious
disease in some geographic area, and watches how the numbers change with
time. The resulting patterns differ greatly among diseases. For certain
diseases, like malaria or hookworm, new cases appear any month of any year
in an affected area. So-called epidemic diseases, though, produce no cases for
a long time, then a whole wave of cases, then no more cases again for a while.
Among such epidemic diseases, influenza is one personally familiar to
most Americans, certain years being particularly bad years for us (but great
years for the influenza virus). Cholera epidemics come at longer intervals, the
1991 Peruvian epidemic being the first one to reach the New World during the
20th century. Although today’s influenza and cholera epidemics make front-
page stories, epidemics used to be far more terrifying before the rise of
modern medicine. The greatest single epidemic in human history was the one
of influenza that killed 21 million people at the end of the First World War.
The Black Death (bubonic plague) killed one-quarter of Europe’s population
between 1346 and 1352, with death tolls ranging up to 70 percent in some
cities. When the Canadian Pacific Railroad was being built through
Saskatchewan in the early 1880s, that province’s Native Americans, who had
previously had little exposure to whites and their germs, died of tuberculosis
at the incredible rate of 9 percent per year.
The infectious diseases that visit us as epidemics, rather than as a steady
trickle of cases, share several characteristics. First, they spread quickly and
efficiently from an infected person to nearby healthy people, with the result
that the whole population gets exposed within a short time. Second, they’re
“acute” illnesses: within a short time, you either die or recover completely.
Third, the fortunate ones of us who do recover develop antibodies that leave
us immune against a recurrence of the disease for a long time, possibly for the
rest of our life. Finally, these diseases tend to be restricted to humans; the
microbes causing them tend not to live in the soil or in other animals. All four
of these traits apply to what Americans think of as the familiar acute epidemic
diseases of childhood, including measles, rubella, mumps, pertussis, and
smallpox.
The reason why the combination of those four traits tends to make a
disease run in epidemics is easy to understand. In simplified form, here’s what
happens. The rapid spread of microbes, and the rapid course of symptoms,
mean that everybody in a local human population is quickly infected and soon
thereafter is either dead or else recovered and immune. No one is left alive
who could still be infected. But since the microbe can’t survive except in the
bodies of living people, the disease dies out, until a new crop of babies
reaches the susceptible age—and until an infectious person arrives from the
outside to start a new epidemic.
A classic illustration of how such diseases occur as epidemics is the
history of measles on the isolated Atlantic islands called the Faeroes. A severe
epidemic of measles reached the Faeroes in 1781 and then died out, leaving
the islands measles free until an infected carpenter arrived on a ship from
Denmark in 1846. Within three months, almost the whole Faeroes population
(7,782 people) had gotten measles and then either died or recovered, leaving
the measles virus to disappear once again until the next epidemic. Studies
show that measles is likely to die out in any human population numbering
fewer than half a million people. Only in larger populations can the disease
shift from one local area to another, thereby persisting until enough babies
have been born in the originally infected area that measles can return there.
What’s true for measles in the Faeroes is true of our other familiar acute
infectious diseases throughout the world. To sustain themselves, they need a
human population that is sufficiently numerous, and sufficiently densely
packed, that a numerous new crop of susceptible children is available for
infection by the time the disease would otherwise be waning. Hence measles
and similar diseases are also known as crowd diseases.
OBVIOUSLY, CROWD DISEASES could not sustain themselves in small bands of
hunter-gatherers and slash-and-burn farmers. As tragic modern experience
with Amazonian Indians and Pacific Islanders confirms, almost an entire
tribelet may be wiped out by an epidemic brought by an outside visitor—
because no one in the tribelet had any antibodies against the microbe. For
example, in the winter of 1902 a dysentery epidemic brought by a sailor on
the whaling ship Active killed 51 out of the 56 Sadlermiut Eskimos, a very
isolated band of people living on Southampton Island in the Canadian Arctic.
In addition, measles and some of our other “childhood” diseases are more
likely to kill infected adults than children, and all adults in the tribelet are
susceptible. (In contrast, modern Americans rarely contract measles as adults,
because most of them get either measles or the vaccine against it as children.)
Having killed most of the tribelet, the epidemic then disappears. The small
population size of tribelets explains not only why they can’t sustain epidemics
introduced from the outside, but also why they never could evolve epidemic
diseases of their own to give back to visitors.
That’s not to say, though, that small human populations are free from all
infectious diseases. They do have infections, but only of certain types. Some
are caused by microbes capable of maintaining themselves in animals or in
the soil, with the result that the disease doesn’t die out but remains constantly
available to infect people. For example, the yellow fever virus is carried by
African wild monkeys, whence it can always infect rural human populations
of Africa, whence it was carried by the transatlantic slave trade to infect New
World monkeys and people.
Still other infections of small human populations are chronic diseases
such as leprosy and yaws. Since the disease may take a very long time to kill
its victim, the victim remains alive as a reservoir of microbes to infect other
members of the tribelet. For instance, the Karimui Basin of the New Guinea
highlands, where I worked in the 1960s, was occupied by an isolated
population of a few thousand people, suffering from the world’s highest
incidence of leprosy—about 40 percent! Finally, small human populations are
also susceptible to nonfatal infections against which we don’t develop
immunity, with the result that the same person can become reinfected after
recovering. That happens with hookworm and many other parasites.
All these types of diseases, characteristic of small isolated populations,
must be the oldest diseases of humanity. They were the ones we could evolve
and sustain through the early millions of years of our evolutionary history,
when the total human population was tiny and fragmented. These diseases are
also shared with, or similar to the diseases of, our closest wild relatives, the
African great apes. In contrast, the crowd diseases, which we discussed
earlier, could have arisen only with the buildup of large, dense human
populations. That buildup began with the rise of agriculture starting about
10,000 years ago and then accelerated with the rise of cities starting several
thousand years ago. In fact, the first attested dates for many familiar
infectious diseases are surprisingly recent: around 1600 B.C. for smallpox (as
deduced from pockmarks on an Egyptian mummy), 400 B.C. for mumps, 200
B.C. for leprosy, A.D. 1840 for epidemic polio, and 1959 for AIDS.
WHY DID THE rise of agriculture launch the evolution of our crowd
infectious diseases? One reason just mentioned is that agriculture sustains
much higher human population densities than does the hunting-gathering
lifestyle—on the average, 10 to 100 times higher. In addition, hunter-
gatherers frequently shift camp and leave behind their own piles of feces with
accumulated microbes and worm larvae. But farmers are sedentary and live
amid their own sewage, thus providing microbes with a short path from one
person’s body into another’s drinking water.
Some farming populations make it even easier for their own fecal bacteria
and worms to infect new victims, by gathering their feces and urine and
spreading them as fertilizer on the fields where people work. Irrigation
agriculture and fish farming provide ideal living conditions for the snails
carrying schistosomiasis and for flukes that burrow through our skin as we
wade through the feces-laden water. Sedentary farmers become surrounded
not only by their feces but also by disease transmitting rodents, attracted by
the farmers’ stored food. The forest clearings made by African farmers also
provide ideal breeding habitats for malaria-transmitting mosquitoes.
If the rise of farming was thus a bonanza for our microbes, the rise of
cities was a greater one, as still more densely packed human populations
festered under even worse sanitation conditions. Not until the beginning of
the 20th century did Europe’s urban populations finally become self-
sustaining: before then, constant immigration of healthy peasants from the
countryside was necessary to make up for the constant deaths of city dwellers
from crowd diseases. Another bonanza was the development of world trade
routes, which by Roman times effectively joined the populations of Europe,
Asia, and North Africa into one giant breeding ground for microbes. That’s
when smallpox finally reached Rome, as the Plague of Antoninus, which
killed millions of Roman citizens between A.D. 165 and 180.
Similarly, bubonic plague first appeared in Europe as the Plague of
Justinian (A.D. 542–43). But plague didn’t begin to hit Europe with full force
as the Black Death epidemics until A.D. 1346, when a new route for overland
trade with China provided rapid transit, along Eurasia’s east-west axis, for
flea-infested furs from plague-ridden areas of Central Asia to Europe. Today,
our jet planes have made even the longest intercontinental flights briefer than
the duration of any human infectious disease. That’s how an Aerolineas
Argentinas airplane, stopping in Lima (Peru) in 1991, managed to deliver
dozens of cholera-infected people that same day to my city of Los Angeles,
over 3,000 miles from Lima. The explosive increase in world travel by
Americans, and in immigration to the United States, is turning us into another
melting pot—this time, of microbes that we previously dismissed as just
causing exotic diseases in far-off countries.
THUS, WHEN THE human population became sufficiently large and
concentrated, we reached the stage in our history at which we could at last
evolve and sustain crowd diseases confined to our own species. But that
conclusion presents a paradox: such diseases could never have existed before
then! Instead, they had to evolve as new diseases. Where did those new
diseases come from?
Evidence has recently been emerging from molecular studies of the
disease-causing microbes themselves. For many of the microbes responsible
for our unique diseases, molecular biologists can now identify the microbe’s
closest relatives. These also prove to be agents of crowd infectious diseases—
but ones confined to various species of our domestic animals and pets!
Among animals, too, epidemic diseases require large, dense populations and
don’t afflict just any animal: they’re confined mainly to social animals
providing the necessary large populations. Hence when we domesticated
social animals, such as cows and pigs, they were already afflicted by epidemic
diseases just waiting to be transferred to us.
For example, measles virus is most closely related to the virus causing
rinderpest. That nasty epidemic disease affects cattle and many wild cud-
chewing mammals, but not humans. Measles in turn doesn’t afflict cattle. The
close similarity of the measles virus to the rinderpest virus suggests that the
latter transferred from cattle to humans and then evolved into the measles
virus by changing its properties to adapt to us. That transfer is not at all
surprising, considering that many peasant farmers live and sleep close to cows
and their feces, urine, breath, sores, and blood. Our intimacy with cattle has
been going on for the 9,000 years since we domesticated them—ample time
for the rinderpest virus to discover us nearby. As Table 11.1 illustrates, others
of our familiar infectious diseases can similarly be traced back to diseases of
our animal friends.
GIVEN OUR PROXIMITY to the animals we love, we must be getting constantly
bombarded by their microbes. Those invaders get winnowed by natural
selection, and only a few of them succeed in establishing themselves as
human diseases. A quick survey of current diseases lets us trace out four
stages in the evolution of a specialized human disease from an animal
precursor.
The first stage is illustrated by dozens of diseases that we now and then
pick up directly from our pets and domestic animals. They include cat-scratch
fever from our cats, leptospirosis from our dogs, psittacosis from our chickens
and parrots, and brucellosis from our cattle. We’re similarly liable to pick up
diseases from wild animals, such as the tularemia that hunters can get from
skinning wild rabbits. All those microbes are still at an early stage in their
evolution into specialized human pathogens. They still don’t get transmitted
directly from one person to another, and even their transfer to us from animals
remains uncommon.
TABLE 11.1 Deadly Gifts from Our Animal Friends
Human Disease
Animal with Most Closely Related Pathogen
Measles
cattle (rinderpest)
Tuberculosis
cattle
Smallpox
cattle (cowpox) or other livestock with related pox viruses
Flu
pigs and ducks
Pertussis
pigs, dogs
Falciparum malaria birds (chickens and ducks?)
In the second stage a former animal pathogen evolves to the point where
it does get transmitted directly between people and causes epidemics.
However, the epidemic dies out for any of several reasons, such as being
cured by modern medicine, or being stopped when everybody around has
already been infected and either becomes immune or dies. For example, a
previously unknown fever termed O’nyong-nyong fever appeared in East
Africa in 1959 and proceeded to infect several million Africans. It probably
arose from a virus of monkeys and was transmitted to humans by mosquitoes.
The fact that patients recovered quickly and became immune to further attack
helped the new disease die out quickly. Closer to home for Americans, Fort
Bragg fever was the name applied to a new leptospiral disease that broke out
in the United States in the summer of 1942 and soon disappeared.
A fatal disease vanishing for another reason was New Guinea’s laughing
sickness, transmitted by cannibalism and caused by a slow-acting virus from
which no one has ever recovered. Kuru was on its way to exterminating New
Guinea’s Foré tribe of 20,000 people, until the establishment of Australian
government control around 1959 ended cannibalism and thereby the
transmission of kuru. The annals of medicine are full of accounts of diseases
that sound like no disease known today, but that once caused terrifying
epidemics and then disappeared as mysteriously as they had come. The
“English sweating sickness,” which swept and terrified Europe between 1485
and 1552, and the “Picardy sweats” of 18th- and 19th-century France, are just
two of the many epidemic illnesses that vanished long before modern
medicine had devised methods for identifying the responsible microbes.
A third stage in the evolution of our major diseases is represented by
former animal pathogens that did establish themselves in humans, that have
not (not yet?) died out, and that may or may not still become major killers of
humanity. The future remains very uncertain for Lassa fever, caused by a
virus derived probably from rodents. Lassa fever was first observed in 1969 in
Nigeria, where it causes a fatal illness so contagious that Nigerian hospitals
have been closed down if even a single case appears. Better established is
Lyme disease, caused by a spirochete that we get from the bite of ticks carried
by mice and deer. Although the first known human cases in the United States
appeared only as recently as 1962, Lyme disease is already reaching epidemic
proportions in many parts of our country. The future of AIDS, derived from
monkey viruses and first documented in humans around 1959, is even more
secure (from the virus’s perspective).
The final stage of this evolution is represented by the major, long-
established epidemic diseases confined to humans. These diseases must have
been the evolutionary survivors of far more pathogens that tried to make the
jump to us from animals—and mostly failed.
What is actually going on in those stages, as an exclusive disease of
animals transforms itself into an exclusive disease of humans? One
transformation involves a change of intermediate vector: when a microbe
relying on some arthropod vector for transmission switches to a new host, the
microbe may be forced to find a new arthropod as well. For example, typhus
was initially transmitted between rats by rat fleas, which sufficed for a while
to transfer typhus from rats to humans. Eventually, typhus microbes
discovered that human body lice offered a much more efficient method of
traveling directly between humans. Now that Americans have mostly
deloused themselves, typhus has discovered a new route into us: by infecting
eastern North American flying squirrels and then transferring to people whose
attics harbor flying squirrels.
In short, diseases represent evolution in progress, and microbes adapt by
natural selection to new hosts and vectors. But compared with cows’ bodies,
ours offer different immune defenses, lice, feces, and chemistries. In that new
environment, a microbe must evolve new ways to live and to propagate itself.
In several instructive cases doctors or veterinarians have actually been able to
observe microbes evolving those new ways.
The best-studied case involves what happened when myxomatosis hit
Australian rabbits. The myxo virus, native to a wild species of Brazilian
rabbit, had been observed to cause a lethal epidemic in European domestic
rabbits, which are a different species. Hence the virus was intentionally
introduced to Australia in 1950 in the hopes of ridding the continent of its
plague of European rabbits, foolishly introduced in the nineteenth century. In
the first year, myxo produced a gratifying (to Australian farmers) 99.8 percent
mortality rate in infected rabbits. Unfortunately for the farmers, the death rate
then dropped in the second year to 90 percent and eventually to 25 percent,
frustrating hopes of eradicating rabbits completely from Australia. The
problem was that the myxo virus evolved to serve its own interests, which
differed from ours as well as from those of the rabbits. The virus changed so
as to kill fewer rabbits and to permit lethally infected ones to live longer
before dying. As a result, a less lethal myxo virus spreads baby viruses to
more rabbits than did the original, highly virulent myxo.
For a similar example in humans, we have only to consider the surprising
evolution of syphilis. Today, our two immediate associations to syphilis are
genital sores and a very slowly developing disease, leading to the death of
many untreated victims only after many years. However, when syphilis was
first definitely recorded in Europe in 1495, its pustules often covered the body
from the head to the knees, caused flesh to fall off people’s faces, and led to
death within a few months. By 1546, syphilis had evolved into the disease
with the symptoms so well known to us today. Apparently, just as with
myxomatosis, those syphilis spirochetes that evolved so as to keep their
victims alive for longer were thereby able to transmit their spirochete
offspring into more victims.
THE IMPORTANCE OF lethal microbes in human history is well illustrated by
Europeans’ conquest and depopulation of the New World. Far more Native
Americans died in bed from Eurasian germs than on the battlefield from
European guns and swords. Those germs undermined Indian resistance by
killing most Indians and their leaders and by sapping the survivors’ morale.
For instance, in 1519 Cortés landed on the coast of Mexico with 600
Spaniards, to conquer the fiercely militaristic Aztec Empire with a population
of many millions. That Cortés reached the Aztec capital of Tenochtitlán,
escaped with the loss of “only” two-thirds of his force, and managed to fight
his way back to the coast demonstrates both Spanish military advantages and
the initial naïveté of the Aztecs. But when Cortés’s next onslaught came, the
Aztecs were no longer naive and fought street by street with the utmost
tenacity. What gave the Spaniards a decisive advantage was smallpox, which
reached Mexico in 1520 with one infected slave arriving from Spanish Cuba.
The resulting epidemic proceeded to kill nearly half of the Aztecs, including
Emperor Cuitláhuac. Aztec survivors were demoralized by the mysterious
illness that killed Indians and spared Spaniards, as if advertising the
Spaniards’ invincibility. By 1618, Mexico’s initial population of about 20
million had plummeted to about 1.6 million.
Pizarro had similarly grim luck when he landed on the coast of Peru in
1531 with 168 men to conquer the Inca Empire of millions. Fortunately for
Pizarro and unfortunately for the Incas, smallpox had arrived overland around
1526, killing much of the Inca population, including both the emperor Huayna
Capac and his designated successor. As we saw in Chapter 3, the result of the
throne’s being left vacant was that two other sons of Huayna Capac,
Atahuallpa and Huascar, became embroiled in a civil war that Pizarro
exploited to conquer the divided Incas.
When we in the United States think of the most populous New World
societies existing in 1492, only those of the Aztecs and the Incas tend to come
to our minds. We forget that North America also supported populous Indian
societies in the most logical place, the Mississippi Valley, which contains
some of our best farmland today. In that case, however, conquistadores
contributed nothing directly to the societies’ destruction; Eurasian germs,
spreading in advance, did everything. When Hernando de Soto became the
first European conquistador to march through the southeastern United States,
in 1540, he came across Indian town sites abandoned two years earlier
because the inhabitants had died in epidemics. These epidemics had been
transmitted from coastal Indians infected by Spaniards visiting the coast. The
Spaniards’ microbes spread to the interior in advance of the Spaniards
themselves.
De Soto was still able to see some of the densely populated Indian towns
lining the lower Mississippi. After the end of his expedition, it was a long
time before Europeans again reached the Mississippi Valley, but Eurasian
microbes were now established in North America and kept spreading. By the
time of the next appearance of Europeans on the lower Mississippi, that of
French settlers in the late 1600s, almost all of those big Indian towns had
vanished. Their relics are the great mound sites of the Mississippi Valley.
Only recently have we come to realize that many of the mound-building
societies were still largely intact when Columbus reached the New World, and
that they collapsed (probably as a result of disease) between 1492 and the
systematic European exploration of the Mississippi.
When I was young, American schoolchildren were taught that North
America had originally been occupied by only about one million Indians.
That low number was useful in justifying the white conquest of what could be
viewed as an almost empty continent. However, archaeological excavations,
and scrutiny of descriptions left by the very first European explorers on our
coasts, now suggest an initial number of around 20 million Indians. For the
New World as a whole, the Indian population decline in the century or two
following Columbus’s arrival is estimated to have been as large as 95 percent.
The main killers were Old World germs to which Indians had never been
exposed, and against which they therefore had neither immune nor genetic
resistance. Smallpox, measles, influenza, and typhus competed for top rank
among the killers. As if these had not been enough, diphtheria, malaria,
mumps, pertussis, plague, tuberculosis, and yellow fever came up close
behind. In countless cases, whites were actually there to witness the
destruction occurring when the germs arrived. For example, in 1837 the
Mandan Indian tribe, with one of the most elaborate cultures in our Great
Plains, contracted smallpox from a steamboat traveling up the Missouri River
from St. Louis. The population of one Mandan village plummeted from 2,000
to fewer than 40 within a few weeks.
WHILE OVER A dozen major infectious diseases of Old World origins
became established in the New World, perhaps not a single major killer
reached Europe from the Americas. The sole possible exception is syphilis,
whose area of origin remains controversial. The one-sidedness of that
exchange of germs becomes even more striking when we recall that large,
dense human populations are a prerequisite for the evolution of our crowd
infectious diseases. If recent reappraisals of the pre-Columbian New World
population are correct, it was not far below the contemporary population of
Eurasia. Some New World cities like Tenochtitlán were among the world’s
most populous cities at the time. Why didn’t Tenochtitlán have awful germs
waiting for the Spaniards?
One possible contributing factor is that the rise of dense human
populations began somewhat later in the New World than in the Old World.
Another is that the three most densely populated American centers—the
Andes, Mesoamerica, and the Mississippi Valley—never became connected
by regular fast trade into one huge breeding ground for microbes, in the way
that Europe, North Africa, India, and China became linked in Roman times.
Those factors still don’t explain, though, why the New World apparently
ended up with no lethal crowd epidemics at all. (Tuberculosis DNA has been
reported from the mummy of a Peruvian Indian who died 1,000 years ago, but
the identification procedure used did not distinguish human tuberculosis from
a closely related pathogen ( Mycobacterium bovis) that is widespread in wild
animals.)
Instead, what must be the main reason for the failure of lethal crowd
epidemics to arise in the Americas becomes clear when we pause to ask a
simple question. From what microbes could they conceivably have evolved?
We’ve seen that Eurasian crowd diseases evolved out of diseases of Eurasian
herd animals that became domesticated. Whereas many such animals existed
in Eurasia, only five animals of any sort became domesticated in the
Americas: the turkey in Mexico and the U.S. Southwest, the llama / alpaca
and the guinea pig in the Andes, the Muscovy duck in tropical South
America, and the dog throughout the Americas.
In turn, we also saw that this extreme paucity of domestic animals in the
New World reflects the paucity of wild starting material. About 80 percent of
the big wild mammals of the Americas became extinct at the end of the last
Ice Age, around 13,000 years ago. The few domesticates that remained to
Native Americans were not likely sources of crowd diseases, compared with
cows and pigs. Muscovy ducks and turkeys don’t live in enormous flocks, and
they’re not cuddly species (like young lambs) with which we have much
physical contact. Guinea pigs may have contributed a trypanosome infection
like Chagas’ disease or leishmaniasis to our catalog of woes, but that’s
uncertain. Initially, most surprising is the absence of any human disease
derived from llamas (or alpacas), which it’s tempting to consider the Andean
equivalent of Eurasian livestock. However, llamas had four strikes against
them as a source of human pathogens: they were kept in smaller herds than
were sheep and goats and pigs; their total numbers were never remotely as
large as those of the Eurasian populations of domestic livestock, since llamas
never spread beyond the Andes; people don’t drink (and get infected by)
llama milk; and llamas aren’t kept indoors, in close association with people.
In contrast, human mothers in the New Guinea highlands often nurse piglets,
and pigs as well as cows are frequently kept inside the huts of peasant
farmers.
THE HISTORICAL IMPORTANCE of animal-derived diseases extends far beyond
the collision of the Old and the New Worlds. Eurasian germs played a key
role in decimating native peoples in many other parts of the world, including
Pacific islanders, Aboriginal Australians, and the Khoisan peoples (Hottentots
and Bushmen) of southern Africa. Cumulative mortalities of these previously
unexposed peoples from Eurasian germs ranged from 50 percent to 100
percent. For instance, the Indian population of Hispaniola declined from
around 8 million, when Columbus arrived in A.D. 1492, to zero by 1535.
Measles reached Fiji with a Fijian chief returning from a visit to Australia in
1875, and proceeded to kill about one-quarter of all Fijians then alive (after
most Fijians had already been killed by epidemics beginning with the first
European visit, in 1791). Syphilis, gonorrhea, tuberculosis, and influenza
arriving with Captain Cook in 1779, followed by a big typhoid epidemic in
1804 and numerous “minor” epidemics, reduced Hawaii’s population from
around half a million in 1779 to 84,000 in 1853, the year when smallpox
finally reached Hawaii and killed around 10,000 of the survivors. These
examples could be multiplied almost indefinitely.
However, germs did not act solely to Europeans’ advantage. While the
New World and Australia did not harbor native epidemic diseases awaiting
Europeans, tropical Asia, Africa, Indonesia, and New Guinea certainly did.
Malaria throughout the tropical Old World, cholera in tropical Southeast Asia,
and yellow fever in tropical Africa were (and still are) the most notorious of
the tropical killers. They posed the most serious obstacle to European
colonization of the tropics, and they explain why the European colonial
partitioning of New Guinea and most of Africa was not accomplished until
nearly 400 years after European partitioning of the New World began.
Furthermore, once malaria and yellow fever did become transmitted to the
Americas by European ship traffic, they emerged as the major impediment to
colonization of the New World tropics as well. A familiar example is the role
of those two diseases in aborting the French effort, and nearly aborting the
ultimately successful American effort, to construct the Panama Canal.
Bearing all these facts in mind, let’s try to regain our sense of perspective
about the role of germs in answering Yali’s question. There is no doubt that
Europeans developed a big advantage in weaponry, technology, and political
organization over most of the non-European peoples that they conquered. But
that advantage alone doesn’t fully explain how initially so few European
immigrants came to supplant so much of the native population of the
Americas and some other parts of the world. That might not have happened
without Europe’s sinister gift to other continents—the germs evolving from
Eurasians’ long intimacy with domestic animals.
CHAPTER 12
BLUEPRINTS AND BORROWED LETTERS
NINETEENTH-CENTURY AUTHORS TENDED TO INTERPRET history as a
progression from savagery to civilization. Key hallmarks of this transition
included the development of agriculture, metallurgy, complex technology,
centralized government, and writing. Of these, writing was traditionally the
one most restricted geographically: until the expansions of Islam and of
colonial Europeans, it was absent from Australia, Pacific islands,
subequatorial Africa, and the whole New World except for a small part of
Mesoamerica. As a result of that confined distribution, peoples who pride
themselves on being civilized have always viewed writing as the sharpest
distinction raising them above “barbarians” or “savages.”
Knowledge brings power. Hence writing brings power to modern
societies, by making it possible to transmit knowledge with far greater
accuracy and in far greater quantity and detail, from more distant lands and
more remote times. Of course, some peoples (notably the Incas) managed to
administer empires without writing, and “civilized” peoples don’t always
defeat “barbarians,” as Roman armies facing the Huns learned. But the
European conquests of the Americas, Siberia, and Australia illustrate the
typical recent outcome.
Writing marched together with weapons, microbes, and centralized
political organization as a modern agent of conquest. The commands of the
monarchs and merchants who organized colonizing fleets were conveyed in
writing. The fleets set their courses by maps and written sailing directions
prepared by previous expeditions. Written accounts of earlier expeditions
motivated later ones, by describing the wealth and fertile lands awaiting the
conquerors. The accounts taught subsequent explorers what conditions to
expect, and helped them prepare themselves. The resulting empires were
administered with the aid of writing. While all those types of information
were also transmitted by other means in preliterate societies, writing made the
transmission easier, more detailed, more accurate, and more persuasive.
Why, then, did only some peoples and not others develop writing, given
its overwhelming value? For example, why did no traditional hunters-
gatherers evolve or adopt writing? Among island empires, why did writing
arise in Minoan Crete but not in Polynesian Tonga? How many separate times
did writing evolve in human history, under what circumstances, and for what
uses? Of those peoples who did develop it, why did some do so much earlier
than others? For instance, today almost all Japanese and Scandinavians are
literate but most Iraqis are not: why did writing nevertheless arise nearly four
thousand years earlier in Iraq?
The diffusion of writing from its sites of origin also raises important
questions. Why, for instance, did it spread to Ethiopia and Arabia from the
Fertile Crescent, but not to the Andes from Mexico? Did writing systems
spread by being copied, or did existing systems merely inspire neighboring
peoples to invent their own systems? Given a writing system that works well
for one language, how do you devise a system for a different language?
Similar questions arise whenever one tries to understand the origins and
spread of many other aspects of human culture—such as technology, religion,
and food production. The historian interested in such questions about writing
has the advantage that they can often be answered in unique detail by means
of the written record itself. We shall therefore trace writing’s development not
only because of its inherent importance, but also for the general insights into
cultural history that it provides.
THE THREE BASIC strategies underlying writing systems differ in the size of
the speech unit denoted by one written sign: either a single basic sound, a
whole syllable, or a whole word. Of these, the one employed today by most
peoples is the alphabet, which ideally would provide a unique sign (termed a
letter) for each basic sound of the language (a phoneme). Actually, most
alphabets consist of only about 20 or 30 letters, and most languages have
more phonemes than their alphabets have letters. For example, English
transcribes about 40 phonemes with a mere 26 letters. Hence most
alphabetically written languages, including English, are forced to assign
several different phonemes to the same letter and to represent some phonemes
by combinations of letters, such as the English two-letter combinations sh and
th (each represented by a single letter in the Russian and Greek alphabets,
respectively).
The second strategy uses so-called logograms, meaning that one written
sign stands for a whole word. That’s the function of many signs of Chinese
writing and of the predominant Japanese writing system (termed kanji).
Before the spread of alphabetic writing, systems making much use of
logograms were more common and included Egyptian hieroglyphs, Maya
glyphs, and Sumerian cuneiform.
The third strategy, least familiar to most readers of this book, uses a sign
for each syllable. In practice, most such writing systems (termed syllabaries)
provide distinct signs just for syllables of one consonant followed by one
vowel (like the syllables of the word “fa-mi-ly”), and resort to various tricks
in order to write other types of syllables by means of those signs. Syllabaries
were common in ancient times, as exemplified by the Linear B writing of
Mycenaean Greece. Some syllabaries persist today, the most important being
the kana syllabary that the Japanese use for telegrams, bank statements, and
texts for blind readers.
I’ve intentionally termed these three approaches strategies rather than
writing systems. No actual writing system employs one strategy exclusively.
Chinese writing is not purely logographic, nor is English writing purely
alphabetic. Like all alphabetic writing systems, English uses many logograms,
such as numerals, $, %, and + : that is, arbitrary signs, not made up of
phonetic elements, representing whole words. “Syllabic” Linear B had many
logograms, and “logographic” Egyptian hieroglyphs included many syllabic
signs as well as a virtual alphabet of individual letters for each consonant.
INVENTING A WRITING system from scratch must have been incomparably
more difficult than borrowing and adapting one. The first scribes had to settle
on basic principles that we now take for granted. For example, they had to
figure out how to decompose a continuous utterance into speech units,
regardless of whether those units were taken as words, syllables, or
phonemes. They had to learn to recognize the same sound or speech unit
through all our normal variations in speech volume, pitch, speed, emphasis,
phrase grouping, and individual idiosyncrasies of pronunciation. They had to
decide that a writing system should ignore all of that variation. They then had
to devise ways to represent sounds by symbols.
Somehow, the first scribes solved all those problems, without having in
front of them any example of the final result to guide their efforts. That task
was evidently so difficult that there have been only a few occasions in history
when people invented writing entirely on their own. The two indisputably
independent inventions of writing were achieved by the Sumerians of
Mesopotamia somewhat before 3000 B.C. and by Mexican Indians before 600
B.C. (Figure 12.1); Egyptian writing of 3000 B.C. and Chinese writing (by 1300
B.C.) may also have arisen independently. Probably all other peoples who have
developed writing since then have borrowed, adapted, or at least been inspired
by existing systems.
The independent invention that we can trace in greatest detail is history’s
oldest writing system, Sumerian cuneiform (Figure 12.1). For thousands of
years before it jelled, people in some farming villages of the Fertile Crescent
had been using clay tokens of various simple shapes for accounting purposes,
such as recording numbers of sheep and amounts of grain. In the last centuries
before 3000 B.C., developments in accounting technology, format, and signs
rapidly led to the first system of writing. One such technological innovation
was the use of flat clay tablets as a convenient writing surface. Initially, the
clay was scratched with pointed tools, which gradually yielded to reed
styluses for neatly pressing a mark into the tablet. Developments in format
included the gradual adoption of conventions whose necessity is now
universally accepted: that writing should be organized into ruled rows or
columns (horizontal rows for the Sumerians, as for modern Europeans); that
the lines should be read in a constant direction (left to right for Sumerians, as
for modern Europeans); and that the lines should be read from top to bottom
of the tablet rather than vice versa.
But the crucial change involved the solution of the problem basic to
virtually all writing systems: how to devise agreed-on visible marks that
represent actual spoken sounds, rather than only ideas or else words
independent of their pronunciation. Early stages in the development of the
solution have been detected especially in thousands of clay tablets excavated
from the ruins of the former Sumerian city of Uruk, on the Euphrates River
about 200 miles southeast of modern Baghdad. The first Sumerian writing
signs were recognizable pictures of the object referred to (for instance, a
picture of a fish or a bird). Naturally, those pictorial signs consisted mainly of
numerals plus nouns for visible objects; the resulting texts were merely
accounting reports in a telegraphic shorthand devoid of grammatical
elements. Gradually, the forms of the signs became more abstract, especially
when the pointed writing tools were replaced by reed styluses. New signs
were created by combining old signs to produce new meanings: for example,
the sign for head was combined with the sign for bread in order to produce a
sign signifying eat.
The earliest Sumerian writing consisted of nonphonetic logograms. That’s
to say, it was not based on the specific sounds of the Sumerian language, and
it could have been pronounced with entirely different sounds to yield the
same meaning in any other language—just as the numeral sign 4 is variously
pronounced four, chetwíre, neljä, and empat by speakers of English, Russian,
Finnish, and Indonesian, respectively. Perhaps the most important single step
in the whole history of writing was the Sumerians’ introduction of phonetic
representation, initially by writing an abstract noun (which could not be
readily drawn as a picture) by means of the sign for a depictable noun that had
the same phonetic pronunciation. For instance, it’s easy to draw a
recognizable picture of arrow, hard to draw a recognizable picture of life, but
both are pronounced ti in Sumerian, so a picture of an arrow came to mean
either arrow or life. The resulting ambiguity was resolved by the addition of a
silent sign called a determinative, to indicate the category of nouns to which
the intended object belonged. Linguists term this decisive innovation, which
also underlies puns today, the rebus principle.
Once Sumerians had hit upon this phonetic principle, they began to use it
for much more than just writing abstract nouns. They employed it to write
syllables or letters constituting grammatical endings. For instance, in English
it’s not obvious how to draw a picture of the common syllable - tion, but we
could instead draw a picture illustrating the verb shun, which has the same
pronunciation. Phonetically interpreted signs were also used to “spell out”
longer words, as a series of pictures each depicting the sound of one syllable.
That’s as if an English speaker were to write the word believe as a picture of a
bee followed by a picture of a leaf. Phonetic signs also permitted scribes to
use the same pictorial sign for a set of related words (such as tooth, speech,
and speaker), but to resolve the ambiguity with an additional phonetically
interpreted sign (such as selecting the sign for two, each, or peak).
Thus, Sumerian writing came to consist of a complex mixture of three
types of signs: logograms, referring to a whole word or name; phonetic signs,
used in effect for spelling syllables, letters, grammatical elements, or parts of
words; and determinatives, which were not pronounced but were used to
resolve ambiguities. Nevertheless, the phonetic signs in Sumerian writing fell
far short of a complete syllabary or alphabet. Some Sumerian syllables lacked
any written signs; the same sign could be pronounced in different ways; and
the same sign could variously be read as a word, a syllable, or a letter.
Besides Sumerian cuneiform, the other certain instance of independent
origins of writing in human history comes from Native American societies of
Mesoamerica, probably southern Mexico. Mesoamerican writing is believed
to have arisen independently of Old World writing, because there is no
convincing evidence for pre-Norse contact of New World societies with Old
World societies possessing writing. In addition, the forms of Mesoamerican
writing signs were entirely different from those of any Old World script.
About a dozen Mesoamerican scripts are known, all or most of them
apparently related to each other (for example, in their numerical and
calendrical systems), and most of them still only partially deciphered. At the
moment, the earliest preserved Mesoamerican script is from the Zapotec area
of southern Mexico around 600 B.C., but by far the best-understood one is of
the Lowland Maya region, where the oldest known written date corresponds
to A.D. 292.
Despite its independent origins and distinctive sign forms, Maya writing
is organized on principles basically similar to those of Sumerian writing and
other western Eurasian writing systems that Sumerian inspired. Like
Sumerian, Maya writing used both logograms and phonetic signs. Logograms
for abstract words were often derived by the rebus principle. That is, an
abstract word was written with the sign for another word pronounced
similarly but with a different meaning that could be readily depicted. Like the
signs of Japan’s kana and Mycenaean Greece’s Linear B syllabaries, Maya
phonetic signs were mostly signs for syllables of one consonant plus one
vowel (such as ta, te, ti, to, tu). Like letters of the early Semitic alphabet,
Maya syllabic signs were derived from pictures of the object whose
pronunciation began with that syllable (for example, the Maya syllabic sign
“ne” resembles a tail, for which the Maya word is neh).
All of these parallels between Mesoamerican and ancient western
Eurasian writing testify to the underlying universality of human creativity.
While Sumerian and Mesoamerican languages bear no special relation to each
other among the world’s languages, both raised similar basic issues in
reducing them to writing. The solutions that Sumerians invented before 3000
B.C. were reinvented, halfway around the world, by early Mesoamerican
Indians before 600 B.C.
WITH THE POSSIBLE exceptions of the Egyptian, Chinese, and Easter Island
writing to be considered later, all other writing systems devised anywhere in
the world, at any time, appear to have been descendants of systems modified
from or at least inspired by Sumerian or early Mesoamerican writing. One
reason why there were so few independent origins of writing is the great
difficulty of inventing it, as we have already discussed. The other reason is
that other opportunities for the independent invention of writing were
preempted by Sumerian or early Mesoamerican writing and their derivatives.
We know that the development of Sumerian writing took at least
hundreds, possibly thousands, of years. As we shall see, the prerequisites for
those developments consisted of several features of human society that
determined whether a society would find writing useful, and whether the
society could support the necessary specialist scribes. Many other human
societies besides those of the Sumerians and early Mexicans—such as those
of ancient India, Crete, and Ethiopia—evolved these prerequisites. However,
the Sumerians and early Mexicans happened to have been the first to evolve
them in the Old World and the New World, respectively. Once the Sumerians
and early Mexicans had invented writing, the details or principles of their
writing spread rapidly to other societies, before they could go through the
necessary centuries or millennia of independent experimentation with writing
themselves. Thus, that potential for other, independent experiments was
preempted or aborted.
The spread of writing has occurred by either of two contrasting methods,
which find parallels throughout the history of technology and ideas. Someone
invents something and puts it to use. How do you, another would-be user,
then design something similar for your own use, knowing that other people
have already got their own model built and working?
Such transmission of inventions assumes a whole spectrum of forms. At
the one end lies “blueprint copying,” when you copy or modify an available
detailed blueprint. At the opposite end lies “idea diffusion,” when you receive
little more than the basic idea and have to reinvent the details. Knowing that it
can be done stimulates you to try to do it yourself, but your eventual specific
solution may or may not resemble that of the first inventor.
To take a recent example, historians are still debating whether blueprint
copying or idea diffusion contributed more to Russia’s building of an atomic
bomb. Did Russia’s bomb-building efforts depend critically on blueprints of
the already constructed American bomb, stolen and transmitted to Russia by
spies? Or was it merely that the revelation of America’s A-bomb at Hiroshima
at last convinced Stalin of the feasibility of building such a bomb, and that
Russian scientists then reinvented the principles in an independent crash
program, with little detailed guidance from the earlier American effort?
Similar questions arise for the history of the development of wheels,
pyramids, and gunpowder. Let’s now examine how blueprint copying and
idea diffusion contributed to the spread of writing systems.
TODAY, PROFESSIONAL LINGUISTS design writing systems for unwritten
languages by the method of blueprint copying. Most such tailor-made systems
modify existing alphabets, though some instead design syllabaries. For
example, missionary linguists are working on modified Roman alphabets for
hundreds of New Guinea and Native American languages. Government
linguists devised the modified Roman alphabet adopted in 1928 by Turkey for
writing Turkish, as well as the modified Cyrillic alphabets designed for many
tribal languages of Russia.
In a few cases, we also know something about the individuals who
designed writing systems by blueprint copying in the remote past. For
instance, the Cyrillic alphabet itself (the one still used today in Russia) is
descended from an adaptation of Greek and Hebrew letters devised by Saint
Cyril, a Greek missionary to the Slavs in the ninth century A.D. The first
preserved texts for any Germanic language (the language family that includes
English) are in the Gothic alphabet created by Bishop Ulfilas, a missionary
living with the Visigoths in what is now Bulgaria in the fourth century A.D.
Like Saint Cyril’s invention, Ulfilas’s alphabet was a mishmash of letters
borrowed from different sources: about 20 Greek letters, about five Roman
letters, and two letters either taken from the runic alphabet or invented by
Ulfilas himself. Much more often, we know nothing about the individuals
responsible for devising famous alphabets of the past. But it’s still possible to
compare newly emerged alphabets of the past with previously existing ones,
and to deduce from letter forms which existing ones served as models. For the
same reason, we can be sure that the Linear B syllabary of Mycenaean Greece
had been adapted by around 1400 B.C. from the Linear A syllabary of Minoan
Crete.
At all of the hundreds of times when an existing writing system of one
language has been used as a blueprint to adapt to a different language, some
problems have arisen, because no two languages have exactly the same sets of
sounds. Some inherited letters or signs may simply be dropped, when the
sounds that those letters represent in the lending language do not exist in the
borrowing language. For example, Finnish lacks the sounds that many other
European languages express by the letters b, c, f, g, w, x, and z, so the Finns
dropped these letters from their version of the Roman alphabet. There has also
been a frequent reverse problem, of devising letters to represent “new” sounds
present in the borrowing language but absent in the lending language. That
problem has been solved in several different ways: such as using an arbitrary
combination of two or more letters (like the English th to represent a sound
for which the Greek and runic alphabets used a single letter); adding a small
distinguishing mark to an existing letter (like the Spanish tilde ñ, the German
umlaut ö, and the proliferation of marks dancing around Polish and Turkish
letters); co-opting existing letters for which the borrowing language had no
use (such as modern Czechs recycling the letter c of the Roman alphabet to
express the Czech sound ts); or just inventing a new letter (as our medieval
ancestors did when they created the new letters j, u, and w).
The Roman alphabet itself was the end product of a long sequence of
blueprint copying. Alphabets apparently arose only once in human history:
among speakers of Semitic languages, in the area from modern Syria to the
Sinai, during the second millennium B.C. All of the hundreds of historical and
now existing alphabets were ultimately derived from that ancestral Semitic
alphabet, in a few cases (such as the Irish ogham alphabet) by idea diffusion,
but in most by actual copying and modification of letter forms.
That evolution of the alphabet can be traced back to Egyptian
hieroglyphs, which included a complete set of 24 signs for the 24 Egyptian
consonants. The Egyptians never took the logical (to us) next step of
discarding all their logograms, determinatives, and signs for pairs and trios of
consonants, and using just their consonantal alphabet. Starting around 1700
B.C., though, Semites familiar with Egyptian hieroglyphs did begin to
experiment with that logical step.
Restricting signs to those for single consonants was only the first of three
crucial innovations that distinguished alphabets from other writing systems.
The second was to help users memorize the alphabet by placing the letters in
a fixed sequence and giving them easy-to-remember names. Our English
names are mostly meaningless monosyllables (“a,” “bee,” “cee,” “dee,” and
so on). But the Semitic names did possess meaning in Semitic languages: they
were the words for familiar objects (’aleph = ox, beth = house, gimel = camel,
daleth = door, and so on). These Semitic words were related “acrophonically”
to the Semitic consonants to which they refer: that is, the first letter of the
word for the object was also the letter named for the object (’ a, b, g, d, and so
on). In addition, the earliest forms of the Semitic letters appear in many cases
to have been pictures of those same objects. All these features made the
forms, names, and sequence of Semitic alphabet letters easy to remember.
Many modern alphabets, including ours, retain with minor modifications that
original sequence (and, in the case of Greek, even the letters’ original names:
alpha, beta, gamma, delta, and so on) over 3,000 years later. One minor
modification that readers will already have noticed is that the Semitic and
Greek g became the Roman and English c, while the Romans invented a new
g in its present position.
The third and last innovation leading to modern alphabets was to provide
for vowels. Already in the early days of the Semitic alphabet, experiments
began with methods for writing vowels by adding small extra letters to
indicate selected vowels, or else by dots, lines, or hooks sprinkled over the
consonantal letters. In the eighth century B.C. the Greeks became the first
people to indicate all vowels systematically by the same types of letters used
for consonants. Greeks derived the forms of their vowel letters
by “co-opting” five letters used in the Phoenician alphabet for consonantal
sounds lacking in Greek.
From those earliest Semitic alphabets, one line of blueprint copying and
evolutionary modification led via early Arabian alphabets to the modern
Ethiopian alphabet. A far more important line evolved by way of the Aramaic
alphabet, used for official documents of the Persian Empire, into the modern
Arabic, Hebrew, Indian, and Southeast Asian alphabets. But the line most
familiar to European and American readers is the one that led via the
Phoenicians to the Greeks by the early eighth century B.C., thence to the
Etruscans in the same century, and in the next century to the Romans, whose
alphabet with slight modifications is the one used to print this book. Thanks
to their potential advantage of combining precision with simplicity, alphabets
have now been adopted in most areas of the modern world.
WHILE BLUEPRINT COPYING and modification are the most straightforward
option for transmitting technology, that option is sometimes unavailable.
Blueprints may be kept secret, or they may be unreadable to someone not
already steeped in the technology. Word may trickle through about an
invention made somewhere far away, but the details may not get transmitted.
Perhaps only the basic idea is known: someone has succeeded, somehow, in
achieving a certain final result. That knowledge may nevertheless inspire
others, by idea diffusion, to devise their own routes to such a result.
A striking example from the history of writing is the origin of the
syllabary devised in Arkansas around 1820 by a Cherokee Indian named
Sequoyah, for writing the Cherokee language. Sequoyah observed that white
people made marks on paper, and that they derived great advantage by using
those marks to record and repeat lengthy speeches. However, the detailed
operations of those marks remained a mystery to him, since (like most
Cherokees before 1820) Sequoyah was illiterate and could neither speak nor
read English. Because he was a blacksmith, Sequoyah began by devising an
accounting system to help him keep track of his customers’ debts. He drew a
picture of each customer; then he drew circles and lines of various sizes to
represent the amount of money owed.
Around 1810, Sequoyah decided to go on to design a system for writing
the Cherokee language. He again began by drawing pictures, but gave them
up as too complicated and too artistically demanding. He next started to
invent separate signs for each word, and again became dissatisfied when he
had coined thousands of signs and still needed more.
Finally, Sequoyah realized that words were made up of modest numbers
of different sound bites that recurred in many different words—what we
would call syllables. He initially devised 200 syllabic signs and gradually
reduced them to 85, most of them for combinations of one consonant and one
vowel.
As one source of the signs themselves, Sequoyah practiced copying the
letters from an English spelling book given to him by a schoolteacher. About
two dozen of his Cherokee syllabic signs were taken directly from those
letters, though of course with completely changed meanings, since Sequoyah
did not know the English meanings. For example, he chose the shapes D, R,
b, h to represent the Cherokee syllables a, e, si, and ni, respectively, while the
shape of the numeral 4 was borrowed for the syllable se. He coined other
signs by modifying English letters, such as designing the signs , , and to
represent the syllables yu, sa, and na, respectively. Still other signs were
entirely of his creation, such as , , and for ho, li, and nu, respectively.
Sequoyah’s syllabary is widely admired by professional linguists for its good
fit to Cherokee sounds, and for the ease with which it can be learned. Within a
short time, the Cherokees achieved almost 100 percent literacy in the
syllabary, bought a printing press, had Sequoyah’s signs cast as type, and
began printing books and newspapers.
Cherokee writing remains one of the best-attested examples of a script
that arose through idea diffusion. We know that Sequoyah received paper and
other writing materials, the idea of a writing system, the idea of using separate
marks, and the forms of several dozen marks. Since, however, he could
neither read nor write English, he acquired no details or even principles from
the existing scripts around him. Surrounded by alphabets he could not
understand, he instead independently reinvented a syllabary, unaware that the
Minoans of Crete had already invented another syllabary 3,500 years
previously.
SEQUOYAH’S EXAMPLE CAN serve as a model for how idea diffusion probably
led to many writing systems of ancient times as well. The han’gul alphabet
devised by Korea’s King Sejong in A.D. 1446 for the Korean language was
evidently inspired by the block format of Chinese characters and by the
alphabetic principle of Mongol or Tibetan Buddhist writing. However, King
Sejong invented the forms of han’gul letters and several unique features of his
alphabet, including the grouping of letters by syllables into square blocks, the
use of related letter shapes to represent related vowel or consonant sounds,
and shapes of consonant letters that depict the position in which the lips or
tongue are held to pronounce that consonant. The ogham alphabet used in
Ireland and parts of Celtic Britain from around the fourth century A.D.
similarly adopted the alphabetic principle (in this case, from existing
European alphabets) but again devised unique letter forms, apparently based
on a five-finger system of hand signals.
We can confidently attribute the han’gul and ogham alphabets to idea
diffusion rather than to independent invention in isolation, because we know
that both societies were in close contact with societies possessing writing and
because it is clear which foreign scripts furnished the inspiration. In contrast,
we can confidently attribute Sumerian cuneiform and the earliest
Mesoamerican writing to independent invention, because at the times of their
first appearances there existed no other script in their respective hemispheres
that could have inspired them. Still debatable are the origins of writing on
Easter Island, in China, and in Egypt.
The Polynesians living on Easter Island, in the Pacific Ocean, had a
unique script of which the earliest preserved examples date back only to about
A.D. 1851, long after Europeans reached Easter in 1722. Perhaps writing arose
independently on Easter before the arrival of Europeans, although no
examples have survived. But the most straightforward interpretation is to take
the facts at face value, and to assume that Easter Islanders were stimulated to
devise a script after seeing the written proclamation of annexation that a
Spanish expedition handed to them in the year 1770.
As for Chinese writing, first attested around 1300 B.C. but with possible
earlier precursors, it too has unique local signs and some unique principles,
and most scholars assume that it evolved independently. Writing had
developed before 3000 B.C. in Sumer, 4,000 miles west of early Chinese urban
centers, and appeared by 2200 B.C. in the Indus Valley, 2,600 miles west, but
no early writing systems are known from the whole area between the Indus
Valley and China. Thus, there is no evidence that the earliest Chinese scribes
could have had knowledge of any other writing system to inspire them.
Egyptian hieroglyphics, the most famous of all ancient writing systems,
are also usually assumed to be the product of independent invention, but the
alternative interpretation of idea diffusion is more feasible than in the case of
Chinese writing. Hieroglyphic writing appeared rather suddenly, in nearly
full-blown form, around 3000 B.C. Egypt lay only 800 miles west of Sumer,
with which Egypt had trade contacts. I find it suspicious that no evidence of a
gradual development of hieroglyphs has come down to us, even though
Egypt’s dry climate would have been favorable for preserving earlier
experiments in writing, and though the similarly dry climate of Sumer has
yielded abundant evidence of the development of Sumerian cuneiform for at
least several centuries before 3000 B.C. Equally suspicious is the appearance
of several other, apparently independently designed, writing systems in Iran,
Crete, and Turkey (so-called proto-Elamite writing, Cretan pictographs, and
Hieroglyphic Hittite, respectively), after the rise of Sumerian and Egyptian
writing. Although each of those systems used distinctive sets of signs not
borrowed from Egypt or Sumer, the peoples involved could hardly have been
unaware of the writing of their neighboring trade partners.
It would be a remarkable coincidence if, after millions of years of human
existence without writing, all those Mediterranean and Near Eastern societies
had just happened to hit independently on the idea of writing within a few
centuries of each other. Hence a possible interpretation seems to me idea
diffusion, as in the case of Sequoyah’s syllabary. That is, Egyptians and other
peoples may have learned from Sumerians about the idea of writing and
possibly about some of the principles, and then devised other principles and
all the specific forms of the letters for themselves.
LET US NOW return to the main question with which we began this chapter:
why did writing arise in and spread to some societies, but not to many others?
Convenient starting points for our discussion are the limited capabilities, uses,
and users of early writing systems.
Early scripts were incomplete, ambiguous, or complex, or all three. For
example, the oldest Sumerian cuneiform writing could not render normal
prose but was a mere telegraphic shorthand, whose vocabulary was restricted
to names, numerals, units of measure, words for objects counted, and a few
adjectives. That’s as if a modern American court clerk were forced to write
“John 27 fat sheep,” because English writing lacked the necessary words and
grammar to write “We order John to deliver the 27 fat sheep that he owes to
the government.” Later Sumerian cuneiform did become capable of rendering
prose, but it did so by the messy system that I’ve already described, with
mixtures of logograms, phonetic signs, and unpronounced determinatives
totaling hundreds of separate signs. Linear B, the writing of Mycenaean
Greece, was at least simpler, being based on a syllabary of about 90 signs plus
logograms. Offsetting that virtue, Linear B was quite ambiguous. It omitted
any consonant at the end of a word, and it used the same sign for several
related consonants (for instance, one sign for both l and r, another for p and b and ph, and still another for g and k and kh). We know how confusing we find it when native-born Japanese people speak English without distinguishing l
and r: imagine the confusion if our alphabet did the same while similarly
homogenizing the other consonants that I mentioned! It’s as if we were to
spell the words “rap,” “lap,” “lab,” and “laugh” identically.
A related limitation is that few people ever learned to write these early
scripts. Knowledge of writing was confined to professional scribes in the
employ of the king or temple. For instance, there is no hint that Linear B was
used or understood by any Mycenaean Greek beyond small cadres of palace
bureaucrats. Since individual Linear B scribes can be distinguished by their
handwriting on preserved documents, we can say that all preserved Linear B
documents from the palaces of Knossos and Pylos are the work of a mere 75
and 40 scribes, respectively.
The uses of these telegraphic, clumsy, ambiguous early scripts were as
restricted as the number of their users. Anyone hoping to discover how
Sumerians of 3000 B.C. thought and felt is in for a disappointment. Instead, the
first Sumerian texts are emotionless accounts of palace and temple
bureaucrats. About 90 percent of the tablets in the earliest known Sumerian
archives, from the city of Uruk, are clerical records of goods paid in, workers
given rations, and agricultural products distributed. Only later, as Sumerians
progressed beyond logograms to phonetic writing, did they begin to write
prose narratives, such as propaganda and myths.
Mycenaean Greeks never even reached that propaganda-and-myths stage.
One-third of all Linear B tablets from the palace of Knossos are accountants’
records of sheep and wool, while an inordinate proportion of writing at the
palace of Pylos consists of records of flax. Linear B was inherently so
ambiguous that it remained restricted to palace accounts, whose context and
limited word choices made the interpretation clear. Not a trace of its use for
literature has survived. The Iliad and Odyssey were composed and transmitted
by nonliterate bards for nonliterate listeners, and not committed to writing
until the development of the Greek alphabet hundreds of years later.
Similarly restricted uses characterize early Egyptian, Mesoamerican, and
Chinese writing. Early Egyptian hieroglyphs recorded religious and state
propaganda and bureaucratic accounts. Preserved Maya writing was similarly
devoted to propaganda, births and accessions and victories of kings, and
astronomical observations of priests. The oldest preserved Chinese writing of
the late Shang Dynasty consists of religious divination about dynastic affairs,
incised into so-called oracle bones. A sample Shang text: “The king, reading
the meaning of the crack [in a bone cracked by heating], said: ‘If the child is
born on a keng day, it will be extremely auspicious.’”
To us today, it is tempting to ask why societies with early writing systems
accepted the ambiguities that restricted writing to a few functions and a few
scribes. But even to pose that question is to illustrate the gap between ancient
perspectives and our own expectations of mass literacy. The intended
restricted uses of early writing provided a positive disincentive for devising
less ambiguous writing systems. The kings and priests of ancient Sumer
wanted writing to be used by professional scribes to record numbers of sheep
owed in taxes, not by the masses to write poetry and hatch plots. As the
anthropologist Claude Lévi-Strauss put it, ancient writing’s main function was
“to facilitate the enslavement of other human beings.” Personal uses of
writing by nonprofessionals came only much later, as writing systems grew
simpler and more expressive.
For instance, with the fall of Mycenaean Greek civilization, around 1200
B.C., Linear B disappeared, and Greece returned to an age of preliteracy. When
writing finally returned to Greece, in the eighth century B.C., the new Greek
writing, its users, and its uses were very different. The writing was no longer
an ambiguous syllabary mixed with logograms but an alphabet borrowed
from the Phoenician consonantal alphabet and improved by the Greek
invention of vowels. In place of lists of sheep, legible only to scribes and read
only in palaces, Greek alphabetic writing from the moment of its appearance
was a vehicle of poetry and humor, to be read in private homes. For instance,
the first preserved example of Greek alphabetic writing, scratched onto an
Athenian wine jug of about 740 B.C., is a line of poetry announcing a dancing
contest: “Whoever of all dancers performs most nimbly will win this vase as a
prize.” The next example is three lines of dactylic hexameter scratched onto a
drinking cup: “I am Nestor’s delicious drinking cup. Whoever drinks from
this cup swiftly will the desire of fair-crowned Aphrodite seize him.” The
earliest preserved examples of the Etruscan and Roman alphabets are also
inscriptions on drinking cups and wine containers. Only later did the
alphabet’s easily learned vehicle of private communication become co-opted
for public or bureaucratic purposes. Thus, the developmental sequence of uses
for alphabetic writing was the reverse of that for the earlier systems of
logograms and syllabaries.
THE LIMITED USES and users of early writing suggest why writing appeared
so late in human evolution. All of the likely or possible independent
inventions of writing (in Sumer, Mexico, China, and Egypt), and all of the
early adaptations of those invented systems (for example, those in Crete, Iran,
Turkey, the Indus Valley, and the Maya area), involved socially stratified
societies with complex and centralized political institutions, whose necessary
relation to food production we shall explore in a later chapter. Early writing
served the needs of those political institutions (such as record keeping and
royal propaganda), and the users were full-time bureaucrats nourished by
stored food surpluses grown by food-producing peasants. Writing was never
developed or even adopted by hunter-gatherer societies, because they lacked
both the institutional uses of early writing and the social and agricultural
mechanisms for generating the food surpluses required to feed scribes.
Thus, food production and thousands of years of societal evolution
following its adoption were as essential for the evolution of writing as for the
evolution of microbes causing human epidemic diseases. Writing arose
independently only in the Fertile Crescent, Mexico, and probably China
precisely because those were the first areas where food production emerged in
their respective hemispheres. Once writing had been invented by those few
societies, it then spread, by trade and conquest and religion, to other societies
with similar economies and political organizations.
While food production was thus a necessary condition for the evolution or
early adoption of writing, it was not a sufficient condition. At the beginning
of this chapter, I mentioned the failure of some food-producing societies with
complex political organization to develop or adopt writing before modern
times. Those cases, initially so puzzling to us moderns accustomed to viewing
writing as indispensable to a complex society, included one of the world’s
largest empires as of A.D. 1520, the Inca Empire of South America. They also
included Tonga’s maritime proto-empire, the Hawaiian state emerging in the
late 18th century, all of the states and chiefdoms of subequatorial Africa and
sub-Saharan West Africa before the arrival of Islam, and the largest native
North American societies, those of the Mississippi Valley and its tributaries.
Why did all those societies fail to acquire writing, despite their sharing
prerequisites with societies that did do so?
Here we have to remind ourselves that the vast majority of societies with
writing acquired it by borrowing it from neighbors or by being inspired by
them to develop it, rather than by independently inventing it themselves. The
societies without writing that I just mentioned are ones that got a later start on
food production than did Sumer, Mexico, and China. (The only uncertainty in
this statement concerns the relative dates for the onset of food production in
Mexico and in the Andes, the eventual Inca realm.) Given enough time, the
societies lacking writing might also have eventually developed it on their
own. Had they been located nearer to Sumer, Mexico, and China, they might
instead have acquired writing or the idea of writing from those centers, just as
did India, the Maya, and most other societies with writing. But they were too
far from the first centers of writing to have acquired it before modern times.
The importance of isolation is most obvious for Hawaii and Tonga, both
of which were separated by at least 4,000 miles of ocean from the nearest
societies with writing. The other societies illustrate the important point that
distance as the crow flies is not an appropriate measure of isolation for
humans. The Andes, West Africa’s kingdoms, and the mouth of the
Mississippi River lay only about 1,200, 1,500, and 700 miles, respectively,
from societies with writing in Mexico, North Africa, and Mexico,
respectively. These distances are considerably less than the distances the
alphabet had to travel from its homeland on the eastern shores of the
Mediterranean to reach Ireland, Ethiopia, and Southeast Asia within 2,000
years of its invention. But humans are slowed by ecological and water barriers
that crows can fly over. The states of North Africa (with writing) and West
Africa (without writing) were separated from each other by Saharan desert
unsuitable for agriculture and cities. The deserts of northern Mexico similarly
separated the urban centers of southern Mexico from the chiefdoms of the
Mississippi Valley. Communication between southern Mexico and the Andes
required either a sea voyage or else a long chain of overland contacts via the
narrow, forested, never urbanized Isthmus of Darien. Hence the Andes, West
Africa, and the Mississippi Valley were effectively rather isolated from
societies with writing.
That’s not to say that those societies without writing were totally isolated.
West Africa eventually did receive Fertile Crescent domestic animals across
the Sahara, and later accepted Islamic influence, including Arabic writing.
Corn diffused from Mexico to the Andes and, more slowly, from Mexico to
the Mississippi Valley. But we already saw in Chapter 10 that the north-south
axes and ecological barriers within Africa and the Americas retarded the
diffusion of crops and domestic animals. The history of writing illustrates
strikingly the similar ways in which geography and ecology influenced the
spread of human inventions.
CHAPTER 13
NECESSITY’S MOTHER
ON JULY 3, 1908, ARCHAEOLOGISTS EXCAVATING THE ancient Minoan palace at
Phaistos, on the island of Crete, chanced upon one of the most remarkable
objects in the history of technology. At first glance it seemed
unprepossessing: just a small, flat, unpainted, circular disk of hard-baked clay,
6½ inches in diameter. Closer examination showed each side to be covered
with writing, resting on a curved line that spiraled clockwise in five coils
from the disk’s rim to its center. A total of 241 signs or letters was neatly
divided by etched vertical lines into groups of several signs, possibly
constituting words. The writer must have planned and executed the disk with
care, so as to start writing at the rim and fill up all the available space along
the spiraling line, yet not run out of space on reaching the center (Chapter 13).
Ever since it was unearthed, the disk has posed a mystery for historians of
writing. The number of distinct signs (45) suggests a syllabary rather than an
alphabet, but it is still undeciphered, and the forms of the signs are unlike
those of any other known writing system. Not another scrap of the strange
script has turned up in the 89 years since its discovery. Thus, it remains
unknown whether it represents an indigenous Cretan script or a foreign import
to Crete.
For historians of technology, the Phaistos disk is even more baffling; its
estimated date of 1700 B.C. makes it by far the earliest printed document in the
world. Instead of being etched by hand, as were all texts of Crete’s later
Linear A and Linear B scripts, the disk’s signs were punched into soft clay
(subsequently baked hard) by stamps that bore a sign as raised type. The
printer evidently had a set of at least 45 stamps, one for each sign appearing
on the disk. Making these stamps must have entailed a great deal of work, and
they surely weren’t manufactured just to print this single document. Whoever
used them was presumably doing a lot of writing. With those stamps, their
owner could make copies much more quickly and neatly than if he or she had
written out each of the script’s complicated signs at each appearance.
The Phaistos disk anticipates humanity’s next efforts at printing, which
similarly used cut type or blocks but applied them to paper with ink, not to
clay without ink. However, those next efforts did not appear until 2,500 years
later in China and 3,100 years later in medieval Europe. Why was the disk’s
precocious technology not widely adopted in Crete or elsewhere in the ancient
Mediterranean? Why was its printing method invented around 1700 B.C. in
Crete and not at some other time in Mesopotamia, Mexico, or any other
ancient center of writing? Why did it then take thousands of years to add the
ideas of ink and a press and arrive at a printing press? The disk thus
constitutes a threatening challenge to historians. If inventions are as
idiosyncratic and unpredictable as the disk seems to suggest, then efforts to
generalize about the history of technology may be doomed from the outset.
Technology, in the form of weapons and transport, provides the direct
means by which certain peoples have expanded their realms and conquered
other peoples. That makes it the leading cause of history’s broadest pattern.
But why were Eurasians, rather than Native Americans or sub-Saharan
Africans, the ones to invent firearms, oceangoing ships, and steel equipment?
The differences extend to most other significant technological advances, from
printing presses to glass and steam engines. Why were all those inventions
Eurasian? Why were all New Guineans and Native Australians in A.D. 1800
still using stone tools like ones discarded thousands of years ago in Eurasia
and most of Africa, even though some of the world’s richest copper and iron
deposits are in New Guinea and Australia, respectively? All those facts
explain why so many laypeople assume that Eurasians are superior to other
peoples in inventiveness and intelligence.
If, on the other hand, no such difference in human neurobiology exists to
account for continental differences in technological development, what does
account for them? An alternative view rests on the heroic theory of invention.
Technological advances seem to come disproportionately from a few very
rare geniuses, such as Johannes Gutenberg, James Watt, Thomas Edison, and
the Wright brothers. They were Europeans, or descendants of European
emigrants to America. So were Archimedes and other rare geniuses of ancient
times. Could such geniuses have equally well been born in Tasmania or
Namibia? Does the history of technology depend on nothing more than
accidents of the birthplaces of a few inventors?
Still another alternative view holds that it is a matter not of individual
inventiveness but of the receptivity of whole societies to innovation. Some
societies seem hopelessly conservative, inward looking, and hostile to change.
That’s the impression of many Westerners who have attempted to help Third
World peoples and ended up discouraged. The people seem perfectly
intelligent as individuals; the problem seems instead to lie with their societies.
How else can one explain why the Aborigines of northeastern Australia failed
to adopt bows and arrows, which they saw being used by Torres Straits
islanders with whom they traded? Might all the societies of an entire
continent be unreceptive, thereby explaining technology’s slow pace of
development there? In this chapter we shall finally come to grips with a
central problem of this book: the question of why technology did evolve at
such different rates on different continents.
THE STARTING POINT for our discussion is the common view expressed in the
saying “Necessity is the mother of invention.” That is, inventions supposedly
arise when a society has an unfulfilled need: some technology is widely
recognized to be unsatisfactory or limiting. Would-be inventors, motivated by
the prospect of money or fame, perceive the need and try to meet it. Some
inventor finally comes up with a solution superior to the existing,
unsatisfactory technology. Society adopts the solution if it is compatible with
the society’s values and other technologies.
Quite a few inventions do conform to this commonsense view of
necessity as invention’s mother. In 1942, in the middle of World War II, the
U.S. government set up the Manhattan Project with the explicit goal of
inventing the technology required to build an atomic bomb before Nazi
Germany could do so. That project succeeded in three years, at a cost of $2
billion (equivalent to over $20 billion today). Other instances are Eli
Whitney’s 1794 invention of his cotton gin to replace laborious hand cleaning
of cotton grown in the U.S. South, and James Watt’s 1769 invention of his
steam engine to solve the problem of pumping water out of British coal
mines.
These familiar examples deceive us into assuming that other major
inventions were also responses to perceived needs. In fact, many or most
inventions were developed by people driven by curiosity or by a love of
tinkering, in the absence of any initial demand for the product they had in
mind. Once a device had been invented, the inventor then had to find an
application for it. Only after it had been in use for a considerable time did
consumers come to feel that they “needed” it. Still other devices, invented to
serve one purpose, eventually found most of their use for other, unanticipated
purposes. It may come as a surprise to learn that these inventions in search of
a use include most of the major technological breakthroughs of modern times,
ranging from the airplane and automobile, through the internal combustion
engine and electric light bulb, to the phonograph and transistor. Thus,
invention is often the mother of necessity, rather than vice versa.
A good example is the history of Thomas Edison’s phonograph, the most
original invention of the greatest inventor of modern times. When Edison
built his first phonograph in 1877, he published an article proposing ten uses
to which his invention might be put. They included preserving the last words
of dying people, recording books for blind people to hear, announcing clock
time, and teaching spelling. Reproduction of music was not high on Edison’s
list of priorities. A few years later Edison told his assistant that his invention
had no commercial value. Within another few years he changed his mind and
did enter business to sell phonographs—but for use as office dictating
machines. When other entrepreneurs created jukeboxes by arranging for a
phonograph to play popular music at the drop of a coin, Edison objected to
this debasement, which apparently detracted from serious office use of his
invention. Only after about 20 years did Edison reluctantly concede that the
main use of his phonograph was to record and play music.
The motor vehicle is another invention whose uses seem obvious today.
However, it was not invented in response to any demand. When Nikolaus
Otto built his first gas engine, in 1866, horses had been supplying people’s
land transportation needs for nearly 6,000 years, supplemented increasingly
by steam-powered railroads for several decades. There was no crisis in the
availability of horses, no dissatisfaction with railroads.
Because Otto’s engine was weak, heavy, and seven feet tall, it did not
recommend itself over horses. Not until 1885 did engines improve to the point
that Gottfried Daimler got around to installing one on a bicycle to create the
first motorcycle; he waited until 1896 to build the first truck.
In 1905, motor vehicles were still expensive, unreliable toys for the rich.
Public contentment with horses and railroads remained high until World War
I, when the military concluded that it really did need trucks. Intensive postwar
lobbying by truck manufacturers and armies finally convinced the public of
its own needs and enabled trucks to begin to supplant horse-drawn wagons in
industrialized countries. Even in the largest American cities, the changeover
took 50 years.
Inventors often have to persist at their tinkering for a long time in the
absence of public demand, because early models perform too poorly to be
useful. The first cameras, typewriters, and television sets were as awful as
Otto’s seven-foot-tall gas engine. That makes it difficult for an inventor to
foresee whether his or her awful prototype might eventually find a use and
thus warrant more time and expense to develop it. Each year, the United
States issues about 70,000 patents, only a few of which ultimately reach the
stage of commercial production. For each great invention that ultimately
found a use, there are countless others that did not. Even inventions that meet
the need for which they were initially designed may later prove more valuable
at meeting unforeseen needs. While James Watt designed his steam engine to
pump water from mines, it soon was supplying power to cotton mills, then
(with much greater profit) propelling locomotives and boats.
THUS, THE COMMONSENSE view of invention that served as our starting point
reverses the usual roles of invention and need. It also overstates the
importance of rare geniuses, such as Watt and Edison. That “heroic theory of
invention,” as it is termed, is encouraged by patent law, because an applicant
for a patent must prove the novelty of the invention submitted. Inventors
thereby have a financial incentive to denigrate or ignore previous work. From
a patent lawyer’s perspective, the ideal invention is one that arises without
any precursors, like Athene springing fully formed from the forehead of Zeus.
In reality, even for the most famous and apparently decisive modern
inventions, neglected precursors lurked behind the bald claim “X invented Y.”
For instance, we are regularly told, “James Watt invented the steam engine in
1769,” supposedly inspired by watching steam rise from a teakettle’s spout.
Unfortunately for this splendid fiction, Watt actually got the idea for his
particular steam engine while repairing a model of Thomas Newcomen’s
steam engine, which Newcomen had invented 57 years earlier and of which
over a hundred had been manufactured in England by the time of Watt’s
repair work. Newcomen’s engine, in turn, followed the steam engine that the
Englishman Thomas Savery patented in 1698, which followed the steam
engine that the Frenchman Denis Papin designed (but did not build) around
1680, which in turn had precursors in the ideas of the Dutch scientist
Christiaan Huygens and others. All this is not to deny that Watt greatly
improved Newcomen’s engine (by incorporating a separate steam condenser
and a double-acting cylinder), just as Newcomen had greatly improved
Savery’s.
Similar histories can be related for all modern inventions that are
adequately documented. The hero customarily credited with the invention
followed previous inventors who had had similar aims and had already
produced designs, working models, or (as in the case of the Newcomen steam
engine) commercially successful models. Edison’s famous “invention” of the
incandescent light bulb on the night of October 21, 1879, improved on many
other incandescent light bulbs patented by other inventors between 1841 and
1878. Similarly, the Wright brothers’ manned powered airplane was preceded
by the manned unpowered gliders of Otto Lilienthal and the unmanned
powered airplane of Samuel Langley; Samuel Morse’s telegraph was
preceded by those of Joseph Henry, William Cooke, and Charles Wheatstone;
and Eli Whitney’s gin for cleaning short-staple (inland) cotton extended gins
that had been cleaning long-staple (Sea Island) cotton for thousands of years.
All this is not to deny that Watt, Edison, the Wright brothers, Morse, and
Whitney made big improvements and thereby increased or inaugurated
commercial success. The form of the invention eventually adopted might have
been somewhat different without the recognized inventor’s contribution. But
the question for our purposes is whether the broad pattern of world history
would have been altered significantly if some genius inventor had not been
born at a particular place and time. The answer is clear: there has never been
any such person. All recognized famous inventors had capable predecessors
and successors and made their improvements at a time when society was
capable of using their product. As we shall see, the tragedy of the hero who
perfected the stamps used for the Phaistos disk was that he or she devised
something that the society of the time could not exploit on a large scale.
MY EXAMPLES SO far have been drawn from modern technologies, because
their histories are well known. My two main conclusions are that technology
develops cumulatively, rather than in isolated heroic acts, and that it finds
most of its uses after it has been invented, rather than being invented to meet
a foreseen need. These conclusions surely apply with much greater force to
the undocumented history of ancient technology. When Ice Age hunter-
gatherers noticed burned sand and limestone residues in their hearths, it was
impossible for them to foresee the long, serendipitous accumulation of
discoveries that would lead to the first Roman glass windows (around A.D. 1),
by way of the first objects with surface glazes (around 4000 B.C.), the first
free-standing glass objects of Egypt and Mesopotamia (around 2500 B.C.), and
the first glass vessels (around 1500 B.C.).
We know nothing about how those earliest known surface glazes
themselves were developed. Nevertheless, we can infer the methods of
prehistoric invention by watching technologically “primitive” people today,
such as the New Guineans with whom I work. I already mentioned their
knowledge of hundreds of local plant and animal species and each species’
edibility, medical value, and other uses. New Guineans told me similarly
about dozens of rock types in their environment and each type’s hardness,
color, behavior when struck or flaked, and uses. All of that knowledge is
acquired by observation and by trial and error. I see that process of
“invention” going on whenever I take New Guineans to work with me in an
area away from their homes. They constantly pick up unfamiliar things in the
forest, tinker with them, and occasionally find them useful enough to bring
home. I see the same process when I am abandoning a campsite, and local
people come to scavenge what is left. They play with my discarded objects
and try to figure out whether they might be useful in New Guinea society.
Discarded tin cans are easy: they end up reused as containers. Other objects
are tested for purposes very different from the one for which they were
manufactured. How would that yellow number 2 pencil look as an ornament,
inserted through a pierced ear-lobe or nasal septum? Is that piece of broken
glass sufficiently sharp and strong to be useful as a knife? Eureka!
The raw substances available to ancient peoples were natural materials
such as stone, wood, bone, skins, fiber, clay, sand, limestone, and minerals, all
existing in great variety. From those materials, people gradually learned to
work particular types of stone, wood, and bone into tools; to convert
particular clays into pottery and bricks; to convert certain mixtures of sand,
limestone, and other “dirt” into glass; and to work available pure soft metals
such as copper and gold, then to extract metals from ores, and finally to work
hard metals such as bronze and iron.
A good illustration of the histories of trial and error involved is furnished
by the development of gunpowder and gasoline from raw materials.
Combustible natural products inevitably make themselves noticed, as when a
resinous log explodes in a campfire. By 2000 B.C., Mesopotamians were
extracting tons of petroleum by heating rock asphalt. Ancient Greeks
discovered the uses of various mixtures of petroleum, pitch, resins, sulfur, and
quicklime as incendiary weapons, delivered by catapults, arrows, firebombs,
and ships. The expertise at distillation that medieval Islamic alchemists
developed to produce alcohols and perfumes also let them distill petroleum
into fractions, some of which proved to be even more powerful incendiaries.
Delivered in grenades, rockets, and torpedoes, those incendiaries played a key
role in Islam’s eventual defeat of the Crusaders. By then, the Chinese had
observed that a particular mixture of sulfur, charcoal, and saltpeter, which
became known as gunpowder, was especially explosive. An Islamic chemical
treatise of about A.D. 1100 describes seven gunpowder recipes, while a treatise
from A.D. 1280 gives more than 70 recipes that had proved suitable for diverse
purposes (one for rockets, another for cannons).
As for postmedieval petroleum distillation, 19th-century chemists found
the middle distillate fraction useful as fuel for oil lamps. The chemists
discarded the most volatile fraction (gasoline) as an unfortunate waste product
—until it was found to be an ideal fuel for internal-combustion engines. Who
today remembers that gasoline, the fuel of modern civilization, originated as
yet another invention in search of a use?
ONCE AN INVENTOR has discovered a use for a new technology, the next step
is to persuade society to adopt it. Merely having a bigger, faster, more
powerful device for doing something is no guarantee of ready acceptance.
Innumerable such technologies were either not adopted at all or adopted only
after prolonged resistance. Notorious examples include the U.S. Congress’s
rejection of funds to develop a supersonic transport in 1971, the world’s
continued rejection of an efficiently designed typewriter keyboard, and
Britain’s long reluctance to adopt electric lighting. What is it that promotes an
invention’s acceptance by a society?
Let’s begin by comparing the acceptability of different inventions within
the same society. It turns out that at least four factors influence acceptance.
The first and most obvious factor is relative economic advantage
compared with existing technology. While wheels are very useful in modern
industrial societies, that has not been so in some other societies. Ancient
Native Mexicans invented wheeled vehicles with axles for use as toys, but not
for transport. That seems incredible to us, until we reflect that ancient
Mexicans lacked domestic animals to hitch to their wheeled vehicles, which
therefore offered no advantage over human porters.
A second consideration is social value and prestige, which can override
economic benefit (or lack thereof). Millions of people today buy designer
jeans for double the price of equally durable generic jeans—because the
social cachet of the designer label counts for more than the extra cost.
Similarly, Japan continues to use its horrendously cumbersome kanji writing
system in preference to efficient alphabets or Japan’s own efficient kana
syllabary—because the prestige attached to kanji is so great.
Still another factor is compatibility with vested interests. This book, like
probably every other typed document you have ever read, was typed with a
QWERTY keyboard, named for the left-most six letters in its upper row.
Unbelievable as it may now sound, that keyboard layout was designed in
1873 as a feat of anti-engineering. It employs a whole series of perverse tricks
designed to force typists to type as slowly as possible, such as scattering the
commonest letters over all keyboard rows and concentrating them on the left
side (where right-handed people have to use their weaker hand). The reason
behind all of those seemingly counterproductive features is that the
typewriters of 1873 jammed if adjacent keys were struck in quick succession,
so that manufacturers had to slow down typists. When improvements in
typewriters eliminated the problem of jamming, trials in 1932 with an
efficiently laid-out keyboard showed that it would let us double our typing
speed and reduce our typing effort by 95 percent. But QWERTY keyboards
were solidly entrenched by then. The vested interests of hundreds of millions
of QWERTY typists, typing teachers, typewriter and computer salespeople,
and manufacturers have crushed all moves toward keyboard efficiency for
over 60 years.
While the story of the QWERTY keyboard may sound funny, many
similar cases have involved much heavier economic consequences. Why does
Japan now dominate the world market for transistorized electronic consumer
products, to a degree that damages the United States’s balance of payments
with Japan, even though transistors were invented and patented in the United
States? Because Sony bought transistor licensing rights from Western Electric
at a time when the American electronics consumer industry was churning out
vacuum tube models and reluctant to compete with its own products. Why
were British cities still using gas street lighting into the 1920s, long after U.S.
and German cities had converted to electric street lighting? Because British
municipal governments had invested heavily in gas lighting and placed
regulatory obstacles in the way of the competing electric light companies.
The remaining consideration affecting acceptance of new technologies is
the ease with which their advantages can be observed. In A.D. 1340, when
firearms had not yet reached most of Europe, England’s earl of Derby and earl
of Salisbury happened to be present in Spain at the battle of Tarifa, where
Arabs used cannons against the Spaniards. Impressed by what they saw, the
earls introduced cannons to the English army, which adopted them
enthusiastically and already used them against French soldiers at the battle of
Crécy six years later.
THUS, WHEELS, DESIGNER jeans, and QWERTY keyboards illustrate the varied
reasons why the same society is not equally receptive to all inventions.
Conversely, the same invention’s reception also varies greatly among
contemporary societies. We are all familiar with the supposed generalization
that rural Third World societies are less receptive to innovation than are
Westernized industrial societies. Even within the industrialized world, some
areas are much more receptive than others. Such differences, if they existed
on a continental scale, might explain why technology developed faster on
some continents than on others. For instance, if all Aboriginal Australian
societies were for some reason uniformly resistant to change, that might
account for their continued use of stone tools after metal tools had appeared
on every other continent. How do differences in receptivity among societies
arise?
A laundry list of at least 14 explanatory factors has been proposed by
historians of technology. One is long life expectancy, which in principle
should give prospective inventors the years necessary to accumulate technical
knowledge, as well as the patience and security to embark on long
development programs yielding delayed rewards. Hence the greatly increased
life expectancy brought by modern medicine may have contributed to the
recently accelerating pace of invention.
The next five factors involve economics or the organization of society: (1)
The availability of cheap slave labor in classical times supposedly
discouraged innovation then, whereas high wages or labor scarcity now
stimulate the search for technological solutions. For example, the prospect of
changed immigration policies that would cut off the supply of cheap Mexican
seasonal labor to Californian farms was the immediate incentive for the
development of a machine-harvestable variety of tomatoes in California. (2)
Patents and other property laws, protecting ownership rights of inventors,
reward innovation in the modern West, while the lack of such protection
discourages it in modern China. (3) Modern industrial societies provide
extensive opportunities for technical training, as medieval Islam did and
modern Zaire does not. (4) Modern capitalism is, and the ancient Roman
economy was not, organized in a way that made it potentially rewarding to
invest capital in technological development. (5) The strong individualism of
U.S. society allows successful inventors to keep earnings for themselves,
whereas strong family ties in New Guinea ensure that someone who begins to
earn money will be joined by a dozen relatives expecting to move in and be
fed and supported.
Another four suggested explanations are ideological, rather than
economic or organizational: (1) Risk-taking behavior, essential for efforts at
innovation, is more widespread in some societies than in others. (2) The
scientific outlook is a unique feature of post-Renaissance European society
that has contributed heavily to its modern technological preeminence. (3)
Tolerance of diverse views and of heretics fosters innovation, whereas a
strongly traditional outlook (as in China’s emphasis on ancient Chinese
classics) stifles it. (4) Religions vary greatly in their relation to technological
innovation: some branches of Judaism and Christianity are claimed to be
especially compatible with it, while some branches of Islam, Hinduism, and
Brahmanism may be especially incompatible with it.
All ten of these hypotheses are plausible. But none of them has any
necessary association with geography. If patent rights, capitalism, and certain
religions do promote technology, what selected for those factors in
postmedieval Europe but not in contemporary China or India?
At least the direction in which those ten factors influence technology
seems clear. The remaining four proposed factors—war, centralized
government, climate, and resource abundance—appear to act inconsistently:
sometimes they stimulate technology, sometimes they inhibit it. (1)
Throughout history, war has often been a leading stimulant of technological
innovation. For instance, the enormous investments made in nuclear weapons
during World War II and in airplanes and trucks during World War I launched
whole new fields of technology. But wars can also deal devastating setbacks
to technological development. (2) Strong centralized government boosted
technology in late-19th-century Germany and Japan, and crushed it in China
after A.D. 1500. (3) Many northern Europeans assume that technology thrives
in a rigorous climate where survival is impossible without technology, and
withers in a benign climate where clothing is unnecessary and bananas
supposedly fall off the trees. An opposite view is that benign environments
leave people free from the constant struggle for existence, free to devote
themselves to innovation. (4) There has also been debate over whether
technology is stimulated by abundance or by scarcity of environmental
resources. Abundant resources might stimulate the development of inventions
utilizing those resources, such as water mill technology in rainy northern
Europe, with its many rivers—but why didn’t water mill technology progress
more rapidly in even rainier New Guinea? The destruction of Britain’s forests
has been suggested as the reason behind its early lead in developing coal
technology, but why didn’t deforestation have the same effect in China?
This discussion does not exhaust the list of reasons proposed to explain
why societies differ in their receptivity to new technology. Worse yet, all of
these proximate explanations bypass the question of the ultimate factors
behind them. This may seem like a discouraging setback in our attempt to
understand the course of history, since technology has undoubtedly been one
of history’s strongest forces. However, I shall now argue that the diversity of
independent factors behind technological innovation actually makes it easier,
not harder, to understand history’s broad pattern.
FOR THE PURPOSES of this book, the key question about the laundry list is
whether such factors differed systematically from continent to continent and
thereby led to continental differences in technological development. Most
laypeople and many historians assume, expressly or tacitly, that the answer is
yes. For example, it is widely believed that Australian Aborigines as a group
shared ideological characteristics contributing to their technological
backwardness: they were (or are) supposedly conservative, living in an
imagined past Dreamtime of the world’s creation, and not focused on
practical ways to improve the present. A leading historian of Africa
characterized Africans as inward looking and lacking Europeans’ drive for
expansion.
But all such claims are based on pure speculation. There has never been a
study of many societies under similar socioeconomic conditions on each of
two continents, demonstrating systematic ideological differences between the
two continents’ peoples. The usual reasoning is instead circular: because
technological differences exist, the existence of corresponding ideological
differences is inferred.
In reality, I regularly observe in New Guinea that native societies there
differ greatly from each other in their prevalent outlooks. Just like
industrialized Europe and America, traditional New Guinea has conservative
societies that resist new ways, living side by side with innovative societies
that selectively adopt new ways. The result, with the arrival of Western
technology, is that the more entrepreneurial societies are now exploiting
Western technology to overwhelm their conservative neighbors.
For example, when Europeans first reached the highlands of eastern New
Guinea, in the 1930s, they “discovered” dozens of previously uncontacted
Stone Age tribes, of which the Chimbu tribe proved especially aggressive in
adopting Western technology. When Chimbus saw white settlers planting
coffee, they began growing coffee themselves as a cash crop. In 1964 I met a
50-year-old Chimbu man, unable to read, wearing a traditional grass skirt, and
born into a society still using stone tools, who had become rich by growing
coffee, used his profits to buy a sawmill for $100,000 cash, and bought a fleet
of trucks to transport his coffee and timber to market. In contrast, a
neighboring highland people with whom I worked for eight years, the Daribi,
are especially conservative and uninterested in new technology. When the
first helicopter landed in the Daribi area, they briefly looked at it and just
went back to what they had been doing; the Chimbus would have been
bargaining to charter it. As a result, Chimbus are now moving into the Daribi
area, taking it over for plantations, and reducing the Daribi to working for
them.
On every other continent as well, certain native societies have proved
very receptive, adopted foreign ways and technology selectively, and
integrated them successfully into their own society. In Nigeria the Ibo people
became the local entrepreneurial equivalent of New Guinea’s Chimbus. Today
the most numerous Native American tribe in the United States is the Navajo,
who on European arrival were just one of several hundred tribes. But the
Navajo proved especially resilient and able to deal selectively with
innovation. They incorporated Western dyes into their weaving, became
silversmiths and ranchers, and now drive trucks while continuing to live in
traditional dwellings.
Among the supposedly conservative Aboriginal Australians as well, there
are receptive societies along with conservative ones. At the one extreme, the
Tasmanians continued to use stone tools superseded tens of thousands of
years earlier in Europe and replaced in most of mainland Australia too. At the
opposite extreme, some aboriginal fishing groups of southeastern Australia
devised elaborate technologies for managing fish populations, including the
construction of canals, weirs, and standing traps.
Thus, the development and reception of inventions vary enormously from
society to society on the same continent. They also vary over time within the
same society. Nowadays, Islamic societies in the Middle East are relatively
conservative and not at the forefront of technology. But medieval Islam in the
same region was technologically advanced and open to innovation. It
achieved far higher literacy rates than contemporary Europe; it assimilated the
legacy of classical Greek civilization to such a degree that many classical
Greek books are now known to us only through Arabic copies; it invented or
elaborated windmills, tidal mills, trigonometry, and lateen sails; it made major
advances in metallurgy, mechanical and chemical engineering, and irrigation
methods; and it adopted paper and gunpowder from China and transmitted
them to Europe. In the Middle Ages the flow of technology was
overwhelmingly from Islam to Europe, rather than from Europe to Islam as it
is today. Only after around A.D. 1500 did the net direction of flow begin to
reverse.
Innovation in China too fluctuated markedly with time. Until around A.D.
1450, China was technologically much more innovative and advanced than
Europe, even more so than medieval Islam. The long list of Chinese
inventions includes canal lock gates, cast iron, deep drilling, efficient animal
harnesses, gunpowder, kites, magnetic compasses, movable type, paper,
porcelain, printing (except for the Phaistos disk), sternpost rudders, and
wheelbarrows. China then ceased to be innovative for reasons about which we
shall speculate in the Epilogue. Conversely, we think of western Europe and
its derived North American societies as leading the modern world in
technological innovation, but technology was less advanced in western
Europe than in any other “civilized” area of the Old World until the late
Middle Ages.
Thus, it is untrue that there are continents whose societies have tended to
be innovative and continents whose societies have tended to be conservative.
On any continent, at any given time, there are innovative societies and also
conservative ones. In addition, receptivity to innovation fluctuates in time
within the same region.
On reflection, these conclusions are precisely what one would expect if a
society’s innovativeness is determined by many independent factors. Without
a detailed knowledge of all of those factors, innovativeness becomes
unpredictable. Hence social scientists continue to debate the specific reasons
why receptivity changed in Islam, China, and Europe, and why the Chimbus,
Ibos, and Navajo were more receptive to new technology than were their
neighbors. To the student of broad historical patterns, though, it makes no
difference what the specific reasons were in each of those cases. The myriad
factors affecting innovativeness make the historian’s task paradoxically easier,
by converting societal variation in innovativeness into essentially a random
variable. That means that, over a large enough area (such as a whole
continent) at any particular time, some proportion of societies is likely to be
innovative.
WHERE DO INNOVATIONS actually come from? For all societies except the
few past ones that were completely isolated, much or most new technology is
not invented locally but is instead borrowed from other societies. The relative
importance of local invention and of borrowing depends mainly on two
factors: the ease of invention of the particular technology, and the proximity
of the particular society to other societies.
Some inventions arose straightforwardly from a handling of natural raw
materials. Such inventions developed on many independent occasions in
world history, at different places and times. One example, which we have
already considered at length, is plant domestication, with at least nine
independent origins. Another is pottery, which may have arisen from
observations of the behavior of clay, a very widespread natural material, when
dried or heated. Pottery appeared in Japan around 14,000 years ago, in the
Fertile Crescent and China by around 10,000 years ago, and in Amazonia,
Africa’s Sahel zone, the U.S. Southeast, and Mexico thereafter.
An example of a much more difficult invention is writing, which does not
suggest itself by observation of any natural material. As we saw in Chapter
12, it had only a few independent origins, and the alphabet arose apparently
only once in world history. Other difficult inventions include the water wheel,
rotary quern, tooth gearing, magnetic compass, windmill, and camera
obscura, all of which were invented only once or twice in the Old World and
never in the New World.
Such complex inventions were usually acquired by borrowing, because
they spread more rapidly than they could be independently invented locally. A
clear example is the wheel, which is first attested around 3400 B.C. near the
Black Sea, and then turns up within the next few centuries over much of
Europe and Asia. All those early Old World wheels are of a peculiar design: a
solid wooden circle constructed of three planks fastened together, rather than
a rim with spokes. In contrast, the sole wheels of Native American societies
(depicted on Mexican ceramic vessels) consisted of a single piece, suggesting
a second independent invention of the wheel—as one would expect from
other evidence for the isolation of New World from Old World civilizations.
No one thinks that that same peculiar Old World wheel design appeared
repeatedly by chance at many separate sites of the Old World within a few
centuries of each other, after 7 million years of wheelless human history.
Instead, the utility of the wheel surely caused it to diffuse rapidly east and
west over the Old World from its sole site of invention. Other examples of
complex technologies that diffused east and west in the ancient Old World,
from a single West Asian source, include door locks, pulleys, rotary querns,
windmills—and the alphabet. A New World example of technological
diffusion is metallurgy, which spread from the Andes via Panama to
Mesoamerica.
When a widely useful invention does crop up in one society, it then tends
to spread in either of two ways. One way is that other societies see or learn of
the invention, are receptive to it, and adopt it. The second is that societies
lacking the invention find themselves at a disadvantage vis-à-vis the inventing
society, and they become overwhelmed and replaced if the disadvantage is
sufficiently great. A simple example is the spread of muskets among New
Zealand’s Maori tribes. One tribe, the Ngapuhi, adopted muskets from
European traders around 1818. Over the course of the next 15 years, New
Zealand was convulsed by the so-called Musket Wars, as musketless tribes
either acquired muskets or were subjugated by tribes already armed with
them. The outcome was that musket technology had spread throughout the
whole of New Zealand by 1833: all surviving Maori tribes now had muskets.
When societies do adopt a new technology from the society that invented
it, the diffusion may occur in many different contexts. They include peaceful
trade (as in the spread of transistors from the United States to Japan in 1954),
espionage (the smuggling of silkworms from Southeast Asia to the Mideast in
A.D. 552), emigration (the spread of French glass and clothing manufacturing
techniques over Europe by the 200,000 Huguenots expelled from France in
1685), and war. A crucial case of the last was the transfer of Chinese
papermaking techniques to Islam, made possible when an Arab army defeated
a Chinese army at the battle of Talas River in Central Asia in A.D. 751, found
some papermakers among the prisoners of war, and brought them to
Samarkand to set up paper manufacture.
In Chapter 12 we saw that cultural diffusion can involve either detailed
“blueprints” or just vague ideas stimulating a reinvention of details. While
Chapter 12 illustrated those alternatives for the spread of writing, they also
apply to the diffusion of technology. The preceding paragraph gave examples
of blueprint copying, whereas the transfer of Chinese porcelain technology to
Europe provides an instance of long-drawn-out idea diffusion. Porcelain, a
fine-grained translucent pottery, was invented in China around the 7th century
A.D. When it began to reach Europe by the Silk Road in the 14th century (with
no information about how it was manufactured), it was much admired, and
many unsuccessful attempts were made to imitate it. Not until 1707 did the
German alchemist Johann Böttger, after lengthy experiments with processes
and with mixing various minerals and clays together, hit upon the solution
and established the now famous Meissen porcelain works. More or less
independent later experiments in France and England led to Sèvres,
Wedgwood, and Spode porcelains. Thus, European potters had to reinvent
Chinese manufacturing methods for themselves, but they were stimulated to
do so by having models of the desired product before them.
DEPENDING ON THEIR geographic location, societies differ in how readily
they can receive technology by diffusion from other societies. The most
isolated people on Earth in recent history were the Aboriginal Tasmanians,
living without oceangoing watercraft on an island 100 miles from Australia,
itself the most isolated continent. The Tasmanians had no contact with other
societies for 10,000 years and acquired no new technology other than what
they invented themselves. Australians and New Guineans, separated from the
Asian mainland by the Indonesian island chain, received only a trickle of
inventions from Asia. The societies most accessible to receiving inventions by
diffusion were those embedded in the major continents. In these societies
technology developed most rapidly, because they accumulated not only their
own inventions but also those of other societies. For example, medieval
Islam, centrally located in Eurasia, acquired inventions from India and China
and inherited ancient Greek learning.
The importance of diffusion, and of geographic location in making it
possible, is strikingly illustrated by some otherwise incomprehensible cases of
societies that abandoned powerful technologies. We tend to assume that
useful technologies, once acquired, inevitably persist until superseded by
better ones. In reality, technologies must be not only acquired but also
maintained, and that too depends on many unpredictable factors. Any society
goes through social movements or fads, in which economically useless things
become valued or useful things devalued temporarily. Nowadays, when
almost all societies on Earth are connected to each other, we cannot imagine a
fad’s going so far that an important technology would actually be discarded. A
society that temporarily turned against a powerful technology would continue
to see it being used by neighboring societies and would have the opportunity
to reacquire it by diffusion (or would be conquered by neighbors if it failed to
do so). But such fads can persist in isolated societies.
A famous example involves Japan’s abandonment of guns. Firearms
reached Japan in A.D. 1543, when two Portuguese adventurers armed with
harquebuses (primitive guns) arrived on a Chinese cargo ship. The Japanese
were so impressed by the new weapon that they commenced indigenous gun
production, greatly improved gun technology, and by A.D. 1600 owned more
and better guns than any other country in the world.
But there were also factors working against the acceptance of firearms in
Japan. The country had a numerous warrior class, the samurai, for whom
swords rated as class symbols and works of art (and as means for subjugating
the lower classes). Japanese warfare had previously involved single combats
between samurai swordsmen, who stood in the open, made ritual speeches,
and then took pride in fighting gracefully. Such behavior became lethal in the
presence of peasant soldiers ungracefully blasting away with guns. In
addition, guns were a foreign invention and grew to be despised, as did other
things foreign in Japan after 1600. The samurai-controlled government began
by restricting gun production to a few cities, then introduced a requirement of
a government license for producing a gun, then issued licenses only for guns
produced for the government, and finally reduced government orders for
guns, until Japan was almost without functional guns again.
Contemporary European rulers also included some who despised guns
and tried to restrict their availability. But such measures never got far in
Europe, where any country that temporarily swore off firearms would be
promptly overrun by gun-toting neighboring countries. Only because Japan
was a populous, isolated island could it get away with its rejection of the
powerful new military technology. Its safety in isolation came to an end in
1853, when the visit of Commodore Perry’s U.S. fleet bristling with cannons
convinced Japan of its need to resume gun manufacture.
That rejection and China’s abandonment of oceangoing ships (as well as
of mechanical clocks and water-driven spinning machines) are well-known
historical instances of technological reversals in isolated or semi-isolated
societies. Other such reversals occurred in prehistoric times. The extreme case
is that of Aboriginal Tasmanians, who abandoned even bone tools and fishing
to become the society with the simplest technology in the modern world
(Chapter 15). Aboriginal Australians may have adopted and then abandoned
bows and arrows. Torres Islanders abandoned canoes, while Gaua Islanders
abandoned and then readopted them. Pottery was abandoned throughout
Polynesia. Most Polynesians and many Melanesians abandoned the use of
bows and arrows in war. Polar Eskimos lost the bow and arrow and the kayak,
while Dorset Eskimos lost the bow and arrow, bow drill, and dogs.
These examples, at first so bizarre to us, illustrate well the roles of
geography and of diffusion in the history of technology. Without diffusion,
fewer technologies are acquired, and more existing technologies are lost.
BECAUSE TECHNOLOGY BEGETS more technology, the importance of an
invention’s diffusion potentially exceeds the importance of the original
invention. Technology’s history exemplifies what is termed an autocatalytic
process: that is, one that speeds up at a rate that increases with time, because
the process catalyzes itself. The explosion of technology since the Industrial
Revolution impresses us today, but the medieval explosion was equally
impressive compared with that of the Bronze Age, which in turn dwarfed that
of the Upper Paleolithic.
One reason why technology tends to catalyze itself is that advances
depend upon previous mastery of simpler problems. For example, Stone Age
farmers did not proceed directly to extracting and working iron, which
requires high-temperature furnaces. Instead, iron ore metallurgy grew out of
thousands of years of human experience with natural outcrops of pure metals
soft enough to be hammered into shape without heat (copper and gold). It also
grew out of thousands of years of development of simple furnaces to make
pottery, and then to extract copper ores and work copper alloys (bronzes) that
do not require as high temperatures as does iron. In both the Fertile Crescent
and China, iron objects became common only after about 2,000 years of
experience of bronze metallurgy. New World societies had just begun making
bronze artifacts and had not yet started making iron ones at the time when the
arrival of Europeans truncated the New World’s independent trajectory.
The other main reason for autocatalysis is that new technologies and
materials make it possible to generate still other new technologies by
recombination. For instance, why did printing spread explosively in medieval
Europe after Gutenberg printed his Bible in A.D. 1455, but not after that
unknown printer printed the Phaistos disk in 1700 B.C.? The explanation is
partly that medieval European printers were able to combine six technological
advances, most of which were unavailable to the maker of the Phaistos disk.
Of those advances—in paper, movable type, metallurgy, presses, inks, and
scripts—paper and the idea of movable type reached Europe from China.
Gutenberg’s development of typecasting from metal dies, to overcome the
potentially fatal problem of nonuniform type size, depended on many
metallurgical developments: steel for letter punches, brass or bronze alloys
(later replaced by steel) for dies, lead for molds, and a tin-zinc-lead alloy for
type. Gutenberg’s press was derived from screw presses in use for making
wine and olive oil, while his ink was an oil-based improvement on existing
inks. The alphabetic scripts that medieval Europe inherited from three
millennia of alphabet development lent themselves to printing with movable
type, because only a few dozen letter forms had to be cast, as opposed to the
thousands of signs required for Chinese writing.
In all six respects, the maker of the Phaistos disk had access to much less
powerful technologies to combine into a printing system than did Gutenberg.
The disk’s writing medium was clay, which is much bulkier and heavier than
paper. The metallurgical skills, inks, and presses of 1700 B.C. Crete were more
primitive than those of A.D. 1455 Germany, so the disk had to be punched by
hand rather than by cast movable type locked into a metal frame, inked, and
pressed. The disk’s script was a syllabary with more signs, of more complex
form, than the Roman alphabet used by Gutenberg. As a result, the Phaistos
disk’s printing technology was much clumsier, and offered fewer advantages
over writing by hand, than Gutenberg’s printing press. In addition to all those
technological drawbacks, the Phaistos disk was printed at a time when
knowledge of writing was confined to a few palace or temple scribes. Hence
there was little demand for the disk maker’s beautiful product, and little
incentive to invest in making the dozens of hand punches required. In
contrast, the potential mass market for printing in medieval Europe induced
numerous investors to lend money to Gutenberg.
HUMAN TECHNOLOGY DEVELOPED from the first stone tools, in use by two
and a half million years ago, to the 1996 laser printer that replaced my already
outdated 1992 laser printer and that was used to print this book’s manuscript.
The rate of development was undetectably slow at the beginning, when
hundreds of thousands of years passed with no discernible change in our stone
tools and with no surviving evidence for artifacts made of other materials.
Today, technology advances so rapidly that it is reported in the daily
newspaper.
In this long history of accelerating development, one can single out two
especially significant jumps. The first, occurring between 100,000 and 50,000
years ago, probably was made possible by genetic changes in our bodies:
namely, by evolution of the modern anatomy permitting modern speech or
modern brain function, or both. That jump led to bone tools, single-purpose
stone tools, and compound tools. The second jump resulted from our adoption
of a sedentary lifestyle, which happened at different times in different parts of
the world, as early as 13,000 years ago in some areas and not even today in
others. For the most part, that adoption was linked to our adoption of food
production, which required us to remain close to our crops, orchards, and
stored food surpluses.
Sedentary living was decisive for the history of technology, because it
enabled people to accumulate nonportable possessions. Nomadic hunter-
gatherers are limited to technology that can be carried. If you move often and
lack vehicles or draft animals, you confine your possessions to babies,
weapons, and a bare minimum of other absolute necessities small enough to
carry. You can’t be burdened with pottery and printing presses as you shift
camp. That practical difficulty probably explains the tantalizingly early
appearance of some technologies, followed by a long delay in their further
development. For example, the earliest attested precursors of ceramics are
fired clay figurines made in the area of modern Czechoslovakia 27,000 years
ago, long before the oldest known fired clay vessels (from Japan 14,000 years
ago). The same area of Czechoslovakia at the same time has yielded the
earliest evidence for weaving, otherwise not attested until the oldest known
basket appears around 13,000 years ago and the oldest known woven cloth
around 9,000 years ago. Despite these very early first steps, neither pottery
nor weaving took off until people became sedentary and thereby escaped the
problem of transporting pots and looms.
Besides permitting sedentary living and hence the accumulation of
possessions, food production was decisive in the history of technology for
another reason. It became possible, for the first time in human evolution, to
develop economically specialized societies consisting of non-food-producing
specialists fed by food-producing peasants. But we already saw, in Part 2 of
this book, that food production arose at different times in different continents.
In addition, as we’ve seen in this chapter, local technology depends, for both
its origin and its maintenance, not only on local invention but also on the
diffusion of technology from elsewhere. That consideration tended to cause
technology to develop most rapidly on continents with few geographic and
ecological barriers to diffusion, either within that continent or on other
continents. Finally, each society on a continent represents one more
opportunity to invent and adopt a technology, because societies vary greatly
in their innovativeness for many separate reasons. Hence, all other things
being equal, technology develops fastest in large productive regions with
large human populations, many potential inventors, and many competing
societies.
Let us now summarize how variations in these three factors—time of
onset of food production, barriers to diffusion, and human population size—
led straightforwardly to the observed intercontinental differences in the
development of technology. Eurasia (effectively including North Africa) is the
world’s largest landmass, encompassing the largest number of competing
societies. It was also the landmass with the two centers where food production
began the earliest: the Fertile Crescent and China. Its east–west major axis
permitted many inventions adopted in one part of Eurasia to spread relatively
rapidly to societies at similar latitudes and climates elsewhere in Eurasia. Its
breadth along its minor axis (north–south) contrasts with the Americas’
narrowness at the Isthmus of Panama. It lacks the severe ecological barriers
transecting the major axes of the Americas and Africa. Thus, geographic and
ecological barriers to diffusion of technology were less severe in Eurasia than
in other continents. Thanks to all these factors, Eurasia was the continent on
which technology started its post-Pleistocene acceleration earliest and
resulted in the greatest local accumulation of technologies.
North and South America are conventionally regarded as separate
continents, but they have been connected for several million years, pose
similar historical problems, and may be considered together for comparison
with Eurasia. The Americas form the world’s second-largest landmass,
significantly smaller than Eurasia. However, they are fragmented by
geography and by ecology: the Isthmus of Panama, only 40 miles wide,
virtually transects the Americas geographically, as do the isthmus’s Darien
rain forests and the northern Mexican desert ecologically. The latter desert
separated advanced human societies of Mesoamerica from those of North
America, while the isthmus separated advanced societies of Mesoamerica
from those of the Andes and Amazonia. In addition, the main axis of the
Americas is north-south, forcing most diffusion to go against a gradient of
latitude (and climate) rather than to operate within the same latitude. For
example, wheels were invented in Mesoamerica, and llamas were
domesticated in the central Andes by 3000 B.C., but 5,000 years later the
Americas’ sole beast of burden and sole wheels had still not encountered each
other, even though the distance separating Mesoamerica’s Maya societies
from the northern border of the Inca Empire (1,200 miles) was far less than
the 6,000 miles separating wheel- and horse-sharing France and China. Those
factors seem to me to account for the Americas’ technological lag behind
Eurasia.
Sub-Saharan Africa is the world’s third largest landmass, considerably
smaller than the Americas. Throughout most of human history it was far more
accessible to Eurasia than were the Americas, but the Saharan desert is still a
major ecological barrier separating sub-Saharan Africa from Eurasia plus
North Africa. Africa’s north-south axis posed a further obstacle to the
diffusion of technology, both between Eurasia and sub-Saharan Africa and
within the sub-Saharan region itself. As an illustration of the latter obstacle,
pottery and iron metallurgy arose in or reached sub-Saharan Africa’s Sahel
zone (north of the equator) at least as early as they reached western Europe.
However, pottery did not reach the southern tip of Africa until around A.D. 1,
and metallurgy had not yet diffused overland to the southern tip by the time
that it arrived there from Europe on ships.
Finally, Australia is the smallest continent. The very low rainfall and
productivity of most of Australia makes it effectively even smaller as regards
its capacity to support human populations. It is also the most isolated
continent. In addition, food production never arose indigenously in Australia.
Those factors combined to leave Australia the sole continent still without
metal artifacts in modern times.
Table 13.1 translates these factors into numbers, by comparing the
continents with respect to their areas and their modern human populations.
The continents’ populations 10,000 years ago, just before the rise of food
production, are not known but surely stood in the same sequence, since many
of the areas producing the most food today would also have been productive
areas for hunter-gatherers 10,000 years ago. The differences in population are
glaring: Eurasia’s (including North Africa’s) is nearly 6 times that of the
Americas, nearly 8 times that of Africa’s, and 230 times that of Australia’s.
Larger populations mean more inventors and more competing societies. Table
13.1 by itself goes a long way toward explaining the origins of guns and steel
in Eurasia.
TABLE 13.1 Human Populations of the Continents
Continent
1990 Population Area (square miles)
Eurasia and North Africa
4,120,000,000
24,200,000
(Eurasia)
(4,000,000,000) (21,500,000)
(North Africa)
(120,000,000)
(2,700,000)
North America and South America 736,000,000
16,400,000
Sub-Saharan Africa
535,000,000
9,100,000
Australia
18,000,000
3,000,000
All these effects that continental differences in area, population, ease of
diffusion, and onset of food production exerted on the rise of technology
became exaggerated, because technology catalyzes itself. Eurasia’s
considerable initial advantage thereby was translated into a huge lead as of
A.D. 1492—for reasons of Eurasia’s distinctive geography rather than of
distinctive human intellect. The New Guineans whom I know include
potential Edisons. But they directed their ingenuity toward technological
problems appropriate to their situations: the problems of surviving without
any imported items in the New Guinea jungle, rather than the problem of
inventing phonographs.
CHAPTER 14
FROM EGALITARIANISM TO KLEPTOCRACY
IN 1979, WHILE I WAS FLYING WITH MISSIONARY FRIENDS over a remote swamp-
filled basin of New Guinea, I noticed a few huts many miles apart. The pilot
explained to me that, somewhere in that muddy expanse below us, a group of
Indonesian crocodile hunters had recently come across a group of New
Guinea nomads. Both groups had panicked, and the encounter had ended with
the Indonesians shooting several of the nomads.
My missionary friends guessed that the nomads belonged to an
uncontacted group called the Fayu, known to the outside world only through
accounts by their terrified neighbors, a missionized group of erstwhile
nomads called the Kirikiri. First contacts between outsiders and New Guinea
groups are always potentially dangerous, but this beginning was especially
inauspicious. Nevertheless, my friend Doug flew in by helicopter to try to
establish friendly relations with the Fayu. He returned, alive but shaken, to
tell a remarkable story.
It turned out that the Fayu normally lived as single families, scattered
through the swamp and coming together once or twice each year to negotiate
exchanges of brides. Doug’s visit coincided with such a gathering, of a few
dozen Fayu. To us, a few dozen people constitute a small, ordinary gathering,
but to the Fayu it was a rare, frightening event. Murderers suddenly found
themselves face-to-face with their victim’s relatives. For example, one Fayu
man spotted the man who had killed his father. The son raised his ax and
rushed at the murderer but was wrestled to the ground by friends; then the
murderer came at the prostrate son with an ax and was also wrestled down.
Both men were held, screaming in rage, until they seemed sufficiently
exhausted to be released. Other men periodically shouted insults at each other,
shook with anger and frustration, and pounded the ground with their axes.
That tension continued for the several days of the gathering, while Doug
prayed that the visit would not end in violence.
The Fayu consist of about 400 hunter-gatherers, divided into four clans
and wandering over a few hundred square miles. According to their own
account, they had formerly numbered about 2,000, but their population had
been greatly reduced as a result of Fayu killing Fayu. They lacked political
and social mechanisms, which we take for granted, to achieve peaceful
resolution of serious disputes. Eventually, as a result of Doug’s visit, one
group of Fayu invited a courageous husband-and-wife missionary couple to
live with them. The couple has now resided there for a dozen years and
gradually persuaded the Fayu to renounce violence. The Fayu are thereby
being brought into the modern world, where they face an uncertain future.
Many other previously uncontacted groups of New Guineans and
Amazonian Indians have similarly owed to missionaries their incorporation
into modern society. After the missionaries come teachers and doctors,
bureaucrats and soldiers. The spreads of government and of religion have thus
been linked to each other throughout recorded history, whether the spread has
been peaceful (as eventually with the Fayu) or by force. In the latter case it is
often government that organizes the conquest, and religion that justifies it.
While nomads and tribespeople occasionally defeat organized governments
and religions, the trend over the past 13,000 years has been for the nomads
and tribespeople to lose.
At the end of the last Ice Age, much of the world’s population lived in
societies similar to that of the Fayu today, and no people then lived in a much
more complex society. As recently as A.D. 1500, less than 20 percent of the
world’s land area was marked off by boundaries into states run by bureaucrats
and governed by laws. Today, all land except Antarctica’s is so divided.
Descendants of those societies that achieved centralized government and
organized religion earliest ended up dominating the modern world. The
combination of government and religion has thus functioned, together with
germs, writing, and technology, as one of the four main sets of proximate
agents leading to history’s broadest pattern. How did government and religion
arise?
FAYU BANDS AND modern states represent opposite extremes along the
spectrum of human societies. Modern American society and the Fayu differ in
the presence or absence of a professional police force, cities, money,
distinctions between rich and poor, and many other political, economic, and
social institutions. Did all of those institutions arise together, or did some
arise before others? We can infer the answer to this question by comparing
modern societies at different levels of organization, by examining written
accounts or archaeological evidence about past societies, and by observing
how a society’s institutions change over time.
Cultural anthropologists attempting to describe the diversity of human
societies often divide them into as many as half a dozen categories. Any such
attempt to define stages of any evolutionary or developmental continuum—
whether of musical styles, human life stages, or human societies—is doubly
doomed to imperfection. First, because each stage grows out of some previous
stage, the lines of demarcation are inevitably arbitrary. (For example, is a 19-
year-old person an adolescent or a young adult?) Second, developmental
sequences are not invariant, so examples pigeonholed under the same stage
are inevitably heterogeneous. (Brahms and Liszt would turn in their graves to
know that they are now grouped together as composers of the romantic
period.) Nevertheless, arbitrarily delineated stages provide a useful shorthand
for discussing the diversity of music and of human societies, provided one
bears in mind the above caveats. In that spirit, we shall use a simple
classification based on just four categories—band, tribe, chiefdom, and state
(see Table 14.1)—to understand societies.
Bands are the tiniest societies, consisting typically of 5 to 80 people, most
or all of them close relatives by birth or by marriage. In effect, a band is an
extended family or several related extended families. Today, bands still living
autonomously are almost confined to the most remote parts of New Guinea
and Amazonia, but within modern times there were many others that have
only recently fallen under state control or been assimilated or exterminated.
They include many or most African Pygmies, southern African San hunter-
gatherers (so-called Bushmen), Aboriginal Australians, Eskimos (Inuit), and
Indians of some resource-poor areas of the Americas such as Tierra del Fuego
and the northern boreal forests. All those modern bands are or were nomadic
hunter-gatherers rather than settled food producers. Probably all humans lived
in bands until at least 40,000 years ago, and most still did as recently as
11,000 years ago.
TABLE 14.1 Types of Societies
Band
Tribe
Chiefdom
State
Membership
Number of people
dozens
hundreds
thousands
over 50,000
fixed: 1 or more
fixed: many villages
Settlement pattern
nomadic
fixed: 1 village
villages
and cities
class and
Basis of relationships
kin
kin-based clans
class and residence
residence
Ethnicities and languages
1
1
1
1 or more
Government
Decision making,
“egalitarian” or
centralized,
“egalitarian”
centralized
leadership
big-man
hereditary
none, or 1 or 2
Bureaucracy
none
none
many levels
levels
Monopoly of force and
no
no
yes
yes
information
Conflict resolution
informal
informal
centralized
laws, judges
no
paramount
Hierarchy of settlement
no
no
capital
village
Religion
Justifies kleptocracy?
no
no
yes
yes
no
Economy
Food production
no
no
yes
yes
intensive
intensive
Division of labor
no
no
no
yes
yes
redistributive
Exchanges
reciprocal
reciprocal
redistributive (“taxes”)
(“tribute”)
Control of land
band
clan
chief
various
Society
Stratified
no
no
yes, by kin
yes, not by kin
Slavery
no
no
small-scale
large-scale
Luxury goods for elite
no
no
yes
yes
Public architecture
no
no
no
yes
yes
Indigenous literacy
no
no
no
often
A horizontal arrow indicates that the attribute varies between less and more complex societies of
that type.
Bands lack many institutions that we take for granted in our own society.
They have no permanent single base of residence. The band’s land is used
jointly by the whole group, instead of being partitioned among subgroups or
individuals. There is no regular economic specialization, except by age and
sex: all able-bodied individuals forage for food. There are no formal
institutions, such as laws, police, and treaties, to resolve conflicts within and
between bands. Band organization is often described as “egalitarian”: there is
no formalized social stratification into upper and lower classes, no formalized
or hereditary leadership, and no formalized monopolies of information and
decision making. However, the term “egalitarian” should not be taken to
mean that all band members are equal in prestige and contribute equally to
decisions. Rather, the term merely means that any band “leadership” is
informal and acquired through qualities such as personality, strength,
intelligence, and fighting skills.
My own experience with bands comes from the swampy lowland area of
New Guinea where the Fayu live, a region known as the Lakes Plains. There,
I still encounter extended families of a few adults with their dependent
children and elderly, living in crude temporary shelters along streams and
traveling by canoe and on foot. Why do peoples of the Lakes Plains continue
to live as nomadic bands, when most other New Guinea peoples, and almost
all other peoples elsewhere in the world, now live in settled larger groups?
The explanation is that the region lacks dense local concentrations of
resources that would permit many people to live together, and that (until the
arrival of missionaries bringing crop plants) it also lacked native plants that
could have permitted productive farming. The bands’ food staple is the sago
palm tree, whose core yields a starchy pith when the palm reaches maturity.
The bands are nomadic, because they must move when they have cut the
mature sago trees in an area. Band numbers are kept low by diseases
(especially malaria), by the lack of raw materials in the swamp (even stone
for tools must be obtained by trade), and by the limited amount of food that
the swamp yields for humans. Similar limitations on the resources accessible
to existing human technology prevail in the regions of the world recently
occupied by other bands.
Our closest animal relatives, the gorillas and chimpanzees and bonobos of
Africa, also live in bands. All humans presumably did so too, until improved
technology for extracting food allowed some hunter-gatherers to settle in
permanent dwellings in some resource-rich areas. The band is the political,
economic, and social organization that we inherited from our millions of years
of evolutionary history. Our developments beyond it all took place within the
last few tens of thousands of years.
THE FIRST OF those stages beyond the band is termed the tribe, which differs
in being larger (typically comprising hundreds rather than dozens of people)
and usually having fixed settlements. However, some tribes and even
chiefdoms consist of herders who move seasonally.
Tribal organization is exemplified by New Guinea highlanders, whose
political unit before the arrival of colonial government was a village or else a
close-knit cluster of villages. This political definition of “tribe” is thus often
much smaller than what linguists and cultural anthropologists would define as
a tribe—namely, a group that shares language and culture. For example, in
1964 I began to work among a group of highlanders known as the Foré. By
linguistic and cultural standards, there were then 12,000 Foré, speaking two
mutually intelligible dialects and living in 65 villages of several hundred
people each. But there was no political unity whatsoever among villages of
the Foré language group. Each hamlet was involved in a kaleidoscopically
changing pattern of war and shifting alliances with all neighboring hamlets,
regardless of whether the neighbors were Foré or speakers of a different
language.
Tribes, recently independent and now variously subordinated to national
states, still occupy much of New Guinea, Melanesia, and Amazonia. Similar
tribal organization in the past is inferred from archaeological evidence of
settlements that were substantial but lacked the archaeological hallmarks of
chiefdoms that I shall explain below. That evidence suggests that tribal
organization began to emerge around 13,000 years ago in the Fertile Crescent
and later in some other areas. A prerequisite for living in settlements is either
food production or else a productive environment with especially
concentrated resources that can be hunted and gathered within a small area.
That’s why settlements, and by inference tribes, began to proliferate in the
Fertile Crescent at that time, when climate changes and improved technology
combined to permit abundant harvests of wild cereals.
Besides differing from a band by virtue of its settled residence and its
larger numbers, a tribe also differs in that it consists of more than one
formally recognized kinship group, termed clans, which exchange marriage
partners. Land belongs to a particular clan, not to the whole tribe. However,
the number of people in a tribe is still low enough that everyone knows
everyone else by name and relationships.
For other types of human groups as well, “a few hundred” seems to be an
upper limit for group size compatible with everyone’s knowing everybody. In
our state society, for instance, school principals are likely to know all their
students by name if the school contains a few hundred children, but not if it
contains a few thousand children. One reason why the organization of human
government tends to change from that of a tribe to that of a chiefdom in
societies with more than a few hundred members is that the difficult issue of
conflict resolution between strangers becomes increasingly acute in larger
groups. A fact further diffusing potential problems of conflict resolution in
tribes is that almost everyone is related to everyone else, by blood or marriage
or both. Those ties of relationships binding all tribal members make police,
laws, and other conflict-resolving institutions of larger societies unnecessary,
since any two villagers getting into an argument will share many kin, who
apply pressure on them to keep it from becoming violent. In traditional New
Guinea society, if a New Guinean happened to encounter an unfamiliar New
Guinean while both were away from their respective villages, the two
engaged in a long discussion of their relatives, in an attempt to establish some
relationship and hence some reason why the two should not attempt to kill
each other.
Despite all of these differences between bands and tribes, many
similarities remain. Tribes still have an informal, “egalitarian” system of
government. Information and decision making are both communal. In the
New Guinea highlands, I have watched village meetings where all adults in
the village were present, sitting on the ground, and individuals made
speeches, without any appearance of one person’s “chairing” the discussion.
Many highland villages do have someone known as the “big-man,” the most
influential man of the village. But that position is not a formal office to be
filled and carries only limited power. The big-man has no independent
decision-making authority, knows no diplomatic secrets, and can do no more
than attempt to sway communal decisions. Big-men achieve that status by
their own attributes; the position is not inherited.
Tribes also share with bands an “egalitarian” social system, without
ranked lineages or classes. Not only is status not inherited; no member of a
traditional tribe or band can become disproportionately wealthy by his or her
own efforts, because each individual has debts and obligations to many
others. It is therefore impossible for an outsider to guess, from appearances,
which of all the adult men in a village is the big-man: he lives in the same
type of hut, wears the same clothes or ornaments, or is as naked, as everyone
else.
Like bands, tribes lack a bureaucracy, police force, and taxes. Their
economy is based on reciprocal exchanges between individuals or families,
rather than on a redistribution of tribute paid to some central authority.
Economic specialization is slight: full-time crafts specialists are lacking, and
every able-bodied adult (including the big-man) participates in growing,
gathering, or hunting food. I recall one occasion when I was walking past a
garden in the Solomon Islands, saw a man digging and waving at me in the
distance, and realized to my astonishment that it was a friend of mine named
Faletau. He was the most famous wood carver of the Solomons, an artist of
great originality—but that did not free him of the necessity to grow his own
sweet potatoes. Since tribes thus lack economic specialists, they also lack
slaves, because there are no specialized menial jobs for a slave to perform.
Just as musical composers of the classical period range from C. P. E. Bach
to Schubert and thereby cover the whole spectrum from baroque composers to
romantic composers, tribes also shade into bands at one extreme and into
chiefdoms at the opposite extreme. In particular, a tribal big-man’s role in
dividing the meat of pigs slaughtered for feasts points to the role of chiefs in
collecting and redistributing food and goods—now reconstrued as tribute—in
chiefdoms. Similarly, presence or absence of public architecture is supposedly
one of the distinctions between tribes and chiefdoms, but large New Guinea
villages often have cult houses (known as haus tamburan, on the Sepik River)
that presage the temples of chiefdoms.
ALTHOUGH A FEW bands and tribes survive today on remote and ecologically
marginal lands outside state control, fully independent chiefdoms had
disappeared by the early twentieth century, because they tended to occupy
prime land coveted by states. However, as of A.D. 1492, chiefdoms were still
widespread over much of the eastern United States, in productive areas of
South and Central America and sub-Saharan Africa that had not yet been
subsumed under native states, and in all of Polynesia. The archaeological
evidence discussed below suggests that chiefdoms arose by around 5500 B.C.
in the Fertile Crescent and by around 1000 B.C. in Mesoamerica and the
Andes. Let us consider the distinctive features of chiefdoms, very different
from modern European and American states and, at the same time, from
bands and simple tribal societies.
As regards population size, chiefdoms were considerably larger than
tribes, ranging from several thousand to several tens of thousands of people.
That size created serious potential for internal conflict because, for any person
living in a chiefdom, the vast majority of other people in the chiefdom were
neither closely related by blood or marriage nor known by name. With the rise
of chiefdoms around 7,500 years ago, people had to learn, for the first time in
history, how to encounter strangers regularly without attempting to kill them.
Part of the solution to that problem was for one person, the chief, to
exercise a monopoly on the right to use force. In contrast to a tribe’s big-man,
a chief held a recognized office, filled by hereditary right. Instead of the
decentralized anarchy of a village meeting, the chief was a permanent
centralized authority, made all significant decisions, and had a monopoly on
critical information (such as what a neighboring chief was privately
threatening, or what harvest the gods had supposedly promised). Unlike big-
men, chiefs could be recognized from afar by visible distinguishing features,
such as a large fan worn over the back on Rennell Island in the Southwest
Pacific. A commoner encountering a chief was obliged to perform ritual
marks of respect, such as (on Hawaii) prostrating oneself. The chief’s orders
might be transmitted through one or two levels of bureaucrats, many of whom
were themselves low-ranked chiefs. However, in contrast to state bureaucrats,
chiefdom bureaucrats had generalized rather than specialized roles. In
Polynesian Hawaii the same bureaucrats (termed konohiki) extracted tribute
and oversaw irrigation and organized labor corvées for the chief, whereas
state societies have separate tax collectors, water district managers, and draft
boards.
A chiefdom’s large population in a small area required plenty of food,
obtained by food production in most cases, by hunting-gathering in a few
especially rich areas. For example, American Indians of the Pacific Northwest
coast, such as the Kwakiutl, Nootka, and Tlingit Indians, lived under chiefs in
villages without any agriculture or domestic animals, because the rivers and
sea were so rich in salmon and halibut. The food surpluses generated by some
people, relegated to the rank of commoners, went to feed the chiefs, their
families, bureaucrats, and crafts specialists, who variously made canoes,
adzes, or spittoons or worked as bird catchers or tattooers.
Luxury goods, consisting of those specialized crafts products or else rare
objects obtained by long-distance trade, were reserved for chiefs. For
example, Hawaiian chiefs had feather cloaks, some of them consisting of tens
of thousands of feathers and requiring many human generations for their
manufacture (by commoner cloak makers, of course). That concentration of
luxury goods often makes it possible to recognize chiefdoms archaeologically,
by the fact that some graves (those of chiefs) contain much richer goods than
other graves (those of commoners), in contrast to the egalitarian burials of
earlier human history. Some ancient complex chiefdoms can also be
distinguished from tribal villages by the remains of elaborate public
architecture (such as temples) and by a regional hierarchy of settlements, with
one site (the site of the paramount chief) being obviously larger and having
more administrative buildings and artifacts than other sites.
Like tribes, chiefdoms consisted of multiple hereditary lineages living at
one site. However, whereas the lineages of tribal villages are equal-ranked
clans, in a chiefdom all members of the chief’s lineage had hereditary
perquisites. In effect, the society was divided into hereditary chief and
commoner classes, with Hawaiian chiefs themselves subdivided into eight
hierarchically ranked lineages, each concentrating its marriages within its
own lineage. Furthermore, since chiefs required menial servants as well as
specialized craftspeople, chiefdoms differed from tribes in having many jobs
that could be filled by slaves, typically obtained by capture in raids.
The most distinctive economic feature of chiefdoms was their shift from
reliance solely on the reciprocal exchanges characteristic of bands and tribes,
by which A gives B a gift while expecting that B at some unspecified future
time will give a gift of comparable value to A. We modern state dwellers
indulge in such behavior on birthdays and holidays, but most of our flow of
goods is achieved instead by buying and selling for money according to the
law of supply and demand. While continuing reciprocal exchanges and
without marketing or money, chiefdoms developed an additional new system
termed a redistributive economy. A simple example would involve a chief
receiving wheat at harvest time from every farmer in the chiefdom, then
throwing a feast for everybody and serving bread or else storing the wheat
and gradually giving it out again in the months between harvests. When a
large portion of the goods received from commoners was not redistributed to
them but was retained and consumed by the chiefly lineages and craftspeople,
the redistribution became tribute, a precursor of taxes that made its first
appearance in chiefdoms. From the commoners the chiefs claimed not only
goods but also labor for construction of public works, which again might
return to benefit the commoners (for example, irrigation systems to help feed
everybody) or instead benefited mainly the chiefs (for instance, lavish tombs).
We have been talking about chiefdoms generically, as if they were all the
same. In fact, chiefdoms varied considerably. Larger ones tended to have
more powerful chiefs, more ranks of chiefly lineages, greater distinctions
between chiefs and commoners, more retention of tribute by the chiefs, more
layers of bureaucrats, and grander public architecture. For instance, societies
on small Polynesian islands were effectively rather similar to tribal societies
with a big-man, except that the position of chief was hereditary. The chief’s
hut looked like any other hut, there were no bureaucrats or public works, the
chief redistributed most goods he received back to the commoners, and land
was controlled by the community. But on the largest Polynesian islands, such
as Hawaii, Tahiti, and Tonga, chiefs were recognizable at a glance by their
ornaments, public works were erected by large labor forces, most tribute was
retained by the chiefs, and all land was controlled by them. A further
gradation among societies with ranked lineages was from those where the
political unit was a single autonomous village, to those consisting of a
regional assemblage of villages in which the largest village with a paramount
chief controlled the smaller villages with lesser chiefs.
BY NOW, IT should be obvious that chiefdoms introduced the dilemma
fundamental to all centrally governed, nonegalitarian societies. At best, they
do good by providing expensive services impossible to contract for on an
individual basis. At worst, they function unabashedly as kleptocracies,
transferring net wealth from commoners to upper classes. These noble and
selfish functions are inextricably linked, although some governments
emphasize much more of one function than of the other. The difference
between a kleptocrat and a wise statesman, between a robber baron and a
public benefactor, is merely one of degree: a matter of just how large a
percentage of the tribute extracted from producers is retained by the elite, and
how much the commoners like the public uses to which the redistributed
tribute is put. We consider President Mobutu of Zaire a kleptocrat because he
keeps too much tribute (the equivalent of billions of dollars) and redistributes
too little tribute (no functioning telephone system in Zaire). We consider
George Washington a statesman because he spent tax money on widely
admired programs and did not enrich himself as president. Nevertheless,
George Washington was born into wealth, which is much more unequally
distributed in the United States than in New Guinea villages.
For any ranked society, whether a chiefdom or a state, one thus has to ask:
why do the commoners tolerate the transfer of the fruits of their hard labor to
kleptocrats? This question, raised by political theorists from Plato to Marx, is
raised anew by voters in every modern election. Kleptocracies with little
public support run the risk of being overthrown, either by downtrodden
commoners or by upstart would-be replacement kleptocrats seeking public
support by promising a higher ratio of services rendered to fruits stolen. For
example, Hawaiian history was repeatedly punctuated by revolts against
repressive chiefs, usually led by younger brothers promising less oppression.
This may sound funny to us in the context of old Hawaii, until we reflect on
all the misery still being caused by such struggles in the modern world.
What should an elite do to gain popular support while still maintaining a
more comfortable lifestyle than commoners? Kleptocrats throughout the ages
have resorted to a mixture of four solutions:
1. Disarm the populace, and arm the elite. That’s much easier in these
days of high-tech weaponry, produced only in industrial plants and easily
monopolized by an elite, than in ancient times of spears and clubs easily made
at home.
2. Make the masses happy by redistributing much of the tribute received,
in popular ways. This principle was as valid for Hawaiian chiefs as it is for
American politicians today.
3. Use the monopoly of force to promote happiness, by maintaining
public order and curbing violence. This is potentially a big and
underappreciated advantage of centralized societies over noncentralized ones.
Anthropologists formerly idealized band and tribal societies as gentle and
nonviolent, because visiting anthropologists observed no murder in a band of
25 people in the course of a three-year study. Of course they didn’t: it’s easy
to calculate that a band of a dozen adults and a dozen children, subject to the
inevitable deaths occurring anyway for the usual reasons other than murder,
could not perpetuate itself if in addition one of its dozen adults murdered
another adult every three years. Much more extensive long-term information
about band and tribal societies reveals that murder is a leading cause of death.
For example, I happened to be visiting New Guinea’s Iyau people at a time
when a woman anthropologist was interviewing Iyau women about their life
histories. Woman after woman, when asked to name her husband, named
several sequential husbands who had died violent deaths. A typical answer
went like this: “My first husband was killed by Elopi raiders. My second
husband was killed by a man who wanted me, and who became my third
husband. That husband was killed by the brother of my second husband,
seeking to avenge his murder.” Such biographies prove common for so-called
gentle tribespeople and contributed to the acceptance of centralized authority
as tribal societies grew larger.
4. The remaining way for kleptocrats to gain public support is to construct
an ideology or religion justifying kleptocracy. Bands and tribes already had
supernatural beliefs, just as do modern established religions. But the
supernatural beliefs of bands and tribes did not serve to justify central
authority, justify transfer of wealth, or maintain peace between unrelated
individuals. When supernatural beliefs gained those functions and became
institutionalized, they were thereby transformed into what we term a religion.
Hawaiian chiefs were typical of chiefs elsewhere, in asserting divinity, divine
descent, or at least a hotline to the gods. The chief claimed to serve the people
by interceding for them with the gods and reciting the ritual formulas required
to obtain rain, good harvests, and success in fishing.
Chiefdoms characteristically have an ideology, precursor to an
institutionalized religion, that buttresses the chief’s authority. The chief may
either combine the offices of political leader and priest in a single person, or
may support a separate group of kleptocrats (that is, priests) whose function is
to provide ideological justification for the chiefs. That is why chiefdoms
devote so much collected tribute to constructing temples and other public
works, which serve as centers of the official religion and visible signs of the
chief’s power.
Besides justifying the transfer of wealth to kleptocrats, institutionalized
religion brings two other important benefits to centralized societies. First,
shared ideology or religion helps solve the problem of how unrelated
individuals are to live together without killing each other—by providing them
with a bond not based on kinship. Second, it gives people a motive, other than
genetic self-interest, for sacrificing their lives on behalf of others. At the cost
of a few society members who die in battle as soldiers, the whole society
becomes much more effective at conquering other societies or resisting
attacks.
THE POLITICAL, ECONOMIC, and social institutions most familiar to us today
are those of states, which now rule all of the world’s land area except for
Antarctica. Many early states and all modern ones have had literate elites, and
many modern states have literate masses as well. Vanished states tended to
leave visible archaeological hallmarks, such as ruins of temples with
standardized designs, at least four levels of settlement sizes, and pottery styles
covering tens of thousands of square miles. We thereby know that states arose
around 3700 B.C. in Mesopotamia and around 300 B.C. in Mesoamerica, over
2,000 years ago in the Andes, China, and Southeast Asia, and over 1,000
years ago in West Africa. In modern times the formation of states out of
chiefdoms has been observed repeatedly. Thus, we possess much more
information about past states and their formation than about past chiefdoms,
tribes, and bands.
Protostates extend many features of large paramount (multivillage)
chiefdoms. They continue the increase in size from bands to tribes to
chiefdoms. Whereas chiefdoms’ populations range from a few thousand to a
few tens of thousands, the populations of most modern states exceed one
million, and China’s exceeds one billion. The paramount chief’s location may
become the state’s capital city. Other population centers of states outside the
capital may also qualify as true cities, which are lacking in chiefdoms. Cities
differ from villages in their monumental public works, palaces of rulers,
accumulation of capital from tribute or taxes, and concentration of people
other than food producers.
Early states had a hereditary leader with a title equivalent to king, like a
super paramount chief and exercising an even greater monopoly of
information, decision making, and power. Even in democracies today, crucial
knowledge is available to only a few individuals, who control the flow of
information to the rest of the government and consequently control decisions.
For instance, in the Cuban Missile Crisis of 1962, information and discussions
that determined whether nuclear war would engulf half a billion people were
initially confined by President Kennedy to a ten-member executive committee
of the National Security Council that he himself appointed; then he limited
final decisions to a four-member group consisting of himself and three of his
cabinet ministers.
Central control is more far-reaching, and economic redistribution in the
form of tribute (renamed taxes) more extensive, in states than in chiefdoms.
Economic specialization is more extreme, to the point where today not even
farmers remain self-sufficient. Hence the effect on society is catastrophic
when state government collapses, as happened in Britain upon the removal of
Roman troops, administrators, and coinage between A.D. 407 and 411. Even
the earliest Mesopotamian states exercised centralized control of their
economies. Their food was produced by four specialist groups (cereal
farmers, herders, fishermen, and orchard and garden growers), from each of
which the state took the produce and to each of which it gave out the
necessary supplies, tools, and foods other than the type of food that this group
produced. The state supplied seeds and plow animals to the cereal farmers,
took wool from the herders, exchanged the wool by long-distance trade for
metal and other essential raw materials, and paid out food rations to the
laborers who maintained the irrigation systems on which the farmers
depended.
Many, perhaps most, early states adopted slavery on a much larger scale
than did chiefdoms. That was not because chiefdoms were more kindly
disposed toward defeated enemies but because the greater economic
specialization of states, with more mass production and more public works,
provided more uses for slave labor. In addition, the larger scale of state
warfare made more captives available.
A chiefdom’s one or two levels of administration are greatly multiplied in
states, as anyone who has seen an organizational chart of any government
knows. Along with the proliferation of vertical levels of bureaucrats, there is
also horizontal specialization. Instead of konohiki carrying out every aspect of
administration for a Hawaiian district, state governments have several
separate departments, each with its own hierarchy, to handle water
management, taxes, military draft, and so on. Even small states have more
complex bureaucracies than large chiefdoms. For instance, the West African
state of Maradi had a central administration with over 130 titled offices.
Internal conflict resolution within states has become increasingly
formalized by laws, a judiciary, and police. The laws are often written,
because many states (with conspicuous exceptions, such as that of the Incas)
have had literate elites, writing having been developed around the same time
as the formation of the earliest states in both Mesopotamia and Mesoamerica.
In contrast, no early chiefdom not on the verge of statehood developed
writing.
Early states had state religions and standardized temples. Many early
kings were considered divine and were accorded special treatment in
innumerable respects. For example, the Aztec and Inca emperors were both
carried about in litters; servants went ahead of the Inca emperor’s litter and
swept the ground clear; and the Japanese language includes special forms of
the pronoun “you” for use only in addressing the emperor. Early kings were
themselves the head of the state religion or else had separate high priests. The
Mesopotamian temple was the center not only of religion but also of
economic redistribution, writing, and crafts technology.
All these features of states carry to an extreme the developments that led
from tribes to chiefdoms. In addition, though, states have diverged from
chiefdoms in several new directions. The most fundamental such distinction is
that states are organized on political and territorial lines, not on the kinship
lines that defined bands, tribes, and simple chiefdoms. Furthermore, bands
and tribes always, and chiefdoms usually, consist of a single ethnic and
linguistic group. States, though—especially so-called empires formed by
amalgamation or conquest of states—are regularly multiethnic and
multilingual. State bureaucrats are not selected mainly on the basis of kinship,
as in chiefdoms, but are professionals selected at least partly on the basis of
training and ability. In later states, including most today, the leadership often
became nonhereditary, and many states abandoned the entire system of formal
hereditary classes carried over from chiefdoms.
OVER THE PAST 13,000 years the predominant trend in human society has
been the replacement of smaller, less complex units by larger, more complex
ones. Obviously, that is no more than an average long-term trend, with
innumerable shifts in either direction: 1,000 amalgamations for 999 reversals.
We know from our daily newspaper that large units (for instance, the former
USSR, Yugoslavia, and Czechoslovakia) can disintegrate into smaller units,
as did Alexander of Macedon’s empire over 2,000 years ago. More complex
units don’t always conquer less complex ones but may succumb to them, as
when the Roman and Chinese Empires were overrun by “barbarian” and
Mongol chiefdoms, respectively. But the long-term trend has still been toward
large, complex societies, culminating in states.
Obviously, too, part of the reason for states’ triumphs over simpler entities
when the two collide is that states usually enjoy an advantage of weaponry
and other technology, and a large numerical advantage in population. But
there are also two other potential advantages inherent in chiefdoms and states.
First, a centralized decision maker has the advantage at concentrating troops
and resources. Second, the official religions and patriotic fervor of many
states make their troops willing to fight suicidally.
The latter willingness is one so strongly programmed into us citizens of
modern states, by our schools and churches and governments, that we forget
what a radical break it marks with previous human history. Every state has its
slogan urging its citizens to be prepared to die if necessary for the state:
Britain’s “For King and Country,” Spain’s “Por Dios y España,” and so on.
Similar sentiments motivated 16th-century Aztec warriors: “There is nothing
like death in war, nothing like the flowery death so precious to Him [the
Aztec national god Huitzilopochtli] who gives life: far off I see it, my heart
yearns for it!”
Such sentiments are unthinkable in bands and tribes. In all the accounts
that my New Guinea friends have given me of their former tribal wars, there
has been not a single hint of tribal patriotism, of a suicidal charge, or of any
other military conduct carrying an accepted risk of being killed. Instead, raids
are initiated by ambush or by superior force, so as to minimize at all costs the
risk that one might die for one’s village. But that attitude severely limits the
military options of tribes, compared with state societies. Naturally, what
makes patriotic and religious fanatics such dangerous opponents is not the
deaths of the fanatics themselves, but their willingness to accept the deaths of
a fraction of their number in order to annihilate or crush their infidel enemy.
Fanaticism in war, of the type that drove recorded Christian and Islamic
conquests, was probably unknown on Earth until chiefdoms and especially
states emerged within the last 6,000 years.
HOW DID SMALL, noncentralized, kin-based societies evolve into large
centralized ones in which most members are not closely related to each other?
Having reviewed the stages in this transformation from bands to states, we
now ask what impelled societies thus to transform themselves.
At many moments in history, states have arisen independently—or, as
cultural anthropologists say, “pristinely,” that is, in the absence of any
preexisting surrounding states. Pristine state origins took place at least once,
possibly many times, on each of the continents except Australia and North
America. Prehistoric states included those of Mesopotamia, North China, the
Nile and Indus Valleys, Mesoamerica, the Andes, and West Africa. Native
states in contact with European states have arisen from chiefdoms repeatedly
in the last three centuries in Madagascar, Hawaii, Tahiti, and many parts of
Africa. Chiefdoms have arisen pristinely even more often, in all of the same
regions and in North America’s Southeast and Pacific Northwest, the
Amazon, Polynesia, and sub-Saharan Africa. All these origins of complex
societies give us a rich database for understanding their development.
Of the many theories addressing the problem of state origins, the simplest
denies that there is any problem to solve. Aristotle considered states the
natural condition of human society, requiring no explanation. His error was
understandable, because all the societies with which he would have been
acquainted—Greek societies of the fourth century B.C.—were states.
However, we now know that, as of A.D. 1492, much of the world was instead
organized into chiefdoms, tribes, or bands. State formation does demand an
explanation.
The next theory is the most familiar one. The French philosopher Jean-
Jacques Rousseau speculated that states are formed by a social contract, a
rational decision reached when people calculated their self-interest, came to
the agreement that they would be better off in a state than in simpler societies,
and voluntarily did away with their simpler societies. But observation and
historical records have failed to uncover a single case of a state’s being
formed in that ethereal atmosphere of dispassionate farsightedness. Smaller
units do not voluntarily abandon their sovereignty and merge into larger units.
They do so only by conquest, or under external duress.
A third theory, still popular with some historians and economists, sets out
from the undoubted fact that, in both Mesopotamia and North China and
Mexico, large-scale irrigation systems began to be constructed around the
time that states started to emerge. The theory also notes that any big, complex
system for irrigation or hydraulic management requires a centralized
bureaucracy to construct and maintain it. The theory then turns an observed
rough correlation in time into a postulated chain of cause and effect.
Supposedly, Mesopotamians and North Chinese and Mexicans foresaw the
advantages that a large-scale irrigation system would bring them, even though
there was at the time no such system within thousands of miles (or anywhere
on Earth) to illustrate for them those advantages. Those farsighted people
chose to merge their inefficient little chiefdoms into a larger state capable of
blessing them with large-scale irrigation.
However, this “hydraulic theory” of state formation is subject to the same
objections leveled against social contract theories in general. More
specifically, it addresses only the final stage in the evolution of complex
societies. It says nothing about what drove the progression from bands to
tribes to chiefdoms during all the millennia before the prospect of large-scale
irrigation loomed up on the horizon. When historical or archaeological dates
are examined in detail, they fail to support the view of irrigation as the driving
force for state formation. In Mesopotamia, North China, Mexico, and
Madagascar, small-scale irrigation systems already existed before the rise of
states. Construction of large-scale irrigation systems did not accompany the
emergence of states but came only significantly later in each of those areas. In
most of the states formed over the Maya area of Mesoamerica and the Andes,
irrigation systems always remained small-scale ones that local communities
could build and maintain themselves. Thus, even in those areas where
complex systems of hydraulic management did emerge, they were a
secondary consequence of states that must have formed for other reasons.
What seems to me to point to a fundamentally correct view of state
formation is an undoubted fact of much wider validity than the correlation
between irrigation and the formation of some states—namely, that the size of
the regional population is the strongest single predictor of societal complexity.
As we have seen, bands number a few dozen individuals, tribes a few
hundred, chiefdoms a few thousand to a few tens of thousands, and states
generally over about 50,000. In addition to that coarse correlation between
regional population size and type of society (band, tribe, and so on), there is a
finer trend, within each of those categories, between population and societal
complexity: for instance, that chiefdoms with large populations prove to be
the most centralized, stratified, and complex ones.
These correlations suggest strongly that regional population size or
population density or population pressure has something to do with the
formation of complex societies. But the correlations do not tell us precisely
how population variables function in a chain of cause and effect whose
outcome is a complex society. To trace out that chain, let us now remind
ourselves how large dense populations themselves arise. Then we can
examine why a large but simple society could not maintain itself. With that as
background, we shall finally return to the question of how a simpler society
actually becomes more complex as the regional population increases.
WE HAVE SEEN that large or dense populations arise only under conditions
of food production, or at least under exceptionally productive conditions for
hunting-gathering. Some productive hunter-gatherer societies reached the
organizational level of chiefdoms, but none reached the level of states: all
states nourish their citizens by food production. These considerations, along
with the just mentioned correlation between regional population size and
societal complexity, have led to a protracted chicken-or-egg debate about the
causal relations between food production, population variables, and societal
complexity. Is it intensive food production that is the cause, triggering
population growth and somehow leading to a complex society? Or are large
populations and complex societies instead the cause, somehow leading to
intensification of food production?
Posing the question in that either-or form misses the point. Intensified
food production and societal complexity stimulate each other, by
autocatalysis. That is, population growth leads to societal complexity, by
mechanisms that we shall discuss, while societal complexity in turn leads to
intensified food production and thereby to population growth. Complex
centralized societies are uniquely capable of organizing public works
(including irrigation systems), long-distance trade (including the importation
of metals to make better agricultural tools), and activities of different groups
of economic specialists (such as feeding herders with farmers’ cereal, and
transferring the herders’ livestock to farmers for use as plow animals). All of
these capabilities of centralized societies have fostered intensified food
production and hence population growth throughout history.
In addition, food production contributes in at least three ways to specific
features of complex societies. First, it involves seasonally pulsed inputs of
labor. When the harvest has been stored, the farmers’ labor becomes available
for a centralized political authority to harness—in order to build public works
advertising state power (such as the Egyptian pyramids), or to build public
works that could feed more mouths (such as Polynesian Hawaii’s irrigation
systems or fishponds), or to undertake wars of conquest to form larger
political entities.
Second, food production may be organized so as to generate stored food
surpluses, which permit economic specialization and social stratification. The
surpluses can be used to feed all tiers of a complex society: the chiefs,
bureaucrats, and other members of the elite; the scribes, craftspeople, and
other non-food-producing specialists; and the farmers themselves, during
times that they are drafted to construct public works.
Finally, food production permits or requires people to adopt sedentary
living, which is a prerequisite for accumulating substantial possessions,
developing elaborate technology and crafts, and constructing public works.
The importance of fixed residence to a complex society explains why
missionaries and governments, whenever they make first contact with
previously uncontacted nomadic tribes or bands in New Guinea or the
Amazon, universally have two immediate goals. One goal, of course, is the
obvious one of “pacifying” the nomads: that is, dissuading them from killing
missionaries, bureaucrats, or each other. The other goal is to induce the
nomads to settle in villages, so that the missionaries and bureaucrats can find
the nomads, bring them services such as medical care and schools, and
proselytize and control them.
THUS, FOOD PRODUCTION, which increases population size, also acts in many
ways to make features of complex societies possible. But that doesn’t prove
that food production and large populations make complex societies inevitable.
How can we account for the empirical observation that band or tribal
organization just does not work for societies of hundreds of thousands of
people, and that all existing large societies have complex centralized
organization? We can cite at least four obvious reasons.
One reason is the problem of conflict between unrelated strangers. That
problem grows astronomically as the number of people making up the society
increases. Relationships within a band of 20 people involve only 190 two-
person interactions (20 people times 19 divided by 2), but a band of 2,000
would have 1,999,000 dyads. Each of those dyads represents a potential time
bomb that could explode in a murderous argument. Each murder in band and
tribal societies usually leads to an attempted revenge killing, starting one
more unending cycle of murder and countermurder that destabilizes the
society.
In a band, where everyone is closely related to everyone else, people
related simultaneously to both quarreling parties step in to mediate quarrels.
In a tribe, where many people are still close relatives and everyone at least
knows everybody else by name, mutual relatives and mutual friends mediate
the quarrel. But once the threshold of “several hundred,” below which
everyone can know everyone else, has been crossed, increasing numbers of
dyads become pairs of unrelated strangers. When strangers fight, few people
present will be friends or relatives of both combatants, with self-interest in
stopping the fight. Instead, many onlookers will be friends or relatives of only
one combatant and will side with that person, escalating the two-person fight
into a general brawl. Hence a large society that continues to leave conflict
resolution to all of its members is guaranteed to blow up. That factor alone
would explain why societies of thousands can exist only if they develop
centralized authority to monopolize force and resolve conflicts.
A second reason is the growing impossibility of communal decision
making with increasing population size. Decision making by the entire adult
population is still possible in New Guinea villages small enough that news
and information quickly spread to everyone, that everyone can hear everyone
else in a meeting of the whole village, and that everyone who wants to speak
at the meeting has the opportunity to do so. But all those prerequisites for
communal decision making become unattainable in much larger communities.
Even now, in these days of microphones and loudspeakers, we all know that a
group meeting is no way to resolve issues for a group of thousands of people.
Hence a large society must be structured and centralized if it is to reach
decisions effectively.
A third reason involves economic considerations. Any society requires
means to transfer goods between its members. One individual may happen to
acquire more of some essential commodity on one day and less on another.
Because individuals have different talents, one individual consistently tends
to wind up with an excess of some essentials and a deficit of others. In small
societies with few pairs of members, the resulting necessary transfers of
goods can be arranged directly between pairs of individuals or families, by
reciprocal exchanges. But the same mathematics that makes direct pairwise
conflict resolution inefficient in large societies makes direct pairwise
economic transfers also inefficient. Large societies can function economically
only if they have a redistributive economy in addition to a reciprocal
economy. Goods in excess of an individual’s needs must be transferred from
the individual to a centralized authority, which then redistributes the goods to
individuals with deficits.
A final consideration mandating complex organization for large societies
has to do with population densities. Large societies of food producers have
not only more members but also higher population densities than do small
bands of hunter-gatherers. Each band of a few dozen hunters occupies a large
territory, within which they can acquire most of the resources essential to
them. They can obtain their remaining necessities by trading with neighboring
bands during intervals between band warfare. As population density
increases, the territory of that band-sized population of a few dozen would
shrink to a small area, with more and more of life’s necessities having to be
obtained outside the area. For instance, one couldn’t just divide Holland’s
16,000 square miles and 16,000,000 people into 800,000 individual
territories, each encompassing 13 acres and serving as home to an
autonomous band of 20 people who remained self-sufficient confined within
their 13 acres, occasionally taking advantage of a temporary truce to come to
the borders of their tiny territory in order to exchange some trade items and
brides with the next band. Such spatial realities require that densely populated
regions support large and complexly organized societies.
Considerations of conflict resolution, decision making, economics, and
space thus converge in requiring large societies to be centralized. But
centralization of power inevitably opens the door—for those who hold the
power, are privy to information, make the decisions, and redistribute the
goods—to exploit the resulting opportunities to reward themselves and their
relatives. To anyone familiar with any modern grouping of people, that’s
obvious. As early societies developed, those acquiring centralized power
gradually established themselves as an elite, perhaps originating as one of
several formerly equal-ranked village clans that became “more equal” than
the others.
THOSE ARE THE reasons why large societies cannot function with band
organization and instead are complex kleptocracies. But we are still left with
the question of how small, simple societies actually evolve or amalgamate
into large, complex ones. Amalgamation, centralized conflict resolution,
decision making, economic redistribution, and kleptocratic religion don’t just
develop automatically through a Rousseauesque social contract. What drives
the amalgamation?
In part, the answer depends upon evolutionary reasoning. I said at the
outset of this chapter that societies classified in the same category are not all
identical to each other, because humans and human groups are infinitely
diverse. For example, among bands and tribes, the big-men of some are
inevitably more charismatic, powerful, and skilled in reaching decisions than
the big-men of others. Among large tribes, those with stronger big-men and
hence greater centralization tend to have an advantage over those with less
centralization. Tribes that resolve conflicts as poorly as did the Fayu tend to
blow apart again into bands, while ill-governed chiefdoms blow apart into
smaller chiefdoms or tribes. Societies with effective conflict resolution, sound
decision making, and harmonious economic redistribution can develop better
technology, concentrate their military power, seize larger and more productive
territories, and crush autonomous smaller societies one by one.
Thus, competition between societies at one level of complexity tends to
lead to societies on the next level of complexity if conditions permit. Tribes
conquer or combine with tribes to reach the size of chiefdoms, which conquer
or combine with other chiefdoms to reach the size of states, which conquer or
combine with other states to become empires. More generally, large units
potentially enjoy an advantage over individual small units if—and that’s a big
“if”—the large units can solve the problems that come with their larger size,
such as perennial threats from upstart claimants to leadership, commoner
resentment of kleptocracy, and increased problems associated with economic
integration.
The amalgamation of smaller units into larger ones has often been
documented historically or archaeologically. Contrary to Rousseau, such
amalgamations never occur by a process of unthreatened little societies freely
deciding to merge, in order to promote the happiness of their citizens. Leaders
of little societies, as of big ones, are jealous of their independence and
prerogatives. Amalgamation occurs instead in either of two ways: by merger
under the threat of external force, or by actual conquest. Innumerable
examples are available to illustrate each mode of amalgamation.
Merger under the threat of external force is well illustrated by the
formation of the Cherokee Indian confederation in the U.S. Southeast. The
Cherokees were originally divided into 30 or 40 independent chiefdoms, each
consisting of a village of about 400 people. Increasing white settlement led to
conflicts between Cherokees and whites. When individual Cherokees robbed
or assaulted white settlers and traders, the whites were unable to discriminate
among the different Cherokee chiefdoms and retaliated indiscriminately
against any Cherokees, either by military action or by cutting off trade. In
response, the Cherokee chiefdoms gradually found themselves compelled to
join into a single confederacy in the course of the 18th century. Initially, the
larger chiefdoms in 1730 chose an overall leader, a chief named Moytoy, who
was succeeded in 1741 by his son. The first task of these leaders was to
punish individual Cherokees who attacked whites, and to deal with the white
government. Around 1758 the Cherokees regularized their decision making
with an annual council modeled on previous village councils and meeting at
one village (Echota), which thereby became a de facto “capital.” Eventually,
the Cherokees became literate (as we saw in Chapter 12) and adopted a
written constitution.
The Cherokee confederacy was thus formed not by conquest but by the
amalgamation of previously jealous smaller entities, which merged only when
threatened with destruction by powerful external forces. In much the same
way, in an example of state formation described in every American history
textbook, the white American colonies themselves, one of which (Georgia)
had precipitated the formation of the Cherokee state, were impelled to form a
nation of their own when threatened with the powerful external force of the
British monarchy. The American colonies were initially as jealous of their
autonomy as the Cherokee chiefdoms, and their first attempt at amalgamation
under the Articles of Confederation (1781) proved unworkable because it
reserved too much autonomy to the ex-colonies. Only further threats, notably
Shays’s Rebellion of 1786 and the unsolved burden of war debt, overcame the
ex-colonies’ extreme reluctance to sacrifice autonomy and pushed them into
adopting our current strong federal constitution in 1787. The 19th-century
unification of Germany’s jealous principalities proved equally difficult. Three
early attempts (the Frankfurt Parliament of 1848, the restored German
Confederation of 1850, and the North German Confederation of 1866) failed
before the external threat of France’s declaration of war in 1870 finally led to
the princelets’ surrendering much of their power to a central imperial German
government in 1871.
The other mode of formation of complex societies, besides merger under
threat of external force, is merger by conquest. A well-documented example is
the origin of the Zulu state, in southeastern Africa. When first observed by
white settlers, the Zulus were divided into dozens of little chiefdoms. During
the late 1700s, as population pressure rose, fighting between the chiefdoms
became increasingly intense. Among all those chiefdoms, the ubiquitous
problem of devising centralized power structures was solved most
successfully by a chief called Dingiswayo, who gained ascendancy of the
Mtetwa chiefdom by killing a rival around 1807. Dingiswayo developed a
superior centralized military organization by drafting young men from all
villages and grouping them into regiments by age rather than by their village.
He also developed superior centralized political organization by abstaining
from slaughter as he conquered other chiefdoms, leaving the conquered
chief’s family intact, and limiting himself to replacing the conquered chief
himself with a relative willing to cooperate with Dingiswayo. He developed
superior centralized conflict resolution by expanding the adjudication of
quarrels. In that way Dingiswayo was able to conquer and begin the
integration of 30 other Zulu chiefdoms. His successors strengthened the
resulting embryonic Zulu state by expanding its judicial system, policing, and
ceremonies.
This Zulu example of a state formed by conquest can be multiplied almost
indefinitely. Native states whose formation from chiefdoms happened to be
witnessed by Europeans in the 18th and 19th centuries include the Polynesian
Hawaiian state, the Polynesian Tahitian state, the Merina state of Madagascar,
Lesotho and Swazi and other southern African states besides that of the Zulus,
the Ashanti state of West Africa, and the Ankole and Buganda states of
Uganda. The Aztec and Inca Empires were formed by 15th-century
conquests, before Europeans arrived, but we know much about their
formation from Indian oral histories transcribed by early Spanish settlers. The
formation of the Roman state and the expansion of the Macedonian Empire
under Alexander were described in detail by contemporary classical authors.
All these examples illustrate that wars, or threats of war, have played a
key role in most, if not all, amalgamations of societies. But wars, even
between mere bands, have been a constant fact of human history. Why is it,
then, that they evidently began causing amalgamations of societies only
within the past 13,000 years? We had already concluded that the formation of
complex societies is somehow linked to population pressure, so we should
now seek a link between population pressure and the outcome of war. Why
should wars tend to cause amalgamations of societies when populations are
dense but not when they are sparse? The answer is that the fate of defeated
peoples depends on population density, with three possible outcomes:
Where population densities are very low, as is usual in regions occupied
by hunter-gatherer bands, survivors of a defeated group need only move
farther away from their enemies. That tends to be the result of wars between
nomadic bands in New Guinea and the Amazon.
Where population densities are moderate, as in regions occupied by food-
producing tribes, no large vacant areas remain to which survivors of a
defeated band can flee. But tribal societies without intensive food production
have no employment for slaves and do not produce large enough food
surpluses to be able to yield much tribute. Hence the victors have no use for
survivors of a defeated tribe, unless to take the women in marriage. The
defeated men are killed, and their territory may be occupied by the victors.
Where population densities are high, as in regions occupied by states or
chiefdoms, the defeated still have nowhere to flee, but the victors now have
two options for exploiting them while leaving them alive. Because chiefdoms
and state societies have economic specialization, the defeated can be used as
slaves, as commonly happened in biblical times. Alternatively, because many
such societies have intensive food production systems capable of yielding
large surpluses, the victors can leave the defeated in place but deprive them of
political autonomy, make them pay regular tribute in food or goods, and
amalgamate their society into the victorious state or chiefdom. This has been
the usual outcome of battles associated with the founding of states or empires
throughout recorded history. For example, the Spanish conquistadores wished
to exact tribute from Mexico’s defeated native populations, so they were very
interested in the Aztec Empire’s tribute lists. It turned out that the tribute
received by the Aztecs each year from subject peoples had included 7,000
tons of corn, 4,000 tons of beans, 4,000 tons of grain amaranth, 2,000,000
cotton cloaks, and huge quantities of cacao beans, war costumes, shields,
feather headdresses, and amber.
Thus, food production, and competition and diffusion between societies,
led as ultimate causes, via chains of causation that differed in detail but that
all involved large dense populations and sedentary living, to the proximate
agents of conquest: germs, writing, technology, and centralized political
organization. Because those ultimate causes developed differently on different
continents, so did those agents of conquest. Hence those agents tended to
arise in association with each other, but the association was not strict: for
example, an empire arose without writing among the Incas, and writing with
few epidemic diseases among the Aztecs. Dingiswayo’s Zulus illustrate that
each of those agents contributed somewhat independently to history’s pattern.
Among the dozens of Zulu chiefdoms, the Mtetwa chiefdom enjoyed no
advantage whatsoever of technology, writing, or germs over the other
chiefdoms, which it nevertheless succeeded in defeating. Its advantage lay
solely in the spheres of government and ideology. The resulting Zulu state
was thereby enabled to conquer a fraction of a continent for nearly a century.
PART FOUR
AROUND THE WORLD IN FIVE CHAPTERS
CHAPTER 15
YALI’S PEOPLE
WHEN MY WIFE, MARIE, AND I WERE VACATIONING IN Australia one summer,
we decided to visit a site with well-preserved Aboriginal rock paintings in the
desert near the town of Menindee. While I knew of the Australian desert’s
reputation for dryness and summer heat, I had already spent long periods
working under hot, dry conditions in the Californian desert and New Guinea
savanna, so I considered myself experienced enough to deal with the minor
challenges we would face as tourists in Australia. Carrying plenty of drinking
water, Marie and I set off at noon on a hike of a few miles to the paintings.
The trail from the ranger station led uphill, under a cloudless sky, through
open terrain offering no shade whatsoever. The hot, dry air that we were
breathing reminded me of how it had felt to breathe while sitting in a Finnish
sauna. By the time we reached the cliff site with the paintings, we had
finished our water. We had also lost our interest in art, so we pushed on uphill,
breathing slowly and regularly. Presently I noticed a bird that was
unmistakably a species of babbler, but it seemed enormous compared with
any known babbler species. At that point, I realized that I was experiencing
heat hallucinations for the first time in my life. Marie and I decided that we
had better head straight back.
Both of us stopped talking. As we walked, we concentrated on listening to
our breathing, calculating the distance to the next landmark, and estimating
the remaining time. My mouth and tongue were now dry, and Marie’s face
was red. When we at last reached the air-conditioned ranger station, we
sagged into chairs next to the water cooler, drank down the cooler’s last half-
gallon of water, and asked the ranger for another bottle. Sitting there
exhausted, both physically and emotionally, I reflected that the Aborigines
who had made those paintings had somehow spent their entire lives in that
desert without air-conditioned retreats, managing to find food as well as
water.
To white Australians, Menindee is famous as the base camp for two
whites who had suffered worse from the desert’s dry heat over a century
earlier: the Irish policeman Robert Burke and the English astronomer William
Wills, ill-fated leaders of the first European expedition to cross Australia from
south to north. Setting out with six camels packing food enough for three
months, Burke and Wills ran out of provisions while in the desert north of
Menindee. Three successive times, they encountered and were rescued by
well-fed Aborigines whose home was that desert, and who plied the explorers
with fish, fern cakes, and roasted fat rats. But then Burke foolishly shot his
pistol at one of the Aborigines, whereupon the whole group fled. Despite their
big advantage over the Aborigines in possessing guns with which to hunt,
Burke and Wills starved, collapsed, and died within a month after the
Aborigines’ departure.
My wife’s and my experience at Menindee, and the fate of Burke and
Wills, made vivid for me the difficulties of building a human society in
Australia. Australia stands out from all the other continents: the differences
between Eurasia, Africa, North America, and South America fade into
insignificance compared with the differences between Australia and any of
those other landmasses. Australia is by far the driest, smallest, flattest, most
infertile, climatically most unpredictable, and biologically most impoverished
continent. It was the last continent to be occupied by Europeans. Until then, it
had supported the most distinctive human societies, and the least numerous
human population, of any continent.
Australia thus provides a crucial test of theories about intercontinental
differences in societies. It had the most distinctive environment, and also the
most distinctive societies. Did the former cause the latter? If so, how?
Australia is the logical continent with which to begin our around-the-world
tour, applying the lessons of Parts 2 and 3 to understanding the differing
histories of all the continents.
MOST LAYPEOPLE WOULD describe as the most salient feature of Native
Australian societies their seeming “backwardness.” Australia is the sole
continent where, in modern times, all native peoples still lived without any of
the hallmarks of so-called civilization—without farming, herding, metal,
bows and arrows, substantial buildings, settled villages, writing, chiefdoms,
or states. Instead, Australian Aborigines were nomadic or seminomadic
hunter-gatherers, organized into bands, living in temporary shelters or huts,
and still dependent on stone tools. During the last 13,000 years less cultural
change has accumulated in Australia than in any other continent. The
prevalent European view of Native Australians was already typified by the
words of an early French explorer, who wrote, “They are the most miserable
people of the world, and the human beings who approach closest to brute
beasts.”
Yet, as of 40,000 years ago, Native Australian societies enjoyed a big
head start over societies of Europe and the other continents. Native
Australians developed some of the earliest known stone tools with ground
edges, the earliest hafted stone tools (that is, stone ax heads mounted on
handles), and by far the earliest watercraft, in the world. Some of the oldest
known painting on rock surfaces comes from Australia. Anatomically modern
humans may have settled Australia before they settled western Europe. Why,
despite that head start, did Europeans end up conquering Australia, rather than
vice versa?
Within that question lies another. During the Pleistocene Ice Ages, when
much ocean water was sequestered in continental ice sheets and sea level
dropped far below its present stand, the shallow Arafura Sea now separating
Australia from New Guinea was low, dry land. With the melting of ice sheets
between around 12,000 and 8,000 years ago, sea level rose, that low land
became flooded, and the former continent of Greater Australia became
sundered into the two hemi-continents of Australia and New Guinea (Figure
15.1 on Chapter 15).
The human societies of those two formerly joined landmasses were in
modern times very different from each other. In contrast to everything that I
just said about Native Australians, most New Guineans, such as Yali’s people,
were farmers and swineherds. They lived in settled villages and were
organized politically into tribes rather than bands. All New Guineans had
bows and arrows, and many used pottery. New Guineans tended to have much
more substantial dwellings, more seaworthy boats, and more numerous and
more varied utensils than did Australians. As a consequence of being food
producers instead of hunter-gatherers, New Guineans lived at much higher
average population densities than Australians: New Guinea has only one-tenth
of Australia’s area but supported a native population several times that of
Australia’s.
Why did the human societies of the larger landmass derived from
Pleistocene Greater Australia remain so “backward” in their development,
while the societies of the smaller landmass “advanced” much more rapidly?
Why didn’t all those New Guinea innovations spread to Australia, which is
separated from New Guinea by only 90 miles of sea at Torres Strait? From the
perspective of cultural anthropology, the geographic distance between
Australia and New Guinea is even less than 90 miles, because Torres Strait is
sprinkled with islands inhabited by farmers using bows and arrows and
culturally resembling New Guineans. The largest Torres Strait island lies only
10 miles from Australia. Islanders carried on a lively trade with Native
Australians as well as with New Guineans. How could two such different
cultural universes maintain themselves across a calm strait only 10 miles wide
and routinely traversed by canoes?
Compared with Native Australians, New Guineans rate as culturally
“advanced.” But most other modern people consider even New Guineans
“backward.” Until Europeans began to colonize New Guinea in the late 19th
century, all New Guineans were nonliterate, dependent on stone tools, and
politically not yet organized into states or (with few exceptions) chiefdoms.
Granted that New Guineans had “progressed” beyond Native Australians,
why had they not yet “progressed” as far as many Eurasians, Africans, and
Native Americans? Thus, Yali’s people and their Australian cousins pose a
puzzle inside a puzzle.
When asked to account for the cultural “backwardness” of Aboriginal
Australian society, many white Australians have a simple answer: supposed
deficiencies of the Aborigines themselves. In facial structure and skin color,
Aborigines certainly look different from Europeans, leading some late-19th
century authors to consider them a missing link between apes and humans.
How else can one account for the fact that white English colonists created a
literate, food-producing, industrial democracy, within a few decades of
colonizing a continent whose inhabitants after more than 40,000 years were
still nonliterate hunter-gatherers? It is especially striking that Australia has
some of the world’s richest iron and aluminum deposits, as well as rich
reserves of copper, tin, lead, and zinc. Why, then, were Native Australians
still ignorant of metal tools and living in the Stone Age?
It seems like a perfectly controlled experiment in the evolution of human
societies. The continent was the same; only the people were different. Ergo,
the explanation for the differences between Native Australian and European-
Australian societies must lie in the different people composing them. The
logic behind this racist conclusion appears compelling. We shall see, however,
that it contains a simple error.
AS THE FIRST step in examining this logic, let us examine the origins of the
peoples themselves. Australia and New Guinea were both occupied by at least
40,000 years ago, at a time when they were both still joined as Greater
Australia. A glance at a map (Figure 15.1) suggests that the colonists must
have originated ultimately from the nearest continent, Southeast Asia, by
island hopping through the Indonesian Archipelago. This conclusion is
supported by genetic relationships between modern Australians, New
Guineans, and Asians, and by the survival today of a few populations of
somewhat similar physical appearance in the Philippines, Malay Peninsula,
and Andaman Islands off Myanmar.
Once the colonists had reached the shores of Greater Australia, they
spread quickly over the whole continent to occupy even its farthest reaches
and most inhospitable habitats. By 40,000 years ago, fossils and stone tools
attest to their presence in Australia’s southwestern corner; by 35,000 years
ago, in Australia’s southeastern corner and Tasmania, the corner of Australia
most remote from the colonists’ likely beachhead in western Australia or New
Guinea (the parts nearest Indonesia and Asia); and by 30,000 years ago, in the
cold New Guinea highlands. All of those areas could have been reached
overland from a western beachhead. However, the colonization of both the
Bismarck and the Solomon Archipelagoes northeast of New Guinea, by
35,000 years ago, required further overwater crossings of dozens of miles.
The occupation could have been even more rapid than that apparent spread of
dates from 40,000 to 30,000 years ago, since the various dates hardly differ
within the experimental error of the radiocarbon method.
At the Pleistocene times when Australia and New Guinea were initially
occupied, the Asian continent extended eastward to incorporate the modern
islands of Borneo, Java, and Bali, nearly 1,000 miles nearer to Australia and
New Guinea than Southeast Asia’s present margin. However, at least eight
channels up to 50 miles wide still remained to be crossed in getting from
Borneo or Bali to Pleistocene Greater Australia. Forty thousand years ago,
those crossings may have been achieved by bamboo rafts, low-tech but
seaworthy watercraft still in use in coastal South China today. The crossings
must nevertheless have been difficult, because after that initial landfall by
40,000 years ago the archaeological record provides no compelling evidence
of further human arrivals in Greater Australia from Asia for tens of thousands
of years. Not until within the last few thousand years do we encounter the
next firm evidence, in the form of the appearance of Asian-derived pigs in
New Guinea and Asian-derived dogs in Australia.
Thus, the human societies of Australia and New Guinea developed in
substantial isolation from the Asian societies that founded them. That
isolation is reflected in languages spoken today. After all those millennia of
isolation, neither modern Aboriginal Australian languages nor the major
group of modern New Guinea languages (the so-called Papuan languages)
exhibit any clear relationships with any modern Asian languages.
The isolation is also reflected in genes and physical anthropology. Genetic
studies suggest that Aboriginal Australians and New Guinea highlanders are
somewhat more similar to modern Asians than to peoples of other continents,
but the relationship is not a close one. In skeletons and physical appearance,
Aboriginal Australians and New Guineans are also distinct from most
Southeast Asian populations, as becomes obvious if one compares photos of
Australians or New Guineans with those of Indonesians or Chinese. Part of
the reason for all these differences is that the initial Asian colonists of Greater
Australia have had a long time in which to diverge from their stay-at-home
Asian cousins, with only limited genetic exchanges during most of that time.
But probably a more important reason is that the original Southeast Asian
stock from which the colonists of Greater Australia were derived has by now
been largely replaced by other Asians expanding out of China.
Aboriginal Australians and New Guineans have also diverged genetically,
physically, and linguistically from each other. For instance, among the major
(genetically determined) human blood groups, groups B of the so-called ABO
system and S of the MNS system occur in New Guinea as well as in most of
the rest of the world, but both are virtually absent in Australia. The tightly
coiled hair of most New Guineans contrasts with the straight or wavy hair of
most Australians. Australian languages and New Guinea’s Papuan languages
are unrelated not only to Asian languages but also to each other, except for
some spread of vocabulary in both directions across Torres Strait.
All that divergence of Australians and New Guineans from each other
reflects lengthy isolation in very different environments. Since the rise of the
Arafura Sea finally separated Australia and New Guinea from each other
around 10,000 years ago, gene exchange has been limited to tenuous contact
via the chain of Torres Strait islands. That has allowed the populations of the
two hemi-continents to adapt to their own environments. While the savannas
and mangroves of coastal southern New Guinea are fairly similar to those of
northern Australia, other habitats of the hemi-continents differ in almost all
major respects.
Here are some of the differences. New Guinea lies nearly on the equator,
while Australia extends far into the temperate zones, reaching almost 40
degrees south of the equator. New Guinea is mountainous and extremely
rugged, rising to 16,500 feet and with glaciers capping the highest peaks,
while Australia is mostly low and flat—94 percent of its area lies below 2,000
feet of elevation. New Guinea is one of the wettest areas on Earth, Australia
one of the driest. Most of New Guinea receives over 100 inches of rain
annually, and much of the highlands receives over 200 inches, while most of
Australia receives less than 20 inches. New Guinea’s equatorial climate varies
only modestly from season to season and year to year, but Australia’s climate
is highly seasonal and varies from year to year far more than that of any other
continent. As a result, New Guinea is laced with permanent large rivers, while
Australia’s permanently flowing rivers are confined in most years to eastern
Australia, and even Australia’s largest river system (the Murray-Darling) has
ceased flowing for months during droughts. Most of New Guinea’s land area
is clothed in dense rain forest, while most of Australia’s supports only desert
and open dry woodland.
New Guinea is covered with young fertile soil, as a consequence of
volcanic activity, glaciers repeatedly advancing and retreating and scouring
the highlands, and mountain streams carrying huge quantities of silt to the
lowlands. In contrast, Australia has by far the oldest, most infertile, most
nutrient-leached soils of any continent, because of Australia’s little volcanic
activity and its lack of high mountains and glaciers. Despite having only one-
tenth of Australia’s area, New Guinea is home to approximately as many
mammal and bird species as is Australia—a result of New Guinea’s equatorial
location, much higher rainfall, much greater range of elevations, and greater
fertility. All of those environmental differences influenced the two hemi-
continents’ very disparate cultural histories, which we shall now consider.
THE EARLIEST AND most intensive food production, and the densest
populations, of Greater Australia arose in the highland valleys of New Guinea
at altitudes between 4,000 and 9,000 feet above sea level. Archaeological
excavations uncovered complex systems of drainage ditches dating back to
9,000 years ago and becoming extensive by 6,000 years ago, as well as
terraces serving to retain soil moisture in drier areas. The ditch systems were
similar to those still used today in the highlands to drain swampy areas for use
as gardens. By around 5,000 years ago, pollen analyses testify to widespread
deforestation of highland valleys, suggesting forest clearance for agriculture.
Today, the staple crops of highland agriculture are the recently introduced
sweet potato, along with taro, bananas, yams, sugarcane, edible grass stems,
and several leafy vegetables. Because taro, bananas, and yams are native to
Southeast Asia, an undoubted site of plant domestication, it used to be
assumed that New Guinea highland crops other than sweet potatoes arrived
from Asia. However, it was eventually realized that the wild ancestors of
sugarcane, the leafy vegetables, and the edible grass stems are New Guinea
species, that the particular types of bananas grown in New Guinea have New
Guinea rather than Asian wild ancestors, and that taro and some yams are
native to New Guinea as well as to Asia. If New Guinea agriculture had really
had Asian origins, one might have expected to find highland crops derived
unequivocally from Asia, but there are none. For those reasons it is now
generally acknowledged that agriculture arose indigenously in the New
Guinea highlands by domestication of New Guinea wild plant species.
New Guinea thus joins the Fertile Crescent, China, and a few other
regions as one of the world’s centers of independent origins of plant
domestication. No remains of the crops actually being grown in the highlands
6,000 years ago have been preserved in archaeological sites. However, that is
not surprising, because modern highland staple crops are plant species that do
not leave archaeologically visible residues except under exceptional
conditions. Hence it seems likely that some of them were also the founding
crops of highland agriculture, especially as the ancient drainage systems
preserved are so similar to the modern drainage systems used for growing
taro.
The three unequivocally foreign elements in New Guinea highland food
production as seen by the first European explorers were chickens, pigs, and
sweet potatoes. Chickens and pigs were domesticated in Southeast Asia and
introduced around 3,600 years ago to New Guinea and most other Pacific
islands by Austronesians, a people of ultimately South Chinese origin whom
we shall discuss in Chapter 17. (Pigs may have arrived earlier.) As for the
sweet potato, native to South America, it apparently reached New Guinea
only within the last few centuries, following its introduction to the Philippines
by Spaniards. Once established in New Guinea, the sweet potato overtook
taro as the highland’s leading crop, because of its shorter time required to
reach maturity, higher yields per acre, and greater tolerance of poor soil
conditions.
The development of New Guinea highland agriculture must have
triggered a big population explosion thousands of years ago, because the
highlands could have supported only very low population densities of hunter-
gatherers after New Guinea’s original megafauna of giant marsupials had
been exterminated. The arrival of the sweet potato triggered a further
explosion in recent centuries. When Europeans first flew over the highlands
in the 1930s, they were astonished to see below them a landscape similar to
Holland’s. Broad valleys were completely deforested and dotted with villages,
and drained and fenced fields for intensive food production covered entire
valley floors. That landscape testifies to the population densities achieved in
the highlands by farmers with stone tools.
Steep terrain, persistent cloud cover, malaria, and risk of drought at lower
elevations confine New Guinea highland agriculture to elevations above about
4,000 feet. In effect, the New Guinea highlands are an island of dense farming
populations thrust up into the sky and surrounded below by a sea of clouds.
Lowland New Guineans on the seacoast and rivers are villagers depending
heavily on fish, while those on dry ground away from the coast and rivers
subsist at low densities by slash-and-burn agriculture based on bananas and
yams, supplemented by hunting and gathering. In contrast, lowland New
Guinea swamp dwellers live as nomadic hunter-gatherers dependent on the
starchy pith of wild sago palms, which are very productive and yield three
times more calories per hour of work than does gardening. New Guinea
swamps thus provide a clear instance of an environment where people
remained hunter-gatherers because farming could not compete with the
hunting-gathering lifestyle.
The sago eaters persisting in lowland swamps exemplify the nomadic
hunter-gatherer band organization that must formerly have characterized all
New Guineans. For all the reasons that we discussed in Chapters 13 and 14,
the farmers and the fishing peoples were the ones to develop more-complex
technology, societies, and political organization. They live in permanent
villages and tribal societies, often led by a big-man. Some of them construct
large, elaborately decorated, ceremonial houses. Their great art, in the form of
wooden statues and masks, is prized in museums around the world.
NEW GUINEA THUS became the part of Greater Australia with the most-
advanced technology, social and political organization, and art. However,
from an urban American or European perspective, New Guinea still rates as
“primitive” rather than “advanced.” Why did New Guineans continue to use
stone tools instead of developing metal tools, remain nonliterate, and fail to
organize themselves into chiefdoms and states? It turns out that New Guinea
had several biological and geographic strikes against it.
First, although indigenous food production did arise in the New Guinea
highlands, we saw in Chapter 8 that it yielded little protein. The dietary
staples were low-protein root crops, and production of the sole domesticated
animal species (pigs and chickens) was too low to contribute much to
people’s protein budgets. Since neither pigs nor chickens can be harnessed to
pull carts, highlanders remained without sources of power other than human
muscle power, and also failed to evolve epidemic diseases to repel the
eventual European invaders.
A second restriction on the size of highland populations was the limited
available area: the New Guinea highlands have only a few broad valleys,
notably the Wahgi and Baliem Valleys, capable of supporting dense
populations. Still a third limitation was the reality that the mid-montane zone
between 4,000 and 9,000 feet was the sole altitudinal zone in New Guinea
suitable for intensive food production. There was no food production at all in
New Guinea alpine habitats above 9,000 feet, little on the hillslopes between
4,000 and 1,000 feet, and only low-density slash-and-burn agriculture in the
lowlands. Thus, large-scale economic exchanges of food, between
communities at different altitudes specializing in different types of food
production, never developed in New Guinea. Such exchanges in the Andes,
Alps, and Himalayas not only increased population densities in those areas,
by providing people at all altitudes with a more balanced diet, but also
promoted regional economic and political integration.
For all these reasons, the population of traditional New Guinea never
exceeded 1,000,000 until European colonial governments brought Western
medicine and the end of intertribal warfare. Of the approximately nine world
centers of agricultural origins that we discussed in Chapter 5, New Guinea
remained the one with by far the smallest population. With a mere 1,000,000
people, New Guinea could not develop the technology, writing, and political
systems that arose among populations of tens of millions in China, the Fertile
Crescent, the Andes, and Mesoamerica.
New Guinea’s population is not only small in aggregate, but also
fragmented into thousands of micropopulations by the rugged terrain: swamps
in much of the lowlands, steep-sided ridges and narrow canyons alternating
with each other in the highlands, and dense jungle swathing both the lowlands
and the highlands. When I am engaged in biological exploration in New
Guinea, with teams of New Guineans as field assistants, I consider excellent
progress to be three miles per day even if we are traveling over existing trails.
Most highlanders in traditional New Guinea never went more than 10 miles
from home in the course of their lives.
Those difficulties of terrain, combined with the state of intermittent
warfare that characterized relations between New Guinea bands or villages,
account for traditional New Guinea’s linguistic, cultural, and political
fragmentation. New Guinea has by far the highest concentration of languages
in the world: 1,000 out of the world’s 6,000 languages, crammed into an area
only slightly larger than that of Texas, and divided into dozens of language
families and isolated languages as different from each other as English is from
Chinese. Nearly half of all New Guinea languages have fewer than 500
speakers, and even the largest language groups (still with a mere 100,000
speakers) were politically fragmented into hundreds of villages, fighting as
fiercely with each other as with speakers of other languages. Each of those
microsocieties alone was far too small to support chiefs and craft specialists,
or to develop metallurgy and writing.
Besides a small and fragmented population, the other limitation on
development in New Guinea was geographic isolation, restricting the inflow
of technology and ideas from elsewhere. New Guinea’s three neighbors were
all separated from New Guinea by water gaps, and until a few thousand years
ago they were all even less advanced than New Guinea (especially the New
Guinea highlands) in technology and food production. Of those three
neighbors, Aboriginal Australians remained hunter-gatherers with almost
nothing to offer New Guineans that New Guineans did not already possess.
New Guinea’s second neighbor was the much smaller islands of the Bismarck
and the Solomon Archipelagoes to the east. That left, as New Guinea’s third
neighbor, the islands of eastern Indonesia. But that area, too, remained a
cultural backwater occupied by hunter-gatherers for most of its history. There
is no item that can be identified as having reached New Guinea via Indonesia,
after the initial colonization of New Guinea over 40,000 years ago, until the
time of the Austronesian expansion around 1600 B.C.
With that expansion, Indonesia became occupied by food producers of
Asian origins, with domestic animals, with agriculture and technology at least
as complex as New Guinea’s, and with navigational skills that served as a
much more efficient conduit from Asia to New Guinea. Austronesians settled
on islands west and north and east of New Guinea, and in the far west and on
the north and southeast coasts of New Guinea itself. Austronesians introduced
pottery, chickens, and probably dogs and pigs to New Guinea. (Early
archaeological surveys claimed pig bones in the New Guinea highlands by
4000 B.C., but those claims have not been confirmed.) For at least the last
thousand years, trade connected New Guinea to the technologically much
more advanced societies of Java and China. In return for exporting bird of
paradise plumes and spices, New Guineans received Southeast Asian goods,
including even such luxury items as Dong Son bronze drums and Chinese
porcelain.
With time, the Austronesian expansion would surely have had more
impact on New Guinea. Western New Guinea would eventually have been
incorporated politically into the sultanates of eastern Indonesia, and metal
tools might have spread through eastern Indonesia to New Guinea. But—that
hadn’t happened by A.D. 1511, the year the Portuguese arrived in the
Moluccas and truncated Indonesia’s separate train of developments. When
Europeans reached New Guinea soon thereafter, its inhabitants were still
living in bands or in fiercely independent little villages, and still using stone
tools.
WHILE THE NEW Guinea hemi-continent of Greater Australia thus developed
both animal husbandry and agriculture, the Australian hemi-continent
developed neither. During the Ice Ages Australia had supported even more
big marsupials than New Guinea, including diprotodonts (the marsupial
equivalent of cows and rhinoceroses), giant kangaroos, and giant wombats.
But all those marsupial candidates for animal husbandry disappeared in the
wave of extinctions (or exterminations) that accompanied human colonization
of Australia. That left Australia, like New Guinea, with no domesticable
native mammals. The sole foreign domesticated mammal adopted in Australia
was the dog, which arrived from Asia (presumably in Austronesian canoes)
around 1500 B.C. and established itself in the wild in Australia to become the
dingo. Native Australians kept captive dingos as companions, watchdogs, and
even as living blankets, giving rise to the expression “five-dog night” to mean
a very cold night. But they did not use dingos / dogs for food, as did
Polynesians, or for cooperative hunting of wild animals, as did New
Guineans.
Agriculture was another nonstarter in Australia, which is not only the
driest continent but also the one with the most infertile soils. In addition,
Australia is unique in that the overwhelming influence on climate over most
of the continent is an irregular nonannual cycle, the ENSO (acronym for El
Niño Southern Oscillation), rather than the regular annual cycle of the seasons
so familiar in most other parts of the world. Unpredictable severe droughts
last for years, punctuated by equally unpredictable torrential rains and floods.
Even today, with Eurasian crops and with trucks and railroads to transport
produce, food production in Australia remains a risky business. Herds build
up in good years, only to be killed off by drought. Any incipient farmers in
Aboriginal Australia would have faced similar cycles in their own
populations. If in good years they had settled in villages, grown crops, and
produced babies, those large populations would have starved and died off in
drought years, when the land could support far fewer people.
The other major obstacle to the development of food production in
Australia was the paucity of domesticable wild plants. Even modern European
plant geneticists have failed to develop any crop except macadamia nuts from
Australia’s native wild flora. The list of the world’s potential prize cereals—
the 56 wild grass species with the heaviest grains—includes only two
Australian species, both of which rank near the bottom of the list (grain
weight only 13 milligrams, compared with a whopping 40 milligrams for the
heaviest grains elsewhere in the world). That’s not to say that Australia had
no potential crops at all, or that Aboriginal Australians would never have
developed indigenous food production. Some plants, such as certain species
of yams, taro, and arrowroot, are cultivated in southern New Guinea but also
grow wild in northern Australia and were gathered by Aborigines there. As
we shall see, Aborigines in the climatically most favorable areas of Australia
were evolving in a direction that might have eventuated in food production.
But any food production that did arise indigenously in Australia would have
been limited by the lack of domesticable animals, the poverty of domesticable
plants, and the difficult soils and climate.
Nomadism, the hunter-gatherer lifestyle, and minimal investment in
shelter and possessions were sensible adaptations to Australia’s ENSO-driven
resource unpredictability. When local conditions deteriorated, Aborigines
simply moved to an area where conditions were temporarily better. Rather
than depending on just a few crops that could fail, they minimized risk by
developing an economy based on a great variety of wild foods, not all of
which were likely to fail simultaneously. Instead of having fluctuating
populations that periodically outran their resources and starved, they
maintained smaller populations that enjoyed an abundance of food in good
years and a sufficiency in bad years.
The Aboriginal Australian substitute for food production has been termed
“firestick farming.” The Aborigines modified and managed the surrounding
landscape in ways that increased its production of edible plants and animals,
without resorting to cultivation. In particular, they intentionally burned much
of the landscape periodically. That served several purposes: the fires drove
out animals that could be killed and eaten immediately; fires converted dense
thickets into open parkland in which people could travel more easily; the
parkland was also an ideal habitat for kangaroos, Australia’s prime game
animal; and the fires stimulated the growth both of new grass on which
kangaroos fed and of fern roots on which Aborigines themselves fed.
We think of Australian Aborigines as desert people, but most of them
were not. Instead, their population densities varied with rainfall (because it
controls the production of terrestrial wild plant and animal foods) and with
abundance of aquatic foods in the sea, rivers, and lakes. The highest
population densities of Aborigines were in Australia’s wettest and most
productive regions: the Murray-Darling river system of the Southeast, the
eastern and northern coasts, and the southwestern corner. Those areas also
came to support the densest populations of European settlers in modern
Australia. The reason we think of Aborigines as desert people is simply that
Europeans killed or drove them out of the most desirable areas, leaving the
last intact Aboriginal populations only in areas that Europeans didn’t want.
Within the last 5,000 years, some of those productive regions witnessed
an intensification of Aboriginal food-gathering methods, and a buildup of
Aboriginal population density. Techniques were developed in eastern
Australia for rendering abundant and starchy, but extremely poisonous, cycad
seeds edible, by leaching out or fermenting the poison. The previously
unexploited highlands of southeastern Australia began to be visited regularly
during the summer, by Aborigines feasting not only on cycad nuts and yams
but also on huge hibernating aggregations of a migratory moth called the
bogong moth, which tastes like a roasted chestnut when grilled. Another type
of intensified food-gathering activity that developed was the freshwater eel
fisheries of the Murray-Darling river system, where water levels in marshes
fluctuate with seasonal rains. Native Australians constructed elaborate
systems of canals up to a mile and a half long, in order to enable eels to
extend their range from one marsh to another. Eels were caught by equally
elaborate weirs, traps set in dead-end side canals, and stone walls across
canals with a net placed in an opening of the wall. Traps at different levels in
the marsh came into operation as the water level rose and fell. While the
initial construction of those “fish farms” must have involved a lot of work,
they then fed many people. Nineteenth-century European observers found
villages of a dozen Aboriginal houses at the eel farms, and there are
archaeological remains of villages of up to 146 stone houses, implying at least
seasonally resident populations of hundreds of people.
Still another development in eastern and northern Australia was the
harvesting of seeds of a wild millet, belonging to the same genus as the
broomcorn millet that was a staple of early Chinese agriculture. The millet
was reaped with stone knives, piled into haystacks, and threshed to obtain the
seeds, which were then stored in skin bags or wooden dishes and finally
ground with millstones. Several of the tools used in this process, such as the
stone reaping knives and grindstones, were similar to the tools independently
invented in the Fertile Crescent for processing seeds of other wild grasses. Of
all the food-acquiring methods of Aboriginal Australians, millet harvesting is
perhaps the one most likely to have evolved eventually into crop production.
Along with intensified food gathering in the last 5,000 years came new
types of tools. Small stone blades and points provided more length of sharp
edge per pound of tool than the large stone tools they replaced. Hatchets with
ground stone edges, once present only locally in Australia, became
widespread. Shell fishhooks appeared within the last thousand years.
WHY DID AUSTRALIA not develop metal tools, writing, and politically
complex societies? A major reason is that Aborigines remained hunter-
gatherers, whereas, as we saw in Chapters 12–14, those developments arose
elsewhere only in populous and economically specialized societies of food
producers. In addition, Australia’s aridity, infertility, and climatic
unpredictability limited its hunter-gatherer population to only a few hundred
thousand people. Compared with the tens of millions of people in ancient
China or Mesoamerica, that meant that Australia had far fewer potential
inventors, and far fewer societies to experiment with adopting innovations.
Nor were its several hundred thousand people organized into closely
interacting societies. Aboriginal Australia instead consisted of a sea of very
sparsely populated desert separating several more productive ecological
“islands,” each of them holding only a fraction of the continent’s population
and with interactions attenuated by the intervening distance. Even within the
relatively moist and productive eastern side of the continent, exchanges
between societies were limited by the 1,900 miles from Queensland’s tropical
rain forests in the northeast to Victoria’s temperate rain forests in the
southeast, a geographic and ecological distance as great as that from Los
Angeles to Alaska.
Some apparent regional or continentwide regressions of technology in
Australia may stem from the isolation and relatively few inhabitants of its
population centers. The boomerang, that quintessential Australian weapon,
was abandoned in the Cape York Peninsula of northeastern Australia. When
encountered by Europeans, the Aborigines of southwestern Australia did not
eat shellfish. The function of the small stone points that appear in Australian
archaeological sites around 5,000 years ago remains uncertain: while an easy
explanation is that they may have been used as spearpoints and barbs, they are
suspiciously similar to the stone points and barbs used on arrows elsewhere in
the world. If they really were so used, the mystery of bows and arrows being
present in modern New Guinea but absent in Australia might be compounded:
perhaps bows and arrows actually were adopted for a while, then abandoned,
across the Australian continent. All these examples remind us of the
abandonment of guns in Japan, of bows and arrows and pottery in most of
Polynesia, and of other technologies in other isolated societies (Chapter 13).
The most extreme losses of technology in the Australian region took place
on the island of Tasmania, 130 miles off the coast of southeastern Australia.
At Pleistocene times of low sea level, the shallow Bass Strait now separating
Tasmania from Australia was dry land, and the people occupying Tasmania
were part of the human population distributed continuously over an expanded
Australian continent. When the strait was at last flooded around 10,000 years
ago, Tasmanians and mainland Australians became cut off from each other
because neither group possessed watercraft capable of negotiating Bass Strait.
Thereafter, Tasmania’s population of 4,000 hunter-gatherers remained out of
contact with all other humans on Earth, living in an isolation otherwise known
only from science fiction novels.
When finally encountered by Europeans in A.D. 1642, the Tasmanians had
the simplest material culture of any people in the modern world. Like
mainland Aborigines, they were hunter-gatherers without metal tools. But
they also lacked many technologies and artifacts widespread on the mainland,
including barbed spears, bone tools of any type, boomerangs, ground or
polished stone tools, hafted stone tools, hooks, nets, pronged spears, traps,
and the practices of catching and eating fish, sewing, and starting a fire. Some
of these technologies may have arrived or been invented in mainland
Australia only after Tasmania became isolated, in which case we can conclude
that the tiny Tasmanian population did not independently invent these
technologies for itself. Others of these technologies were brought to Tasmania
when it was still part of the Australian mainland, and were subsequently lost
in Tasmania’s cultural isolation. For example, the Tasmanian archaeological
record documents the disappearance of fishing, and of awls, needles, and
other bone tools, around 1500 B.C. On at least three smaller islands (Flinders,
Kangaroo, and King) that were isolated from Australia or Tasmania by rising
sea levels around 10,000 years ago, human populations that would initially
have numbered around 200 to 400 died out completely.
Tasmania and those three smaller islands thus illustrate in extreme form a
conclusion of broad potential significance for world history. Human
populations of only a few hundred people were unable to survive indefinitely
in complete isolation. A population of 4,000 was able to survive for 10,000
years, but with significant cultural losses and significant failures to invent,
leaving it with a uniquely simplified material culture. Mainland Australia’s
300,000 hunter-gatherers were more numerous and less isolated than the
Tasmanians but still constituted the smallest and most isolated human
population of any of the continents. The documented instances of
technological regression on the Australian mainland, and the example of
Tasmania, suggest that the limited repertoire of Native Australians compared
with that of peoples of other continents may stem in part from the effects of
isolation and population size on the development and maintenance of
technology—like those effects on Tasmania, but less extreme. By implication,
the same effects may have contributed to differences in technology between
the largest continent (Eurasia) and the next smaller ones (Africa, North
America, and South America).
WHY DIDN’T MORE-ADVANCED technology reach Australia from its
neighbors, Indonesia and New Guinea? As regards Indonesia, it was separated
from northwestern Australia by water and was very different from it
ecologically. In addition, Indonesia itself was a cultural and technological
backwater until a few thousand years ago. There is no evidence of any new
technology or introduction reaching Australia from Indonesia, after
Australia’s initial colonization 40,000 years ago, until the dingo appeared
around 1500 B.C.
The dingo reached Australia at the peak of the Austronesian expansion
from South China through Indonesia. Austronesians succeeded in settling all
the islands of Indonesia, including the two closest to Australia—Timor and
Tanimbar (only 275 and 205 miles from modern Australia, respectively).
Since Austronesians covered far greater sea distances in the course of their
expansion across the Pacific, we would have to assume that they repeatedly
reached Australia, even if we did not have the evidence of the dingo to prove
it. In historical times northwestern Australia was visited each year by sailing
canoes from the Macassar district on the Indonesian island of Sulawesi
(Celebes), until the Australian government stopped the visits in 1907.
Archaeological evidence traces the visits back until around A.D. 1000, and
they may well have been going on earlier. The main purpose of the visits was
to obtain sea cucumbers (also known as bêche-de-mer or trepang), starfish
relatives exported from Macassar to China as a reputed aphrodisiac and prized
ingredient of soups.
Naturally, the trade that developed during the Macassans’ annual visits
left many legacies in northwestern Australia. The Macassans planted tamarind
trees at their coastal campsites and sired children by Aboriginal women.
Cloth, metal tools, pottery, and glass were brought as trade goods, though
Aborigines never learned to manufacture those items themselves. Aborigines
did acquire from the Macassans some loan words, some ceremonies, and the
practices of using dugout sailing canoes and smoking tobacco in pipes.
But none of these influences altered the basic character of Australian
society. More important than what happened as a result of the Macassan visits
is what did not happen. The Macassans did not settle in Australia—
undoubtedly because the area of northwestern Australia facing Indonesia is
much too dry for Macassan agriculture. Had Indonesia faced the tropical rain
forests and savannas of northeastern Australia, the Macassans could have
settled, but there is no evidence that they ever traveled that far. Since the
Macassans thus came only in small numbers and for temporary visits and
never penetrated inland, just a few groups of Australians on a small stretch of
coast were exposed to them. Even those few Australians got to see only a
fraction of Macassan culture and technology, rather than a full Macassan
society with rice fields, pigs, villages, and workshops. Because the
Australians remained nomadic hunter-gatherers, they acquired only those few
Macassan products and practices compatible with their lifestyle. Dugout
sailing canoes and pipes, yes; forges and pigs, no.
Apparently much more astonishing than Australians’ resistance to
Indonesian influence is their resistance to New Guinea influence. Across the
narrow ribbon of water known as Torres Strait, New Guinea farmers who
spoke New Guinea languages and had pigs, pottery, and bows and arrows
faced Australian hunter-gatherers who spoke Australian languages and lacked
pigs, pottery, and bows and arrows. Furthermore, the strait is not an open-
water barrier but is dotted with a chain of islands, of which the largest
(Muralug Island) lies only 10 miles from the Australian coast. There were
regular trading visits between Australia and the islands, and between the
islands and New Guinea. Many Aboriginal women came as wives to Muralug
Island, where they saw gardens and bows and arrows. How was it that those
New Guinea traits did not get transmitted to Australia?
This cultural barrier at Torres Strait is astonishing only because we may
mislead ourselves into picturing a full-fledged New Guinea society with
intensive agriculture and pigs 10 miles off the Australian coast. In reality,
Cape York Aborigines never saw a mainland New Guinean. Instead, there
was trade between New Guinea and the islands nearest New Guinea, then
between those islands and Mabuiag Island halfway down the strait, then
between Mabuiag Island and Badu Island farther down the strait, then
between Badu Island and Muralug Island, and finally between Muralug and
Cape York.
New Guinea society became attenuated along that island chain. Pigs were
rare or absent on the islands. Lowland South New Guineans along Torres
Strait practiced not the intensive agriculture of the New Guinea highlands but
a slash-and-burn agriculture with heavy reliance on seafoods, hunting, and
gathering. The importance of even those slash-and-burn practices decreased
from southern New Guinea toward Australia along the island chain. Muralug
Island itself, the island nearest Australia, was dry, marginal for agriculture,
and supported only a small human population, which subsisted mainly on
seafood, wild yams, and mangrove fruits.
The interface between New Guinea and Australia across Torres Strait was
thus reminiscent of the children’s game of telephone, in which children sit in
a circle, one child whispers a word into the ear of the second child, who
whispers what she thinks she has just heard to the third child, and the word
finally whispered by the last child back to the first child bears no resemblance
to the initial word. In the same way, trade along the Torres Strait islands was a
telephone game that finally presented Cape York Aborigines with something
very different from New Guinea society. In addition, we should not imagine
that relations between Muralug Islanders and Cape York Aborigines were an
uninterrupted love feast at which Aborigines eagerly sopped up culture from
island teachers. Trade instead alternated with war for the purposes of head-
hunting and capturing women to become wives.
Despite the dilution of New Guinea culture by distance and war, some
New Guinea influence did manage to reach Australia. Intermarriage carried
New Guinea physical features, such as coiled rather than straight hair, down
the Cape York Peninsula. Four Cape York languages had phonemes unusual
for Australia, possibly because of the influence of New Guinea languages.
The most important transmissions were of New Guinea shell fishhooks, which
spread far into Australia, and of New Guinea outrigger canoes, which spread
down the Cape York Peninsula. New Guinea drums, ceremonial masks,
funeral posts, and pipes were also adopted on Cape York. But Cape York
Aborigines did not adopt agriculture, in part because what they saw of it on
Muralug Island was so watered-down. They did not adopt pigs, of which there
were few or none on the islands, and which they would in any case have been
unable to feed without agriculture. Nor did they adopt bows and arrows,
remaining instead with their spears and spear-throwers.
Australia is big, and so is New Guinea. But contact between those two big
landmasses was restricted to those few small groups of Torres Strait islanders
with a highly attenuated New Guinea culture, interacting with those few small
groups of Cape York Aborigines. The latter groups’ decisions, for whatever
reason, to use spears rather than bows and arrows, and not to adopt certain
other features of the diluted New Guinea culture they saw, blocked
transmission of those New Guinea cultural traits to all the rest of Australia.
As a result, no New Guinea trait except shell fishhooks spread far into
Australia. If the hundreds of thousands of farmers in the cool New Guinea
highlands had been in close contact with the Aborigines in the cool highlands
of southeastern Australia, a massive transfer of intensive food production and
New Guinea culture to Australia might have followed. But the New Guinea
highlands are separated from the Australian highlands by 2,000 miles of
ecologically very different landscape. The New Guinea highlands might as
well have been the mountains of the moon, as far as Australians’ chances of
observing and adopting New Guinea highland practices were concerned.
In short, the persistence of Stone Age nomadic hunter-gatherers in
Australia, trading with Stone Age New Guinea farmers and Iron Age
Indonesian farmers, at first seems to suggest singular obstinacy on the part of
Native Australians. On closer examination, it merely proves to reflect the
ubiquitous role of geography in the transmission of human culture and
technology.
IT REMAINS FOR us to consider the encounters of New Guinea’s and
Australia’s Stone Age societies with Iron Age Europeans. A Portuguese
navigator “discovered” New Guinea in 1526, Holland claimed the western
half in 1828, and Britain and Germany divided the eastern half in 1884. The
first Europeans settled on the coast, and it took them a long time to penetrate
into the interior, but by 1960 European governments had established political
control over most New Guineans.
The reasons that Europeans colonized New Guinea, rather than vice versa,
are obvious. Europeans were the ones who had the oceangoing ships and
compasses to travel to New Guinea; the writing systems and printing presses
to produce maps, descriptive accounts, and administrative paperwork useful
in establishing control over New Guinea; the political institutions to organize
the ships, soldiers, and administration; and the guns to shoot New Guineans
who resisted with bow and arrow and clubs. Yet the number of European
settlers was always very small, and today New Guinea is still populated
largely by New Guineans. That contrasts sharply with the situation in
Australia, the Americas, and South Africa, where European settlement was
numerous and lasting and replaced the original native population over large
areas. Why was New Guinea different?
A major factor was the one that defeated all European attempts to settle
the New Guinea lowlands until the 1880s: malaria and other tropical diseases,
none of them an acute epidemic crowd infection as discussed in Chapter 11.
The most ambitious of those failed lowland settlement plans, organized by the
French marquis de Rays around 1880 on the nearby island of New Ireland,
ended with 930 out of the 1,000 colonists dead within three years. Even with
modern medical treatments available today, many of my American and
European friends in New Guinea have been forced to leave because of
malaria, hepatitis, or other diseases, while my own health legacy of New
Guinea has been a year of malaria and a year of dysentery.
As Europeans were being felled by New Guinea lowland germs, why
were Eurasian germs not simultaneously felling New Guineans? Some New
Guineans did become infected, but not on the massive scale that killed off
most of the native peoples of Australia and the Americas. One lucky break for
New Guineans was that there were no permanent European settlements in
New Guinea until the 1880s, by which time public health discoveries had
made progress in bringing smallpox and other infectious diseases of European
populations under control. In addition, the Austronesian expansion had
already been bringing a stream of Indonesian settlers and traders to New
Guinea for 3,500 years. Since Asian mainland infectious diseases were well
established in Indonesia, New Guineans thereby gained long exposure and
built up much more resistance to Eurasian germs than did Aboriginal
Australians.
The sole part of New Guinea where Europeans do not suffer from severe
health problems is the highlands, above the altitudinal ceiling for malaria. But
the highlands, already occupied by dense populations of New Guineans, were
not reached by Europeans until the 1930s. By then, the Australian and Dutch
colonial governments were no longer willing to open up lands for white
settlement by killing native people in large numbers or driving them off their
lands, as had happened during earlier centuries of European colonialism.
The remaining obstacle to European would-be settlers was that European
crops, livestock, and subsistence methods do poorly everywhere in the New
Guinea environment and climate. While introduced tropical American crops
such as squash, corn, and tomatoes are now grown in small quantities, and tea
and coffee plantations have been established in the highlands of Papua New
Guinea, staple European crops, like wheat, barley, and peas, have never taken
hold. Introduced cattle and goats, kept in small numbers, suffer from tropical
diseases, just as do European people themselves. Food production in New
Guinea is still dominated by the crops and agricultural methods that New
Guineans perfected over the course of thousands of years.
All those problems of disease, rugged terrain, and subsistence contributed
to Europeans’ leaving eastern New Guinea (now the independent nation of
Papua New Guinea) occupied and governed by New Guineans, who
nevertheless use English as their official language, write with the alphabet,
live under democratic governmental institutions modeled on those of
England, and use guns manufactured overseas. The outcome was different in
western New Guinea, which Indonesia took over from Holland in 1963 and
renamed Irian Jaya province. The province is now governed by Indonesians,
for Indonesians. Its rural population is still overwhelmingly New Guinean,
but its urban population is Indonesian, as a result of government policy aimed
at encouraging Indonesian immigration. Indonesians, with their long history
of exposure to malaria and other tropical diseases shared with New Guineans,
have not faced as potent a germ barrier as have Europeans. They are also
better prepared than Europeans for subsisting in New Guinea, because
Indonesian agriculture already included bananas, sweet potatoes, and some
other staple crops of New Guinea agriculture. The ongoing changes in Irian
Jaya represent the continuation, backed by a centralized government’s full
resources, of the Austronesian expansion that began to reach New Guinea
3,500 years ago. Indonesians are modern Austronesians.
EUROPEANS COLONIZED AUSTRALIA, rather than Native Australians colonizing
Europe, for the same reasons that we have just seen in the case of New
Guinea. However, the fates of New Guineans and of Aboriginal Australians
were very different. Today, Australia is populated and governed by 20 million
non-Aborigines, most of them of European descent, plus increasing numbers
of Asians arriving since Australia abandoned its previous White Australia
immigration policy in 1973. The Aboriginal population declined by 80
percent, from around 300,000 at the time of European settlement to a
minimum of 60,000 in 1921. Aborigines today form an underclass of
Australian society. Many of them live on mission stations or government
reserves, or else work for whites as herdsmen on cattle stations. Why did
Aborigines fare so much worse than New Guineans?
The basic reason is Australia’s suitability (in some areas) for European
food production and settlement, combined with the role of European guns,
germs, and steel in clearing Aborigines out of the way. While I already
stressed the difficulties posed by Australia’s climate and soils, its most
productive or fertile areas can nevertheless support European farming.
Agriculture in the Australian temperate zone is now dominated by the
Eurasian temperate-zone staple crops of wheat (Australia’s leading crop),
barley, oats, apples, and grapes, along with sorghum and cotton of African
Sahel origins and potatoes of Andean origins. In tropical areas of northeastern
Australia (Queensland) beyond the optimal range of Fertile Crescent crops,
European farmers introduced sugarcane of New Guinea origins, bananas and
citrus fruit of tropical Southeast Asian origins, and peanuts of tropical South
American origins. As for livestock, Eurasian sheep made it possible to extend
food production to arid areas of Australia unsuitable for agriculture, and
Eurasian cattle joined crops in moister areas.
Thus, the development of food production in Australia had to await the
arrival of non-native crops and animals domesticated in climatically similar
parts of the world too remote for their domesticates to reach Australia until
brought by transoceanic shipping. Unlike New Guinea, most of Australia
lacked diseases serious enough to keep out Europeans. Only in tropical
northern Australia did malaria and other tropical diseases force Europeans to
abandon their 19th-century attempts at settlement, which succeeded only with
the development of 20th-century medicine.
Australian Aborigines, of course, stood in the way of European food
production, especially because what was potentially the most productive
farmland and dairy country initially supported Australia’s densest populations
of Aboriginal hunter-gatherers. European settlement reduced the number of
Aborigines by two means. One involved shooting them, an option that
Europeans considered more acceptable in the 19th and late 18th centuries than
when they entered the New Guinea highlands in the 1930s. The last large-
scale massacre, of 31 Aborigines, occurred at Alice Springs in 1928. The
other means involved European-introduced germs to which Aborigines had
had no opportunity to acquire immunity or to evolve genetic resistance.
Within a year of the first European settlers’ arrival at Sydney, in 1788, corpses
of Aborigines who had died in epidemics became a common sight. The
principal recorded killers were smallpox, influenza, measles, typhoid, typhus,
chicken pox, whooping cough, tuberculosis, and syphilis.
In these two ways, independent Aboriginal societies were eliminated in
all areas suitable for European food production. The only societies that
survived more or less intact were those in areas of northern and western
Australia useless to Europeans. Within one century of European colonization,
40,000 years of Aboriginal traditions had been mostly swept away.
WE CAN NOW return to the problem that I posed near the beginning of this
chapter. How, except by postulating deficiencies in the Aborigines
themselves, can one account for the fact that white English colonists
apparently created a literate, food-producing, industrial democracy, within a
few decades of colonizing a continent whose inhabitants after more than
40,000 years were still nonliterate nomadic hunter-gatherers? Doesn’t that
constitute a perfectly controlled experiment in the evolution of human
societies, forcing us to a simple racist conclusion?
The resolution of this problem is simple. White English colonists did not
create a literate, food-producing, industrial democracy in Australia. Instead,
they imported all of the elements from outside Australia: the livestock, all of
the crops (except macadamia nuts), the metallurgical knowledge, the steam
engines, the guns, the alphabet, the political institutions, even the germs. All
these were the end products of 10,000 years of development in Eurasian
environments. By an accident of geography, the colonists who landed at
Sydney in 1788 inherited those elements. Europeans have never learned to
survive in Australia or New Guinea without their inherited Eurasian
technology. Robert Burke and William Wills were smart enough to write, but
not smart enough to survive in Australian desert regions where Aborigines
were living.
The people who did create a society in Australia were Aboriginal
Australians. Of course, the society that they created was not a literate, food-
producing, industrial democracy. The reasons follow straightforwardly from
features of the Australian environment.
CHAPTER 16
HOW CHINA BECAME CHINESE
IMMIGRATION, AFFIRMATIVE ACTION, MULTILINGUALISM, ethnic diversity—my
state of California was among the pioneers of these controversial policies and
is now pioneering a backlash against them. A glance into the classrooms of
the Los Angeles public school system, where my sons are being educated,
fleshes out the abstract debates with the faces of children. Those children
represent over 80 languages spoken in the home, with English-speaking
whites in the minority. Every single one of my sons’ playmates has at least
one parent or grandparent who was born outside the United States; that’s true
of three of my own sons’ four grandparents. But immigration is merely
restoring the diversity that America held for thousands of years. Before
European settlement, the mainland United States was home to hundreds of
Native American tribes and languages and came under control of a single
government only within the last hundred years.
In these respects the United States is a thoroughly “normal” country. All
but one of the world’s six most populous nations are melting pots that
achieved political unification recently, and that still support hundreds of
languages and ethnic groups. For example, Russia, once a small Slavic state
centered on Moscow, did not even begin its expansion beyond the Ural
Mountains until A.D. 1582. From then until the 19th century, Russia proceeded
to swallow up dozens of non-Slavic peoples, many of which retain their
original language and cultural identity. Just as American history is the story of
how our continent’s expanse became American, Russia’s history is the story
of how Russia became Russian. India, Indonesia, and Brazil are also recent
political creations (or re-creations, in the case of India), home to about 850,
670, and 210 languages, respectively.
The great exception to this rule of the recent melting pot is the world’s
most populous nation, China. Today, China appears politically, culturally, and
linguistically monolithic, at least to laypeople. It was already unified
politically in 221 B.C. and has remained so for most of the centuries since
then. From the beginnings of literacy in China, it has had only a single writing
system, whereas modern Europe uses dozens of modified alphabets. Of
China’s 1.2 billion people, over 800 million speak Mandarin, the language
with by far the largest number of native speakers in the world. Some 300
million others speak seven other languages as similar to Mandarin, and to
each other, as Spanish is to Italian. Thus, not only is China not a melting pot,
but it seems absurd to ask how China became Chinese. China has been
Chinese, almost from the beginnings of its recorded history.
We take this seeming unity of China so much for granted that we forget
how astonishing it is. One reason why we should not have expected such
unity is genetic. While a coarse racial classification of world peoples lumps
all Chinese people as so-called Mongoloids, that category conceals much
more variation than the differences between Swedes, Italians, and Irish within
Europe. In particular, North and South Chinese are genetically and physically
rather different: North Chinese are most similar to Tibetans and Nepalese,
while South Chinese are similar to Vietnamese and Filipinos. My North and
South Chinese friends can often distinguish each other at a glance by physical
appearance: the North Chinese tend to be taller, heavier, paler, with more
pointed noses, and with smaller eyes that appear more “slanted” (because of
what is termed their epicanthic fold).
North and South China differ in environment and climate as well: the
north is drier and colder; the south, wetter and hotter. Genetic differences
arising in those differing environments imply a long history of moderate
isolation between peoples of North and South China. How did those peoples
nevertheless end up with the same or very similar languages and cultures?
China’s apparent linguistic near-unity is also puzzling in view of the
linguistic disunity of other long-settled parts of the world. For instance, we
saw in the last chapter that New Guinea, with less than one-tenth of China’s
area and with only about 40,000 years of human history, has a thousand
languages, including dozens of language groups whose differences are far
greater than those among the eight main Chinese languages. Western Europe
has evolved or acquired about 40 languages just in the 6,000–8,000 years
since the arrival of Indo-European languages, including languages as different
as English, Finnish, and Russian. Yet fossils attest to human presence in
China for over half a million years. What happened to the tens of thousands of
distinct languages that must have arisen in China over that long time span?
These paradoxes hint that China too was once diverse, as all other
populous nations still are. China differs only by having been unified much
earlier. Its “Sinification” involved the drastic homogenization of a huge
region in an ancient melting pot, the repopulation of tropical Southeast Asia,
and the exertion of a massive influence on Japan, Korea, and possibly even
India. Hence the history of China offers the key to the history of all of East
Asia. This chapter will tell the story of how China did become Chinese.
A CONVENIENT STARTING point is a detailed linguistic map of China (see
Figure 16.1). A glance at it is an eye-opener to all of us accustomed to
thinking of China as monolithic. It turns out that, in addition to China’s eight
“big” languages—Mandarin and its seven close relatives (often referred to
collectively simply as “Chinese”), with between 11 million and 800 million
speakers each—China also has over 130 “little” languages, many of them
with just a few thousand speakers. All these languages, “big” and “little,” fall
into four language families, which differ greatly in the compactness of their
distributions.
At the one extreme, Mandarin and its relatives, which constitute the
Chinese subfamily of the Sino-Tibetan language family, are distributed
continuously from North to South China. One could walk through China,
from Manchuria in the north to the Gulf of Tonkin in the south, while
remaining entirely within land occupied by native speakers of Mandarin and
its relatives. The other three families have fragmented distributions, being
spoken by “islands” of people surrounded by a “sea” of speakers of Chinese
and other language families.
Especially fragmented is the distribution of the Miao-Yao (alias Hmong-
Mien) family, which consists of 6 million speakers divided among about five
languages, bearing the colorful names of Red Miao, White Miao (alias Striped
Miao), Black Miao, Green Miao (alias Blue Miao), and Yao. Miao-Yao
speakers live in dozens of small enclaves, all surrounded by speakers of other
language families and scattered over an area of half a million square miles,
extending from South China to Thailand. More than 100,000 Miao-speaking
refugees from Vietnam have carried this language family to the United States,
where they are better known under the alternative name of Hmong.
Another fragmented language group is the Austroasiatic family, whose
most widely spoken languages are Vietnamese and Cambodian. The 60
million Austroasiatic speakers are scattered from Vietnam in the east to the
Malay Peninsula in the south and to northern India in the west. The fourth and
last of China’s language families is the Tai-Kadai family (including Thai and
Lao), whose 50 million speakers are distributed from South China southward
into Peninsular Thailand and west to Myanmar (Figure 16.1).
Naturally, Miao-Yao speakers did not acquire their current fragmented
distribution as a result of ancient helicopter flights that dropped them here and
there over the Asian landscape. Instead, one might guess that they once had a
more nearly continuous distribution, which became fragmented as speakers of
other language families expanded or induced Miao-Yao speakers to abandon
their tongues. In fact, much of that process of linguistic fragmentation
occurred within the past 2,500 years and is well documented historically. The
ancestors of modern speakers of Thai, Lao, and Burmese all moved south
from South China and adjacent areas to their present locations within
historical times, successively inundating the settled descendants of previous
migrations. Speakers of Chinese languages were especially vigorous in
replacing and linguistically converting other ethnic groups, whom Chinese
speakers looked down upon as primitive and inferior. The recorded history of
China’s Zhou Dynasty, from 1100 to 221 B.C., describes the conquest and
absorption of most of China’s non-Chinese-speaking population by Chinese-
speaking states.
We can use several types of reasoning to try to reconstruct the linguistic
map of East Asia as of several thousand years ago. First, we can reverse the
historically known linguistic expansions of recent millennia. Second, we can
reason that modern areas with just a single language or related language
group occupying a large, continuous area testify to a recent geographic
expansion of that group, such that not enough historical time has elapsed for it
to differentiate into many languages. Finally, we can reason conversely that
modern areas with a high diversity of languages within a given language
family lie closer to the early center of distribution of that language family.
Using those three types of reasoning to turn back the linguistic clock, we
conclude that North China was originally occupied by speakers of Chinese
and other Sino-Tibetan languages; that different parts of South China were
variously occupied by speakers of Miao-Yao, Austroasiatic, and Tai-Kadai
languages; and that Sino-Tibetan speakers have replaced most speakers of
those other families over South China. An even more drastic linguistic
upheaval must have swept over tropical Southeast Asia to the south of China
—in Thailand, Myanmar, Laos, Cambodia, Vietnam, and Peninsular
Malaysia. Whatever languages were originally spoken there must now be
entirely extinct, because all of the modern languages of those countries appear
to be recent invaders, mainly from South China or, in a few cases, from
Indonesia. Since Miao-Yao languages barely survived into the present, we
might also guess that South China once harbored still other language families
besides Miao-Yao, Austroasiatic, and Tai-Kadai, but that those other families
left no modern surviving languages. As we shall see, the Austronesian
language family (to which all Philippine and Polynesian languages belong)
may have been one of those other families that vanished from the Chinese
mainland, and that we know only because it spread to Pacific islands and
survived there.
These language replacements in East Asia remind us of the spread of
European languages, especially English and Spanish, into the New World,
formerly home to a thousand or more Native American languages. We know
from our recent history that English did not come to replace U.S. Indian
languages merely because English sounded musical to Indians’ ears. Instead,
the replacement entailed English-speaking immigrants’ killing most Indians
by war, murder, and introduced diseases, and the surviving Indians’ being
pressured into adopting English, the new majority language. The immediate
causes of that language replacement were the advantages in technology and
political organization, stemming ultimately from the advantage of an early
rise of food production, that invading Europeans held over Native Americans.
Essentially the same processes accounted for the replacement of Aboriginal
Australian languages by English, and of subequatorial Africa’s original
Pygmy and Khoisan languages by Bantu languages.
Hence East Asia’s linguistic upheavals raise a corresponding question:
what enabled Sino-Tibetan speakers to spread from North China to South
China, and speakers of Austroasiatic and the other original South China
language families to spread south into tropical Southeast Asia? Here, we must
turn to archaeology for evidence of the technological, political, and
agricultural advantages that some Asians evidently gained over other Asians.
AS EVERYWHERE ELSE in the world, the archaeological record in East Asia for
most of human history reveals only the debris of hunter-gatherers using
unpolished stone tools and lacking pottery. The first East Asian evidence for
something different comes from China, where crop remains, bones of
domestic animals, pottery, and polished (Neolithic) stone tools appear by
around 7500 B.C. That date is within a thousand years of the beginning of the
Neolithic Age and food production in the Fertile Crescent. But because the
previous millennium in China is poorly known archaeologically, one cannot
decide at present whether the origins of Chinese food production were
contemporaneous with those in the Fertile Crescent, slightly earlier, or
slightly later. At the least, we can say that China was one of the world’s first
centers of plant and animal domestication.
China may actually have encompassed two or more independent centers
of origins of food production. I already mentioned the ecological differences
between China’s cool, dry north and warm, wet south. At a given latitude,
there are also ecological distinctions between the coastal lowlands and the
interior uplands. Different wild plants are native to these disparate
environments and would thus have been variously available to incipient
farmers in various parts of China. In fact, the earliest identified crops were
two drought-resistant species of millet in North China, but rice in South
China, suggesting the possibility of separate northern and southern centers of
plant domestication.
Chinese sites with the earliest evidence of crops also contained bones of
domestic pigs, dogs, and chickens. These domestic animals and crops were
gradually joined by China’s many other domesticates. Among the animals,
water buffalo were most important (for pulling plows), while silkworms,
ducks, and geese were others. Familiar later Chinese crops include soybeans,
hemp, citrus fruit, tea, apricots, peaches, and pears. In addition, just as
Eurasia’s east–west axis permitted many of these Chinese animals and crops
to spread westward in ancient times, West Asian domesticates also spread
eastward to China and became important there. Especially significant western
contributions to ancient China’s economy have been wheat and barley, cows
and horses, and (to a lesser extent) sheep and goats.
As elsewhere in the world, in China food production gradually led to the
other hallmarks of “civilization” discussed in Chapters 11–14. A superb
Chinese tradition of bronze metallurgy had its origins in the third millennium
B.C. and eventually resulted in China’s developing by far the earliest cast-iron
production in the world, around 500 B.C. The following 1,500 years saw the
outpouring of Chinese technological inventions, mentioned in Chapter 13,
that included paper, the compass, the wheelbarrow, and gunpowder. Fortified
towns emerged in the third millennium B.C., with cemeteries whose great
variation between unadorned and luxuriously furnished graves bespeaks
emerging class differences. Stratified societies whose rulers could mobilize
large labor forces of commoners are also attested by huge urban defensive
walls, big palaces, and eventually the Grand Canal (the world’s longest canal,
over 1,000 miles long), linking North and South China. Writing is preserved
from the second millennium B.C. but probably arose earlier. Our
archaeological knowledge of China’s emerging cities and states then becomes
supplemented by written accounts of China’s first dynasties, going back to the
Xia Dynasty, which arose around 2000 B.C.
As for food production’s more sinister by-product of infectious diseases,
we cannot specify where within the Old World most major diseases of Old
World origin arose. However, European writings from Roman and medieval
times clearly describe the arrival of bubonic plague and possibly smallpox
from the east, so these germs could be of Chinese or East Asian origin.
Influenza (derived from pigs) is even more likely to have arisen in China,
since pigs were domesticated so early and became so important there.
China’s size and ecological diversity spawned many separate local
cultures, distinguishable archaeologically by their differing styles of pottery
and artifacts. In the fourth millennium B.C. those local cultures expanded
geographically and began to interact, compete with each other, and coalesce.
Just as exchanges of domesticates between ecologically diverse regions
enriched Chinese food production, exchanges between culturally diverse
regions enriched Chinese culture and technology, and fierce competition
between warring chiefdoms drove the formation of ever larger and more
centralized states (Chapter 14).
While China’s east–west gradient retarded crop diffusion, the gradient
was less of a barrier there than in the Americas or Africa, because China’s
east–west distances were smaller; and because China’s is transected neither
by desert, as is Africa and northern Mexico, nor by a narrow isthmus, as is
Central America. Instead, China’s long east–west rivers (the Yellow River in
the north, the Yangtze River in the south) facilitated diffusion of crops and
technology between the coast and inland, while its broad east–west expanse
and relatively gentle terrain, which eventually permitted those two river
systems to be joined by canals, facilitated east–west exchanges. All these
geographic factors contributed to the early cultural and political unification of
China, whereas western Europe, with a similar area but a more rugged terrain
and no such unifying rivers, has resisted cultural and political unification to
this day.
Some developments spread from south to north in China, especially iron
smelting and rice cultivation. But the predominant direction of spread was
from north to south. That trend is clearest for writing: in contrast to western
Eurasia, which produced a plethora of early writing systems, such as
Sumerian cuneiform, Egyptian hieroglyphics, Hittite, Minoan, and the
Semitic alphabet, China developed just a single well-attested writing system.
It was perfected in North China, spread and preempted or replaced any other
nascent system, and evolved into the writing still used in China today. Other
major features of North Chinese societies that spread southward were bronze
technology, Sino-Tibetan languages, and state formation. All three of China’s
first three dynasties, the Xia and Shang and Zhou Dynasties, arose in North
China in the second millennium B.C.
Preserved writings of the first millennium B.C. show that ethnic Chinese
already tended then (as many still do today) to feel culturally superior to non-
Chinese “barbarians,” while North Chinese tended to regard even South
Chinese as barbarians. For example, a late Zhou Dynasty writer of the first
millennium B.C. described China’s other peoples as follows: “The people of
those five regions—the Middle states and the Rong, Yi, and other wild tribes
around them—had all their several natures, which they could not be made to
alter. The tribes on the east were called Yi. They had their hair unbound, and
tattooed their bodies. Some of them ate their food without its being cooked by
fire.” The Zhou author went on to describe wild tribes to the south, west, and
north as indulging in equally barbaric practices, such as turning their feet
inward, tattooing their foreheads, wearing skins, living in caves, not eating
cereals, and, of course, eating their food raw.
States organized by or modeled on that Zhou Dynasty of North China
spread to South China during the first millennium B.C., culminating in China’s
political unification under the Qin Dynasty in 221 B.C. Its cultural unification
accelerated during that same period, as literate “civilized” Chinese states
absorbed, or were copied by, the illiterate “barbarians.” Some of that cultural
unification was ferocious: for instance, the first Qin emperor condemned all
previously written historical books as worthless and ordered them burned,
much to the detriment of our understanding of early Chinese history and
writing. Those and other draconian measures must have contributed to the
spread of North China’s Sino-Tibetan languages over most of China, and to
reducing the Miao-Yao and other language families to their present
fragmented distributions.
Within East Asia, China’s head start in food production, technology,
writing, and state formation had the consequence that Chinese innovations
also contributed heavily to developments in neighboring regions. For
instance, until the fourth millennium B.C. most of tropical Southeast Asia was
still occupied by hunter-gatherers making pebble and flake stone tools
belonging to what is termed the Hoabinhian tradition, named after the site of
Hoa Binh, in Vietnam. Thereafter, Chinese-derived crops, Neolithic
technology, village living, and pottery similar to that of South China spread
into tropical Southeast Asia, probably accompanied by South China’s
language families. The historical southward expansions of Burmese, Laotians,
and Thais from South China completed the Sinification of tropical Southeast
Asia. All those modern peoples are recent offshoots of their South Chinese
cousins.
So overwhelming was this Chinese steamroller that the former peoples of
tropical Southeast Asia have left behind few traces in the region’s modern
populations. Just three relict groups of hunter-gatherers—the Semang
Negritos of the Malay Peninsula, the Andaman Islanders, and the Veddoid
Negritos of Sri Lanka—remain to suggest that tropical Southeast Asia’s
former inhabitants may have been dark-skinned and curly-haired, like modern
New Guineans and unlike the light-skinned, straight-haired South Chinese
and the modern tropical Southeast Asians who are their offshoots. Those
relict Negritos of Southeast Asia may be the last survivors of the source
population from which New Guinea was colonized. The Semang Negritos
persisted as hunter-gatherers trading with neighboring farmers but adopted an
Austroasiatic language from those farmers—much as, we shall see, Philippine
Negrito and African Pygmy hunter-gatherers adopted languages from their
farmer trading partners. Only on the remote Andaman Islands do languages
unrelated to the South Chinese language families persist—the last linguistic
survivors of what must have been hundreds of now extinct aboriginal
Southeast Asian languages.
Even Korea and Japan were heavily influenced by China, although their
geographic isolation from it ensured that they did not lose their languages or
physical and genetic distinctness, as did tropical Southeast Asia. Korea and
Japan adopted rice from China in the second millennium B.C., bronze
metallurgy by the first millennium B.C., and writing in the first millennium
A.D. China also transmitted West Asian wheat and barley to Korea and Japan.
In thus describing China’s seminal role in East Asian civilization, we
should not exaggerate. It is not the case that all cultural advances in East Asia
stemmed from China and that Koreans, Japanese, and tropical Southeast
Asians were noninventive barbarians who contributed nothing. The ancient
Japanese developed some of the oldest pottery in the world and settled as
hunter-gatherers in villages subsisting on Japan’s rich seafood resources, long
before the arrival of food production. Some crops were probably domesticated
first or independently in Japan, Korea, and tropical Southeast Asia.
But China’s role was nonetheless disproportionate. For example, the
prestige value of Chinese culture is still so great in Japan and Korea that
Japan has no thought of discarding its Chinese-derived writing system despite
its drawbacks for representing Japanese speech, while Korea is only now
replacing its clumsy Chinese-derived writing with its wonderful indigenous
han’gul alphabet. That persistence of Chinese writing in Japan and Korea is a
vivid 20th-century legacy of plant and animal domestication in China nearly
10,000 years ago. Thanks to the achievements of East Asia’s first farmers,
China became Chinese, and peoples from Thailand to (as we shall see in the
next chapter) Easter Island became their cousins.
CHAPTER 17
SPEEDBOAT TO POLYNESIA
PACIFIC ISLAND HISTORY IS ENCAPSULATED FOR ME IN AN incident that happened
when three Indonesian friends and I walked into a store in Jayapura, the
capital of Indonesian New Guinea. My friends’ names were Achmad, Wiwor,
and Sauakari, and the store was run by a merchant named Ping Wah. Achmad,
an Indonesian government officer, was acting as the boss, because he and I
were organizing an ecological survey for the government and had hired
Wiwor and Sauakari as local assistants. But Achmad had never before been in
a New Guinea mountain forest and had no idea what supplies to buy. The
results were comical.
At the moment that my friends entered the store, Ping Wah was reading a
Chinese newspaper. When he saw Wiwor and Sauakari, he kept reading it but
then shoved it out of sight under the counter as soon as he noticed Achmad.
Achmad picked up an ax head, causing Wiwor and Sauakari to laugh, because
he was holding it upside down. Wiwor and Sauakari showed him how to hold
it correctly and to test it. Achmad and Sauakari then looked at Wiwor’s bare
feet, with toes splayed wide from a lifetime of not wearing shoes. Sauakari
picked out the widest available shoes and held them against Wiwor’s feet, but
the shoes were still too narrow, sending Achmad and Sauakari and Ping Wah
into peals of laughter. Achmad picked up a plastic comb with which to comb
out his straight, coarse black hair. Glancing at Wiwor’s tough, tightly coiled
hair, he handed the comb to Wiwor. It immediately stuck in Wiwor’s hair,
then broke as soon as Wiwor pulled on the comb. Everyone laughed,
including Wiwor. Wiwor responded by reminding Achmad that he should buy
lots of rice, because there would be no food to buy in New Guinea mountain
villages except sweet potatoes, which would upset Achmad’s stomach—more
hilarity.
Despite all the laughter, I could sense the underlying tensions. Achmad
was Javan, Ping Wah Chinese, Wiwor a New Guinea highlander, and Sauakari
a New Guinea lowlander from the north coast. Javans dominate the
Indonesian government, which annexed western New Guinea in the 1960s
and used bombs and machine guns to crush New Guinean opposition.
Achmad later decided to stay in town and to let me do the forest survey alone
with Wiwor and Sauakari. He explained his decision to me by pointing to his
straight, coarse hair, so unlike that of New Guineans, and saying that New
Guineans would kill anyone with hair like his if they found him far from army
backup.
Ping Wah had put away his newspaper because importation of Chinese
writing is nominally illegal in Indonesian New Guinea. In much of Indonesia
the merchants are Chinese immigrants. Latent mutual fear between the
economically dominant Chinese and politically dominant Javans erupted in
1966 in a bloody revolution, when Javans slaughtered hundreds of thousands
of Chinese. As New Guineans, Wiwor and Sauakari shared most New
Guineans’ resentment of Javan dictatorship, but they also scorned each other’s
groups. Highlanders dismiss lowlanders as effete sago eaters, while
lowlanders dismiss highlanders as primitive big-heads, referring both to their
massive coiled hair and to their reputation for arrogance. Within a few days of
my setting up an isolated forest camp with Wiwor and Sauakari, they came
close to fighting each other with axes.
Tensions among the groups that Achmad, Wiwor, Sauakari, and Ping Wah
represent dominate the politics of Indonesia, the world’s fourth-most-
populous nation. These modern tensions have roots going back thousands of
years. When we think of major overseas population movements, we tend to
focus on those since Columbus’s discovery of the Americas, and on the
resulting replacements of non-Europeans by Europeans within historic times.
But there were also big overseas movements long before Columbus, and
prehistoric replacements of non-European peoples by other non-European
peoples. Wiwor, Achmad, and Sauakari represent three prehistorical waves of
people that moved overseas from the Asian mainland into the Pacific.
Wiwor’s highlanders are probably descended from an early wave that had
colonized New Guinea from Asia by 40,000 years ago. Achmad’s ancestors
arrived in Java ultimately from the South China coast, around 4,000 years
ago, completing the replacement there of people related to Wiwor’s ancestors.
Sauakari’s ancestors reached New Guinea around 3,600 years ago, as part of
that same wave from the South China coast, while Ping Wah’s ancestors still
occupy China.
The population movement that brought Achmad’s and Sauakari’s
ancestors to Java and New Guinea, respectively, termed the Austronesian
expansion, was among the biggest population movements of the last 6,000
years. One prong of it became the Polynesians, who populated the most
remote islands of the Pacific and were the greatest seafarers among Neolithic
peoples. Austronesian languages are spoken today as native languages over
more than half of the globe’s span, from Madagascar to Easter Island. In this
book on human population movements since the end of the Ice Ages, the
Austronesian expansion occupies a central place, as one of the most important
phenomena to be explained. Why did Austronesian people, stemming
ultimately from mainland China, colonize Java and the rest of Indonesia and
replace the original inhabitants there, instead of Indonesians colonizing China
and replacing the Chinese? Having occupied all of Indonesia, why were the
Austronesians then unable to occupy more than a narrow coastal strip of the
New Guinea lowlands, and why were they completely unable to displace
Wiwor’s people from the New Guinea highlands? How did the descendants of
Chinese emigrants become transformed into Polynesians?
TODAY, THE POPULATION of Java, most other Indonesian islands (except the
easternmost ones), and the Philippines is rather homogeneous. In appearance
and genes those islands’ inhabitants are similar to South Chinese, and even
more similar to tropical Southeast Asians, especially those of the Malay
Peninsula. Their languages are equally homogeneous: while 374 languages
are spoken in the Philippines and western and central Indonesia, all of them
are closely related and fall within the same sub-subfamily (Western Malayo-
Polynesian) of the Austronesian language family. Austronesian languages
reached the Asian mainland on the Malay Peninsula and in small pockets in
Vietnam and Cambodia, near the westernmost Indonesian islands of Sumatra
and Borneo, but they occur nowhere else on the mainland (Figure 17.1). Some
Austronesian words borrowed into English include “taboo” and “tattoo” (from
a Polynesian language), “boondocks” (from the Tagalog language of the
Philippines), and “amok,” “batik,” and “orangutan” (from Malay).
That genetic and linguistic uniformity of Indonesia and the Philippines is
initially as surprising as is the predominant linguistic uniformity of China.
The famous Java Homo erectus fossils prove that humans have occupied at
least western Indonesia for a million years. That should have given ample
time for humans to evolve genetic and linguistic diversity and tropical
adaptations, such as dark skins like those of many other tropical peoples—but
instead Indonesians and Filipinos have light skins.
It is also surprising that Indonesians and Filipinos are so similar to
tropical Southeast Asians and South Chinese in other physical features
besides light skins and in their genes. A glance at a map makes it obvious that
Indonesia offered the only possible route by which humans could have
reached New Guinea and Australia 40,000 years ago, so one might naively
have expected modern Indonesians to be like modern New Guineans and
Australians. In reality, there are only a few New Guinean–like populations in
the Philippine / western Indonesia area, notably the Negritos living in
mountainous areas of the Philippines. As is also true of the three New
Guinean-like relict populations that I mentioned in speaking of tropical
Southeast Asia (Chapter 16), the Philippine Negritos could be relicts of
populations ancestral to Wiwor’s people before they reached New Guinea.
Even those Negritos speak Austronesian languages similar to those of their
Filipino neighbors, implying that they too (like Malaysia’s Semang Negritos
and Africa’s Pygmies) have lost their original language.
All these facts suggest strongly that either tropical Southeast Asians or
South Chinese speaking Austronesian languages recently spread through the
Philippines and Indonesia, replacing all the former inhabitants of those islands
except the Philippine Negritos, and replacing all the original island languages.
That event evidently took place too recently for the colonists to evolve dark
skins, distinct language families, or genetic distinctiveness or diversity. Their
languages are of course much more numerous than the eight dominant
Chinese languages of mainland China, but are no more diverse. The
proliferation of many similar languages in the Philippines and Indonesia
merely reflects the fact that the islands never underwent a political and
cultural unification, as did China.
Details of language distributions provide valuable clues to the route of
this hypothesized Austronesian expansion. The whole Austronesian language
family consists of 959 languages, divided among four subfamilies. But one of
those subfamilies, termed Malayo-Polynesian, comprises 945 of those 959
languages and covers almost the entire geographic range of the Austronesian
family. Before the recent overseas expansion of Europeans speaking Indo-
European languages, Austronesian was the most widespread language family
in the world. That suggests that the Malayo-Polynesian subfamily
differentiated recently out of the Austronesian family and spread far from the
Austronesian homeland, giving rise to many local languages, all of which are
still closely related because there has been too little time to develop large
linguistic differences. For the location of that Austronesian homeland, we
should therefore look not to MalayoPolynesian but to the other three
Austronesian subfamilies, which differ considerably more from each other
and from Malayo-Polynesian than the sub-subfamilies of Malayo-Polynesian
differ among each other.
It turns out that those three other subfamilies have coincident
distributions, all of them tiny compared with the distribution of Malayo-
Polynesian. They are confined to aborigines of the island of Taiwan, lying
only 90 miles from the South China mainland. Taiwan’s aborigines had the
island largely to themselves until mainland Chinese began settling in large
numbers within the last thousand years. Still more mainlanders arrived after
1945, especially after the Chinese Communists defeated the Chinese
Nationalists in 1949, so that aborigines now constitute only 2 percent of
Taiwan’s population. The concentration of three out of the four Austronesian
subfamilies on Taiwan suggests that, within the present Austronesian realm,
Taiwan is the homeland where Austronesian languages have been spoken for
the most millennia and have consequently had the longest time in which to
diverge. All other Austronesian languages, from those on Madagascar to
those on Easter Island, would then stem from a population expansion out of
Taiwan.
WE CAN NOW turn to archaeological evidence. While the debris of ancient
village sites does not include fossilized words along with bones and pottery, it
does reveal movements of people and cultural artifacts that could be
associated with languages. Like the rest of the world, most of the present
Austronesian realm—Taiwan, the Philippines, Indonesia, and many Pacific
islands—was originally occupied by hunter-gatherers lacking pottery,
polished stone tools, domestic animals, and crops. (The sole exceptions to this
generalization are the remote islands of Madagascar, eastern Melanesia,
Polynesia, and Micronesia, which were never reached by hunter-gatherers and
remained empty of humans until the Austronesian expansion.) The first
archaeological signs of something different within the Austronesian realm
come from—Taiwan. Beginning around the fourth millennium B.C., polished
stone tools and a distinctive decorated pottery style (so-called Ta-p’en-k’eng
pottery) derived from earlier South China mainland pottery appeared on
Taiwan and on the opposite coast of the South China mainland. Remains of
rice and millet at later Taiwanese sites provide evidence of agriculture.
Ta-p’en-k’eng sites of Taiwan and the South China coast are full of fish
bones and mollusk shells, as well as of stone net sinkers and adzes suitable for
hollowing out a wooden canoe. Evidently, those first Neolithic occupants of
Taiwan had watercraft adequate for deep-sea fishing and for regular sea traffic
across Taiwan Strait, separating that island from the China coast. Thus,
Taiwan Strait may have served as the training ground where mainland
Chinese developed the open-water maritime skills that would permit them to
expand over the Pacific.
One specific type of artifact linking Taiwan’s Ta-p’en-k’eng culture to
later Pacific island cultures is a bark beater, a stone implement used for
pounding the fibrous bark of certain tree species into rope, nets, and clothing.
Once Pacific peoples spread beyond the range of wool-yielding domestic
animals and fiber plant crops and hence of woven clothing, they became
dependent on pounded bark “cloth” for their clothing. Inhabitants of Rennell
Island, a traditional Polynesian island that did not become Westernized until
the 1930s, told me that Westernization yielded the wonderful side benefit that
the island became quiet. No more sounds of bark beaters everywhere,
pounding out bark cloth from dawn until after dusk every day!
Within a millennium or so after the Ta-p’en-k’eng culture reached
Taiwan, archaeological evidence shows that cultures obviously derived from
it spread farther and farther from Taiwan to fill up the modern Austronesian
realm (Figure 17.2). The evidence includes ground stone tools, pottery, bones
of domestic pigs, and crop remains. For example, the decorated Ta-p’en-k’eng
pottery on Taiwan gave way to undecorated plain or red pottery, which has
also been found at sites in the Philippines and on the Indonesian islands of
Celebes and Timor. This cultural “package” of pottery, stone tools, and
domesticates appeared around 3000 B.C. in the Philippines, around 2500 B.C.
on the Indonesian islands of Celebes and North Borneo and Timor, around
2000 B.C. on Java and Sumatra, and around 1600 B.C. in the New Guinea
region. There, as we shall see, the expansion assumed a speedboat pace, as
bearers of the cultural package raced eastward into the previously uninhabited
Pacific Ocean beyond the Solomon Archipelago. The last phases of the
expansion, during the millennium after A.D. 1, resulted in the colonization of
every Polynesian and Micronesian island capable of supporting humans.
Astonishingly, it also swept westward across the Indian Ocean to the east
coast of Africa, resulting in the colonization of the island of Madagascar.
At least until the expansion reached coastal New Guinea, travel between
islands was probably by double-outrigger sailing canoes, which are still
widespread throughout Indonesia today. That boat design represented a major
advance over the simple dugout canoes prevalent among traditional peoples
living on inland waterways throughout the world. A dugout canoe is just what
its name implies: a solid tree trunk “dug out” (that is, hollowed out), and its
ends shaped, by an adze. Since the canoe is as round-bottomed as the trunk
from which it was carved, the least imbalance in weight distribution tips the
canoe toward the overweighted side. Whenever I’ve been paddled in dugouts
up New Guinea rivers by New Guineans, I have spent much of the trip in
terror: it seemed that every slight movement of mine risked capsizing the
canoe and spilling out me and my binoculars to commune with crocodiles.
New Guineans manage to look secure while paddling dugouts on calm lakes
and rivers, but not even New Guineans can use a dugout in seas with modest
waves. Hence some stabilizing device must have been essential not only for
the Austronesian expansion through Indonesia but even for the initial
colonization of Taiwan.
The solution was to lash two smaller logs (“outriggers”) parallel to the
hull and several feet from it, one on each side, connected to the hull by poles
lashed perpendicular to the hull and outriggers. Whenever the hull starts to tip
toward one side, the buoyancy of the outrigger on that side prevents the
outrigger from being pushed under the water and hence makes it virtually
impossible to capsize the vessel. The invention of the double-outrigger sailing
canoe may have been the technological breakthrough that triggered the
Austronesian expansion from the Chinese mainland.
TWO STRIKING COINCIDENCES between archaeological and linguistic evidence
support the inference that the people bringing a Neolithic culture to Taiwan,
the Philippines, and Indonesia thousands of years ago spoke Austronesian
languages and were ancestral to the Austronesian speakers still inhabiting
those islands today. First, both types of evidence point unequivocally to the
colonization of Taiwan as the first stage of the expansion from the South
China coast, and to the colonization of the Philippines and Indonesia from
Taiwan as the next stage. If the expansion had proceeded from tropical
Southeast Asia’s Malay Peninsula to the nearest Indonesian island of Sumatra,
then to other Indonesian islands, and finally to the Philippines and Taiwan, we
would find the deepest divisions (reflecting the greatest time depth) of the
Austronesian language family among the modern languages of the Malay
Peninsula and Sumatra, and the languages of Taiwan and the Philippines
would have differentiated only recently within a single subfamily. Instead, the
deepest divisions are in Taiwan, and the languages of the Malay Peninsula
and Sumatra fall together in the same sub-sub-subfamily: a recent branch of
the Western Malayo-Polynesian sub-subfamily, which is in turn a fairly recent
branch of the Malayo-Polynesian subfamily. Those details of linguistic
relationships agree perfectly with the archaeological evidence that the
colonization of the Malay Peninsula was recent, and followed rather than
preceded the colonization of Taiwan, the Philippines, and Indonesia.
The other coincidence between archaeological and linguistic evidence
concerns the cultural baggage that ancient Austronesians used. Archaeology
provides us with direct evidence of culture in the form of pottery, pig and fish
bones, and so on. One might initially wonder how a linguist, studying only
modern languages whose unwritten ancestral forms remain unknown, could
ever figure out whether Austronesians living on Taiwan 6,000 years ago had
pigs. The solution is to reconstruct the vocabularies of vanished ancient
languages (so-called protolanguages) by comparing vocabularies of modern
languages derived from them.
For instance, the words meaning “sheep” in many languages of the Indo-
European language family, distributed from Ireland to India, are quite similar:
“avis,” “avis,” “ovis,” “oveja,” “ovtsa,” “owis,” and “oi” in Lithuanian,
Sanskrit, Latin, Spanish, Russian, Greek, and Irish, respectively. (The English
“sheep” is obviously from a different root, but English retains the original
root in the word “ewe.”) Comparison of the sound shifts that the various
modern Indo-European languages have undergone during their histories
suggests that the original form was “owis” in the ancestral Indo-European
language spoken around 6,000 years ago. That unwritten ancestral language is
termed Proto-Indo-European.
Evidently, Proto-Indo-Europeans 6,000 years ago had sheep, in agreement
with archaeological evidence. Nearly 2,000 other words of their vocabulary
can similarly be reconstructed, including words for “goat,” “horse,” “wheel,”
“brother,” and “eye.” But no Proto-Indo-European word can be reconstructed
for “gun,” which uses different roots in different modern Indo-European
languages: “gun” in English, “fusil” in French, “ruzhyo” in Russian, and so
on. That shouldn’t surprise us: people 6,000 years ago couldn’t possibly have
had a word for guns, which were invented only within the past 1,000 years.
Since there was thus no inherited shared root meaning “gun,” each Indo-
European language had to invent or borrow its own word when guns were
finally invented.
Proceeding in the same way, we can compare modern Taiwanese,
Philippine, Indonesian, and Polynesian languages to reconstruct a Proto-
Austronesian language spoken in the distant past. To no one’s surprise, that
reconstructed Proto-Austronesian language had words with meanings such as
“two,” “bird,” “ear,” and “head louse”: of course, Proto-Austronesians could
count to 2, knew of birds, and had ears and lice. More interestingly, the
reconstructed language had words for “pig,” “dog,” and “rice,” which must
therefore have been part of Proto-Austronesian culture. The reconstructed
language is full of words indicating a maritime economy, such as “outrigger
canoe,” “sail,” “giant clam,” “octopus,” “fish trap,” and “sea turtle.” This
linguistic evidence regarding the culture of Proto-Austronesians, wherever
and whenever they lived, agrees well with the archaeological evidence
regarding the pottery-making, sea-oriented, food-producing people living on
Taiwan around 6,000 years ago.
The same procedure can be applied to reconstruct Proto-Malayo-
Polynesian, the ancestral language spoken by Austronesians after emigrating
from Taiwan. Proto-Malayo-Polynesian contains words for many tropical
crops like taro, breadfruit, bananas, yams, and coconuts, for which no word
can be reconstructed in Proto-Austronesian. Thus, the linguistic evidence
suggests that many tropical crops were added to the Austronesian repertoire
after the emigration from Taiwan. This conclusion agrees with archaeological
evidence: as colonizing farmers spread southward from Taiwan (lying about
23 degrees north of the equator) toward the equatorial tropics, they came to
depend increasingly on tropical root and tree crops, which they proceeded to
carry with them out into the tropical Pacific.
How could those Austronesian-speaking farmers from South China via
Taiwan replace the original hunter-gatherer population of the Philippines and
western Indonesia so completely that little genetic and no linguistic evidence
of that original population survived? The reasons resemble the reasons why
Europeans replaced or exterminated Native Australians within the last two
centuries, and why South Chinese replaced the original tropical Southeast
Asians earlier: the farmers’ much denser populations, superior tools and
weapons, more developed watercraft and maritime skills, and epidemic
diseases to which the farmers but not the hunter-gatherers had some
resistance. On the Asian mainland Austronesian-speaking farmers were able
similarly to replace some of the former hunter-gatherers of the Malay
Peninsula, because Austronesians colonized the peninsula from the south and
east (from the Indonesian islands of Sumatra and Borneo) around the same
time that Austroasiatic-speaking farmers were colonizing the peninsula from
the north (from Thailand). Other Austronesians managed to establish
themselves in parts of southern Vietnam and Cambodia to become the
ancestors of the modern Chamic minority of those countries.
However, Austronesian farmers could spread no farther into the Southeast
Asian mainland, because Austroasiatic and Tai-Kadai farmers had already
replaced the former hunter-gatherers there, and because Austronesian farmers
had no advantage over Austroasiatic and Tai-Kadai farmers. Although we
infer that Austronesian speakers originated from coastal South China,
Austronesian languages today are not spoken anywhere in mainland China,
possibly because they were among the hundreds of former Chinese languages
eliminated by the southward expansion of Sino-Tibetan speakers. But the
language families closest to Austronesian are thought to be Tai-Kadai,
Austroasiatic, and Miao-Yao. Thus, while Austronesian languages in China
may not have survived the onslaught of Chinese dynasties, some of their sister
and cousin languages did.
WE HAVE NOW followed the initial stages of the Austronesian expansion for
2,500 miles from the South China coast, through Taiwan and the Philippines,
to western and central Indonesia. In the course of that expansion,
Austronesians came to occupy all habitable areas of those islands, from the
seacoast to the interior, and from the lowlands to the mountains. By 1500 B.C.
their familiar archaeological hallmarks, including pig bones and plain red-
slipped pottery, show that they had reached the eastern Indonesian island of
Halmahera, less than 200 miles from the western end of the big mountainous
island of New Guinea. Did they proceed to overrun that island, just as they
had already overrun the big mountainous islands of Celebes, Borneo, Java,
and Sumatra?
They did not, as a glance at the faces of most modern New Guineans
makes obvious, and as detailed studies of New Guinean genes confirm. My
friend Wiwor and all other New Guinea highlanders differ obviously from
Indonesians, Filipinos, and South Chinese in their dark skins, tightly coiled
hair, and face shapes. Most lowlanders from New Guinea’s interior and south
coast resemble the highlanders except that they tend to be taller. Geneticists
have failed to find characteristic Austronesian gene markers in blood samples
from New Guinea highlanders.
But peoples of New Guinea’s north and east coasts, and of the Bismarck
and Solomon Archipelagoes north and east of New Guinea, present a more
complex picture. In appearance, they are variably intermediate between
highlanders like Wiwor and Indonesians like Achmad, though on the average
considerably closer to Wiwor. For instance, my friend Sauakari from the north
coast has wavy hair intermediate between Achmad’s straight hair and
Wiwor’s coiled hair, and skin somewhat paler than Wiwor’s, though
considerably darker than Achmad’s. Genetically, the Bismarck and Solomon
islanders and north coastal New Guineans are about 15 percent Austronesian
and 85 percent like New Guinea highlanders. Hence Austronesians evidently
reached the New Guinea region but failed completely to penetrate the island’s
interior and were genetically diluted by New Guinea’s previous residents on
the north coast and islands.
Modern languages tell essentially the same story but add detail. In
Chapter 15 I explained that most New Guinea languages, termed Papuan
languages, are unrelated to any language families elsewhere in the world.
Without exception, every language spoken in the New Guinea mountains, the
whole of southwestern and south-central lowland New Guinea, including the
coast, and the interior of northern New Guinea is a Papuan language. But
Austronesian languages are spoken in a narrow strip immediately on the north
and southeast coasts. Most languages of the Bismarck and Solomon islands
are Austronesian: Papuan languages are spoken only in isolated pockets on a
few islands.
Austronesian languages spoken in the Bismarcks and Solomons and north
coastal New Guinea are related, as a separate sub-sub-subfamily termed
Oceanic, to the sub-sub-subfamily of languages spoken on Halmahera and the
west end of New Guinea. That linguistic relationship confirms, as one would
expect from a map, that Austronesian speakers of the New Guinea region
arrived by way of Halmahera. Details of Austronesian and Papuan languages
and their distributions in North New Guinea testify to long contact between
the Austronesian invaders and the Papuan-speaking residents. Both the
Austronesian and the Papuan languages of the region show massive
influences of each other’s vocabularies and grammars, making it difficult to
decide whether certain languages are basically Austronesian languages
influenced by Papuan ones or the reverse. As one travels from village to
village along the north coast or its fringing islands, one passes from a village
with an Austronesian language to a village with a Papuan language and then
to another Austronesian-speaking village, without any genetic discontinuity at
the linguistic boundaries.
All this suggests that descendants of Austronesian invaders and of
original New Guineans have been trading, intermarrying, and acquiring each
other’s genes and languages for several thousand years on the North New
Guinea coast and its islands. That long contact transferred Austronesian
languages more effectively than Austronesian genes, with the result that most
Bismarck and Solomon islanders now speak Austronesian languages, even
though their appearance and most of their genes are still Papuan. But neither
the genes nor the languages of the Austronesians penetrated New Guinea’s
interior. The outcome of their invasion of New Guinea was thus very different
from the outcome of their invasion of Borneo, Celebes, and other big
Indonesian islands, where their steamroller eliminated almost all traces of the
previous inhabitants’ genes and languages. To understand what happened in
New Guinea, let us now turn to the evidence from archaeology.
AROUND 1600 B.C., almost simultaneously with their appearance on
Halmahera, the familiar archaeological hallmarks of the Austronesian
expansion—pigs, chickens, dogs, red-slipped pottery, and adzes of ground
stone and of giant clamshells—appear in the New Guinea region. But two
features distinguish the Austronesians’ arrival there from their earlier arrival
in the Philippines and Indonesia.
The first feature consists of pottery designs, which are aesthetic features
of no economic significance but which do let archaeologists immediately
recognize an early Austronesian site. Whereas most early Austronesian
pottery in the Philippines and Indonesia was undecorated, pottery in the New
Guinea region was finely decorated with geometric designs arranged in
horizontal bands. In other respects the pottery preserved the red slip and the
vessel forms characteristic of earlier Austronesian pottery in Indonesia.
Evidently, Austronesian settlers in the New Guinea region got the idea of
“tattooing” their pots, perhaps inspired by geometric designs that they had
already been using on their bark cloth and body tattoos. This style is termed
Lapita pottery, after an archaeological site named Lapita, where it was
described.
The much more significant distinguishing feature of early Austronesian
sites in the New Guinea region is their distribution. In contrast to those in the
Philippines and Indonesia, where even the earliest known Austronesian sites
are on big islands like Luzon and Borneo and Celebes, sites with Lapita
pottery in the New Guinea region are virtually confined to small islets
fringing remote larger islands. To date, Lapita pottery has been found at only
one site (Aitape) on the north coast of New Guinea itself, and at a couple of
sites in the Solomons. Most Lapita sites of the New Guinea region are in the
Bismarcks, on islets off the coast of the larger Bismarck islands, occasionally
on the coasts of the larger islands themselves. Since (as we shall see) the
makers of Lapita pottery were capable of sailing thousands of miles, their
failure to transfer their villages a few miles to the large Bismarck islands, or a
few dozen miles to New Guinea, was certainly not due to inability to get
there.
The basis of Lapita subsistence can be reconstructed from the garbage
excavated by archaeologists at Lapita sites. Lapita people depended heavily
on seafood, including fish, porpoises, sea turtles, sharks, and shellfish. They
had pigs, chickens, and dogs and ate the nuts of many trees (including
coconuts). While they probably also ate the usual Austronesian root crops,
such as taro and yams, evidence of those crops is hard to obtain, because hard
nut shells are much more likely than soft roots to persist for thousands of
years in garbage heaps.
Naturally, it is impossible to prove directly that the people who made
Lapita pots spoke an Austronesian language. However, two facts make this
inference virtually certain. First, except for the decorations on the pots, the
pots themselves and their associated cultural paraphernalia are similar to the
cultural remains found at Indonesian and Philippine sites ancestral to modern
Austronesian-speaking societies. Second, Lapita pottery also appears on
remote Pacific islands with no previous human inhabitants, with no evidence
of a major second wave of settlement subsequent to that bringing Lapita pots,
and where the modern inhabitants speak an Austronesian language (more of
this below). Hence Lapita pottery may be safely assumed to mark
Austronesians’ arrival in the New Guinea region.
What were those Austronesian pot makers doing on islets adjacent to
bigger islands? They were probably living in the same way as modern pot
makers lived until recently on islets in the New Guinea region. In 1972 I
visited such a village on Malai Islet, in the Siassi island group, off the
medium-sized island of Umboi, off the larger Bismarck island of New Britain.
When I stepped ashore on Malai in search of birds, knowing nothing about
the people there, I was astonished by the sight that greeted me. Instead of the
usual small village of low huts, surrounded by large gardens sufficient to feed
the village, and with a few canoes drawn up on the beach, most of the area of
Malai was occupied by two-story wooden houses side by side, leaving no
ground available for gardens—the New Guinea equivalent of downtown
Manhattan. On the beach were rows of big canoes. It turned out that Malai
islanders, besides being fishermen, were also specialized potters, carvers, and
traders, who lived by making beautifully decorated pots and wooden bowls,
transporting them in their canoes to larger islands and exchanging their wares
for pigs, dogs, vegetables, and other necessities. Even the timber for Malai
canoes was obtained by trade from villagers on nearby Umboi Island, since
Malai does not have trees big enough to be fashioned into canoes.
In the days before European shipping, trade between islands in the New
Guinea region was monopolized by such specialized groups of canoe-building
potters, skilled in sailing without navigational instruments, and living on
offshore islets or occasionally in mainland coastal villages. By the time I
reached Malai in 1972, those indigenous trade networks had collapsed or
contracted, partly because of competition from European motor vessels and
aluminum pots, partly because the Australian colonial government forbade
long-distance canoe voyaging after some accidents in which traders were
drowned. I would guess that the Lapita potters were the inter-island traders of
the New Guinea region in the centuries after 1600 B.C.
The spread of Austronesian languages to the north coast of New Guinea
itself, and over even the largest Bismarck and Solomon islands, must have
occurred mostly after Lapita times, since Lapita sites themselves were
concentrated on Bismarck islets. Not until around A.D. 1 did pottery derived
from the Lapita style appear on the south side of New Guinea’s southeast
peninsula. When Europeans began exploring New Guinea in the late 19th
century, all the remainder of New Guinea’s south coast still supported
populations only of Papuan-language speakers, even though Austronesian-
speaking populations were established not only on the southeastern peninsula
but also on the Aru and Kei Islands (lying 70–80 miles off western New
Guinea’s south coast). Austronesians thus had thousands of years in which to
colonize New Guinea’s interior and its southern coast from nearby bases, but
they never did so. Even their colonization of North New Guinea’s coastal
fringe was more linguistic than genetic: all northern coastal peoples remained
predominantly New Guineans in their genes. At most, some of them merely
adopted Austronesian languages, possibly in order to communicate with the
long-distance traders who linked societies.
THUS, THE OUTCOME of the Austronesian expansion in the New Guinea
region was opposite to that in Indonesia and the Philippines. In the latter
region the indigenous population disappeared—presumably driven off, killed,
infected, or assimilated by the invaders. In the former region the indigenous
population mostly kept the invaders out. The invaders (the Austronesians)
were the same in both cases, and the indigenous populations may also have
been genetically similar to each other, if the original Indonesian population
supplanted by Austronesians really was related to New Guineans, as I
suggested earlier. Why the opposite outcomes?
The answer becomes obvious when one considers the differing cultural
circumstances of Indonesia’s and New Guinea’s indigenous populations.
Before Austronesians arrived, most of Indonesia was thinly occupied by
hunter-gatherers lacking even polished stone tools. In contrast, food
production had already been established for thousands of years in the New
Guinea highlands, and probably in the New Guinea lowlands and in the
Bismarcks and Solomons as well. The New Guinea highlands supported some
of the densest populations of Stone Age people anywhere in the modern
world.
Austronesians enjoyed few advantages in competing with those
established New Guinean populations. Some of the crops on which
Austronesians subsisted, such as taro, yams, and bananas, had probably
already been independently domesticated in New Guinea before
Austronesians arrived. The New Guineans readily integrated Austronesian
chickens, dogs, and especially pigs into their food-producing economies. New
Guineans already had polished stone tools. They were at least as resistant to
tropical diseases as were Austronesians, because they carried the same five
types of genetic protections against malaria as did Austronesians, and some or
all of those genes evolved independently in New Guinea. New Guineans were
already accomplished seafarers, although not as accomplished as the makers
of Lapita pottery. Tens of thousands of years before the arrival of
Austronesians, New Guineans had colonized the Bismarck and Solomon
Archipelagoes, and a trade in obsidian (a volcanic stone suitable for making
sharp tools) was thriving in the Bismarcks at least 18,000 years before the
Austronesians arrived. New Guineans even seem to have expanded recently
westward against the Austronesian tide, into eastern Indonesia, where
languages spoken on the islands of North Halmahera and of Timor are typical
Papuan languages related to some languages of western New Guinea.
In short, the variable outcomes of the Austronesian expansion strikingly
illustrate the role of food production in human population movements.
Austronesian food-producers migrated into two regions (New Guinea and
Indonesia) occupied by resident peoples who were probably related to each
other. The residents of Indonesia were still hunter-gatherers, while the
residents of New Guinea were already food producers and had developed
many of the concomitants of food production (dense populations, disease
resistance, more advanced technology, and so on). As a result, while the
Austronesian expansion swept away the original Indonesians, it failed to
make much headway in the New Guinea region, just as it also failed to make
headway against Austroasiatic and Tai-Kadai food producers in tropical
Southeast Asia.
We have now traced the Austronesian expansion through Indonesia and
up to the shores of New Guinea and tropical Southeast Asia. In Chapter 19 we
shall trace it across the Indian Ocean to Madagascar, while in Chapter 15 we
saw that ecological difficulties kept Austronesians from establishing
themselves in northern and western Australia. The expansion’s remaining
thrust began when the Lapita potters sailed far eastward into the Pacific
beyond the Solomons, into an island realm that no other humans had reached
previously. Around 1200 B.C. Lapita potsherds, the familiar triumvirate of pigs
and chickens and dogs, and the usual other archaeological hallmarks of
Austronesians appeared on the Pacific archipelagoes of Fiji, Samoa, and
Tonga, over a thousand miles east of the Solomons. Early in the Christian era,
most of those same hallmarks (with the notable exception of pottery)
appeared on the islands of eastern Polynesia, including the Societies and
Marquesas. Further long overwater canoe voyages brought settlers north to
Hawaii, east to Pitcairn and Easter Islands, and southwest to New Zealand.
The native inhabitants of most of those islands today are the Polynesians, who
thus are the direct descendants of the Lapita potters. They speak Austronesian
languages closely related to those of the New Guinea region, and their main
crops are the Austronesian package that included taro, yams, bananas,
coconuts, and breadfruit.
With the occupation of the Chatham Islands off New Zealand around A.D.
1400, barely a century before European “explorers” entered the Pacific, the
task of exploring the Pacific was finally completed by Asians. Their tradition
of exploration, lasting tens of thousands of years, had begun when Wiwor’s
ancestors spread through Indonesia to New Guinea and Australia. It ended
only when it had run out of targets and almost every habitable Pacific island
had been occupied.
TO ANYONE INTERESTED in world history, human societies of East Asia and
the Pacific are instructive, because they provide so many examples of how
environment molds history. Depending on their geographic homeland, East
Asian and Pacific peoples differed in their access to domesticable wild plant
and animal species and in their connectedness to other peoples. Again and
again, people with access to the prerequisites for food production, and with a
location favoring diffusion of technology from elsewhere, replaced peoples
lacking these advantages. Again and again, when a single wave of colonists
spread out over diverse environments, their descendants developed in separate
ways, depending on those environmental differences.
For instance, we have seen that South Chinese developed indigenous food
production and technology, received writing and still more technology and
political structures from North China, and went on to colonize tropical
Southeast Asia and Taiwan, largely replacing the former inhabitants of those
areas. Within Southeast Asia, among the descendants or relatives of those
food-producing South Chinese colonists, the Yumbri in the mountain rain
forests of northeastern Thailand and Laos reverted to living as hunter-
gatherers, while the Yumbri’s close relatives the Vietnamese (speaking a
language in the same sub-subfamily of Austroasiatic as the Yumbri language)
remained food producers in the rich Red Delta and established a vast metal-
based empire. Similarly, among Austronesian emigrant farmers from Taiwan
and Indonesia, the Punan in the rain forests of Borneo were forced to turn
back to the hunter-gatherer lifestyle, while their relatives living on Java’s rich
volcanic soils remained food producers, founded a kingdom under the
influence of India, adopted writing, and built the great Buddhist monument at
Borobudur. The Austronesians who went on to colonize Polynesia became
isolated from East Asian metallurgy and writing and hence remained without
writing or metal. As we saw in Chapter 2, though, Polynesian political and
social organization and economies underwent great diversification in different
environments. Within a millennium, East Polynesian colonists had reverted to
hunting-gathering on the Chathams while building a protostate with intensive
food production on Hawaii.
When Europeans at last arrived, their technological and other advantages
enabled them to establish temporary colonial domination over most of
tropical Southeast Asia and the Pacific islands. However, indigenous germs
and food producers prevented Europeans from settling most of this region in
significant numbers. Within this area, only New Zealand, New Caledonia, and
Hawaii—the largest and most remote islands, lying farthest from the equator
and hence in the most nearly temperate (Europe-like) climates—now support
large European populations. Thus, unlike Australia and the Americas, East
Asia and most Pacific islands remain occupied by East Asian and Pacific
peoples.
CHAPTER 18
HEMISPHERES COLLIDING
THE LARGEST POPULATION REPLACEMENT OF THE LAST 13,000 years has been
the one resulting from the recent collision between Old World and New World
societies. Its most dramatic and decisive moment, as we saw in Chapter 3,
occurred when Pizarro’s tiny army of Spaniards captured the Inca emperor
Atahuallpa, absolute ruler of the largest, richest, most populous, and
administratively and technologically most advanced Native American state.
Atahuallpa’s capture symbolizes the European conquest of the Americas,
because the same mix of proximate factors that caused it was also responsible
for European conquests of other Native American societies. Let us now return
to that collision of hemispheres, applying what we have learned since Chapter
3. The basic question to be answered is: why did Europeans reach and
conquer the lands of Native Americans, instead of vice versa? Our starting
point will be a comparison of Eurasian and Native American societies as of
A.D. 1492, the year of Columbus’s “discovery” of the Americas.
OUR COMPARISON BEGINS with food production, a major determinant of local
population size and societal complexity—hence an ultimate factor behind the
conquest. The most glaring difference between American and Eurasian food
production involved big domestic mammal species. In Chapter 9 we
encountered Eurasia’s 13 species, which became its chief source of animal
protein (meat and milk), wool, and hides, its main mode of land transport of
people and goods, its indispensable vehicles of warfare, and (by drawing
plows and providing manure) a big enhancer of crop production. Until
waterwheels and windmills began to replace Eurasia’s mammals in medieval
times, they were also the major source of its “industrial” power beyond
human muscle power—for example, for turning grindstones and operating
water lifts. In contrast, the Americas had only one species of big domestic
mammal, the llama / alpaca, confined to a small area of the Andes and the
adjacent Peruvian coast. While it was used for meat, wool, hides, and goods
transport, it never yielded milk for human consumption, never bore a rider,
never pulled a cart or a plow, and never served as a power source or vehicle of
warfare.
That’s an enormous set of differences between Eurasian and Native
American societies—due largely to the Late Pleistocene extinction
(extermination?) of most of North and South America’s former big wild
mammal species. If it had not been for those extinctions, modern history
might have taken a different course. When Cortés and his bedraggled
adventurers landed on the Mexican coast in 1519, they might have been
driven into the sea by thousands of Aztec cavalry mounted on domesticated
native American horses. Instead of the Aztecs’ dying of smallpox, the
Spaniards might have been wiped out by American germs transmitted by
disease-resistant Aztecs. American civilizations resting on animal power
might have been sending their own conquistadores to ravage Europe. But
those hypothetical outcomes were foreclosed by mammal extinctions
thousands of years earlier.
Those extinctions left Eurasia with many more wild candidates for
domestication than the Americas offered. Most candidates disqualify
themselves as potential domesticates for any of half a dozen reasons. Hence
Eurasia ended up with its 13 species of big domestic mammals and the
Americas with just its one very local species. Both hemispheres also had
domesticated species of birds and small mammals—the turkey, guinea pig,
and Muscovy duck very locally and the dog more widely in the Americas;
chickens, geese, ducks, cats, dogs, rabbits, honeybees, silkworms, and some
others in Eurasia. But the significance of all those species of small domestic
animals was trivial compared with that of the big ones.
Eurasia and the Americas also differed with respect to plant food
production, though the disparity here was less marked than for animal food
production. In 1492 agriculture was widespread in Eurasia. Among the few
Eurasian hunter-gatherers lacking both crops and domestic animals were the
Ainu of northern Japan, Siberian societies without reindeer, and small hunter-
gatherer groups scattered through the forests of India and tropical Southeast
Asia and trading with neighboring farmers. Some other Eurasian societies,
notably the Central Asian pastoralists and the reindeer-herding Lapps and
Samoyeds of the Arctic, had domestic animals but little or no agriculture.
Virtually all other Eurasian societies engaged in agriculture as well as in
herding animals.
Agriculture was also widespread in the Americas, but hunter-gatherers
occupied a larger fraction of the Americas’ area than of Eurasia’s. Those
regions of the Americas without food production included all of northern
North America and southern South America, the Canadian Great Plains, and
all of western North America except for small areas of the U.S. Southwest
that supported irrigation agriculture. It is striking that the areas of Native
America without food production included what today, after Europeans’
arrival, are some of the most productive farmlands and pastures of both North
and South America: the Pacific states of the United States, Canada’s wheat
belt, the pampas of Argentina, and the Mediterranean zone of Chile. The
former absence of food production in these lands was due entirely to their
local paucity of domesticable wild animals and plants, and to geographic and
ecological barriers that prevented the crops and the few domestic animal
species of other parts of the Americas from arriving. Those lands became
productive not only for European settlers but also, in some cases, for Native
Americans, as soon as Europeans introduced suitable domestic animals and
crops. For instance, Native American societies became renowned for their
mastery of horses, and in some cases of cattle and sheepherding, in parts of
the Great Plains, the western United States, and the Argentine pampas. Those
mounted plains warriors and Navajo sheepherders and weavers now figure
prominently in white Americans’ image of American Indians, but the basis for
that image was created only after 1492. These examples demonstrate that the
sole missing ingredients required to sustain food production in large areas of
the Americas were domestic animals and crops themselves.
In those parts of the Americas that did support Native American
agriculture, it was constrained by five major disadvantages vis-à-vis Eurasian
agriculture: widespread dependence on protein-poor corn, instead of Eurasia’s
diverse and protein-rich cereals; hand planting of individual seeds, instead of
broadcast sowing; tilling by hand instead of plowing by animals, which
enables one person to cultivate a much larger area, and which also permits
cultivation of some fertile but tough soils and sods that are difficult to till by
hand (such as those of the North American Great Plains); lack of animal
manuring to increase soil fertility; and just human muscle power, instead of
animal power, for agricultural tasks such as threshing, grinding, and
irrigation. These differences suggest that Eurasian agriculture as of 1492 may
have yielded on the average more calories and protein per person-hour of
labor than Native American agriculture did.
SUCH DIFFERENCES IN food production constituted a major ultimate cause of
the disparities between Eurasian and Native American societies. Among the
resulting proximate factors behind the conquest, the most important included
differences in germs, technology, political organization, and writing. Of these,
the one linked most directly to the differences in food production was germs.
The infectious diseases that regularly visited crowded Eurasian societies, and
to which many Eurasians consequently developed immune or genetic
resistance, included all of history’s most lethal killers: smallpox, measles,
influenza, plague, tuberculosis, typhus, cholera, malaria, and others. Against
that grim list, the sole crowd infectious diseases that can be attributed with
certainty to pre-Columbian Native American societies were nonsyphilitic
treponemas. (As I explained in Chapter 11, it remains uncertain whether
syphilis arose in Eurasia or in the Americas, and the claim that human
tuberculosis was present in the Americas before Columbus is in my opinion
unproven.)
This continental difference in harmful germs resulted paradoxically from
the difference in useful livestock. Most of the microbes responsible for the
infectious diseases of crowded human societies evolved from very similar
ancestral microbes causing infectious diseases of the domestic animals with
which food producers began coming into daily close contact around 10,000
years ago. Eurasia harbored many domestic animal species and hence
developed many such microbes, while the Americas had very few of each.
Other reasons why Native American societies evolved so few lethal microbes
were that villages, which provide ideal breeding grounds for epidemic
diseases, arose thousands of years later in the Americas than in Eurasia; and
that the three regions of the New World supporting urban societies (the
Andes, Mesoamerica, and the U.S. Southeast) were never connected by fast,
high-volume trade on the scale that brought plague, influenza, and possibly
smallpox to Europe from Asia. As a result, even malaria and yellow fever, the
infectious diseases that eventually became major obstacles to European
colonization of the American tropics, and that posed the biggest barrier to the
construction of the Panama Canal, are not American diseases at all but are
caused by microbes of Old World tropical origin, introduced to the Americas
by Europeans.
Rivaling germs as proximate factors behind Europe’s conquest of the
Americas were the differences in all aspects of technology. These differences
stemmed ultimately from Eurasia’s much longer history of densely populated,
economically specialized, politically centralized, interacting and competing
societies dependent on food production. Five areas of technology may be
singled out:
First, metals—initially copper, then bronze, and finally iron—were used
for tools in all complex Eurasian societies as of 1492. In contrast, although
copper, silver, gold, and alloys were used for ornaments in the Andes and
some other parts of the Americas, stone and wood and bone were still the
principal materials for tools in all Native American societies, which made
only limited local use of copper tools.
Second, military technology was far more potent in Eurasia than in the
Americas. European weapons were steel swords, lances, and daggers,
supplemented by small firearms and artillery, while body armor and helmets
were also made of solid steel or else of chain mail. In place of steel, Native
Americans used clubs and axes of stone or wood (occasionally copper in the
Andes), slings, bows and arrows, and quilted armor, constituting much less
effective protection and weaponry. In addition, Native American armies had
no animals to oppose to horses, whose value for assaults and fast transport
gave Europeans an overwhelming advantage until some Native American
societies themselves adopted them.
Third, Eurasian societies enjoyed a huge advantage in their sources of
power to operate machines. The earliest advance over human muscle power
was the use of animals—cattle, horses, and donkeys—to pull plows and to
turn wheels for grinding grain, raising water, and irrigating or draining fields.
Waterwheels appeared in Roman times and then proliferated, along with tidal
mills and windmills, in the Middle Ages. Coupled to systems of geared
wheels, those engines harnessing water and wind power were used not only to
grind grain and move water but also to serve myriad manufacturing purposes,
including crushing sugar, driving blast furnace bellows, grinding ores, making
paper, polishing stone, pressing oil, producing salt, producing textiles, and
sawing wood. It is conventional to define the Industrial Revolution arbitrarily
as beginning with the harnessing of steam power in 18th-century England, but
in fact an industrial revolution based on water and wind power had begun
already in medieval times in many parts of Europe. As of 1492, all of those
operations to which animal, water, and wind power were being applied in
Eurasia were still being carried out by human muscle power in the Americas.
Long before the wheel began to be used in power conversion in Eurasia, it
had become the basis of most Eurasian land transport—not only for animal-
drawn vehicles but also for human-powered wheelbarrows, which enabled
one or more people, still using just human muscle power, to transport much
greater weights than they could have otherwise. Wheels were also adopted in
Eurasian pottery making and in clocks. None of those uses of the wheel was
adopted in the Americas, where wheels are attested only in Mexican ceramic
toys.
The remaining area of technology to be mentioned is sea transport. Many
Eurasian societies developed large sailing ships, some of them capable of
sailing against the wind and crossing the ocean, equipped with sextants,
magnetic compasses, sternpost rudders, and cannons. In capacity, speed,
maneuverability, and seaworthiness, those Eurasian ships were far superior to
the rafts that carried out trade between the New World’s most advanced
societies, those of the Andes and Mesoamerica. Those rafts sailed with the
wind along the Pacific coast. Pizarro’s ship easily ran down and captured such
a raft on his first voyage toward Peru.
IN ADDITION TO their germs and technology, Eurasian and Native American
societies differed in their political organization. By late medieval or
Renaissance times, most of Eurasia had come under the rule of organized
states. Among these, the Habsburg, Ottoman, and Chinese states, the Mogul
state of India, and the Mongol state at its peak in the 13th century started out
as large polyglot amalgamations formed by the conquest of other states. For
that reason they are generally referred to as empires. Many Eurasian states
and empires had official religions that contributed to state cohesion, being
invoked to legitimize the political leadership and to sanction wars against
other peoples. Tribal and band societies in Eurasia were largely confined to
the Arctic reindeer herders, the Siberian hunter-gatherers, and the hunter-
gatherer enclaves in the Indian subcontinent and tropical Southeast Asia.
The Americas had two empires, those of the Aztecs and Incas, which
resembled their Eurasian counterparts in size, population, polyglot makeup,
official religions, and origins in the conquest of smaller states. In the
Americas those were the sole two political units capable of mobilizing
resources for public works or war on the scale of many Eurasian states,
whereas seven European states (Spain, Portugal, England, France, Holland,
Sweden, and Denmark) had the resources to acquire American colonies
between 1492 and 1666. The Americas also held many chiefdoms (some of
them virtually small states) in tropical South America, Mesoamerica beyond
Aztec rule, and the U.S. Southeast. The rest of the Americas was organized
only at the tribal or band level.
The last proximate factor to be discussed is writing. Most Eurasian states
had literate bureaucracies, and in some a significant fraction of the populace
other than bureaucrats was also literate. Writing empowered European
societies by facilitating political administration and economic exchanges,
motivating and guiding exploration and conquest, and making available a
range of information and human experience extending into remote places and
times. In contrast, use of writing in the Americas was confined to the elite in a
small area of Mesoamerica. The Inca Empire employed an accounting system
and mnemonic device based on knots (termed quipu), but it could not have
approached writing as a vehicle for transmitting detailed information.
THUS, EURASIAN SOCIETIES in the time of Columbus enjoyed big advantages
over Native American societies in food production, germs, technology
(including weapons), political organization, and writing. These were the main
factors tipping the outcome of the post-Columbian collisions. But those
differences as of A.D. 1492 represent just one snapshot of historical trajectories
that had extended over at least 13,000 years in the Americas, and over a much
longer time in Eurasia. For the Americas, in particular, the 1492 snapshot
captures the end of the independent trajectory of Native Americans. Let us
now trace out the earlier stages of those trajectories.
Table 18.1 summarizes approximate dates of the appearance of key
developments in the main “homelands” of each hemisphere (the Fertile
Crescent and China in Eurasia, the Andes and Amazonia and Mesoamerica in
the Americas). It also includes the trajectory for the minor New World
homeland of the eastern United States, and that for England, which is not a
homeland at all but is listed to illustrate how rapidly developments spread
from the Fertile Crescent.
This table is sure to horrify any knowledgeable scholar, because it reduces
exceedingly complex histories to a few seemingly precise dates. In reality, all
of those dates are merely attempts to label arbitrary points along a continuum.
For example, more significant than the date of the first metal tool found by
some archaeologist is the time when a significant fraction of all tools was
made of metal, but how common must metal tools be to rate as “widespread”?
Dates for the appearance of the same development may differ among different
parts of the same homeland. For instance, within the Andean region pottery
appears about 1,300 years earlier in coastal Ecuador (3100 B.C.) than in Peru
(1800 B.C.). Some dates, such as those for the rise of chiefdoms, are more
difficult to infer from the archaeological record than are dates of artifacts like
pottery or metal tools. Some of the dates in Table 18.1 are very uncertain,
especially those for the onset of American food production. Nevertheless, as
long as one understands that the table is a simplification, it is useful for
comparing continental histories.
The table suggests that food production began to provide a large fraction
of human diets around 5,000 years earlier in the Eurasian homelands than in
those of the Americas. A caveat must be mentioned immediately: while there
is no doubt about the antiquity of food production in Eurasia, there is
controversy about its onset in the Americas. In particular, archaeologists often
cite considerably older claimed dates for domesticated plants at Coxcatlán
Cave in Mexico, at Guitarrero Cave in Peru, and at some other American sites
than the dates given in the table. Those claims are now being reevaluated for
several reasons: recent direct radiocarbon dating of crop remains themselves
has in some cases been yielding younger dates; the older dates previously
reported were based instead on charcoal thought to be contemporaneous with
the plant remains, but possibly not so; and the status of some of the older
plant remains as crops or just as collected wild plants is uncertain. Still, even
if plant domestication did begin earlier in the Americas than the dates shown
in Table 18.1, agriculture surely did not provide the basis for most human
calorie intake and sedentary existence in American homelands until much
later than in Eurasian homelands.
TABLE 18.1 Historical Trajectories of Eurasia and the Americas
Approximate Date of Adoption
Eurasia
Fertile Crescent
China
England
Plant domestication
8500 B.C.
by 7500 B.C. 3500 B.C.
Animal domestication
8000 B.C.
by 7500 B.C. 3500 B.C.
Pottery
7000 B.C.
by 7500 B.C. 3500 B.C.
Villages
9000 B.C.
by 7500 B.C. 3000 B.C.
Chiefdoms
5500 B.C.
4000 B.C.
2500 B.C.
Widespread metal tools or artifacts (copper and/or bronze) 4000 B.C.
2000 B.C.
2000 B.C.
States
3700 B.C.
2000 B.C.
A.D. 500
Writing
3200 B.C.
by 1300 B.C. A.D. 43
Widespread iron tools
900 B.C.
500 B.C.
650 B.C.
Native America
Andes
Amazonia
Mesoamerica
Eastern U.S.
by 3000 B.C.
3000 B.C.
by 3000 B.C.
2500 B.C.
3500 B.C.
?
500 B.C.
—
3100–1800 B.C.
6000 B.C.
1500 B.C.
2500 B.C.
3100–1800 B.C.
6000 B.C.
1500 B.C.
500 B.C.
by 1500 B.C.
A.D. 1
1500 B.C.
200 B.C.
A.D. 1000
—
—
—
A.D. 1
—
300 B.C.
—
—
—
600 B.C.
—
—
—
—
—
This table gives approximate dates of widespread adoption of significant developments in three
Eurasian and four Native American areas. Dates for animal domestication neglect dogs, which were
domesticated earlier than food-producing animals in both Eurasia and the Americas. Chiefdoms are
inferred from archaeological evidence, such as ranked burials, architecture, and settlement patterns.
The table greatly simplifies a complex mass of historical facts: see the text for some of the many
important caveats.
As we saw in Chapters 5 and 10, only a few relatively small areas of each
hemisphere acted as a “homeland” where food production first arose and from
which it then spread. Those homelands were the Fertile Crescent and China in
Eurasia, and the Andes and Amazonia, Mesoamerica, and the eastern United
States in the Americas. The rate of spread of key developments is especially
well understood for Europe, thanks to the many archaeologists at work there.
As Table 18.1 summarizes for England, once food production and village
living had arrived from the Fertile Crescent after a long lag (5,000 years), the
subsequent lag for England’s adoption of chiefdoms, states, writing, and
especially metal tools was much shorter: 2,000 years for the first widespread
metal tools of copper and bronze, and only 250 years for widespread iron
tools. Evidently, it was much easier for one society of already sedentary
farmers to “borrow” metallurgy from another such society than for nomadic
hunter-gatherers to “borrow” food production from sedentary farmers (or to
be replaced by the farmers).
WHY WERE THE trajectories of all key developments shifted to later dates in
the Americas than in Eurasia? Four groups of reasons suggest themselves: the
later start, more limited suite of wild animals and plants available for
domestication, greater barriers to diffusion, and possibly smaller or more
isolated areas of dense human populations in the Americas than in Eurasia.
As for Eurasia’s head start, humans have occupied Eurasia for about a
million years, far longer than they have lived in the Americas. According to
the archaeological evidence discussed in Chapter 1, humans entered the
Americas at Alaska only around 12,000 B.C., spread south of the Canadian ice
sheets as Clovis hunters a few centuries before 11,000 B.C., and reached the
southern tip of South America by 10,000 B.C. Even if the disputed claims of
older human occupation sites in the Americas prove valid, those postulated
pre-Clovis inhabitants remained for unknown reasons very sparsely
distributed and did not launch a Pleistocene proliferation of hunter-gatherer
societies with expanding populations, technology, and art as in the Old World.
Food production was already arising in the Fertile Crescent only 1,500 years
after the time when Clovis-derived hunter-gatherers were just reaching
southern South America.
Several possible consequences of that Eurasian head start deserve
consideration. First, could it have taken a long time after 11,000 B.C. for the
Americas to fill up with people? When one works out the likely numbers
involved, one finds that this effect would make only a trivial contribution to
the Americas’ 5,000-year lag in food-producing villages. The calculations
given in Chapter 1 tell us that even if a mere 100 pioneering Native
Americans had crossed the Canadian border into the lower United States and
increased at a rate of only 1 percent per year, they would have saturated the
Americas with hunter-gatherers within 1,000 years. Spreading south at a mere
one mile per month, those pioneers would have reached the southern tip of
South America only 700 years after crossing the Canadian border. Those
postulated rates of spread and of population increase are very low compared
with actual known rates for peoples occupying previously uninhabited or
sparsely inhabited lands. Hence the Americas were probably fully occupied
by hunter-gatherers within a few centuries of the arrival of the first colonists.
Second, could a large part of the 5,000-year lag have represented the time
that the first Americans required to become familiar with the new local plant
species, animal species, and rock sources that they encountered? If we can
again reason by analogy with New Guinean and Polynesian hunter-gatherers
and farmers occupying previously unfamiliar environments—such as Maori
colonists of New Zealand or Tudawhe colonists of New Guinea’s Karimui
Basin—the colonists probably discovered the best rock sources and learned to
distinguish useful from poisonous wild plants and animals in much less than a
century.
Third, what about Eurasians’ head start in developing locally appropriate
technology? The early farmers of the Fertile Crescent and China were heirs to
the technology that behaviorally modern Homo sapiens had been developing
to exploit local resources in those areas for tens of thousands of years. For
instance, the stone sickles, underground storage pits, and other technology
that hunter-gatherers of the Fertile Crescent had been evolving to utilize wild
cereals were available to the first cereal farmers of the Fertile Crescent. In
contrast, the first settlers of the Americas arrived in Alaska with equipment
appropriate to the Siberian Arctic tundra. They had to invent for themselves
the equipment suitable to each new habitat they encountered. That technology
lag may have contributed significantly to the delay in Native American
developments.
An even more obvious factor behind the delay was the wild animals and
plants available for domestication. As I discussed in Chapter 6, when hunter-
gatherers adopt food production, it is not because they foresee the potential
benefits awaiting their remote descendants but because incipient food
production begins to offer advantages over the hunter-gatherer lifestyle. Early
food production was less competitive with hunting-gathering in the Americas
than in the Fertile Crescent or China, partly owing to the Americas’ virtual
lack of domesticable wild mammals. Hence early American farmers remained
dependent on wild animals for animal protein and necessarily remained part-
time hunter-gatherers, whereas in both the Fertile Crescent and China animal
domestication followed plant domestication very closely in time to create a
food producing package that quickly won out over hunting-gathering. In
addition, Eurasian domestic animals made Eurasian agriculture itself more
competitive by providing fertilizer, and eventually by drawing plows.
Features of American wild plants also contributed to the lesser
competitiveness of Native American food production. That conclusion is
clearest for the eastern United States, where less than a dozen crops were
domesticated, including small-seeded grains but no large-seeded grains,
pulses, fiber crops, or cultivated fruit or nut trees. It is also clear for
Mesoamerica’s staple grain of corn, which spread to become a dominant crop
elsewhere in the Americas as well. Whereas the Fertile Crescent’s wild wheat
and barley evolved into crops with minimal changes and within a few
centuries, wild teosinte may have required several thousand years to evolve
into corn, having to undergo drastic changes in its reproductive biology and
energy allocation to seed production, loss of the seed’s rock-hard casings, and
an enormous increase in cob size.
As a result, even if one accepts the recently postulated later dates for the
onset of Native American plant domestication, about 1,500 or 2,000 years
would have elapsed between that onset (about 3000–2500 B.C.) and
widespread year-round villages (1800–500 B.C.) in Mesoamerica, the inland
Andes, and the eastern United States. Native American farming served for a
long time just as a small supplement to food acquisition by hunting-gathering,
and supported only a sparse population. If one accepts the traditional, earlier
dates for the onset of American plant domestication, then 5,000 years instead
of 1,500 or 2,000 years elapsed before food production supported villages. In
contrast, villages were closely associated in time with the rise of food
production in much of Eurasia. (The hunter-gatherer lifestyle itself was
sufficiently productive to support villages even before the adoption of
agriculture in parts of both hemispheres, such as Japan and the Fertile
Crescent in the Old World, and coastal Ecuador and Amazonia in the New
World.) The limitations imposed by locally available domesticates in the New
World are well illustrated by the transformations of Native American societies
themselves when other crops or animals arrived, whether from elsewhere in
the Americas or from Eurasia. Examples include the effects of corn’s arrival
in the eastern United States and Amazonia, the llama’s adoption in the
northern Andes after its domestication to the south, and the horse’s
appearance in many parts of North and South America.
In addition to Eurasia’s head start and wild animal and plant species,
developments in Eurasia were also accelerated by the easier diffusion of
animals, plants, ideas, technology, and people in Eurasia than in the
Americas, as a result of several sets of geographic and ecological factors.
Eurasia’s east–west major axis, unlike the Americas’ north–south major axis,
permitted diffusion without change in latitude and associated environmental
variables. In contrast to Eurasia’s consistent east–west breadth, the New
World was constricted over the whole length of Central America and
especially at Panama. Not least, the Americas were more fragmented by areas
unsuitable for food production or for dense human populations. These
ecological barriers included the rain forests of the Panamanian isthmus
separating Mesoamerican societies from Andean and Amazonian societies;
the deserts of northern Mexico separating Mesoamerica from U.S.
southwestern and southeastern societies; dry areas of Texas separating the
U.S. Southwest from the Southeast; and the deserts and high mountains
fencing off U.S. Pacific coast areas that would otherwise have been suitable
for food production. As a result, there was no diffusion of domestic animals,
writing, or political entities, and limited or slow diffusion of crops and
technology, between the New World centers of Mesoamerica, the eastern
United States, and the Andes and Amazonia.
Some specific consequences of these barriers within the Americas deserve
mention. Food production never diffused from the U.S. Southwest and
Mississippi Valley to the modern American breadbaskets of California and
Oregon, where Native American societies remained hunter-gatherers merely
because they lacked appropriate domesticates. The llama, guinea pig, and
potato of the Andean highlands never reached the Mexican highlands, so
Mesoamerica and North America remained without domestic mammals
except for dogs. Conversely, the domestic sunflower of the eastern United
States never reached Mesoamerica, and the domestic turkey of Mesoamerica
never made it to South America or the eastern United States. Mesoamerican
corn and beans took 3,000 and 4,000 years, respectively, to cover the 700
miles from Mexico’s farmlands to the eastern U.S. farmlands. After corn’s
arrival in the eastern United States, seven centuries more passed before the
development of a corn variety productive in North American climates
triggered the Mississippian emergence. Corn, beans, and squash may have
taken several thousand years to spread from Mesoamerica to the U.S.
Southwest. While Fertile Crescent crops spread west and east sufficiently fast
to preempt independent domestication of the same species or else
domestication of closely related species elsewhere, the barriers within the
Americas gave rise to many such parallel domestications of crops.
As striking as these effects of barriers on crop and livestock diffusion are
the effects on other features of human societies. Alphabets of ultimately
eastern Mediterranean origin spread throughout all complex societies of
Eurasia, from England to Indonesia, except for areas of East Asia where
derivatives of the Chinese writing system took hold. In contrast, the New
World’s sole writing systems, those of Mesoamerica, never spread to the
complex Andean and eastern U.S. societies that might have adopted them.
The wheels invented in Mesoamerica as parts of toys never met the llamas
domesticated in the Andes, to generate wheeled transport for the New World.
From east to west in the Old World, the Macedonian Empire and the Roman
Empire both spanned 3,000 miles, the Mongol Empire 6,000 miles. But the
empires and states of Mesoamerica had no political relations with, and
apparently never even heard of, the chiefdoms of the eastern United States
700 miles to the north or the empires and states of the Andes 1,200 miles to
the south.
The greater geographic fragmentation of the Americas compared with
Eurasia is also reflected in distributions of languages. Linguists agree in
grouping all but a few Eurasian languages into about a dozen language
families, each consisting of up to several hundred related languages. For
example, the Indo-European language family, which includes English as well
as French, Russian, Greek, and Hindi, comprises about 144 languages. Quite
a few of those families occupy large contiguous areas—in the case of Indo-
European, the area encompassing most of Europe east through much of
western Asia to India. Linguistic, historical, and archaeological evidence
combines to make clear that each of these large, contiguous distributions
stems from a historical expansion of an ancestral language, followed by
subsequent local linguistic differentiation to form a family of related
languages (Table 18.2). Most such expansions appear to be attributable to the
advantages that speakers of the ancestral language, belonging to food-
producing societies, held over hunter-gatherers. We already discussed such
historical expansions in Chapters 16 and 17 for the Sino-Tibetan,
Austronesian, and other East Asian language families. Among major
expansions of the last millennium are those that carried Indo-European
languages from Europe to the Americas and Australia, the Russian language
from eastern Europe across Siberia, and Turkish (a language of the Altaic
family) from Central Asia westward to Turkey.
With the exception of the Eskimo-Aleut language family of the American
Arctic and the Na-Dene language family of Alaska, northwestern Canada, and
the U.S. Southwest, the Americas lack examples of large-scale language
expansions widely accepted by linguists. Most linguists specializing in Native
American languages do not discern large, clear-cut groupings other than
Eskimo-Aleut and Na-Dene. At most, they consider the evidence sufficient
only to group other Native American languages (variously estimated to
number from 600 to 2,000) into a hundred or more language groups or
isolated languages. A controversial minority view is that of the linguist Joseph
Greenberg, who groups all Native American languages other than Eskimo-
Aleut and Na-Dene languages into a single large family, termed Amerind,
with about a dozen subfamilies.
TABLE 18.2 Language Expansions in the Old World
Inferred
Language Family or
Expansion
Ultimate Driving Force
Date
Language
6000 or 4000
Ukraine or Anatolia
Europe, C.
food production or horse-
Indo-European
B.C.
Asia, India
based pastoralism
6000 B.C.–
Elamo-Dravidian
Iran
India
food production
2000 B.C.
4000 B.C.–
Tibetan Plateau, N. China
S.
Sino-Tibetan
food production
present
China, tropical S.E. Asia
3000 B.C.–
Austronesian
S. China
Indonesia, Pacific islands food production
1000 B.C.
3000
B.C.–A.D.
Bantu
Nigeria and Cameroon
S. Africa
food production
1000
3000
Austroasiatic
S. China
tropical S.E. Asia, India
food production
B.C.–A.D. 1
1000
B.C.–A.D.
Tai-Kadai, Miao-Yao S. China
tropical S.E. Asia
food production
1500
A.D. 892
Hungarian
Ural Mts.
Hungary
horse-based pastoralism
A.D.
Altaic (Mongol,
Asian steppes
Europe, Turkey,
1000–A.D.
horse-based pastoralism
Turkish)
China, India
1300
A.D.
1480–A.D.
Russian
European Russia
Asiatic Siberia
food production
1638
Some of Greenberg’s subfamilies, and some groupings recognized by
more-traditional linguists, may turn out to be legacies of New World
population expansions driven in part by food production. These legacies may
include the Uto-Aztecan languages of Mesoamerica and the western United
States, the Oto-Manguean languages of Mesoamerica, the Natchez-
Muskogean languages of the U.S. Southeast, and the Arawak languages of the
West Indies. But the difficulties that linguists have in agreeing on groupings
of Native American languages reflect the difficulties that complex Native
American societies themselves faced in expanding within the New World.
Had any food-producing Native American peoples succeeded in spreading far
with their crops and livestock and rapidly replacing hunter-gatherers over a
large area, they would have left legacies of easily recognized language
families, as in Eurasia, and the relationships of Native American languages
would not be so controversial.
Thus, we have identified three sets of ultimate factors that tipped the
advantage to European invaders of the Americas: Eurasia’s long head start on
human settlement; its more effective food production, resulting from greater
availability of domesticable wild plants and especially of animals; and its less
formidable geographic and ecological barriers to intracontinental diffusion. A
fourth, more speculative ultimate factor is suggested by some puzzling non-
inventions in the Americas: the non-inventions of writing and wheels in
complex Andean societies, despite a time depth of those societies
approximately equal to that of complex Mesoamerican societies that did make
those inventions; and wheels’ confinement to toys and their eventual
disappearance in Mesoamerica, where they could presumably have been
useful in human-powered wheelbarrows, as in China. These puzzles remind
one of equally puzzling non-inventions, or else disappearances of inventions,
in small isolated societies, including Aboriginal Tasmania, Aboriginal
Australia, Japan, Polynesian islands, and the American Arctic. Of course, the
Americas in aggregate are anything but small: their combined area is fully 76
percent that of Eurasia, and their human population as of A.D. 1492 was
probably also a large fraction of Eurasia’s. But the Americas, as we have seen,
are broken up into “islands” of societies with tenuous connections to each
other. Perhaps the histories of Native American wheels and writing exemplify
the principles illustrated in a more extreme form by true island societies.
AFTER AT LEAST 13,000 years of separate developments, advanced American
and Eurasian societies finally collided within the last thousand years. Until
then, the sole contacts between human societies of the Old and the New
Worlds had involved the hunter-gatherers on opposite sides of the Bering
Strait.
There were no Native American attempts to colonize Eurasia, except at
the Bering Strait, where a small population of Inuit (Eskimos) derived from
Alaska established itself across the strait on the opposite Siberian coast. The
first documented Eurasian attempt to colonize the Americas was by the Norse
at Arctic and sub-Arctic latitudes (Figure 18.1). Norse from Norway
colonized Iceland in A.D. 874, then Norse from Iceland colonized Greenland
in A.D. 986, and finally Norse from Greenland repeatedly visited the
northeastern coast of North America between about A.D. 1000 and 1350. The
sole Norse archaeological site discovered in the Americas is on
Newfoundland, possibly the region described as Vinland in Norse sagas, but
these also mention landings evidently farther north, on the coasts of Labrador
and Baffin Island.
Iceland’s climate permitted herding and extremely limited agriculture, and
its area was sufficient to support a Norse-derived population that has persisted
to this day. But most of Greenland is covered by an ice cap, and even the two
most favorable coastal fjords were marginal for Norse food production. The
Greenland Norse population never exceeded a few thousand. It remained
dependent on imports of food and iron from Norway, and of timber from the
Labrador coast. Unlike Easter Island and other remote Polynesian islands,
Greenland could not support a self-sufficient food-producing society, though
it did support self-sufficient Inuit hunter-gatherer populations before, during,
and after the Norse occupation period. The populations of Iceland and
Norway themselves were too small and too poor for them to continue their
support of the Greenland Norse population.
In the Little Ice Age that began in the 13th century, the cooling of the
North Atlantic made food production in Greenland, and Norse voyaging to
Greenland from Norway or Iceland, even more marginal than before. The
Greenlanders’ last known contact with Europeans came in 1410 with an
Icelandic ship that arrived after being blown off course. When Europeans
finally began again to visit Greenland in 1577, its Norse colony no longer
existed, having evidently disappeared without any record during the 15th
century.
But the coast of North America lay effectively beyond the reach of ships
sailing directly from Norway itself, given Norse ship technology of the period
A.D. 986–1410. The Norse visits were instead launched from the Greenland
colony, separated from North America only by the 200-mile width of Davis
Strait. However, the prospect of that tiny marginal colony’s sustaining an
exploration, conquest, and settlement of the Americas was nil. Even the sole
Norse site located on Newfoundland apparently represents no more than a
winter camp occupied by a few dozen people for a few years. The Norse
sagas describe attacks on their Vinland camp by people termed Skraelings,
evidently either Newfoundland Indians or Dorset Eskimos.
The fate of the Greenland colony, medieval Europe’s most remote
outpost, remains one of archaeology’s romantic mysteries. Did the last
Greenland Norse starve to death, attempt to sail off, intermarry with Eskimos,
or succumb to disease or Eskimo arrows? While those questions of proximate
cause remain unanswered, the ultimate reasons why Norse colonization of
Greenland and America failed are abundantly clear. It failed because the
source (Norway), the targets (Greenland and Newfoundland), and the time
(A.D. 984–1410) guaranteed that Europe’s potential advantages of food
production, technology, and political organization could not be applied
effectively. At latitudes too high for much food production, the iron tools of a
few Norse, weakly supported by one of Europe’s poorer states, were no match
for the stone, bone, and wooden tools of Eskimo and Indian hunter-gatherers,
the world’s greatest masters of Arctic survival skills.
THE SECOND EURASIAN attempt to colonize the Americas succeeded because
it involved a source, target, latitude, and time that allowed Europe’s potential
advantages to be exerted effectively. Spain, unlike Norway, was rich and
populous enough to support exploration and subsidize colonies. Spanish
landfalls in the Americas were at subtropical latitudes highly suitable for food
production, based at first mostly on Native American crops but also on
Eurasian domestic animals, especially cattle and horses. Spain’s transatlantic
colonial enterprise began in 1492, at the end of a century of rapid
development of European oceangoing ship technology, which by then
incorporated advances in navigation, sails, and ship design developed by Old
World societies (Islam, India, China, and Indonesia) in the Indian Ocean. As a
result, ships built and manned in Spain itself were able to sail to the West
Indies; there was nothing equivalent to the Greenland bottleneck that had
throttled Norse colonization. Spain’s New World colonies were soon joined
by those of half a dozen other European states.
The first European settlements in the Americas, beginning with the one
founded by Columbus in 1492, were in the West Indies. The island Indians,
whose estimated population at the time of their “discovery” exceeded a
million, were rapidly exterminated on most islands by disease, dispossession,
enslavement, warfare, and casual murder. Around 1508 the first colony was
founded on the American mainland, at the Isthmus of Panama. Conquest of
the two large mainland empires, those of the Aztecs and Incas, followed in
1519–1520 and 1532–1533, respectively. In both conquests European-
transmitted epidemics (probably smallpox) made major contributions, by
killing the emperors themselves, as well as a large fraction of the population.
The overwhelming military superiority of even tiny numbers of mounted
Spaniards, together with their political skills at exploiting divisions within the
native population, did the rest. European conquest of the remaining native
states of Central America and northern South America followed during the
16th and 17th centuries.
As for the most advanced native societies of North America, those of the
U.S. Southeast and the Mississippi River system, their destruction was
accomplished largely by germs alone, introduced by early European explorers
and advancing ahead of them. As Europeans spread throughout the Americas,
many other native societies, such as the Mandans of the Great Plains and the
Sadlermiut Eskimos of the Arctic, were also wiped out by disease, without
need for military action. Populous native societies not thereby eliminated
were destroyed in the same way the Aztecs and Incas had been—by full-scale
wars, increasingly waged by professional European soldiers and their native
allies. Those soldiers were backed by the political organizations initially of
the European mother countries, then of the European colonial governments in
the New World, and finally of the independent neo-European states that
succeeded the colonial governments.
Smaller native societies were destroyed more casually, by small-scale
raids and murders carried out by private citizens. For instance, California’s
native hunter-gatherers initially numbered about 200,000 in aggregate, but
they were splintered among a hundred tribelets, none of which required a war
to be defeated. Most of those tribelets were killed off or dispossessed during
or soon after the California gold rush of 1848–52, when large numbers of
immigrants flooded the state. As one example, the Yahi tribelet of northern
California, numbering about 2,000 and lacking firearms, was destroyed in
four raids by armed white settlers: a dawn raid on a Yahi village carried out
by 17 settlers on August 6, 1865; a massacre of Yahis surprised in a ravine in
1866; a massacre of 33 Yahis tracked to a cave around 1867; and a final
massacre of about 30 Yahis trapped in another cave by 4 cowboys around
1868. Many Amazonian Indian groups were similarly eliminated by private
settlers during the rubber boom of the late 19th and early 20th centuries. The
final stages of the conquest are being played out in the present decade, as the
Yanomamo and other Amazonian Indian societies that remain independent are
succumbing to disease, being murdered by miners, or being brought under
control by missionaries or government agencies.
The end result has been the elimination of populous Native American
societies from most temperate areas suitable for European food production
and physiology. In North America those that survived as sizable intact
communities now live mostly on reservations or other lands considered
undesirable for European food production and mining, such as the Arctic and
arid areas of the U.S. West. Native Americans in many tropical areas have
been replaced by immigrants from the Old World tropics (especially black
Africans, along with Asian Indians and Javanese in Suriname).
In parts of Central America and the Andes, the Native Americans were
originally so numerous that, even after epidemics and wars, much of the
population today remains Native American or mixed. That is especially true
at high altitudes in the Andes, where genetically European women have
physiological difficulties even in reproducing, and where native Andean crops
still offer the most suitable basis for food production. However, even where
Native Americans do survive, there has been extensive replacement of their
culture and languages with those of the Old World. Of the hundreds of Native
American languages originally spoken in North America, all except 187 are
no longer spoken at all, and 149 of these last 187 are moribund in the sense
that they are being spoken only by old people and no longer learned by
children. Of the approximately 40 New World nations, all now have an Indo-
European language or creole as the official language. Even in the countries
with the largest surviving Native American populations, such as Peru,
Bolivia, Mexico, and Guatemala, a glance at photographs of political and
business leaders shows that they are disproportionately Europeans, while
several Caribbean nations have black African leaders and Guyana has had
Asian Indian leaders.
The original Native American population has been reduced by a debated
large percentage: estimates for North America range up to 95 percent. But the
total human population of the Americas is now approximately ten times what
it was in 1492, because of arrivals of Old World peoples (Europeans,
Africans, and Asians). The Americas’ population now consists of a mixture of
peoples originating from all continents except Australia. That demographic
shift of the last 500 years—the most massive shift on any continent except
Australia—has its ultimate roots in developments between about 11,000 B.C.
and A.D. 1.
CHAPTER 19
HOW AFRICA BECAME BLACK
NO MATTER HOW MUCH ONE HAS READ ABOUT AFRICA beforehand, one’s first
impressions from actually being there are overwhelming. On the streets of
Windhoek, capital of newly independent Namibia, I saw black Herero people,
black Ovambos, whites, and Namas, different again from both blacks and
whites. They were no longer mere pictures in a textbook, but living humans in
front of me. Outside Windhoek, the last of the formerly widespread Kalahari
Bushmen were struggling for survival. But what most surprised me in
Namibia was a street sign: one of downtown Windhoek’s main roads was
called Goering Street!
Surely, I thought, no country could be so dominated by unrepentant Nazis
as to name a street after the notorious Nazi Reichskommissar and founder of
the Luftwaffe, Hermann Goering! No, it turned out that the street instead
commemorated Hermann’s father, Heinrich Goering, founding
Reichskommissar of the former German colony of South-West Africa, which
became Namibia. But Heinrich was also a problematic figure, for his legacy
included one of the most vicious attacks by European colonists on Africans,
Germany’s 1904 war of extermination against the Hereros. Today, while
events in neighboring South Africa command more of the world’s attention,
Namibia as well is struggling to deal with its colonial past and establish a
multiracial society. Namibia illustrated for me how inseparable Africa’s past
is from its present.
Most Americans and many Europeans equate native Africans with blacks,
white Africans with recent intruders, and African racial history with the story
of European colonialism and slave trading. There is an obvious reason why
we focus on those particular facts: blacks are the sole native Africans familiar
to most Americans, because they were brought in large numbers as slaves to
the United States. But very different peoples may have occupied much of
modern black Africa until as recently as a few thousand years ago, and so-
called African blacks themselves are heterogeneous. Even before the arrival
of white colonialists, Africa already harbored not just blacks but (as we shall
see) five of the world’s six major divisions of humanity, and three of them are
confined as natives to Africa. One-quarter of the world’s languages are
spoken only in Africa. No other continent approaches this human diversity.
Africa’s diverse peoples resulted from its diverse geography and its long
prehistory. Africa is the only continent to extend from the northern to the
southern temperate zone, while also encompassing some of the world’s driest
deserts, largest tropical rain forests, and highest equatorial mountains.
Humans have lived in Africa far longer than anywhere else: our remote
ancestors originated there around 7 million years ago, and anatomically
modern Homo sapiens may have arisen there since then. The long interactions
between Africa’s many peoples generated its fascinating prehistory, including
two of the most dramatic population movements of the past 5,000 years—the
Bantu expansion and the Indonesian colonization of Madagascar. All of those
past interactions continue to have heavy consequences, because the details of
who arrived where before whom are shaping Africa today.
How did those five divisions of humanity get to be where they are now in
Africa? Why were blacks the ones who came to be so widespread, rather than
the four other groups whose existence Americans tend to forget? How can we
ever hope to wrest the answers to those questions from Africa’s preliterate
past, lacking the written evidence that teaches us about the spread of the
Roman Empire? African prehistory is a puzzle on a grand scale, still only
partly solved. As it turns out, the story has some little-appreciated but striking
parallels with the American prehistory that we encountered in the preceding
chapter.
THE FIVE MAJOR human groups to which Africa was already home by A.D.
1000 are those loosely referred to by laypeople as blacks, whites, African
Pygmies, Khoisan, and Asians. Figure 19.1 depicts their distributions, while
the portraits in Chapter 14 will remind you of their striking differences in skin
color, hair form and color, and facial features. Blacks were formerly confined
to Africa, Pygmies and Khoisan still live only there, while many more whites
and Asians live outside Africa than in it. These five groups constitute or
represent all the major divisions of humanity except for Aboriginal
Australians and their relatives.
Many readers may already be protesting: don’t stereotype people by
classifying them into arbitrary “races”! Yes, I acknowledge that each of these
so-called major groups is very diverse. To lump people as different as Zulus,
Somalis, and Ibos under the single heading of “blacks” ignores the differences
between them. We ignore equally big differences when we lump Africa’s
Egyptians and Berbers with each other and with Europe’s Swedes under the
single heading of “whites.” In addition, the divisions between blacks, whites,
and the other major groups are arbitrary, because each such group shades into
others: all human groups on Earth have mated with humans of every other
group that they encountered. Nevertheless, as we’ll see, recognizing these
major groups is still so useful for understanding history that I’ll use the group
names as shorthand, without repeating the above caveats in every sentence.
Of the five African groups, representatives of many populations of blacks
and whites are familiar to Americans and Europeans and need no physical
description. Blacks occupied the largest area of Africa even as of A.D. 1400:
the southern Sahara and most of sub-Saharan Africa (see Figure 19.1). While
American blacks of African descent originated mainly from Africa’s west
coastal zone, similar peoples traditionally occupied East Africa as well, north
to the Sudan and south to the southeast coast of South Africa itself. Whites,
ranging from Egyptians and Libyans to Moroccans, occupied Africa’s north
coastal zone and the northern Sahara. Those North Africans would hardly be
confused with blue-eyed blond-haired Swedes, but most laypeople would still
call them “whites” because they have lighter skin and straighter hair than
peoples to the south termed “blacks.” Most of Africa’s blacks and whites
depended on farming or herding, or both, for their living.
In contrast, the next two groups, the Pygmies and Khoisan, include
hunter-gatherers without crops or livestock. Like blacks, Pygmies have dark
skins and tightly curled hair. However, Pygmies differ from blacks in their
much smaller size, more reddish and less black skins, more extensive facial
and body hair, and more prominent foreheads, eyes, and teeth. Pygmies are
mostly hunter-gatherers living in groups widely scattered through the Central
African rain forest and trading with (or working for) neighboring black
farmers.
The Khoisan make up the group least familiar to Americans, who are
unlikely even to have heard of their name. Formerly distributed over much of
southern Africa, they consisted not only of small-sized hunter-gatherers,
known as San, but also of larger herders, known as Khoi. (These names are
now preferred to the better-known terms Hottentot and Bushmen.) Both the
Khoi and the San look (or looked) quite unlike African blacks: their skins are
yellowish, their hair is very tightly coiled, and the women tend to accumulate
much fat in their buttocks (termed “steatopygia”). As a distinct group, the
Khoi have been greatly reduced in numbers: European colonists shot,
displaced, or infected many of them, and most of the survivors interbred with
Europeans to produce the populations variously known in South Africa as
Coloreds or Basters. The San were similarly shot, displaced, and infected, but
a dwindling small number have preserved their distinctness in Namibian
desert areas unsuitable for agriculture, as depicted some years ago in the
widely seen film The Gods Must Be Crazy.
The northern distribution of Africa’s whites is unsurprising, because
physically similar peoples live in adjacent areas of the Near East and Europe.
Throughout recorded history, people have been moving back and forth
between Europe, the Near East, and North Africa. I’ll therefore say little more
about Africa’s whites in this chapter, since their origins aren’t mysterious.
Instead, the mystery involves blacks, Pygmies, and Khoisan, whose
distributions hint at past population upheavals. For instance, the present
fragmented distribution of the 200,000 Pygmies, scattered amid 120 million
blacks, suggests that Pygmy hunters were formerly widespread through the
equatorial forests until displaced and isolated by the arrival of black farmers.
The Khoisan area of southern Africa is surprisingly small for a people so
distinct in anatomy and language. Could the Khoisan, too, have been
originally more widespread until their more northerly populations were
somehow eliminated?
I’ve saved the biggest anomaly for last. The large island of Madagascar
lies only 250 miles off the East African coast, much closer to Africa than to
any other continent, and separated by the whole expanse of the Indian Ocean
from Asia and Australia. Madagascar’s people prove to be a mixture of two
elements. Not surprisingly, one element is African blacks, but the other
consists of people instantly recognizable, from their appearance, as tropical
Southeast Asians. Specifically, the language spoken by all the people of
Madagascar—Asians, blacks, and mixed—is Austronesian and very similar to
the Ma’anyan language spoken on the Indonesian island of Borneo, over
4,000 miles across the open Indian Ocean from Madagascar. No other people
remotely resembling Borneans live within thousands of miles of Madagascar.
These Austronesians, with their Austronesian language and modified
Austronesian culture, were already established on Madagascar by the time it
was first visited by Europeans, in 1500. This strikes me as the single most
astonishing fact of human geography for the entire world. It’s as if Columbus,
on reaching Cuba, had found it occupied by blue-eyed, blond-haired
Scandinavians speaking a language close to Swedish, even though the nearby
North American continent was inhabited by Native Americans speaking
Amerindian languages. How on earth could prehistoric people of Borneo,
presumably voyaging in boats without maps or compasses, end up in
Madagascar?
THE CASE OF Madagascar tells us that peoples’ languages, as well as their
physical appearance, can yield important clues to their origins. Just by
looking at the people of Madagascar, we’d have known that some of them
came from tropical Southeast Asia, but we wouldn’t have known from which
area of tropical Southeast Asia, and we’d never have guessed Borneo. What
else can we learn from African languages that we didn’t already know from
African faces?
The mind-boggling complexities of Africa’s 1,500 languages were
clarified by Stanford University’s great linguist Joseph Greenberg, who
recognized that all those languages fall into just five families (see Figure 19.2
for their distribution). Readers accustomed to thinking of linguistics as dull
and technical may be surprised to learn what fascinating contributions Figure
19.2 makes to our understanding of African history.
If we begin by comparing Figure 19.2 with Figure 19.1, we’ll see a rough
correspondence between language families and anatomically defined human
groups: languages of a given language family tend to be spoken by distinct
people. In particular, Afroasiatic speakers mostly prove to be people who
would be classified as whites or blacks, Nilo-Saharan and Niger-Congo
speakers prove to be blacks, Khoisan speakers Khoisan, and Austronesian
speakers Indonesian. This suggests that languages have tended to evolve
along with the people who speak them.
Concealed at the top of Figure 19.2 is our first surprise, a big shock for
Eurocentric believers in the superiority of so-called Western civilization.
We’re taught that Western civilization originated in the Near East, was
brought to brilliant heights in Europe by the Greeks and Romans, and
produced three of the world’s great religions: Christianity, Judaism, and
Islam. Those religions arose among peoples speaking three closely related
languages, termed Semitic languages: Aramaic (the language of Christ and
the Apostles), Hebrew, and Arabic, respectively. We instinctively associate
Semitic peoples with the Near East.
However, Greenberg determined that Semitic languages really form only
one of six or more branches of a much larger language family, Afroasiatic, all
of whose other branches (and other 222 surviving languages) are confined to
Africa. Even the Semitic subfamily itself is mainly African, 12 of its 19
surviving languages being confined to Ethiopia. This suggests that Afroasiatic
languages arose in Africa, and that only one branch of them spread to the
Near East. Hence it may have been Africa that gave birth to the languages
spoken by the authors of the Old and New Testaments and the Koran, the
moral pillars of Western civilization.
The next surprise in Figure 19.2 is a seeming detail on which I didn’t
comment when I just told you that distinct peoples tend to have distinct
languages. Among Africa’s five groups of people—blacks, whites, Pygmies,
Khoisan, and Indonesians—only the Pygmies lack any distinct languages:
each band of Pygmies speaks the same language as the neighboring group of
black farmers. However, if one compares a given language as spoken by
Pygmies with the same language as spoken by blacks, the Pygmy version
seems to contain some unique words with distinctive sounds.
Originally, of course, people as distinctive as the Pygmies, living in a
place as distinctive as the equatorial African rain forest, were surely isolated
enough to develop their own language family. However, today those
languages are gone, and we already saw from Figure 19.1 that the Pygmies’
modern distribution is highly fragmented. Thus, distributional and linguistic
clues combine to suggest that the Pygmy homeland was engulfed by invading
black farmers, whose languages the remaining Pygmies adopted, leaving only
traces of their original languages in some words and sounds. We saw
previously that much the same is true of the Malaysian Negritos (Semang)
and Philippine Negritos, who adopted Austroasiatic and Austronesian
languages, respectively, from the farmers who came to surround them.
The fragmented distribution of Nilo-Saharan languages in Figure 19.2
similarly implies that many speakers of those languages have been engulfed
by speakers of Afroasiatic or Niger-Congo languages. But the distribution of
Khoisan languages testifies to an even more dramatic engulfing. Those
languages are famously unique in the whole world in their use of clicks as
consonants. (If you’ve been puzzled by the name !Kung Bushman, the
exclamation mark is not an expression of premature astonishment; it’s just
how linguists denote a click.) All existing Khoisan languages are confined to
southern Africa, with two exceptions. Those exceptions are two very
distinctive, click-laden Khoisan languages named Hadza and Sandawe,
stranded in Tanzania more than 1,000 miles from the nearest Khoisan
languages of southern Africa.
In addition, Xhosa and a few other Niger-Congo languages of southern
Africa are full of clicks. Even more unexpectedly, clicks or Khoisan words
also appear in two Afroasiatic languages spoken by blacks in Kenya, stranded
still farther from present Khoisan peoples than are the Hadza and Sandawe
peoples of Tanzania. All this suggests that Khoisan languages and peoples
formerly extended far north of their present southern African distribution,
until they too, like the Pygmies, were engulfed by the blacks, leaving only
linguistic legacies of their former presence. That’s a unique contribution of
the linguistic evidence, something we could hardly have guessed just from
physical studies of living people.
I have saved the most remarkable contribution of linguistics for last. If
you look again at Figure 19.2, you’ll see that the Niger-Congo language
family is distributed all over West Africa and most of subequatorial Africa,
apparently giving no clue as to where within that enormous range the family
originated. However, Greenberg recognized that all Niger-Congo languages of
subequatorial Africa belong to a single language subgroup termed Bantu. That
subgroup accounts for nearly half of the 1,032 Niger-Congo languages and
for more than half (nearly 200 million) of the Niger-Congo speakers. But all
those 500 Bantu languages are so similar to each other that they have been
facetiously described as 500 dialects of a single language.
Collectively, the Bantu languages constitute only a single, low-order
subfamily of the Niger-Congo language family. Most of the 176 other
subfamilies are crammed into West Africa, a small fraction of the entire
Niger-Congo range. In particular, the most distinctive Bantu languages, and
the non-Bantu Niger-Congo languages most closely related to Bantu
languages, are packed into a tiny area of Cameroon and adjacent eastern
Nigeria.
Evidently, the Niger-Congo language family arose in West Africa; the
Bantu branch of it arose at the east end of that range, in Cameroon and
Nigeria; and the Bantu then spread out of that homeland over most of
subequatorial Africa. That spread must have begun long ago enough that the
ancestral Bantu language had time to split into 500 daughter languages, but
nevertheless recently enough that all those daughter languages are still very
similar to each other. Since all other Niger-Congo speakers, as well as the
Bantu, are blacks, we couldn’t have inferred who migrated in which direction
just from the evidence of physical anthropology.
To make this type of linguistic reasoning clear, let me give you a familiar
example: the geographic origins of the English language. Today, by far the
largest number of people whose first language is English live in North
America, with others scattered over the globe in Britain, Australia, and other
countries. Each of those countries has its own dialects of English. If we knew
nothing else about language distributions and history, we might have guessed
that the English language arose in North America and was carried overseas to
Britain and Australia by colonists.
But all those English dialects form only one low-order subgroup of the
Germanic language family. All the other subgroups—the various
Scandinavian, German, and Dutch languages—are crammed into
northwestern Europe. In particular, Frisian, the other Germanic language most
closely related to English, is confined to a tiny coastal area of Holland and
western Germany. Hence a linguist would immediately deduce correctly that
the English language arose in coastal northwestern Europe and spread around
the world from there. In fact, we know from recorded history that English
really was carried from there to England by invading Anglo-Saxons in the
fifth and sixth centuries A.D.
Essentially the same line of reasoning tells us that the nearly 200 million
Bantu people, now flung over much of the map of Africa, arose from
Cameroon and Nigeria. Along with the North African origins of Semites and
the origins of Madagascar’s Asians, that’s another conclusion that we couldn’t
have reached without linguistic evidence.
We had already deduced, from Khoisan language distributions and the
lack of distinct Pygmy languages, that Pygmies and Khoisan peoples had
formerly ranged more widely, until they were engulfed by blacks. (I’m using
“engulfing” as a neutral all-embracing word, regardless of whether the
process involved conquest, expulsion, interbreeding, killing, or epidemics.)
We’ve now seen, from Niger-Congo language distributions, that the blacks
who did the engulfing were the Bantu. The physical and linguistic evidence
considered so far has let us infer these prehistoric engulfings, but it still hasn’t
solved their mysteries for us. Only the further evidence that I’ll now present
can help us answer two more questions: What advantages enabled the Bantu
to displace the Pygmies and Khoisan? When did the Bantu reach the former
Pygmy and Khoisan homelands?
TO APPROACH THE question about the Bantu’s advantages, let’s examine the
remaining type of evidence from the living present—the evidence derived
from domesticated plants and animals. As we saw in previous chapters, that
evidence is important because food production led to high population
densities, germs, technology, political organization, and other ingredients of
power. Peoples who, by accident of their geographic location, inherited or
developed food production thereby became able to engulf geographically less
endowed people.
When Europeans reached sub-Saharan Africa in the 1400s, Africans were
growing five sets of crops (Figure 19.3), each of them laden with significance
for African history. The first set was grown only in North Africa, extending to
the highlands of Ethiopia. North Africa enjoys a Mediterranean climate,
characterized by rainfall concentrated in the winter months. (Southern
California also experiences a Mediterranean climate, explaining why my
basement and that of millions of other southern Californians often gets
flooded in the winter but infallibly dries out in the summer.) The Fertile
Crescent, where agriculture arose, enjoys that same Mediterranean pattern of
winter rains.
Hence North Africa’s original crops all prove to be ones adapted to
germinating and growing with winter rains, and known from archaeological
evidence to have been first domesticated in the Fertile Crescent beginning
around 10,000 years ago. Those Fertile Crescent crops spread into
climatically similar adjacent areas of North Africa and laid the foundations
for the rise of ancient Egyptian civilization. They include such familiar crops
as wheat, barley, peas, beans, and grapes. These are familiar to us precisely
because they also spread into climatically similar adjacent areas of Europe,
thence to America and Australia, and became some of the staple crops of
temperate-zone agriculture around the world.
As one travels south in Africa across the Saharan desert and reencounters
rain in the Sahel zone just south of the desert, one notices that Sahel rains fall
in the summer rather than in the winter. Even if Fertile Crescent crops adapted
to winter rain could somehow have crossed the Sahara, they would have been
difficult to grow in the summer-rain Sahel zone. Instead, we find two sets of
African crops whose wild ancestors occur just south of the Sahara, and which
are adapted to summer rains and less seasonal variation in day length. One set
consists of plants whose ancestors are widely distributed from west to east
across the Sahel zone and were probably domesticated there. They include,
notably, sorghum and pearl millet, which became the staple cereals of much
of sub-Saharan Africa. Sorghum proved so valuable that it is now grown in
areas with hot, dry climates on all the continents, including in the United
States.
The other set consists of plants whose wild ancestors occur in Ethiopia
and were probably domesticated there in the highlands. Most are still grown
mainly just in Ethiopia and remain unknown to Americans—including
Ethiopia’s narcotic chat, its banana-like ensete, its oily noog, its finger millet
used to brew its national beer, and its tiny-seeded cereal called teff, used to
make its national bread. But every reader addicted to coffee can thank ancient
Ethiopian farmers for domesticating the coffee plant. It remained confined to
Ethiopia until it caught on in Arabia and then around the world, to sustain
today the economies of countries as far-flung as Brazil and Papua New
Guinea.
The next-to-last set of African crops arose from wild ancestors in the wet
climate of West Africa. Some, including African rice, have remained virtually
confined there; others, such as African yams, spread throughout other areas of
sub-Saharan Africa; and two, the oil palm and kola nut, reached other
continents. West Africans were chewing the caffeine-containing nuts of the
latter as a narcotic, long before the Coca-Cola Company enticed first
Americans and then the world to drink a beverage originally laced with its
extracts.
The last batch of African crops is also adapted to wet climates but
provides the biggest surprise of Figure 19.3. Bananas, Asian yams, and taro
were already widespread in sub-Saharan Africa in the 1400s, and Asian rice
was established on the coast of East Africa. But those crops originated in
tropical Southeast Asia. Their presence in Africa would astonish us, if the
presence of Indonesian people on Madagascar had not already alerted us to
Africa’s prehistoric Asian connection. Did Austronesians sailing from Borneo
land on the East African coast, bestow their crops on grateful African farmers,
pick up African fishermen, and sail off into the sunrise to colonize
Madagascar, leaving no other Austronesian traces in Africa?
The remaining surprise is that all of Africa’s indigenous crops—those of
the Sahel, Ethiopia, and West Africa—originated north of the equator. Not a
single African crop originated south of it. This already gives us a hint why
speakers of Niger-Congo languages, stemming from north of the equator,
were able to displace Africa’s equatorial Pygmies and subequatorial Khoisan
people. The failure of the Khoisan and Pygmies to develop agriculture was
due not to any inadequacy of theirs as farmers but merely to the accident that
southern Africa’s wild plants were mostly unsuitable for domestication.
Neither Bantu nor white farmers, heirs to thousands of years of farming
experience, were subsequently able to develop southern African native plants
into food crops.
Africa’s domesticated animal species can be summarized much more
quickly than its plants, because there are so few of them. The sole animal that
we know for sure was domesticated in Africa, because its wild ancestor is
confined there, is a turkeylike bird called the guinea fowl. Wild ancestors of
domestic cattle, donkeys, pigs, dogs, and house cats were native to North
Africa but also to Southwest Asia, so we can’t yet be certain where they were
first domesticated, although the earliest dates currently known for domestic
donkeys and house cats favor Egypt. Recent evidence suggests that cattle may
have been domesticated independently in North Africa, Southwest Asia, and
India, and that all three of those stocks have contributed to modern African
cattle breeds. Otherwise, all the remainder of Africa’s domestic mammals
must have been domesticated elsewhere and introduced as domesticates to
Africa, because their wild ancestors occur only in Eurasia. Africa’s sheep and
goats were domesticated in Southwest Asia, its chickens in Southeast Asia, its
horses in southern Russia, and its camels probably in Arabia.
The most unexpected feature of this list of African domestic animals is
again a negative one. The list includes not a single one of the big wild
mammal species for which Africa is famous and which it possesses in such
abundance—its zebras and wildebeests, its rhinos and hippos, its giraffes and
buffalo. As we’ll see, that reality was as fraught with consequences for
African history as was the absence of native domestic plants in subequatorial
Africa.
This quick tour through Africa’s food staples suffices to show that some
of them traveled a long way from their points of origin, both inside and
outside Africa. In Africa as elsewhere in the world, some peoples were much
“luckier” than others, in the suites of domesticable wild plant and animal
species that they inherited from their environment. By analogy with the
engulfing of Aboriginal Australian hunter-gatherers by British colonists fed
on wheat and cattle, we have to suspect that some of the “lucky” Africans
parlayed their advantage into engulfing their African neighbors. Now, at last,
let’s turn to the archaeological record to find out who engulfed whom when.
WHAT CAN ARCHAEOLOGY can tell us about actual dates and places for the
rise of farming and herding in Africa? Any reader steeped in the history of
Western civilization would be forgiven for assuming that African food
production began in ancient Egypt’s Nile Valley, land of the pharaohs and
pyramids. After all, Egypt by 3000 B.C. was undoubtedly the site of Africa’s
most complex society, and one of the world’s earliest centers of writing. In
fact, though, possibly the earliest archaeological evidence for food production
in Africa comes instead from the Sahara.
Today, of course, much of the Sahara is so dry that it cannot support even
grass. But between about 9000 and 4000 B.C. the Sahara was more humid,
held numerous lakes, and teemed with game. In that period, Saharans began
to tend cattle and make pottery, then to keep sheep and goats, and they may
also have been starting to domesticate sorghum and millet. Saharan
pastoralism precedes the earliest known date (5200 B.C.) for the arrival of food
production in Egypt, in the form of a full package of Southwest Asian winter
crops and livestock. Food production also arose in West Africa and Ethiopia,
and by around 2500 B.C. cattle herders had already crossed the modern border
from Ethiopia into northern Kenya.
While those conclusions rest on archaeological evidence, there is also an
independent method for dating the arrival of domestic plants and animals: by
comparing the words for them in modern languages. Comparisons of terms
for plants in southern Nigerian languages of the Niger-Congo family show
that the words fall into three groups. First are cases in which the word for a
particular crop is very similar in all those southern Nigerian languages. Those
crops prove to be ones like West African yams, oil palm, and kola nut—plants
that were already believed on botanical and other evidence to be native to
West Africa and first domesticated there. Since those are the oldest West
African crops, all modern southern Nigerian languages inherited the same
original set of words for them.
Next come crops whose names are consistent only among the languages
falling within a small subgroup of those southern Nigerian languages. Those
crops turn out to be ones believed to be of Indonesian origin, such as bananas
and Asian yams. Evidently, those crops reached southern Nigeria only after
languages began to break up into subgroups, so each subgroup coined or
received different names for the new plants, which the modern languages of
only that particular subgroup inherited. Last come crop names that aren’t
consistent within language groups at all, but instead follow trade routes.
These prove to be New World crops like corn and peanuts, which we know
were introduced into Africa after the beginnings of transatlantic ship traffic
(A.D. 1492) and diffused since then along trade routes, often bearing their
Portuguese or other foreign names.
Thus, even if we possessed no botanical or archaeological evidence
whatsoever, we would still be able to deduce from the linguistic evidence
alone that native West African crops were domesticated first, that Indonesian
crops arrived next, and that finally the European introductions came in. The
UCLA historian Christopher Ehret has applied this linguistic approach to
determining the sequence in which domestic plants and animals became
utilized by the people of each African language family. By a method termed
glottochronology, based on calculations of how rapidly words tend to change
over historical time, comparative linguistics can even yield estimated dates
for domestications or crop arrivals.
Putting together direct archaeological evidence of crops with the more
indirect linguistic evidence, we deduce that the people who were
domesticating sorghum and millet in the Sahara thousands of years ago spoke
languages ancestral to modern Nilo-Saharan languages. Similarly, the people
who first domesticated wet-country crops of West Africa spoke languages
ancestral to the modern Niger-Congo languages. Finally, speakers of ancestral
Afroasiatic languages may have been involved in domesticating the crops
native to Ethiopia, and they certainly introduced Fertile Crescent crops to
North Africa.
Thus, the evidence derived from plant names in modern African
languages permits us to glimpse the existence of three languages being
spoken in Africa thousands of years ago: ancestral Nilo-Saharan, ancestral
Niger-Congo, and ancestral Afroasiatic. In addition, we can glimpse the
existence of ancestral Khoisan from other linguistic evidence, though not that
of crop names (because ancestral Khoisan people domesticated no crops).
Now surely, since Africa harbors 1,500 languages today, it is big enough to
have harbored more than four ancestral languages thousands of years ago. But
all those other languages must have disappeared—either because the people
speaking them survived but lost their original language, like the Pygmies, or
because the people themselves disappeared.
The survival of modern Africa’s four native language families (that is, the
four other than the recently arrived Austronesian language of Madagascar)
isn’t due to the intrinsic superiority of those languages as vehicles for
communication. Instead, it must be attributed to a historical accident:
ancestral speakers of Nilo-Saharan, Niger-Congo, and Afroasiatic happened
to be living at the right place and time to acquire domestic plants and animals,
which let them multiply and either replace other peoples or impose their
language. The few modern Khoisan speakers survived mainly because of their
isolation in areas of southern Africa unsuitable for Bantu farming.
BEFORE WE TRACE Khoisan survival beyond the Bantu tide, let’s see what
archaeology tells us about Africa’s other great prehistoric population
movement—the Austronesian colonization of Madagascar. Archaeologists
exploring Madagascar have now proved that Austronesians had arrived at
least by A.D. 800, possibly as early as A.D. 300. There the Austronesians
encountered (and proceeded to exterminate) a strange world of living animals
as distinctive as if they had come from another planet, because those animals
had evolved on Madagascar during its long isolation. They included giant
elephant birds, primitive primates called lemurs as big as gorillas, and pygmy
hippos. Archaeological excavations of the earliest human settlements on
Madagascar yield remains of iron tools, livestock, and crops, so the colonists
were not just a small canoeload of fishermen blown off course; they formed a
full-fledged expedition. How did that prehistoric 4,000-mile expedition come
about?
One hint is in an ancient book of sailors’ directions, the Periplus of the
Erythrean Sea, written by an anonymous merchant living in Egypt around A.D.
100. The merchant describes an already thriving sea trade connecting India
and Egypt with the coast of East Africa. With the spread of Islam after A.D.
800, Indian Ocean trade becomes well documented archaeologically by
copious quantities of Mideastern (and occasionally even Chinese!) products
such as pottery, glass, and porcelain in East African coastal settlements. The
traders waited for favorable winds to let them cross the Indian Ocean directly
between East Africa and India. When the Portuguese navigator Vasco da
Gama became the first European to sail around the southern cape of Africa
and reached the Kenya coast in 1498, he encountered Swahili trading
settlements and picked up a pilot who guided him on that direct route to India.
But there was an equally vigorous sea trade from India eastward, between
India and Indonesia. Perhaps the Austronesian colonists of Madagascar
reached India from Indonesia by that eastern trade route and then fell in with
the westward trade route to East Africa, where they joined with Africans and
discovered Madagascar. That union of Austronesians and East Africans lives
on today in Madagascar’s basically Austronesian language, which contains
loan words from coastal Kenyan Bantu languages. But there are no
corresponding Austronesian loan words in Kenyan languages, and other
traces of Austronesians are very thin on the ground in East Africa: mainly just
Africa’s possible legacy of Indonesian musical instruments (xylophones and
zithers) and, of course, the Austronesian crops that became so important in
African agriculture. Hence one wonders whether Austronesians, instead of
taking the easier route to Madagascar via India and East Africa, somehow
(incredibly) sailed straight across the Indian Ocean, discovered Madagascar,
and only later got plugged into East African trade routes. Thus, some mystery
remains about Africa’s most surprising fact of human geography.
WHAT CAN ARCHAEOLOGY tell us about the other great population movement
in recent African prehistory—the Bantu expansion? We saw from the twin
evidence of modern peoples and their languages that sub-Saharan Africa was
not always a black continent, as we think of it today. Instead, this evidence
suggested that Pygmies had once been widespread in the rain forest of Central
Africa, while Khoisan peoples had been widespread in drier parts of
subequatorial Africa. Can archaeology test those assumptions?
In the case of the Pygmies, the answer is “not yet,” merely because
archaeologists have yet to discover ancient human skeletons from the Central
African forests. For the Khoisan, the answer is “yes.” In Zambia, to the north
of the modern Khoisan range, archaeologists have found skulls of people
possibly resembling the modern Khoisan, as well as stone tools resembling
those that Khoisan peoples were still making in southern Africa at the time
Europeans arrived.
As for how the Bantu came to replace those northern Khoisan,
archaeological and linguistic evidence suggest that the expansion of ancestral
Bantu farmers from West Africa’s inland savanna south into its wetter coastal
forest may have begun as early as 3000 B.C. (Figure 19.4). Words still
widespread in all Bantu languages show that, already then, the Bantu had
cattle and wet-climate crops such as yams, but that they lacked metal and
were still engaged in much fishing, hunting, and gathering. They even lost
their cattle to disease borne by tsetse flies in the forest. As they spread into
the Congo Basin’s equatorial forest zone, cleared gardens, and increased in
numbers, they began to engulf the Pygmy hunter-gatherers and compress
them into the forest itself.
By soon after 1000 B.C. the Bantu had emerged from the eastern side of
the forest into the more open country of East Africa’s Rift Valley and Great
Lakes. Here they encountered a melting pot of Afroasiatic and Nilo-Saharan
farmers and herders growing millet and sorghum and raising livestock in drier
areas, along with Khoisan hunter-gatherers. Thanks to their wet-climate crops
inherited from their West African homeland, the Bantu were able to farm in
wet areas of East Africa unsuitable for all those previous occupants. By the
last centuries B.C. the advancing Bantu had reached the East African coast.
In East Africa the Bantu began to acquire millet and sorghum (along with
the Nilo-Saharan names for those crops), and to reacquire cattle, from their
Nilo-Saharan and Afroasiatic neighbors. They also acquired iron, which had
just begun to be smelted in Africa’s Sahel zone. The origins of ironworking in
sub-Saharan Africa soon after 1000 B.C. are still unclear. That early date is
suspiciously close to dates for the arrival of Near Eastern ironworking
techniques in Carthage, on the North African coast. Hence historians often
assume that knowledge of metallurgy reached sub-Saharan Africa from the
north. On the other hand, copper smelting had been going on in the West
African Sahara and Sahel since at least 2000 B.C. That could have been the
precursor to an independent African discovery of iron metallurgy.
Strengthening that hypothesis, the iron-smelting techniques of smiths in sub-
Saharan Africa were so different from those of the Mediterranean as to
suggest independent development: African smiths discovered how to produce
high temperatures in their village furnaces and manufacture steel over 2,000
years before the Bessemer furnaces of 19th-century Europe and America.
With the addition of iron tools to their wet-climate crops, the Bantu had
finally put together a military-industrial package that was unstoppable in the
subequatorial Africa of the time. In East Africa they still had to compete
against numerous Nilo-Saharan and Afroasiatic Iron Age farmers. But to the
south lay 2,000 miles of country thinly occupied by Khoisan hunter-gatherers,
lacking iron and crops. Within a few centuries, in one of the swiftest
colonizing advances of recent prehistory, Bantu farmers had swept all the way
to Natal, on the east coast of what is now South Africa.
It’s easy to oversimplify what was undoubtedly a rapid and dramatic
expansion, and to picture all Khoisan in the way being trampled by onrushing
Bantu hordes. In reality, things were more complicated. Khoisan peoples of
southern Africa had already acquired sheep and cattle a few centuries ahead
of the Bantu advance. The first Bantu pioneers probably were few in number,
selected wet-forest areas suitable for their yam agriculture, and leapfrogged
over drier areas, which they left to Khoisan herders and hunter-gatherers.
Trading and marriage relationships were undoubtedly established between
those Khoisan and the Bantu farmers, each occupying different adjacent
habitats, just as Pygmy hunter-gatherers and Bantu farmers still do today in
equatorial Africa. Only gradually, as the Bantu multiplied and incorporated
cattle and dry-climate cereals into their economy, did they fill in the
leapfrogged areas. But the eventual result was still the same: Bantu farmers
occupying most of the former Khoisan realm; the legacy of those former
Khoisan inhabitants reduced to clicks in scattered non-Khoisan languages, as
well as buried skulls and stone tools waiting for archaeologists to discover;
and the Khoisan-like appearance of some southern African Bantu peoples.
What actually happened to all those vanished Khoisan populations? We
don’t know. All we can say for sure is that, in places where Khoisan peoples
had lived for perhaps tens of thousands of years, there are now Bantu. We can
only venture a guess, by analogy with witnessed events in modern times when
steel-toting white farmers collided with stone tool–using hunter-gatherers of
Aboriginal Australia and Indian California. There, we know that hunter-
gatherers were rapidly eliminated in a combination of ways: they were driven
out, men were killed or enslaved, women were appropriated as wives, and
both sexes became infected with epidemics of the farmers’ diseases. An
example of such a disease in Africa is malaria, which is borne by mosquitoes
that breed around farmers’ villages, and to which the invading Bantu had
already developed genetic resistance but Khoisan hunter-gatherers probably
had not.
However, Figure 19.1, of recent African human distributions, reminds us
that the Bantu did not overrun all the Khoisan, who did survive in southern
African areas unsuitable for Bantu agriculture. The southernmost Bantu
people, the Xhosa, stopped at the Fish River on South Africa’s south coast,
500 miles east of Cape Town. It’s not that the Cape of Good Hope itself is too
dry for agriculture: it is, after all, the breadbasket of modern South Africa.
Instead, the Cape has a Mediterranean climate of winter rains, in which the
Bantu summer-rain crops do not grow. By 1652, the year the Dutch arrived at
Cape Town with their winter-rain crops of Near Eastern origin, the Xhosa had
still not spread beyond the Fish River.
That seeming detail of plant geography had enormous implications for
politics today. One consequence was that, once South African whites had
quickly killed or infected or driven off the Cape’s Khoisan population, whites
could claim correctly that they had occupied the Cape before the Bantu and
thus had prior rights to it. That claim needn’t be taken seriously, since the
prior rights of the Cape Khoisan didn’t inhibit whites from dispossessing
them. The much heavier consequence was that the Dutch settlers in 1652 had
to contend only with a sparse population of Khoisan herders, not with a dense
population of steel-equipped Bantu farmers. When whites finally spread east
to encounter the Xhosa at the Fish River in 1702, a period of desperate
fighting began. Even though Europeans by then could supply troops from
their secure base at the Cape, it took nine wars and 175 years for their armies,
advancing at an average rate of less than one mile per year, to subdue the
Xhosa. How could whites have succeeded in establishing themselves at the
Cape at all, if those first few arriving Dutch ships had faced such fierce
resistance?
Thus, the problems of modern South Africa stem at least in part from a
geographic accident. The homeland of the Cape Khoisan happened to contain
few wild plants suitable for domestication; the Bantu happened to inherit
summer-rain crops from their ancestors of 5,000 years ago; and Europeans
happened to inherit winter-rain crops from their ancestors of nearly 10,000
years ago. Just as the sign “Goering Street” in the capital of newly
independent Namibia reminded me, Africa’s past has stamped itself deeply on
Africa’s present.
THAT’S HOW THE Bantu were able to engulf the Khoisan, instead of vice
versa. Now let’s turn to the remaining question in our puzzle of African
prehistory: why Europeans were the ones to colonize sub-Saharan Africa.
That it was not the other way around is especially surprising, because Africa
was the sole cradle of human evolution for millions of years, as well as
perhaps the homeland of anatomically modern Homo sapiens. To these
advantages of Africa’s enormous head start were added those of highly
diverse climates and habitats and of the world’s highest human diversity. An
extraterrestrial visiting Earth 10,000 years ago might have been forgiven for
predicting that Europe would end up as a set of vassal states of a sub-Saharan
African empire.
The proximate reasons behind the outcome of Africa’s collision with
Europe are clear. Just as in their encounter with Native Americans, Europeans
entering Africa enjoyed the triple advantage of guns and other technology,
widespread literacy, and the political organization necessary to sustain
expensive programs of exploration and conquest. Those advantages
manifested themselves almost as soon as the collisions started: barely four
years after Vasco da Gama first reached the East African coast, in 1498, he
returned with a fleet bristling with cannons to compel the surrender of East
Africa’s most important port, Kilwa, which controlled the Zimbabwe gold
trade. But why did Europeans develop those three advantages before sub-
Saharan Africans could?
As we have discussed, all three arose historically from the development of
food production. But food production was delayed in sub-Saharan Africa
(compared with Eurasia) by Africa’s paucity of domesticable native animal
and plant species, its much smaller area suitable for indigenous food
production, and its north–south axis, which retarded the spread of food
production and inventions. Let’s examine how those factors operated.
First, as regards domestic animals, we’ve already seen that those of sub-
Saharan Africa came from Eurasia, with the possible exception of a few from
North Africa. As a result, domestic animals did not reach sub-Saharan Africa
until thousands of years after they began to be utilized by emerging Eurasian
civilizations. That’s initially surprising, because we think of Africa as the
continent of big wild mammals. But we saw in Chapter 9 that a wild animal,
to be domesticated, must be sufficiently docile, submissive to humans, cheap
to feed, and immune to diseases and must grow rapidly and breed well in
captivity. Eurasia’s native cows, sheep, goats, horses, and pigs were among
the world’s few large wild animal species to pass all those tests. Their African
equivalents—such as the African buffalo, zebra, bush pig, rhino, and
hippopotamus—have never been domesticated, not even in modern times.
It’s true, of course, that some large African animals have occasionally
been tamed. Hannibal enlisted tamed African elephants in his unsuccessful
war against Rome, and ancient Egyptians may have tamed giraffes and other
species. But none of those tamed animals was actually domesticated—that is,
selectively bred in captivity and genetically modified so as to become more
useful to humans. Had Africa’s rhinos and hippos been domesticated and
ridden, they would not only have fed armies but also have provided an
unstoppable cavalry to cut through the ranks of European horsemen. Rhino-
mounted Bantu shock troops could have overthrown the Roman Empire. It
never happened.
A second factor is a corresponding, though less extreme, disparity
between sub-Saharan Africa and Eurasia in domesticable plants. The Sahel,
Ethiopia, and West Africa did yield indigenous crops, but many fewer
varieties than grew in Eurasia. Because of the limited variety of wild starting
material suitable for plant domestication, even Africa’s earliest agriculture
may have begun several thousand years later than that of the Fertile Crescent.
Thus, as far as plant and animal domestication was concerned, the head
start and high diversity lay with Eurasia, not with Africa. A third factor is that
Africa’s area is only about half that of Eurasia. Furthermore, only about one-
third of its area falls within the sub-Saharan zone north of the equator that
was occupied by farmers and herders before 1000 B.C. Today, the total
population of Africa is less than 700 million, compared with 4 billion for
Eurasia. But, all other things being equal, more land and more people mean
more competing societies and inventions, hence a faster pace of development.
The remaining factor behind Africa’s slower rate of post-Pleistocene
development compared with Eurasia’s is the different orientation of the main
axes of these continents. Like that of the Americas, Africa’s major axis is
north–south, whereas Eurasia’s is east–west (Figure 10.1). As one moves
along a north–south axis, one traverses zones differing greatly in climate,
habitat, rainfall, day length, and diseases of crops and livestock. Hence, crops
and animals domesticated or acquired in one part of Africa had great
difficulty in moving to other parts. In contrast, crops and animals moved
easily between Eurasian societies thousands of miles apart but at the same
latitude and sharing similar climates and day lengths.
The slow passage or complete halt of crops and livestock along Africa’s
north–south axis had important consequences. For example, the
Mediterranean crops that became Egypt’s staples require winter rains and
seasonal variation in day length for their germination. Those crops were
unable to spread south of the Sudan, beyond which they encountered summer
rains and little or no seasonal variation in daylight. Egypt’s wheat and barley
never reached the Mediterranean climate at the Cape of Good Hope until
European colonists brought them in 1652, and the Khoisan never developed
agriculture. Similarly, the Sahel crops adapted to summer rain and to little or
no seasonal variation in day length were brought by the Bantu into southern
Africa but could not grow at the Cape itself, thereby halting the advance of
Bantu agriculture. Bananas and other tropical Asian crops for which Africa’s
climate is eminently suitable, and which today are among the most productive
staples of tropical African agriculture, were unable to reach Africa by land
routes. They apparently did not arrive until the first millennium A.D., long
after their domestication in Asia, because they had to wait for large-scale boat
traffic across the Indian Ocean.
Africa’s north–south axis also seriously impeded the spread of livestock.
Equatorial Africa’s tsetse flies, carrying trypanosomes to which native
African wild mammals are resistant, proved devastating to introduced
Eurasian and North African species of livestock. The cows that the Bantu
acquired from the tsetse-free Sahel zone failed to survive the Bantu expansion
through the equatorial forest. Although horses had already reached Egypt
around 1800 B.C. and transformed North African warfare soon thereafter, they
did not cross the Sahara to drive the rise of cavalry-mounted West African
kingdoms until the first millennium A.D., and they never spread south through
the tsetse fly zone. While cattle, sheep, and goats had already reached the
northern edge of the Serengeti in the third millennium B.C., it took more than
2,000 years beyond that for livestock to cross the Serengeti and reach
southern Africa.
Similarly slow in spreading down Africa’s north–south axis was human
technology. Pottery, recorded in the Sudan and Sahara around 8000 B.C., did
not reach the Cape until around A.D. 1. Although writing developed in Egypt
by 3000 B.C. and spread in an alphabetized form to the Nubian kingdom of
Meroe, and although alphabetic writing reached Ethiopia (possibly from
Arabia), writing did not arise independently in the rest of Africa, where it was
instead brought in from the outside by Arabs and Europeans.
In short, Europe’s colonization of Africa had nothing to do with
differences between European and African peoples themselves, as white
racists assume. Rather, it was due to accidents of geography and
biogeography—in particular, to the continents’ different areas, axes, and
suites of wild plant and animal species. That is, the different historical
trajectories of Africa and Europe stem ultimately from differences in real
estate.
EPILOGUE
THE FUTURE OF HUMAN HISTORY AS A
SCIENCE
YALI’S QUESTION WENT TO THE HEART OF THE CURRENT human condition, and
of post-Pleistocene human history. Now that we have completed this brief
tour over the continents, how shall we answer Yali?
I would say to Yali: the striking differences between the long-term
histories of peoples of the different continents have been due not to innate
differences in the peoples themselves but to differences in their environments.
I expect that if the populations of Aboriginal Australia and Eurasia could have
been interchanged during the Late Pleistocene, the original Aboriginal
Australians would now be the ones occupying most of the Americas and
Australia, as well as Eurasia, while the original Aboriginal Eurasians would
be the ones now reduced to downtrodden population fragments in Australia.
One might at first be inclined to dismiss this assertion as meaningless,
because the experiment is imaginary and my claim about its outcome cannot
be verified. But historians are nevertheless able to evaluate related hypotheses
by retrospective tests. For instance, one can examine what did happen when
European farmers were transplanted to Greenland or the U.S. Great Plains,
and when farmers stemming ultimately from China emigrated to the Chatham
Islands, the rain forests of Borneo, or the volcanic soils of Java or Hawaii.
These tests confirm that the same ancestral peoples either ended up extinct, or
returned to living as hunter-gatherers, or went on to build complex states,
depending on their environments. Similarly, Aboriginal Australian hunter-
gatherers, variously transplanted to Flinders Island, Tasmania, or southeastern
Australia, ended up extinct, or as hunter-gatherers with the modern world’s
simplest technology, or as canal builders intensively managing a productive
fishery, depending on their environments.
Of course, the continents differ in innumerable environmental features
affecting trajectories of human societies. But a mere laundry list of every
possible difference does not constitute an answer to Yali’s question. Just four
sets of differences appear to me to be the most important ones.
The first set consists of continental differences in the wild plant and
animal species available as starting materials for domestication. That’s
because food production was critical for the accumulation of food surpluses
that could feed non-food-producing specialists, and for the buildup of large
populations enjoying a military advantage through mere numbers even before
they had developed any technological or political advantage. For both of
those reasons, all developments of economically complex, socially stratified,
politically centralized societies beyond the level of small nascent chiefdoms
were based on food production.
But most wild animal and plant species have proved unsuitable for
domestication: food production has been based on relatively few species of
livestock and crops. It turns out that the number of wild candidate species for
domestication varied greatly among the continents, because of differences in
continental areas and also (in the case of big mammals) in Late Pleistocene
extinctions. These extinctions were much more severe in Australia and the
Americas than in Eurasia or Africa. As a result, Africa ended up biologically
somewhat less well endowed than the much larger Eurasia, the Americas still
less so, and Australia even less so, as did Yali’s New Guinea (with one-
seventieth of Eurasia’s area and with all of its original big mammals extinct in
the Late Pleistocene).
On each continent, animal and plant domestication was concentrated in a
few especially favorable homelands accounting for only a small fraction of
the continent’s total area. In the case of technological innovations and
political institutions as well, most societies acquire much more from other
societies than they invent themselves. Thus, diffusion and migration within a
continent contribute importantly to the development of its societies, which
tend in the long run to share each other’s developments (insofar as
environments permit) because of the processes illustrated in such simple form
by Maori New Zealand’s Musket Wars. That is, societies initially lacking an
advantage either acquire it from societies possessing it or (if they fail to do
so) are replaced by those other societies.
Hence a second set of factors consists of those affecting rates of diffusion
and migration, which differed greatly among continents. They were most
rapid in Eurasia, because of its east–west major axis and its relatively modest
ecological and geographical barriers. The reasoning is straightforward for
movements of crops and livestock, which depend strongly on climate and
hence on latitude. But similar reasoning also applies to the diffusion of
technological innovations, insofar as they are best suited without modification
to specific environments. Diffusion was slower in Africa and especially in the
Americas, because of those continents’ east–west major axes and geographic
and ecological barriers. It was also difficult in traditional New Guinea, where
rugged terrain and the long backbone of high mountains prevented any
significant progress toward political and linguistic unification.
Related to these factors affecting diffusion within continents is a third set
of factors influencing diffusion between continents, which may also help
build up a local pool of domesticates and technology. Ease of intercontinental
diffusion has varied, because some continents are more isolated than others.
Within the last 6,000 years it has been easiest from Eurasia to sub-Saharan
Africa, supplying most of Africa’s species of livestock. But interhemispheric
diffusion made no contribution to Native America’s complex societies,
isolated from Eurasia at low latitudes by broad oceans, and at high latitudes
by geography and by a climate suitable just for hunting-gathering. To
Aboriginal Australia, isolated from Eurasia by the water barriers of the
Indonesian Archipelago, Eurasia’s sole proven contribution was the dingo.
The fourth and last set of factors consists of continental differences in
area or total population size. A larger area or population means more potential
inventors, more competing societies, more innovations available to adopt—
and more pressure to adopt and retain innovations, because societies failing to
do so will tend to be eliminated by competing societies. That fate befell
African pygmies and many other hunter-gatherer populations displaced by
farmers. Conversely, it also befell the stubborn, conservative Greenland Norse
farmers, replaced by Eskimo hunter-gatherers whose subsistence methods and
technology were far superior to those of the Norse under Greenland
conditions. Among the world’s landmasses, area and the number of
competing societies were largest for Eurasia, much smaller for Australia and
New Guinea and especially for Tasmania. The Americas, despite their large
aggregate area, were fragmented by geography and ecology and functioned
effectively as several poorly connected smaller continents.
Those four sets of factors constitute big environmental differences that
can be quantified objectively and that are not subject to dispute. While one
can contest my subjective impression that New Guineans are on the average
smarter than Eurasians, one cannot deny that New Guinea has a much smaller
area and far fewer big animal species than Eurasia. But mention of these
environmental differences invites among historians the label “geographic
determinism,” which raises hackles. The label seems to have unpleasant
connotations, such as that human creativity counts for nothing, or that we
humans are passive robots helplessly programmed by climate, fauna, and
flora. Of course these fears are misplaced. Without human inventiveness, all
of us today would still be cutting our meat with stone tools and eating it raw,
like our ancestors of a million years ago. All human societies contain
inventive people. It’s just that some environments provide more starting
materials, and more favorable conditions for utilizing inventions, than do
other environments.
THESE ANSWERS TO Yali’s question are longer and more complicated than
Yali himself would have wanted. Historians, however, may find them too
brief and oversimplified. Compressing 13,000 years of history on all
continents into a 400-page book works out to an average of about one page
per continent per 150 years, making brevity and simplification inevitable. Yet
the compression brings a compensating benefit: long-term comparisons of
regions yield insights that cannot be won from short-term studies of single
societies.
Naturally, a host of issues raised by Yali’s question remain unresolved. At
present, we can put forward some partial answers plus a research agenda for
the future, rather than a fully developed theory. The challenge now is to
develop human history as a science, on a par with acknowledged historical
sciences such as astronomy, geology, and evolutionary biology. Hence it
seems appropriate to conclude this book by looking to the future of the
discipline of history, and by outlining some of the unresolved issues.
The most straightforward extension of this book will be to quantify
further, and thus to establish more convincingly the role of, intercontinental
differences in the four sets of factors that appear to be most important. To
illustrate differences in starting materials for domestication, I provided
numbers for each continent’s total of large wild terrestrial mammalian
herbivores and omnivores (Table 9.2) and of large-seeded cereals (Table 8.1).
One extension would be to assemble corresponding numbers for large-seeded
legumes (pulses), such as beans, peas, and vetches. In addition, I mentioned
factors disqualifying big mammalian candidates for domestication, but I did
not tabulate how many candidates are disqualified by each factor on each
continent. It would be interesting to do so, especially for Africa, where a
higher percentage of candidates is disqualified than in Eurasia: which
disqualifying factors are most important in Africa, and what has selected for
their high frequency in African mammals? Quantitative data should also be
assembled to test my preliminary calculations suggesting differing rates of
diffusion along the major axes of Eurasia, the Americas, and Africa.
A SECOND EXTENSION will be to smaller geographic scales and shorter time
scales than those of this book. For instance, the following obvious question
has probably occurred to readers already: why, within Eurasia, were European
societies, rather than those of the Fertile Crescent or China or India, the ones
that colonized America and Australia, took the lead in technology, and
became politically and economically dominant in the modern world? A
historian who had lived at any time between 8500 B.C. and A.D. 1450, and who
had tried then to predict future historical trajectories, would surely have
labeled Europe’s eventual dominance as the least likely outcome, because
Europe was the most backward of those three Old World regions for most of
those 10,000 years. From 8500 B.C. until the rise of Greece and then Italy after
500 B.C., almost all major innovations in western Eurasia—animal
domestication, plant domestication, writing, metallurgy, wheels, states, and so
on—arose in or near the Fertile Crescent. Until the proliferation of water mills
after about A.D. 900, Europe west or north of the Alps contributed nothing of
significance to Old World technology or civilization; it was instead a recipient
of developments from the eastern Mediterranean, Fertile Crescent, and China.
Even from A.D. 1000 to 1450 the flow of science and technology was
predominantly into Europe from the Islamic societies stretching from India to
North Africa, rather than vice versa. During those same centuries China led
the world in technology, having launched itself on food production nearly as
early as the Fertile Crescent did.
Why, then, did the Fertile Crescent and China eventually lose their
enormous leads of thousands of years to late-starting Europe? One can, of
course, point to proximate factors behind Europe’s rise: its development of a
merchant class, capitalism, and patent protection for inventions, its failure to
develop absolute despots and crushing taxation, and its Greco-Judeo-Christian
tradition of critical empirical inquiry. Still, for all such proximate causes one
must raise the question of ultimate cause: why did these proximate factors
themselves arise in Europe, rather than in China or the Fertile Crescent?
For the Fertile Crescent, the answer is clear. Once it had lost the head start
that it had enjoyed thanks to its locally available concentration of
domesticable wild plants and animals, the Fertile Crescent possessed no
further compelling geographic advantages. The disappearance of that head
start can be traced in detail, as the westward shift in powerful empires. After
the rise of Fertile Crescent states in the fourth millennium B.C., the center of
power initially remained in the Fertile Crescent, rotating between empires
such as those of Babylon, the Hittites, Assyria, and Persia. With the Greek
conquest of all advanced societies from Greece east to India under Alexander
the Great in the late fourth century B.C., power finally made its first shift
irrevocably westward. It shifted farther west with Rome’s conquest of Greece
in the second century B.C., and after the fall of the Roman Empire it eventually
moved again, to western and northern Europe.
The major factor behind these shifts becomes obvious as soon as one
compares the modern Fertile Crescent with ancient descriptions of it. Today,
the expressions “Fertile Crescent” and “world leader in food production” are
absurd. Large areas of the former Fertile Crescent are now desert, semidesert,
steppe, or heavily eroded or salinized terrain unsuited for agriculture. Today’s
ephemeral wealth of some of the region’s nations, based on the single
nonrenewable resource of oil, conceals the region’s long-standing
fundamental poverty and difficulty in feeding itself.
In ancient times, however, much of the Fertile Crescent and eastern
Mediterranean region, including Greece, was covered with forest. The
region’s transformation from fertile woodland to eroded scrub or desert has
been elucidated by paleobotanists and archaeologists. Its woodlands were
cleared for agriculture, or cut to obtain construction timber, or burned as
firewood or for manufacturing plaster. Because of low rainfall and hence low
primary productivity (proportional to rainfall), regrowth of vegetation could
not keep pace with its destruction, especially in the presence of overgrazing
by abundant goats. With the tree and grass cover removed, erosion proceeded
and valleys silted up, while irrigation agriculture in the low-rainfall
environment led to salt accumulation. These processes, which began in the
Neolithic era, continued into modern times. For instance, the last forests near
the ancient Nabataean capital of Petra, in modern Jordan, were felled by the
Ottoman Turks during construction of the Hejaz railroad just before World
War I.
Thus, Fertile Crescent and eastern Mediterranean societies had the
misfortune to arise in an ecologically fragile environment. They committed
ecological suicide by destroying their own resource base. Power shifted
westward as each eastern Mediterranean society in turn undermined itself,
beginning with the oldest societies, those in the east (the Fertile Crescent).
Northern and western Europe has been spared this fate, not because its
inhabitants have been wiser but because they have had the good luck to live in
a more robust environment with higher rainfall, in which vegetation regrows
quickly. Much of northern and western Europe is still able to support
productive intensive agriculture today, 7,000 years after the arrival of food
production. In effect, Europe received its crops, livestock, technology, and
writing systems from the Fertile Crescent, which then gradually eliminated
itself as a major center of power and innovation.
That is how the Fertile Crescent lost its huge early lead over Europe. Why
did China also lose its lead? Its falling behind is initially surprising, because
China enjoyed undoubted advantages: a rise of food production nearly as
early as in the Fertile Crescent; ecological diversity from North to South
China and from the coast to the high mountains of the Tibetan plateau, giving
rise to a diverse set of crops, animals, and technology; a large and productive
expanse, nourishing the largest regional human population in the world; and
an environment less dry or ecologically fragile than the Fertile Crescent’s,
allowing China still to support productive intensive agriculture after nearly
10,000 years, though its environmental problems are increasing today and are
more serious than western Europe’s.
These advantages and head start enabled medieval China to lead the
world in technology. The long list of its major technological firsts includes
cast iron, the compass, gunpowder, paper, printing, and many others
mentioned earlier. It also led the world in political power, navigation, and
control of the seas. In the early 15th century it sent treasure fleets, each
consisting of hundreds of ships up to 400 feet long and with total crews of up
to 28,000, across the Indian Ocean as far as the east coast of Africa, decades
before Columbus’s three puny ships crossed the narrow Atlantic Ocean to the
Americas’ east coast. Why didn’t Chinese ships proceed around Africa’s
southern cape westward and colonize Europe, before Vasco da Gama’s own
three puny ships rounded the Cape of Good Hope eastward and launched
Europe’s colonization of East Asia? Why didn’t Chinese ships cross the
Pacific to colonize the Americas’ west coast? Why, in brief, did China lose its
technological lead to the formerly so backward Europe?
The end of China’s treasure fleets gives us a clue. Seven of those fleets
sailed from China between A.D. 1405 and 1433. They were then suspended as
a result of a typical aberration of local politics that could happen anywhere in
the world: a power struggle between two factions at the Chinese court (the
eunuchs and their opponents). The former faction had been identified with
sending and captaining the fleets. Hence when the latter faction gained the
upper hand in a power struggle, it stopped sending fleets, eventually
dismantled the shipyards, and forbade oceangoing shipping. The episode is
reminiscent of the legislation that strangled development of public electric
lighting in London in the 1880s, the isolationism of the United States between
the First and Second World Wars, and any number of backward steps in any
number of countries, all motivated by local political issues. But in China there
was a difference, because the entire region was politically unified. One
decision stopped fleets over the whole of China. That one temporary decision
became irreversible, because no shipyards remained to turn out ships that
would prove the folly of that temporary decision, and to serve as a focus for
rebuilding other shipyards.
Now contrast those events in China with what happened when fleets of
exploration began to sail from politically fragmented Europe. Christopher
Columbus, an Italian by birth, switched his allegiance to the duke of Anjou in
France, then to the king of Portugal. When the latter refused his request for
ships in which to explore westward, Columbus turned to the duke of Medina-
Sedonia, who also refused, then to the count of Medina-Celi, who did
likewise, and finally to the king and queen of Spain, who denied Columbus’s
first request but eventually granted his renewed appeal. Had Europe been
united under any one of the first three rulers, its colonization of the Americas
might have been stillborn.
In fact, precisely because Europe was fragmented, Columbus succeeded
on his fifth try in persuading one of Europe’s hundreds of princes to sponsor
him. Once Spain had thus launched the European colonization of America,
other European states saw the wealth flowing into Spain, and six more joined
in colonizing America. The story was the same with Europe’s cannon, electric
lighting, printing, small firearms, and innumerable other innovations: each
was at first neglected or opposed in some parts of Europe for idiosyncratic
reasons, but once adopted in one area, it eventually spread to the rest of
Europe.
These consequences of Europe’s disunity stand in sharp contrast to those
of China’s unity. From time to time the Chinese court decided to halt other
activities besides overseas navigation: it abandoned development of an
elaborate water-driven spinning machine, stepped back from the verge of an
industrial revolution in the 14th century, demolished or virtually abolished
mechanical clocks after leading the world in clock construction, and retreated
from mechanical devices and technology in general after the late 15th century.
Those potentially harmful effects of unity have flared up again in modern
China, notably during the madness of the Cultural Revolution in the 1960s
and 1970s, when a decision by one or a few leaders closed the whole
country’s school systems for five years.
China’s frequent unity and Europe’s perpetual disunity both have a long
history. The most productive areas of modern China were politically joined
for the first time in 221 B.C. and have remained so for most of the time since
then. China has had only a single writing system from the beginnings of
literacy, a single dominant language for a long time, and substantial cultural
unity for two thousand years. In contrast, Europe has never come remotely
close to political unification: it was still splintered into 1,000 independent
statelets in the 14th century, into 500 statelets in A.D. 1500, got down to a
minimum of 25 states in the 1980s, and is now up again to nearly 40 at the
moment that I write this sentence. Europe still has 45 languages, each with its
own modified alphabet, and even greater cultural diversity. The disagreements
that continue today to frustrate even modest attempts at European unification
through the European Economic Community (EEC) are symptomatic of
Europe’s ingrained commitment to disunity.
Hence the real problem in understanding China’s loss of political and
technological preeminence to Europe is to understand China’s chronic unity
and Europe’s chronic disunity. The answer is again suggested by maps (see
Backmatter). Europe has a highly indented coastline, with five large
peninsulas that approach islands in their isolation, and all of which evolved
independent languages, ethnic groups, and governments: Greece, Italy, Iberia,
Denmark, and Norway / Sweden. China’s coastline is much smoother, and
only the nearby Korean Peninsula attained separate importance. Europe has
two islands (Britain and Ireland) sufficiently big to assert their political
independence and to maintain their own languages and ethnicities, and one of
them (Britain) big and close enough to become a major independent European
power. But even China’s two largest islands, Taiwan and Hainan, have each
less than half the area of Ireland; neither was a major independent power until
Taiwan’s emergence in recent decades; and Japan’s geographic isolation kept
it until recently much more isolated politically from the Asian mainland than
Britain has been from mainland Europe. Europe is carved up into independent
linguistic, ethnic, and political units by high mountains (the Alps, Pyrenees,
Carpathians, and Norwegian border mountains), while China’s mountains east
of the Tibetan plateau are much less formidable barriers. China’s heartland is
bound together from east to west by two long navigable river systems in rich
alluvial valleys (the Yangtze and Yellow Rivers), and it is joined from north to
south by relatively easy connections between these two river systems
(eventually linked by canals). As a result, China very early became dominated
by two huge geographic core areas of high productivity, themselves only
weakly separated from each other and eventually fused into a single core.
Europe’s two biggest rivers, the Rhine and Danube, are smaller and connect
much less of Europe. Unlike China, Europe has many scattered small core
areas, none big enough to dominate the others for long, and each the center of
chronically independent states.
Once China was finally unified, in 221 B.C., no other independent state
ever had a chance of arising and persisting for long in China. Although
periods of disunity returned several times after 221 B.C., they always ended in
reunification. But the unification of Europe has resisted the efforts of such
determined conquerors as Charlemagne, Napoleon, and Hitler; even the
Roman Empire at its peak never controlled more than half of Europe’s area.
Thus, geographic connectedness and only modest internal barriers gave
China an initial advantage. North China, South China, the coast, and the
interior contributed different crops, livestock, technologies, and cultural
features to the eventually unified China. For example, millet cultivation,
bronze technology, and writing arose in North China, while rice cultivation
and cast-iron technology emerged in South China. For much of this book I
have emphasized the diffusion of technology that takes place in the absence of
formidable barriers. But China’s connectedness eventually became a
disadvantage, because a decision by one despot could and repeatedly did halt
innovation. In contrast, Europe’s geographic balkanization resulted in dozens
or hundreds of independent, competing statelets and centers of innovation. If
one state did not pursue some particular innovation, another did, forcing
neighboring states to do likewise or else be conquered or left economically
behind. Europe’s barriers were sufficient to prevent political unification, but
insufficient to halt the spread of technology and ideas. There has never been
one despot who could turn off the tap for all of Europe, as of China.
These comparisons suggest that geographic connectedness has exerted
both positive and negative effects on the evolution of technology. As a result,
in the very long run, technology may have developed most rapidly in regions
with moderate connectedness, neither too high nor too low. Technology’s
course over the last 1,000 years in China, Europe, and possibly the Indian
subcontinent exemplifies those net effects of high, moderate, and low
connectedness, respectively.
Naturally, additional factors contributed to history’s diverse courses in
different parts of Eurasia. For instance, the Fertile Crescent, China, and
Europe differed in their exposure to the perennial threat of barbarian
invasions by horse-mounted pastoral nomads of Central Asia. One of those
nomad groups (the Mongols) eventually destroyed the ancient irrigation
systems of Iran and Iraq, but none of the Asian nomads ever succeeded in
establishing themselves in the forests of western Europe beyond the
Hungarian plains. Environmental factors also include the Fertile Crescent’s
geographically intermediate location, controlling the trade routes linking
China and India to Europe, and China’s more remote location from Eurasia’s
other advanced civilizations, making China a gigantic virtual island within a
continent. China’s relative isolation is especially relevant to its adoption and
then rejection of technologies, so reminiscent of the rejections on Tasmania
and other islands (Chapters 13 and 15). But this brief discussion may at least
indicate the relevance of environmental factors to smaller-scale and shorter-
term patterns of history, as well as to history’s broadest pattern.
The histories of the Fertile Crescent and China also hold a salutary lesson
for the modern world: circumstances change, and past primacy is no
guarantee of future primacy. One might even wonder whether the
geographical reasoning employed throughout this book has at last become
wholly irrelevant in the modern world, now that ideas diffuse everywhere
instantly on the Internet and cargo is routinely airfreighted overnight between
continents. It might seem that entirely new rules apply to competition
between the world’s peoples, and that as a result new powers are emerging—
such as Taiwan, Korea, Malaysia, and especially Japan.
On reflection, though, we see that the supposedly new rules are just
variations on the old ones. Yes, the transistor, invented at Bell Labs in the
eastern United States in 1947, leapt 8,000 miles to launch an electronics
industry in Japan—but it did not make the shorter leap to found new
industries in Zaire or Paraguay. The nations rising to new power are still ones
that were incorporated thousands of years ago into the old centers of
dominance based on food production, or that have been repopulated by
peoples from those centers. Unlike Zaire or Paraguay, Japan and the other
new powers were able to exploit the transistor quickly because their
populations already had a long history of literacy, metal machinery, and
centralized government. The world’s two earliest centers of food production,
the Fertile Crescent and China, still dominate the modern world, either
through their immediate successor states (modern China), or through states
situated in neighboring regions influenced early by those two centers (Japan,
Korea, Malaysia, and Europe), or through states repopulated or ruled by their
overseas emigrants (the United States, Australia, Brazil). Prospects for world
dominance of sub-Saharan Africans, Aboriginal Australians, and Native
Americans remain dim. The hand of history’s course at 8000 B.C. lies heavily
on us.
AMONG OTHER FACTORS relevant to answering Yali’s question, cultural
factors and influences of individual people loom large. To take the former
first, human cultural traits vary greatly around the world. Some of that
cultural variation is no doubt a product of environmental variation, and I have
discussed many examples in this book. But an important question concerns
the possible significance of local cultural factors unrelated to the
environment. A minor cultural feature may arise for trivial, temporary local
reasons, become fixed, and then predispose a society toward more important
cultural choices, as is suggested by applications of chaos theory to other fields
of science. Such cultural processes are among history’s wild cards that would
tend to make history unpredictable.
As one example, I mentioned in Chapter 13 the QWERTY keyboard for
typewriters. It was adopted initially, out of many competing keyboard
designs, for trivial specific reasons involving early typewriter construction in
America in the 1860s, typewriter salesmanship, a decision in 1882 by a
certain Ms. Longley who founded the Shorthand and Typewriter Institute in
Cincinnati, and the success of Ms. Longley’s star typing pupil Frank
McGurrin, who thrashed Ms. Longley’s non-QWERTY competitor Louis
Taub in a widely publicized typing contest in 1888. The decision could have
gone to another keyboard at any of numerous stages between the 1860s and
the 1880s; nothing about the American environment favored the QWERTY
keyboard over its rivals. Once the decision had been made, though, the
QWERTY keyboard became so entrenched that it was also adopted for
computer keyboard design a century later. Equally trivial specific reasons,
now lost in the remote past, may have lain behind the Sumerian adoption of a
counting system based on 12 instead of 10 (leading to our modern 60-minute
hour, 24-hour day, 12-month year, and 360-degree circle), in contrast to the
widespread Mesoamerican counting system based on 20 (leading to its
calendar using two concurrent cycles of 260 named days and a 365-day year).
Those details of typewriter, clock, and calendar design have not affected
the competitive success of the societies adopting them. But it is easy to
imagine how they could have. For example, if the QWERTY keyboard of the
United States had not been adopted elsewhere in the world as well—say, if
Japan or Europe had adopted the much more efficient Dvorak keyboard—that
trivial decision in the 19th century might have had big consequences for the
competitive position of 20th-century American technology.
Similarly, a study of Chinese children suggested that they learn to write
more quickly when taught an alphabetic transcription of Chinese sounds
(termed pinyin) than when taught traditional Chinese writing, with its
thousands of signs. It has been suggested that the latter arose because of their
convenience for distinguishing the large numbers of Chinese words
possessing differing meanings but the same sounds (homophones). If so, the
abundance of homophones in the Chinese language may have had a large
impact on the role of literacy in Chinese society, yet it seems unlikely that
there was anything in the Chinese environment selecting for a language rich
in homophones. Did a linguistic or cultural factor account for the otherwise
puzzling failure of complex Andean civilizations to develop writing? Was
there anything about India’s environment predisposing toward rigid
socioeconomic castes, with grave consequences for the development of
technology in India? Was there anything about the Chinese environment
predisposing toward Confucian philosophy and cultural conservatism, which
may also have profoundly affected history? Why was proselytizing religion
(Christianity and Islam) a driving force for colonization and conquest among
Europeans and West Asians but not among Chinese?
These examples illustrate the broad range of questions concerning cultural
idiosyncrasies, unrelated to environment and initially of little significance,
that might evolve into influential and long-lasting cultural features. Their
significance constitutes an important unanswered question. It can best be
approached by concentrating attention on historical patterns that remain
puzzling after the effects of major environmental factors have been taken into
account.
WHAT ABOUT THE effects of idiosyncratic individual people? A familiar
modern example is the narrow failure, on July 20, 1944, of the assassination
attempt against Hitler and of a simultaneous uprising in Berlin. Both had been
planned by Germans who were convinced that the war could not be won and
who wanted to seek peace then, at a time when the eastern front between the
German and Russian armies still lay mostly within Russia’s borders. Hitler
was wounded by a time bomb in a briefcase placed under a conference table;
he might have been killed if the case had been placed slightly closer to the
chair where he was sitting. It is likely that the modern map of Eastern Europe
and the Cold War’s course would have been significantly different if Hitler
had indeed been killed and if World War II had ended then.
Less well known but even more fateful was a traffic accident in the
summer of 1930, over two years before Hitler’s seizure of power in Germany,
when a car in which he was riding in the “death seat” (right front passenger
seat) collided with a heavy trailer truck. The truck braked just in time to avoid
running over Hitler’s car and crushing him. Because of the degree to which
Hitler’s psychopathology determined Nazi policy and success, the form of an
eventual World War II would probably have been quite different if the truck
driver had braked one second later.
One can think of other individuals whose idiosyncrasies apparently
influenced history as did Hitler’s: Alexander the Great, Augustus, Buddha,
Christ, Lenin, Martin Luther, the Inca emperor Pachacuti, Mohammed,
William the Conqueror, and the Zulu king Shaka, to name a few. To what
extent did each really change events, as opposed to “just” happening to be the
right person in the right place at the right time? At the one extreme is the view
of the historian Thomas Carlyle: “Universal history, the history of what man
[ sic] has accomplished in this world, is at bottom the History of the Great
Men who have worked here.” At the opposite extreme is the view of the
Prussian statesman Otto von Bismarck, who unlike Carlyle had long firsthand
experience of politics’ inner workings: “The statesman’s task is to hear God’s
footsteps marching through history, and to try to catch on to His coattails as
He marches past.”
Like cultural idiosyncrasies, individual idiosyncrasies throw wild cards
into the course of history. They may make history inexplicable in terms of
environmental forces, or indeed of any generalizable causes. For the purposes
of this book, however, they are scarcely relevant, because even the most
ardent proponent of the Great Man theory would find it difficult to interpret
history’s broadest pattern in terms of a few Great Men. Perhaps Alexander the
Great did nudge the course of western Eurasia’s already literate, food-
producing, iron-equipped states, but he had nothing to do with the fact that
western Eurasia already supported literate, food-producing, iron-equipped
states at a time when Australia still supported only nonliterate hunter-gatherer
tribes lacking metal tools. Nevertheless, it remains an open question how
wide and lasting the effects of idiosyncratic individuals on history really are.
THE DISCIPLINE OF history is generally not considered to be a science, but
something closer to the humanities. At best, history is classified among the
social sciences, of which it rates as the least scientific. While the field of
government is often termed “political science” and the Nobel Prize in
economics refers to “economic science,” history departments rarely if ever
label themselves “Department of Historical Science.” Most historians do not
think of themselves as scientists and receive little training in acknowledged
sciences and their methodologies. The sense that history is nothing more than
a mass of details is captured in numerous aphorisms: “History is just one
damn fact after another,” “History is more or less bunk,” “There is no law of
history any more than of a kaleidoscope,” and so on.
One cannot deny that it is more difficult to extract general principles from
studying history than from studying planetary orbits. However, the difficulties
seem to me not fatal. Similar ones apply to other historical subjects whose
place among the natural sciences is nevertheless secure, including astronomy,
climatology, ecology, evolutionary biology, geology, and paleontology.
People’s image of science is unfortunately often based on physics and a few
other fields with similar methodologies. Scientists in those fields tend to be
ignorantly disdainful of fields to which those methodologies are inappropriate
and which must therefore seek other methodologies—such as my own
research areas of ecology and evolutionary biology. But recall that the word
“science” means “knowledge” (from the Latin scire, “to know,” and scientia,
“knowledge”), to be obtained by whatever methods are most appropriate to
the particular field. Hence I have much empathy with students of human
history for the difficulties they face.
Historical sciences in the broad sense (including astronomy and the like)
share many features that set them apart from nonhistorical sciences such as
physics, chemistry, and molecular biology. I would single out four:
methodology, causation, prediction, and complexity.
In physics the chief method for gaining knowledge is the laboratory
experiment, by which one manipulates the parameter whose effect is in
question, executes parallel control experiments with that parameter held
constant, holds other parameters constant throughout, replicates both the
experimental manipulation and the control experiment, and obtains
quantitative data. This strategy, which also works well in chemistry and
molecular biology, is so identified with science in the minds of many people
that experimentation is often held to be the essence of the scientific method.
But laboratory experimentation can obviously play little or no role in many of
the historical sciences. One cannot interrupt galaxy formation, start and stop
hurricanes and ice ages, experimentally exterminate grizzly bears in a few
national parks, or rerun the course of dinosaur evolution. Instead, one must
gain knowledge in these historical sciences by other means, such as
observation, comparison, and so-called natural experiments (to which I shall
return in a moment).
Historical sciences are concerned with chains of proximate and ultimate
causes. In most of physics and chemistry the concepts of “ultimate cause,”
“purpose,” and “function” are meaningless, yet they are essential to
understanding living systems in general and human activities in particular. For
instance, an evolutionary biologist studying Arctic hares whose fur color turns
from brown in summer to white in winter is not satisfied with identifying the
mundane proximate causes of fur color in terms of the fur pigments’
molecular structures and biosynthetic pathways. The more important
questions involve function (camouflage against predators?) and ultimate
cause (natural selection starting with an ancestral hare population with
seasonally unchanging fur color?). Similarly, a European historian is not
satisfied with describing the condition of Europe in both 1815 and 1918 as
having just achieved peace after a costly pan-European war. Understanding
the contrasting chains of events leading up to the two peace treaties is
essential to understanding why an even more costly pan-European war broke
out again within a few decades of 1918 but not of 1815. But chemists do not
assign a purpose or function to a collision of two gas molecules, nor do they
seek an ultimate cause for the collision.
Still another difference between historical and nonhistorical sciences
involves prediction. In chemistry and physics the acid test of one’s
understanding of a system is whether one can successfully predict its future
behavior. Again, physicists tend to look down on evolutionary biology and
history, because those fields appear to fail this test. In historical sciences, one
can provide a posteriori explanations (e.g., why an asteroid impact on Earth
66 million years ago may have driven dinosaurs but not many other species to
extinction), but a priori predictions are more difficult (we would be uncertain
which species would be driven to extinction if we did not have the actual past
event to guide us). However, historians and historical scientists do make and
test predictions about what future discoveries of data will show us about past
events.
The properties of historical systems that complicate attempts at prediction
can be described in several alternative ways. One can point out that human
societies and dinosaurs are extremely complex, being characterized by an
enormous number of independent variables that feed back on each other. As a
result, small changes at a lower level of organization can lead to emergent
changes at a higher level. A typical example is the effect of that one truck
driver’s braking response, in Hitler’s nearly fatal traffic accident of 1930, on
the lives of a hundred million people who were killed or wounded in World
War II. Although most biologists agree that biological systems are in the end
wholly determined by their physical properties and obey the laws of quantum
mechanics, the systems’ complexity means, for practical purposes, that that
deterministic causation does not translate into predictability. Knowledge of
quantum mechanics does not help one understand why introduced placental
predators have exterminated so many Australian marsupial species, or why
the Allied Powers rather than the Central Powers won World War I.
Each glacier, nebula, hurricane, human society, and biological species,
and even each individual and cell of a sexually reproducing species, is unique,
because it is influenced by so many variables and made up of so many
variable parts. In contrast, for any of the physicist’s elementary particles and
isotopes and of the chemist’s molecules, all individuals of the entity are
identical to each other. Hence physicists and chemists can formulate universal
deterministic laws at the macroscopic level, but biologists and historians can
formulate only statistical trends. With a very high probability of being correct,
I can predict that, of the next 1,000 babies born at the University of California
Medical Center, where I work, not fewer than 480 or more than 520 will be
boys. But I had no means of knowing in advance that my own two children
would be boys. Similarly, historians note that tribal societies may have been
more likely to develop into chiefdoms if the local population was sufficiently
large and dense and if there was potential for surplus food production than if
that was not the case. But each such local population has its own unique
features, with the result that chiefdoms did emerge in the highlands of
Mexico, Guatemala, Peru, and Madagascar, but not in those of New Guinea or
Guadalcanal.
Still another way of describing the complexity and unpredictability of
historical systems, despite their ultimate determinacy, is to note that long
chains of causation may separate final effects from ultimate causes lying
outside the domain of that field of science. For example, the dinosaurs may
have been exterminated by the impact of an asteroid whose orbit was
completely determined by the laws of classical mechanics. But if there had
been any paleontologists living 67 million years ago, they could not have
predicted the dinosaurs’ imminent demise, because asteroids belong to a field
of science otherwise remote from dinosaur biology. Similarly, the Little Ice
Age of A.D. 1300–1500 contributed to the extinction of the Greenland Norse,
but no historian, and probably not even a modern climatologist, could have
predicted the Little Ice Age.
THUS, THE DIFFICULTIES historians face in establishing cause-and-effect
relations in the history of human societies are broadly similar to the
difficulties facing astronomers, climatologists, ecologists, evolutionary
biologists, geologists, and paleontologists. To varying degrees, each of these
fields is plagued by the impossibility of performing replicated, controlled
experimental interventions, the complexity arising from enormous numbers of
variables, the resulting uniqueness of each system, the consequent
impossibility of formulating universal laws, and the difficulties of predicting
emergent properties and future behavior. Prediction in history, as in other
historical sciences, is most feasible on large spatial scales and over long
times, when the unique features of millions of small-scale brief events
become averaged out. Just as I could predict the sex ratio of the next 1,000
newborns but not the sexes of my own two children, the historian can
recognize factors that made inevitable the broad outcome of the collision
between American and Eurasian societies after 13,000 years of separate
developments, but not the outcome of the 1960 U.S. presidential election. The
details of which candidate said what during a single televised debate in
October 1960 could have given the electoral victory to Nixon instead of to
Kennedy, but no details of who said what could have blocked the European
conquest of Native Americans.
How can students of human history profit from the experience of
scientists in other historical sciences? A methodology that has proved useful
involves the comparative method and so-called natural experiments. While
neither astronomers studying galaxy formation nor human historians can
manipulate their systems in controlled laboratory experiments, they both can
take advantage of natural experiments, by comparing systems differing in the
presence or absence (or in the strong or weak effect) of some putative
causative factor. For example, epidemiologists, forbidden to feed large
amounts of salt to people experimentally, have still been able to identify
effects of high salt intake by comparing groups of humans who already differ
greatly in their salt intake; and cultural anthropologists, unable to provide
human groups experimentally with varying resource abundances for many
centuries, still study long-term effects of resource abundance on human
societies by comparing recent Polynesian populations living on islands
differing naturally in resource abundance. The student of human history can
draw on many more natural experiments than just comparisons among the
five inhabited continents. Comparisons can also utilize large islands that have
developed complex societies in a considerable degree of isolation (such as
Japan, Madagascar, Native American Hispaniola, New Guinea, Hawaii, and
many others), as well as societies on hundreds of smaller islands and regional
societies within each of the continents.
Natural experiments in any field, whether in ecology or human history,
are inherently open to potential methodological criticisms. Those include
confounding effects of natural variation in additional variables besides the one
of interest, as well as problems in inferring chains of causation from observed
correlations between variables. Such methodological problems have been
discussed in great detail for some of the historical sciences. In particular,
epidemiology, the science of drawing inferences about human diseases by
comparing groups of people (often by retrospective historical studies), has for
a long time successfully employed formalized procedures for dealing with
problems similar to those facing historians of human societies. Ecologists
have also devoted much attention to the problems of natural experiments, a
methodology to which they must resort in many cases where direct
experimental interventions to manipulate relevant ecological variables would
be immoral, illegal, or impossible. Evolutionary biologists have recently been
developing ever more sophisticated methods for drawing conclusions from
comparisons of different plants and animals of known evolutionary histories.
In short, I acknowledge that it is much more difficult to understand human
history than to understand problems in fields of science where history is
unimportant and where fewer individual variables operate. Nevertheless,
successful methodologies for analyzing historical problems have been worked
out in several fields. As a result, the histories of dinosaurs, nebulas, and
glaciers are generally acknowledged to belong to fields of science rather than
to the humanities. But introspection gives us far more insight into the ways of
other humans than into those of dinosaurs. I am thus optimistic that historical
studies of human societies can be pursued as scientifically as studies of
dinosaurs—and with profit to our own society today, by teaching us what
shaped the modern world, and what might shape our future.
WHO ARE THE JAPANESE ?
AMONG MODERN WORLD POWERS, THE MOST DISTINCTIVE IN their culture and
environment are the Japanese people. The origins of their language are among
the most disputed questions of linguistics: for not a single other one of the
world’s major languages are the affinities to other languages still in doubt.
Who are the Japanese, when and whence did they come to Japan, and how did
they evolve their unique speech? These questions are central to the self-image
of the Japanese, and to how they are viewed by other peoples. Japan’s rising
dominance and its sometimes touchy relations with its neighbors make it
more important than ever to strip away persistent myths and to find answers.
My minimal coverage of Japan in previous editions of Guns, Germs, and
Steel constituted the most important geographic lacuna of my book. New
information about Japanese genetics and language origins, accumulating since
the book’s first publication, now encourages me to test how Japan fits into my
overall framework.
The search for answers is difficult because the evidence is so conflicting.
On the one hand, the Japanese people are biologically undistinctive, being
very similar in their appearance and genes to other East Asians, especially to
Koreans. As the Japanese are fond of stressing, they are culturally and
biologically rather homogeneous: there is little difference among people from
different parts of Japan, except for a very different people called the Ainu on
Japan’s nothernmost island of Hokkaido. All these facts seem to suggest that
the Japanese reached Japan recently from the East Asian mainland and
displaced the Ainu, who represent the original inhabitants. But, if that were
true, you might expect the Japanese language to show obvious close affinities
to some East Asian mainland language, just as the English language is closely
related to other Germanic languages because Anglo-Saxons from the
continent took over England as recently as the 6th century A.D. How can we
resolve this contradiction between Japan’s presumably ancient language and
all that other evidence for recent origins?
Four conflicting theories, each of them popular in some countries and
unpopular in others, have been proposed. Most popular in Japan is the view
that the Japanese gradually evolved from ancient Ice Age people who
occupied Japan long before 20,000 B.C. Also widespread in Japan is the theory
that the Japanese are descended from horse-riding Central Asian nomads who
passed through Korea to conquer Japan in the 4th century A.D. but who were
emphatically not Koreans. A theory favored by many Western archaeologists
and Koreans, and unpopular in some circles in Japan, is that the Japanese are
descendants of immigrants from Korea who arrived with rice paddy
agriculture around 400 B.C. Finally, peoples named in these other three
theories could have mixed to form the modern Japanese.
When similar questions arise about the origins of other peoples, they can
be discussed dispassionately. That is not true of questions about Japanese
origins. It was a remarkable achievement that Japan, unlike so many other
non-European countries, preserved its political independence and culture
while emerging from isolation and creating an industrialized society in the
late 19th century. Now, the Japanese people are understandably concerned
about maintaining their traditions in the face of massive Western cultural
influence. They want to believe that their language and culture are so unique
as to have required uniquely complex developmental processes, unlike those
operating elsewhere in the world. To acknowledge that the Japanese language
is related to any other language seems to constitute a surrender of cultural
identity.
Until 1946, Japanese schools taught a myth of Japanese history based on
the earliest Japanese chronicles of A.D. 712 and 720. Those chronicles
describe how the sun goddess Amaterasu, born from the left eye of the creator
god Izanagi, sent her grandson Ninigi to earth on the Japanese island of
Kyushu to wed an earthly deity. Ninigi’s great-grandson Jimmu, aided by a
dazzling sacred bird that rendered his enemies helpless, became the first
emperor of Japan in 660 B.C. To fill the gap between 660 B.C. and the earliest
historically documented Japanese monarchs, the chronicles invented 13 other,
equally fictitious emperors.
Before the end of World War II, when Emperor Hirohito finally told the
Japanese people that he was not of divine descent, Japanese archaeologists
and historians had to make their interpretations conform to this account.
Although they have more freedom of interpretation today, constraints remain.
Japan’s most important archaeological monuments—the 158 gigantic kofun
tombs constructed between A.D. 300 and 686, and thought to contain the
remains of ancestral emperors and their families—are still the property of the
Imperial Household Agency. Excavation of the tombs is forbidden because it
would constitute desecration—and it might also shed undesired light on
where Japan’s imperial family really came from (e.g., perhaps Korea?).
Whereas archaeological deposits in the United States were left by peoples
(Native Americans) unrelated to most modern Americans, deposits in Japan,
no matter how ancient, are believed to have been left by ancestors of the
modern Japanese themselves. Hence archaeology in Japan is supported by
astronomically large budgets and draws public attention to a degree
inconceivable anywhere else in the world. Each year, Japanese archaeologists
excavate over 10,000 digs and employ up to 50,000 field workers. Twenty
times more Neolithic sites have thereby been discovered in Japan than in the
whole of China. Accounts of excavations appear almost daily on TV and on
the front page of Japan’s largest newspapers. Determined to prove that the
ancestors of the modern Japanese came to Japan in the remote past,
archaeologists reporting on the excavations emphasize how different Japan’s
ancient inhabitants were from contemporary peoples elsewhere, but how
similar they were to the Japanese today. For instance, an archaeologist
lecturing about a site 2,000 years old would draw attention to the garbage pits
into which the site’s inhabitants threw their raw garbage, illustrating that
Japanese at those distant times already practiced the cleanliness on which
their supposed descendants pride themselves today.
What makes it especially difficult to discuss Japanese archaeology
dispassionately is that Japanese interpretations of their past affect their present
behavior. Among East Asian peoples, who brought culture to whom, who is
culturally superior and who is a barbarian, and who has historic claims to
whose land? For instance, there is much archaeological evidence for
exchanges of people and material objects between Japan and Korea in the
period A.D. 300–700. The Japanese interpret this to mean that Japan conquered
Korea then and brought Korean slaves and artisans to Japan; the Korean
interpretation is instead that Korea conquered Japan, and that the founders of
the Japanese imperial family were Korean.
Hence when Japan sent troops to Korea and annexed it in 1910, Japanese
military leaders celebrated the annexation as “the restoration of the legitimate
arrangement of antiquity.” For the next 35 years, Japanese occupation forces
tried to eradicate Korean culture and to replace Korean with the Japanese
language in schools. Korean families that have lived in Japan for several
generations still find it difficult to acquire Japanese citizenship. “Nose tombs”
in Japan still contain the noses cut off of 20,000 Koreans and brought to Japan
as trophies of a 16th-century Japanese invasion of that country. Not
surprisingly, loathing of the Japanese is widespread in Korea, and contempt
for Koreans is widespread in Japan.
As just one example of how seemingly arcane archaeological disputes can
arouse passion, consider the best-known archaeological relic of pre-chronicle
Japan: the Eta-Funayama sword of the 5th century A.D., designated a national
treasure and held in the Tokyo National Museum. Inlaid in silver on the iron
sword is an inscription in Chinese characters, one of the oldest surviving
samples of writing in Japan, referring to a Great King and an official serving
him and a Korean scribe named Ch-oan. Several of the Chinese characters are
incomplete, rusted, or missing and must be guessed at. Japanese scholars
traditionally took the missing characters to mean that the king was the
Japanese emperor Mizuha-wake of the Beautiful Teeth named in the 8th-
century Japanese chronicles. In 1966, however, the Korean historian Kim
Sokhyong shocked Japanese scholars with the suggestion that the missing
name was actually King Kaero of Korea, and that the named official was one
of his Korean vassals who were then occupying parts of Japan. What really
was “the legitimate arrangement of antiquity”?
Today, Japan and Korea are both economic powerhouses, facing each
other across Tsushima Strait, and viewing each other through poisoned lenses
of false myths and real past atrocities. It bodes ill for the future of East Asia if
these two great peoples cannot find common ground. A correct understanding
of who the Japanese people really are, and how they diverged from the
closely related Korean people, will be essential to finding that common
ground.
STARTING POINTS FOR UNDERSTANDING JAPAN’S UNIQUE culture are its unique
geography and environment. At first, Japan might seem to be geographically
very similar to Britain, both being large archipelagoes flanking the Eurasian
continent on the east and the west respectively. But there are detailed
differences that prove important: Japan is somewhat larger and more distant.
Japan’s area of 146,000 square miles is half again greater than Britain’s, and
nearly equal to California’s. Britain lies only 22 miles from the French coast,
but Japan lies 110 miles from the closest point of the Asian mainland (South
Korea), and is 180 miles from mainland Russia and 460 miles from mainland
China.
Perhaps as a result, Britain throughout its history has been much more
closely enmeshed with mainland Europe than has Japan with mainland Asia.
For instance, since the time of Christ there have been four successful
invasions of Britain from the continent, but none of Japan (unless Korea
really did conquer pre-chronicle Japan). Conversely, British troops have
fought on the continent in every century since the Norman Conquest of A.D.
1066, whereas before the late 19th century mainland Asia was always free of
Japanese troops except for Korea during pre-chronicle times and the last
decade of the 16th century. Thus, details of geography have made Japan more
isolated and, therefore, even more distinctive culturally than Britain.
As for Japan’s climate, its rainfall, ranging up to 160 inches per year,
makes it the wettest temperate country in the world. Furthermore, in contrast
to the winter rains prevailing over much of Europe, Japan’s rains are
concentrated in the summer growing season. That combination of high
rainfall and summer rains gives Japan the highest plant productivity of any
nation in the temperate zones. Half of its farmland is devoted to labor-
intensive, high-yield, irrigated rice agriculture, facilitated by abundant rivers
flowing from the wet mountains onto sloping lowland plains. While 80
percent of Japan’s land area consists of mountains unsuitable for agriculture
and only 14 percent is farmland, per square mile of that farmland Japan
supports a population density eight times that of Britain’s. In fact, in
proportion to its available area of farmland, Japan is the most densely
populated major society in the world.
Japan’s high rainfall also ensures that its forest regenerates quickly after
logging. Despite thousand of years of dense human occupation, everyone’s
first impression of Japan is of its greenness, because more than 70 percent of
its land area is still covered by forest (compared with only 10 percent for
Britain). Conversely, all that forest means that there is no native grassland or
natural pasture. Traditionally, the sole animal raised on a large scale for food
in Japan has been the pig; sheep and goats have never been significant, and
cattle were raised for pulling plows and carts but not for food. Japanese-raised
beef remains a luxury food of the wealthy few, selling for up to $100 per
pound.
Japanese forest composition varies with latitude and altitude: evergreen
leafy forest in the south at low altitude, deciduous leafy forest in central
Japan, and coniferous forest in the north and at high altitude. For prehistoric
humans the most productive forest was the deciduous leafy forest because of
its abundance of edible nuts, such as walnuts, chestnuts, horse chestnuts,
acorns, and beechnuts. Like Japanese forests, Japanese waters are
outstandingly productive. The lakes, rivers, Inland Sea, Sea of Japan to the
west, and Pacific Ocean to the east teem with fish such as salmon, trout, tuna,
sardines, mackerel, herring, and cod. Today, Japan is the largest catcher,
importer, and consumer of fish in the world. Japanese waters are also rich in
clams and oysters and other shellfish, crabs and shrimp and crayfish, and
edible seaweeds. As we shall see, that high productivity of the land, fresh
water, and seas was a key to Japan’s prehistory.
BEFORE WE TURN TO THE EVIDENCE OF ARCHEOLOGY, let us consider the
evidence of Japanese origins from biology, linguistics, early portraits, and
recorded history. The conflicts between these four familiar types of evidence
are what make Japanese origins so controversial.
From southwest to northeast, the four main Japanese islands are Kyushu,
Shikoku, Honshu (the largest island), and Hokkaido. Until large-scale
Japanese immigration into Hokkaido in the late 19th century, that island (plus
northern Honshu) was inhabited in historic times mainly by Ainu, living as
hunter-gatherers with only limited agriculture, while the Japanese occupied
the other three islands. In their genes and skeletons as well as in external
appearance, the Japanese are very similar to other East Asians, including
North Chinese, East Siberians, and especially Koreans. Even my Japanese and
Korean friends say that they sometimes have difficulty guessing whether
someone is Japanese or Korean just by looking at his or her face.
As for the Ainu, their distinctive appearance has resulted in more being
written about the origins and relationships than about any other single people
on earth. Ainu men have a luxuriant beard and the most profuse body hair of
any people. That fact, coupled with some other inherited traits such as their
fingerprint patterns and their type of ear wax, has often led to their being
classified as Causcasoids (so-called white people) who somehow migrated
east through Eurasia and ended up in Japan. In their overall genetic makeup,
though, the Ainu are related to other East Asians, including the Japanese,
Koreans, and Okinawans. Perhaps their distinctive external appearance
involves relatively few genes that arose through sexual selection after they
migrated from mainland Asia and became isolated on the Japanese
archipelago. The distinctive appearance and hunter-gatherer lifestyle of the
Ainu, and the undistinctive appearance and the intensive agricultural lifestyle
of the Japanese, are frequently taken to suggest the straightforward
interpretation that the Ainu are descended from Japan’s original hunter-
gatherer inhabitants, and that the Japanese are more recent invaders from the
Asian mainland.
But this view is difficult to reconcile with the distinctiveness of the
Japanese language, which everyone agrees does not bear a detailed close
relation to any other language in the world (in the way that French is close to
Spanish). Insofar as anything can be said about its relationships, many
scholars consider it to be an isolate member of Asia’s Altaic language family,
which consists of Turkic languages, Mongolian languages, and the Tungus
languages of East Siberia. Korean is also often considered to be an isolated
member of this family, and within the family Japanese and Korean may be
more related to each other than to other Altaic languages. However, the
similarities between Japanese and Korean are confined to general
grammatical features and about 15 percent of their basic vocabulary, rather
than the detailed shared features of grammar and vocabulary that link French
to Spanish. If one accepts that Japanese and Koreans are indeed related,
however distantly, that sharing of 15 percent of their vocabulary suggests that
the two languages began to diverge from each other over 5,000 years ago,
rather than the mere 2,000 years or less during which French and Spanish
have been diverging. As for the Ainu language, its relationships are
thoroughly in doubt; it may not have any special relationship to Japanese.
After biology and language, our third type of evidence about Japanese
origins comes from ancient portraits. The earliest preserved likenesses of
Japan’s inhabitants are statues called haniwa, erected outside tombs around
1,500 years ago. Especially in their eye shapes, those statues unmistakably
depict East Asians, such as modern Japanese or Koreans. They do not
resemble the heavily bearded Ainu. If the Japanese did replace the Ainu over
Japan south of Hokkaido, that replacement must have occurred before A.D.
500. After the Japanese established trading posts on Hokkaido in 1615, they
proceeded to treat the Hokkaido Ainu much as white Americans treated
Native Americans. The Ainu were conquered, rounded up into reserves,
forced to work for trading posts, driven off land desired by Japanese farmers,
and killed when they revolted. When Japan annexed Hokkaido in 1869,
Japanese schoolteachers made determined efforts to expunge the Ainu culture
and language. Today, the language is virtually extinct, and probably no
purebred Ainu survives.
Our earliest written information about Japan comes from Chinese
chronicles, because China developed literacy long before it spread from China
to either Korea or Japan. From 108 B.C. until A.D. 313 China occupied a
settlement in North Korea and exchanged envoys with Japan. In the resulting
Chinese accounts of various peoples referred to as “Eastern Barbarians,”
Japan is described under the name of Wa, whose inhabitants were said to be
divided into over a hundred little states that fought a lot with each other. Only
a few Koreans of Japanese inscriptions before the year A.D. 700 have been
preserved, but extensive chronicles were written in A.D. 712 and 720 in Japan
and later in Korea. While these Japanese and Korean chronicles purport to
relate histories of earlier periods, they are full of obvious fabrications
designed to glorify and legitimize ruling families—such as the Japanese
account of their emperor’s descent from the sun goddess Amaterasu.
Nevertheless, the chronicles suffice to make clear that there was massive
influence of Korea itself, and of China via Korea, on Japan, leading to the
introduction of Buddhism, writing, metallurgy, other crafts, and bureaucratic
methods into Japan. The chronicles are also full of accounts of Koreans in
Japan and of Japanese in Korea—interpreted by Japanese or Korean
historians respectively as evidence of Japanese conquest of Korea or the
reverse.
WE HAVE THUS SEEN THAT THE ANCESTORS OF THE Japanese reached Japan
before they had writing, and that their biology would suggest a recent arrival
but their language seemingly suggests arrival at least 5,000 years ago. Let us
now turn to the evidence of archaeology in an attempt to solve this puzzle. We
shall see that ancient Japanese societies were among the most remarkable in
the world.
Shallow seas now surround much of Japan and coastal East Asia. Hence
those seas became dry land during the Ice Ages, when much ocean water was
locked up in glaciers and sea level lay at about 500 feet below its present
stand. At those times, Japan’s northernmost island of Hokkaido was
connected by a land bridge over what is now Sakhalin Island to the Russian
mainland; Japan’s southernmost island of Kyushu was connected by another
land bridge to South Korea over what is now Tsushima Strait; all of the main
Japanese islands were connected to one another; and much of the expanse of
what are now the Yellow Sea and the East China Sea consisted of land
extensions of mainland China. Hence it comes as no surprise that the
mammals walking out to Japan in those land bridge days included not only
the ancestors of modern Japan’s bears and monkeys but also ancient humans,
long before boats had been invented. Stone tools indicate human arrival as
early as half a million years ago. Ancient stone tools of northern Japan
resemble those of Siberia and northern China, but those of southern Japan
resemble those of Korea and southern China, suggesting that both the
northern and southern land bridges were used.
Ice Age Japan was not a great place to live. Even though most of Japan
escaped the glaciers that blanketed Britain and Canada, Japan was still cold,
dry, and extensively covered with conifer and birch forests offering little plant
food to humans. Those drawbacks make the precocity of the Ice Age Japanese
all the more impressive: around 30,000 years ago, they were among the
earliest people in the world to develop stone tools with edges ground to a
sharp edge instead of just chipped or flaked. In the archaeology of Britain,
edge-ground tools are considered a big cultural advance that separates the
Neolithic (New Stone Age) from the Paleolithic (Old Stone Age), but they
don’t show up in Britain until agriculture’s arrival less than 7,000 years ago.
Around 13,000 years ago, as glaciers melted rapidly all over the world,
conditions in Japan changed spectacularly for the better, as far as humans
were concerned. Temperatures, rainfall, and humidity all increased, raising
plant productivity to the modern high levels for which Japan is preeminent
among temperate-zone countries. Deciduous leafy forests full of nut trees,
which had been confined to southern Japan during the Ice Ages, expanded
northward at the expense of coniferous forest, thereby replacing a forest type
that had been rather sterile for humans with a much more productive forest
type. The rise in sea level severed the land bridges, converted Japan from
being a piece of the Asian continent to being a big archipelago, turned what
had been a plain into rich shallow seas, and created thousands of miles of
productive new coastline with innumerable islands, bays, tidal flats, and
estuaries, all teeming with seafood.
The end of the Ice Age was accompanied by the first of the two most
decisive changes in Japanese history: the invention of pottery. For the first
time in human experience, people now had watertight containers readily
available in any desired shape. With their new ability to boil, steam, or
simmer food, they gained access to abundant food resources that had
previously been difficult to utilize: leafy vegetables, which would burn or
dehydrate if cooked over a fire; shellfish, which could now be opened easily;
and toxic or bitter, but otherwise nutritious, foods like acorns and horse
chestnuts, which could now have their toxins leached out by soaking. Soft
boiled foods could be fed to small children, permitting the kids to be weaned
earlier and their mothers to produce babies at shorter birth intervals. Toothless
old people, the repositories of information in a preliterate society, could now
be fed and live longer. All those momentous consequences of pottery
triggered a population explosion, causing Japan’s population to climb from an
estimated few thousand people to a quarter of a million.
Naturally, the Japanese were not the sole ancient people with pottery: it
was invented independently at many different times and places around the
ancient world. But the world’s oldest known pottery was made in Japan
12,700 years ago. When those radiocarbon dates were announced in 1960, not
even Japanese scientists could at first believe them. In the usual experience of
archaeologists, inventions are supposed to flow from mainlands to islands,
and small peripheral societies aren’t supposed to contribute revolutionary
advances to the rest of the world. Especially in the experience of Japanese
archaeologists, China is regarded as the source of cultural breakthroughs in
East Asia, such as agriculture, writing, metallurgy, and everything else of
significance. Today, nearly 40 years after those early dates for pottery in
Japan were measured, archaeologists are still reeling from the carbon 14
shock, as it is termed. Other early pottery has been found in China and in
eastern Russia (near Vladivostok). Asian archaeologists are racing to beat the
Japanese record. (In fact, I just heard rumors that the Chinese and Russians
are close to beating it.) But the Japanese still hold the world record, with
pottery thousands of years older than the oldest from the Fertile Crescent or
Europe.
The prejudice that islanders are supposed to learn from superior
continentals wasn’t the sole reason why record-breaking Japanese pottery
caused such a shock. In addition, those first Japanese potters were clearly
hunter-gatherers, and that also violated established views. Mostly, pottery is
owned by sedentary societies: what nomad wants to carry a collection of
heavy pots, as well as weapons and the baby, every time he or she shifts
camp? Hence hunter-gatherers usually don’t have pottery, because most
sedentary societies elsewhere in the world arose only with the adoption of
agriculture. But the Japanese environment is so productive that it was one of
the few locations where people could settle down and make pottery while still
living as hunter-gatherers. Pottery helped those Japanese hunter-gatherers to
exploit their environment’s rich food resources more than 10,000 years before
intensive agriculture reached Japan. In contrast, pottery wasn’t adopted in the
Fertile Crescent until about a thousand years after the adoption of agriculture.
Not surprisingly, ancient Japanese pottery was technologically simple by
today’s standards. It lacked glazes, was made by hand rather than on potters’
wheels, was baked in open fires rather than in kilns, and was fired at
relatively low temperatures. But, as time went on, it came to be made in an
incredible profusion of shapes that rate as great art by the standards of any
era. Much of it was decorated by rolling or pressing a cord on the clay while it
was still soft. Because the Japanese word for “cord marking” is jomon, the
term jomon is applied to the pottery itself, to the ancient Japanese people who
made it, and to that whole period in Japanese prehistory beginning with the
invention of pottery and ending only 10,000 years later.
The earliest Jomon pottery of 12,700 years ago comes from Kyushu, the
southernmost Japanese island. Thereafter, pottery spread north, reaching the
vicinity of modern Tokyo around 9,500 years ago and the northernmost island
of Hokkaido by 7,000 years ago. Pottery’s northward spread followed the
northward spread of deciduous forest rich in nuts, suggesting that the food
explosion was what permitted sedentary living and the pottery explosion.
Reinforcing that interpretation of a single invention of pottery in the south
and a spread from that one source, the style of the earliest Jomon pottery is
fairly uniform over the whole of Japan. With time, a few dozen regional styles
developed over the 1,500-mile length of the Japanese archipelago.
HOW DID JOMON PEOPLE MAKE THEIR LIVING? WE have abundant evidence
from the garbage that they left behind at hundreds of thousands of excavated
archaeological sites and huge shell mounds distributed all over Japan. It turns
out that they were hunters, gatherers, and fishing people enjoying a
remarkably diverse and well-balanced diet that modern nutritionists would
applaud.
One major food category was nuts, especially chestnuts and walnuts, plus
horse chestnuts and acorns leached free of their bitter poisons. Nuts could be
harvested in autumn in prodigious quantities and then stored for the winter in
underground storage pits up to six feet deep and six feet wide. Other plant
foods included berries, fruits, seeds, leaves, shoots, bulbs, and roots. In all,
archaeologists sifting through Jomon garbage have identified 64 species of
edible plants.
Then as now, Japan’s inhabitants were also among the world’s leading
consumers of seafood. Tuna were harpooned in the open ocean; porpoises
were driven into shallow water and clubbed or speared, just as they are in
Japan today; seals were killed on the beaches; seasonal runs of salmon were
exploited in the rivers; a wide variety of fish were netted, captured in weirs,
and caught on fishhooks carved out of deer antlers; and shellfish, crabs, and
seaweed were gathered in the intertidal zone or harvested by divers. Jomon
skeletons show a high incidence of what pathologists term auditory exostosis,
meaning abnormal bone growth in the ears as often observed in divers today.
Among land animals hunted, wild boar and deer were the commonest
prey, followed by mountain goat and bear. These game animals were caught
in pit traps, shot with bow and arrow, and run down with dogs. Pig bones
appeared in Jomon times on offshore islands where pigs do not occur
naturally, making one wonder whether Jomon people were starting to
experiment with pig domestication.
The most debated question about Jomon subsistence concerns the possible
contribution of agriculture. Jomon sites often contain remains of edible plants
that are native to Japan as wild species but are also grown as crops today,
including adzuki bean, mung bean, and barnyard millet. The remains from
Jomon times do not clearly show morphological features distinguishing the
crops from their wild ancestors, so we do not know whether these plants were
gathered in the wild or were being intentionally grown. Sites also have debris
of edible or useful plant species that are not native to Japan, and that must
have been introduced for their value from the Asian mainland, such as
buckwheat, melons, bottle gourd, hemp, and shiso or beefsteak plant (used for
seasoning). Around 1200 B.C., toward the end of the Jomon period, a few
grains of rice, barley, foxtail millet, and broomcorn millet, the staple cereals
of East Asia, began to appear. All of these tantalizing clues make it likely that
Jomon people were starting to practice some slash-and-burn agriculture, but it
was evidently in a casual way that made only a minor contribution to their
diet.
I don’t mean to leave the impression that every one of these foods that
I’ve mentioned was eaten everywhere throughout Jomon Japan. In the nut-
rich forests of northern Japan, nut storage pits were especially important,
along with seal hunting and sea fishing. In the nut-poor southwest, shellfish
assumed a greater role. But diversity still characterizes local Jomon diets and
even individual Jomon meals. For instance, as shown by preserved remains of
meals, Jomon people blended chestnut and walnut flour, pig and deer meat
and blood, and bird eggs in various proportions to produce either a high-
carbohydrate Mrs. Jomon’s cookie or a high-protein McJomonburger. Recent
Ainu hunter-gatherers kept a ceramic stewpot simmering constantly on the
fire and threw all types of foods together into it; their Jomon predecessors,
living at the same sites and eating the same foods, may have done the same.
I mentioned that their pottery (including heavy pieces up to three feet tall)
suggests Jomon hunter-gatherers to have been sedentary rather than nomadic.
Further evidence of fixed residence comes from their heavy stone tools,
remains of substantial semi-underground houses with signs of remodeling, big
village sites of a hundred or more dwellings, and cemeteries. All of these
features distinguish Jomon people from observed modern hunter-gatherers
who shift base every few weeks, build only shelters, and burden themselves
with few and easily portable possessions. This sedentary lifestyle was made
possible by the diversity of resource-rich habitats available to Jomon people
within a short distance of one central site: inland forests, rivers, seashores,
bays, and open oceans.
Jomon people lived at some of the highest population densities ever
estimated for hunter-gatherers, especially in central and northern Japan with
its nut-rich forests, salmon runs, and productive seas. Estimates of the total
population of Jomon Japan at its peak are 250,000—trivial of course
compared with modern Japan’s, but impressive for hunter-gatherers. Their
closest rivals in modern times would have been American Indians of the
Pacific Northwest coast and of California, subsisting similarly off of nut-rich
forests, salmon runs, and productive seas—a striking case of convergent
evolution of human societies.
With all this stress on what Jomon people did have, we need to be clear as
well about what they did not have. They had no intensive agriculture, and
only questionably any agriculture at all. Apart from dogs (and questionably
pigs), they had no domestic animals. They had no metal tools, no writing, and
no weaving. Jomon villages and cemeteries do not consist of a few richly
decorated houses and graves contrasting with numerous spartan ones but are
instead rather uniform—suggesting that there was little social stratification
into chiefs and commoners. The regional variation in pottery styles suggests
little progress toward political centralization and unification. All of these
negative features contrast with features of contemporary societies only a few
hundred miles distant from Jomon Japan in mainland China and Korea—and
with the changes that swept over Japan itself after 400 B.C.
Despite its distinctiveness even in East Asia at that time, Jomon Japan
was not a completely isolated universe. Distribution of pottery and of
obsidian (a very hard volcanic rock favored for stone tools) shows that Jomon
watercraft connected the Izu island chain stretching 180 miles south from
Tokyo. Pottery, obsidian, and fishhooks similarly testify to some Jomon trade
with Korea, Russia, and Okinawa—as does the arrival of the half-a-dozen
Asian mainland crops that I already mentioned. But archaeologists studying
Jomon Japan have found little evidence of direct imports from China, in
contrast to China’s big influence on subsequent Japanese history. Compared
with later eras, what is impressive about Jomon Japan is not that some contact
with the outside world did occur but that it had so little influence on Jomon
society. Jomon Japan was a conservative miniature universe that maintained
its isolation and changed surprisingly little over the course of 10,000 years—
an island of stability in a fragile, rapidly changing contemporary world.
To place the distinctiveness of Jomon Japan in a contemporary
perspective, let us remind ourselves of what human societies were like on the
Asian mainland a few hundred miles west of Japan in 400 B.C., just as the
Jomon lifestyle was about to come to an end. China consisted then of
kingdoms with rich elites and poorer commoners, living in walled towns, and
on the verge of political unification to become the world’s largest empire.
Beginning around 7500 B.C., China had developed intensive agriculture based
on millets in the north and rice in the south, and with domestic pigs, chickens,
and water buffalo. China had had writing for at least 900 years, and metal
tools for at least 1,500 years, and had just invented the world’s first cast-iron
production. Those Chinese developments were also spreading to Korea,
which had already had agriculture for several thousand years (including rice
since 2200 B.C.) and metal since 1000 B.C.
Given all of these developments going on for thousands of years just
across Tsushima Strait and the East China Sea from Japan, it seems at first
astonishing that Japan was still occupied in 400 B.C. by people who had some
trade with Korea but remained preliterate, stone-tool-using hunter-gatherers.
Throughout human history, centralized states with metal weapons and armies
supported by dense agricultural populations have swept away sparser
populations of stone-tool-using hunter-gatherers. How did Jomon Japan
survive so long?
To understand the answer to this paradox, we have to remember that, until
400 B.C., the frontier of Tsushima Strait separated not rich farmers from poor
hunter-gatherers but poor farmers from rich hunter-gatherers. China itself and
Jomon Japan were not in direct contact. Instead, Japan’s trade contacts, such
as they were, involved Korea. But rice had been domesticated in warm
southern China and spread only slowly northward to much cooler Korea,
because it took a long time to develop new, cold-resistant strains of rice. Early
rice agriculture in Korea used dry-field methods rather than irrigated paddies
and was not particularly productive. Hence early Korean agriculture could not
compete with Jomon hunting and gathering. Jomon people themselves would
have seen no advantage warranting adoption of Korean agriculture, insofar as
they were aware of its existence; and poor Korean farmers possessed no
advantages enabling them to force their way into Japan. As we shall see, the
advantages finally reversed suddenly and dramatically.
I ALREADY MENTIONED THAT THE INVENTION OF POTTERY in Kyushu around
12,700 years ago and the resulting Jomon population explosion were the first
of two decisive changes in Japanese history. The other decisive change, which
triggered a second population explosion, began around 400 B.C. with the
arrival of a new lifestyle (and people?) from South Korea. This second
transition poses in acute form our question about who the Japanese are. Does
the transition mark the replacement of Jomon people with immigrants from
Korea, ancestral to the modern Japanese? Or does it merely mark Japan’s
original Jomon inhabitants continuing to occupy Japan while learning
valuable new tricks?
The new lifestyle appeared first on the north coast of Japan’s
southwesternmost island of Kyushu, immediately across Tsushima Strait from
South Korea. The most important new elements were Japan’s first metal tools,
of iron, and its first undisputed full-scale agriculture. That agriculture came in
the form of irrigated rice fields, complete with canals, dams, banks, paddies,
and rice residues revealed by archaeological excavations. Archaeologists term
the new lifestyle “Yayoi,” after a district of Tokyo where in 1884 its
characteristic pottery was first recognized. Unlike Jomon pottery, Yayoi
pottery had shapes very similar to those of contemporary South Korean
pottery. Among the new Yayoi culture’s many other elements that were
unmistakably Korean but previously foreign to Japan were bronze objects,
weaving, glass beads, underground rice storage pits, the custom of burying
remains of dead people in jars, and Korean styles of tools and houses.
Although rice was the most important Yayoi crop, 27 other crops new to
Japan plus unquestionably domesticated pigs were grown as well. Yayoi
farmers may have practiced double-cropping, with paddies irrigated for rice
production in the summer, then the same fields drained for dry-land
cultivation of millets, barley, and wheat in the winter. Inevitably, this highly
productive system of intensive agriculture triggered an immediate population
explosion in Kyushu, where archaeologists have identified far more Yayoi
sites than Jomon sites, even though the Jomon period lasted 14 times longer.
In virtually no time, Yayoi farming jumped from Kyushu to the adjacent
main islands of Shikoku and Honshu, reaching the Tokyo area within 200
years and the northern tip of Honshu (1,000 miles from the first Yayoi
settlements on Kyushu) in another century. The earliest Yayoi sites on Kyushu
contained pots both in the new Yayoi styles and in the old Jomon styles, but
the latter dropped out as Yayoi culture and pottery spread north through
Honshu. However, some elements of Jomon culture did not vanish
completely. Yayoi farmers continued to use some Jomon types of chipped-
stone tools, which had already been completely replaced by metal tools in
Korea and China. Some Yayoi houses were in Korean styles, some in Jomon
styles. Particularly as Yayoi culture spread north of Tokyo, into cooler areas
where rice farming was less productive and where Jomon hunter-gatherers
had lived in the highest population densities, a mixed Yayoi/Jomon culture
arose, with fishhooks made of metal but in Jomon shapes, and with pots made
in modified Yayoi forms but with Jomon cord marking. After briefly
occupying the cold northern tip of Honshu, Yayoi farmers abandoned that
area, presumably because rice farming just could not compete there with the
Jomon hunter-gatherer lifestyle. For the next 2,000 years, northern Honshu
remained a frontier zone, beyond which the northernmost Japanese island of
Hokkaido and its Ainu hunter-gatherers were not even considered part of the
Japanese state until their annexation in the 19th century.
Yayoi iron tools were initially imported from Korea in enormous
quantities, until domestic Japanese iron smelting and production began after
several centuries. It also took several centuries for Yayoi Japan to exhibit the
first signs of social stratification, as reflected especially in cemeteries. After
about 100 B.C., separate parts of cemeteries began to be set aside for the
graves of what was evidently an emerging elite class, marked by luxury goods
imported from China, such as beautiful jade objects and bronze mirrors. As
the Yayoi population explosion continued, and as all the best swamps or
irrigable plains suitable for wet rice agriculture began to be filled up,
archaeological evidence for war became more and more frequent: mass
production of arrowheads, defensive moats surrounding the villages, and
buried skeletons pierced by projectile points. These hallmarks of war in Yayoi
Japan corroborate the earliest accounts of Japan in Chinese chronicles, which
describe the land of Wa and its hundred little political units fighting with one
another.
In the period from A.D. 300 to 700, both archaeological excavations and
frustratingly ambiguous accounts in later chronicles let us glimpse dimly the
emergence of a politically unified Japan. Before A.D. 300, elite tombs were
small and exhibited a regional diversity of styles. Beginning around A.D. 300,
increasingly enormous earth mound tombs termed kofun, in the shape of a
keyhole, were constructed in Honshu’s Kinai region and then appeared over
the whole former Yayoi culture area, from Kyushu to North Honshu. Why the
Kinai region? Perhaps because it contains some of Japan’s best agricultural
land, where super-expensive Kobe beef is raised today, and where Japan’s
ancient capital was located at Kyoto until the capital’s shift to Tokyo in 1868.
Kofun tombs are up to 1,500 feet long and over 100 feet high, making
them possibly the largest earth mound tombs in the ancient world. The
prodigious amount of labor required to construct them, and the uniformity of
their style over Japan, imply powerful rulers who commanded a huge labor
force and were in the process of achieving Japan’s political unification. Those
kofun that have been excavated contain lavish burial goods, but excavation of
all of the largest ones is still forbidden because they are believed to contain
the ancestors of the Japanese imperial line. This visible evidence of political
centralization that the kofun provide reinforces the accounts of Kofun era
Japanese emperors written down much later in Japanese and Korean
chronicles. Massive Korean influences on Japan during the Kofun era—
whether through Korean conquest of Japan (the Korean view) or Japanese
conquest of Korea (the Japanese view)—transmitted Buddhism, writing,
horse riding, and new ceramic and metallurgical techniques to Japan from the
Asian mainland.
Finally, with the completion of Japan’s first chronicle, in A.D. 712, partly
myth and partly rewritings of true events, Japan emerges into the full light of
history. As of 712, the people inhabiting Japan were at last unquestionably
Japanese, and their language (termed Old Japanese) was unquestionably
ancestral to modern Japanese. Japan’s Emperor Akihito, who reigns today, is
the 82nd direct descendant of the emperor under whom that first chronicle of
A.D. 712 was written. He is traditionally considered the 125th direct
descendant of the legendary first emperor, Jimmu, the great-great-great-
grandson of the sun goddess Amaterasu.
JAPANESE CULTURE UNDERWENT far more radical change in the 700 years of
the Yayoi era than in the ten millennia of Jomon times. The contrast between
Jomon stability (alias conservatism) and radical Yayoi change is the most
striking feature of Japanese history. Obviously, something momentous
happened at 400 B.C. What was it? Were the Jomon people, the Yayoi people,
or a mixture of them the ancestors of the modern Japanese? Japan’s
population increased by the astonishing factor of 70 during Yayoi times: what
caused that change? A passionate debate has raged around three alternative
hypotheses.
One theory is that Jomon hunter-gatherers themselves gradually evolved
into the modern Japanese. Because they had already been living a settled
existence in villages for thousands of years, they may have been pre-adapted
to accepting agriculture. At the Yayoi transition, perhaps nothing more
happened than that Jomon society received cold-resistant rice seeds and
information about paddy irrigation from Korea, enabling people to produce
more food and increase their numbers. This theory appeals to some modern
Japanese, because it minimizes the unwelcome contribution of Korean genes
to the Japanese gene pool, and because it portrays the Japanese people as
uniquely Japanese for at least the last 12,000 years.
A second theory, unappealing to those Japanese who prefer the first
theory, argues instead that the Yayoi transition represents a massive influx of
immigrants from Korea, carrying Korean farming practices, culture, and
genes. Kyushu would have seemed a paradise to Korean rice farmers, because
it is warmer and swampier than Korea and hence a better place to grow rice.
According to one estimate, Yayoi Japan received several million immigrants
from Korea, utterly swamping out the genetic contribution of Jomon people
(thought to have numbered around 75,000 just before the Yayoi transition). If
so, modern Japanese are descendants of Korean immigrants who developed a
modified culture of their own over the last 2,000 years.
The last theory accepts the evidence for immigration from Korea but
denies that it was massive. Instead, highly productive agriculture enabled a
modest number of immigrant rice farmers to reproduce much faster than
Jomon hunter-gatherers and eventually to outnumber them. For instance,
suppose that a mere 5,000 Koreans had come to Kyushu, but that rice
agriculture had enabled them to feed babies and to increase their numbers at a
rate of 1 percent per year. That rate is much higher than observed for hunter-
gatherer populations but is easily attained by farmers: Kenya’s population is
now growing at 4.5 percent per year. In 700 years those 5,000 immigrants
would have left 5,000,000 descendants, again swamping out the Jomon
people. Like the second theory, this one considers modern Japanese to be
slightly modified Koreans but dispenses with the need for large-scale
immigration.
By comparison with similar transitions elsewhere in the world, the second
or third theory seems to me more plausible than the first theory. Over the last
12,000 years, agriculture arose at not more than nine places over the face of
the earth: China, the Fertile Crescent, and a few other places. Twelve
thousand years ago, everybody on earth was a hunter-gatherer; now almost all
of us are farmers or else are fed by farmers. The spread of farming from those
few sites of origin usually did not occur as a result of the hunter-gatherers’
elsewhere adopting farming; hunter-gatherers tend to be conservative, as
Jomon people evidently were from 10,700 to 400 B.C. Instead, farming spread
mainly through farmers’ outbreeding hunters, developing more potent
technology, and then killing the hunters or driving them off of all lands
suitable for agriculture. In modern times, European farmers thereby replaced
western North American Indian hunters, Aboriginal Australians, and the San
people of South Africa. Stone-tool-using farmers similarly replaced hunters
prehistorically throughout Europe, Southeast Asia, and Indonesia. Compared
with the only modest advantage that farmers enjoyed over hunters in these
prehistoric expansions, Korean farmers of 400 B.C. would have enjoyed an
enormous advantage over Jomon hunters, because the Koreans already
possessed iron tools and a highly developed form of intensive agriculture.
Which of the three theories is correct for Japan? The only direct way to
answer this question is to compare Jomon and Yayoi skeletons and genes with
those of modern Japanese and Ainu. Measurements have now been made of
many series of skeletons. In addition, within recent years, molecular
geneticists have begun to extract DNA from ancient human skeletons and to
compare the genes of Japan’s ancient and modern populations. What one
finds is that Jomon and Yayoi skeletons are on the average readily
distinguishable. Jomon people tended to have shorter stature, relatively longer
forearms and lower legs, more wide-set eyes, shorter and wider faces, and
much more pronounced facial “topography” with strikingly raised brow
ridges, noses, and bridges of the nose. Yayoi people averaged an inch or two
taller, had close-set eyes, high and narrow faces, and flat brow ridges and
noses. Some skeletons of the Yayoi period were still Jomon-like in
appearance, but that is to be expected by almost any theory of the
Jomon/Yayoi transition. By the time of the Kofun period, all Japanese
skeletons except those of the Ainu formed a homogeneous group, resembling
modern Japanese and Koreans.
In all of these respects, Jomon skulls differ from those of modern
Japanese and are most similar to those of modern Ainu, while Yayoi skulls
most resemble those of modern Japanese. On the assumption that modern
Japanese people arose as a mixture of a Korean-like Yayoi population with an
Ainu-like Jomon population, geneticists have attempted to calculate the
relative contributions of the two gene pools. The resulting conclusion is that
the Korean/Yayoi contribution was generally dominant. The Ainu/Jomon
contribution was lowest in southwest Japan, where most Korean immigrants
would have arrived and Jomon populations were sparse, and relatively greater
in northern Japan, where forests were richer in nuts, Jomon population
densities were highest, and Yayoi rice agriculture was least successful.
Thus, immigrants from Korea really did make a big contribution to the
modern Japanese, though we cannot yet say whether that was because of
massive immigration or else modest immigration amplified by a high rate of
population increase. The Ainu are more nearly the descendants of Japan’s
ancient Jomon inhabitants, mixed with Korean genes of Yayoi colonists and
of the modern Japanese.
Given the overwhelming advantage that rice agriculture finally gave to
Korean farmers over Jomon hunters, one has to wonder why the farmers
achieved victory so suddenly, after making little headway in Japan for
thousands of years after farming reached Korea. I already mentioned that
early Korean farming was relatively unproductive and resulted only in poor
farmers outclassed by rich hunters. What finally tipped the balance to the
farmers and triggered the Yayoi transition was probably a combination of four
factors coming together: the development of irrigated rice agriculture, instead
of less productive dry-field rice agriculture; the continuing improvement of
rice strains adapted to a cool climate; the growth in Korea’s farming
population, putting pressure on Koreans to emigrate; and the development of
iron tools for efficiently mass-producing the wooden shovels, hoes, and other
tools needed for rice paddy agriculture. The fact that iron and intensive
farming reached Japan simultaneously is unlikely to be a coincidence.
I BEGAN THIS PIECE BY MENTIONING A TRANSPARENT interpretation for how the
distinctive-looking Ainu and the undistinctive-looking Japanese came to share
Japan. On the face of it, these facts would appear to suggest that the Ainu are
descended from Japan’s original inhabitants, and that the Japanese are
descended from more recent arrivals. We have now seen that the combined
evidence of archaeology, physical anthropology, and genetics supports this
view.
But I also mentioned at the outset a potent objection that causes most
people (especially the Japanese themselves) to seek other interpretations. If
the Japanese really are recent arrivals from Korea, you might expect the
Japanese and Korean languages to be very similar to each other. More
generally, if the Japanese people arose recently from some mixture, on the
island of Kyushu, of original Ainu-like Jomon inhabitants with Yayoi
invaders from Korea, the Japanese language might show close affinities to
both the Korean and the Ainu languages. Instead, Japanese and Ainu have no
demonstrable relationship, and the relationship between Japanese and Korean
is distant. How could this be so if the mixing occurred a mere 2,400 years
ago? I suggest the following resolution of this paradox: that the languages of
the Kyushu Jomon residents and of the Yayoi invaders were in fact unlikely to
have been very similar to the modern Ainu and Korean languages,
respectively.
Taking first the Ainu language, that language as we know it is the one that
was spoken in recent times by the Ainu on the northern Japanese island of
Hokkaido. Therefore, Hokkaido’s Jomon inhabitants also probably spoke an
Ainu-like language, but the Jomon inhabitants of Kyushu surely did not. From
the southern tip of Kyushu to the northern tip of Hokkaido, the Japanese
archipelago is nearly 1,500 miles long. We know that in Jomon times it
supported great regional diversity of subsistence techniques and of pottery
styles and was never unified politically. During the 10,000 years of Jomon
occupation, Jomon people would have evolved correspondingly great
linguistic diversity. Their languages may even have been already diverse over
12,000 years ago, if the northern and southern Jomon people arrived over land
bridges from Russia and Korea, respectively, as the archaeological evidence
seems to indicate.
In fact, many Japanese place-names on Hokkaido and northern Honshu
include the Ainu words for “river” ( nai or betsu) or cape ( shiri), but such
Ainu-like names do not occur farther south in Japan. This suggests that Yayoi
and Japanese pioneers adopted many local Jomon place-names, just as white
Americans did from Native Americans (think of “Massachusetts,”
“Mississippi,” and so on), but that Ainu was the Jomon language only of
northernmost Japan. The Jomon language of Kyushu may instead have shared
a common ancestor with the Austronesian language family, which includes
Polynesian and Indonesian languages and the Aboriginal languages of
Taiwan. As many linguists have pointed out, the Japanese language shows
some influence of Austronesian languages in the shared preference for so-
called open syllables (a consonant followed by a vowel, as in “Hi-ro-hi-to”).
Ancient Taiwanese were great seafarers whose descendants spread out far to
the south, east, and west; some of them may also have spread north, to
Kyushu.
That is, the modern Ainu language of Hokkaido is not a model for the
ancient Jomon language of Kyushu. By the same token, the modern Korean
language may be a poor model for the ancient Yayoi language of Korean
immigrants in 400 B.C. In the centuries before Korea became unified
politically in A.D. 676, it consisted of three kingdoms. The modern Korean
language is derived from the language of the kingdom of Silla, the kingdom
that emerged triumphant and unified Korea, but Silla was not the kingdom
that had had close contact with Japan in the preceding centuries. Early Korean
chronicles tell us that the different kingdoms had different languages. While
the languages of the kingdoms defeated by Silla are poorly known, the few
preserved words of one of those kingdoms (Koguryo) are much more similar
to the corresponding Old Japanese words than are the corresponding modern
Korean words. Korean languages may have been even more diverse in 400
B.C., before political unification had reached the stage of three kingdoms. I
suspect that the Korean language that was carried to Japan in 400 B.C., and
that evolved into modern Japanese, was quite different from the Silla
language that evolved into modern Korean. Hence we should not be surprised
that modern Japanese and Korean people resemble each other far more in
their appearance and genes than in their languages.
This conclusion is likely to be equally unpopular in Japan and in Korea,
because of the current mutual dislike of those two peoples. History gives them
good reason to dislike each other: especially, for Koreans to dislike Japanese.
Like Arabs and Jews, Koreans and Japanese are peoples joined by blood, yet
locked in traditional enmity. But enmity is mutually destructive, in East Asia
and in the Middle East. Reluctant as Japanese and Koreans are to admit it,
they are like twin brothers who shared their formative years. The political
future of East Asia depends in large part on their success in rediscovering
those ancient bonds between them.
2003 AFTERWORD: Guns, Germs, and Steel Today
G UNS, GERMS, AND STEEL (GGS) IS ABOUT WHY THE RISE OF complex human
societies unfolded differently on different continents over the last 13,000
years. I finished revising the manuscript in 1996, and it was published in
1997. Since then, I have been involved mostly in work on other projects,
especially on my next book about collapses of societies. Hence seven years’
distance in time and focus now separates me from GGS’s writing. How does
the book look in retrospect, and what has happened to change or extend its
conclusions since its publication? To my admittedly biased eye, the book’s
central message has survived well, and the most interesting developments
since its publication have involved four extensions of the story to the modern
world and to recent history.
My main conclusion was that societies developed differently on different
continents because of differences in continental environments, not in human
biology. Advanced technology, centralized political organization, and other
features of complex societies could emerge only in dense sedentary
populations capable of accumulating food surpluses—populations that
depended for their food on the rise of agriculture that began around 8,500 B.C.
But the domesticable wild plant and animal species essential for that rise of
agriculture were distributed very unevenly over the continents. The most
valuable domesticable wild species were concentrated in only nine small
areas of the globe, which thus became the earliest homelands of agriculture.
The original inhabitants of those homelands thereby gained a head start
toward developing guns, germs, and steel. The languages and genes of those
homeland inhabitants, as well as their livestock, crops, technologies, and
writing systems, became dominant in the ancient and modern world.
Discoveries in the last half-dozen years, by archaeologists, geneticists,
linguists, and other specialists, have enriched our understanding of this story,
without changing its main outlines. Let me mention three examples. One of
the biggest gaps in GGS’s geographic coverage involved Japan, about whose
prehistory I had little to say in 1996. Recent genetic evidence now suggests
that the modern Japanese people are the product of an agricultural expansion
similar to others discussed in GGS: an expansion of Korean farmers,
beginning around 400 B.C., into southwestern Japan and then advancing
northeast up the Japanese archipelago. The immigrants brought intensive rice
agriculture and metal tools, and they mixed with the original Japanese
population (related to the modern Ainu) to produce the modern Japanese,
much as expanding Fertile Crescent farmers mixed with Europe’s original
hunter/gatherer population to produce modern Europeans.
As another example, archaeologists originally assumed that Mexican
corn, beans, and squashes reached the southeastern United States by the most
direct route via northeastern Mexico and eastern Texas. But it is now
becoming clear that this route was too dry for farming; those crops instead
took a longer route, spreading from Mexico into the southwestern United
States to trigger the rise of Anasazi societies there, and then spreading east
from New Mexico and Colorado through river valleys of the Great Plains into
the southeastern United States.
As a final example, in Chapter 10 I contrasted the frequency of repeated
independent domestications and slow spreads of the same or related plants
along the Americas’ north/south axis with the predominantly single
domestications and rapid east/west spreads of Eurasian crops. Even more
examples of those two contrasting patterns have continued to turn up, but it
now appears that most or all of Eurasia’s Big Five domestic mammals also
underwent repeated independent domestications in different parts of Eurasia
—unlike Eurasia’s plants, but like the Americas’ plants.
These and other discoveries add details, which continue to fascinate me,
to our understanding of how agriculture’s rise triggered the rise of
agriculturally based complex societies in the ancient world. However, the
biggest advances building on GGS have involved extensions into areas that
were not the book’s main focus. Since publication, thousands of people have
written, phoned, e-mailed, or buttonholed me to tell me of parallels or
contrasts that they noticed between the ancient continental processes of GGS
and the modern or recent processes that they study. I’ll tell you about four of
these revelations: briefly, the illuminating example of New Zealand’s Musket
Wars; the perennial question “Why Europe, not China?” in more detail,
parallels between competition in the ancient world and in the modern business
world; and GGS’s relevance to why some societies today are rich while others
are poor.
IN 1996 I DEVOTED one brief paragraph (in Chapter 13) to a phenomenon in
19th-century New Zealand history termed the Musket Wars, as an illustration
of how powerful new technologies spread. The Musket Wars were a
complicated, poorly understood series of tribal wars among New Zealand’s
indigenous Maori people, between 1818 and the 1830s—wars by which
European guns spread among tribes that had previously fought one another
with stone and wooden weapons. Two books published since then have
increased our understanding of that chaotic period of New Zealand history,
placed it in a broader historical context, and made its relevance to GGS even
clearer.
In the early 1800s, European traders, missionaries, and whalers began to
visit New Zealand, which had been occupied 600 years previously by
Polynesian farmers and fishermen known as Maoris. The first European
visitors were concentrated at New Zealand’s northern end. Those northern
Maori tribes with the earliest access to Europeans thereby became the first
tribes to acquire muskets, which gave them a big military advantage over all
the other tribes lacking muskets. They used that advantage to settle scores
with neighboring tribes that were their traditional enemies. But they also used
muskets for a new type of warfare: long-distance raids against Maori tribes
hundreds of miles away, carried out in order to outdo rivals in acquiring
slaves and prestige.
At least as important as European muskets in making long-distance raids
feasible were European-introduced potatoes (originating in South America),
which yielded many more tons of food per acre or per farmer than did
traditional Maori agriculture based on sweet potatoes. The main limitation
that had previously prevented Maoris from undertaking long raids had been
the twin problems of feeding warriors away from home for a long time, and
feeding the at-home population of women and children dependent on the
would-be warriors to stay home and grow sweet potatoes. Potatoes solved that
bottleneck. Hence a less heroic term for the Musket Wars would be the Potato
Wars.
Whatever they are called, the Musket/Potato Wars proved very
destructive, killing about one-quarter of the original Maori population. The
highest body counts arose when a tribe with lots of muskets and potatoes
attacked a tribe with few or none. Of the tribes not among the first to acquire
muskets and potatoes, some were virtually exterminated before they could
acquire them, while others made determined efforts to acquire them and
thereby restore the previous military equilibrium. One episode in these wars
was the conquest and mass killing of Moriori tribes by Maori tribes, as
described in Chapter 2.
The Musket/Potato Wars illustrate the main process running through the
history of the last 10,000 years: human groups with guns, germs, and steel, or
with earlier technological and military advantages, spreading at the expense
of other groups, until either the latter groups became replaced or everyone
came to share the new advantages. Recent history furnishes innumerable
examples as Europeans expanded to other continents. In many places the non-
European locals never got a chance to acquire guns and ended up losing their
lives or their freedom. However, Japan did succeed in acquiring (actually,
reacquiring) guns, preserved its independence, and within 50 years used its
new guns to defeat a European power in the Russo-Japanese war of 1904–5.
North American Plains Indians, South American Araucanian Indians, New
Zealand’s Maoris, and Ethiopians acquired guns and used them to hold off
European conquest for a long time, though they were ultimately defeated.
Today, Third World countries are doing their best to catch up with the First
World by acquiring the latter’s technological and agricultural advantages.
Such spreads of technology and agriculture, arising ultimately from
competition between human groups, must have happened at innumerable
other times and places over the past 10,000 years.
In that sense, there was nothing unusual about New Zealand’s
Musket/Potato Wars. While those wars were a purely local phenomenon
confined to New Zealand, they are of worldwide interest because they furnish
such a clear example, so narrowly confined in space and time, of so many
other similar local phenomena. Within about two decades following their
introduction to the northern end of New Zealand, muskets and potatoes had
spread 900 miles to the southern end of New Zealand. In the past, agriculture,
writing, and improved pre-gun weapons took much longer to spread much
greater distances, but the underlying social processes of population
replacement and competition were essentially the same. Now we are
wondering whether nuclear weapons will proliferate around the world by the
same often-violent process, from the eight countries that presently possess
them.
A SECOND AREA of active discussion since 1997 falls under a heading that
could be termed “Why Europe, not China?” Most of GGS concerned
differences between continents: i.e., the question of why some Eurasians
rather than Aboriginal Australians, sub-Saharan Africans, or Native
Americans were the ones to expand over the world within the past
millennium. However, I realized that many readers would also wonder “Why,
among Eurasians, was it Europeans rather than Chinese or some other group
that expanded?” I knew that my readers would not let me get away with
concluding GGS without saying anything about this obvious question.
Hence I briefly considered it in the book’s epilogue. I suggested that the
underlying reason behind Europe’s overtaking China was something deeper
than the proximate factors suggested by most historians (e.g., China’s
Confucianism vs. Europe’s Judeo-Christian tradition, the rise of western
science, the rise of European mercantilism and capitalism, Britain’s
deforestation coupled with its coal deposits, etc.). Behind these and other
proximate factors, I saw an “Optimal Fragmentation Principle”: ultimate
geographic factors that led to China becoming unified early and mostly
remaining unified thereafter, while Europe remained constantly fragmented.
Europe’s fragmentation did, and China’s unity didn’t, foster the advance of
technology, science, and capitalism by fostering competition between states
and providing innovators with alternative sources of support and havens from
persecution.
Historians have subsequently pointed out to me that Europe’s
fragmentation, China’s unity, and Europe’s and China’s relative strengths
were all more complex than depicted in my account. The geographic
boundaries of the political/social spheres that could usefully be grouped as
“Europe” or “China” fluctuated over the centuries. China led Europe in
technology at least until the 15th century and might do so again in the future,
in which case the question “Why Europe, not China?” might only refer to an
ephemeral phenomenon without deep explanation. Political fragmentation has
more complex effects than only providing a constructive forum for
competition: for instance, competition can be destructive as well as
constructive (think of World Wars I and II). Fragmentation itself is a
multifaceted rather than a monolithic concept: its effect on innovation
depends on factors such as the freedom with which ideas and people can
move across the boundaries between fragments, and whether the fragments
are distinct or just clones of each other. Whether fragmentation is “optimal”
may also vary with the measure of optimality used; a degree of political
fragmentation that is optimal for technological innovation may not be optimal
for economic productivity, political stability, or human happiness.
My sense is that a large majority of social scientists still favors proximate
explanations for the different courses of European and Chinese history. For
example, in a thoughtful recent essay Jack Goldstone stressed the importance
of Europe’s (especially Britain’s) “engine science,” meaning the applications
of science to the development of machines and engines. Goldstone wrote,
“Two problems faced all pre-industrial economies in regard to energy: amount
and concentration. The amount of mechanical energy available to any pre-
industrial economy was limited to water flows, animals or people who could
be fed, and wind that could be captured. In any geographically fixed area, this
amount was strictly limited.…It is difficult to overstate the advantage given to
the first economy or military/political power to devise a means to extract
useful work from the energy in fossil fuels…. [It was] the application of
steam power to spinning, to water and surface transport, to brick-making,
grain-threshing, iron-making, shoveling, construction, and all sorts of
manufacturing processes that transformed Britain’s economy…. It thus may
be that, far from a necessary development of European civilization, the rich
development of engine science was the chance outcome of specific, even if
highly contingent, circumstances that happened to arise in 17th- and 18th-
century Britain.” If this reasoning is correct, then a search for deep
geographic or ecological explanations will not be profitable.
The opposite minority view, similar to my view expressed in the epilogue
of GGS, has been argued in detail by Graeme Lang: “Differences between
Europe and China in ecology and geography helped to explain the very
different fates of science in the two regions. First, [rainfall] agriculture in
Europe provided no role for the state, which remained far from local
communities most of the time, and when the agricultural revolution in Europe
produced a growing agricultural surplus, this allowed the growth of relatively
autonomous towns along with urban institutions such as universities prior to
the rise of the centralized states in the late Middle Ages. [Irrigation and water-
control] agriculture in China, by contrast, favored the early development of
intrusive and coercive states in the major river valleys, while towns and their
institutions never achieved the degree of local autonomy found in Europe.
Second, the geography of China, unlike that of Europe, did not favor the
prolonged survival of independent states. Instead, China’s geography
facilitated eventual conquest and unification over a vast area, followed by
long periods of relative stability under imperial rule. The resulting state
system suppressed most of the conditions required for the emergence of
modern science…. The explanation outlined above is certainly
oversimplified. However, one of the advantages of this kind of account is that
it escapes the circularity which often creeps into explanations which do not go
deeper than social or cultural differences between Europe and China. Such
explanations can always be challenged with a further question: why were
Europe and China different with regard to those social or cultural factors?
Explanations rooted ultimately in geography and ecology, however, have
reached bedrock.”
It remains a challenge for historians to reconcile these different
approaches to answering the question “Why Europe, not China.” The answer
may have important consequences for how best to govern China and Europe
today. For example, from Lang’s and my perspective, the disaster of China’s
Cultural Revolution of the 1960s and 1970s, when a few misguided leaders
were able to close the school systems of the world’s largest country for five
years, may not be a unique one-time-only aberration, but may presage more
such disasters in the future unless China can introduce far more
decentralization into its political system. Conversely, Europe, in its rush
toward political and economic unity today, will have to devote much thought
to how to avoid dismantling the underlying reason behind its successes of the
last five centuries.
THE THIRD RECENT extension of GGS’s message to the modern world was to
me the most unexpected one. Soon after the book’s publication, it was
reviewed favorably by Bill Gates, and then I began receiving letters from
other business people and economists who pointed out possible parallels
between the histories of entire human societies discussed in GGS and the
histories of groups in the business world. This correspondence concerned the
following broad question: what is the best way to organize human groups,
organizations, and businesses so as to maximize productivity, creativity,
innovation, and wealth? Should your group have a centralized direction (in
the extreme, a dictator), or should there be diffuse leadership or even
anarchy? Should your collection of people be organized into a single group,
or broken down into a small or large number of groups? Should you maintain
open communication between your groups, or erect walls of secrecy between
them? Should you erect protectionist tariff walls against the outside, or should
you expose your business to free competition?
These questions arise at many different levels and for many types of
groups. They apply to the organization of entire countries: remember the
perennial arguments about whether the best form of government is a benign
dictatorship, a federal system, or an anarchical free-for-all. The same
questions arise about the organization of different companies within the same
industry. How can we account for the fact that Microsoft has been so
successful recently, while IBM, which was formerly successful, fell behind
but then drastically changed its organization and improved its success? How
can we explain the different successes of different industrial belts? When I
was a boy growing up in Boston, Route 128, the industrial belt around
Boston, led the world in scientific creativity and imagination. But Route 128
has fallen behind, and now Silicon Valley is the center of innovation. The
relations of businesses to one another in Silicon Valley and on Route 128 are
very different, possibly resulting in those different outcomes.
Of course, there are also the famous differences between the
productivities of the economies of whole countries, such as Japan, the United
States, France, and Germany. Actually, though, there are big differences
between the productivity and wealth of different business sectors even within
the same country. For example, the Korean steel industry is equal in
efficiency to ours, but all other Korean industries lag behind their American
counterparts. What is it about the different organization of these various
Korean industries that accounts for their differences in productivity within the
same country?
Obviously, answers to these questions about differences in organizational
success depend partly on the idiosyncrasies of individuals. For example, the
success of Microsoft has surely had something to do with the personal talents
of Bill Gates. Even with a superior corporate organization, Microsoft would
not be successful with an ineffectual leader. Nevertheless, one can still ask: all
other things being equal, or else in the long run, or else on the average, what
form of organization of human groups is best?
My comparison of the histories of China, the Indian subcontinent, and
Europe in the epilogue of GGS suggested an answer to this question as
applied to technological innovation in whole countries. As explained in the
preceding section, I inferred that competition between different political
entities spurred innovation in geographically fragmented Europe, and that the
lack of such competition held innovation back in unified China. Would that
mean that a higher degree of political fragmentation than Europe’s would be
even better? Probably not: India was geographically even more fragmented
than Europe, but less innovative technologically. This suggested to me the
Optimal Fragmentation Principle: innovation proceeds most rapidly in a
society with some optimal intermediate degree of fragmentation: a too-unified
society is at a disadvantage, and so is a too-fragmented society.
This inference rang a bell with Bill Lewis and other executives of
McKinsey Global Institute, a leading consulting firm based in Washington,
D.C., which carries out comparative studies of the economies of countries and
industries all over the world. The executives were so struck by the parallels
between their business experience and my historical inferences that they
presented a copy of GGS to each of the firm’s several hundred partners, and
they presented me with copies of their reports on the economies of the United
States, France, Germany, Korea, Japan, Brazil, and other countries. They, too,
detected a key role of competition and group size in spurring innovation. Here
are some of the conclusions that I gleaned from conversations with McKinsey
executives and from their reports:
We Americans often fantasize that German and Japanese industries are
super-efficient, exceeding American industries in productivity. In reality,
that’s not true: on the average across all industries, America’s industrial
productivity is higher than that in either Japan or Germany. But those average
figures conceal big differences among the industries of each country, related
to differences in organization—and those differences are very instructive. Let
me give you two examples from McKinsey case studies on the German beer
industry and the Japanese food-processing industry.
Germans make wonderful beer. Every time that my wife and I fly to
Germany for a visit, we carry with us an empty suitcase, so we can fill it with
bottles of German beer to bring back to the United States and enjoy over the
following year. Yet the productivity of the German beer industry is only 43
percent that of the U.S. beer industry. Meanwhile, the German metalworking
and steel industries are equal in productivity to their American counterparts.
Since the Germans are evidently perfectly capable of organizing industries
well, why can’t they do so when it comes to beer?
It turns out that the German beer industry suffers from small-scale
production. There are a thousand tiny beer companies in Germany, shielded
from competition with one another because each German brewery has
virtually a local monopoly, and they are also shielded from competition with
imports. The United States has 67 major beer breweries, producing 23 billion
liters of beer per year. All of Germany’s 1,000 breweries combined produce
only half as much. Thus the average U.S. brewery produces 31 times more
beer than the average German brewery.
This fact results from local tastes and German government policies.
German beer drinkers are fiercely loyal to their local brand, so there are no
national brands in Germany analogous to our Budweiser, Miller, or Coors.
Instead, most German beer is consumed within 30 miles of the factory where
it is brewed. Therefore, the German beer industry cannot profit from
economies of scale. In the beer business, as in other businesses, production
costs decrease greatly with scale. The bigger the refrigerating unit for making
beer, and the longer the assembly line for filling bottles with beer, the lower
the cost of manufacturing beer. Those tiny German beer companies are
relatively inefficient. There’s no competition; there are just a thousand local
monopolies.
The local beer loyalties of individual German drinkers are reinforced by
German laws that make it hard for foreign beers to compete in the German
market. The German government has so-called beer purity laws that specify
exactly what can go into beer. Not surprisingly, those government purity
specifications are based on what German breweries put into beer, and not on
what American, French, and Swedish breweries like to put into beer. Because
of those laws, not much foreign beer gets exported to Germany, and because
of inefficiency and high prices much less of that wonderful German beer than
you would otherwise expect gets sold abroad. (Before you object that German
Löwenbräu beer is widely available in the United States, please read the label
on the next bottle of Löwenbräu that you drink here: it’s not produced in
Germany but in North America, under license, in big factories with North
American productivity and efficiencies of scale.)
The German soap industry and consumer electronics industry are
similarly inefficient; their companies are not exposed to competition with one
another, nor are they exposed to foreign competition, and so they do not
acquire the best practices of international industry. (When is the last time that
you bought an imported TV set made in Germany?) But those disadvantages
are not shared by the German metal and steel industries, in which big German
companies have to compete with one another and internationally, and thus are
forced to acquire the best international practices.
My other favorite example from the McKinsey reports concerns the
Japanese food-processing industry. We Americans tend to be paranoid about
Japanese efficiency, and it is indeed formidable in some industries—but not in
food-processing. The efficiency of the Japanese food-processing industry is a
miserable 32 percent that of ours. There are 67,000 food-processing
companies in Japan, compared to only 21,000 in the United States, which has
twice Japan’s population—so the average U.S. food-processing company is
six times bigger than its Japanese counterpart. Why does the Japanese food-
processing industry, like the German beer industry, consist of small
companies with local monopolies? Basically, the answer is the same two
reasons: local taste and government policies.
The Japanese are fanatics for fresh food. A container of milk in a U.S.
supermarket bears only one date: the expiration date. When my wife and I
visited a Tokyo supermarket with one of my wife’s Japanese cousins, we were
surprised to discover that in Japan a milk container bears three dates: the date
the milk was manufactured, the date it arrived at the supermarket, and the
expiration date. Milk production in Japan always starts at one minute past
midnight, so that the milk that goes to market in the morning can be labeled
as today’s milk. If the milk were produced at 11:59 P.M., the date on the
container would have to indicate that the milk was made yesterday, and no
Japanese consumer would buy it.
As a result, Japanese food-processing companies enjoy local monopolies.
A milk producer in northern Japan cannot hope to compete in southern Japan,
because transporting milk there would take an extra day or two, a fatal
disadvantage in the eyes of consumers. These local monopolies are reinforced
by the Japanese government, which obstructs the import of foreign processed
food by imposing a 10-day quarantine, among other restrictions. (Imagine
how Japanese consumers who shun food labeled as only one day old feel
about food 10 days old.) Hence Japanese food-producing companies are not
exposed to either domestic or foreign competition, and they don’t learn the
best international methods for producing food. Partly as a result, food prices
in Japan are very high: the best beef costs $200 a pound, while chicken costs
$25 a pound.
Some other Japanese industries are organized very differently from the
food processors. For instance, Japanese steel, metal, car, car parts, camera,
and consumer electronic companies compete fiercely and have higher
productivities than their U.S. counterparts. But the Japanese soap, beer, and
computer industries, like the Japanese food-processing industry, are not
exposed to competition, do not apply the best practices, and thus have lower
productivities than the corresponding industries in the United States. (If you
look around your house, you are likely to find that your TV set and camera,
and possibly also your car, are Japanese, but that your computer and soap are
not.)
Finally, let’s apply these lessons to comparing different industrial belts or
businesses within the United States. Since the publication of GGS, I’ve spent
much time talking with people from Silicon Valley and from Route 128, and
they tell me that these two industrial belts are quite different in terms of
corporate ethos. Silicon Valley consists of lots of companies that are fiercely
competitive with one another. Nevertheless, there is much collaboration—a
free flow of ideas, people, and information among companies. In contrast, I’m
told, the businesses of Route 128 are much more secretive and insulated from
one another, like Japanese milk-producing companies.
What about the contrast between Microsoft and IBM? Since GGS was
published, I’ve acquired friends at Microsoft and have learned about that
corporation’s distinctive organization. Microsoft has lots of units, each
comprised of 5 to 10 people, with free communication among units, and the
units are not micromanaged; they are allowed a great deal of freedom in
pursuing their own ideas. That unusual organization at Microsoft—which in
essence is broken into many competing semi-independent units—contrasts
with the organization at IBM, which until some years ago consisted of much
more insulated groups and resulted in IBM’s loss of competitive ability. Then
IBM acquired a new chief executive officer who changed things drastically:
IBM now has a more Microsoft-like organization, and I’m told that IBM’s
innovativeness has improved as a result.
All of this suggests that we may be able to extract a general principle
about group organization. If your goal is innovation and competitive ability,
you don’t want either excessive unity or excessive fragmentation. Instead,
you want your country, industry, industrial belt, or company to be broken up
into groups that compete with one another while maintaining relatively free
communication—like the U.S. federal government system, with its built-in
competition between our 50 states.
THE REMAINING EXTENSION of GGS has been into one of the central questions
of world economics: why are some countries (like the United States and
Switzerland) rich, while other countries (like Paraguay and Mali) are poor?
Per-capita gross national products (GNP) of the world’s richest countries are
more than 100 times those of the poorest countries. This is not just a
challenging theoretical question giving employment to economics professors,
but also one with important policy implications. If we could identify the
answers, then poor countries could concentrate on changing the things that
keep them poor and on adopting the things that make other countries rich.
Obviously, part of the answer depends on differences in human
institutions. The clearest evidence for this view comes from pairs of countries
that divide essentially the same environment but have very different
institutions and, associated with those institutions, different per-capita GNPs.
Four flagrant examples are the comparison of South Korea with North Korea,
the former West Germany with the former East Germany, the Dominican
Republic with Haiti, and Israel with its Arab neighbors. Among the many
“good institutions” often invoked to explain the greater wealth of the first-
named country of each of these pairs are effective rule of law, enforcement of
contracts, protection of private property rights, lack of corruption, low
frequency of assassinations, openness to trade and to flow of capital,
incentives for investment, and so on.
Undoubtedly, good institutions are indeed part of the answer to the
different wealths of nations. Many, perhaps most, economists go further and
believe that good institutions are overwhelmingly the most important
explanation. Many governments, agencies, and foundations base their
policies, foreign aid, and loans on this explanation, by making the
development of good institutions in poor countries their top priority.
But there is increasing recognition that this good-institutions view is
incomplete—not wrong, just incomplete—and that other important factors
need addressing if poor countries are to become rich. This recognition has its
own policy implications. One cannot just introduce good institutions to poor
countries like Paraguay and Mali and expect those countries to adopt the
institutions and achieve the per-capita GNPs of the United States and
Switzerland. The criticisms of the good-institutions view are of two main
types. One type recognizes the importance of other proximate variables
besides good institutions, such as public health, soil- and climate-imposed
limits on agricultural productivity, and environmental fragility. The other type
concerns the origin of good institutions.
According to the latter criticism, it is not enough to consider good
institutions as a proximate influence whose origins are of no further practical
interest. Good institutions are not a random variable that could have popped
up anywhere around the globe, in Denmark or in Somalia, with equal
probability. Instead, it seems to me that, in the past, good institutions always
arose because of a long chain of historical connections from ultimate causes
rooted in geography to the proximate dependent variables of the institutions.
We must understand that chain if we hope, now, to produce good institutions
quickly in countries lacking them.
At the time that I wrote GGS, I commented, “The nations rising to new
power [today] are still ones that were incorporated thousands of years ago
into the old centers of dominance based on food production, or that have been
repopulated by peoples from those centers…. The hand of history’s course at
8,000 B.C. lies heavily on us.” Two new papers by economists (Olsson and
Hibbs, and Bockstette, Chanda, and Putterman) have subjected this postulated
heavy hand of history to detailed tests. It turns out that countries in regions
with long histories of state societies or agriculture have higher per-capita GNP
than countries with short histories, even after other variables have been
controlled. The effect explains a large fraction of the variance in GNP. Even
just among countries with still-low or recently low GNPs, countries in regions
with long histories of state societies or agriculture, like South Korea, Japan,
and China, have higher growth rates than countries with short histories, such
as New Guinea and the Philippines, even though some of the countries with
short histories are much richer in natural resources.
There are many obvious reasons for these effects of history, such as that
long experience of state societies and agriculture implies experienced
administrators, experience with market economies, and so on. Statistically,
part of that ultimate effect of history proves to be mediated by the familiar
proximate causes of good institutions. But there is still a large effect of history
remaining after one controls for the usual measures of good institutions.
Hence there must be other mediating proximate mechanisms as well. Thus a
key problem will be to understand the detailed chain of causation from a long
history of state societies and agriculture to modern economic growth, in order
to help developing countries advance up that chain more quickly.
In short, the themes of GGS seem to me to be not only a driving force in
the ancient world but also a ripe area for study in the modern world.
ACKNOWLEDGMENTS
IT IS A PLEASURE FOR ME TO ACKNOWLEDGE THE CONTRIBUTIONS of many people
to this book. My teachers at Roxbury Latin School introduced me to the
fascination of history. My great debt to my many New Guinea friends will be
obvious from the frequency with which I cite their experiences. I owe an
equally great debt (and absolution from responsibility for my errors) to my
many scientist friends and professional colleagues, who patiently explained
the subtleties of their subjects and read my drafts. In particular, Peter
Bellwood, Kent Flannery, Patrick Kirch, and my wife, Marie Cohen, read the
whole manuscript, and Charles Heiser, Jr., David Keightley, Bruce Smith,
Richard Yarnell, and Daniel Zohary each read several chapters. Earlier
versions of several of the chapters appeared as articles in Discover magazine
and in Natural History magazine. The National Geographic Society, World
Wildlife Fund, and University of California at Los Angeles supported my
fieldwork on Pacific islands. I have been fortunate to have John Brockman
and Katinka Matson as my agents, Lori Iversen and Lori Rosen as my
research assistants and secretaries, Ellen Modecki as my illustrator, and as my
editors Donald Lamm at W. W. Norton, Neil Belton and Will Sulkin at
Jonathan Cape, Willi Köhler at Fischer, Marc Zabludoff and Mark Wheeler
and Polly Shulman at Discover, and Ellen Goldensohn and Alan Ternes at
Natural History.
FURTHER READINGS
THESE SUGGESTIONS ARE FOR THOSE INTERESTED IN READING further. Hence, in
addition to key books and papers, I have favored references that provide
comprehensive listings of the earlier literature. A journal title (in italics) is
followed by the volume number, followed after a colon by the first and last
page numbers, and then the year of publication in parentheses.
Prologue
Among references relevant to most chapters of this book is an enormous
compendium of human gene frequencies entitled The History and Geography
of Human Genes, by L. Luca Cavalli-Sforza, Paolo Menozzi, and Alberto
Piazza (Princeton: Princeton University Press, 1994). This remarkable book
approximates a history of everything about everybody, because the authors
begin their accounts of each continent with a convenient summary of the
continent’s geography, ecology, and environment, followed by the prehistory,
history, languages, physical anthropology, and culture of its peoples. L. Luca
Cavalli-Sforza and Francisco Cavalli-Sforza, The Great Human Diasporas
(Reading, Mass.: Addison-Wesley, 1995), covers similar material but is
written for the general reader rather than for specialists.
Another convenient source is a series of five volumes entitled The
Illustrated History of Humankind, ed. Göran Burenhult (San Francisco:
HarperCollins, 1993–94). The five individual volumes in this series are
entitled, respectively, The First Humans, People of the Stone Age, Old World
Civilizations, New World and Pacific Civilizations, and Traditional Peoples
Today.
Several series of volumes published by Cambridge University Press
(Cambridge, England, various dates) provide histories of particular regions or
eras. One series consists of books entitled The Cambridge History of [X],
where X is variously Africa, Early Inner Asia, China, India, Iran, Islam,
Japan, Latin America, Poland, and Southeast Asia. Another series is The
Cambridge Encyclopedia of [X], where X is variously Africa, China, Japan,
Latin America and the Caribbean, Russia and the former Soviet Union,
Australia, the Middle East and North Africa, and India, Pakistan, and adjacent
countries. Still other series include The Cambridge Ancient History, The
Cambridge Medieval History, The Cambridge Modern History, The
Cambridge Economic History of Europe, and The Cambridge Economic
History of India.
Three encyclopedic accounts of the world’s languages are Barbara
Grimes, Ethnologue: Languages of the World, 13th ed. (Dallas: Summer
Institute of Linguistics, 1996), Merritt Ruhlen, A Guide to the World’s
Languages (Stanford: Stanford University Press, 1987), and C. F. Voegelin
and F. M. Voegelin, Classification and Index of the World’s Languages (New
York: Elsevier, 1977).
Among large-scale comparative histories, Arnold Toynbee, A Study of
History, 12 vols. (London: Oxford University Press, 1934–54), stands out. An
excellent history of Eurasian civilization, especially western Eurasian
civilization, is William McNeill, The Rise of the West (Chicago: University of
Chicago Press, 1991). The same author’s A World History (New York: Oxford
University Press, 1979), despite its title, also maintains a focus on western
Eurasian civilization, as does V. Gordon Childe, What Happened in History,
rev. ed. (Baltimore: Penguin Books, 1954). Another comparative history with
a focus on western Eurasia, C. D. Darlington, The Evolution of Man and
Society (New York: Simon and Schuster, 1969), is by a biologist who
recognizes some of the same links between continental history and
domestication that I discuss. Two books by Alfred Crosby are distinguished
studies of the European overseas expansion with emphasis on its
accompanying plants, animals, and germs: The Columbian Exchange:
Biological Consequences of 1492 (Westport, Conn.: Greenwood, 1972) and
Ecological Imperialism: The Biological Expansion of Europe, 900–1900
(Cambridge: Cambridge University Press, 1986). Marvin Harris, Cannibals
and Kings: The Origins of Cultures (New York: Vintage Books, 1978), and
Marshall Sahlins and Elman Service, eds., Evolution and Culture (Ann Arbor:
University of Michigan Press, 1960), are comparative histories from the
perspective of cultural anthropologists. Ellen Semple, Influences of
Geographic Environment (New York: Holt, 1911), is an example of earlier
efforts to study geographic influences on human societies. Other important
historical studies are listed under further readings for the Epilogue. My book
The Third Chimpanzee (New York: HarperCollins, 1992), especially its
chapter 14, on the comparative histories of Eurasia and the Americas,
provided the starting point for my thinking about the present book.
The best-known or most notorious recent entrant into the debate about
group differences in intelligence is Richard Herrnstein and Charles Murray,
The Bell Curve: Intelligence and Class Structure in American Life (New
York: Free Press, 1994).
Chapter 1
Excellent books about early human evolution include Richard Klein, The
Human Career (Chicago: University of Chicago Press, 1989), Roger Lewin,
Bones of Contention (New York: Simon and Schuster, 1989), Paul Mellars
and Chris Stringer, eds., The Human Revolution: Behavioural and Biological
Perspectives on the Origins of Modern Humans (Edinburgh: Edinburgh
University Press, 1989), Richard Leakey and Roger Lewin, Origins
Reconsidered (New York: Doubleday, 1992), D. Tab Rasmussen, ed., The
Origin and Evolution of Humans and Humanness (Boston: Jones and Bartlett,
1993), Matthew Nitecki and Doris Nitecki, eds., Origins of Anatomically
Modern Humans (New York: Plenum, 1994), and Chris Stringer and Robin
McKie, African Exodus (London: Jonathan Cape, 1996). Three popular books
dealing specifically with the Neanderthals are Christopher Stringer and Clive
Gamble, In Search of the Neanderthals (New York: Thames and Hudson,
1993), Erik Trinkaus and Pat Shipman, The Neandertals (New York: Knopf,
1993), and Ian Tattersall, The Last Neanderthal (New York: Macmillan,
1995).
Genetic evidence of human origins is the subject of the two books by L.
Luca Cavalli-Sforza et al. already cited under the Prologue, and of chapter 1
of my book The Third Chimpanzee. Two technical papers with recent
advances in the genetic evidence are J. L. Mountain and L. L. Cavalli-Sforza,
“Inference of human evolution through cladistic analysis of nuclear DNA
restriction polymorphism,” Proceedings of the National Academy of Sciences
91:6515–19 (1994), and D. B. Goldstein et al., “Genetic absolute dating based
on microsatellites and the origin of modern humans,” ibid. 92:6723–27
(1995).
References to the human colonization of Australia, New Guinea, and the
Bismarck and Solomon Archipelagoes, and to extinctions of large animals
there, are listed under further readings for Chapter 15. In particular, Tim
Flannery, The Future Eaters (New York: Braziller, 1995), discusses those
subjects in clear, understandable terms and explains the problems with claims
of very recent survival of extinct big Australian mammals.
The standard text on Late Pleistocene and Recent extinctions of large
animals is Paul Martin and Richard Klein, eds., Quaternary Extinctions
(Tucson: University of Arizona Press, 1984). More recent updates are Richard
Klein, “The impact of early people on the environment: The case of large
mammal extinctions,” pp. 13–34 in J. E. Jacobsen and J. Firor, Human Impact
on the Environment (Boulder, Colo.: Westview Press, 1992), and Anthony
Stuart, “Mammalian extinctions in the Late Pleistocene of Northern Eurasia
and North America,” Biological Reviews 66:453–62 (1991). David Steadman
summarizes recent evidence that extinction waves accompanied human
settlement of Pacific islands in his paper “Prehistoric extinctions of Pacific
island birds: Biodiversity meets zooarchaeology,” Science 267:1123–31
(1995).
Popular accounts of the settlement of the Americas, the accompanying
extinctions of large mammals, and the resulting controversies are Brian
Fagan, The Great Journey: The Peopling of Ancient America (New York:
Thames and Hudson, 1987), and chapter 18 of my book The Third
Chimpanzee, both of which provide many other references. Ronald Carlisle,
ed., Americans before Columbus: Ice-Age Origins (Pittsburgh: University of
Pittsburgh, 1988), includes a chapter by J. M. Adovasio and his colleagues on
pre-Clovis evidence at the Meadowcroft site. Papers by C. Vance Haynes, Jr.,
an expert on the Clovis horizon and reported pre-Clovis sites, include
“Contributions of radiocarbon dating to the geochronology of the peopling of
the New World,” pp. 354–74 in R. E. Taylor, A. Long, and R. S. Kra, eds.,
Radiocarbon after Four Decades (New York: Springer, 1992), and “Clovis-
Folson geochronology and climate change,” pp. 219–36 in Olga Soffer and N.
D. Praslov, eds., From Kostenki to Clovis: Upper Paleolithic Paleo-Indian
Adaptations (New York: Plenum, 1993). Pre-Clovis claims for the Pedra
Furada site are argued by N. Guidon and G. Delibrias, “Carbon-14 dates point
to man in the Americas 32,000 years ago,” Nature 321:769–71 (1986), and
David Meltzer et al., “On a Pleistocene human occupation at Pedra Furada,
Brazil,” Antiquity 68:695–714 (1994). Other publications relevant to the pre-
Clovis debate include T. D. Dillehay et al., “Earliest hunters and gatherers of
South America,” Journal of World Prehistory 6:145–204 (1992), T. D.
Dillehay, Monte Verde: A Late Pleistocene Site in Chile (Washington, D.C.;
Smithsonian Institution Press, 1989), T. D. Dillehay and D. J. Meltzer, eds.,
The First Americans: Search and Research (Boca Raton: CRC Press, 1991),
Thomas Lynch, “Glacial-age man in South America?—a critical review,”
American Antiquity 55:12–36 (1990), John Hoffecker et al., “The colonization
of Beringia and the peopling of the New World,” Science 259:46–53 (1993),
and A. C. Roosevelt et al., “Paleoindian cave dwellers in the Amazon: The
peopling of the Americas,” Science 272:373–84 (1996).
Chapter 2
Two outstanding books explicitly concerned with cultural differences
among Polynesian islands are Patrick Kirch, The Evolution of the Polynesian
Chiefdoms (Cambridge: Cambridge University Press, 1984), and the same
author’s The Wet and the Dry (Chicago: University of Chicago Press, 1994).
Much of Peter Bellwood’s The Polynesians, rev. ed. (London: Thames and
Hudson, 1987), also deals with this problem. Notable books dealing with
specific Polynesian islands include Michael King, Moriori (Auckland:
Penguin, 1989), on the Chatham Islands, Patrick Kirch, Feathered Gods and
Fishhooks (Honolulu: University of Hawaii Press, 1985), on Hawaii, Patrick
Kirch and Marshall Sahlins, Anahulu (Chicago: University of Chicago Press,
1992), also on Hawaii, Jo Anne Van Tilburg, Easter Island (Washington,
D.C.: Smithsonian Institution Press, 1994), and Paul Bahn and John Flenley,
Easter Island, Earth Island (London: Thames and Hudson, 1992).
Chapter 3
My account of Pizarro’s capture of Atahuallpa combines the eyewitness
accounts by Francisco Pizarro’s brothers Hernando Pizarro and Pedro Pizarro
and by Pizarro’s companions Miguel de Estete, Cristóbal de Mena, Ruiz de
Arce, and Francisco de Xerez. The accounts by Hernando Pizarro, Miguel de
Estete, and Francisco de Xerez have been translated by Clements Markham,
Reports on the Discovery of Peru, Hakluyt Society, 1st ser., vol. 47 (New
York, 1872); Pedro Pizarro’s account, by Philip Means, Relation of the
Discovery and Conquest of the Kingdoms of Peru (New York: Cortés Society,
1921); and Cristóbal de Mena’s account, by Joseph Sinclair, The Conquest of
Peru, as Recorded by a Member of the Pizarro Expedition (New York, 1929).
The account by Ruiz de Arce was reprinted in Boletín de la Real Academia de
Historia (Madrid) 102:327–84 (1933). John Hemming’s excellent The
Conquest of the Incas (San Diego: Harcourt Brace Jovanovich, 1970) gives a
full account of the capture and indeed of the whole conquest, with an
extensive bibliography. A 19th-century account of the conquest, William H.
Prescott’s History of the Conquest of Peru (New York, 1847), is still highly
readable and ranks among the classics of historical writing. Corresponding
modern and classic 19th-century accounts of the Spanish conquest of the
Aztecs are, respectively, Hugh Thomas, Conquest: Montezuma, Cortés, and
the Fall of Old Mexico (New York: Simon and Schuster, 1993), and William
Prescott, History of the Conquest of Mexico (New York, 1843). Contemporary
eyewitness accounts of the conquest of the Aztecs were written by Cortés
himself (reprinted as Hernando Cortés, Five Letters of Cortés to the Emperor
[New York: Norton, 1969]) and by many of Cortés’s companions (reprinted in
Patricia de Fuentes, ed., The Conquistadors [Norman: University of
Oklahoma Press, 1993]).
Chapters 4–10
References for these seven chapters on food production will be combined,
since many of the references apply to more than one of them.
Five important sources, all of them excellent and fact-filled, address the
question how food production evolved from the hunter-gatherer lifestyle:
Kent Flannery, “The origins of agriculture,” Annual Reviews of Anthropology
2:271–310 (1973); Jack Harlan, Crops and Man, 2nd ed. (Madison, Wis.:
American Society of Agronomy, 1992); Richard MacNeish, The Origins of
Agriculture and Settled Life (Norman: University of Oklahoma Press, 1992);
David Rindos, The Origins of Agriculture: An Evolutionary Perspective (San
Diego: Academic Press, 1984); and Bruce Smith, The Emergence of
Agriculture (New York: Scientific American Library, 1995). Notable older
references about food production in general include two multi-author
volumes: Peter Ucko and G. W. Dimbleby, eds., The Domestication and
Exploitation of Plants and Animals (Chicago: Aldine, 1969), and Charles
Reed, ed., Origins of Agriculture (The Hague: Mouton, 1977). Carl Sauer,
Agricultural Origins and Dispersals (New York: American Geographical
Society, 1952), is a classic early comparison of Old World and New World
food production, while Erich Isaac, Geography of Domestication (Englewood
Cliffs, N. J.: Prentice-Hall, 1970), addresses the questions of where, when,
and how regarding plant and animal domestication.
Among references specifically about plant domestication, Daniel Zohary
and Maria Hopf, Domestication of Plants in the Old World, 2nd ed. (Oxford:
Oxford University Press, 1993), stands out. It provides the most detailed
account of plant domestication available for any part of the world. For each
significant crop grown in western Eurasia, the book summarizes
archaeological and genetic evidence about its domestication and subsequent
spread.
Among important multi-author books on plant domestication are C.
Wesley Cowan and Patty Jo Watson, eds., The Origins of Agriculture
(Washington, D.C.: Smithsonian Institution Press, 1992), David Harris and
Gordon Hillman, eds., Foraging and Farming: The Evolution of Plant
Exploitation (London: Unwin Hyman, 1989), and C. Barigozzi, ed., The
Origin and Domestication of Cultivated Plants (Amsterdam: Elsevier, 1986).
Two engaging popular accounts of plant domestication by Charles Heiser, Jr.,
are Seed to Civilization: The Story of Food, 3rd ed. (Cambridge: Harvard
University Press, 1990), and Of Plants and People (Norman: University of
Oklahoma Press, 1985). J. Smartt and N. W. Simmonds, ed., Evolution of
Crop Plants, 2nd ed. (London: Longman, 1995), is the standard reference
volume summarizing information about all of the world’s major crops and
many minor ones. Three excellent papers describe the changes that evolve
automatically in wild plants under human cultivation: Mark Blumler and
Roger Byrne, “The ecological genetics of domestication and the origins of
agriculture,” Current Anthropology 32:23–54 (1991); Charles Heiser, Jr.,
“Aspects of unconscious selection and the evolution of domesticated plants,”
Euphytica 37:77–81 (1988); and Daniel Zohary, “Modes of evolution in
plants under domestication,” in W. F. Grant, ed., Plant Biosystematics
(Montreal: Academic Press, 1984). Mark Blumler, “Independent inventionism
and recent genetic evidence on plant domestication,” Economic Botany
46:98–111 (1992), evaluates the evidence for multiple domestications of the
same wild plant species, as opposed to single origins followed by spread.
Among writings of general interest in connection with animal
domestication, the standard encyclopedic reference work to the world’s wild
mammals is Ronald Nowak, ed., Walker’s Mammals of the World, 5th ed.
(Baltimore: Johns Hopkins University Press, 1991). Juliet Clutton-Brock,
Domesticated Animals from Early Times (London: British Museum [Natural
History], 1981), gives an excellent summary of all important domesticated
mammals. I. L. Mason, ed., Evolution of Domesticated Animals (London:
Longman, 1984), is a multi-author volume discussing each significant
domesticated animal individually. Simon Davis, The Archaeology of Animals
(New Haven: Yale University Press, 1987), provides an excellent account of
what can be learned from mammal bones in archaeological sites. Juliet
Clutton-Brock, ed., The Walking Larder (London: Unwin-Hyman, 1989),
presents 31 papers about how humans have domesticated, herded, hunted, and
been hunted by animals around the world. A comprehensive book in German
about domesticated animals is Wolf Herre and Manfred Röhrs, Haustiere
zoologisch gesehen (Stuttgart: Fischer, 1990). Stephen Budiansky, The
Covenant of the Wild (New York: William Morrow, 1992), is a popular
account of how animal domestication evolved automatically from
relationships between humans and animals. An important paper on how
domestic animals became used for plowing, transport, wool, and milk is
Andrew Sheratt, “Plough and pastoralism: Aspects of the secondary products
revolution,” pp. 261–305 in Ian Hodder et al., eds., Pattern of the Past
(Cambridge: Cambridge University Press, 1981).
Accounts of food production in particular areas of the world include a
deliciously detailed mini-encyclopedia of Roman agricultural practices, Pliny,
Natural History, vols. 17–19 (Latin text side-by-side with English translation
in the Loeb Classical Library edition [Cambridge: Harvard University Press,
1961]); Albert Ammerman and L. L. Cavalli-Sforza, The Neolithic Transition
and the Genetics of Populations in Europe (Princeton: Princeton University
Press, 1984), analyzing the spread of food production from the Fertile
Crescent westward across Europe; Graeme Barker, Prehistoric Farming in
Europe (Cambridge: Cambridge University Press, 1985), and Alasdair
Whittle, Neolithic Europe: A Survey (Cambridge: Cambridge University
Press, 1985), for Europe; Donald Henry, From Foraging to Agriculture: The
Levant at the End of the Ice Age (Philadelphia: University of Pennsylvania
Press, 1989), for the lands bordering the eastern shore of the Mediterranean;
and D. E. Yen, “Domestication: Lessons from New Guinea,” pp. 558–69 in
Andrew Pawley, ed., Man and a Half (Auckland: Polynesian Society, 1991),
for New Guinea. Edward Schafer, The Golden Peaches of Samarkand
(Berkeley: University of California Press, 1963), describes the animals,
plants, and other things imported into China during the T’ang dynasty.
The following are accounts of plant domestication and crops in specific
parts of the world. For Europe and the Fertile Crescent: Willem van Zeist et
al., eds., Progress in Old World Palaeoethnobotany (Rotterdam: Balkema,
1991), and Jane Renfrew, Paleoethnobotany (London: Methuen, 1973). For
the Harappan civilization of the Indus Valley, and for the Indian subcontinent
in general: Steven Weber, Plants and Harappan Subsistence (New Delhi:
American Institute of Indian Studies, 1991). For New World crops: Charles
Heiser, Jr., “New perspectives on the origin and evolution of New World
domesticated plants: Summary,” Economic Botany 44(3 suppl.): 111–16
(1990), and the same author’s “Origins of some cultivated New World
plants,” Annual Reviews of Ecology and Systematics 10:309–26 (1979). For a
Mexican site that may document the transition from hunting-gathering to
early agriculture in Mesoamerica: Kent Flannery, ed., Guilá Naquitz (New
York: Academic Press, 1986). For an account of crops grown in the Andes
during Inca times, and their potential uses today: National Research Council,
Lost Crops of the Incas (Washington, D.C.: National Academy Press, 1989).
For plant domestication in the eastern and / or southwestern United States:
Bruce Smith, “Origins of agriculture in eastern North America,” Science
246:1566–71 (1989); William Keegan, ed., Emergent Horticultural
Economies of the Eastern Woodlands (Carbondale: Southern Illinois
University, 1987); Richard Ford, ed., Prehistoric Food Production in North
America (Ann Arbor: University of Michigan Museum of Anthropology,
1985); and R. G. Matson, The Origins of Southwestern Agriculture (Tucson:
University of Arizona Press, 1991). Bruce Smith, “The origins of agriculture
in the Americas,” Evolutionary Anthropology 3:174–84 (1995), discusses the
revisionist view, based on accelerator mass spectrometry dating of very small
plant samples, that the origins of agriculture in the Americas were much more
recent than previously believed.
The following are accounts of animal domestication and livestock in
specific parts of the world. For central and eastern Europe: S. Bökönyi,
History of Domestic Mammals in Central and Eastern Europe (Budapest:
Akadémiai Kiadó, 1974). For Africa: Andrew Smith, Pastoralism in Africa
(London: Hurst, 1992). For the Andes: Elizabeth Wing, “Domestication of
Andean mammals,” pp. 246–64 in F. Vuilleumier and M. Monasterio, eds.,
High Altitude Tropical Biogeography (Oxford: Oxford University Press,
1986).
References on specific important crops include the following. Thomas
Sodestrom et al., eds., Grass Systematics and Evolution (Washington, D.C.:
Smithsonian Institution Press, 1987), is a comprehensive multi-author account
of grasses, the plant group that gave rise to our cereals, now the world’s most
important crops. Hugh Iltis, “From teosinte to maize: The catastrophic sexual
transmutation,” Science 222:886–94 (1983), gives an account of the drastic
changes in reproductive biology involved in the evolution of corn from
teosinte, its wild ancestor. Yan Wenming, “China’s earliest rice agricultural
remains,” Indo-Pacific Prehistory Association Bulletin 10:118–26 (1991),
discusses early rice domestication in South China. Two books by Charles
Heiser, Jr., are popular accounts of particular crops: The Sunflower (Norman:
University of Oklahoma Press, 1976) and The Gourd Book (Norman:
University of Oklahoma Press, 1979).
Many papers or books are devoted to accounts of particular domesticated
animal species. R. T. Loftus et al., “Evidence for two independent
domestications of cattle,” Proceedings of the National Academy of Sciences
U.S.A. 91:2757–61 (1994), uses evidence from mitochondrial DNA to
demonstrate that cattle were domesticated independently in western Eurasia
and in the Indian subcontinent. For horses: Juliet Clutton-Brock, Horse Power
(Cambridge: Harvard University Press, 1992), Richard Meadow and Hans-
Peter Uerpmann, eds., Equids in the Ancient World (Wiesbaden: Reichert,
1986), Matthew J. Kust, Man and Horse in History (Alexandria, Va.: Plutarch
Press, 1983), and Robin Law, The Horse in West African History (Oxford:
Oxford University Press, 1980). For pigs: Colin Groves, Ancestors for the
Pigs: Taxonomy and Phylogeny of the Genus Sus (Technical Bulletin no. 3,
Department of Prehistory, Research School of Pacific Studies, Australian
National University [1981]). For llamas: Kent Flannery, Joyce Marcus, and
Robert Reynolds, The Flocks of the Wamani (San Diego: Academic Press,
1989). For dogs: Stanley Olsen, Origins of the Domestic Dog (Tucson:
University of Arizona Press, 1985). John Varner and Jeannette Varner, Dogs
of the Conquest (Norman: University of Oklahoma Press, 1983), describes the
Spaniards’ use of dogs as military weapons to kill Indians during the Spanish
conquests of the Americas. Clive Spinnage, The Natural History of Antelopes
(New York: Facts on File, 1986), gives an account of the biology of antelopes,
and hence a starting point for trying to understand why none of these
seemingly obvious candidates for domestication was actually domesticated.
Derek Goodwin, Domestic Birds (London: Museum Press, 1965), summarizes
the bird species that have been domesticated, and R. A. Donkin, The Muscovy
Duck Cairina moschata domestica (Rotterdam: Balkema, 1989), discusses one
of the sole two bird species domesticated in the New World.
Finally, the complexities of calibrating radiocarbon dates are discussed by
G. W. Pearson, “How to cope with calibration,” Antiquity 61:98–103 (1987),
R. E. Taylor, eds., Radiocarbon after Four Decades: An Interdisciplinary
Perspective (New York: Springer, 1992), M. Stuiver et al., “Calibration,”
Radiocarbon 35:1–244 (1993), S. Bowman, “Using radiocarbon: An update,”
Antiquity 68:838–43 (1994), and R. E. Taylor, M. Stuiver, and C. Vance
Haynes, Jr., “Calibration of the Late Pleistocene radiocarbon time scale:
Clovis and Folsom age estimates,” Antiquity vol. 70 (1996).
Chapter 11
For a gripping account of the impact of disease on a human population,
nothing can match Thucydides’ account of the plague of Athens, in book 2 of
his Peloponnesian War (available in many translations).
Three classic accounts of disease in history are Hans Zinsser, Rats, Lice,
and History (Boston: Little, Brown, 1935), Geddes Smith, A Plague on Us
(New York: Commonwealth Fund, 1941), and William McNeill, Plagues and
Peoples (Garden City, N.Y.: Doubleday, 1976). The last book, written by a
distinguished historian rather than by a physician, has been especially
influential in bringing historians to recognize the impacts of disease, as have
been the two books by Alfred Crosby listed under the further readings for the
Prologue.
Friedrich Vogel and Arno Motulsky, Human Genetics, 2nd ed. (Berlin:
Springer, 1986), the standard textbook on human genetics, is a convenient
reference for natural selection of human populations by disease, and for the
development of genetic resistance against specific diseases. Roy Anderson
and Robert May, Infectious Diseases of Humans (Oxford: Oxford University
Press, 1992), is a clear mathematical treatment of disease dynamics,
transmission, and epidemiology. MacFarlane Burnet, Natural History of
Infectious Disease (Cambridge: Cambridge University Press, 1953), is a
classic by a distinguished medical researcher, while Arno Karlen, Man and
Microbes (New York: Putnam, 1995), is a recent popular account.
Books and articles specifically concerned with the evolution of human
infectious diseases include Aidan Cockburn, Infectious Diseases: Their
Evolution and Eradication (Springfield, Ill.: Thomas, 1967); the same
author’s “Where did our infectious diseases come from?” pp. 103–13 in
Health and Disease in Tribal Societies, CIBA Foundation Symposium, no. 49
(Amsterdam: Elsevier, 1977); George Williams and Randolph Nesse, “The
dawn of Darwinian medicine,” Quarterly Reviews of Biology 66:1–62 (1991);
and Paul Ewald, Evolution of Infectious Disease (New York: Oxford
University Press, 1994).
Francis Black, “Infectious diseases in primitive societies,” Science
187:515–18 (1975), discusses the differences between endemic and acute
diseases in their impact on, and maintenance in, small isolated societies.
Frank Fenner, “Myxoma virus and Oryctolagus cuniculus: Two colonizing
species,” pp. 485–501 in H. G. Baker and G. L. Stebbins, eds., Genetics of
Colonizing Species (New York: Academic Press, 1965), describes the spread
and evolution of Myxoma virus among Australian rabbits. Peter Panum,
Observations Made during the Epidemic of Measles on the Faroe Islands in
the Year 1846 (New York: American Public Health Association, 1940),
illustrates how the arrival of an acute epidemic disease in an isolated
nonresistant population quickly kills or immunizes the whole population.
Francis Black, “Measles endemicity in insular populations: Critical
community size and its evolutionary implication,” Journal of Theoretical
Biology 11:207–11 (1966), uses such measles epidemics to calculate the
minimum size of population required to maintain measles. Andrew Dobson,
“The population biology of parasite-induced changes in host behavior,”
Quarterly Reviews of Biology 63:139–65 (1988), discusses how parasites
enhance their own transmission by changing the behavior of their host. Aidan
Cockburn and Eve Cockburn, eds., Mummies, Diseases, and Ancient Cultures
(Cambridge: Cambridge University Press, 1983), illustrates what can be
learned from mummies about past impacts of diseases.
As for accounts of disease impacts on previously unexposed populations,
Henry Dobyns, Their Number Became Thinned (Knoxville: University of
Tennessee Press, 1983), marshals evidence for the view that European-
introduced diseases killed up to 95 percent of all Native Americans.
Subsequent books or articles arguing that controversial thesis include John
Verano and Douglas Ubelaker, eds., Disease and Demography in the
Americas (Washington, D.C.: Smithsonian Institution Press, 1992); Ann
Ramenofsky, Vectors of Death (Albuquerque: University of New Mexico
Press, 1987); Russell Thornton, American Indian Holocaust and Survival
(Norman: University of Oklahoma Press, 1987); and Dean Snow,
“Microchronology and demographic evidence relating to the size of the pre-
Columbian North American Indian population,” Science 268:1601–4 (1995).
Two accounts of depopulation caused by European-introduced diseases
among Hawaii’s Polynesian population are David Stannard, Before the
Horror: The Population of Hawaii on the Eve of Western Contact (Honolulu:
University of Hawaii Press, 1989), and O. A. Bushnell, The Gifts of
Civilization: Germs and Genocide in Hawaii (Honolulu: University of Hawaii
Press, 1993). The near-extermination of the Sadlermiut Eskimos by a
dysentery epidemic in the winter of 1902–3 is described by Susan Rowley,
“The Sadlermiut: Mysterious or misunderstood?” pp. 361–84 in David
Morrison and Jean-Luc Pilon, eds., Threads of Arctic Prehistory (Hull:
Canadian Museum of Civilization, 1994). The reverse phenomenon, of
European deaths due to diseases encountered overseas, is discussed by Philip
Curtin, Death by Migration: Europe’s Encounter with the Tropical World in
the 19th Century (Cambridge: Cambridge University Press, 1989).
Among accounts of specific diseases, Stephen Morse, ed., Emerging
Viruses (New York: Oxford University Press, 1993), contains many valuable
chapters on “new” viral diseases of humans; so does Mary Wilson et al., eds.,
Disease in Evolution, Annals of the New York Academy of Sciences, vol. 740
(New York, 1995). References for other diseases include the following. For
bubonic plague: Colin McEvedy, “Bubonic plague,” Scientific American
258(2): 118–23 (1988). For cholera: Norman Longmate, King Cholera
(London: Hamish Hamilton, 1966). For influenza: Edwin Kilbourne,
Influenza (New York: Plenum, 1987), and Robert Webster et al., “Evolution
and ecology of influenza A viruses,” Microbiological Reviews 56:152–79
(1992). For Lyme disease: Alan Barbour and Durland Fish, “The biological
and social phenomenon of Lyme disease,” Science 260:1610–16 (1993), and
Allan Steere, “Lyme disease: A growing threat to urban populations,”
Proceedings of the National Academy of Sciences 91:2378–83 (1994).
For the evolutionary relationships of human malarial parasites: Thomas
McCutchan et al., “Evolutionary relatedness of Plasmodium species as
determined by the structure of DNA,” Science 225:808–11 (1984), and A. P.
Waters et al., “Plasmodium falciparum appears to have arisen as a result of
lateral transfer between avian and human hosts,” Proceedings of the National
Academy of Sciences 88:3140–44 (1991). For the evolutionary relationships
of measles virus: E. Norrby et al., “Is rinderpest virus the archevirus of the
Morbillivirus genus?” Intervirology 23:228–32 (1985), and Keith Murray et
al., “A morbillivirus that caused fatal disease in horses and humans,” Science
268:94–97 (1995). For pertussis, also known as whooping cough: R. Gross et
al., “Genetics of pertussis toxin,” Molecular Microbiology 3:119–24 (1989).
For smallpox: Donald Hopkins, Princes and Peasants: Smallpox in History
(Chicago: University of Chicago Press, 1983); F. Vogel and M. R.
Chakravartti, “ABO blood groups and smallpox in a rural population of West
Bengal and Bihar (India),” Human Genetics 3:166–80 (1966); and my article
“A pox upon our genes,” Natural History 99(2): 26–30 (1990). For
monkeypox, a disease related to smallpox: Zden k Je ek and Frank Fenner,
Human Monkeypox (Basel: Karger, 1988). For syphilis: Claude Quétel,
History of Syphilis (Baltimore: Johns Hopkins University Press, 1990). For
tuberculosis: Guy Youmans, Tuberculosis (Philadelphia: Saunders, 1979). For
the claim that human tuberculosis was present in Native Americans before
Columbus’s arrival: in favor, Wilmar Salo et al., “Identification of
Mycobacterium tuberculosis DNA in a pre-Columbian Peruvian mummy,”
Proceedings of the National Academy of Sciences 91:2091–94 (1994);
opposed, William Stead et al., “When did Mycobacterium tuberculosis
infection first occur in the New World?” American Journal of Respiratory
Critical Care Medicine 151:1267–68 (1995).
Chapter 12
Books providing general accounts of writing and of particular writing
systems include David Diringer, Writing (London: Thames and Hudson,
1982), I. J. Gelb, A Study of Writing, 2nd ed. (Chicago: University of Chicago
Press, 1963), Geoffrey Sampson, Writing Systems (Stanford: Stanford
University Press, 1985), John DeFrancis, Visible Speech (Honolulu:
University of Hawaii Press, 1989), Wayne Senner, ed., The Origins of Writing
(Lincoln: University of Nebraska Press, 1991), and J. T. Hooker, ed., Reading
the Past (London: British Museum Press, 1990). A comprehensive account of
significant writing systems, with plates depicting texts in each system, is
David Diringer, The Alphabet, 3rd ed., 2 vols. (London: Hutchinson, 1968).
Jack Goody, The Domestication of the Savage Mind (Cambridge: Cambridge
University Press, 1977), and Robert Logan, The Alphabet Effect (New York:
Morrow, 1986), discuss the impact of literacy in general and of the alphabet in
particular. Uses of early writing are discussed by Nicholas Postgate et al.,
“The evidence for early writing: Utilitarian or ceremonial?” Antiquity 69:459–
80 (1995).
Exciting accounts of decipherments of previously illegible scripts are
given by Maurice Pope, The Story of Decipherment (London: Thames and
Hudson, 1975), Michael Coe, Breaking the Maya Code (New York: Thames
and Hudson, 1992), John Chadwick, The Decipherment of Linear B
(Cambridge: Cambridge University Press, 1992), Yves Duhoux, Thomas
Palaima, and John Bennet, eds., Problems in Decipherment (Louvain-la-
Neuve: Peeters, 1989), and John Justeson and Terrence Kaufman, “A
decipherment of epi-Olmec hieroglyphic writing,” Science 259:1703–11
(1993).
Denise Schmandt-Besserat’s two-volume Before Writing (Austin:
University of Texas Press, 1992) presents her controversial reconstruction of
the origins of Sumerian writing from clay tokens over the course of nearly
5,000 years. Hans Nissen et al., eds., Archaic Bookkeeping (Chicago:
University of Chicago Press, 1994), describes Mesopotamian tablets that
represent the earliest stages of cuneiform itself. Joseph Naveh, Early History
of the Alphabet (Leiden: Brill, 1982), traces the emergence of alphabets in the
eastern Mediterranean region. The remarkable Ugaritic alphabet is the subject
of Gernot Windfuhr, “The cuneiform signs of Ugarit,” Journal of Near
Eastern Studies 29:48–51 (1970). Joyce Marcus, Mesoamerican Writing
Systems: Propaganda, Myth, and History in Four Ancient Civilizations
(Princeton: Princeton University Press, 1992), and Elizabeth Boone and
Walter Mignolo, Writing without Words (Durham: Duke University Press,
1994), describe the development and uses of Mesoamerican writing systems.
William Boltz, The Origin and Early Development of the Chinese Writing
System (New Haven: American Oriental Society, 1994), and the same author’s
“Early Chinese writing,” World Archaeology 17:420–36 (1986), do the same
for China. Finally, Janet Klausner, Sequoyah’s Gift (New York: HarperCollins,
1993), is an account readable by children, but equally interesting to adults, of
Sequoyah’s development of the Cherokee syllabary.
Chapter 13
The standard detailed history of technology is the eight-volume A History
of Technology, by Charles Singer et al. (Oxford: Clarendon Press, 1954–84).
One-volume histories are Donald Cardwell, The Fontana History of
Technology (London: Fontana Press, 1994), Arnold Pacey, Technology in
World Civilization (Cambridge: MIT Press, 1990), and Trevor Williams, The
History of Invention (New York: Facts on File, 1987). R. A. Buchanan, The
Power of the Machine (London: Penguin Books, 1994), is a short history of
technology focusing on the centuries since A.D. 1700. Joel Mokyr, The Lever
of Riches (New York: Oxford University Press, 1990), discusses why the rate
of development of technology has varied with time and place. George Basalla,
The Evolution of Technology (Cambridge: Cambridge University Press,
1988), presents an evolutionary view of technological change. Everett Rogers,
Diffusion of Innovations, 3rd ed. (New York: Free Press, 1983), summarizes
modern research on the transfer of innovations, including the QWERTY
keyboard. David Holloway, Stalin and the Bomb (New Haven: Yale
University Press, 1994), dissects the relative contributions of blueprint
copying, idea diffusion (by espionage), and independent invention to the
Soviet atomic bomb.
Preeminent among regional accounts of technology is the series Science
and Civilization in China, by Joseph Needham (Cambridge: Cambridge
University Press), of which 5 volumes in 16 parts have appeared since 1954,
with a dozen more parts on the way. Ahmad al-Hassan and Donald Hill,
Islamic Technology (Cambridge: Cambridge University Press, 1992), and K.
D. White, Greek and Roman Technology (London: Thames and Hudson,
1984), summarize technology’s history for those cultures.
Two conspicuous examples of somewhat isolated societies adopting and
then abandoning technologies potentially useful in competition with other
societies involve Japan’s abandonment of firearms, after their adoption in A.D.
1543, and China’s abandonment of its large oceangoing fleets after A.D. 1433.
The former case is described by Noel Perrin, Giving Up the Gun (Boston:
Hall, 1979), and the latter by Louise Levathes, When China Ruled the Seas
(New York: Simon and Schuster, 1994). An essay entitled “The disappearance
of useful arts,” pp. 190–210 in W. H. B. Rivers, Psychology and Ethnology
(New York: Harcourt, Brace, 1926), gives similar examples among Pacific
islanders.
Articles on the history of technology will be found in the quarterly journal
Technology and Culture, published by the Society for the History of
Technology since 1959. John Staudenmaier, Technology’s Storytellers
(Cambridge: MIT Press, 1985), analyzes the papers in its first twenty years.
Specific fields providing material for those interested in the history of
technology include electric power, textiles, and metallurgy. Thomas Hughes,
Networks of Power (Baltimore: Johns Hopkins University Press, 1983),
discusses the social, economic, political, and technical factors in the
electrification of Western society from 1880 to 1930. Dava Sobel, Longitude
(New York: Walker, 1995), describes the development of John Harrison’s
chronometers that solved the problem of determining longitude at sea. E. J.
W. Barber, Prehistoric Textiles (Princeton: Princeton University Press, 1991),
sets out the history of cloth in Eurasia from its beginnings more than 9,000
years ago. Accounts of the history of metallurgy over wide regions or even
over the world include Robert Maddin, The Beginning of the Use of Metals
and Alloys (Cambridge: MIT Press, 1988), Theodore Wertime and James
Muhly, eds., The Coming of the Age of Iron (New Haven: Yale University
Press, 1980), R. D. Penhallurick, Tin in Antiquity (London: Institute of
Metals, 1986), James Muhly, “Copper and Tin,” Transactions of the
Connecticut Academy of Arts and Sciences 43:155–535 (1973), and Alan
Franklin, Jacqueline Olin, and Theodore Wertime, The Search for Ancient Tin
(Washington, D.C.: Smithsonian Institution Press, 1978). Accounts of
metallurgy for local regions include R. F. Tylecote, The Early History of
Metallurgy in Europe (London: Longman, 1987), and Donald Wagner, Iron
and Steel in Ancient China (Leiden: Brill, 1993).
Chapter 14
The fourfold classification of human societies into bands, tribes,
chiefdoms, and states owes much to two books by Elman Service: Primitive
Social Organization (New York: Random House, 1962) and Origins of the
State and Civilization (New York: Norton, 1975). A related classification of
societies, using different terminology, is Morton Fried, The Evolution of
Political Society (New York: Random House, 1967). Three important review
articles on the evolution of states and societies are Kent Flannery, “The
cultural evolution of civilizations,” Annual Review of Ecology and
Systematics 3:399–426 (1972), the same author’s “Prehistoric social
evolution,” pp. 1–26 in Carol and Melvin Ember, eds., Research Frontiers in
Anthropology (Englewood Cliffs: Prentice-Hall, 1995), and Henry Wright,
“Recent research on the origin of the state,” Annual Review of Anthropology
6:379–97 (1977). Robert Carneiro, “A theory of the origin of the state,”
Science 169:733–38 (1970), argues that states arise through warfare under
conditions in which land is ecologically limiting. Karl Wittfogel, Oriental
Despotism (New Haven: Yale University Press, 1957), relates state origins to
large-scale irrigation and hydraulic management. Three essays in On the
Evolution of Complex Societies, by William Sanders, Henry Wright, and
Robert Adams (Malibu: Undena, 1984), present differing views of state
origins, while Robert Adams, The Evolution of Urban Society (Chicago:
Aldine, 1966), contrasts state origins in Mesopotamia and Mesoamerica.
Among studies of the evolution of societies in specific parts of the world,
sources for Mesopotamia include Robert Adams, Heartland of Cities
(Chicago: University of Chicago Press, 1981), and J. N. Postgate, Early
Mesopotamia (London: Routledge, 1992); for Mesoamerica, Richard Blanton
et al., Ancient Mesoamerica (Cambridge: Cambridge University Press, 1981),
and Joyce Marcus and Kent Flannery, Zapotec Civilization (London: Thames
and Hudson, 1996); for the Andes, Richard Burger, Chavin and the Origins of
Andean Civilization (New York: Thames and Hudson, 1992), and Jonathan
Haas et al., eds., The Origins and Development of the Andean State
(Cambridge: Cambridge University Press, 1987); for American chiefdoms,
Robert Drennan and Carlos Uribe, eds., Chiefdoms in the Americas (Lanham,
Md.: University Press of America, 1987); for Polynesian societies, the books
cited under Chapter 2; and for the Zulu state, Donald Morris, The Washing of
the Spears (London: Jonathan Cape, 1966).
Chapter 15
Books covering the prehistory of both Australia and New Guinea include
Alan Thorne and Robert Raymond, Man on the Rim: The Peopling of the
Pacific (North Ryde: Angus and Robertson, 1989), J. Peter White and James
O’Connell, A Prehistory of Australia, New Guinea, and Sahul (Sydney:
Academic Press, 1982), Jim Allen et al., eds., Sunda and Sahul (London:
Academic Press, 1977), M. A. Smith et al., eds., Sahul in Review (Canberra:
Australian National University, 1993), and Tim Flannery, The Future Eaters
(New York: Braziller, 1995). The first and third of these books discuss the
prehistory of island Southeast Asia as well. A recent account of the history of
Australia itself is Josephine Flood, Archaeology of the Dreamtime, rev. ed.
(Sydney: Collins, 1989). Some additional key papers on Australian prehistory
are Rhys Jones, “The fifth continent: Problems concerning the human
colonization of Australia,” Annual Reviews of Anthropology 8:445–66 (1979),
Richard Roberts et al., “Thermoluminescence dating of a 50,000-year-old
human occupation site in northern Australia,” Nature 345:153–56 (1990), and
Jim Allen and Simon Holdaway, “The contamination of Pleistocene
radiocarbon determinations in Australia,” Antiquity 69:101–12 (1995). Robert
Attenborough and Michael Alpers, eds., Human Biology in Papua New
Guinea (Oxford: Clarendon Press, 1992), summarizes New Guinea
archaeology as well as languages and genetics.
As for the prehistory of Northern Melanesia (the Bismarck and Solomon
Archipelagoes, northeast and east of New Guinea), discussion will be found
in the above-cited books by Thorne and Raymond, Flannery, and Allen et al.
Papers pushing back the dates for the earliest occupation of Northern
Melanesia include Stephen Wickler and Matthew Spriggs, “Pleistocene
human occupation of the Solomon Islands, Melanesia,” Antiquity 62:703–6
(1988), Jim Allen et al., “Pleistocene dates for the human occupation of New
Ireland, Northern Melanesia,” Nature 331:707–9 (1988), Jim Allen et al.,
“Human Pleistocene adaptations in the tropical island Pacific: Recent
evidence from New Ireland, a Greater Australian outlier,” Antiquity 63:548–
61 (1989), and Christina Pavlides and Chris Gosden, “35,000-year-old sites in
the rainforests of West New Britain, Papua New Guinea,” Antiquity 68:604–
10 (1994). References to the Austronesian expansion around the coast of New
Guinea will be found under further readings for Chapter 17.
Two books on the history of Australia after European colonization are
Robert Hughes, The Fatal Shore (New York: Knopf, 1987), and Michael
Cannon, The Exploration of Australia (Sydney: Reader’s Digest, 1987).
Aboriginal Australians themselves are the subject of Richard Broome,
Aboriginal Australians (Sydney: Allen and Unwin, 1982), and Henry
Reynolds, Frontier (Sydney: Allen and Unwin, 1987). An incredibly detailed
history of New Guinea, from the earliest written records until 1902, is the
three-volume work by Arthur Wichmann, Entdeckungsgeschichte von Neu-
Guinea (Leiden: Brill, 1909–12). A shorter and more readable account is
Gavin Souter, New Guinea: The Last Unknown (Sydney: Angus and
Robertson, 1964). Bob Connolly and Robin Anderson, First Contact (New
York: Viking, 1987), movingly describes the first encounters of highland New
Guineans with Europeans.
For detailed accounts of New Guinea’s Papuan (i.e., non-Austronesian)
languages, see Stephen Wurm, Papuan Languages of Oceania (Tübingen:
Guntet Narr, 1982), and William Foley, The Papuan Languages of New
Guinea (Cambridge: Cambridge University Press, 1986); and of Australian
languages, see Stephen Wurm, Languages of Australia and Tasmania (The
Hague: Mouton, 1972), and R. M. W. Dixon, The Languages of Australia
(Cambridge: Cambridge University Press, 1980).
An entrance into the literature on plant domestication and origins of food
production in New Guinea can be found in Jack Golson, “Bulmer phase II:
Early agriculture in the New Guinea highlands,” pp. 484–91 in Andrew
Pawley, ed., Man and a Half (Auckland: Polynesian Society, 1991), and D. E.
Yen, “Polynesian cultigens and cultivars: The question of origin,” pp. 67–95
in Paul Cox and Sandra Banack, eds., Islands, Plants, and Polynesians
(Portland: Dioscorides Press, 1991).
Numerous articles and books are devoted to the fascinating problem of
why trading visits of Indonesians and of Torres Strait islanders to Australia
produced only limited cultural change. C. C. Macknight, “Macassans and
Aborigines,” Oceania 42:283–321 (1972), discusses the Macassan visits,
while D. Walker, ed., Bridge and Barrier: The Natural and Cultural History
of Torres Strait (Canberra: Australian National University, 1972), discusses
connections at Torres Strait. Both connections are also discussed in the above-
cited books by Flood, White and O’Connell, and Allen et al.
Early eyewitness accounts of the Tasmanians are reprinted in N. J. B.
Plomley, The Baudin Expedition and the Tasmanian Aborigines 1802 (Hobart:
Blubber Head Press, 1983), N. J. B. Plomley, Friendly Mission: The
Tasmanian Journals and Papers of George Augustus Robinson, 1829–1834
(Hobart: Tasmanian Historical Research Association, 1966), and Edward
Duyker, The Discovery of Tasmania: Journal Extracts from the Expeditions of
Abel Janszoon Tasman and Marc-Joseph Marion Dufresne, 1642 and 1772
(Hobart: St. David’s Park Publishing, 1992). Papers debating the effects of
isolation on Tasmanian society include Rhys Jones, “The Tasmanian
Paradox,” pp. 189–284 in R. V. S. Wright, ed., Stone Tools as Cultural
Markers (Canberra: Australian Institute of Aboriginal Studies, 1977); Rhys
Jones, “Why did the Tasmanians stop eating fish?” pp. 11–48 in R. Gould,
ed., Explorations in Ethnoarchaeology (Albuquerque: University of New
Mexico Press, 1978); D. R. Horton, “Tasmanian adaptation,” Mankind 12:28–
34 (1979); I. Walters, “Why did the Tasmanians stop eating fish?: A
theoretical consideration,” Artefact 6:71–77 (1981); and Rhys Jones,
“Tasmanian Archaeology,” Annual Reviews of Anthropology 24:423–46
(1995). Results of Robin Sim’s archaeological excavations on Flinders Island
are described in her article “Prehistoric human occupation on the King and
Furneaux Island regions, Bass Strait,” pp. 358–74 in Marjorie Sullivan et al.,
eds., Archaeology in the North (Darwin: North Australia Research Unit,
1994).
Chapters 16 and 17
Relevant readings cited under previous chapters include those on East
Asian food production (Chapters 4–10), Chinese writing (Chapter 12),
Chinese technology (Chapter 13), and New Guinea and the Bismarcks and
Solomons in general (Chapter 15). James Matisoff, “Sino-Tibetan linguistics:
Present state and future prospects,” Annual Reviews of Anthropology 20:469–
504 (1991), reviews Sino-Tibetan languages and their wider relationships.
Takeru Akazawa and Emoke Szathmáry, eds., Prehistoric Mongoloid
Dispersals (Oxford: Oxford University Press, 1996), and Dennis Etler,
“Recent developments in the study of human biology in China: A review,”
Human Biology 64:567–85 (1992), discuss evidence of Chinese or East Asian
relationships and dispersal. Alan Thorne and Robert Raymond, Man on the
Rim (North Ryde: Angus and Robertson, 1989), describes the archaeology,
history, and culture of Pacific peoples, including East Asians and Pacific
islanders. Adrian Hill and Susan Serjeantson, eds., The Colonization of the
Pacific: A Genetic Trail (Oxford: Clarendon Press, 1989), interprets the
genetics of Pacific islanders, Aboriginal Australians, and New Guineans in
terms of their inferred colonization routes and histories. Evidence based on
tooth structure is interpreted by Christy Turner III, “Late Pleistocene and
Holocene population history of East Asia based on dental variation,”
American Journal of Physical Anthropology 73:305–21 (1987), and “Teeth
and prehistory in Asia,” Scientific American 260 (2):88–96 (1989).
Among regional accounts of archaeology, China is covered by Kwangchih
Chang, The Archaeology of Ancient China, 4th ed. (New Haven: Yale
University Press, 1987), David Keightley, ed., The Origins of Chinese
Civilization (Berkeley: University of California Press, 1983), and David
Keightley, “Archaeology and mentality: The making of China,”
Representations 18:91–128 (1987). Mark Elvin, The Pattern of the Chinese
Past (Stanford: Stanford University Press, 1973), examines China’s history
since its political unification. Convenient archaeological accounts of
Southeast Asia include Charles Higham, The Archaeology of Mainland
Southeast Asia (Cambridge: Cambridge University Press, 1989); for Korea,
Sarah Nelson, The Archaeology of Korea (Cambridge: Cambridge University
Press, 1993); for Indonesia, the Philippines, and tropical Southeast Asia, Peter
Bellwood, Prehistory of the Indo-Malaysian Archipelago (Sydney: Academic
Press, 1985); for peninsular Malaysia, Peter Bellwood, “Cultural and
biological differentiation in Peninsular Malaysia: The last 10,000 years,”
Asian Perspectives 32:37–60 (1993); for the Indian subcontinent, Bridget and
Raymond Allchin, The Rise of Civilization in India and Pakistan (Cambridge:
Cambridge University Press, 1982); for Island Southeast Asia and the Pacific
with special emphasis on Lapita, a series of five articles in Antiquity 63:547–
626 (1989) and Patrick Kirch, The Lapita Peoples: Ancestors of the Oceanic
World (London: Basil Blackwell, 1996); and for the Austronesian expansion
as a whole, Andrew Pawley and Malcolm Ross, “Austronesian historical
linguistics and culture history,” Annual Reviews of Anthropology 22:425–59
(1993), and Peter Bellwood et al., The Austronesians: Comparative and
Historical Perspectives (Canberra: Australian National University, 1995).
Geoffrey Irwin, The Prehistoric Exploration and Colonization of the
Pacific (Cambridge: Cambridge University Press, 1992), is an account of
Polynesian voyaging, navigation, and colonization. The dating of the
settlement of New Zealand and eastern Polynesia is debated by Atholl
Anderson, “The chronology of colonisation in New Zealand,” Antiquity
65:767–95 (1991), and “Current approaches in East Polynesian colonisation
research,” Journal of the Polynesian Society 104:110–32 (1995), and Patrick
Kirch and Joanna Ellison, “Palaeoenvironmental evidence for human
colonization of remote Oceanic islands,” Antiquity 68:310–21 (1994).
Chapter 18
Many relevant further readings for this chapter will be found listed under
those for other chapters: under Chapter 3 for the conquests of the Incas and
Aztecs, Chapters 4–10 for plant and animal domestication, Chapter 11 for
infectious diseases, Chapter 12 for writing, Chapter 13 for technology,
Chapter 14 for political institutions, and Chapter 16 for China. Convenient
worldwide comparisons of dates for the onset of food production will be
found in Bruce Smith, The Emergence of Agriculture (New York: Scientific
American Library, 1995).
Some discussions of the historical trajectories summarized in Table 18.1,
other than references given under previous chapters, are as follows. For
England: Timothy Darvill, Prehistoric Britain (London: Batsford, 1987). For
the Andes: Jonathan Haas et al., The Origins and Development of the Andean
State (Cambridge: Cambridge University Press, 1987); Michael Moseley, The
Incas and Their Ancestors (New York: Thames and Hudson, 1992); and
Richard Burger, Chavin and the Origins of Andean Civilization (New York:
Thames and Hudson, 1992). For Amazonia: Anna Roosevelt, Parmana (New
York: Academic Press, 1980), and Anna Roosevelt et al., “Eighth millennium
pottery from a prehistoric shell midden in the Brazilian Amazon,” Science
254:1621–24 (1991). For Mesoamerica: Michael Coe, Mexico, 3rd ed. (New
York: Thames and Hudson, 1984), and Michael Coe, The Maya, 3rd ed. (New
York: Thames and Hudson, 1984). For the eastern United States: Vincas
Steponaitis, “Prehistoric archaeology in the southeastern United States, 1970–
1985,” Annual Reviews of Anthropology 15:363–404 (1986); Bruce Smith,
“The archaeology of the southeastern United States: From Dalton to de Soto,
10,500–500 B.P.,” Advances in World Archaeology 5:1–92 (1986); William
Keegan, ed., Emergent Horticultural Economies of the Eastern Woodlands
(Carbondale: Southern Illinois University, 1987); Bruce Smith, “Origins of
agriculture in eastern North America,” Science 246:1566–71 (1989); Bruce
Smith, The Mississippian Emergence (Washington, D.C.: Smithsonian
Institution Press, 1990); and Judith Bense, Archaeology of the Southeastern
United States (San Diego: Academic Press, 1994). A compact reference on
Native Americans of North America is Philip Kopper, The Smithsonian Book
of North American Indians before the Coming of the Europeans (Washington,
D.C.: Smithsonian Institution Press, 1986). Bruce Smith, “The origins of
agriculture in the Americas,” Evolutionary Anthropology 3:174–84 (1995),
discusses the controversy over early versus late dates for the onset of New
World food production.
Anyone inclined to believe that New World food production and societies
were limited by the culture or psychology of Native Americans themselves,
rather than by limitations of the wild species available to them for
domestication, should consult three accounts of the transformation of Great
Plains Indian societies by the arrival of the horse: Frank Row, The Indian and
the Horse (Norman: University of Oklahoma Press, 1955), John Ewers, The
Blackfeet: Raiders on the Northwestern Plains (Norman: University of
Oklahoma Press, 1958), and Ernest Wallace and E. Adamson Hoebel, The
Comanches: Lords of the South Plains (Norman: University of Oklahoma
Press, 1986).
Among discussions of the spread of language families in relation to the
rise of food production, a classic account for Europe is Albert Ammerman
and L. L. Cavalli-Sforza, The Neolithic Transition and the Genetics of
Populations in Europe (Princeton: Princeton University Press, 1984), while
Peter Bellwood, “The Austronesian dispersal and the origin of languages,”
Scientific American 265(1): 88–93 (1991), does the same for the Austronesian
realm. Studies citing examples from around the world are the two books by L.
L. Cavalli-Sforza et al. and the book by Merritt Ruhlen cited as further
readings for the Prologue. Two books with diametrically opposed
interpretations of the Indo-European expansion provide entrances into that
controversial literature: Colin Renfrew, Archaeology and Language: The
Puzzle of Indo-European Origins (Cambridge: Cambridge University Press,
1987), and J. P. Mallory, In Search of the Indo-Europeans (London: Thames
and Hudson, 1989). Sources on the Russian expansion across Siberia are
George Lantzeff and Richard Pierce, Eastward to Empire (Montreal: McGill-
Queens University Press, 1973), and W. Bruce Lincoln, The Conquest of a
Continent (New York: Random House, 1994).
As for Native American languages, the majority view that recognizes
many separate language families is exemplified by Lyle Campbell and
Marianne Mithun, The Languages of Native America (Austin: University of
Texas, 1979). The opposing view, lumping all Native American languages
other than Eskimo-Aleut and Na-Dene languages into the Amerind family, is
presented by Joseph Greenberg, Language in the Americas (Stanford:
Stanford University Press, 1987), and Merritt Ruhlen, A Guide to the World’s
Languages, vol. 1 (Stanford: Stanford University Press, 1987).
Standard accounts of the origin and spread of the wheel for transport in
Eurasia are M. A. Littauer and J. H. Crouwel, Wheeled Vehicles and Ridden
Animals in the Ancient Near East (Leiden: Brill, 1979), and Stuart Piggott,
The Earliest Wheeled Transport (London: Thames and Hudson, 1983).
Books on the rise and demise of the Norse colonies in Greenland and
America include Finn Gad, The History of Greenland, vol. 1 (Montreal:
McGill-Queens University Press, 1971), G. J. Marcus, The Conquest of the
North Atlantic (New York: Oxford University Press, 1981), Gwyn Jones, The
Norse Atlantic Saga, 2nd ed. (New York: Oxford University Press, 1986), and
Christopher Morris and D. James Rackham, eds., Norse and Later Settlement
and Subsistence in the North Atlantic (Glasgow: University of Glasgow,
1992). Two volumes by Samuel Eliot Morison provide masterly accounts of
early European voyaging to the New World: The European Discovery of
America: The Northern Voyages, A.D. 500–1600 (New York: Oxford
University Press, 1971) and The European Discovery of America: The
Southern Voyages, A.D. 1492–1616 (New York: Oxford University Press,
1974). The beginnings of Europe’s overseas expansion are treated by Felipe
Fernández-Armesto, Before Columbus: Exploration and Colonization from
the Mediterranean to the Atlantic, 1229–1492 (London: Macmillan
Education, 1987). Not to be missed is Columbus’s own day-by-day account of
history’s most famous voyage, reprinted as Oliver Dunn and James Kelley, Jr.,
The Diario of Christopher Columbus’s First Voyage to America, 1492–1493
(Norman: University of Oklahoma Press, 1989).
As an antidote to this book’s mostly dispassionate account of how peoples
conquered or slaughtered other peoples, read the classic account of the
destruction of the Yahi tribelet of northern California and the emergence of
Ishi, its solitary survivor: Theodora Kroeber, Ishi in Two Worlds (Berkeley:
University of California Press, 1961). The disappearance of native languages
in the Americas and elsewhere is the subject of Robert Robins and Eugenius
Uhlenbeck, Endangered Languages (Providence: Berg, 1991), Joshua
Fishman, Reversing Language Shift (Clevedon: Multilingual Matters, 1991),
and Michael Krauss, “The world’s languages in crisis,” Language 68:4–10
(1992).
Chapter 19
Books on the archaeology, prehistory, and history of the African continent
include Roland Oliver and Brian Fagan, Africa in the Iron Age (Cambridge:
Cambridge University Press, 1975), Roland Oliver and J. D. Fage, A Short
History of Africa, 5th ed. (Harmondsworth: Penguin, 1975), J. D. Fage, A
History of Africa (London: Hutchinson, 1978), Roland Oliver, The African
Experience (London: Weidenfeld and Nicolson, 1991), Thurstan Shaw et al.,
eds., The Archaeology of Africa: Food, Metals, and Towns (New York:
Routledge, 1993), and David Phillipson, African Archaeology, 2nd ed.
(Cambridge: Cambridge University Press, 1993). Correlations between
linguistic and archaeological evidence of Africa’s past are summarized by
Christopher Ehret and Merrick Posnansky, eds., The Archaeological and
Linguistic Reconstruction of African History (Berkeley: University of
California Press, 1982). The role of disease is discussed by Gerald Hartwig
and K. David Patterson, eds., Disease in African History (Durham: Duke
University Press, 1978).
As for food production, many of the listed further readings for Chapters
4–10 discuss Africa. Also of note are Christopher Ehret, “On the antiquity of
agriculture in Ethiopia,” Journal of African History 20:161–77 (1979); J.
Desmond Clark and Steven Brandt, eds., From Hunters to Farmers: The
Causes and Consequences of Food Production in Africa (Berkeley:
University of California Press, 1984); Art Hansen and Della McMillan, eds.,
Food in Sub-Saharan Africa (Boulder, Colo.: Rienner, 1986); Fred Wendorf
et al., “Saharan exploitation of plants 8,000 years B.P.,” Nature 359:721–24
(1992); Andrew Smith, Pastoralism in Africa (London: Hurst, 1992); and
Andrew Smith, “Origin and spread of pastoralism in Africa,” Annual Reviews
of Anthropology 21:125–41 (1992).
For information about Madagascar, two starting points are Robert Dewar
and Henry Wright, “The culture history of Madagascar,” Journal of World
Prehistory 7:417–66 (1993), and Pierre Verin, The History of Civilization in
North Madagascar (Rotterdam: Balkema, 1986). A detailed study of the
linguistic evidence about the source for the colonization of Madagascar is
Otto Dahl, Migration from Kalimantan to Madagascar (Oslo: Norwegian
University Press, 1991). Possible musical evidence for Indonesian contact
with East Africa is described by A. M. Jones, Africa and Indonesia: The
Evidence of the Xylophone and Other Musical and Cultural Factors (Leiden:
Brill, 1971). Important evidence about the early settlement of Madagascar
comes from dated bones of now extinct animals as summarized by Robert
Dewar, “Extinctions in Madagascar: The loss of the subfossil fauna,” pp.
574–93 in Paul Martin and Richard Klein, eds., Quaternary Extinctions
(Tucson: University of Arizona Press, 1984). A tantalizing subsequent fossil
discovery is reported by R. D. E. MacPhee and David Burney, “Dating of
modified femora of extinct dwarf Hippopotamus from Southern Madagascar,”
Journal of Archaeological Science 18:695–706 (1991). The onset of human
colonization is assessed from paleobotanical evidence by David Burney,
“Late Holocene vegetational change in Central Madagascar,” Quaternary
Research 28:130–43 (1987).
Epilogue
Links between environmental degradation and the decline of civilization
in Greece are explored by Tjeerd van Andel et al., “Five thousand years of
land use and abuse in the southern Argolid,” Hesperia 55:103–28 (1986),
Tjeerd van Andel and Curtis Runnels, Beyond the Acropolis: A Rural Greek
Past (Stanford: Stanford University Press, 1987), and Curtis Runnels,
“Environmental degradation in ancient Greece,” Scientific American 272(3):
72–75 (1995). Patricia Fall et al., “Fossil hyrax middens from the Middle
East: A record of paleovegetation and human disturbance,” pp. 408–27 in
Julio Betancourt et al., eds., Packrat Middens (Tucson: University of Arizona
Press, 1990), does the same for the decline of Petra, as does Robert Adams,
Heartland of Cities (Chicago: University of Chicago Press, 1981), for
Mesopotamia.
A stimulating interpretation of the differences between the histories of
China, India, Islam, and Europe is provided by E. L. Jones, The European
Miracle, 2nd ed. (Cambridge: Cambridge University Press, 1987). Louise
Levathes, When China Ruled the Seas (New York: Simon and Schuster,
1994), describes the power struggle that led to the suspension of China’s
treasure fleets. The further readings for Chapters 16 and 17 provide other
references for early Chinese history.
The impact of Central Asian nomadic pastoralists on Eurasia’s complex
civilizations of settled farmers is discussed by Bennett Bronson, “The role of
barbarians in the fall of states,” pp. 196–218 in Norman Yoffee and George
Cowgill, eds., The Collapse of Ancient States and Civilizations (Tucson:
University of Arizona Press, 1988).
The possible relevance of chaos theory to history is discussed by Michael
Shermer in the paper “Exorcising Laplace’s demon: Chaos and antichaos,
history and metahistory,” History and Theory 34:59–83 (1995). Shermer’s
paper also provides a bibliography for the triumph of the QWERTY keyboard,
as does Everett Rogers, Diffusion of Innovations, 3rd ed. (New York: Free
Press, 1983).
An eyewitness account of the traffic accident that nearly killed Hitler in
1930 will be found in the memoirs of Otto Wagener, a passenger in Hitler’s
car. Those memoirs have been edited by Henry Turner, Jr., as a book, Hitler:
Memoirs of a Confidant (New Haven: Yale University Press, 1978). Turner
goes on to speculate on what might have happened if Hitler had died in 1930,
in his chapter “Hitler’s impact on history,” in David Wetzel, ed., German
History: Ideas, Institutions, and Individuals (New York: Praeger, 1996).
The many distinguished books by historians interested in problems of
long-term history include Sidney Hook, The Hero in History (Boston: Beacon
Press, 1943), Patrick Gardiner, ed., Theories of History (New York: Free
Press, 1959), Fernand Braudel, Civilization and Capitalism (New York:
Harper and Row, 1979), Fernand Braudel, On History (Chicago: University of
Chicago Press, 1980), Peter Novick, That Noble Dream (Cambridge:
Cambridge University Press, 1988), and Henry Hobhouse, Forces of Change
(London: Sedgewick and Jackson, 1989).
Several writings by the biologist Ernst Mayr discuss the differences
between historical and nonhistorical sciences, with particular reference to the
contrast between biology and physics, but much of what Mayr says is also
applicable to human history. His views will be found in his Evolution and the
Diversity of Life (Cambridge: Harvard University Press, 1976), chap. 25, and
in Towards a New Philosophy of Biology (Cambridge: Harvard University
Press, 1988), chaps. 1–2.
The methods by which epidemiologists reach cause-and-effect
conclusions about human diseases, without resorting to laboratory
experiments on people, are discussed in standard epidemiology texts, such as
A. M. Lilienfeld and D. E. Lilienfeld, Foundations of Epidemiology, 3rd ed.
(New York: Oxford University Press, 1994). Uses of natural experiments are
considered from the viewpoint of an ecologist in my chapter “Overview:
Laboratory experiments, field experiments, and natural experiments,” pp. 3–
22 in Jared Diamond and Ted Case, eds., Community Ecology (New York:
Harper and Row; 1986). Paul Harvey and Mark Pagel, The Comparative
Method in Evolutionary Biology (Oxford: Oxford University Press, 1991),
analyzes how to extract conclusions by comparing species.
2003 Afterword
Two articles and one book summarize discoveries of the last half-dozen
years about domestication of plants and animals, spreads of language
families, and the relation of the spreads of language families to food
production: Jared Diamond, “Evolution, consequences and the future of plant
and animal domestication,” Nature 418:34–41 (2002); Jared Diamond and
Peter Bellwood, “The first agricultural expansions: archaeology, languages,
and people,” Science, in press; and Peter Bellwood and Colin Renfrew,
Examining the Language/Farming Dispersal Hypothesis (Cambridge:
McDonald Institute for Archaeological Research, 2002). Those two articles
and that book give references to the detailed recent literature. A recent book-
length account of the role of agricultural expansion in the origins of the
modern Japanese people is Mark Hudson’s Ruins of Identity: Ethnogenesis in
the Japanese Islands (Honolulu: University of Hawaii Press, 1999).
For a detailed account of New Zealand’s Musket Wars, see the book by
R.D. Crosby, The Musket Wars: a History of Inter-Iwi Conflict 1806–45
(Auckland: Reed, 1999). Those wars are summarized much more briefly but
placed in a larger context in two books by James Belich: The New Zealand
Wars and the Victorian Interpretation of Racial Conflict (Auckland: Penguin,
1986) and Making Peoples: A History of the New Zealanders (Auckland:
Penguin, 1996).
Two recent efforts by social scientists to identify proximate causes behind
Europe’s and China’s divergence include an article by Jack Goldstone,
“Efflorescences and economic growth in world history: rethinking the ‘rise of
the West’ and the Industrial Revolution,” Journal of World History 13:323–89
(2002), and a book by Kenneth Pomeranz, The Great Divergence: China,
Europe, and the Making of the Modern World Economy (Princeton: Princeton
University Press, 2000). The opposite approach, the search for ultimate
causes, is exemplified by a recent article by Graeme Lang, “State systems and
the origins of modern science: a comparison of Europe and China,” East-West
Dialog 2:16–30 (1997), and by a book by David Cosandey, Le Secret de
I’Occident (Paris: Arléa, 1997). Those articles by Goldstone and by Lang are
the sources of my quotations above.
The two papers analyzing the connection between economic indicators of
modern wealth or growth rate, on the one hand, and long history of state
societies or agriculture, on the other hand, are: Ola Olsson and Douglas
Hibbs, “Biogeography and long-term economic development,” in press in
European Economic Review; and Valerie Bockstette, Areendam Chanda, and
Louis Putterman, “States and markets: the advantage of an early start,”
Journal of Economic Growth 7:351–73 (2002).
CREDITS
Chapter 12: J. Beckett/K. Perkins, American Museum of Natural History.
Negative 2A17202.
Chapter 12: Courtesy of V.I.P. Publishing.
Chapter 12: Courtesy of Myoung Soon Kim and Christie Kim.
Chapter 12: The Metropolitan Museum of Art.
Chapter 13: Heracleion Museum, Hellenic Republic Ministry of Culture.
BETWEEN PP. 92 AND 93
Plates 1 and 8. Irven DeVore, Anthro-Photo.
Plates 2–5. Courtesy of the author.
Plate 6. P. McLanahan, American Museum of Natural History. Negative
337549.
Plate 7. Richard Gould, American Museum of Natural History. Negative
332911.
Plate 9. J. W. Beattie, American Museum of Natural History. Negative 12.
Plate 10. Bogoras, American Museum of Natural History. Negative 2975.
Plate 11. AP/Wide World Photos.
Plate 12. Judith Ferster, Anthro-Photo.
Plate 13. R. H. Beck, American Museum of Natural History. Negative
107814.
Plate 14. Dan Hrdy, Anthro-Photo.
Plate 15. Rodman Wanamaker, American Museum of Natural History.
Negative 316824.
Plate 16. Marjorie Shostak, Anthro-Photo.
BETWEEN PP. 278 AND 279
Plate 17. Boris Malkin, Anthro-Photo.
Plate 18. Napoleon Chagnon, Anthro-Photo.
Plate 19. Kirschner, American Museum of Natural History. Negative
235230.
Plates 20, 22, 24, 30, and 32. AP/Wide World Photos.
Plate 21. Gladstone, Anthro-Photo.
Plate 23. Above, AP/Wide World Photos. Below, W. B., American
Museum of Natural History. Negative 2A13829.
Plate 25. Marjorie Shostak, Anthro-Photo.
Plate 26. Irven DeVore, Anthro-Photo.
Plate 27. Steve Winn, Anthro-Photo.
Plate 28. J.B. Thorpe, American Museum of Natural History. Negative
336181.
Plates 29 and 31. J. F. E. Bloss, Anthro-Photo.
INDEX
Page numbers listed correspond to the print edition of this book. You can use your
device’s search function to locate particular terms in the text.
Page numbers in italics refer to illustrations, maps, and tables.
Aboriginal Australians:
band societies of, 256, 285
in coastal and river areas, 149, 297
cultural diffusion barriers for, 298, 300–303, 391
current underclass status of, 306
in desert environment, 284, 297
Eurasian diseases and, 205, 304, 307
European conquest of, 98, 297, 305–7, 374, 429
evolutionary ancestry of, 288–89, 302–3
as hunter-gatherers, 98, 108, 149, 285, 295, 296–98, 301
languages of, 289, 303, 314
plant management practices of, 102–3, 149, 296, 297
population size of, 284, 286, 297, 298, 300, 306
racist theories on, 286–88, 307
rock paintings of, 283, 285
technological innovation and, 241, 242, 247, 298–303
villages constructed by, 149, 298
watercraft developed by, 285
weapons used by, 299, 303
wild foods of, 284, 296–98
accelerator mass spectrometry, 92
acorns, 110, 111, 113, 123
Africa, 361–85
Asian links with, 362, 363, 365
Bantu expansion in, 98, 128, 157, 370–71, 377–82, 379 , 391
diffusion barriers in, 227, 228, 250–52, 383–84
diseases in, 189, 196, 200, 205, 381
domesticated animals in, 94, 156, 156, 157, 168, 179, 373, 390
early crops of, 120–21 , 122, 128, 371–75, 372
European conquest of, 179–80, 381, 382–85
five population groups in, 362–65
human evolution in, 35–37, 38, 40, 49, 362, 382
languages of, 314, 362, 365–70, 367 , 374, 375–76
nomadic bands in, 256
non-animal protein sources in, 122
North, Eurasian culture related to, 155
north–south axis of, 179, 181, 251, 383–84
population levels in, 252, 383
racial oppression in, 179–80
spread of food production in, 94, 128, 173, 179, 181
state formation in, 278–79, 280
technological receptivity of societies in, 242
Afroasiatic language family, 366, 368, 376, 378
agave, 121 , 122
agriculture, see crop cultivation; plants
AIDS, 189, 191, 192, 196, 200
Ainu, 159, 164, 341, 411–12, 416–18, 423, 430–31, 432, 434
airplanes, 197, 232, 235
Akihito, Emperor of Japan, 427
Alexander of Macedon, 269, 279, 394, 403, 404
almonds, 109, 113, 123–24
alpacas, 153 , 154, 155, 161 , 171, 204
alphabets, 183, 207–8, 216–18, 220, 225, 226, 244, 248, 309, 319, 352, 385
Altaic Language family, 417
amaranths, 173, 181
Amazonian cultures, 96 , 120–21, 195, 255, 256, 359
American colonies, unification of, 278
Americas:
animal extinctions in, 46, 156, 168, 204, 340, 390
barriers to cultural diffusion in, 170–73, 251, 351–52, 355, 391
current population level of, 359
diseases brought to, 195, 201–3, 342
domestic animals in, 74, 137, 152, 171, 204, 251, 340, 349
historical trajectory of key developments in, 345–55, 347
human presence in, 35, 38, 44–49, 65
modern band societies in, 256
native peoples of, see Native Americans
Norse voyages to, 355–57, 356
north–south axis of, 169, 180, 182, 251, 351, 434
onset of food production in, 92, 94, 95, 96, 347–48, 347, 349–50
population density of, 252
population replacement in, 339, 358–59
spread of food production inhibited in, 170–73, 351–52, 434
technology diffusion in, 244, 251
Americas, European conquest of, 65–78
Atahuallpa’s capture and, 66–72, 74–75, 339
centralized political structure and, 76, 358
horses used in, 74
infectious diseases and, 74–75, 189, 202, 340, 358–59
literacy as factor in, 76–78
maritime technology for, 75–76, 357
progress of, 358–59
weaponry in, 72, 74, 358
amygdalin, 109, 113
Anasazi societies, 434
Andaman Islanders, 318
Andes:
crops of, 94, 96 , 122, 171, 178, 180, 351, 359
survival of Native American population in, 359
animal extinctions:
in Americas, 46, 156, 168, 204, 340, 390
in Australia/New Guinea, 41–43, 46, 156, 168, 292, 295, 390
breeding programs for prevention of, 162
climatic theory of, 46
of domestic mammals’ ancestors, 153 , 168
food production intensified after, 105–6
in Polynesia, 58
animals:
latitude-related climate adaptations of, 177
taming vs. domestication of, 154, 158–59
territorial behavior of, 167
animals, domestic, 151–68
animal extinctions and, 43, 46, 390
crop production enhanced by, 84–85, 94, 122, 315, 341
earliest, 35, 136, 346–47
evolutionary alterations in, 154–55
human diseases and, 82–83 , 87–88, 157, 187–89, 198–200, 199 , 205, 316, 340, 342
initial sites for domestication of, 94–98, 96 , 135–36, 137, 315–16, 373, 434
for land transport, 86–87, 237
as military advantage, 74, 87, 343
morphology of wild species vs., 91, 154–55
number of potential candidates for, 43, 127
as pets, 157, 158–59, 160, 188, 198
as power source, 340, 343
regional differences in, 151–68, 340
as source of fibers, 86, 122, 154, 158, 163–64
wildlife decline as motive for, 105–6
wild mammals’ suitability for, 126–27
see also mammals
anisakiasis, 190
Anna Karenina (Tolstoy), 151, 162, 168
annual plants, 115, 131, 134
Antarctica, 44, 255
antelope, 161, 165–66, 167
apes, human evolution from, 36, 196
apples, 110, 112, 116, 119, 128–29, 146, 149–50, 178
apricots, 116, 178
Arafura Sea, 285, 290
architecture, public, 258 , 262, 263, 267
Aristotle, 271
arrowroot, 296
Asia:
China’s language families in, 310–14, 312 , 313
expansion to Australia/New Guinea from, 40–42, 50–51, 288
initial human presence in, 36, 39
prehistoric coastline of, 287 , 289
asses, North African, 165
Atahuallpa, 66–72, 73, 74–76, 77, 82, 202, 339
atoll types, 57
atomic bomb, 215, 232
aurochs, 154, 157, 160
Australia:
Aboriginals of, see Aboriginal Australians
desert environment in, 283–84, 297
domesticates imported to, 181, 252, 295, 306, 391
environmental limits on food production in, 171, 295–96
European conquest of, 285, 305–7, 359
geological/climatic conditions in, 290, 295–96, 298, 336
mineral resources of, 286
Murray-Darling river system of, 290, 297
New Guinea’s separation from, 285, 286, 287, 289–90, 303
population size of, 252, 306
rabbit eradication efforts in, 201
sheep raised in, 188
Australia/New Guinea:
geographic isolation of, 246, 252
human presence in, 38, 40–44, 48, 50–51, 285, 288, 295
large-animal extinctions in, 41–43, 46, 156, 168, 292, 295, 390
mineral resources of, 231, 286
separation of, 285, 287 , 289–90
stone tools used in, 231
Australopithecus africanus, 36
Austroasiatic language group, 311, 314, 318–19, 330–31, 337, 354, 368, 376, 377, 431–32
Austronesian expansion, 322–38
crops spread by, 330, 335–36
linguistic evidence of, 236, 258 , 322–25, 323 , 328–31, 366, 368
to Madagascar, 326, 362, 365–66, 373, 376–77
to New Guinea, 294–95, 304, 305, 322, 331–36
pottery developments of, 325, 326, 331, 333–35, 336
routes of, 98, 300–301, 304, 325–26, 327 , 328, 331, 332, 336–37, 373, 377
watercraft for, 326–28, 337
automobiles, invention of, 232, 233
axis orientations, 169–83, 170, 250, 251, 315, 351, 383–84
of Africa, 179, 181, 251, 383–84
of Americas, 169, 180, 182, 251, 351
of Eurasia, 169, 176–79, 315, 351, 383–84
latitude-related climate conditions and, 176–79, 384
technological diffusion and, 182–83
Aztec Empire, 345
Spanish conquest of, 201, 340, 358
tribute collected by, 280
warrior religious ideology of, 270
Bali, as part of prehistoric Asia, 288
Bali cattle, 154
Balkans, Fertile Crescent food production spread to, 180
bamboo rafts, 289
bananas, 113, 116, 117, 121, 122, 127, 142, 178, 291, 292, 305, 330, 372, 375, 384
band societies, 195, 254–59, 266, 274–76
banteng, 154, 154 , 157, 161
Bantu:
food production spread by, 128, 179, 182, 378, 381–82, 384
geographic origins of, 369
iron metallurgy of, 378–80
languages of, 314, 354 , 369–70, 377, 378
subequatorial-African expansion of, 98, 157, 370–71, 377–82, 379
bark beaters, 326
barley:
domestication of, 115, 118, 131–32
as founder crop, 120 , 134, 136, 140, 175
nutritive value of, 120, 133, 145
spread of, 315, 318
Bar-Yosef, Ofer, 140
beans, 105, 113, 120, 120, 121 , 145, 172, 173, 180, 181, 351, 434
bears, 159, 164
beer industry, 441–42
bees, 152
beets, 117
Bering land bridge, 38, 41, 44, 45
berries, 109, 110, 111, 112, 123, 124, 146
birds, domesticated, 152, 158, 198, 199 , 373
birth intervals, 85
Bismarck, Otto von, 403
bison, 157, 158, 160, 161
bitter vetch, 136
Black Death (bubonic plague), 188–89, 190, 194, 197, 203, 316, 342
Blackfoot Indians, 81, 82
black pepper, 178
blueberries, 109, 112, 123, 146
blueprint copying:
new writing systems developed through, 215–18
technological diffusion by, 245
Blumler, Mark, 134, 135 , 147
bogong moth, 297
bone tools, 39, 86
bonobos, 36, 259
boomerangs, 299
Borneo:
Austronesian influence in, 322, 326, 327, 330, 333, 366
hunter-gatherer reversion in, 338
as part of prehistoric Asia, 287 , 288
Böttger, Johann, 245
bows and arrows, 247, 285, 299, 303, 343
Brahms, Johannes, 256
brain size:
of domestic vs. wild animals, 154
human, 36, 37, 39
brain structure, language skills and, 39
breadfruit, 122, 142, 330
Britain, Roman authority withdrawn from, 268
broadcast seeding, 122, 341
bronze, 248, 315, 317, 318
bubonic plague (Black Death), 188–89, 190, 194, 197, 203, 316, 342
buffalo:
African, 157, 164, 374, 383
see also water buffalo
bureaucrats, 85–86, 258, 262–63, 268, 269
burial practices, 38, 39, 315
Burke, Robert, 284, 307
business world, organization in, 439–44
cabbage, 113, 117
Cajamarca, Atahuallpa captured at, 66–72, 74
calibrated radiocarbon dates, 35 n, 92
California, University of, at Davis, 110
camels, 152, 153, 156, 159, 161, 374
canal technology, 149, 242, 243, 297, 316
Candia, Pedro de, 68
cannibalism:
disease transmitted through, 190, 199
protein deficiency linked to, 143
canoes, 247, 301, 326–28, 337
capitalism, 239, 437
carbon 14/carbon 12 ratios, 91, 92
see also radiocarbon dating
Carlyle, Thomas, 403
carnivores, unsuitability of as domesticates, 162–63
cashews, 123
cassava (manioc), 121 , 122, 127, 171
cassowary, 142, 158
cast-iron production, 243, 315–16
cats, 152, 167, 198, 373–74
cattle (cows), 93–94, 135, 136, 153, 154, 160, 161 , 162, 179, 198, 199, 341, 373, 374, 384
cavalry, foot soldiers routed by, 74
cave paintings, 39, 47, 50
centralized societies:
in chiefdoms, 262–63, 264, 267
economy controlled by, 268
information flow limited in, 262, 267–68
public order maintained under, 265–66
religious support for efforts of, 266, 344
technological advancement under, 240
cereals:
Australian paucity of, 296
domestication of, 106, 107, 118, 120, 121
as founder crops, 136, 137, 140
nutritive value of, 120, 122, 127, 133
productive yields from, 131–32
sites for onset of cultivation of, 128, 371, 378
technologies developed for farming of, 106–7, 298
tropical climates and, 143
Chalcuchima, 77
charcoal residues, radiocarbon dating of, 91–92
Charles V, Holy Roman Emperor, 66, 71, 72
Chatham Islands:
Austronesian expansion to, 337
hunter-gatherers in, 54, 338
Maori conquest of Moriori settlements on, 52–56
cheetahs, 159, 163
chenopods, 173, 180, 181
Cherokee confederacy, 278
Cherokee language, writing system developed for, 218–20, 219
cherries, 116, 119
chicken pox, 307
chickens, domestication of, 152, 178, 291, 374
chickpeas, 93, 120 , 136
chiefdoms, 257, 258 , 262–65, 266, 267–69, 270, 278–79, 280, 346–47
chili peppers, 172, 173, 180, 181
Chimbu tribe, 241
chimpanzees, 36, 259
China:
agricultural techniques developed in, 119, 178
Austronesian expansion from, 326, 328
crops cultivated in, 120–21, 122, 132, 178, 297, 315, 429, 438
cultural exchanges in, 315–16
cultural expansion from, 178, 311–14, 319, 418, 424
developmental lead lost by, 393–94, 395–401, 437–39, 440
domestic animals of, 152, 178, 315–16
earliest evidence of human presence in, 309
environmental/climatic variety in, 309, 315, 395
ethnic attitudes in, 317–18
genetic diversity in, 309
geographical connectedness of, 397–400, 399, 439
gunpowder developed in, 236, 243, 316
innovations vs. conservatism in, 242–43, 247, 395–97
linguistic history of, 309–14, 318, 324, 330–31, 354, 397, 402
New Guinea immigrants from, 321
non-animal protein sources in, 122
North vs. South, 309, 315, 316–17
political unity of, 309, 318, 324, 396–98, 437–39, 440
pottery from, 243, 244, 245
printing development in, 231, 243, 248
as site of food production origins, 94, 95, 96, 182, 315–16, 395
writing system of, 208, 209, 210, 213, 220, 222, 222 , 225, 226–27, 248, 309, 316, 317, 319, 352,
397, 402
China, People’s Republic of:
Cultural Revolution in, 439
population of, 267, 309
cholera, 188–89, 191–92, 194, 197, 205, 342
cities:
infectious diseases spread in, 197
villages vs., 267
citrus fruit, 178
clans, 260
climate:
animal extinctions and, 43, 46
of Australia vs. New Guinea, 290
biodiversity linked to, 134
cold, 39, 43–44, 45, 357
crop diffusion and, 176–79, 182
of drought cycles, 295–96
latitude-related features of, 176–79
Mediterranean, 131, 133–36, 133 , 177, 384
plant habitats expanded by global changes in, 106
seasonal variation in, 290, 295–96, 371
technological innovation vs., 240
Clovis hunters, 44–49, 348
coffee cultivation, 178, 241, 372
cold climates, human survival in, 39, 43–44, 45, 357
Colledge, Susan, 139
Columbus, Christopher, 65, 76, 189, 202, 203, 205, 321, 339, 395, 396
communal decision process, 261, 275
condors, 162
conflict resolution, societal systems for, 254–55, 257 , 260–61, 269, 274–75
conquest:
diseases spread by, 74–75, 189, 201–4, 342, 358
literacy as factor in, 76–78, 206–7
religious justification for, 67, 69–70, 71, 86, 255, 267, 270, 344, 402
state amalgamation through, 277, 278–80
suicidal fighting in support of, 270
tribute gained from, 280
continental differences in cultural development, 433
axis orientation, 169, 170 , 171, 176–83, 383–84
diffusion factors in, 390–91
domestication potential in, 390
environmental adaptation and, 50–51
geographical connectedness and, 397–400, 399
initial settlement dates and, 49–50
mutability of, 400–401
role of idiosyncracy in, 401–3
total area and population size in, 391–92
Cook, James, 205
Cooke, William, 235
copper metallurgy, 248, 343, 380
corn:
for animal feed, 162
cultivation of, 120 , 121 , 127, 142
diffusion of, 105, 145, 180–81, 351–52, 375
domestication of, 112, 132, 137, 178
nutritive value of, 120, 133, 145, 341
wild ancestors of, 109, 132
Cortés, Hernán, 72, 76, 77, 87, 201, 340
cotton, 86, 114, 121, 122, 173, 181
cotton gin, 232, 235
cowpeas, 120 , 122
cows, (cattle), 93–94, 135, 136, 153, 154, 160, 161 , 162, 179, 198, 199, 341, 373, 374, 384
crafts specialists, 261, 263
Cro-Magnons, 39–40
crop cultivation:
in Africa, 371–75, 372
in altitude range over short distance, 134–35
continental diffusion of, 171–83, 341, 434
in diverse Polynesian environments, 59
domestic animals used in, 84–85, 94, 122, 315–16, 341–42
eight founder crops of, 136
for fibers, 113, 122
of former weeds, 119
of fruit and nut trees, 118–19, 123, 149–50
in Japan, 415, 425–26, 429
latitude-related climate features and, 176–77
linguistic evidence of, 375
by Native Americans vs. Eurasians, 340–42, 434
natural selection and, 111–13, 114–15, 117
nutritive value and, 119–20, 122, 143, 341
for oils, 113–14
preemptive domestication and, 171–73
reproductive biology and, 116
storage factors and, 118, 131
tools and technologies for, 84–85, 106–7, 122, 150, 341–42
twelve leading species for, 127, 131
Cuban Missile Crisis, 268
cucumber, 121, 178
Cuitláhuac, 75, 202
cult houses, 262
cumin, 178
cuneiform, 208, 209–13, 212, 222, 224, 226–27, 229
Custer, George, 72
cycad nuts, 297
Cyril, Saint, 216
Cyrillic alphabets, 216
Daimler, Gottfried, 233
Darwin, Charles, 117, 124
dates (fruit), 118, 128
dating, radiocarbon, 47, 91–93
deer, 165–66, 167, 200
desert environment, 283–84
determinatives, 211
Dingiswayo, 278–79, 280
dingos, 295, 300–301
diphtheria, 203
diseases, infectious:
domestic animals in spread of, 82–83 , 87–88, 157, 187–89, 198–200, 199 , 204–5, 316, 340, 342
epidemics of, 193–205, 342
European conquests furthered by, 74–75, 189, 201–4, 342, 358
food production related to development of, 82–83, 187
four evolutionary stages of, 198–201
genetic defenses against, 192–93
germ transmission strategies and, 190–92, 193, 201
immune defenses against, 192–93, 196, 304, 342, 381
population and, 194–97
sexual transmission of, 188, 191
symptoms of bodily responses to, 191–93
of tropical climates, 75, 189, 205, 342–43
dodo, extinction of, 42
dogs, 136, 146, 152, 155, 158, 160, 163, 166, 204, 295, 373
infectious diseases and, 191, 198, 199
Domestication of Plants in the Old World (Zohary and Hopf), 174
dominance hierarchy, 166, 167
donkeys, 154, 161, 165, 374
drought cycles, 295–96
ducks, 152, 199, 204
dugout canoes, 301, 326–28
dysentery, 195, 304
eagles, 158–59
Easter Island:
giant statues of, 63
writing system of, 213, 221
eastern United States, Native American groups in:
early crops domesticated by, 94, 96 , 120–21, 122, 144–45, 172
indigenous biota vs. imported additions cultivated by, 141, 144–47, 149–50
east–west axes, continental diffusion along, 169, 170, 171, 176–79, 315, 351, 383–84
economy:
centralized control of, 268, 275
of non-food-producers, 256–58, 261, 268, 273, 280
of rich vs. poor nations, 444–46
redistributive, 258, 263–64, 265, 275
productivity in, 440, 441–43
Edison, Thomas, 231, 233, 234, 235
eel fisheries, 298
eggplants, 113
Egypt, ancient:
food production in, 97, 171, 174, 374, 375, 384
hieroglyphs of, 208, 209, 210 , 213, 217, 221, 222–23, 223, 225, 226, 385
Ehret, Christopher, 375
eland, 160–61, 165–66
electric lighting, 234–35, 237, 238
elephants, 154, 159, 163, 383
elk (red deer), 160, 161, 165
English Channel, 41
English language, geographic history of, 370
ENSO (El Niño Southern Oscillation), 296
equids, 165
Eskimos:
Arctic survival skills of, 357
in band societies, 256
Eurasian colonization and, 355
European diseases contracted by, 358
technologies abandoned by, 247
Eta-Funayama sword, 414
Ethiopia:
crops of, 120–21, 122, 174, 178, 372, 376
food production begun in, 94, 96 , 97, 375
guns acquired in, 436
writing system in, 218, 385
ethnic diversity, political incorporation of, 308–9
ethnobiology, 137–40
Eurasia:
defined, 155
diseases from, 189, 197, 204
east-west axis of, 169, 176–79, 315, 351, 383–84, 434
European dominance of, 393–401
food production in Americas vs., 339–42, 434
historical trajectory of key developments in, 345–55, 346
language expansions of, 87, 352, 354 , 359
large-animal extinctions in, 44
large-mammal domestication in, 151–68, 434
population density of, 252
as site for technological innovation, 231, 250–53, 343–44
spread of food production in, 171, 173, 175, 182, 183, 434
technological diffusion in, 244, 245, 246, 248, 250–51
Europe:
Cro-Magnon dominance in, 39, 40
Eurasia dominated by, 393–401
genetic immunities evolved in, 193
infectious diseases from, 74–75, 189
initial human presence in, 36, 48
language replacements from, 314
literate tradition in, 76–78, 345
maritime technology of, 75–76, 344
New World conquered by, 65–78, 189, 339, 340, 358–59
onset of food production in, 96 , 97, 98–99, 104, 434
Pacific conquests of, 338
political/geographic fragmentation of, 396–400, 399, 437–39, 440–41
pottery making in, 98
tropical diseases of colonists from, 189, 205, 304, 343
Fayu bands, 254–55, 256, 258, 277
ferrets, 152, 167
Fertile Crescent (Near East) (Southwestern Asia):
climate of, 131, 133–34
crops of, 118–19, 120–21, 128, 129
developmental lead lost by, 393–95, 400
diffusion process for food package from, 171, 172–80, 174, 182, 375, 376
domesticates spread from, 95, 97, 98
elements of food production package in, 136
environmental and biotic advantages for food production onset in, 129–38
evolution and social organization in, 260, 262
hunter-gatherer decline in, 137
mammal domesticates in, 135–36, 137
map of, 130
natural biodiversity of, 133–36
pottery from, 244
sequence of crop development in, 118–19
as site for food production’s origin, 93, 94–95, 96, 315, 429
technological spread from, 175
topographical variety in, 134–35
fertilizer, 84, 197
fiber production, 86, 114, 121 , 122
figs, 118, 128
Fiji Islands:
European diseases in, 75, 205
guns introduced in, 73
Finnish language, 216
fire:
early human use of, 37
for land management, 297
firearms, see guns
firestick farming, 297
fish, diseases from, 190
fish farms, 197, 297–98
fishing skills, 38–39, 242
Flannery, Tim, 50–51
flax, 86, 114, 115, 121 , 122, 127, 136, 175
food-processing industry, 442–43
food production:
archaeological evidence of, 90–94
combination of local flora needed for, 129–50
continental axis orientations in spread of, 169, 171, 176–83
defined, 82
dense population supported through, 84–85, 107–8, 187, 196, 273–75, 415
diffusion of, 169–83
east–west axis of Eurasian diffusion for, 169, 176–79, 315, 351, 384
evolution of epidemic diseases related to, 82–83, 187–88, 196, 342
geographic differences in history of, 89–90, 94–99, 95 , 390–91
hunting-gathering competition with, 54, 82, 142, 147–48, 350
from indigenous biota vs. imported additions, 141–47
local ethnobiological knowledge in, 139–41
military advantages and, 82–88
non-food-producing specialists enabled by, 85–86, 250, 273
nutrition levels in, 107, 293
onset of, 89–99, 149–50, 169–70, 250, 291, 315
population replacements concurrent with onset of, 97–99, 330, 336, 337
in pre-Columbian America vs. Eurasia, 339–42
sedentary lifestyle, 85, 196, 250, 274
state control of, 267
surplus management for, 85–86, 273
technological developments linked to onset of, 106–7, 250–51, 252, 343–44, 349
writing development and, 226–27
Foré, 138, 199, 259–60
founder domesticates, 95–97, 96, 136
fruit:
onset of cultivation of, 128–29
seed dispersal by, 110–11
seedless mutations of, 113, 116
fur animals, 152
Galton, Francis, 159, 162
Gama, Vasco da, 377, 382, 395
gasoline, extraction of, 236
Gates, Bill, 439, 440
gaur, 154, 154, 157, 161
gazelles, 137, 159, 166
geese, 152
geographic determinism, 392
Germany:
beer industry in, 441–42
unification of, 278
germination, natural inhibitors of, 115–16
germs, evolution of, 87–88, 198–201
see also diseases, infectious
giraffes, 159, 374, 383
glass, 231, 235
glottochronology, 376
goats, 135, 136, 153 , 154, 159, 161 , 166, 179, 374, 384
Goering, Heinrich, 361
Goering, Hermann, 361
Goldstone, Jack, 438
gonorrhea, 205
goosefoot, 120 , 145, 172
gorillas, 36, 162, 163, 259
gourds, 121
government:
communal decision process used for, 261, 275
spread of religion linked with, 255–56
grafting, 119, 150
Grand Canyon, 46
grapes, 110, 113, 116, 118, 128–29, 146
grasses, 291
early cultivation of, 119, 120, 140
worldwide survey of, 119, 120 , 147
see also cereals
Great Leap Forward, 39, 39–40, 51
Greek alphabet, 208, 217, 218, 225–26
Greenberg, Joseph, 353, 366, 368, 369
Greenland, Norse settlement in, 355–56, 357, 391
grizzly bears, 159, 164
groundnuts, 120 , 122
guanaco, 153
guinea fowl, 373
guinea pigs, 152, 154, 171, 180, 204
gunpowder, 216, 236, 242, 243, 316
guns, 73, 231, 238, 244–47, 299, 329, 435–36
Gutenberg, Johannes, 231, 248–49
Halmahera, 323 , 327, 331, 332
han’gu˘l alphabet, 220–21, 221, 319
haniwa statues, 417
Hannibal, 154, 383
harquebuses, 73
Harris, David, 139
haus tamburan, 262
Hawaii:
chiefdoms of, 262–65, 266, 268, 279
epidemic disease in, 205
food production in, 338
isolation of, 227
political unification of, 62–63, 64, 279
hemp, 86, 114, 121 , 122
Henry, Joseph, 235
hepatitis, 304
herd animals, social characteristics of, 165–67
herders:
seasonal movement of, 259
in sub-Saharan Africa, 94, 97–98, 108, 157, 380
hereditary social position, 262, 263, 267, 269
hieroglyphs, 208, 209, 210, 215, 217, 221, 222, 223, 225
Hillman, Gordon, 139
Hirohito, Emperor of Japan, 413
hippopotamus, 164, 374, 383
history, as science, 392, 403–9
Hitler, Adolf, 398, 403, 406
Hmong-Mien (Miao-Yao) language family, 310–11, 314, 318, 330, 354
Hobbes, Thomas, 100
Hokkaido, 412, 416, 418, 421, 431
Homo erectus, 36–37, 323
Homo habilis, 36
Homo sapiens, 37–38
honeybees, 152
Honshu, 416, 426, 434
hookworms, 191, 193, 196
Hopewell culture, 146
Hopf, Maria, 173, 174
horses:
in Africa, 179, 384
in Americas, 156, 341, 343
domestication of, 153 , 154, 157, 161 , 165, 374
Eurasian diffusion of, 87, 251
for long-distance transport, 87
military use of, 74, 87, 157–58, 343
motor vehicles vs., 233, 234
social dominance in bands of, 166
Huascar, 75, 202
Huayna Capac, Emperor of Incas, 75, 202
humans:
biological evolution of, 36–40
geographical colonization patterns of, 36–51
hunter-gatherers:
African herders vs., 98, 108, 157
in band societies, 256, 259
in chiefdoms, 263, 273
disease vulnerability of, 195
end of, 82, 108
ethnobiological knowledge of, 137–39
farmers reverted to, 54, 104–5
food producers’ conquest and replacement of, 98, 108, 330, 336, 337
food production in competition with, 107–8, 142, 147–48, 350
in 1492 Eurasia vs. Americas, 341, 344
in Japan, 164–65, 416, 417, 420–21, 424, 426, 428
landscape management practiced by, 102–3
in modern New Guinea, 142, 292
nutritional status of, 107
population densities of, 44–45, 54, 84–85, 196
sedentary societies of, 86, 131, 137, 139, 423, 428
of Southeast Asia, 318
hunting skills:
animal extinctions and, 41–43, 46, 168
of protohumans, 38–39, 42
Huygens, Christiaan, 234
hydraulic management, 271–72
IBM, 440, 444
Ice Ages, 35
land bridges during, 38, 40–41, 285, 287, 418–19, 431
idea diffusion:
of porcelain technology, 245
writing systems developed through, 215, 218–23
immune system, 192–93, 196
Inca Empire:
animals of, 164
disease epidemics and, 74–75, 358
geographic isolation of, 227, 251
incandescent light bulb, 234–35
Incas, European conquest of:
Atahuallpa’s capture and, 66–72, 74–75
cavalry vs. foot soldiers in conquest of, 74–76
centralized political organization and, 76, 345
literacy as aid to, 76–78
military equipment of, 74–76
India, 441
Chinese influence on, 310
crops cultivated in, 120, 120–21
domesticates from, 178
food production spread to, 173, 175, 182
sea trade routes with, 377
Indo-European languages, regional expansion of, 87, 310, 352–53, 354, 359
Indonesia, 310, 353, 354 , 359
agricultural crops of, 305, 375
Austronesian expansion and, 98, 295, 300, 304, 322, 323–24, 328, 333, 335, 377
colonization of, 41, 50–51
New Guinean cultural influence from, 295
population of, 321
western New Guinea controlled by, 305, 320–22
Industrial Revolution, 117–18, 344
Indus Valley, food production’s development in, 96, 97, 171, 182
influenza, 87, 188–89, 191, 192, 194, 199, 203, 205, 307, 316, 342
information, government control of, 257, 267–68
insects:
diseases transmitted by, 190, 199, 200
domesticated, 152
Irian Jaya, 305
iron metallurgy, 248, 315–16, 317, 346–47 , 378–80
irrigation systems, 197, 264, 271–72, 341
Islam, 242–43, 245
cultural diffusion of, geographic factors in, 246
incendiary weapons in wars of, 236
Iyau, 266
Japan:
agriculture in, 422, 425–26, 429, 434
Ainu in, 159, 164, 341, 411–12, 416–18, 423, 430–31, 432, 434
archaeology of, 413–14, 418–19
Chinese influence on, 310, 318, 418, 424
cultural isolation of, 246–47, 412, 415, 424–25, 428
enmity toward Koreans in, 414, 432
food-processing industry in, 442–43
food production in, 415–16, 419, 420–22, 430
geography of, 414–16
guns abandoned and reacquired in, 246–47, 299, 436
hunter-gatherer lifestyle in, 104–5, 416, 417, 420–21, 424, 426, 428, 430
Ice Ages in, 412, 418–19, 431
Jomon culture in, 419–25, 428, 429–30, 431–32
Kofun period in, 427, 429
Korean immigration into, 411, 425–26, 428–30, 432, 434
Korean trade with, 413–14, 418, 424, 425
mythic history of, 412–13
political and cultural independence of, 412
population density in, 415, 423, 425, 428
pottery from, 244, 319, 419–21, 423, 425
transistor technology acquired by, 238, 245
warfare in, 426–27
writing systems of, 208, 237, 319, 414
Yayoi culture in, 425–27, 428, 429–30, 432
Japanese language, 411, 412, 417, 431–32
Japanese people:
genetic homogeneity of, 411, 429
Koreans as ancestors of, 411, 428–31
origins of, conflicting theories of, 412, 431
Java:
Asian mainland joined to, 287 , 288
Austronesian expansion to, 322, 326, 327
Java man, 36, 323
jewelry, earliest evidence of, 39
jicama, 121
Job’s tears, 143
Jordan Valley, plants selected for domestication in, 140
Kamehameha I, King of Hawaii, 62–63
kangaroos, 142, 156, 295, 297
Kennedy, John F., 268
Khoisan peoples:
Bantu encroachment on, 179, 370, 378–82
genetic background of, 362–65
as hunter-gatherers vs. herders, 97, 108, 157, 380–81
language family of, 314, 366, 368–69, 370, 376
no crops domesticated by, 179, 373, 376, 384
white decimation of, 75, 179–80, 205, 381
Kingdon, Jonathan, 50–51
Kirikiri, 254
Kislev, Mordechai, 140
kleptocracies, four sustaining strategies for, 265–67
knotweed, 120 , 145
koalas, 162
kofun, tombs, 413, 427
kola nuts, 373, 375
konohiki, 263, 268
Korea:
Chinese influence in, 310, 318, 418, 424
enmity toward Japanese in, 414, 432
immigration to Japan from, 411, 425–26, 428–30, 432, 434
Japanese trade with, 413–14, 418, 424, 425
steel industry in, 440
Korean hemorrhagic fever, 191
Korean language, 431–32
kuru (laughing sickness), 190, 199
Kyushu, 416, 418, 421, 425, 431
labor force:
diversity of, 61
seasonal shifts in, 273
technological innovation related to size of, 239
Lang, Graeme, 438–39
Langley, Samuel, 235
languages:
for address of royalty, 269
African diversity of, 365–70, 367 , 374, 375–76
anatomical basis for, 39
ancestral vs. modern, 329
of Australia/New Guinea vs. Asia, 289
of Austronesian family, 311, 314, 318, 330, 337, 354 , 376, 377
of China and Southeast Asia, 309–14, 312, 313 , 317–18
click sounds in, 369
cultural history deduced from, 329
Eurasian movements of, 87, 352, 354 , 359
of Native Americans, 314, 352–53, 359
political connectedness reflected in, 397
replacements of, 311–14, 317–18
writing systems modified for differences in, 207, 216–17
Lapita pottery, 333–35, 336
Lascaux Cave, 39
Lassa fever, 200
latitude, climate features related to, 176–79
laughing sickness (kuru), 190, 199
leather, 86
leeks, 119
legumes, 118, 119
lentils, 115, 119, 120 , 136
leprosy, 196
lettuce, 117, 118, 119
Lévi-Strauss, Claude, 225
Lewis, Bill, 441
Lilienthal, Otto, 235
lima beans, 113, 120, 172, 181
Linearbandkeramik culture, 84
Linear B writing system, 208, 216, 224–25, 230
lions, 163, 164
Liszt, Franz, 256
literacy, see writing systems
llamas, 153, 154, 155, 161, 171, 180, 204, 251, 351
logograms, 208, 224–25, 226
Los Angeles public schools, ethnic diversity in, 308
luxury goods, 258 , 263
Lyme disease, 200
macadamia nuts, 123, 296, 307
Macassans, Australian Aboriginal contact with, 301
Madagascar, Austronesian expansion to, 326, 362, 365–66, 373, 376–77
Malai Islet, 334
malaria, 188–89, 192, 193, 203, 205, 259
altitudinal ceiling for, 305
European susceptibility to, 304, 306, 342, 343
immunity to, 336, 381
transmission of, 190, 197
Malayo-Polynesian language group, 322, 323, 324–25, 328, 330
Malay Peninsula, Austronesian expansion to, 322, 323, 327 , 328, 330
mammals:
aquatic, 152
extinctions of, 46, 153, 168, 204, 340, 390
large species of, 41, 46
as milk sources, 84
small, as domesticates, 152, 159
mammals, large, domesticated, 151–68
African paucity of, 373–74, 383, 384
aquatic, 152
captive breeding suitability of, 163
continental diffusion of, 171, 179, 180
diet requirements of, 162–63
domestication dates for, 137, 159, 161
fourteen ancient species of, 153–54 , 154, 155–57, 159, 340
growth rates of, 163
modern efforts at developing, 157, 160–61
rapid acceptance of Eurasian species of, 157
size of, 154
social characteristics of, 166–67
suitability criteria for, 126–27, 151, 159–68, 383
temperament as factor in, 164–66
unequal global distribution of candidates for, 136, 155–57, 168, 340, 373–74, 390, 393
Manco, 74
Mandan Indians, 203, 358
Mandarin, 309, 310
Manhattan Project, 232
manioc (cassava), 121 , 122, 127, 171
manure, 84, 340, 342
Maori:
British defeat of, 86
Moriori conquered by, 52–56
muskets adopted by, 244–45, 390–91, 435–36
New Zealand colonized by, 45, 50
maritime technology:
of Austronesian expansion, 300, 326–28
Eurasian origins of, 231
of European expansion, 75–76, 344
marsupials, extinct, 292, 295
Marx, Karl, 265
Matthew, Saint, 168
Maya societies:
barriers to cultural diffusion from, 251
writing system developed by, 208, 213, 225
maygrass, 121, 145
McKinsey Global Institute, 441, 442
Meadowcroft rock shelter, 47, 48
measles, 87, 188–89, 194–95, 198, 199, 203, 205, 307, 342
Mediterranean region:
climate typified by, 131, 133–36, 133 , 177, 384
earliest evidence of watercraft in, 41
megafauna, extinctions of, 41–43, 46, 156, 168, 204, 340
melons, 111, 113, 121 , 175
Mena, Cristóbal de, 76
Merina state, 279
Mesoamerica:
barriers to cultural diffusion from, 251, 351–52
diffusion to and from South America in, 171, 173, 180
domestic animals in, 137, 152, 171, 204
early crops of, 96, 120–21 , 122, 172, 434
languages of, 353
non-animal protein sources in, 122
as site of food production’s origin, 94, 96
social organization begun in, 262
technological advances in, 237, 355
writing systems developed in, 183, 208, 209, 212–13, 214, 221, 225, 226–27, 345, 352
metallurgy, 175, 244, 248, 251, 315–16, 337, 343, 346, 346–47, 348
Miao-Yao (Hmong-Mein) language family, 310–11, 314, 318, 330, 354
Microsoft, 440, 444
military:
Eurasian technological advantages in, 342
food supplies for, 86
ideology of patriotic suicide in, 270
religious motivation of, 67, 69–70, 71, 86, 255, 267, 270, 344
milk production, 84, 154, 175
millets, 120 , 122, 178, 298, 315, 371, 372, 374, 378
missionaries, 255, 274
mithan, 154
Mobotu Sese Seko, 265
Mongol Empire, 352
Mongoloids, 309
monkey viruses, 189, 196, 199, 200
monoculture fields, 122
Monte Verde site, 48
Montezuma, 75, 77
moose, 160, 161
Moriori society, Maori conquest of, 52–56
Morse, Samuel, 235
mosquitoes, 190, 197, 199, 381
moths:
as food, 297
natural selection for industrial melanism in, 117
motor vehicles, invention of, 233
Mtetwa chiefdom, 278, 280
mumps, 194, 196, 203
Muralug Island, 302
murder, among band and tribal societies, 254–55, 266
Murray-Darling river system, 290, 297
Muscovy ducks, 152, 204
mushrooms, 109, 138
muskets, 244–45, 390–91, 435–36
musk ox, 160
mustard seeds, 114, 140
myxomatosis, 200–201
Namibia, colonial history of, 361–62
Native Americans:
crops cultivated by, 105, 341, 434
cultural diversity of, 308
disease epidemics among, 74–75, 189, 191, 194, 195, 201–3, 342, 358
domestic animals of, 158, 204, 340, 341
of Eastern U.S., 94, 96, 120–21 , 122, 141, 144–47, 149–50, 172
Eurasian food production vs., 339–42, 434
European conquest of, 65–78, 81–82, 189, 202, 314, 339–60, 429
geographic/ecological isolation of, 171–73, 227, 228, 246–47, 352, 391
guns acquired by, 436
as hunter-gatherers, 81, 98, 108, 263, 341, 348, 349–50, 352
independent inventions of, 237, 244, 352, 354
innovation vs. tradition among, 242
languages of, 314, 352–53, 359
of Mississippi Valley, 202, 227
population levels of, 203, 204, 359
technological disadvantages of, 343–44
of West Indies, 204, 358
writing systems of, 208, 209, 212–13, 214, 218–20, 219, 225, 228, 345
see also specific Native American groups
natural selection:
for disease immunity, 192–93
human crop cultivation vs., 111–13, 115, 117, 124–25
Nature, 47
Navajo, 158, 242, 341
Neanderthals, 38, 39, 43
Near East, see Fertile Crescent
Negritos, 318, 324, 368
Newcomen, Thomas, 234
New Guinea:
Australia separated from, 285, 286, 287 , 289–90
Austronesian expansion to, 294–95, 304, 305, 322, 331–36
diverse languages of, 289, 309–10, 323, 331–32
domestic animals in, 143, 291, 293, 295, 302
early crops of, 120–21 , 122, 142
environmental conditions in, 141, 290, 293
European presence in, 286, 295, 303–5
food production in, 141–44, 286, 290–93, 295, 305
geographic barriers to cultural diffusion in, 293–94, 391
giant marsupials exterminated on, 292
indigenous biota vs. imported additions cultivated in, 143–44
indigenous fauna of, 141, 143, 290
Indonesian province of, 305, 320–22
initial human presence in, 41–44, 141, 288
intertribal warfare in, 294
onset of food production in, 94, 95 , 96 , 291, 293
political fragmentation in, 294
population density of, 286, 292, 293
Torres Strait population from, 301–3
see also Australia/New Guinea
New Guinea, modern:
art of, 292
band societies of, 254–55, 256, 258–59, 286
Chinese immigrants as, 321
diseases of, 196, 199, 304
eastern vs. western, 305
ethnic tensions among, 320–22
ethnobiological expertise of, 138, 140, 142, 144
European colonization among, 286
evolutionary ancestry of, 288–89, 318, 321–22, 331
highland agriculture vs. lowland subsistence for, 142–44, 292, 302, 321
innovative vs. conservative cultures in, 241–42
languages of, 259, 289, 294
Native Australians vs., 285–86, 290
pets kept by, 158, 160
population distribution in, 292, 293–94
stone tools of, 37, 286
tribal groups of, 138, 199, 259–62, 266, 286, 293
New Zealand:
Austronesian expansion to, 337
geological diversity of, 57
Maori ancestors in, 45, 53
mineral resources of, 57, 63, 64
Musket Wars in, 244–45, 390–91, 435–36
Niger-Congo language family, 366, 368–70, 373, 375, 376
Nilo-Saharan language family, 366, 368, 376, 378
Ninan Cuyuchi, 75
Norse, North Atlantic expansion efforts of, 355–57, 356, 391
North Africa, Eurasian culture related to, 155
north–south continental axes, 169, 171, 179–83, 251, 351, 383–84
nuts, 109, 110, 113, 123, 142, 146, 373, 419, 430
oak trees, 110, 113, 123, 146
oats, 119, 178
oca, 121, 122
ogham alphabet, 217, 220–21
olives, 110, 114, 118, 128
onagers, 165
On the Origin of Species (Darwin), 124
Optimal Fragmentation Principle, 437, 440–41
oranges, 113, 116
Otto, Nikolaus, 233
Pacific Northwest, hunter-gather chiefdoms in, 263
papermaking techniques, 243, 245, 248, 316
Papin, Denis, 234
Papuan languages, 289, 332
Patagonia, Clovis hunter-gatherer expansion to, 44
patent law, 233, 239
patriotism, conquest furthered by, 270
peaches, 116, 178
peanuts, 120, 171, 375
pears, 119
peas, 91, 110, 112, 114, 115, 118, 120 , 122, 136, 175, 176
pecans, 110, 123, 146
peccaries, 151, 157
Pedra Furada, cave paintings at, 47
pertussis (whooping cough), 191, 194, 199, 203
petroleum products, 236–37
pets, 157, 158–59, 160, 188, 198
Phaistos disk, 229–30, 230 , 235, 243, 248–49
Philippines:
Austronesian expansion to, 98, 323–24, 328, 333, 335
crops brought to, 143
food production’s spread from, 171
languages of, 314, 318, 322, 323–24, 328, 329, 368
phonemes, 208, 209
phonograph, invention of, 233
pigs:
domestication of, 135, 136, 153, 154, 157, 159, 161 , 291, 295, 316, 373, 415
human diseases and, 190, 204
pineapples, 116
Pizarro, Francisco, 66–73, 74–77, 82, 87, 202, 339, 344
Pizarro, Hernando, 67, 68, 77
Pizarro, Juan, 68
Pizarro, Pedro, 67
plague, bubonic (Black Death), 188–89, 190, 194, 197, 203, 316, 342
plants:
atmospheric carbon absorbed by, 91
dioecious species of, 116
reproductive processes of, 116, 132–33
self-fertilization of, 116, 119
plants, domestication of, 109–50
alterations undergone through, 91, 110–18, 131–32, 141
in China, 315
defined, 109
earliest known dates for, 35, 94–95, 96 , 346–47, 347–48
hermaphroditic selfers and, 132–33
initial sites for, 93, 94–98, 96 , 244, 291, 373, 375, 433–34
interspecific hybrids, 133
in Japan, 415
local ethnobiological knowledge employed in, 137–39
modern lack of major additions to, 127–28
natural selection process vs., 111–13, 114–15, 117, 124–25
in New Guinea, 291
nutritional yield and, 84, 119–20, 133, 137, 144, 145, 341
preemptive domestication and, 171–73
prehistoric climate change and, 106
regional potential for required variety of, 128–50, 383, 393
single vs. multiple instances of, 172, 175, 181, 434
size increases in, 112
variations in ease of, 118–24, 132
wild mutants in, 114–16, 124, 171–72
plants, wild:
almonds, 109, 113
of Australia, 296
berries, 109, 111
bitterness of, 109, 112–13
cereals, 106–7, 131, 132
domestication potential of, 126–28, 131–33
food production’s onset based on entire regional assortment of, 128–31
germination inhibition in, 114–15
grass species, 134, 135 , 147
local ethnobiological knowledge of, 137–40
number of species of, 127
poisonous, 109, 112–13, 138
Plato, 265
Pleistocene Era, end of, 35
plow animals, 84–85, 122, 315, 341–42
plums, 116, 119, 146
pneumonia, 188
polio, 196
political systems:
centralized, 76, 240, 262–68, 344, 433
ethnic diversity encompassed by, 308–9
of Eurasian societies vs. Native Americans, 344–45
of Inca Empire, 76, 345
of kleptocracies, 265–67
Polynesian diversity of, 60–63
population density and, 60–61, 62, 274
in sedentary societies, 85–86
Spanish conquests enabled by, 76
spread of religion linked with, 255–56
technological advancement and, 240, 437–39
units of, 60
writing development linked to, 224–27
see also social organization
Polynesian islands:
Austronesian expansion to, 322
chiefdoms in, 262–65, 266, 267, 269, 279
domestic animals on, 58, 143
human adaptation to diverse environments in, 54–64, 338
languages of, 314, 323, 329
metallurgy and writing absent on, 338
seafaring expertise developed on, 322
spread of food production to, 171, 179
technologies abandoned on, 247, 299
Polynesian islands, environmental variations in:
agricultural development influenced by, 59, 106
in climate, 56
economic specialization and, 61–63
geological types of, 56
isolation and, 58, 60, 62
marine resources of, 57
material culture and, 62
political organization and, 60–63
in size, 57
subsistence practices and, 58–60
in terrain fragmentation, 57–58, 60
pomegranates, 118
poppy cultivation, 96–97, 114, 174, 178
population density:
agricultural productivity vs., 59, 85, 293
bidirectional links of food production with, 84–85, 107–8, 187, 196, 273–75
defeated peoples’ fates tied to, 279
epidemic diseases linked with, 82–83, 194–95, 196–97
labor force diversity and, 61
political organization affected by, 60, 62, 274
of Polynesian environments, 60–61
for sedentary society vs. nomadic peoples, 85
societal complexity related to, 272–75
population growth, 44–45, 55
population size:
for epidemic diseases, 194–97
social organization and, 255–56, 258, 260, 262, 267, 272–80, 433
technological development linked to, 246, 250, 251, 252, 355, 391–92, 433
porcelain, 243, 245
potatoes, 113, 121 , 122, 127, 178, 180, 435–36
sweet, 121, 122, 127, 143–44, 147, 171, 291–92, 305, 435–36
pottery:
in Africa, 251–52, 384–85
of Austronesian expansion, 325, 326, 331, 333–35, 336
from conquering cultures, 98
first appearances of, 244, 250, 319, 346–47
furnace technology for, 248
in Japan, 244, 319, 419–21, 423, 425
Polynesian abandonment of, 247, 299
porcelain, 243, 244, 245
power sources, mechanical, 343–44
preemptive domestication, 171–73, 175
priests, 86, 225, 266, 269
printing methods, 230–31, 243, 248–49
protein sources:
animals as, 136, 143
deficiencies in, 143
non-animal, 120, 136, 143, 145, 341
Proto-Indo-Europeans, 329
protolanguages, 329
public works, 264, 266, 267, 273
pulses, 120, 120 , 121, 127, 136, 143
Punan, 338
Pygmies, 314, 318, 324, 362–65, 368, 370, 373, 376, 378, 380
quince, 178
quinoa, 120 , 122
quipu, 345
Quizo Yupanqui, 74
QWERTY keyboards, 237–38, 401–2
rabbits, 152, 159, 198, 200–201
rabies, 191
radiocarbon dating, 35 n, 47, 91–93
radishes, 119
ragweed, 145
rainfall, crop diffusion and, 182, 183
Recent Era, 35
red deer (elk), 160, 161, 165
redistributive economy, 257, 263–64, 265, 275
reindeer, 153, 154, 166, 167, 341
religion:
conquest justified by, 67, 69–70, 71, 86, 255, 267, 270, 344, 402
kleptocracies supported by, 257 , 266
spread of government linked with, 255–56
state leaders elevated by, 269
technological innovation and, 240
tribal supernatural beliefs institutionalized as, 266
resource supply, technological innovation vs., 240
rhinoceros, 162, 167, 374, 383
rice cultivation, 120, 120 , 127, 133, 143, 145, 315, 317, 318, 373, 415, 430, 434
rinderpest, 198
rock paintings, 283, 285
rodents, 152, 197, 200
Roman alphabet, 216, 218, 226
Roman Empire:
food production in, 178
geographic range of, 352, 398
root crops, 121, 122, 127, 142, 143
Rothschild, Lord Walter, 165
Rousseau, Jean-Jacques, 271, 277
rubella, 191, 194
Russia, Imperial, ethnic diversity incorporated in, 308–9
Russian language, 216, 352–53, 354
Russo-Japanese War, 436
rye, 119
sago palm tree, 142, 259, 292
Sahara, African food production begun in, 374
Sahel zone:
crops of, 120–21, 132, 179, 371, 384
east–west crop diffusion in, 179
metallurgy from, 378
onset of food production in, 94, 96
salmonella, 190
samurai, 246–47
San people, 75, 365
see also Khoisan peoples
Savage, Charlie, 73
Savery, Thomas, 234
schistosomes, 191, 197
science, history as, 392, 403–9
sea transport, see maritime technology; watercraft
sedentary societies, 85
disease transmission in, 196
hunter-gatherers as, 86, 131, 137, 139, 423, 428
population density and, 85
state control and, 274
technological innovation fostered in, 249–50
seeds:
broadcast sowing of, 122, 341
of cereals, 131–32
competitive pressure of farming environment for, 118
crop cultivation for, 113, 131
mutant, 114–16
natural dispersal and germination of, 110–11, 112–13
in pods, 114
Sejong, King of Korea, 220
selfers, 132–33
Semang Negritos, 318, 324, 368
Semitic languages, 217, 367
sense organs, of domestic vs. wild animals, 154
Sepik River, 262
Sequoyah, 218–20, 219, 223
sesame seeds, 114, 178
sheep, 135, 153, 154, 157, 158, 159, 161, 166, 167–68, 174, 179, 188, 341, 374, 384
Shikoku, 416, 426
Siberia, hunter-gatherers in, 98, 341, 344, 355
sickness, earliest evidence of care during, 38
silk production, 152, 245
Sino-Tibetan language family, 310, 314, 315, 317, 318, 354
skull development, human, 37
slash-and-burn agriculture, 292, 294, 302
slavery, 196, 202, 239, 258 , 261, 263, 268, 280, 362
sleeping sickness, 190, 192
smallpox, 74–75, 87, 191, 194
among Aboriginal Australians, 307
domestic animals related to, 188–89, 199
first appearance of, 196, 316
in Hawaii, 208
immunity to, 342
modern control of, 304
Native Americans killed by, 191, 202, 203, 340, 358
as Plague of Antoninus, 197
transmission of, 191, 342
social contract, 271
social organization:
amalgamation of, 276–80
in bands, 195, 254–59, 266, 274–76
in business world, 439–44
in chiefdoms, 257, 258 , 262–65, 267–69, 271, 278–79, 280, 346–47
conflict resolution methods related to size of, 254–55, 260–61, 274–75, 434
disintegration of, 270, 276
food production linked to, 272–75, 434
four categories of, 256–69, 257 , 258
hereditary status in, 262, 263–64, 267–68, 269
of Native Americans vs. Eurasians, 344–45
population size and, 255–56, 257 , 260, 262, 267, 272, 433
as state, 256–80
technological innovation vs., 239
of tribes, 257 , 258, 259–62, 266
wealth distribution and, 261, 263, 265–67
sorghum, 120, 122, 127, 128, 179, 371, 374, 378
Soto, Hernando de, 68, 202
South Africa, food production’s spread inhibited in, 171
South America:
crops of, 122
domestic animals from, 204
sites of food production’s origin in, 94, 96
Southeast Asia:
Austronesian expansion from, see Austronesian expansion
Chinese language families in, 311, 312 , 313 , 314
human ancestors in, 36, 289
prehistoric coastline of, 287 , 289
repopulation of, 310, 318
Southwest Asia, see Fertile Crescent
soybeans, 120, 120 , 127
squashes:
as containers, 145
domestication of, 105, 113, 117, 121 , 145, 172, 173
spread of, 145–46, 173, 180, 181, 352, 434
Stalin, Joseph, 215
state societies, 257, 267–70, 446
in Americas vs. Eurasia, 344–45, 346–47
archaeological evidence of, 267
bureaucrats in, 258 , 263, 268, 269
conditions for formation of, 270–80
military advantages of, 270, 344
multiple ethnicities in, 269–70, 308–9
percentage of globe occupied by, 255, 272
religion in support of, 269, 344
suicidal patriotism and, 270
steam power, 231, 232, 234, 344
steel:
African manufacture of, 380
Eurasian development of, 231
Native Americans conquered with, 72, 74
stone tools:
of Aboriginal Australians, 285
archaeological identification of, 47
of Clovis hunters, 44
earliest use of, 36, 37–38, 249
for farming, 292, 298
in modern societies, 36, 37, 231, 242, 286
standardization of, 39
from volcanic stone, 57, 63
strawberries, 109, 110, 111, 112, 118, 123, 124, 146
street lighting, gas vs. electricity for, 237, 238
sugarcane, 120 , 127, 142, 291
Sumatra, Asian mainland joined to, 287
Sumerian cuneiform, 208, 209–13, 212 , 221, 222, 224, 226–27
sumpweed, 145
sunflowers, 114, 115, 117, 145, 146, 180
supernatural beliefs, religious institutionalization of, 266
supersonic transport, 237
sushi, 190
sweet potatoes, 121 , 122, 127, 143–44, 147, 171, 291–92, 305, 435–36
swords, 73, 246–47, 343
syllabaries, 208, 211, 217, 218–20, 219, 223, 226, 249
syphilis, 191, 201, 203, 205, 208, 307, 342
Tahiti, unification of, 279
Tai-Kadai language group, 311, 314, 330, 336
Taiwan:
Asian mainland connected with, 287
Austronesian expansion begun in, 325–26, 328–29, 330
languages of, 323 , 325, 328, 329
Ta-p’en-k’eng culture on, 325–26
tamarind trees, 301
tannins, 123
Tanzania, languages of, 369
taro, 121, 122, 142, 143, 178, 291, 330, 373
Tasmania:
cultural isolation of, 242, 246, 300
dogs adopted on, 158
first human presence in, 288
technological innovation abandoned in, 247, 300
taxation, 86, 257, 263, 265, 267
technological advances, 229–53, 436–37
applications found after discoveries of, 232–33, 234
autocatalytic tendencies of, 247–49
commercial motivations for, 233–34
cumulative development of, 235
food production linked to, 250–51, 252, 298, 343–44, 349
geographic/ecological factors and, 182–83, 245–47, 251, 252, 400
heroic view of, 231, 234
intercontinental differences in, 250–53, 343–44
local invention vs. diffusion of, 244, 247–48, 349
moderate political connectedness as optimal condition for, 400
necessity as impetus for, 232–34
population size and, 246, 250, 252, 300, 355, 391–92, 433
technological advances, societal receptivity for, 148, 234–35
diffusion conditions for, 244–47, 250, 396–97
economic motives for, 237, 239, 249
historical reversals in, 246–47, 298–300, 396–97, 400
ideological atmosphere for, 239
perceived advantage and, 238
perceived need as motive for, 232–34
prestige considerations in, 237
in single societies or continents, 148, 241–43
standard explanations for, 239–41
trial-and-error examples of, 236
vested interests in opposition to, 238
teff, 120, 372
telegraph, 235
Tell Abu Hureyra, evidence of selection in plants gathered at, 139, 141
temples, 262, 263, 267, 269
teosinte, 132
territorial behavior, 167
Third Chimpanzee, The (Diamond), 39
tobacco, 181, 301
Tolstoy, Leo, 151, 168
Tonga, isolation of, 227
tools:
bones used in, 39, 86
for crop production, 84–85, 106–7, 341–42
metal, 346, 346–47 , 348, 426, 430, 434
natural resources for, 57, 63
see also stone tools
Torres Strait, islands of, 286, 301–3
trade routes:
diseases transmitted along, 197, 342–43
of Indian Ocean, 377, 384
transistor technology, 238, 245, 400
trees:
annual growth rings of, 92
fruit, 114, 118–19, 150, 175
oak, 110, 113, 123, 146
sago palm, 142, 259, 292
tribal organization, 257 , 258, 259–62, 266
tribute, 258, 262, 263, 265, 267, 280
trichinosis, 190
tropical rain forests, latitude limitations for, 177
trypanosome diseases, 157, 179, 204, 384
tsetse flies, 157, 179, 190, 384
tuber crops, 121, 122, 127
tuberculosis, 188–89, 193, 194, 199 , 203, 204, 205, 307, 342
turkeys, 137, 152, 180, 204
turnips, 119
Tutankhamen, 113
typewriter keyboards, letter sequence designed for, 237–38, 401–2
typhoid, 205, 307
typhus, 190, 200, 203, 307, 342
Ulfilas, 216
United States:
beer industry in, 441–42
ethnic diversity of, 308
food-processing industry in, 443
technological innovation in, 440, 444
upright posture, 36
vaccination, 192
Valverde, Vicente de, 69–70
Veddoid Negritos, 318
venereal diseases, 191
vicuña, 163–64
Vietnam, languages spoken in, 311, 322
village life, 35, 346–47
violence, government curbs on, 265–66
volcanic islands, 57, 63
war:
diseases transmitted by, 189
horses used in, 72, 74, 87, 157–58, 233, 343
societal amalgamation fostered by, 276–80
technological advancement affected by, 240, 244–45
Washington, George, 265
water buffalo, 153, 154, 157, 161, 315
watercraft:
for Austronesian expansion, 325–28, 337
canoes, 247, 301, 326–28, 337
cultural reversals of, 247
earliest evidence of, 41, 44, 285
Eurasian vs. Native American, 344
for transatlantic crossing, 357
watermelons, 113, 121, 174
water power, 343–44
Watts, James, 231, 234
wealth distribution:
in chiefdoms, 263
for elite vs. general population, 265–67
in smaller social organizations, 261
weaponry:
of Aboriginal Australians, 299, 303
boomerang, 299
bows and arrows, 247, 286, 299, 303, 343
cultural attitudes about, 246–47
elite monopoly on, 265
European advantage in, 343
guns, 73, 231, 238, 244–47, 299, 329
incendiary, 236
multipiece construction for, 39
muskets, 244–45
of Native Americans, 72, 73
of steel, 73, 343
swords, 73, 246, 343
technology diffusion of, 238, 244–45, 246–47
weaving, 158, 242, 250
wheat:
diffusion of, 315, 318
domestication of, 93, 115, 117, 118, 120, 128, 131–32, 140
ease of germination of, 115
einkorn, 128, 133, 136, 174
emmer, 93, 133, 134, 136, 140
nutritive value of, 120, 133, 136, 144, 145
worldwide production of, 127, 142
Wheatstone, Charles, 235
wheels, 175, 182–83, 237, 244, 251, 343, 344, 352, 354
Whitney, Eli, 232, 235
whooping cough (pertussis), 191, 194, 199, 203
wild boar, 153
wild foods, decline in availability of, 105–6
Wills, William, 284, 307
wind power, 343
wolves, 152, 155, 160, 166
wool, 122, 154, 158, 164
Wright brothers, 231, 235
writing systems, 64, 206–28
accounting records as stimulus for, 209, 218, 345
alphabetic, 183, 208, 216–18, 221, 224, 225–26, 244, 248, 309, 319, 352, 385
blueprint copying and modification of, 215, 216–18
expressive limitations of, 224–26
geographic and ecological barriers to spread of, 226–28, 385
idea diffusion as source for development of, 215, 218–23
independent invention of, 208–13, 210 , 221, 226, 244
language differences and, 207, 216–17
in Mesoamerica vs. Eurasia, 345
as military advantage, 76–78, 206–7
on Phaistos disk, 229–31, 230
phonetic principle employed in, 211, 224, 401
power of information transmittal through, 76–78, 206–7, 345
printing technology and, 229–31, 248–49
sites of origin of, 207, 226–27, 244, 346–47
sociopolitical organization linked to early use of, 224–26
spread of, 175, 183, 207–8, 337, 385
state societies and, 269, 345
three basic strategies used in, 207–8, 212
in western Eurasia vs. China, 317
Wu Li, 222
Xhosa, 381
Yahi Indians, 358–59
yak, 153, 154, 155, 157, 161
Yali, 35, 37, 248, 389–92
yams, 121, 122, 142, 178, 291, 292, 297, 330, 373, 375
yaws, 196
yellow fever, 196, 203, 205, 343
yucca, 121 , 122
Yumbri, 337
Zaire, kleptocratic practices in, 265
zebras, 151, 157, 160, 165, 374, 383
Zhou Dynasty, 311, 317
Zohary, Daniel, 173, 174
zoos, breeding programs at, 162, 163
Zulu state, 278–79, 280
*Throughout this book, dates for about the last 15,000 years will be quoted as so-called calibrated radiocarbon dates, rather than as conventional, uncalibrated radiocarbon dates. The difference between the two types of dates will be explained in Chapter 5. Calibrated dates are the ones believed to
correspond more closely to actual calendar dates. Readers accustomed to uncalibrated dates will need to bear this distinction in mind whenever they find me quoting apparently erroneous dates that are older
than the ones with which they are familiar. For example, the date of the Clovis archaeological horizon in North America is usually quoted as around 9000 B.C. (11,000 years ago), but I quote it instead as
around 11,000 B.C. (13,000 years ago), because the date usually quoted is uncalibrated.
Archive Assistant
Greetings. I am ready to analyze this library entry for you.