Get Smarter

Download as pdf or txt
Download as pdf or txt
You are on page 1of 8

The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.

com/magazine/print/2009/07/get-smarter/7548/

July/August 2009 Print | Close

PAN D E M I C S . GL O B AL W AR M I N G . FO O D S H O R T AG E S . N O M O R E F O S S I L F U E L S . W H AT AR E H U M AN S T O D O ?

T H E S AM E T H I N G T H E S PE C I E S H AS D O N E B E F O R E : E V O L VE T O M E E T T H E C H AL L E N G E . B U T T H I S T I M E W E

D O N ’ T H AVE T O R E L Y O N N AT U R AL E V O L U T I O N T O M AK E U S S M AR T E N O U G H T O S U R V I V E . W E C AN D O I T

O U R S E L VE S , R I GH T N O W , B Y H AR N E S S I N G T E C H N O L O G Y AN D PH AR M AC O L O G Y T O B O O S T O U R

I N T E L L I G E N C E . I S G O O G L E AC T U AL L Y M AK I N G U S S M AR T E R ?

By Jamais Cascio

IMAGE: ANASTASIA VASILAKIS

SEVENTY-FOUR THOUSAND YEARS ago, humanity nearly went extinct. A super-volcano at what’s
now Lake Toba, in Sumatra, erupted with a strength more than a thousand times that of Mount St.
Helens in 1980. Some 800 cubic kilometers of ash filled the skies of the Northern Hemisphere,
lowering global temperatures and pushing a climate already on the verge of an ice age over the edge.
Some scientists speculate that as the Earth went into a deep freeze, the population of Homo sapiens
may have dropped to as low as a few thousand families.

The Mount Toba incident, although unprecedented in magnitude, was part of a broad pattern. For a
period of 2 million years, ending with the last ice age around 10,000 B.C., the Earth experienced a
series of convulsive glacial events. This rapid-fire climate change meant that humans couldn’t rely on
consistent patterns to know which animals to hunt, which plants to gather, or even which predators
might be waiting around the corner.

How did we cope? By getting smarter. The neurophysiologist William Calvin argues persuasively that
modern human cognition—including sophisticated language and the capacity to plan ahead—evolved

1 de 8 19/05/2010 15:01
The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.com/magazine/print/2009/07/get-smarter/7548/

in response to the demands of this long age of turbulence. According to Calvin, the reason we survived
is that our brains changed to meet the challenge: we transformed the ability to target a moving animal
with a thrown rock into a capability for foresight and long-term planning. In the process, we may have
developed syntax and formal structure from our simple language.

Our present century may not be quite as perilous for the human race as an ice age in the aftermath of a
super-volcano eruption, but the next few decades will pose enormous hurdles that go beyond the
climate crisis. The end of the fossil-fuel era, the fragility of the global food web, growing population
density, and the spread of pandemics, as well as the emergence of radically transformative bio- and
nanotechnologies—each of these threatens us with broad disruption or even devastation. And as good
as our brains have become at planning ahead, we’re still biased toward looking for near-term, simple
threats. Subtle, long-term risks, particularly those involving complex, global processes, remain
devilishly hard for us to manage.

But here’s an optimistic scenario for you: if the next several decades are as bad as some of us fear they
could be, we can respond, and survive, the way our species has done time and again: by getting
smarter. But this time, we don’t have to rely solely on natural evolutionary processes to boost our
intelligence. We can do it ourselves.

Most people don’t realize that this process is already under way. In fact, it’s happening all around us,
across the full spectrum of how we understand intelligence. It’s visible in the hive mind of the Internet,
in the powerful tools for simulation and visualization that are jump-starting new scientific disciplines,
and in the development of drugs that some people (myself included) have discovered let them study
harder, focus better, and stay awake longer with full clarity. So far, these augmentations have largely
been outside of our bodies, but they’re very much part of who we are today: they’re physically separate
from us, but we and they are becoming cognitively inseparable. And advances over the next few
decades, driven by breakthroughs in genetic engineering and artificial intelligence, will make today’s
technologies seem primitive. The nascent jargon of the field describes this as “ intelligence
augmentation.” I prefer to think of it as “You+.”

Scientists refer to the 12,000 years or so since the last ice age as the Holocene epoch. It encompasses
the rise of human civilization and our co-evolution with tools and technologies that allow us to grapple
with our physical environment. But if intelligence augmentation has the kind of impact I expect, we
may soon have to start thinking of ourselves as living in an entirely new era. The focus of our
technological evolution would be less on how we manage and adapt to our physical world, and more
on how we manage and adapt to the immense amount of knowledge we’ve created. We can call it the
Nöocene epoch, from Pierre Teilhard de Chardin’s concept of the Nöosphere, a collective
consciousness created by the deepening interaction of human minds. As that epoch draws closer, the
world is becoming a very different place.

OF COURSE, WE’VE been augmenting our ability to think for millennia. When we developed written
language, we significantly increased our functional memory and our ability to share insights and
knowledge across time and space. The same thing happened with the invention of the printing press,
the telegraph, and the radio. The rise of urbanization allowed a fraction of the populace to focus on
more-cerebral tasks—a fraction that grew inexorably as more-complex economic and social practices
demanded more knowledge work, and industrial technology reduced the demand for manual labor.
And caffeine and nicotine, of course, are both classic cognitive-enhancement drugs, primitive though
they may be.

With every technological step forward, though, has come anxiety about the possibility that technology
harms our natural ability to think. These anxieties were given eloquent expression in these pages by

2 de 8 19/05/2010 15:01
The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.com/magazine/print/2009/07/get-smarter/7548/

Nicholas Carr, whose essay “Is Google Making Us Stupid?” (July/August 2008 Atlantic) argued that
the information-dense, hyperlink-rich, spastically churning Internet medium is effectively rewiring
our brains, making it harder for us to engage in deep, relaxed contemplation.

Carr’s fears about the impact of wall-to-wall connectivity on the human intellect echo cyber-theorist
Linda Stone’s description of “continuous partial attention,” the modern phenomenon of having
multiple activities and connections under way simultaneously. We’re becoming so accustomed to
interruption that we’re starting to find focusing difficult, even when we’ve achieved a bit of quiet. It’s
an induced form of ADD—a “continuous partial attention-deficit disorder,” if you will.

There’s also just more information out there—because unlike with previous information media, with
the Internet, creating material is nearly as easy as consuming it. And it’s easy to mistake more voices
for more noise. In reality, though, the proliferation of diverse voices may actually improve our overall
ability to think. In Everything Bad Is Good for You, Steven Johnson argues that the increasing
complexity and range of media we engage with have, over the past century, made us smarter, rather
than dumber, by providing a form of cognitive calisthenics. Even pulp-television shows and video
games have become extraordinarily dense with detail, filled with subtle references to broader subjects,
and more open to interactive engagement. They reward the capacity to make connections and to see
patterns—precisely the kinds of skills we need for managing an information glut.

Scientists describe these skills as our “fluid intelligence”—the ability to find meaning in confusion and
to solve new problems, independent of acquired knowledge. Fluid intelligence doesn’t look much like
the capacity to memorize and recite facts, the skills that people have traditionally associated with
brainpower. But building it up may improve the capacity to think deeply that Carr and others fear
we’re losing for good. And we shouldn’t let the stresses associated with a transition to a new era blind
us to that era’s astonishing potential. We swim in an ocean of data, accessible from nearly anywhere,
generated by billions of devices. We’re only beginning to explore what we can do with this knowledge-
at-a-touch.

Moreover, the technology-induced ADD that’s associated with this new world may be a short-term
problem. The trouble isn’t that we have too much information at our fingertips, but that our tools for
managing it are still in their infancy. Worries about “information overload” predate the rise of the Web
(Alvin Toffler coined the phrase in 1970), and many of the technologies that Carr worries about were
developed precisely to help us get some control over a flood of data and ideas. Google isn’t the
problem; it’s the beginning of a solution.

In any case, there’s no going back. The information sea isn’t going to dry up, and relying on cognitive
habits evolved and perfected in an era of limited information flow—and limited information access—is
futile. Strengthening our fluid intelligence is the only viable approach to navigating the age of constant
connectivity.

WHEN PEOPLE HEAR the phrase intelligence augmentation, they tend to envision people with
computer chips plugged into their brains, or a genetically engineered race of post-human super-
geniuses. Neither of these visions is likely to be realized, for reasons familiar to any Best Buy shopper.
In a world of ongoing technological acceleration, today’s cutting-edge brain implant would be
tomorrow’s obsolete junk—and good luck if the protocols change or you’re on the wrong side of a
“format war” (anyone want a Betamax implant?). And then there’s the question of stability: Would you
want a chip in your head made by the same folks that made your cell phone, or your PC?

Likewise, the safe modification of human genetics is still years away. And even after genetic
modification of adult neurobiology becomes possible, the science will remain in flux; our
understanding of how augmentation works, and what kinds of genetic modifications are possible,

3 de 8 19/05/2010 15:01
The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.com/magazine/print/2009/07/get-smarter/7548/

would still change rapidly. As with digital implants, the brain modification you might undergo one
week could become obsolete the next. Who would want a 2025-vintage brain when you’re competing
against hotshots with Model 2026?

Yet in one sense, the age of the cyborg and the super-genius has already arrived. It just involves
external information and communication devices instead of implants and genetic modification. The
bioethicist James Hughes of Trinity College refers to all of this as “exocortical technology,” but you can
just think of it as “stuff you already own.” Increasingly, we buttress our cognitive functions with our
computing systems, no matter that the connections are mediated by simple typing and pointing. These
tools enable our brains to do things that would once have been almost unimaginable:

• powerful simulations and massive data sets allow physicists to visualize, understand, and debate
models of an 11-dimension universe;

• real-time data from satellites, global environmental databases, and high-resolution models allow
geophysicists to recognize the subtle signs of long-term changes to the planet;

• cross-connected scheduling systems allow anyone to assemble, with a few clicks, a complex,
multimodal travel itinerary that would have taken a human travel agent days to create.

If that last example sounds prosaic, it simply reflects how embedded these kinds of augmentation have
become. Not much more than a decade ago, such a tool was outrageously impressive—and it destroyed
the travel-agent industry.

That industry won’t be the last one to go. Any occupation requiring pattern-matching and the ability to
find obscure connections will quickly morph from the domain of experts to that of ordinary people
whose intelligence has been augmented by cheap digital tools. Humans won’t be taken out of the
loop—in fact, many, many more humans will have the capacity to do something that was once limited
to a hermetic priesthood. Intelligence augmentation decreases the need for specialization and
increases participatory complexity.

As the digital systems we rely upon become faster, more sophisticated, and (with the usual hiccups)
more capable, we’re becoming more sophisticated and capable too. It’s a form of co-evolution: we
learn to adapt our thinking and expectations to these digital systems, even as the system designs
become more complex and powerful to meet more of our needs—and eventually come to adapt to us.

Consider the Twitter phenomenon, which went from nearly invisible to nearly ubiquitous (at least
among the online crowd) in early 2007. During busy periods, the user can easily be overwhelmed by
the volume of incoming messages, most of which are of only passing interest. But there is a tiny
minority of truly valuable posts. (Sometimes they have extreme value, as they did during the October
2007 wildfires in California and the November 2008 terrorist attacks in Mumbai.) At present,
however, finding the most-useful bits requires wading through messages like “My kitty sneezed!” and
“I hate this taco!”

But imagine if social tools like Twitter had a way to learn what kinds of messages you pay attention to,
and which ones you discard. Over time, the messages that you don’t really care about might start to
fade in the display, while the ones that you do want to see could get brighter. Such attention filters—or
focus assistants—are likely to become important parts of how we handle our daily lives. We’ll move
from a world of “continuous partial attention” to one we might call “continuous augmented
awareness.”

As processor power increases, tools like Twitter may be able to draw on the complex simulations and
massive data sets that have unleashed a revolution in science. They could become individualized

4 de 8 19/05/2010 15:01
The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.com/magazine/print/2009/07/get-smarter/7548/

systems that augment our capacity for planning and foresight, letting us play “what-if” with our life
choices: where to live, what to study, maybe even where to go for dinner. Initially crude and clumsy,
such a system would get better with more data and more experience; just as important, we’d get better
at asking questions. These systems, perhaps linked to the cameras and microphones in our mobile
devices, would eventually be able to pay attention to what we’re doing, and to our habits and language
quirks, and learn to interpret our sometimes ambiguous desires. With enough time and complexity,
they would be able to make useful suggestions without explicit prompting.

And such systems won’t be working for us alone. Intelligence has a strong social component; for
example, we already provide crude cooperative information-filtering for each other. In time, our
interactions through the use of such intimate technologies could dovetail with our use of collaborative
knowledge systems (such as Wikipedia), to help us not just to build better data sets, but to filter them
with greater precision. As our capacity to provide that filter gets faster and richer, it increasingly
becomes something akin to collaborative intuition—in which everyone is effectively augmenting
everyone else.

IN PHARMACOLOGY, TOO, the future is already here. One of the most prominent examples is a drug
called modafinil. Developed in the 1970s, modafinil—sold in the U.S. under the brand name Provigil
—appeared on the cultural radar in the late 1990s, when the American military began to test it for
long-haul pilots. Extended use of modafinil can keep a person awake and alert for well over 32 hours
on end, with only a full night’s sleep required to get back to a normal schedule.

While it is FDA-approved only for a few sleep disorders, like narcolepsy and sleep apnea, doctors
increasingly prescribe it to those suffering from depression, to “shift workers” fighting fatigue, and to
frequent business travelers dealing with time-zone shifts. I’m part of the latter group: like more and
more professionals, I have a prescription for modafinil in order to help me overcome jet lag when I
travel internationally. When I started taking the drug, I expected it to keep me awake; I didn’t expect it
to make me feel smarter, but that’s exactly what happened. The change was subtle but clear, once I
recognized it: within an hour of taking a standard 200-mg tablet, I was much more alert, and thinking
with considerably more clarity and focus than usual. This isn’t just a subjective conclusion. A
University of Cambridge study, published in 2003, concluded that modafinil confers a measurable
cognitive-enhancement effect across a variety of mental tasks, including pattern recognition and
spatial planning, and sharpens focus and alertness.

I’m not the only one who has taken advantage of this effect. The Silicon Valley insider webzine Tech
Crunch reported in July 2008 that some entrepreneurs now see modafinil as an important competitive
tool. The tone of the piece was judgmental, but the implication was clear: everybody’s doing it, and if
you’re not, you’re probably falling behind.

This is one way a world of intelligence augmentation emerges. Little by little, people who don’t know
about drugs like modafinil or don’t want to use them will face stiffer competition from the people who
do. From the perspective of a culture immersed in athletic doping wars, the use of such drugs may
seem like cheating. From the perspective of those who find that they’re much more productive using
this form of enhancement, it’s no more cheating than getting a faster computer or a better education.

Modafinil isn’t the only example; on college campuses, the use of ADD drugs (such as Ritalin and
Adderall) as study aids has become almost ubiquitous. But these enhancements are primitive. As the
science improves, we could see other kinds of cognitive-modification drugs that boost recall, brain
plasticity, even empathy and emotional intelligence. They would start as therapeutic treatments, but
end up being used to make us “better than normal.” Eventually, some of these may become over-the-
counter products at your local pharmacy, or in the juice and snack aisles at the supermarket. Spam

5 de 8 19/05/2010 15:01
The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.com/magazine/print/2009/07/get-smarter/7548/

e-mail would be full of offers to make your brain bigger, and your idea production more powerful.

Such a future would bear little resemblance to Brave New World or similar narcomantic nightmares;
we may fear the idea of a population kept doped and placated, but we’re more likely to see a populace
stuck in overdrive, searching out the last bits of competitive advantage, business insight, and radical
innovation. No small amount of that innovation would be directed toward inventing the next, more
powerful cognitive-enhancement technology.

This would be a different kind of nightmare, perhaps, and cause waves of moral panic and legislative
restriction. Safety would be a huge issue. But as we’ve found with athletic doping, if there’s a technique
for beating out rivals (no matter how risky), shutting it down is nearly impossible. This would be yet
another pharmacological arms race—and in this case, the competitors on one side would just keep
getting smarter.

THE MOST RADICAL form of superhuman intelligence, of course, wouldn’t be a mind augmented by
drugs or exocortical technology; it would be a mind that isn’t human at all. Here we move from the
realm of extrapolation to the realm of speculation, since solid predictions about artificial intelligence
are notoriously hard: our understanding of how the brain creates the mind remains far from good
enough to tell us how to construct a mind in a machine.

But while the concept remains controversial, I see no good argument for why a mind running on a
machine platform instead of a biological platform will forever be impossible; whether one might
appear in five years or 50 or 500, however, is uncertain. I lean toward 50, myself. That’s enough time
to develop computing hardware able to run a high-speed neural network as sophisticated as that of a
human brain, and enough time for the kids who will have grown up surrounded by virtual-world
software and household robots—that is, the people who see this stuff not as “Technology,” but as
everyday tools—to come to dominate the field.

Many proponents of developing an artificial mind are sure that such a breakthrough will be the biggest
change in human history. They believe that a machine mind would soon modify itself to get
smarter—and with its new intelligence, then figure out how to make itself smarter still. They refer to
this intelligence explosion as “the Singularity,” a term applied by the computer scientist and science-
fiction author Vernor Vinge. “Within thirty years, we will have the technological means to create
superhuman intelligence,” Vinge wrote in 1993. “Shortly after, the human era will be ended.” The
Singularity concept is a secular echo of Teilhard de Chardin’s “Omega Point,” the culmination of the
Nöosphere at the end of history. Many believers in Singularity—which one wag has dubbed “the
Rapture for nerds”—think that building the first real AI will be the last thing humans do. Some imagine
this moment with terror, others with a bit of glee.

My own suspicion is that a stand-alone artificial mind will be more a tool of narrow utility than
something especially apocalyptic. I don’t think the theory of an explosively self-improving AI is
convincing—it’s based on too many assumptions about behavior and the nature of the mind. Moreover,
AI researchers, after years of talking about this prospect, are already ultra-conscious of the risk of
runaway systems.

More important, though, is that the same advances in processor and process that would produce a
machine mind would also increase the power of our own cognitive-enhancement technologies. As
intelligence augmentation allows us to make ourselves smarter, and then smarter still, AI may turn out
to be just a sideshow: we could always be a step ahead.

SO WHAT’S LIFE like in a world of brain doping, intuition networks, and the occasional artificial
mind?

6 de 8 19/05/2010 15:01
The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.com/magazine/print/2009/07/get-smarter/7548/

Banal.

Not from our present perspective, of course. For us, now, looking a generation ahead might seem
surreal and dizzying. But remember: people living in, say, 2030 will have lived every moment from
now until then—we won’t jump into the future. For someone going from 2009 to 2030 day by day,
most of these changes wouldn’t be jarring; instead, they’d be incremental, almost overdetermined, and
the occasional surprises would quickly blend into the flow of inevitability.

By 2030, then, we’ll likely have grown accustomed to (and perhaps even complacent about) a world
where sophisticated foresight, detailed analysis and insight, and augmented awareness are
commonplace. We’ll have developed a better capacity to manage both partial attention and laser-like
focus, and be able to slip between the two with ease—perhaps by popping the right pill, or eating the
right snack. Sometimes, our augmentation assistants will handle basic interactions on our behalf; that’s
okay, though, because we’ll increasingly see those assistants as extensions of ourselves.

The amount of data we’ll have at our fingertips will be staggering, but we’ll finally have gotten over the
notion that accumulated information alone is a hallmark of intelligence. The power of all of this
knowledge will come from its ability to inform difficult decisions, and to support complex analysis.
Most professions will likely use simulation and modeling in their day-to-day work, from political
decisions to hairstyle options. In a world of augmented intelligence, we will have a far greater
appreciation of the consequences of our actions.

This doesn’t mean we’ll all come to the same conclusions. We’ll still clash with each other’s emotions,
desires, and beliefs. If anything, our arguments will be more intense, buttressed not just by strongly
held opinions but by intricate reasoning. People in 2030 will look back aghast at how ridiculously
unsubtle the political and cultural disputes of our present were, just as we might today snicker at
simplistic advertising from a generation ago.

Conversely, the debates of the 2030s would be remarkable for us to behold. Nuance and multiple
layers will characterize even casual disputes; our digital assistants will be there to catch any references
we might miss. And all of this will be everyday, banal reality. Today, it sounds mind-boggling; by then,
it won’t even merit comment.

What happens if such a complex system collapses? Disaster, of course. But don’t forget that we already
depend upon enormously complex systems that we no longer even think of as technological.
Urbanization, agriculture, and trade were at one time huge innovations. Their collapse (and all of them
are now at risk, in different ways, as we have seen in recent months) would be an even greater
catastrophe than the collapse of our growing webs of interconnected intelligence.

A less apocalyptic but more likely danger derives from the observation made by the science-fiction
author William Gibson: “The future is already here, it’s just unevenly distributed.” The rich, whether
nations or individuals, will inevitably gain access to many augmentations before anyone else. We know
from history, though, that a world of limited access wouldn’t last forever, even as the technology
improved: those who sought to impose limits would eventually face angry opponents with newer,
better systems.

Even as competition provides access to these kinds of technologies, though, development paths won’t
be identical. Some societies may be especially welcoming to biotech boosts; others may prefer to use
digital tools. Some may readily adopt collaborative approaches; others may focus on individual
enhancement. And around the world, many societies will reject the use of intelligence-enhancement
technology entirely, or adopt a cautious wait-and-see posture.

The bad news is that these divergent paths may exacerbate cultural divides created by already

7 de 8 19/05/2010 15:01
The Atlantic :: Magazine :: Get Smarter http://www.theatlantic.com/magazine/print/2009/07/get-smarter/7548/

divergent languages and beliefs. National rivalries often emphasize cultural differences, but for now
we’re all still standard human beings. What happens when different groups quite literally think in very,
very different ways?

The good news, though, is that this diversity of thought can also be a strength. Coping with the various
world-historical dangers we face will require the greatest possible insight, creativity, and innovation.
Our ability to build the future that we want—not just a future we can survive—depends on our capacity
to understand the complex relationships of the world’s systems, to take advantage of the diversity of
knowledge and experience our civilization embodies, and to fully appreciate the implications of our
choices. Such an ability is increasingly within our grasp. The Nöocene awaits.

This article available online at:

http://www.theatlantic.com/magazine/archive/2009/07/get-smarter/7548/

Copyright © 2010 by The Atlantic Monthly Group. All Rights Reserved.

8 de 8 19/05/2010 15:01

You might also like