[Infowarrior] - Is technology eating our brains?

Richard Forno rforno at infowarrior.org
Sat Feb 7 16:17:21 UTC 2009


Is technology eating our brains?

     * Peter Munro
     * February 7, 2009

http://www.watoday.com.au/national/is-technology-eating-our-brains-20090207-80gj.html?page=-1

THERE was a time when technology sought to save us from daily  
drudgery. Labour-saving devices such as automatic washing machines,  
dishwashers, the drive-through carwash and electric drill made lives  
easier by saving us from sweating out mundane tasks. Machines made us  
free to waste as much time as we pleased, and we did.

A classic advertisement in the early 1900s for a hand-operated washer  
boasted that it could "transform Blue Monday into a bright and happy  
day" — saving women (and it was always women) time, labour, nerves and  
strength. Today's technologies, though, seem dedicated to a pursuit  
higher than happiness, even. Google can connect us to a source — any  
source — within a fraction of a second, while mobile phones mean the  
world is more portable, accessible and simultaneously more demanding.  
But in the course of making our lives more convenient, have these  
technologies also made us more stupid?

Modern marvels are less labour-savers than brain-savers. Mobile phones  
remember your partner's number, your parents' and even your own — so  
you don't have to. Technology is equally adept at recalling birthdays  
and anniversaries of relatives and close friends. You don't need to  
think about the path to their homes, because Google or GPS does it for  
you. Take a taxi in Melbourne and you soon discover that navigation,  
that most adventurous of learned human skills, has been outsourced to  
a console on the dashboard.

Arguably, these are piddling concerns. Why bother the brain with dross  
when technology can pick up the slack? But deeper thought, too, seems  
to be skipping away in a ready stream of information. Some argue our  
unique capacity for original thought, innovation and imagination is  
being stultified by the spread of new technology.

Author Nicholas Carr, writing in The Atlantic last year, worried  
someone, or something, had tinkered with his brain, remapping the  
circuits and reprogramming his memory. The influence of the internet  
meant he was not thinking the way he used to. Once he could immerse  
himself in a book and spend hours strolling through prose. "Now my  
concentration often starts to drift after two or three pages. I get  
fidgety, lose the thread, begin looking for something else to do. I  
feel as if I'm always dragging my wayward brain back to the text."

"Is Google making us stupid?" he asked. But the answer was already  
staring at him through the computer screen. "What the net seems to be  
doing is chipping away my capacity for concentration and  
contemplation," he wrote. "My mind now expects to take in information  
the way the net distributes it: in a swiftly moving stream of  
particles. Once I was a scuba diver in the sea of words. Now I zip  
along the surface like a guy on a jet-ski."

I skim, therefore I am. Robert Fitzgerald, associate dean in the  
faculty of education at the University of Canberra, says there is  
indeed a "dumb side" to technology. "My children are immensely good at  
jumping on Google and finding things, but I wonder to what extent  
these are productive searches and to what extent they are hit-and- 
miss," he says.

American media critic Neil Postman once asked if we had known the  
impact the motor vehicle would have on life, would we have embraced it  
so thoroughly. Fitzgerald says it's time we asked the same question of  
computers. "If you look at very early computer applications,  
particularly in the area of education, they were about simple  
cognitive skills such as addition, subtraction and memory devices.  
There was a sense of relieving us from some of the more simple but  
tedious tasks of intellectual function.

"But now we need to recognise some of those routine, tedious tasks are  
quite fundamental to higher-level tasks. Having calculators in schools  
certainly allows children to calculate more quickly, but if they don't  
have an understanding of the equation, if they don't have the capacity  
to establish the answer, then they're at the mercy of technology. If  
it is faulty, they will never know the answer is wrong."

Indeed, Google was proved fallible only last weekend, when a system  
error meant links to all search results were flagged with the warning:  
"This site may harm your computer." Tellingly, the internet behemoth  
initially tried to blame the mishap on human error — but not its own.

If not making us stupid, as such, Google seems to be making us  
intellectually lazy. Its search engine attracts several hundred  
million queries a day, but relatively few users venture beyond the  
first page of results. It is enough to take what comes first and  
fastest, scan through an article and move on. If you slow down while  
skimming across the water, you sink.

American psychologist Maryanne Wolf, author of Proust and the Squid:  
The Story and Science of the Reading Brain, argues we are becoming  
"mere decoders of information" obtained online, rather than  
interpreters. Technology might lead us two ways, she says. Children  
might become so accustomed to immediate, on-screen information they  
fail to probe for deeper levels of insight, imagination and knowledge.  
Or the need to multitask and prioritise vast pools of information  
could see them develop equally, if not more valuable, skills.

Stephanie Trigg, professor of English literature at the University of  
Melbourne, says technology has helped her students become more adept  
at finding and extracting information for study. "I think technology  
is making us more savvy at working out what we need from various  
websites, but the downside is it's starting to affect students'  
capacity to read long works of fiction. You have to train yourself to  
read at different speeds for different purposes," she says. "But I  
don't think their mental faculties are affected by the constant  
temptation to check their mobile phones. I don't think technology is  
making us stupid; maybe it's producing a different form of attention  
and concentration where you become more clever at working out what you  
need and reading between the lines. You get better and faster at  
processing information."

A study by Dublin's Trinity College in 2007 found a quarter of Britons  
don't know their home phone number, while only a third can recall more  
than three birthdays of their immediate family. Six out of 10 claimed  
they suffered "information overload", but more than half admitted they  
used the same password across different bank accounts. Recall was  
worse among the younger set. Only 40 per cent of those aged under 30  
could remember the birthdays of close relatives, against 87 per cent  
of those aged over 50.

Of course, this doesn't denote stupidity. We now need to have these  
numbers and dates committed to memory as much as we need to know, in  
the developed world at least, how to use a hand-operated washer. We  
have outsourced parts of our memory, letting the machines do the  
thinking for us. And some argue releasing our brains of such small fry  
might free us to ponder weightier matters.

Professor Sue Trinidad, dean of teaching and learning at Curtin  
University of Technology, says technologies such as computer games are  
preparing children for success in the 21st century. "Digital natives"  
are developing special skills to sift through information quickly and  
use scanning to effectively pick out what's important to them, she  
writes by email. "These digital natives are in a 3D, multifunctional,  
fast, instant, mobile world where you want it now and get it now."

In this world, spending time or grey matter memorising phone numbers  
and birthdates might be more hindrance than help. But Nicholas Carr,  
for one, argues something much more significant is being lost in the  
rush of technology. "I argue that what the net might be doing is  
rewiring the neural circuitry of our brains in a way that diminishes  
our capacity for concentration, reflection and contemplation," he  
writes on his blog, Rough Type. "The net is sapping us of a form of  
thinking — concentrated, linear, relaxed, reflective, deep — that I  
see as central to human identity."

WHAT is technology doing to our minds? Professor Christos Pantelis,  
scientific director of the Melbourne Neuropsychiatry Centre at the  
University of Melbourne, says the brain is forever changing and being  
moulded, "but whether it's rewiring itself based on technological  
advances, we don't know".

He studies changes in the structure and function of the adolescent  
brain, which is particularly malleable in those areas involved in  
problem-solving, planning and flexible thinking. "The brain is  
changing during adolescence and early adulthood in very specific ways,  
up to about the age of 25. That means there is the potential to modify  
the way the brain is maturing during this critical phase of  
development, and you might hypothesise that what we do, and how we  
interact with the world, will have a direct effect on that," he says.

"From my perspective, I would have thought technology helps to extend  
our abilities. It helps us to look at things in different ways, and in  
that regard I would have considered technological advances are  
actually a plus and assist us in all our endeavours. But you could  
also argue the other way; that training our mind to remember things is  
also a good thing and if we're not doing that so much maybe we're  
missing out somewhere. It's a good hypothesis to test: in what ways  
will the next generation exposed to these technologies see their brain  
changed by them?"

What little study exists in this area is inconclusive. Scientists at  
University College London have found people demonstrate "a form of  
skimming activity" when using the internet for research. More than  
half of e-journal users in the study, published last year, viewed no  
more than three pages of an article or book before "bouncing" to  
another site. Almost two-thirds of users never returned to a source  
they had visited. Little time was spent evaluating information for  
relevance, accuracy or authority. The researchers warned of the  
emergence of a whole new form of reading behaviour, whereby users  
"power browsed" through titles, content pages and abstracts.

But a separate study, published in the American Journal of Geriatric  
Psychiatry last year, instead suggested internet searches enhance  
brain power. Researchers at the University of California, Los Angeles,  
scanned the brains of 24 volunteers aged 55-76, half of whom had no  
prior internet search experience, while they were online or reading a  
book. During online searches, the brains of those who reported using  
the internet regularly showed increased stimulation, particularly in  
those regions associated with complex reasoning and decision-making.  
Within five days, the internet novices showed the same increase in  
activity in their frontal lobes.

The study was led by neuroscientist Gary Small, a professor at UCLA's  
Semel Institute for Neuroscience and Human Behaviour, and author of  
iBrain: Surviving the Technological Alteration of the Modern Mind. He  
argues that as the brain shifts towards, and is energised by, new  
technological skills, it simultaneously drifts away from fundamental  
social skills. The cognitive gains made by younger generations in  
adapting to and processing new technologies come at the cost of such  
age-old social skills as reading facial expressions and body language.

"Our social interactions may become awkward, and we tend to  
misinterpret, and even miss, subtle, non-verbal messages," he writes.  
"The dramatic conclusion would be we're drifting into an autistic  
society, but that would be overshooting."

To where, then, might we be drifting? The dark imaginings of science  
fiction may offer some guide. Nicholas Carr cites Stanley Kubrick's  
film 2001: A Space Odyssey, and that scene where astronaut Dave Bowman  
coolly disconnects the memory circuits that control the artificial  
"brain" of malfunctioning supercomputer HAL. The humans act with  
almost "robot-like efficiency", while it is the machine that expresses  
anguish and loss. "That's the essence of Kubrick's dark prophecy: as  
we come to rely on computers to mediate our understanding of the  
world, it is our own intelligence that flattens into artificial  
intelligence," Carr argues.

Robert Fitzgerald, from the University of Canberra, instead alludes to  
Ridley Scott's Alien trilogy. In a scene in the sequel, Lieutenant  
Ellen Ripley dons a mechanical "exosuit" to fight her alien foe —  
spitting out that memorable line, "Get away from her, you bitch!"

"For me, that exosuit is sort of symbolic for the way technology can  
expand our human capacities," Fitzgerald says.

"But I suspect what we've got at the moment are very small fragments  
of that exosuit, with nothing really fully functioning or connected  
yet. We're really in the very early days in terms of the development  
of new internet technologies. While we have seen quite remarkable  
developments in the rates of blog use or wikis, I suspect five years  
down the track we will not recognise those technologies we're  
currently using — they'll be more intuitive, more integrated, more  
intelligent."

But will we be more intelligent as well?

Our intelligence ultimately might reveal itself in the smarts of those  
same technologies, which have the capacity either to increase the sum  
of deep intelligence or leave us skating on the surface. But here's a  
sobering thought: if the key to human intelligence lies beyond the  
first page of a Google search, or in the last paragraph of a lengthy  
newspaper article, will we ever find it?


More information about the Infowarrior mailing list