Tumgik
#kathleen i lyons
Text
For those of your who are on Facebook, I've made a page for my debut novel, Arboriah Lux.
More books will follow!!
Arboriah Lux will soon be available for pre-order and will be available for purchase on April 26th!
6 notes · View notes
queerlilchinchin · 5 months
Text
WiYS Interest Form
Yet again, I have an idea that's blowing up way faster than people realizing that I have created a form for interested parties.
So below is the interest form for my Write in Your Style streams that will begin at the beginning of next month (December 2023).
I'll be tagging anyone that showed interest so they can see the form but you won't be tagged in any more posts about this unless you fill out the form :) so make sure if you're interested, you fill out the form! It also includes the ability to say you're casually interested but don't want to be tagged. For those people, I just want to see where you'd want to watch the stream mostly. :) Kind of gauge where my audience would be. So make sure you fill it out, even if you don't want to be tagged anymore after this!
Tagging new interested peoples:
@mrblazeflappybird @arijensineink @strugglequill @squintclover @beautitudes @theresthesnitch @schmem14 @hihimissamericanbi @calligraphic-tac @elizababie @sunflowerdrake @werewolf-midoriya @senti-naidar @kathleen-lyons-the-writer @pevensiechase @marmeegle @thefleurographist @in-flvx @creative-girl @soopsiedaisies
13 notes · View notes
arijensineink · 5 months
Text
Last Line Tag
Thankee for the tag @sparrow-orion-writes
Since I actually wrote something fresh yesterday, I wanted to post. I'll cheat a little bit and include the last 2 sentences I wrote.
Mezereon stood, stepped over, and Seujinne forced himself not to flinch as she came to stand close to the bars.  “Well, you had time today,” she said, not without contempt. 
I'll tag @kathleen-lyons-the-writer @thequietsoliloquy @words-after-midnight or anyone who wants to do it!
No pressure, of course <3
7 notes · View notes
Photo
Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media Tumblr media
June Monthly Recap:
Once again, I read a lot in June, 18 books! Slightly more than I did in May, although it felt a lot less, mostly because I ended up DNF’ing a lot (5 different books!) But I still read some excellent books, including Where the Drowned Girls Go and Unraveled. I also made some progress on finishing series that have been hanging around on my TBR for a long time! I’m all caught up on the Wayward Children series, and I finally finished Phedre’s trilogy after 5(!!) years.
Across the Green Grass Fields by Seanan McGuire: 4/5
Unraveled by Courtney Milan: 5/5
New Worlds, Year Two by Marie Brennan: 4/5
Before the Coffee Gets Cold by Toshikazu Kawaguchi: 3.5/5
This Mortal Coil by Emily Suvada: 3.75/5
Hexarchate Stories by Yoon Ha Lee: 5/5
The Prince of Midnight by Laura Kinsale: 2.5/5, dnf
The Goblin Emperor by Katherine Addison: 5/5, re-read
Sorcery of Thorns by Margaret Rogerson: 4.5/5
Paladin’s Strength by T. Kingfisher: 5/5, re-read
The Ruin of Kings by Jenn Lyons: 2/5, dnf
All the Feels by Olivia Dade: 4.25/5
Where the Drowned Girls Go by Seanan McGuire: 5/5
So We Meet Again by Suzanne Park: 2/5, dnf
After Perfect by Maan Gabriel: 1.5/5, dnf
A Lady’s Guide to Ruin by Kathleen Kimmel: 4.5/5
Kushiel’s Avatar by Jacqueline Carey: 4.25/5
Goliath by Tochi Onyebuchi: 3/5, dnf
Edit: even more goals
22 in 2022: 9
Read 100 Books: 82
Read 40% AOC: 38%
Completing Series: 13 series completed/caught up vs. 10 started
Translated Works: 2
Books in Spanish: 0
Numbered TBR: 7
Discworld: 0
Books by an Indigenous Author: 1
Physical TBR: 10
Storygraph Recs: 3
11 notes · View notes
In case you missed the back storey involving all of Bill Clinton’s “sexual misconduct”.
In case you missed the back storey involving all of Bill Clinton’s “sexual misconduct”. It would be nice if the left was consistent, and this is one reason they can’t be taken seriously, we only believe “all women” when accusations are made about conservatives. Where is the #MeToo movement.
My whole way of thinking changed during the Duke Lacrosse shenanigans. I believe now each allegation should be taken seriously and left to a jury to decide, not the public.
"Every survivor of sexual assault deserves to be heard, believed, and supported," – Hillary Clinton
Direct Quotes:
Clinton has faced multiple allegations of sexual assault and harassment, most famously his affair with Monica Lewinsky — which, while consensual in some sense, was nonetheless textbook sexual harassment of a subordinate of a kind that would (or perhaps more accurately, should) get many CEOs fired from their companies
Paula Jones, a former Arkansas state employee, sued Clinton during his presidency for allegedly exposing himself to her when he was governor in 1991. Kathleen Willey claims that Clinton fondled her breast and forced her hand on his crotch in the Oval Office in 1993, when she was a White House volunteer.
Most seriously of all, Juanita Broaddrick claims that Clinton raped her during his 1978 campaign for Arkansas governor.
the Juanita Broaddrick case is the hardest one for admirers of Bill Clinton. Her allegation has never been definitively refuted. Only she and Bill Clinton know what the truth of the matter in the case is. But if one generally believes it's important to believe the victim, it's hard to argue that this case should be an exception.
Juanita Broaddrick gave a lengthy account of her alleged rape in a 1999 Dateline NBC interview
The interview was conducted on January 20, 1999, before the Senate ultimately acquitted Clinton on charges related to his affair with Monica Lewinsky on February 12.
In 1978, Broaddrick was volunteering for Clinton's gubernatorial campaign
they spent only a few minutes chatting by the window -- Clinton pointed to an old jail he wanted to renovate if he became governor -- before he began kissing her. She resisted his advances, she said, but soon he pulled her back onto the bed and forcibly had sex with her. She said she did not scream because everything happened so quickly. Her upper lip was bruised and swollen after the encounter because, she said, he had grabbed onto it with his mouth.
"The last thing he said to me was, 'You better get some ice for that.' And he put on his sunglasses and walked out the door," she recalled.
Several friends of Broaddrick's backed up the story.
"[Hillary] came directly to me as soon as she hit the door. I had been there only a few minutes, I only wanted to make an appearance and leave. She caught me and took my hand and said 'I am so happy to meet you. I want you to know that we appreciate everything you do for Bill.' I started to turn away and she held onto my hand and reiterated her phrase -- looking less friendly and repeated her statement — 'Everything you do for Bill'. I said nothing. She wasn't letting me get away until she made her point. She talked low, the smile faded on the second thank you. I just released her hand from mine and left the gathering." Blumenthal also cites reporting from Conason and Lyons suggesting that Broaddrick had asked the Van Buren Press Argus-Courier, her local paper, to photograph her nursing home in 1990 upon the visit of Gov. Bill Clinton. "This was hardly the attitude of a rape victim toward her predator," Blumenthal writes. This assertion, that a "true" rape victim would cut off all contact with their rapist, is rather misleading and pernicious, and maintaining contact with an alleged assailant is hardly proof that a victim is lying. "It is common for victims to maintain contact with their abusers because they may still feel affection for them even though they hate the abuse," according to the Pennsylvania Coalition Against Rape. "It is also common for some victims to maintain contact in an attempt to regain control over their assault. Others may maintain contact in an attempt to regain a feeling of normalcy."
Given the prevailing view among many progressives — including Hillary Clinton — that one should default to believing rape accusers, the Broaddrick allegation thus poses a problem. Michelle Goldberg, then at Slate, explained the conundrum well in a piece from December 2015:
Today, feminists have repeatedly and convincingly made the case that when women say they’ve been sexually assaulted, we should assume they’re telling the truth. Particularly when it comes to Broaddrick, it’s not easy to square the arguments against believing her with the dominant progressive consensus on trusting victims.
But today, few feminists would find her shifting story disqualifying.
Speaking to Klein, she says she was traumatized and blamed herself for what happened. "I felt responsible. I don’t know if you know the mentality of women and men at that time. But me letting him come to my room? I accepted full blame." In any other context, most feminists today would find this credible. After all, many were outraged when rape skeptics tried to discredit Columbia student Emma Sulkowicz because she’d sent friendly Facebook messages to her alleged rapist after the alleged rape.
There is a crucial tension between "believe
survivors" and the "Juanita Broaddrick is lying" position of some Clinton defenders, lacking further information. When Hillary Clinton tweeted during the campaign that "Every survivor of sexual assault deserves to be heard, believed, and supported," it’s reasonable to ask if that’s true of Juanita Broaddrick, too.
0 notes
Note
Hey who do u take request for
From casualty
Linda Andrews
Lloyd Asike
Tamzin Bayle
Alice Chantry
Kirsty Clements
Jeff Collier
Scarlett Conway
Curtis Cooper
John 'Abs' Denham
Kathleen 'Dixie' Dixon
Madiha 'Mads'Durrani
Polly Emmerson
Jay Faldren
Greg Fallon
Adrian 'Fletch' Fletcher
Zoe Hanna
Ethan Hardy
Jessica Harrison
Lenny Lyons
Tom Kent
Sam Nicholls
May Phelps
Kelsey Phillips
Yuki Reid
Adam Trueman
Max Walker
I think that's all of them but I may have forgotten someone
1 note · View note
hiddenwashington · 1 year
Text
anonymous said : mw from romcoms?
thank you to everyone who suggested films or characters in the mw on our ooc discord the other day! 
for most wanted from romcoms, we would love to see: 
jenna rink, matt flamhaff, and lucy wyman from 13 going on 30; anna scott and will thacker from notting hill; rachel chu, nick young, and astrid leong-tao from crazy rich asians; kathleen kelly and joe fox from you’ve got mail; kat stratford, bianca stratford, patrick verona, and cameron james from 10 things i hate about you; cher horowitz, tai frasier, and dionne davenport from clueless; laura jean, peter, margot, or kitty from to all the boys i’ve loved before; elle evans, lee flynn, noah flynn, and rachel from the kissing booth; margaret tate and andrew paxton from the proposal; harper moore, charlie young, and kirsten stevens from set it up; sasha tran and marcus kim from always be my maybe; loretta sage, alan caprison, abigail fairfax, and jack trainer from the lost city; phil, cate, and jexi from jexi; jung jiwoo and jung jihoo from love & leashes; gabriela diaz and jake taylor from falling inn love; jibran and leilani from the lovebirds; or ellie chu, paul munsky, and aster flores from the half of it.
we also have had a couple of requests for characters from holiday romcoms: amber moore, prince richard, melissa, and count simon from a christmas prince; kate andrich, tom webster, and santa from last christmas; abby sutton, josh barton, and ty walker from the holiday calendar; julie reyes, stuart bale, dorrie, kerry, and addie from let it snow; and brooke winters, cole lyons, and geoffrey lyons from the knight before christmas!
Tumblr media
0 notes
xtremetrust · 2 years
Text
Time manager diary 2015
Tumblr media
#Time manager diary 2015 full
#Time manager diary 2015 plus
#Time manager diary 2015 tv
If the day feels like an accelerating whirl of genres, it’s only to prepare us for the stylistic head-spin that is Gorillaz. IDLES’ politics – celebrating immigration, challenging gender norms and terrifying Tories – melt naturally out from the chaos, where the tumbling punk of ‘Never Fight A Man With A Perm’ lands in a heap of bone-crushing chords and ‘Crawl!’ sounds like a pop song screaming to be let out of punk’s basement and allowed to keep what remains of its skin. In a pretty sailor-style dress, he stares down the camera from the organ stool during the spooky slow-dance of ‘The Beachland Ballroom’, unleashes fire-in-Heaven atmospherics upon ‘A Hymn’ and rolls into the crowd to get them to sing the name-spelling crescendo of a fantastic ‘Danny Nedelko’. As much as singer Joe Talbot indulges in knees-up jogs in circles and bandy-legged flamenco dancing – and who is in charge of delivering lyrics calling cocaine “Charlie Sheen” and imagining “Kathleen Hanna with bear claws grabbing Trump by the pussy” – it’s guitarist Mark Bowen who realty draws the eye. IDLES see Turnstile’s energised eclecticism and raise it a load of cross-dressed crowd-surfing. Later, ‘NO SURPRISE’ exposes bassist ‘Freaky’ Franz Lyons as a fine singer of future R&B before it’s crushed beneath the wheels of ‘Come Back For More’, a belated audition to replace Rage Against The Machine at Reading & Leeds. Singer Brendan Yates throws himself around in all manner of aerobic moves as ‘UNDERWATER BOI’ itself contorts from beachside psych pop to the sound of a hard rock tsunami hitting.
#Time manager diary 2015 tv
On the BBC Radio 6Music Stage, Nia Archives – who picked up Best Producer at the BandLab NME Awards 2022 – bounces and sings along to her new wave jungle beats (sporadically in French) while, on the main stage, Turnstile come on like a crazed mash-up of funk rock, thrash metal, Mars Volta freakouts and breakfast TV workout sessions. Whoever programmed the day certainly embraced the all-encompassing nature of streaming-gen tastes. Still, he relishes his verses of Kanye West tracks such as ‘New God Flow’, ‘Mercy’ and ‘Runaway’ and his music deserves far more credit as the quintessential concoction of rap, twisted soul and East Coast synth work, which he has honed – oh yes – to a T. Besides a few such heartfelt moments, though, T largely seems concerned with the inflationary crisis on the cost of his drug supply, although what he’s pushing hardest today is his new album and tour. “ I’m dealin’ with heartbreak, checkin’ my ego / I’m livin’ with lost faith,” he confesses to the Heavens on ‘Santeria’, addressing the spirit of his late tour manager De’Von Pickett, who was killed in 2015. Over on the main East stage, fellow former NME cover star Pusha T is indulging in his own form of introspection. Most moving is the open-diary soul ballad ‘I Do This All The Time’, wherein Taylor sets herself all manner of self-help tasks and briefly, at the song’s abrupt end, glimpses light at the end of the emotional tunnel. ‘Moody’ concerns “me being a moody bitch”, ‘How Can I Help You’ concludes its supportive tribal-pop intervention with the admission that none of us “ know shit”, and ‘The 345’ bluntly advocates living for the moment because “ it all ends mid-paragraph”.
#Time manager diary 2015 plus
Beneath a backdrop demanding “MALE PILL WHEN?”, and with a deceptively cheery demeanour, she blasts through 40 minutes of forthright and confrontational electronic crank-pop – plus moments of gospel balladry and catchy tropicalia – in unison with her trio of backing singer/dancers.
#Time manager diary 2015 full
Over on the Victoria Park West stage mid-afternoon, Self Esteem’s set is scuppered by several sound desk meltdowns, but she powers on regardless, taking to the stage for a third time singing “ that was not my fuck up” and insisting she plays her full set.Īnd Rebecca Taylor is not a force to be fucked with. 60,000 Londoners somehow find their way to Hackney for the first of 2022’s 10 All Points East shows with next to no help from the transport system and, as a result, no-one’s letting the experience get away. Here, he insists, dwell “the best fucking people in the world”, thanks to one notable cosmopolitan trait: “everybody’s got rhythm!” “Lon-don! Lon-don!” chants Damon Albarn, stopping off midway through a global Gorillaz jaunt to play his “most local gig to where I was born” ever and punch the air in celebration of the City That Never Underpays.
Tumblr media
0 notes
probablespluyt · 2 years
Text
Table of Contents
Praise
Dedication
Title Page
Contents
Acknowledgments
1. What to Do Next
2. Evolving a Good Guess
3. The Janitor’s Dream
4. Evolving Intelligent Animals
5. Syntax as a Foundation of Intelligence
6. Evolution on-the-Fly
7. Shaping Up an Intelligent Act from Humble Origins
8. Prospects for a Superhuman Intelligence
Recommended Reading
Notes
Index
About the Author
Copyright
The Science Masters Series
‘These books admirably achieve what the series sets out to do – to explain exciting science clearly’
Sunday Times
‘The whole series . . . manages to tell powerful stories very simply and very cheaply. What more could you want?’
Science Now
‘The Science Masters series looks set to play a major role in the responsible popularisation of science’
New Scientist
How Brains Think
‘In How Brains Think Mr Calvin takes us on an exhilarating intellectual journey . . . an exquisite distillation of his key ideas. He’s a member of that rare breed of scientists who can translate the arcana of their fields into lay language, and he’s one of the best . . . Mr Calvin, so lyrical and imaginative in his presentation, draws you into his world of neural Darwinism and inspires you to read more’
New York Times
‘Calvin is fizzing with ideas and this is a provocative, stimulating book’
Sunday Times
Dedicated to my late futurist friend, Thomas F. Mandel (1946–1995), whose memes live on.
HOW BRAINS THINK
Evolving Intelligence, Then and Now
WILLIAM H. CALVIN
CONTENTS
Cover
Praise
Dedication
Title Page
Acknowledgments
1 What to Do Next
2 Evolving a Good Guess
3 The Janitor’s Dream
4 Evolving Intelligent Animals
5 Syntax as a Foundation of Intelligence
6 Evolution on-the-Fly
7 Shaping Up an Intelligent Act from Humble Origins
8 Prospects for a Superhuman Intelligence
Recommended Reading
Notes
Index
About the Author
Copyright
ACKNOWLEDGMENTS
Helpful discussions with Derek Bickerton, Iain Davidson, Daniel C. Dennett, Stephen Jay Gould, Katherine Graubard (who suggested the book’s title), Marcel Kinsbourne, Elizabeth Loftus, Jennifer Lund, Don Michael, George Ojemann, Duanne Rumbaugh, Sue Savage-Rumbaugh, Mark Sullivan, and the late Jan Wind are reflected at multiple places in this book. Bonnie Hurren kindly pointed me to the Piagetian definition of intelligence.
The editors at Scientific American, John Rennie, Jonathan Piel, and Michelle Press, were very helpful (a short version of my intelligence argument appeared in their ‘Life in the Universe’ special issue of October 1994; well-tuned paragraphs from it are scattered throughout this book), as was Howard Rheingold at Whole Earth Review (the last part of the last chapter appeared in their Winter 1993 issue).
Among the others I must thank for their editorial suggestions are Lynn Basa, Hoover Chan, Lena Diethelm, Dan Downs, Seymour Graubard, the late Kathleen Johnston of San Francisco, Fritz Newmeyer, Paolo Pignatelli, Doug vanderHoof, Doug Yanega, and The WELL’s writers conference.
Blanche Graubard, as usual, edited the book before it was inflicted on the publisher, and I have again profited from her good sense and style. Jeremiah Lyons and Sara Lippincott edited the book for the Science Masters series and made many excellent suggestions for revision.
CHAPTER 1
WHAT TO DO NEXT
It is perfectly true, as philosophers say, that life must be understood backwards. But they forget the other proposition, that it must be lived forwards.
Sören Kierkegaard, 1843
All organisms with complex nervous systems are faced with the moment-by-moment question that is posed by life: What shall I do next?
Sue Savage-Rumbaugh and Roger Lewin, 1994
Piaget used to say that intelligence is what you use when you don’t know what to do (an apt description of my present predicament as I attempt to write about intelligence). If you’re good at finding the one right answer to life’s multiple-choice questions, you’re smart. But there’s more to being intelligent – a creative aspect, whereby you invent something new ‘on the fly.’ Indeed, various answers occur to your brain, some better than others.
Every time we contemplate the leftovers in the refrigerator, trying to figure out what else needs to be fetched from the grocery store before fixing dinner, we’re exercising an aspect of intelligence not seen in even the smartest ape. The best chefs surprise us with interesting combinations of ingredients, things we would ordinarily never think ‘went together.’ Poets are particularly good at arranging words in ways that overwhelm us with intense meaning. Yet we’re all constructing brand-new utterances hundreds of times every day, recombining words and gestures to get across a novel message. Whenever you set out to speak a sentence that you’ve never spoken before, you have the same creativity problem as the chefs and poets – furthermore, you do all your trial and error inside your brain, in the last second before speaking aloud.
We’ve lately made a lot of progress in locating some aspects of semantics in the brain. Frequently we find verbs in the frontal lobe. Proper names, for some reason, seem to prefer the temporal lobe (its front end; color and tool concepts tend to be found toward the rear of the left temporal lobe). But intelligence is a process, not a place. It’s about improvization, where the ‘sweet spot’ is a moving target. It’s a way, involving many brain regions, by which we grope for new meanings, often ‘consciously.’
The more experienced writers about intelligence, such as IQ researchers, steer clear of the C word. Many of my fellow neuroscientists avoid consciousness as well (some physicists, alas, have been all too happy to fill the vacuum with beginner’s mistakes). Some clinicians unintentionally trivialize consciousness by redefining it as mere arousability (though to talk of the brain stem as the seat of consciousness is to thereby confuse the light switch with the light!). Or we redefine consciousness as mere awareness, or the ‘searchlight’ of selective attention.
They’re all useful lines of inquiry but they leave out that activism of your mental life by which you create – and edit and re-create – yourself. Your intelligent mental life is a fluctuating view of your inner and outer worlds. It’s partly under your control, partly hidden from your introspection, even capricious (every night, during your four or five episodes of dreaming sleep, it is almost totally out of control). This book tries to fathom how this inner life evolves from one second to the next, as you steer yourself from one topic to another, as you create and reject alternatives. It draws from studies of intelligence by psychologists, but even more from ethology, evolutionary biology, linguistics, and the neurosciences.
There used to be some good reasons for avoiding a comprehensive discussion of consciousness and the intellect. A good tactic in science, especially when mechanistic-level explanations don’t help structure your approach to a fuzzy subject, is to fragment the problem into bite-sized pieces – and that is, in some sense, what’s been going on.
A second reason was to avoid trouble by camouflaging the real issues to all but insiders (maintaining deniability, in the modern idiom). Whenever I see words that have everyday meanings but also far more specific connotations used only by insider groups, I am reminded of code names. Several centuries ago, an uncamouflaged mechanistic analogy to mind could get you into big trouble, even in relatively tolerant Western Europe. Admittedly, Julien Offroy de La Mettrie didn’t merely say the wrong thing in casual conversation: this French physician (1709–1751) published a pamphlet in which he wrote of human motivations as if they were analogous to energy-releasing springs inside machines.
That was in 1747; the year before, La Mettrie had fled to Amsterdam from France. He had written a book, it seems, entitled The Natural History of the Soul. The Paris Parliament disliked it to the point of ordering all copies burned.
This time, La Mettrie took the precaution of publishing his pamphlet, entitled Man à Machine, anonymously. The Dutch, considered the most tolerant people in Europe, were scandalized and tried with a vengeance to discover who the pamphlet’s author was. They nearly found out, and so La Mettrie was forced to flee once more – this time to Berlin, where he died four years later, at the age of forty-two.
Though he was clearly ahead of his time, La Mettrie didn’t invent the machine metaphor. That’s usually ascribed to René Descartes (1596–1650), writing a century earlier, in his De Homine. He too had moved to Amsterdam from his native France, at about the same time that Galileo was getting into trouble with the Vatican over the scientific method itself. Descartes didn’t have to flee Holland, as did La Mettrie; he took the precaution, one might say, of publishing his book a dozen years after he was safely dead.
Descartes and his followers weren’t trying to banish all talk of spirits; indeed, one of their characteristic concerns was to identify exactly where in the brain lay the ‘seat of the soul.’ This endeavor was a continuation of a scholastic tradition that focused on the big reservoirs of cerebrospinal fluid inside the brain called the ventricles. Religious scholars of five hundred years ago thought that the subdivisions of the soul were housed in these cavities: memory in one; fantasy, common sense, and imagination in another; rational thought and judgment in a third. Like the bottle with the genie inside, the ventricles were supposedly containers for spirits. Descartes thought that the pineal gland was a better locale for the seat of government, on the grounds that it was one of the few brain structures that didn’t come in pairs.
Here at the fin de millennium, though there are theocratic countries where using code words would still be a good idea, we are generally more at ease when it comes to machine metaphors for mind. We can even discuss principled grounds for disputing any analogy of mind to machine. Minds, the argument goes, are creative and unpredictable; the machines we know are unimaginative but reliable – so machines such as digital computers initially seem like an unreasonable analogy.
Fair enough. But what Descartes established was that it was useful to talk of the brain as if it were a machine. You tend to make progress that way, peeling away the layers of the onion. Even if there is ‘something else’ hidden beneath the obscuring layers, the scientist tentatively assumes that there isn’t anything fundamentally unknowable, in order to test the alternative explanations. This scientific tactic – not to be confused with a scientific conclusion – has produced a revolution in how we see ourselves.
Mechanistic approaches to mind were, for a long time, missing an essential ingredient: a bootstrap mechanism. We’re used to the idea that a fancy artifact such as a watch requires an even fancier watch designer. It’s common sense – just as Aristotle’s physics still is (despite being wrong).
But, ever since Darwin, we’ve known that fancy things can also emerge (indeed, self-organize) from simpler beginnings. Even highly educated people, as the philosopher Daniel Dennett notes in the preface to Darwin’s Dangerous Idea, can be uncomfortable with such bootstrapping notions:
Darwin’s theory of evolution by natural selection has always fascinated me, but over the years I have found a surprising variety of thinkers who cannot conceal their discomfort with his great idea, ranging from nagging skepticism to outright hostility. I have found not just lay people and religious thinkers, but secular philosophers, psychologists, physicists, and even biologists who would prefer, it seems, that Darwin were wrong.
But not all. Only a dozen years after the 1859 publication of On the Origin of Species, the psychologist William James was writing letters to friends about his notion that thought involved a darwinian process in the mind. More than a century later, we are only beginning to flesh out this idea with appropriate brain mechanisms for darwinism. For several decades, we have been talking about the selective survival of overproduced synapses. And that’s only the cardboard version of darwinism, analogous to carving a pattern into a wood block. Now we’re also seeing brain wiring that could operate the full-fledged darwinian process, and probably on the milliseconds-to-minutes timescale of consciousness.
This shaping-up-the-improbable version of darwinism involves generating lots of copies of certain cerebral firing patterns, letting the copies vary somewhat, and then letting those variants compete for dominance over a work space (rather as those variants called bluegrass and crabgrass compete for my backyard). The competition is biased by how well those spatiotemporal firing patterns resonate with the ‘bumps and ruts in the road’ – the memorized patterns stored in the synaptic strengths. Such Darwin Machines are a favorite topic of mine, as you will see, but let us first get some idea of what intelligence is – and isn’t.
A useful tactic for exploring intelligence, one that avoids premature definitions, is the journalist’s who-what-where-when-why-how checklist. I’ll start with what constitutes intelligence and when intelligence is needed, simply because the term is used in so many ways that it is easy to talk at cross-purposes, just as in the case of consciousness. Narrowing intelligence down a little, without throwing out the baby with the bathwater, is the task of the next chapter, after which I’ll tackle levels of explanation and the ‘consciousness’ confusions.
A little ice-age perspective turns out to be important when exploring the evolutionary why aspects of intelligence, particularly in discussing our hominid ancestors. Alaska’s coastline is the best place to see the ice age still in action; Glacier Bay, some fifty miles long, was totally filled with ice only two hundred years ago. Now it’s populated with enough harbor seals, kayaks, and cruise ships to cause traffic jams. In the context of Glacier Bay, I’ll raise the question of how jack-of-all-trades abilities could possibly evolve, when efficiency arguments tell us that a streamlined specialist (the lean, mean machine beloved of economists) always does better in any one climate. The short answer? Just keep changing the climate, abruptly and unpredictably, so that efficiency doesn’t remain the name of the game.
In the fifth chapter, I’ll discuss the mental machinery needed for parsing sentences complicated enough to require syntax. Many observers, myself included, suspect that the big boost in intelligence during hominid evolution was provided by those logical structures needed for a grammatical language (and also useful for other tasks). Chimpanzees and bonobos (these ‘chimpanzees of the pygmies’ are a distinctly different ape, now called by the name that the natives were once said to use) provide some essential perspective for judging the role of language in intelligence and consciousness. Stones and bones are all that’s left of our actual ancestors, but our distant cousins show us what ancestral behaviors might have been like.
The sixth chapter takes up the problems of convergent and divergent thinking in the darwinian context. Small neurobiology meetings, such as one I recently attended down on Monterey Bay, certainly illustrate convergent thinking – all those specialists trying to find the one right answer, as the search for memory mechanisms narrows down. But divergent thinking is what creative people need to discover a scientific theory, or write a poem, or (at a more mundane level) dream up all those wrong answers to use in multiple-choice exams for testing convergent thinking. Whenever a neuroscientist proposes an explanation for a memory storage mechanism, questioners from the audience promptly suggest several alternative explanations – ones they’ve dreamed up on the spot with divergent thinking. So, how do we shape up a novel thought into something of quality, without the equivalent of the guiding hand that shapes up a lump of clay into a pot? The answer may be in the title of chapter 6. ‘Evolution on-the-Fly.’ The same darwinian process that shapes up a new species in millennia – or a new antibody during the several weeks of an immune response – may also shape up ideas on the timescale of thought and action.
In the penultimate chapter, I’m going to venture past the analogy of mental processes to other known darwinian processes and propose how (the mechanistic how of the physiologist) our brains can manipulate representations in such a way as to cause a copying competition, one that can be darwinian and so shape up randomness into a good guess. This descent into cerebral codes (which, like the bar codes in supermarkets, are abstract patterns that stand in for the real thing) and cerebral circuitry (particularly the circuitry of the superficial cortical layers responsible for the brain’s inter-office mail) has provided me with my best glimpse so far of mechanisms for higher intellectual function: how we can guess, speak sentences we’ve never spoken before, and even operate on a metaphorical plane. It even provides some insight into the big step up from proto-language to Universal Grammar.
This cerebral version of a Darwin Machine is what, in my opinion, will most fundamentally change our concept of what a person is. Like the Dodo in Alice in Wonderland, who said it was better to demonstrate the game than to explain it, I will walk you through the darwinian process in some detail, as it shapes up a thought and makes a decision. Trying to describe intelligence is not, I am happy to report, as difficult as describing how to ride a bicycle; still, you will understand the description a lot better if you develop a feel for the process rather than being satisfied with an abstract appreciation (what you’ll get from chapters 6 and 8, if you skip over my favorite chapter).
In the final chapter, I will come back up for air and summarize the crucial elements of higher intelligence described in earlier chapters, focusing on those mechanisms that an exotic or an artificial intelligence would require in order to operate in the range spanning clever chimps to human musical genius. I will conclude with some cautions about any transition to superhuman intelligence, those aspects of arms races that the Red Queen cautioned Alice about – why you have to keep running to stay in the same place.
[One doctrine] depicts man as an induction machine nudged along by external pressures, and deprived of all initiative and spontaneity. The second gives him the Spielraum [room to play] to originate ideas and try them out. Learning about the world means, on the first view, being conditioned by it; on the second view, it means adventuring within it.
J. W. N. Watkins, 1974
CHAPTER 2
EVOLVING A GOOD GUESS
While innate processing, instinctive behavior, internally orchestrated motivation and drive, and innately guided learning are all essential and important elements of an animal’s cognitive repertoire, they are not likely to be part of that more esoteric realm of mental activity that we associate with thinking, judgment, and decision making. But what is thought, and how are we to recognize its operation in other creatures within that most private of organs, the brain? What behavioral criteria can permit us to distinguish between the true thought that we are wont to believe goes into our aesthetic, moral, and practical decision making on one hand, and the intricate programming that can create the illusion of thought in at least certain other animals? Or could it be, as advocates of artificial intelligence suspect, that all thought, including ours, is just the consequence of clever programming?
James L. Gould and Carol Grant Gould, The Animal Mind, 1994
Intelligence gets framed in surprisingly narrow terms most of the time, as if it were some more-is-better number that could be assigned to a person in the manner of a batting average. It has always been measured by a varied series of glimpses of spatial abilities, verbal comprehension, word fluency, number facility, inductive reasoning, perceptual speed, deductive reasoning, rote memory, and the like. In recent decades, there has been a tendency to talk about these various subtests as ‘multiple intelligences.’ Indeed, why conflate these abilities by trying to boil intelligence down to a single number?
The short answer is that the single number seems to tell us something additional – while hazardous when overgeneralized, it’s an interesting bit of information. Here’s why: Doing well on one kind of intelligence subtest never predicts that you’ll do poorly on another; one ability never seems to be at the expense of another. On the other hand, an individual who does well on one such test will often perform better than average on the other subtests.
It’s as if there were some common factor at work, such as test-taking ability. The so-called ‘general factor g’ expresses this interesting correlation between subtests. The psychologist Arthur Jensen likes to point out that the two strongest influences on g are speed (such as how many questions you can answer in a fixed amount of time) and the number of items you can mentally juggle at the same time. Analogy questions A is to B as C is to [D, E, F]) typically require at least six concepts to be kept in mind simultaneously and compared.
Together, they make high IQ sound like a job description for a high-volume short-order cook, juggling the preparation of six different meals at the same time, hour after hour. Thus, high IQ might be without significance for the kind of lives that most people lead, or important only on those occasions demanding a quick versatility. A high IQ is usually necessary to perform well in very complex or fluid jobs (for example, being a doctor), and it’s an advantage in moderately complex ones (secretarial or police work), but it provides little advantage in work that requires only routine, unhurried decision making or simple problem solving (for example, for clerks and cashiers, whose reliability and social skills are likely to be far more important than their IQ).
IQ is certainly one fascinating aspect of intelligence, but it doesn’t subsume the others; we shouldn’t make the mistake of trying to reduce the subject of intelligence to a simple number on a rating scale. That would be like characterizing a football game in terms of one statistic, say, the percent of passes completed. Yes, over the football league as a whole, winning does significantly correlate with that statistic, but there’s a lot more to football than just percent-passes-completed; some teams win without completing a single pass, by emphasizing other strengths. IQ does correlate with ‘winning’ in many environments, but it’s not what the intelligence game is all about, any more than successful passing is what football is all about.
I think of intelligence as the high-end scenery of neurophysiology – the outcome of many aspects of an individual’s brain organization which bear on doing something one has never done before. We may not be able to explain intelligence in all its glory, but we now know some of the elements of an explanation. Some are behavioral, some are neurophysiological, and some are evolutionlike processes that operate in mere seconds. We even know something about the self-organizational principles that lead to emergent stuff – those levels-in-the-making, as when (to anticipate a later chapter) categories and metaphors compete for cerebral territory.
The big issue for understanding intelligence isn’t who has more but what intelligence is, when it’s needed, and how it operates. Some of what intelligence encompasses are cleverness, foresight, speed, creativity, and how many things you can juggle at once. More later.
Did our intelligence arise from having more of what other animals have? Just looking at the brain and judging it by its size, as if it were a contaloupe, is apt to be misleading. Only the outer shell, the cerebral cortex, is markedly involved in making novel associations. Most of the brain’s bulk comes from the insulation around the ‘wires’ that connect one part of the brain to another; the more insulation, the faster the messages flow. As animals become larger and distances greater, more insulation is needed to speed up transmission and keep the reaction times short; this insulation increases the bulk of the white matter even when the number of cortical neurons stays the same.
An orange peel is only a small part of an orange, and our cerebral cortex is even thinner: about 2 mm, the thickness of two dimes. Our cortex is extensively wrinkled; were it to be peeled off and flattened out, it would cover four sheets of typing paper. A chimpanzee’s cortex would fit on one sheet, a monkey’s on a postcard, a rat’s on a stamp. Were we to mark off a fine grid on the flattened surface, we’d find about the same number of neurons in each little grid square in all cortical regions (except primary visual cortex which, in all binocular animals, has lots of additional small neurons). So if you need more neurons for a particular function, you need more cortical surface area.
We tend to talk of demanding visual tasks for food-finding as ‘enlarging’ monkey visual cortex in later generations but not its auditory cortex, with evolution tending to produce a bulge here – and then, when some other selection pressure comes into play, a bump there. But there is now a strong suspicion that any non-olfactory natural selection for more brain space (say vision) results in more brain space for all the other functions as well – that it is often developmentally difficult to make regional enlargements of the brain. So enlarge one, enlarge them all may be the general rule, rather than an exception.
And if one evolutionary route to a free lunch isn’t enough, here’s another: new functions often first appear by making spare-time use of some preexisting part of the brain. Brain regions are, to some extent, multifunctional, resisting our attempts to label them. So, what preexisting functions might be most relevant to the quantum leap in cleverness and foresight during hominid evolution from the apes? Most would say language. I will argue that a ‘core facility’ common to language and to planning hand movements (and used in our spare time for music and dance) has even greater explanatory power than a special facility only for language functions.
Intelligence is sometimes described as a patchwork of know-how and know-what areas in the brain, all those perceptual mechanisms so sensitive to expectations. That is surely true, but if your definition of intelligence is so broad as to include most things that the brain does, such a formulation doesn’t advance your understanding any more than extending consciousness to cover plant life does. Catalogs are not explanations, no matter how interesting the list or how much the topics may need inclusion in an introductory course. It’s not my purpose to eliminate perceptual mechanisms from intelligence but to illuminate the underpinnings of guessing well and those levels of self-organization that produce stratified stability.
The Spanish physician Juan Huarte defined intelligence in 1575 as the ability to learn, exercise judgment, and be imaginative. In the modern literature, intelligence often connotes the capacity for thinking abstractly, for reasoning, and for organizing large quantities of information into meaningful systems. Not only does this sound like academics trying to define themselves, but it aims too high to be a definition that is readily extended to other animals. A better place to start for the what aspects is the animal behavior literature, where good operational definitions of intelligence center on versatility in problem solving.
Bertrand Russell once wryly noted, ‘Animals studied by Americans rush about frantically, with an incredible display of hustle and pep, and at last achieve the desired result by chance. Animals observed by Germans sit still and think, and at last evolve the solution out of their inner consciousness.’ Besides being a British commentary on the scientific fashions of 1927, Russell’s quip about problem-solving cleverness illustrates the usual false dichotomy between insight and random trial and error. Insight is, beyond argument, intelligent behavior. ‘Mere randomness’ is not, in the usual scheme of things; but we are thereby misled – of which more later.
I like Jean Piaget’s emphasis, that intelligence is what you use when you don’t know what to do. This captures the element of novelty, the coping and groping ability needed when there is no ‘right answer,’ when business as usual isn’t likely to suffice. Intelligent improvising. Think of jazz improvisations rather than a highly polished finished product, such as a Mozart or Bach concerto. Intelligence is about the process of improvising and polishing on the timescale of thought and action.
The neurobiologist Horace Barlow frames the issue a little more tightly, and points us toward experimentally testable aspects, by saying that intelligence is all about making a guess – not any old guess, of course, but one that discovers some new underlying order. ‘Guessing well’ neatly covers a lot of ground: finding the solution to a problem or the logic in an argument, happening upon an appropriate analogy, creating a pleasing harmony or a witty reply, correctly predicting what’s likely to happen next.
Indeed, you routinely guess what comes next, even subconsciously – say, in listening to a narrative or a melody. Getting a crying child to fill in the last word of each song line is an amazingly effective distraction, seen in many cultures. Subconscious prediction is often why a joke’s punch line or a P.D.Q. Bach musical parody brings you up short – you are surprised by the mismatch. Being a little wrong can be amusing, but substantial environmental incoherence is unpleasant, as when a day filled with job insecurity, noise, erratic drivers, and too many strangers leaves you frustrated, because of the frequent mismatch between what you expected and what actually happened.
Calvin’s Cure for Environmental Incoherence? Scale back the predictive challenges to a more comfortable level – not all the way into the boredom of surefire predictability but to where you’ll be right half the time. That way, you reassure yourself that you’re still competent at predicting. Perhaps that’s why, after a hard day awash in unpredictability, you tend to seek relief in ritual, music, or sitcoms – anything where you can again take pleasure in frequently guessing what comes next!
One of the beginner’s errors is to equate intelligence with purpose and complexity. Elaborate, complex behaviors initially seem like a reasonable place to look for signs of intelligence. After all, our language and foresight behaviors are surely aspects of intelligent behavior and they’re quite complex.
But many complex behaviors in animals are innate: no learning is needed as they’re wired in from birth. Such behaviors tend to be inflexible and often difficult to perform at will, just as sneezing and blushing are. These stereotyped movement patterns exhibit no more insight or understanding of purpose than does a computer program. They’re a set piece.
Both innate and learned behaviors can be long and complex. Consider, for example, the performance of an idiot savant, a person with enormous detailed recall but poor ability to make good use of the recollected information in a new context, by breaking the pattern into meaningful parts and recombining them. Whale song and insect nest building may be equally unintelligent.
That whales and birds link song sequences together is also not evidence of versatility. The most mindless of behaviors are often linked, the completion of one calling forth the next. Courtship behavior may be followed by intricate nest building, then segue into egg laying, then incubation, then the various stereotyped parental behaviors. Indeed, the more complex and ‘purposeful’ the behavior is, the further it may be from intelligent behavior, simply because natural selection has evolved a surefire way of accomplishing it, with little left to chance. Learning, after all, is usually focused on far simpler things than the complex chains of all-important behaviors.
The animal might understand its own behavior no better than we understand our yawn, or our tendencies to hug and kiss (clearly seen in bonobos and chimpanzees). Most animals in most contexts don’t appear to have much need for ‘understanding’ – in our sense of appreciating the underpinnings – and they don’t attempt innovations except by modest variations and a slow learning process. It’s as if thinking were a little-used backup, too slow and error-prone to be depended on in the normal course of things.
The best indicators of intelligence may be found in the simpler but less predictable problems that confront animals – those rare or novel situations for which evolution has not provided a standard response, so that the animal has to improvise, using its intellectual wherewithal. While we often take ‘intelligence’ to mean both a broad range of abilities and the efficiency with which they’re done, it also implies flexibility and creativity – in the words of the ethologists James and Carol Gould, an ‘ability to slip the bonds of instinct and generate novel solutions to problems.’ That narrows the what field quite a lot.
In tests of convergent thinking there is almost always one conclusion or answer that is regarded as unique, and thinking is to be channeled or controlled in the direction of that answer. . . . In divergent thinking, on the other hand, there is much searching about or going off in various directions. This is most obviously seen when there are no unique conclusions. Divergent thinking . . . is characterized . . . as being less goal-bound. There is freedom to go off in different directions. . . . Rejecting the old solution and striking out in some direction is necessary, and the resourceful organism will more probably succeed.
J. P. Guilford, 1959
Aren’t-they-clever stories are what many people recall when the topic of conversation turns to intelligence. Surely a dog qualifies as intelligent, they will insist. Most such stories turn out to hinge on how well a dog understands English or reads his owner’s mind.
Ethologists and animal psychologists will patiently reply that dogs are very social animals, expert in reading body language. A dog is always looking up to his owner, in the same way that a wild dog looks to the pack leader, asking, ‘What’s next, boss?’ or emotionally seeking reassurance in a juvenile manner, hoping to elicit benevolence. Talking to domesticated dogs plays into these innate tendencies, though your words per se may not carry the message. People don’t realize how much information is conveyed by the tone of voice and body language of the substitute leader (that’s you). If you read today’s newspaper headline to your dog in the same tone of voice, and with the same glances and postures, as you use to ask him to fetch your slippers, it might work just as well in evoking the desired behavior.
In many cases, there isn’t much to confuse the dog. The setting itself (people, places, situations, objects present) provides most of the information the dog needs to respond appropriately to a command. Most dogs have limited repertoires, and it’s therefore easy for them to guess correctly. Training a dog to fetch a dozen different items on command is a more difficult proposition, simply because it becomes harder for the dog to guess your intentions.
If you are confident that your dog understands words per se, you might try getting someone else to speak the words from another room over an intercom; this will eliminate most of the situational cues. Many smart animals cannot pass this severe a test of understanding spoken words, not even some extensively tutored chimpanzees who readily respond to graphical symbols. But dogs do pass the lesser test of performing the desired action most of the time, when the situation is familiar and the choices are obvious from the context.
The size of the response repertoire is one important factor in intelligence. Dogs have many instinctive behaviors, such as herding and alarm barks; they can learn many more. Even their communicative repertoire can reach impressive numbers with extensive training, as the psychologist Stanley Coren observes.
[My pet] dogs have a receptive language of about sixty-five words or phrases and about twenty-five signals or gestures for a total receptive vocabulary of about ninety items. They have a productive language of about twenty-five vocalizations and about thirty-five bodily gestures for a total productive vocabulary of about sixty items. They show no evidence of syntax or grammar. If they were human children, they would be demonstrating the level of language customary at around eighteen to twenty-two months of age. [Bonobos] that have learned [a sign or other symbolic] language can obtain [comprehension] scores equivalent to a child of around thirty months of age.
Speed of learning is also related to intelligence; one reason that dogs and dolphins achieve a wider repertoire of behavior with training is that they learn faster than cats usually do. So, ‘intelligence’ is quite a composite of other things, and many mental abilities are relevant. Perhaps it is making effective combinations of them that better constitutes intelligent behavior.
An animal’s selection of appropriate behavior may be the key to sorting out claims of animal intelligence. In many of the aren’t-they-clever animal stories, the animal isn’t thinking for itself, but merely responding to a command. Piaget’s element of creativity, in the face of an ambiguous task, is usually missing – except during the animal’s playful antics.
The scientific literature on nonhuman intelligence tries to cope with innovation, but since most putatively intelligent animal actions are not repeated actions, it’s hard to avoid a series of anecdotes (indeed, there’s a wonderful book of them about apes, Machiavellian Intelligence). The usual scientific hazards of anecdotal evidence can be somewhat reduced by emphasizing comparisons between species. For example, most dogs can’t untangle their leashes from around trees, but a chimpanzee seems to have what it takes. A leash-style snap fastener on the door will suffice to keep most small monkeys inside their cage, even if they can reach the fastener to fiddle with it. But the great apes can figure the fastener out, so you must use padlocks – and not leave the key lying around! Chimpanzees can practice deception: a chimp can guess what another animal is likely to be thinking, and can exploit this knowledge. But most monkeys don’t seem to have the mental machinery to deceive one another.
To many people, the essence of intelligence is such creative cleverness. When an animal is especially versatile at solving problems or inventing new moves, we consider that behavior to be particularly intelligent. But human intelligence is judged by additional standards.
When I tried cut this ‘creative cleverness’ definition of intelligence on one of my colleagues, he was dubious and started citing examples of the terminally clever.
You know, someone asks you how intelligent a certain person is, and you say, ‘Well, he’s certainly clever’ By this, you mean that he talks a good line – he’s versatile at improvising tactics in the short run but doesn’t follow through on his projects and lacks longer-term virtues, such as strategy, perseverance, and good judgment.
OK, I agreed, it also takes foresight to be truly intelligent. And chimps don’t think much about tomorrow, as far as anyone can tell from their behaviors, even if they occasionally do some planning on the half-hour timescale.
So maybe the flexible future is a human addition to ape intelligence. Intelligence also involves some imagination, I continued, remembering a high IQ group for whom I once gave an after-dinner speech. I had been surprised – in view of the fact that everyone in the audience had scored high on intelligence tests – at how unimaginative one of them was, and then I abruptly realized that I had always thought that IQ and imagination went hand in hand. But imagination contributes to intelligence only when shaped up into something of quality.
Patients with hallucinations are pretty imaginative, too, but that doesn’t necessarily make them highly intelligent.
It just goes to show that IQ measures only some aspects of what we more commonly understand as intelligent behavior. The very nature of IQ exams tends to preclude tests of creativity or the ability to make plans.
If I ever conceive any original idea, it will be because I have been abnormally prone to confuse ideas . . . and have thus found remote analogies and relations which others have not considered! Others rarely make these confusions, and proceed by precise analysis.
Kenneth J. W. Craik, The Nature of Explanation, 1943
Innovative behaviors are usually not new units: instead, they are composed of a novel combination of old elements: a different stimulus evokes a standard behavior, or some new combination of movements is used in response. How is sensory/movement innovation related to intelligence?
The sheer quantity of building-block types could be important. Cataloging the sensory and movement repertoires, as Stanley Coren did for dogs, is a useful exercise as long as one doesn’t take the stimulus-response dichotomy too literally. Sometimes responses appear without apparent triggers; there’s a lot of fiddling around, as when chimps strip the leaves off a branch for no apparent reason. Often the stimulus-response aspect is muted; the animal will seek out sensations as part of shaping the response. With those cautions, consider some classic examples of stimulus-response.
Many animals have sensory templates, which they try out for size (and shape) on what they see, rather like a child trying out a number of cookie cutters on the baked assortment of Christmas cookies, to see which (if any) fits a particular cookie. Baby birds, for example, crouch when a hawk flies overhead, a behavior suggesting that they were born with the image of a hawk wired into their bird brains. The reality is quite different: initially, they crouch when any sort of bird flies above them. They then come to recognize the sorts of birds they see every day; as a shape becomes familiar, they cease the response to it. Because of such habituation, they eventually crouch only in response to infrequently seen shapes, such as exotic birds that are just passing through – and to predators, such as hawks, which are infrequent because there aren’t very many of any species at the top of the food chain.
So the crouch is a response to novelty, not to a pre-wired ‘alarm’ search image. It’s as if the child found a misshapen cookie that none of the cookie cutters fitted, and was thereby distressed.
Composers note that while pure overtones (as from the flute) are relatively soothing, random overtones (as in heavy metal or the raspy voices of some singers, such as Mick Jagger) seem to signal threat or alarm, and I’ve long thought that the disordered sensations produced by nerve injuries are often perceived as painful (rather than merely nonsensical) for the same reason.
Besides sensory templates for familiar sights and sounds, animals also have familiar movement schemas, among which they pick and choose. A cormorant can decide whether to cruise around underwater in search of another meal, or fly away to another pond, or spread its wings to dry (cormorant feathers lack the oil that duck feathers have), or just stand around – presumably by consulting the weightiness of its wings, the fullness of its stomach, its sexual drives, and so forth. Decision making is something that all animals do; it is usually an economistlike weighing of sensations and drives, followed by a standard behavior from its repertoire, as modified by the circumstances.
Of course, we humans often do something similar in deciding on a restaurant, taking into account its menu, parking, cost, travel and waiting time, and ambiance – and somehow comparing all these factors with those of other restaurants. While such weighing of choices seems especially conscious, purposeful, and intentional, choice per se does not imply an extensive mental life – not of the kind we associate with creating novel additions to the list of choices for what to do next (‘Suppose there are any northern Vietnamese restaurants in town?’).
Curious, I took a pencil from my pocket and touched a strand of the [spider] web. Immediately there was a response. The web, plucked by its menacing occupant, began to vibrate until it was a blur. Anything that had brushed claw or wing against that amazing snare would be thoroughly entrapped. As the vibrations slowed, I could see the owner fingering her guidelines for signs of struggle. A pencil point was an intrusion into this universe for which no precedent existed. Spider was circumscribed by spider ideas; its universe was spider universe. All outside was irrational, extraneous, at best raw material for spider. As I proceeded on my way along the gully, like a vast impossible shadow, I realized that in the world of spider I did not exist.
Loren Elseley, The Star Thrower, 1978
Sometimes an animal tries out a new combination of sensory template and movement during play, and finds a use for that combination later on. So perhaps we should add play to our list of intelligence attributes.
Many animals, however, are playful only as juveniles. Being an adult is a serious business, with all those mouths to feed, so adults don’t have the time or inclination to fool around. A long juvenile period, characteristic of apes and humans, surely aids versatility because of the accumulation of useful combinations. In addition, some evolutionary trends, including domestication of animals, tend to cany over juvenile traits into adulthood – so that, too, might increase versatility.
You don’t learn just from your own experiences. You can copy the actions of others, as Japanese monkeys were observed to copy one inventive female’s technique for washing the sand off food. You may avoid what seems to spook others, even if you haven’t been personally threatened by it, and such ‘superstitious’ behavior can be passed on. The original reason for ‘Don’t step on the crack in the sidewalk’ may be lost, but the cultural transmission between generations continues for centuries, sufficient unto itself.
A wide repertoire of ‘good moves,’ of course, makes foresight a lot easier. Foresight initially seems simple, almost too simple to be a requirement for high intelligence. But that’s because we confuse foresight with species-specific seasonal behaviors.
Squirrels hoarding nuts for winter seems to be the standard example of planning ahead in the animal kingdom. And we now know how such things work. The hormone melatonin, released from the pineal gland during the hours of darkness, serves to warn of the approach of winter. Longer and longer nights result in the release of increasing amounts of melatonin, which in turn triggers food hoarding and new fur coats. It doesn’t take much of a brain to do that kind of ‘planning.’
There are, of course, other behaviors created by the brain’s initial wiring which serve to set things up for months ahead. Mating behaviors have the effect of producing offspring after a considerable delay. Seasonal migrations come with innate brain wiring or are learned by juveniles and become mindless adult rituals. Of course, such behavior isn’t the result of planning at all. Seasons are eminently predictable; and over the millennia, plants and animals have been shaped by evolution to sense the signs of approaching winter by means of the innate surefire mechanisms: hoarding nuts probably ‘feels good’ as the days shorten, much as does following the gradient of a sexual pheromone in the air.
Planning on the timescale of a few minutes is seen in some instances, but, as you’ll see, none should probably be called planning. Keeping a movement plan on hold – as when caged monkeys who have watched food being hidden are able to locate it twenty minutes later, when they are let out of their cages – is sometimes called ‘planning.’ But is it simply the memory of an intention? Another disputed type of evidence arises from spatial maneuvering. When bees are kidnapped and carried in a windowless container to a random location several kilometers distant and then released, they quickly set off on the optimal path to an unseen favorite food source. Is this planning, or are they just referencing memories of horizon profiles? Before setting off in the correct direction, they fly a few circles first to get oriented; they may well be scanning the horizon for clues.
Perhaps we should say that planning involves something novel, closer to the way in which we procrastinate, figuring out what can safely be put off until tomorrow (or avoided altogether). Indeed, I’d reserve the term for the assembling of multiple stages of the move in advance of action – not when you organize the later stages after getting the initial moves in motion, which goal-plus-feedback can accomplish.
Alas, there is surprisingly little evidence for this kind of multistage planning in the great apes, even in their frequent behaviors. None of the termite-fishing chimps, as the polymath Jacob Bronowski once pointed out, ‘spends the evening going round and tearing off a nice tidy supply of a dozen probes for tomorrow.’ Although wild chimps often seem to arrive at a distant fruit tree just as the fruit is ripening, how much of that is migration ritual and how much is all-in-advance planning of a unique route?
For most of your movements, such as raising a coffee cup to your lips, there is time for improvisation en route. If the cup is lighter than you remembered, you can correct its trajectory before it hits your nose. Thus, a complete advance plan really isn’t needed; a goal and periodic piecewise elaboration will suffice. You get started in the general direction and then correct your path, just as a moon rocket does. Most ‘planning’ stories involving animals fit into that mold.
Multistage planning is perhaps best seen in an advanced type of social intelligence: making a mental model of someone else’s mental model, then exploiting it. Imagine a chimp that cries ‘food’ in a place where there is no food, and then quietly circles back through the dense forest to where it actually saw the food earlier. While the other chimps beat the bushes at the site of the food cry, the chimp that uttered it gets to eat all the food rather than having to share it.
What’s really difficult is to make a detailed advance plan in response to a unique situation – like those leftovers in the refrigerator and what might go with them. It requires imagining multiple scenarios, whether you are a hunter plotting various approaches to a deer or a futurist spinning three scenarios bracketing what an industry might look like in another decade. Compared to apes, we do a lot of that: we are even capable of occasionally heeding the eighteenth-century admonition of Edmund Burke, ‘The public interest requires doing today those things that men of intelligence and goodwill would wish, five or ten years hence, had been done.’
So multistage planning for novel situations is surely an aspect of intelligence – indeed, one that appears greatly augmented in the transition from the ape brain to the human brain. But knowledge is, I think, a commonplace.
A base of existing knowledge is, of course, required for versatility, foresight, and creativity. You can’t be a poet or scientist without a good vocabulary, but definitions of intelligence that stress knowledge or memory’s synaptic mechanisms really do miss the mark; they’re mistaken reductionism – the practice of reducing something to its fundamental constituents, which for present purposes is carried a few steps too far. This is the mistake, as I explain in the next chapter, that the consciousness physicists often make.
For example, Shakespeare didn’t invent the vocabulary he used. He invented combinations of those words, most notably the metaphors that allow relationships to be imported from one level of discourse to another. In a similar manner, much intelligent behavior consists of new combinations of old things.
Deductive logic is another what aspect of intelligence – at least, of the human variety. Philosophers and physicists have, I suspect, been unduly impressed with the human faculty for logical reasoning. Logic might consist of guessing the underlying order of things, à la Horace Barlow, but only in situations where an unambiguous underlying order exists to be guessed (mathematics being the prime exemplar). Piecewise approximation, as with the guessing needed for long division, could operate subconsciously so rapidly as to seem like a leap to the finished ‘logical’ product. Could it be that logic is more a property of the subject matter than of the mental process – that guessing is the name of the game during mental calculations as well as during creative thinking?
The what list can be extended further, both for what is and what isn’t. But I am going to focus hereafter on Barlow’s guessing-at-order aspect and more generally on Piaget’s improvisation problem of how to proceed when the choice isn’t obvious. I realize that this excludes certain uses of the word ‘intelligence,’ as when we talk of intelligent design or military intelligence, but the guessing aspect buys us such a broad range of intelligence connotations that we will do well to organize analysis around it – provided we can avoid consciousness confusions and inappropriate levels of explanation.
The mixture of hormone-driven aggression, sexual and social lust for power, deceit and gamesmanship, friendship and spite, and good-and ill-natured fun ring familiar chords . . . there is no reasonable way to account for much of primate (and especially chimpanzee) behavior without assuming that these animals understand a great deal about what they are doing and seeking to do, and are inferring almost as much as humans do about the intentions and attitudes of their peers.
James L. Gould and Carol Grant Gould, The Animal Mind, 1994
CHAPTER 3
THE JANITOR’S DREAM
Human consciousness is just about the last surviving mystery. A mystery is a phenomenon that people don’t know how to think about – yet. There have been other great mysteries: the mystery of the origin of the universe, the mystery of life and reproduction, the mystery of the design to be found in nature, the mysteries of time, space, and gravity. These were not just areas of scientific ignorance, but of utter bafflement and wonder. We do not yet have all the answers to any of the questions of cosmology and particle physics, molecular genetics and evolutionary theory, but we do know how to think about them. . . . With consciousness, however, we are still in a terrible muddle. Consciousness stands alone today as a topic that often leaves even the most sophisticated thinkers tongue-tied and confused. And, as with all of the earlier mysteries, there are many who insist – and hope – that there will never be a demystification of consciousness.
Daniel C. Dennett, Consciousness Explained, 1991
As Charles Mingus said about jazz, you can’t improvise from nothing, you have to improvise from something. The Romans’ phrasing was Ex nihilo nihil fit. Creating a novel plan of action has to start somewhere and then refine things. The two greatest examples of creativity in action, species evolution and the immune response, both utilize a darwinian process to shape up crude beginnings into something of quality. But confusions about consciousness (not to mention confusions about levels of mechanisms) usually lead us astray when we attempt to apply darwinism to our mental lives. That’s probably why more than a century passed with so little progress on mental darwinism.
In the last chapter, I discussed something of what intelligence is and isn’t. Here I am going to attempt the same thing for consciousness, hoping to head off repetitions of those arguments that have sidetracked William James’s idea. There is wide overlap between the connotations of consciousness and intelligence, though the C word tends to refer to the waking-aware aspect of our mental lives, while intelligence tends to refer to the imagination or efficiency of our mental lives. Bear in mind that the higher types of intellect may actually require conscious (and therefore subconscious) processing.
How should we approach explaining the unknown? It is well to keep overall strategy in mind, especially whenever attractive shortcuts are offered as explanations by those whom the philosopher Owen Flanagan terms ‘the new mysterians.’ Using Dennett’s epigrammatic definition of a mystery, consider for a moment those physicists who are speculating about how quantum mechanics might have a role in consciousness, might provide ‘free will’ an escape route from ‘determinism’ via quantum mechanical processes down at the subcellular level, in the thin microtubules that often cluster near synapses.
I’m not going to take the space needed to do justice to their best-selling arguments (or rather the arguments of their best-selling books), but when you consider how little they actually encompass (let alone explain) of the wide range of themes involved in consciousness and intelligence, you might feel (as I do) that they’re just another case of ‘much ado about very little.’
Moreover, as studies of chaos and complexity have been teaching us, determinism is really a nonissue, suitable only for cocktail party conversational gambits and hardly in need of a quantum mechanical escape clause. With some notable exceptions (I call them ecclesiastical neuroscientists, after the great Australian neurophysiologist John C. Eccles), neuroscientists seldom talk in this way; indeed, we rarely engage in any sort of word games about consciousness.
It’s not for lack of interest; how the brain works is, after all, our primary preoccupation. Over our beers after a hard day at the neuroscience meetings, we tell each other that while we may not have wide-ranging explanations of consciousness yet, we do know what kinds of explanations don’t work. Word games produce more heat than light, and the same is true of explanations that simply replace one mystery with another.
Neuroscientists know that a useful scientific explanation for our inner life has to explain more than just a catalog of mental capabilities. It also has to explain the characteristic errors that the consciousness physicists ignore – the distortions of illusions, the inventiveness of hallucinations, the snares of delusions, the unreliability of memory, and our propensities to mental illnesses and seizures rarely seen in other animals. An explanation has to be consistent with many facts from the last century of brain research – with what we know about consciousness from studies of sleep, strokes, and mental illness. We have numerous ways of ruling out otherwise attractive ideas; I’ve heard a lot of them in thirty years of doing brain research.
There are various angles along which to cut the cake of our mental lives. I tried focusing on consciousness in The Cerebral Symphony. One reason that I’m going to hereafter avoid a discussion of consciousness in favor of intelligence underpinnings is that considerations of consciousness quickly lead to a passive observer as the end point, rather than someone who explores, who adventures within the world. You can see that in the many ‘consciousness’ connotations you’ll find in a dictionary:
Capable of or marked by thought, will, design, or perception.
Personally felt, as in ‘conscious guilt.’
Perceiving, apprehending, or noticing, with a degree of controlled thought or observation. (In other words, fully aware.)
Having mental faculties undulled by sleep, faintness, or stupor: ‘She became conscious after the anesthesia wore off.’ (In other words, awake.)
Done or acting with critical awareness: ‘He made a conscious effort to avoid the same mistakes.’ (Here, ‘deliberate’ may substitute for ‘conscious.’)
Likely to notice, consider, or appraise: ‘He was a bargain-conscious shopper.’
Marked by concern or interest: ‘She was a budgetconscious manager.’
Marked by strong feelings or notions: ‘They are a raceconscious society.’ (For these last three uses, ‘sensitive’ may be substituted.)
The philosopher Paul M. Churchland has recently made a more useful list, noting that consciousness:
utilizes short-term memory (or, as it is sometimes called, working memory).
is independent of sensory inputs, in that we can think about things not present and imagine unreal things.
displays steerable attention.
has the capacity for alternative interpretations of complex or ambiguous data.
disappears in deep sleep.
reappears in dreaming.
harbors the contents of several sensory modalities within a single unified experience.
Again, this list has the passive-observer focus rather than the explorer focus, but we see the Piagetian notion of intelligence incorporated into a consciousness definition, in the ‘alternative interpretations’ item.
Among scientists, there is a tendency to use consciousness to mean awareness and recognition; for example, Francis Crick and Christof Koch use consciousness when addressing the ‘binding problem’ in object recognition and recall. But just because one word (in English) is used to denote these widely different mental faculties doesn’t mean that they share the same neural mechanism. Other languages, after all, may assign one or another of the aforementioned ‘consciousness’ connotations its own word. Crick’s thalamocortical theory is most useful for thinking about object recognition, but it doesn’t say anything about anticipation or decision making – yet those are often among the connotations of consciousness, the word he uses. It’s easy to overgeneralize, just by the words you choose. This isn’t a criticism: there aren’t any good choices until we understand mechanisms better.
By now, the reader might reasonably conclude that consciousness connotations are some sort of intelligence test that examines one’s ability to float in ambiguity. Debates about consciousness regularly confuse these connotations with one another, the debaters acting as if they believed in the existence of a common underlying entity – ‘a little person inside the head’ – that sees all. To avoid this presumption of a common mechanism for all connotations, we can use different English words for different connotations, such as when we use ‘aware’ and avoid ‘conscious.’ I usually try to do this, but there are also pitfalls when you use alternative words. That’s because of what might be termed back-translation.
Physicians, for example, try to avoid the C word by talking instead about the level of arousal that can be achieved with some shouting and prodding of the patient (coma, stupor, alertness, or full orientation to time and place). That’s fine, until someone tries to translate back into C-word terminology; yes, a person in a coma is unconscious, but to say that consciousness is at the opposite end of the arousability scale may be seriously misleading.
Worse, equating ‘conscious’ with ‘arousable’ tends to be interpreted as ascribing consciousness to any organism that can experience irritation. Since irritability is a basic property of all living tissue, plant as well as animal, this extends consciousness to almost everything except rocks; some nonscientists are already talking about plant consciousness. While this is appealing to some people and appalling to others, scientifically it is simply bad strategy (even if true). If you throw everything into the consciousness pot and mix it up, you reduce your chances of understanding consciousness.
With so many major synonyms (aware, sensitive, awake, arousable, deliberate, and more), you can see why everyone gets a little confused talking about consciousness. One often hears the word’s connotation shift in the course of a single discussion; were this to happen to the word ‘lift,’ with one speaker meaning what hitchhikers get and the other meaning an elevator, we’d burst out laughing. But when we talk about consciousness, we often fail to notice the shift (and debaters even exploit the ambiguity to score points or sidetrack the argument).
And there’s more: at least within the cognitive neuroscience community, consciousness connotations include such aspects of mental life as the focusing of attention, vigilance, mental rehearsal, voluntary actions, subliminal priming, things you didn’t know you knew, imagery, understanding, thinking, decision making, altered states of consciousness, and the development of the concept of self in children – all of which grade over into the subconscious as well, all of which have automatic aspects that our ‘narrator of consciousness’ may fail to notice.
Many people think that the narratives we tell ourselves when awake or dreaming tend to structure our consciousness. Narratives are an important part of our sense of self, and not merely in an autobiographical sense. When we play a role – as when the four-year-old engages in make-believe, playing ‘doctor’ or ‘tea party’ – we must temporarily step outside of ourselves, imagine ourselves in someone else’s place, and act accordingly. (The ability to do this is one of the more useful definitions of a sense of self.)
But narratives are an automatic part of everyday life in our own skins. Starting around the age of three or four, we make stories out of most things. Syntax is often a junior version of narrative: just the word ‘lunch’ in a sentence sends us looking for variants of the verb ‘eat,’ for the food, the place, and persons present. A verb such as ‘give’ sends us searching for the three required nouns we need to fit into roles: an actor, an object given, and a recipient. There are lots of standard relationships, with familiar roles for the players, and we guess from the context what goes into any unfilled gaps. Often we guess well, but dreams illustrate the same kinds of confabulation seen in people with memory disorders, in which bad guesses are unknowingly tolerated.
‘Perception,’ it has been recently said, ‘may be regarded as primarily the modification of an anticipation.’ It is always an active process, conditioned by our expectations and adapted to situations. Instead of talking of seeing and knowing, we might do a little better to talk of seeing and noticing. We notice only when we look for something, and we look when our attention is aroused by some disequilibrium, a difference between our expectation and the incoming message. We cannot take in all we see in a room, but we notice if something has changed.
E. M. Gombrich, Art and Illusion, 1960
A sense of self is thought to go along with a fancy mental life, so let me briefly address the common notion that self-awareness (often called self-consciousness) involves sophisticated, ‘intelligent’ mental structures.
How do you know which muscles to move in order to mimic the action of someone else – say, in order to stick out your tongue in response to seeing such an action? Do you have to see yourself in a mirror first, to make the association between that sight and the muscle commands that will mimic it?
No. Newborn humans can imitate some of the facial expressions they see, without any such experience. This suggests that innate wiring connects at least some sensory templates with their corresponding movement commands – that we’re ‘wired to imitate’ to some extent. Such wiring might explain why some animals can recognize themselves in a mirror, while others treat their mirror image as another animal, to be coaxed or threatened. Chimps, bonobos, and orangutans can recognize themselves either immediately or within a few days’ experience; gorillas, baboons, and most other primates cannot. A capuchin monkey (Cebus are the most intelligent of the New World monkeys and the best tool users) with a full-length mirror in its cage may spend weeks threatening the ‘other animal.’ Ordinarily, one animal would back down after a brief period, acknowledging the other as dominant. But in the case of the mirror monkey, nothing is ever resolved; even if the capuchin tries acting submissive, so does the other animal. Eventually the monkey begins acting so depressed at the unresolved social conflict that the experimenters must remove the mirror.
What might self-recognition involve? Actions produce expectations about what sensory inflow will result from them (so-called efference copy), and so the perfect fit of these sensory predictions with the inputs from your skin and muscles during small movements would provide a way of recognizing yourself in an image. In the case of wild animals, a perfect fit of the image’s movements with internal predictions for facial movements would certainly be unusual, since they rarely see their own face.
The issue of self-consciousness in the animal literature could revolve about something as simple as the attention paid to predictions about facial sensations. That’s part of consciousness considerations, certainly, but hardly the pivot that some would make it. Self-recognition surely involves both Horace Barlow’s guessing right and Jean Piaget’s sophisticated groping, but I’d put it on the list of things that intelligence isn’t. Self-recognition is surely more to the point than quantum fields, however.
Do the enigmas of quantum mechanics really have something to do with such conscious aspects of our mental lives? Or is the invocation of QM in the consciousness context just another mistaken instance of suggesting that one area in which mysterious effects are thought to lurk – chaos, selforganizing automata, fractals, economics, the weather – might be related to another, equally mysterious one? Most such associations certainly conflate the unrelated, and when the two areas are at opposite ends of the spectrum of enigmatic phenomena, the argument is particularly suspicious.
Reducing things to basics – the physicists’ rallying cry – is an excellent scientific strategy, as long as the basics are at an appropriate level of organization. In their reductionist enthusiasm, the consciousness physicists act as if they haven’t heard of one of the broad characteristics of science: levels of explanation (frequently related to levels of mechanism). The cognitive scientist Douglas Hofstadter gives a nice example of levels when he points out that the cause of a traffic jam is not to be found within a single car or its elements. Traffic jams are an example of selforganization, more easily recognized when stop-and-go achieves an extreme form of quasi-stability – the crystallization known as gridlock. An occasional traffic jam may be due to component failure, but faulty spark plugs aren’t a very illuminating level of analysis – not when compared to merging traffic, comfortable car spacing, driver reaction times, traffic signal settings, and the failure of drivers to accelerate for hills.
The more elementary levels of explanation are largely irrelevant to traffic jams – unless they provide useful analogies. Indeed, packing principles, surface-to-volume ratios, crystallization, chaos, and fractals are seen at multiple levels of organization. That the same principle is seen at several levels does not, however, mean that it constitutes a level-spanning mechanism: an analogy does not a mechanism make.
Quasi-stable levels make selforganization easier to spot, especially when building blocks – such as crystals – emerge. Since we are searching for some useful analogies to help explain our mental lives, it is worth examining how levels of explanation have functioned elsewhere. The tumult of random combinations occasionally produces a new form of organization. Some forms, such as the hexagonal cells that appear in the cooking porridge if you forget to stir it, are ephemeral. Other forms may have a ‘ratchet’ that prevents backsliding once some new order is achieved. While crystals are the best known of these quasi-stable forms, molecular conformations are another, and it is even possible that there are quasi-stable forms at intermediate levels – such as microtubule quantum states where the consciousness physicists would like the action to be.
Stratified stability refers to stacking up such quasi-stable levels. Life-forms involve piling up quite a few of them; occasionally they collapse like a house of cards and the higher forms of organization dissolve (which is one way of thinking about death).
Between quantum mechanics and consciousness are perhaps a dozen of these persistent levels of organization: examples include chemical bonds, molecules and their selforganization, molecular biology, genetics, biochemistry, membranes and their ion channels, synapses and their neurotransmitters, the neuron itself, the neural circuit, columns and modules, larger-scale cortical dynamics, and so on. In neuroscience, one is always aware of these levels, because of the intense rivalry between neuroscientists working at adjacent levels.
An occasional alteration in consciousness is due to widespread failures in certain types of synapses. But a more appropriate level of inquiry into consciousness is probably at a level of organization immediately subjacent to that of perception and planning: likely (in my view), cerebral-cortex circuitry and dynamic selforganization involving firing patterns within a constantly shifting quiltwork of postage-stamp-sized cortical regions. Consciousness, in any of its varied connotations, certainly isn’t located down in the basement of chemistry or the subbasement of physics. This attempt to leap, in a single bound, from the subbasement of quantum mechanics to the penthouse of consciousness is what I call the Janitor’s Dream.
Quantum mechanics is probably essential to consciousness in about the same way as crystals were once essential to radios, or spark plugs are still essential to traffic jams. Necessary, but not sufficient. Interesting in its own right, but a subject related only distantly to our mental lives.
Yet, because mind seems ‘different’ from mere matter, many people still assume – despite all the foregoing – that this means some spooky stuff is needed to explain it. But the mind should be seen as something like a crystal – comprised of the same old matter and energy as everything else, just temporarily organized in some complicated way. This is hardly a new idea; witness Percy Bysshe Shelley in the early nineteenth century:
It has been the persuasion of an immense majority of human beings that sensibility and thought [as opposed to matter] are, in their own nature, less susceptible of division and decay, and when the body is resolved into its elements, the principle which animated it will remain perpetual and unchanged. However, it is probable that what we call thought is not an actual being, but no more than the relation between certain parts of that infinitely varied mass, of which the rest of the universe is composed, and which ceases to exist as soon as those parts change their position with respect to each other.
The traffic flow patterns in brains are far more complicated than those in vehicular movement; fortunately, there are in music some similarities that we can exploit for analogies. Understanding consciousness and intelligence will require better metaphors and actual mechanisms, not steps backward into word games or spooky stuff.
Ghosts are another version of spooky stuff, and for our analysis of creative mental life it’s worth looking at what has happened to the ghost concept. Ghosts illustrate the other essential creative aspect of mind, the role of memory.
The very presence of the word ‘ghost’ in most languages suggests that quite a few people have needed to talk about inexplicable things they’ve heard or seen. Why have so many people considered ghosts to be real? Is this where the notion of an incorporeal spirit world got started?
We now know that ghosts appear real because of mistakes made in the brain. Some are trivial, everyday mistakes and others arise from abnormalities in dreaming sleep; a few are stirred up by small epileptic seizures or the pathological processes seen in psychosis. We call them hallucinations; they involve false sounds more often than false sights. The people and pets that they feature are often scrambled a bit, just as they are in the jumble of our nighttime dreams.
Remember that what you see under normal circumstances owes its stability to a mental model that you construct. Your eyes are actually darting all over, producing a retinal image of the scene as jerky as an amateur video, and some of what you thought you saw was instead filled in from memory. In a hallucination, this mental model is carried to an extreme: memories stored inside your brain are interpreted as current sensory input. Sometimes this happens when you are struggling to wake up, when the paralysis of the muscles during dreaming sleep hasn’t worn off as fast as usual. Dream elements appear superimposed on the image of real people walking around the bedroom. Or you might hear a dead relative speak to you with a familiar phrase. Half the brain is awake, and the rest is still dreaming away. With any luck, you realize this and don’t try to place a more exotic construction on it. Each of us, after all, experiences nightly the symptoms of dementia, delusions and hallucinations in the course of our dreaming sleep; we’re accustomed to discounting such things.
Yet hallucinations can also happen when you are lying awake at night, or when you are working during the day. I suspect that many of these ‘ghosts’ are just simple cognitive mistakes, like one that recently happened to me: I heard a distinct crunching sound in the kitchen, which was repeated a moment later. Ah, I thought as I continued typing, the cat is finally eating her dry food. It took another two seconds before ‘Oops, let’s play that again.’ The cat, alas, had been dead for several months, and had had a long period of being fussy about her food. What I had faintly heard turned out to be the automatic defroster on our refrigerator – it’s somewhat more subtle than the racket made by icemakers – and I had routinely made a guess about what the sound meant without fully considering the matter.
We are always guessing, filling in the details when something is heard faintly. A squeaking screen door, blown by the wind, may sound enough like the I-want-food whine of your dear departed dog for you to ‘hear’ the dog again. Once this memory is recalled, it may be very hard to replay the actual sound you heard – and so the fill-in of details from memory becomes the perceived reality. This isn’t unusual; as William James noted a century ago, we do it all the time:
When we listen to a person speaking or [we] read a page of print, much of what we think we see or hear is supplied from our memory. We overlook misprints, imagining the right letters, though we see the wrong ones; and how little we actually hear, when we listen to speech, we realize when we go to a foreign theatre; for there what troubles us is not so much that we cannot understand what the actors say as that we cannot hear their words. The fact is that we hear quite as little under similar conditions at home, only our mind, being fuller of English verbal associations, supplies the requisite material for comprehension upon a much slighter auditory hint.
This fill-in from memory is part of what’s known as categorical perception; we just call it a hallucination when we are unaware of what triggered it. Unless a sound repeats, we may not be able to compare our filled-in perception of it to the original; fortunately, where visual phenomena are concerned, we can often manage a second look and catch the mistake before getting committed to ‘the apparition.’
We now know that suggestibility (it doesn’t even take hypnosis) and stress (it doesn’t even require grieving) can augment our natural tendencies to jump to conclusions, allowing memories to be interpreted as current reality. If I’d been stressed out over something, I might not have searched for an alternative explanation until it was too late to walk into the kitchen and find the sound’s true source. Later, upon recalling that I’d ‘heard’ the dead cat, I might have fallen into the common nonscientific explanations: ‘It was a ghost!’ or ‘I must be losing my mind! Maybe it’s Alzheimer’s!’ Both possibilities are frightening, and both are highly unlikely. But if they’re the only explanations that occur to you, you may make yourself quite unhappy.
Have the scientific explanations eliminated ghosts from our culture? At least for those at the educational level of juveniles, the whole notion of ghosts remains a cheap thrill (for exactly the same reason that dinosaurs are so popular with children: they’re the potent triple combination of big, scary, and safely dead). Temporal-lobe epileptics, before a physician explains their hallucinations to them, don’t think ghosts are funny at all. Grieving relatives may wish, in retrospect, that someone had warned them about meaningless hallucinations.
In this case, science (for those whose education includes it) can eliminate what was once a frightening mystery. Science doesn’t merely empower us, as in seeding better technologies; it also helps prevent trouble in the first place. Knowledge can be like a vaccine, immunizing you against false fears and bad moves.
There’s a second neuroscience ghost story: the philosopher Gilbert Ryle’s lovely phrase ‘the Ghost in the Machine’ refers to the little-person-inside manner in which we commonly refer to the ‘us’ inside our brains. It has led some researchers to talk about the ‘interface’ between ‘mind’ and brain, between the unknowable and the knowable. Descartes’s pineal gland proposal dressed up in modern clothes by the new mysterians?
We’re now making good progress in replacing such pseudo-spirits with better physiological analogies – and even, in some cases, with actual brain mechanisms. Just as an earlier generation of scientists usefully eliminated the external ghosts, I like to think that our currently evolving knowledge about the spirit substitutes will help people think more clearly about themselves and interpret their experiences more reliably, and will help psychiatrists to interpret the symptoms of mental illness.
The consciousness physicists, with their solution in search of a problem, surely aren’t intending to tell yet another ghost story. They’re just having a good time speculating, in the manner of science fiction writers. (Still, consider how odd it would be for neuroscientists to speculate about the enigmas of physics, even those neurophysiologists – and there are many – who once took several courses in quantum mechanics). But why do these physicists take themselves so seriously, when they’re ignoring a dozen levels of organization outside their own specialties? Specialization itself is perhaps part of the answer, and it demonstrates one of the hazards of intelligence.
Specialization in science is all about asking answerable questions, which requires focusing on the details – and that takes a lot of time and energy. None of us really wanted to give up those wonderful debates we carried on as undergraduates about the Big Questions. We cared about those questions. They’re what attracted us to science in the first place. They’re not obsolete, like the ghosts. But the subsequent intellectual development of working scientists sometimes reminds me of what it’s like to be in a canal lock as the water level drops.
At least in Seattle, that’s like being in a giant bathtub with a view of waterfront, fish ladders, mountains, and spectators. Once the plug is pulled, your boat sinks, and your attention is captured by the formation of the whirlpools in the lock, which are bouncing the boats about. They’re fascinating. If you stick an oar in one, you can spawn off secondary whirlpools. Self-similarity theories suggest themselves, and so begins a digression into fractals.
Should you look up from your experiments and your theorizing in this oversized bathtub, the view of your surroundings will have become a rectangular patch of sky. Now you’re looking out from inside a big wet box, whose walls are one or two stories high. In the patch of sunlight on the north wall of the box are some shadows of the people standing topside. As in Plato’s Cave, you start to interpret the shadows on the walls, making imperfect guesses about what’s really happening up there. What appears to be two people slugging each other turns out to be nothing more than one person standing in front of the other and gesturing wildly while carrying on a conversation.
Specialization can be like that – no more big picture, unless you come up for air occasionally and admire the scenery, see the fuller context.
The price of progress is often an unfamiliarity with other levels of organization, except for those just above or below that of your specialty. (A chemist might know biochemistry and quantum mechanics, but not much neuroanatomy.) When you’ve got no data but those supplied by your own mental life, it’s easy to give fanciful interpretations of the shadows on the wall. Still, sometimes that’s the best you can do, and Plato and Descartes did it very well in their day.
But when you can do better, why be satisfied with shadow boxing? Or continue to play word games? A word itself, one eventually realizes, is a very poor approximation of the process it represents. By the end of this book, the reader will, I hope, be able to imagine some neural processes that could result in consciousness – processes that can operate rapidly enough to constitute a quick intelligence.
Describing our mental lives has a well-known hang-up, the old subjectivity snare associated with point of view, but there are two other whirlpools we will also need to navigate around.
The passive observer, poised in the mental middle between sensation and action, is a point of view that leads to all sorts of needless philosophical trouble. Partly, that’s because sensation is only half of the loop, and we thereby ignore sensation’s role in preparing for action. Some of the more elaborate couplings of sensation to action are called ‘cortical reflexes,’ but we also need to understand how thought is coupled to action in an intelligent manner, when we grope for a novel course of action. Ignoring the mental middle, as the behavioral psychologists did a half century ago, is not a long-term solution. What neuroscientists often do is investigate the preparation for movement; that gets us closer to the thought process.
We often talk of our mental activities as being subdivided among sensing, thinking, and acting phases. But trouble arises because few things happen at one point in time and space. All of the interesting actions in the brain involve spatiotemporal patterns of cellular activity – not unlike what constitutes a musical melody, where the space is the keyboard or musical scale. All our sensations are patterns spread out in time and space, such as the sensation from your fingers as you get ready to turn to the next page. So too, all our movements are spatiotemporal patterns involving the different muscles and the times at which they are activated. When you turn the page, you are activating about as many muscles as you use in playing the piano (and unless you get the timing just right, you won’t be able to separate the next page from the rest). Still, we often try to understand mental events by treating them as if they actually occurred at one place and happened at one instant.
But what’s in the mental middle is also a spatiotemporal pattern – the electrical discharges of various neurons – and we shouldn’t count on these discharges being funneled through one point in space (such as a particular neuron) and a decision being made at one point in time (such as the moment when that particular neuron discharges an impulse), as if a perception or thought consisted of playing a single note, once. I know of only one such case in vertebrates (occasionally nature makes things convenient for neurophysiologists): it’s an escape reflex in fish, conveniently channeled through a single large brain-stem neuron, whose discharge initiates a massive tail flip. But higher functions inevitably involve large overlapping committees of cells, whose actions are spread out in time, and that’s a more difficult concept. Understanding higher intellectual function requires us to look at the brain’s spatiotemporal patterns, those melodies of the cerebral cortex.
In addition to the navigational hazards, we need to select our building blocks with care, so that we don’t simply replace one mystery with another. Premature closure is the most obvious hazard in selecting building blocks – sometimes we stop surveying candidate mechanisms too early, as when we explain via spirits or quantum fields.
We must also beware of hazards having to do with the end points of an ‘explanation’: the New Age everything-is-related-to-everything and the reductionistic explanations at an inappropriate level of organization (what the consciousness physicists and ecclesiastical neuroscientists do, in my not-so-humble opinion).
Explaining mental life is a big task, and you may have noticed that this is a reasonably thin book. As noted, instead of further exploring consciousness connotations, I’m going to cut the cake differently, focusing on the structures of our mental lives that are associated with intelligence. Intelligence is all about improvising, creating a wide repertoire of behaviors, ‘good moves’ for various situations. A focus on intelligence covers a lot of the same ground as does a focus on consciousness – but it avoids many of the navigational hazards. Most important, the good-moves repertoire is an end point very different from the snapshots of passive contemplation. Certainly it’s easier to find a continuity between ourselves and the rest of the animal kingdom by addressing the subject of intelligence, compared to the muddle we generate when we try to talk about animal ‘consciousness.’ And so the next task is to take a brief look at where good guessing might have come from, in evolutionary terms.
The paradox of consciousness – that the more consciousness one has, the more layers of processing divide one from the world – is, like so much else in nature, a trade-off. Progressive distancing from the external world is simply the price that is paid for knowing anything about the world at all. The deeper and broader [our] consciousness of the world becomes, the more complex the layers of processing necessary to obtain that consciousness.
Derek Bickerton, Language and Species, 1990
CHAPTER 4
EVOLVING INTELLIGENT ANIMALS
The apes I know behave every living, breathing moment as though they have minds that are very much like my own. They may not think about as many things, or in the depth that I do, and they may not plan as far ahead as I do. Apes make tools and coordinate their actions during the hunting of prey, such as monkeys. But no ape has been observed to plan far enough ahead to combine the skills of tool construction and hunting for a common purpose. Such activities were a prime factor in the lives of early hominids. These greater skills that I have as a human being are the reason that I am able to construct my own shelter, earn my own salary, and follow written laws. They allow me to behave as a civilized person but they do not mean that I think while apes merely react.
Sue Savage-Rumbaugh, 1994
Answering the how questions is often our closest approach to answering a why question. Just remember that the answers to how mechanisms come in two extreme forms, which are sometimes known as proximate and ultimate causation. Even the pros sometimes get them mixed up, only to discover that they’ve been arguing about two sides of the same coin, so I suspect that a few words of background are needed here.
When you ask, ‘How does that work?’ you sometimes mean how in a short-term, mechanical sense – how does something work in one person, right now. But sometimes you mean how in a long-term transformational sense – involving a series of animal populations that change during species evolution. The physiological mechanisms underlying intelligent behavior are the proximate how; the prehistoric mechanisms that evolved our present brains are the other kind of how. You can sometimes ‘explain’ in one sense without even touching upon the other sense of how. Such a false sense of completeness is, of course, a good way to get blindsided.
Furthermore, there are different levels of explanation in both cases. Physiological how questions can be asked at a number of different levels of organization. Both consciousness and intelligence are at the high end of our mental life, but they are frequently confused with more elementary mental processes – with what we use to recognize a friend or tie a shoelace. Such simpler neural mechanisms are, of course, likely to be the foundations from which our abilities to handle logic and metaphor evolved.
Evolutionary how questions also have a number of levels of explanation: just saying that ‘a mutation did it’ isn’t likely to be a useful answer to an evolutionary question involving whole populations. Both physiological and evolutionary answers at multiple levels are needed if we are to understand our own intelligence in any detail. They might even help us appreciate how an artificial or an exotic intelligence could evolve – as opposed to creation from top-down design.
Everyone was admiring the bald eagles as our cruise ship slipped through the narrow passage at the top end of the Strait of Georgia, between Vancouver Island and the mainland of British Columbia. In one eagle nest after another, busy parents were feeding open mouths.
I was watching the raven, myself. It had found a clam and was trying to break open the shell to get at the innards, which were thus far successfully holding the two halves of the shell tightly together. It picked up the clam in its beak, flew several stories high, and dropped the clam on a rocky area of shoreline. This had to be repeated three times before the raven could settle down to pick his meal out of the shattered shell.
Was that instinctive behavior, or learned by observing others, or learned by trial and accidental success, or intelligently innovative? Did some ancestral raven contemplate the problem and then guess the solution? We have a difficult time seeing the intermediate steps between ‘reacting’ and ‘thinking,’ yet we also have an unwarranted faith that ‘more is better’ – that having more behavioral options is better than having fewer.
Nature is full of specialists that do one thing very well, with no frills – like a character actor who only plays one kind of role, never a repertoire. Most animals are specialists. The mountain gorilla, for example, processes fifty monotonous pounds of assorted greenery every single day. The panda’s diet is just as specialized.
In terms of finding what they like to eat, neither gorilla nor panda needs to be any smarter than a horse. Their ancestors may have needed to be intelligent in a different niche, but now the gorilla and the panda have each retreated into a niche that doesn’t require much intelligence. The same is true of the big-brained marine mammals we saw on the Alaskan cruise – animals that now make their living in more or less the same way as the small-brained fish that specialize in eating other fish.
In comparison, a chimpanzee has a varied diet: fruit, termites, leaves – even meat, when it’s lucky enough to catch a small monkey or piglet. So the chimp has to switch around a lot, and that means a lot of mental versatility. But what aids in building up a wide repertoire? One can be born with many movement programs, or learn them, or recombine existing ones in ways that cause novel behaviors to emerge suddenly. Omnivores, such as the octopus, crow, bear, and chimpanzee, have many ‘moves,’ simply because their ancestors had to switch among various food sources. Omnivores need a lot more sensory templates too – images and sounds they are in search of.
The other way to accumulate novel behaviors is through social life and play, in both of which new combinations can be discovered. A long life span ought to help both learned and innovative behaviors accumulate, and a long life span is what even the smartest of the invertebrates, the octopus, lacks. (The octopus is about as smart as a rat in some ways.) Smart animals have arisen from various branches of the vertebrate tree of species – ravens among the birds, marine mammals, bears, the primate line.
If specialization is most commonly the name of the game, however, then what selects for versatility? A fickle environment is one answer – an answer that highlights the environmental factor in natural selection. But let me start with another major contributor to sophistication: social life itself, which involves the sexual-selection aspect of natural selection.
Social intelligence is another aspect of intelligence: I refer not to just mimicry but to the challenges that social life (living in groups) poses – challenges that require innovative problem solving. The British psychologist Nicholas Humphrey, for one, considers social interaction, not tool use, to be of primary importance in hominid evolution.
Certainly a social life is an enormous facilitator of an expanded repertoire of actions. Some animals aren’t around others of their species long enough to partake of observational learning. Except for brief mating opportunities, adult orangutans seldom encounter one another, because their food sources are so sparse that it takes a large area to support a single adult. A mother with one offspring is about the biggest social group (except for the transient alliances formed by adolescent orangs), so there’s not much opportunity for cultural transmission.
Social life, besides facilitating the spread of new techniques, is also full of interpersonal problems to be solved, like pecking orders. You may need to hide food from the view of the dominant animal, in order to keep it for yourself. You need a lot of sensory templates to avoid confusing one individual with another, and a lot of memory to keep track of your past interactions with each of your colleagues. The challenges of social life go well beyond the usual environmental challenges to survive and reproduce that the solitary orang confronts. It would therefore seem that a social life is central to the cultural accumulation of ‘good moves’ – though I suspect nevertheless that a sociable dog lacks the mental potential of the solitary orang.
Natural selection for social intelligence may not involve the usual staying-alive factors commonly stressed in adaptationist arguments. The advantages of social intelligence would instead manifest themselves primarily via what Darwin called sexual selection. Not all adults pass on their genes. In harem-style mating systems, only a few males get the chance to mate, after having outsmarted or outpushed the others. In female-choice mating systems, acceptability as a social companion is likely to be important for males; for example, they need to be good at grooming, willing to share their food, and so forth. The male who can spot approaching estrus better than other males, and who can persuade the female to go off into the bushes with him for the duration of estrus, away from the other males, will stand a much better chance of passing on his genes, even in a promiscuous mating system. (And this female-choice bootstrap might improve more than just intelligence: I argue elsewhere that female choice would have been an excellent setup for improving language abilities, were a female to insist on male language ability at least as good as her own.)
[S]ocial primates are required by the very nature of the system they create and maintain to be calculating beings; they must be able to calculate the consequences of their own behaviour, to calculate the likely behaviour of others, to calculate the balance of advantage and loss – and all this in a context where the evidence on which their calculations are based is ephemeral, ambiguous and liable to change, not the least as a consequence of their own actions. In such a situation, ‘social skill’ goes hand in hand with intellect, and here at last the intellectual faculties required are of the highest order. The game of social plot and counter-plot cannot be played merely on the basis of accumulated knowledge . . . It asks for a level of intelligence which is, I submit, unparalleled in any other sphere of living.
Nicholas Humphrey, Consciousness Regained, 1984
The most frequent environmental stress likely to drive natural selection occurs in the temperate zones. Once a year, there is a period of a few months when plants are largely dormant. Eating grass (which stays nutritious even when dormant) is one strategy for getting through the winter. Another, which is much more demanding of versatile neural mechanisms, involves eating animals that eat grass. The extant wild apes all live very close to the equator; while they may have to cope with a dry season, it’s nothing like winter’s withdrawal of resources.
Climate change is the next most common recurring stress, seen even in the tropics: annual weather patterns shift into a new mode. Multiyear droughts are a familiar example, but sometimes they last for centuries or even millennia. In some cases, there are state-dependent modes of climate. We saw an example in Glacier Bay, just west of Juneau. When explorers passed the mouth of Glacier Bay 200 years ago, they reported that it was full of ice. Now the glaciers have retreated nearly 100 kilometers, and Glacier Bay is open to the sea once more. A series of large glaciers remain in the side valleys, and our ship maneuvered to within a respectful distance of one of these walls of ice: large blocks of it were breaking off and falling into the ocean, even as we watched.
In discussing the local glaciers with a geologist on board, I learned that some were advancing (those are the ones we were taken to see) but that others were in retreat. Advance and retreat at the same time, even in the same valley, and sharing the same climate? What’s going on here? I asked.
It’s as if a glacier can get stuck in ‘advance mode’ for centuries or millennia, even if the climate cools in the meantime. For example, meltwater from a few hot summers can get underneath and erode the craggy connections to the bedrock, and so the glacier, even if the melting were to stop, can slide downhill faster. That in turn causes the ice to fracture rather than flow when going over bumps, and so more vertical cracks open up. Any meltwater ponds on the surface can then drain down to the bedrock, further greasing the skids and accelerating the movement. The tall mountain of ice starts to collapse by spreading sideways. Eventually you may see glacial surges of the mile-a-month variety – but in Glacier Bay, the ice pushes into the ocean, which erodes it away in giant chunks, that in turn may float away to warmer climates to melt.
Later in the trip we saw Hubbard Glacier, a cliff of ice 5 kilometers long and taller than our ship. Great blocks of ice, loosened by the waves, would periodically crash into the sea. Off to the right side of Yakutat Bay, we could see back up Russell Fjord. Only a decade earlier, the entrance to that fjord was blocked by a surge in Hubbard Glacier. The glacier’s advance was faster than the waves could chip it away, so it crept past the mouth of the fjord and dammed it up. Water started rising behind the ice dam, threatening the trapped sea mammals as the salt water became increasingly diluted with the fresh meltwater. When the lake level got up to about two stories above sea level, the ice dam broke.
We know all about glacial surges in Washington State, because they blocked the Columbia River at least 59 times about 13,000 years ago; each time the ice dam broke, a wall of water went racing across the middle of Washington, carving the terrain into scabland as it surged to the sea. (Perhaps the ground-shaking roar warned anyone who was trying to catch salmon in the river valleys to run for the hills.)
Damming up a fjord may have had even more serious consequences. Fjords are often cut off by glacial surges, just as mountain valleys are temporarily dammed by the rubble that avalanches deposit. But dammed-up fjords serve as natural reservoirs for fresh water, and when the ice dam finally breaks, enormous quantities of fresh water flood into the adjacent oceans, a half-year’s worth in only a half-day’s time. It layers over the ocean surface and only later mixes with the salt water. Unfortunately, that freshening of the surface layer could have major consequences in the case of Greenland’s fjords: it is potentially a mechanism for shutting down for a few centuries the North Atlantic Current, which warms Europe – a subject to which I will shortly return.
I tell you all this to point out that there is an enormous asymmetry between the buildup of ice and its subsequent meltdown; this is not at all like the exchange of energy involved in freezing and melting a tray of ice cubes. Buildup mode keeps any cracks filled with new snow and minimizes the greasing of the skids. Melting mode is like a house of cards collapsing in slow motion.
‘Modes of operation’ are familiar to us from cool-fan-heat modes of air-conditioning systems. Not only do glaciers have modes, but so do ocean currents and continental climates – modes that may be triggered in some cases by glacial surges far away. Sometimes annual temperature and rainfall switch back and forth so rapidly that they have major implications for the evolutionary process, giving versatile animals, like the raven, a real advantage over their lean-mean-machine competitors. That’s what this chapter is really about: how the evolutionary crank is turned to yield our kind of versatility – wide repertoires and good guessing get a special kind of boost from a series of climatic instabilities.
Paleoclimatologists have discovered that many parts of the earth suffer fairly abrupt climate changes. Decade-long droughts are one example, and we now know something of the thirty-year cycle by which the Sahara expands and contracts. The El Niño cycle, averaging about six years, now appears to have major effects on North American rainfall.
There have also been dozens of episodes in which forests have disappeared over several decades because of drastic drops in temperature and rainfall. In another abrupt change, the warm rains suddenly return a few centuries later – although the last time that Europe reverted to a Siberian-style climate, more than a thousand years passed before it switched back.
In the 1980s, when confirming evidence of these abrupt climate changes was discovered, we thought they were a peculiarity of the ice ages. (Ice sheets have come and gone during the last 2:5 million years, the major meltoffs occurring about every 100,000 years.) None of the abrupt cooling episodes have occurred in the last 10,000 years.
But it turns out that it’s only our present interglaciation that has been free of them (so far). The warm times after the last major meltoff, 130,000 years ago, were turbulent in comparison to the present interglaciation; that earlier 10,000-year warm period was punctuated with two abrupt cold episodes. One lasted 70 years, the other 750 years. During them, the German pine forests were replaced with scrubs and herbs now characteristic of central Siberia.
We have thus far been spared such civilization-threatening episodes. Climatically, we have been living in unusually stable times.
A climate flip-flop that eliminated fruit trees would be a disaster for regional populations of many monkey species. While it would hurt the more omnivorous as well, they could ‘make do’ with other foods, and their offspring might enjoy a population boom following the crunch, when few competitors would remain.
Such boom times temporarily have enough resources so that most offspring can survive to reproductive age, and this is true even of the odd variants thrown up by the gene shuffles that produce sperm and ova. In ordinary times, such oddities die in childhood, but in a boom time they face little competition; it’s as if the usual competitive rules had been suspended temporarily. When the next crunch comes, some odd variants may have better abilities to ‘make do’ with whatever resources are left. The traditional theme extracted from the darwinian process is survival of the fittest, but here we see that it is the rebound from hard times that promotes the creative aspects of evolution.
Though Africa was cooling and drying as upright posture was becoming established in hominids about four million years ago, brain size didn’t change much. So far, there’s not much evidence that brains enlarged during the climate changes in Africa between 3.0 million and 2.6 million years ago – a period in which many new species of African mammals appeared. This isn’t the place for an extended discussion of all the factors involved in human evolution, but it is important to note that hominid brain size begins to increase between 2.5 million to 2.0 million years ago and continues for an amazing fourfold expansion in cerebral cortex over the apes. This is the period of the ice ages, and although Africa wasn’t a major site of glaciers, the continent probably experienced major fluctuations in climate as the ocean currents rearranged themselves. An ice age is not confined to the Northern Hemisphere; the glaciers in the Andes change at the same time.
The first major episodes of floating ice in the Atlantic occurred at 2.51 million and 2.37 million years ago, with the winter ice pack reaching south to British latitudes. Ice sheets in Antarctica, Greenland, northern Europe, and North America have been with us ever since, melting off occasionally. As noted, we are currently in an interglaciation period, which started about 10,000 years ago. There is a stately rhythm of ice advance and retreat, associated with changes in the earth’s axial tilt and its orbit around the sun.
The season of the earth’s closest approach to the sun varies; perihelion is currently in the first week of January. Perihelion drifts around the calendar, returning to January in 19,000 to 26,000 years, depending on where the other planets are. The configurations of the other planets approximately repeat about every 400,000 years, though they come close to repeating about every 100,000 year. Their gravitational pull causes the shape of the earth’s orbit to vary from near-circular to ellipsoidal. (We’re currently about 3 percent farther away from the sun in July, and so receive 7 percent less heat.) Moreover, the tilt of the earth’s axis varies between 22.0° and 24.6°, a cycle taking 41,000 years. The last maximum tilt was 9,500 years ago; it’s currently 23.4° and declining. The three rhythms combine to contribute to a really major meltoff about every 100,000 years, typically when tilt is maximal and perihelion is also in June; that creates particularly hot summers in the high northern latitudes, where most ice sheets are situated.
Superimposed on the glacial slowness are the aforementioned episodes of abrupt cooling and rewarming. The first one to be discovered happened at a time – 13,000 years ago – when all those orbital factors were combining to produce hot summers in the Northern Hemisphere – indeed, half of the accumulated ice had already melted. The Younger Dryas (named for an arctic plant whose pollen was found deep beneath old lakes in Denmark) began quite suddenly. Ice cores from Greenland’s ice sheets show that it was as sudden as a drought. Annual rainfall decreased, winter storms grew in severity, and the average European temperature dropped by about 7°C (13°F) – all within several decades. This cold snap lasted more than a thousand years until, just as abruptly, the warm rains returned. (With regard to global warming from greenhouse gases, note that the last time an abrupt cooling happened, it was during a major episode of gradual global warming).
The Greenland ice cores go back only one-tenth of the 2.5 million years of the Pleistocene ice ages; only ice from the last 250,000 years remains in Greenland, because the antepenultimate meltoff exposed all the bedrock. But the cores do record the last two major meltoffs – the one that began 130,000 years ago and the most recent one, which began 15,000 years ago and was complete about 8,000 years ago. Most important, one can see annual ‘tree rings’ in the more recent millennia and count the years, sample their oxygen isotopes, and thereby deduce the sea surface temperature at the time the water evaporated in the mid Atlantic before falling as snow in Greenland.
The paleoclimatologists now can see dozens of abrupt events in the last 130,000 years, superimposed upon the glacial slowness – and even occurring during warm periods. Big glacial surges could be one factor – as I discuss in The Ascent of Mind – simply because a lot of fresh water floating on the ocean surface before mixing may well cause major changes in the ocean current that imports a lot of heat into the North Atlantic and helps keep Europe warm in the winter. That’s why I worry about a glacial surge producing an enormous freshwater reservoir in Greenland’s fjords: it could all be released in a day when an ice dam is finally breached. The last time I flew over the extensive fjord system on the east coast of Greenland at 70° north latitude, I was appalled to see fjords that, though open to the ocean, had the bathtub-ring appearance of drawn-down reservoirs. There was an ice-free area extending far above the high-tide line, and everywhere it appeared to be the same height; this suggests an enormous freshwater lake formed sometime since the last ice age, uniformly trimming the ice sheet.
Another cold flip would be devastating to agriculture in Europe, and to the half billion people it supports – and the effects of the Younger Dryas were seen worldwide, even in Australia and Southern California. While another one would threaten civilizations, the cold flips of the past probably played an important role in evolving humans from our apelike ancestors, simply because the episodes happened so rapidly.
A round man cannot be expected to fit into a square hole right away. He must have time to modify his shape.
Mark Twain
Whether or not versatility is important during an animal’s lifespan depends on the timescales: for both the modern traveler and the evolving ape, it’s how fast the weather changes and how long the trip lasts. When the chimpanzees of Uganda arrive at a grove of fruit trees, they often discover that the efficient local monkeys are already speedily stripping the trees of edible fruit. The chimps can turn to termite fishing, or perhaps catch a monkey and eat it, but in practice their population is severely limited by that competition, despite the fact that the chimpanzee’s brain is twice the size of that of its specialist rivals.
Versatility is not always a virtue, and more of it is not always better. As frequent airline travelers know, passengers who have only carry-on bags can get all the available taxicabs, while those burdened with three suitcases await their checked luggage. On the other hand, if the weather is so unpredictable and extreme that everyone has to travel with clothing ranging from swimsuits to Arctic parkas, the jack-of-all-trades has an advantage over the master of one. And so it is with behavioral versatility that allows a species to switch instantly from round to square holes.
Versatility might well require a bigger brain. But you need some pretty good reasons to balance out the disadvantages of a big brain. As the linguist Steven Pinker has noted:
Why would evolution ever have selected for sheer bigness of brain, that bulbous, metabolically greedy organ? A large-brained creature is sentenced to a life that combines all the disadvantages of balancing a watermelon on a broomstick . . . and, for women, passing a large kidney stone every few years. Any selection on brain size itself would have surely favored the pinhead. Selection for more powerful computational abilities (language, perception, reasoning, and so on) must have given us a big brain as a by-product, not the other way around!
How fast things change is important for any incremental-accumulations model of intelligence, whether it involves a bigger brain or merely a rearranged one. In any one climate, a specialist can eventually evolve that outperforms the overburdened generalist; however, anatomical adaptations occur much more slowly than the frequent climatic changes of the ice ages, making it hard for adaptations to ‘track’ the climate. Indeed, the abrupt transitions can occur within the lifetime of a single individual, who either has the reserve abilities needed to survive the crunch, or doesn’t.
This sudden-death-overtime argument applies to many omnivores, not just to our ancestors. But there aren’t any other examples around of fourfold brain enlargements in the last several million years, so an erratic climate by itself isn’t a surefire way of getting a swelled head. Something else was also going on, and the episodes of abrupt climate change probably exaggerated its importance, and kept those lean-mean-machine competitors from outcompeting the jack-of-all-trades types that evolved.
Everyone has a favorite theory for what this ‘something else’ was. (Nick Humphrey would pick social intelligence as the driver, for example.) My candidate is accurate throwing for hunting – handy for getting through the winter by eating animals that eat grass. But most people would pick language. Especially syntax.
[Language comprehension] involves many components of intelligence: recognition of words, decoding them into meanings, segmenting word sequences into grammatical constituents, combining meanings into statements, inferring connections among statements, holding in short-term memory earlier concepts while processing later discourse, inferring the writer’s or speaker’s intentions, schematization of the gist of a passage, and memory retrieval in answering questions about the passage . . . [The reader] constructs a mental representation of the situation and actions being described . . . Readers tend to remember the mental model they constructed from a text, rather than the text itself.
Gordon H. Bower and Daniel G. Morrow, 1990
I often find that a novel, even a well-written and compelling novel, can become a blur to me soon after I’ve finished it. I recollect perfectly the feeling of reading it, the mood I occupied, but I am less sure about the narrative details. It is almost as if the book were, as Wittgenstein said of his propositions, a ladder to be climbed and then discarded after it has served its purpose.
Sven Birkerts, 1994
CHAPTER 5
SYNTAX AS A FOUNDATION OF INTELLIGENCE
It is hard to imagine how a creature without language would think, but one may suspect that a world without any kind of language would in some ways resemble a world without money – a world in which actual commodities, rather than metal or paper symbols for the value of these, would have to be exchanged. How slow and cumbersome the simplest sale would be, and how impossible the more complex ones!
Derek Bickerton, Language and Species, 1990
Humans have some spectacular abilities, compared to our closest cousins among the surviving apes – even the apes that share much of our social intelligence, reassuring touches, and abilities to deceive. We have a syntactic language capable of supporting metaphor and analogical reasoning. We’re always planning ahead, imagining scenarios for the future, and then choosing in ways that take remote contingencies into account. We even have music and dance. What were the steps in transforming a chimpanzeelike creature into a nearly human one? That’s a question which is really central to our humanity.
There’s no doubt that syntax is what human levels of intelligence are mostly about – that without syntax we would be little cleverer than chimpanzees. The neurologist Oliver Sacks’s description of an eleven-year-old deaf boy, reared without sign language for his first ten years, shows what life is like without syntax:
Joseph saw, distinguished, categorized, used; he had no problems with perceptual categorization or generalization, but he could not, it seemed, go much beyond this, hold abstract ideas in mind, reflect, play, plan. He seemed completely literal – unable to juggle images or hypotheses or possibilities, unable to enter an imaginative or figurative realm . . . He seemed, like an animal, or an infant, to be stuck in the present, to be confined to literal and immediate perception, though made aware of this by a consciousness that no infant could have.
Similar cases also illustrate that any intrinsic aptitude for language must be developed by practice during early childhood. Joseph didn’t have the opportunity to observe syntax in operation during his critical years of early childhood: he couldn’t hear spoken language, nor he was ever exposed to the syntax of sign language.
There is thought to be a bioprogram, sometimes called Universal Grammar. It is not the mental grammar itself (after all, each dialect has a different one) but rather the predisposition to discover grammars in one’s surroundings – indeed, particular grammars, out of a much larger set of possible ones. To understand why humans are so intelligent, we need to understand how our ancestors remodeled the ape’s symbolic repertoire and enhanced it by inventing syntax.
Stones and bones are, unfortunately, about all that remain of our ancestors in the last four million years, not their higher intellectual abilities. Other species branched off along the way, but they are no longer around to test. We have to go back six million years before there are living species with whom we shared a common ancestor: the nonhominid branch itself split about three million years ago into the chimpanzee and the much rarer bonobo (the ‘chimpanzee of the Pygmies’). If we want a glimpse at ancestral behaviors, the bonobos are our best chance. They share more behavioral similarities with humans and they’re also much better subjects for studying language than the chimps that starred in the sixties and seventies.
Linguists have a bad habit of claiming that anything lacking syntax isn’t language. That, ahem, is like saying that a Gregorian chant isn’t music, merely because it lacks Bach’s use of the contrapuntal techniques of stretto, parallel voice leading, and mirror inversions of themes. Since linguistics confines itself to ‘Bach and beyond,’ it has primarily fallen to the anthropologists, the ethologists, and the comparative psychologists to be the ‘musicologists,’ to grapple with the problem of what came before syntax. The linguists’ traditional put-down of all such research (‘It isn’t really language, you know’) is a curious category error, since the object of the research is to understand the antecedents of the powerful structuring that syntax provides.
One occasionally gets some help from the well-known ontogeny-recapitulates-phylogeny crib, but human language is acquired so rapidly in early childhood that I suspect a streamlining, one that thoroughly obscures any original staging, rather as freeways tend to obliterate post roads. The fast track starts in infants with the development of phoneme boundaries: prototypes become ‘magnets’ that capture variants. Then there’s a pronounced acquisitiveness for new words in the second year, for inferring patterns of words in the third (kids suddenly start to use past tense -ed and plural -s with consistency, a generalization that occurs without a lot of trial and error), and for narratives and fantasy by the fifth. It is fortunate for us that chimps and bonobos lack such fast-tracking, because it gives us a chance to see, in their development, the intermediate stages that were antecedent to our powerful syntax.
Vervet monkeys in the wild use four different alarm calls, one for each of their typical predators. They also have other vocalizations to call the group together or to warn of the approach of another group of monkeys. Wild chimpanzees use about three dozen different vocalizations, each of them, like those of the vervets, meaningful in itself. A chimp’s loud waa-bark is defiant, angry. A soft cough-bark is, surprisingly, a threat. Wraaa mixes fear with curiosity (‘Weird stuff, this!’) and the soft huu signifies weirdness without hostility (‘What is this stuff?’).
If a waa-wraaa-huu is to mean something different than huu-wraaa-waa, the chimp would have to suspend judgment, ignoring the standard meanings of each call until the whole string had been received and analyzed. This doesn’t happen. Combinations are not used for special meanings.
Humans also have about three dozen units of vocalization, called phonemes – but they’re all meaningless! Even most syllables like ‘ba’ and ‘ga’ are meaningless unless combined with other phonemes to make meaningful words, like ‘bat’ or ‘galaxy.’ Somewhere along the line, our ancestors stripped most speech sounds of their meaning. Only the combinations of sounds now have meaning: we string together meaningless sounds to make meaningful words. That’s not seen anywhere else in the animal kingdom.
Furthermore, there are strings of strings – such as the word phrases that make up this sentence – as if the principle were being repeated on yet another level of organization. Monkeys and apes may repeat an utterance to intensify its meaning (as do many human languages, such as Polynesian), but nonhumans in the wild don’t (so far) string together different sounds to create entirely new meanings.
No one has yet explained how our ancestors got over the hump of replacing one-sound/one-meaning with a sequential combinatorial system of meaningless phonemes, but it’s probably one of the most important transitions that happened during ape-to-human evolution.
The honeybee appears, at least in the context of a simple coordinate system, to have broken out of the mold of one-sign/one-meaning. When she returns to her hive, she performs a ‘waggle dance’ in a figure-8 that communicates information about the location of a food source she has just visited. The angle of the figure-8 axis points toward the food. The duration of the dance is proportional to the distance from the hive: for example, at least in the traditional version of this story, three loops around the figure-8 would suggest 60 meters away to the average Italian honeybee, though 150 meters to a German one – a matter of genes rather than the company in which the bee was reared. Still, the linguists are not very impressed – in his Language and Species, Derek Bickerton notes:
All other creatures can communicate only about things that have evolutionary significance for them, but human beings can communicate about anything . . . Animal calls and signs are structurally holistic [and] cannot be broken down into component parts, as language can . . . Though in themselves the sounds of [human] language are meaningless, they can be recombined in different ways to yield thousands of words, each distinct in meaning . . . In just the same way, a finite stock of words . . . can be combined to produce an infinite number of sentences. Nothing remotely like this is found in animal communication.
With enough experience, various animals can learn a wide range of words, symbols, or human gestures – but one must be careful to distinguish between comprehension and the ability to originate fancy communications. They don’t necessarily go together.
One psychologist’s dog, as I noted earlier, understands about 90 items; the 60 it produces don’t overlap very much in meaning with the receptive ones. A sea lion has learned to comprehend 190 human gestures – but it doesn’t gesture back with anything near the same productivity. Bonobos have learned an even greater number of symbols for words and can combine them with gestures to make requests. A gray parrot has learned, over the course of a decade, a 70-word vocabulary that includes thirty object names, seven colors, five shape adjectives, and a variety of other ‘words’ – and can make requests with some of them.
None of these talented animals is telling stories about who did what to whom; they’re not even discussing the weather. But it is clear that our closest cousins, the chimpanzee and the bonobo, can achieve considerable levels of language comprehension with the aid of skilled teachers who can motivate them. The most accomplished bonobo, under the tutelage of Sue Savage-Rumbaugh and her co-workers, can now interpret sentences it has never heard before – such as ‘Kanzi, go to the office and bring back the red ball’ – about as well as a two-and-a-half-year-old child. Neither bonobo nor child is constructing such sentences, but they can demonstrate by their actions that they understand them. And comprehension comes first, production later, as language develops in children.
I often wonder how many of the limited successes in ape language studies were merely a matter of insufficient motivation; perhaps teachers have to be good enough to substitute for the normal self-motivating acquisitiveness of the young child. Or if the limited successes were from not starting with very young animals. If a bonobo could somehow become motivated in its first two years to comprehend new words at a rate approaching that of the year-old child, might the bonobo then go on to discover patterning of words in the manner of the pre-syntax child? But have it happen slowly enough for us to see distinct stages preceding serious syntax, the ones obscured by the streamlined freeways provided by the present human genome?
All of this animal communicative ability is very impressive, but is it language? The term language is used rather loosely by most people. First of all, it refers to a particular dialect such as English, Frisian, and Dutch (and the German of a thousand years ago, from which each was derived – and, further back, proto-Indo-European). But language also designates the overarching Category of communication systems that are especially elaborate. Bee researchers use language to describe what they see their subjects doing, and chimpanzee researchers do the same. At what point do animal symbolic repertoires become humanlike language?
The answer isn’t obvious. Webster’s Collegiate Dictionary offers ‘a systematic means of communicating ideas or feelings by use of conventionalized signs, sounds, gestures, or marks having understood meanings’ as one definition of language. That would encompass the foregoing examples. Sue Savage-Rumbaugh suggests that the essence of language is ‘the ability to tell another individual something he or she did not already know,’ which, of course, means that the receiving individual is going to have to use some Piagetian guessing-right intelligence in constructing a meaning.
But humanlike language? Linguists will immediately say ‘No, there are rules!’ They will start talking about the rules implied by mental grammar and questioning whether or not these rules are found in any of the nonhuman examples. That some animals such as Kanzi can make use of word order to disambiguate requests does not impress them. The linguist Ray Jackendoff is more diplomatic than most, but has the same bottom line:
A lot of people have taken the issue to be whether the apes have language or not, citing definitions and counter-definitions to support their position. I think this is a silly dispute, often driven by an interest either in reducing the distance between people and animals or in maintaining this distance at all costs. In an attempt to be less doctrinaire, let’s ask: do the apes succeed in communicating? Undoubtedly yes. It even looks as if they succeed in communicating symbolically, which is pretty impressive. But, going beyond that, it does not look as though they are capable of constructing a mental grammar that regiments the symbols coherently. (Again, a matter of degree – maybe there is a little, but nothing near human capacity.) In short, Universal Grammar, or even something remotely like it, appears to be exclusively human.
What, if anything, does this dispute about True Language have to do with intelligence? Judging by what the linguists have discovered about mental structures and the ape-language researchers have discovered about bonobos inventing rules – quite a lot. Let us start simple.
Some utterances are so simple that fancy rules aren’t needed to sort out the elements of the message – most requests such as ‘banana’ and ‘give’ in either sequence get across the message. Simple association suffices. But suppose there are two nouns in a sentence with one verb: how do we associate ‘dog boy bite’ in any order? Not much mental grammar is needed, as boys usually don’t bite dogs. But ‘boy girl touch’ is ambiguous without some rule to help you decide which noun is the actor and which is the acted upon.
A simple convention can decide this, such as the subject-verb-object order (SVO) of most declarative sentences in English (‘The dog bit the boy’) or the SOV of Japanese. In short word phrases, this boils down to the first noun being the actor – a rule that Kanzi probably has absorbed from the way that Savage-Rumbaugh usually phrases requests, such as ‘Touch the ball to the banana.’
You can also tag the words in a phrase in order to designate their role as subject or object, either with conventional inflections or by utilizing special forms called case markings – as when we say ‘he’ to communicate that the person is the subject of the sentence, but ‘him’ when he is the object of the verb or preposition. English once had lots of case markings, such as ‘ye’ for subject and ‘you’ for object, but they now survive mostly in the personal pronouns and in ‘who’/‘whom.’ Special endings can also tip you off about a word’s role in the phrase, as when -ly suggests to you that ‘softly’ modifies the verb and not a noun. In highly inflected languages, such markings are extensively used, making word order a minor consideration in identifying the role a word is intended to play in constructing the mental model of relationships.
[For] us to be able to speak and understand novel sentences, we have to store in our heads not just the words of our language but also the patterns of sentences possible in our language. These patterns, in turn, describe not just patterns of words but also patterns of patterns. Linguists refer to these patterns as the rules of language stored in memory; they refer to the complete collection of rules as the mental grammar of the language, or grammar for short.
Ray Jackendoff, Patterns in the Mind, 1994
The simpler ways of generating word collections, such as pidgins (or my tourist German), are what the linguist Derek Bickerton calls protolanguage. They don’t utilize much in the way of mental rules. The word association (‘boy dog bite’) carries the message, perhaps with some aid from customary word order such as SVO. Linguists would probably classify the ape language achievements, both comprehension and production, as protolanguage.
Children learn a mental grammar by listening to a language (deaf children by observing sign language). They are acquisitive of associations as well as new words, and one fancy set of associations constitutes the mental grammar of a particular language. Starting at about eighteen months of age, children start to figure out the local rules and eventually begin using them in their own sentences. They may not be able to describe the parts of speech, or diagram a sentence, but their ‘language machine’ seems to know all about such matters after a year’s experience.
This biological tendency to discover and imitate order is so strong that deaf playmates may invent their own sign language (‘home sign’) with inflections, if they aren’t properly exposed to one they can model. Bickerton showed that the children of immigrants invent a new language – a creole – out of the pidgin protolanguage they hear their parents speaking. A pidgin is what traders, tourists, and ‘guest workers’ (and, in the old days, slaves) use to communicate when they don’t share a real language. There’s usually a lot of gesturing, and it takes a long time to say a little, because of all those circumlocutions.
In a proper language with lots of rules (that mental grammar), you can pack a lot of meaning into a short sentence. Creoles are indeed proper languages: the children of pidgin speakers take the vocabulary they hear and create some rules for it – a mental grammar. The rules aren’t necessarily any of those they know from simultaneously learning their parents’ native languages. And so a new language emerges, from the mouths of children, as they quickly describe who did what to whom.
Which aspects of language are easy to acquire and which are difficult? Broad categories may be the easiest, as when the child goes through a phase of designating any four-legged animal as ‘doggie’ or any adult male as ‘Daddy.’ Going from the general to the specific is more difficult. But some animals, as we have seen, can eventually learn hundreds of symbolic representations.
A more important issue may be whether new categories can be created that escape old ones. The comparative psychologist Duane Rumbaugh notes that prosimians (lorises, galagos, and so forth) and small monkeys often get trapped by the first set of discrimination rules they are taught, unlike rhesus monkeys and apes, both of which can learn a new set of rules that violates the old one. We too can overlay a new category atop an old one, but it is sometimes difficult: categorical perception (the pigeonholing mentioned earlier, in association with auditory hallucinations) is why some Japanese have such a hard time distinguishing between the English sounds for L and R.
The Japanese language has an intermediate phoneme, a neighbor to both L and R. Those English phonemes are, mistakenly, treated as mere variants on the Japanese phoneme. Because of this ‘capture’ by the traditional category, those Japanese speakers who can’t hear the difference will also have trouble pronouncing them distinctly.
Combining a word with a gesture is somewhat more sophisticated than one-word, one-meaning – and putting a few words together into a string of unique meaning is considerably more difficult. Basic word order is helpful in resolving ambiguities, as when you can’t otherwise tell which noun is the actor and which the acted upon. The SVO declarative sentence of English is only one of the six permutations of those units, and each permutation is found in some human language. Some word orders are more frequently found than others, but the variety suggests that word order is a cultural convention rather than a biological imperative in the manner proposed for Universal Grammar.
Words to indicate points in time (‘tomorrow’ or ‘before’) require more advanced abilities, as do words that indicate a desire for information (‘What’ or ‘Are there’) and the words for possibility (‘might’ or ‘could’). It is worthwhile noting what a pidgin protolanguage lacks: it doesn’t use articles like ‘a’ or ‘the’ which help you know whether a noun refers to a particular object or simply the general class of objects. It doesn’t use inflections (-s, -ly, and the like) or subordinate clauses, and it often omits the verb, which is guessed from the context.
Though they take time to learn, vocabulary and basic word order are nonetheless easier than the other rule-bound parts of language. Indeed, in the studies of Jacqueline S. Johnson and Elissa L. Newport, Asian immigrants who learn English as adults succeed with vocabulary and basic word-order sentences, but have great difficulty with other tasks – tasks that those who arrived as children easily master. At least in English, the who-what-where-when-why-how questions deviate from basic word order: ‘What did John give to Betty?’ is the usual convention (except on quiz shows in which questions mimic the basic word order and use emphasis instead: ‘John gave what to Betty?’). Nonbasic word orders in English are difficult for those who immigrated as adults, and so are other long-range dependencies, such as when plural object names must match plural verbs, despite many intervening adjectives. Not only do adult immigrants commit such grammatical errors, but they can’t detect errors when they hear them. For example, the inflectional system of English alters a noun when it refers to a multiplicity (‘The boy ate three cookie.’ Is that normal English?) and alters a verb when it refers to past time (‘Yesterday the girl pet a dog.’ OK?). Those who arrived in the United States before the age of seven make few such recognition errors as adults, and there is a steady rise in error rate for adults who began learning English between the ages of seven and fifteen – at which age the adult error level is reached (I should emphasize that the linguists were, in all cases, testing immigrants with ten years’ exposure to English, who score normally on vocabulary and the interpretation of basic word-order sentences).
By the age of two or three, children learn the plural rule: add -s. Before that, they treat all nouns as irregular. But even if they had been saying ‘mice,’ once they learn the plural rule they may begin saying ‘mouses’ instead. Eventually they learn to treat the irregular nouns and verbs as special cases, exceptions to the rules. Not only are children becoming acquisitive of the regular rules at about the time of their second birthday but it also appears that the window of opportunity is closing during the school years. It may not be impossible to learn such things as an adult, but simple immersion in an English-language society does not work for adults in the way that it does for children aged two to seven.
Whether you want to call it a bioprogram or a Universal Grammar, learning the hardest aspects of language seems to be made easier by a childhood acquisitiveness that has a biological basis, just as does learning to walk upright. Perhaps this acquisitiveness is specific to language, perhaps it merely looks for intricate patterns in sound and sight and learns to mimic them. A deaf child like Joseph who regularly watched chess games might, for all we know, discover chess patterns instead. In many ways, this pattern-seeking bioprogram looks like an important underpinning for human levels of intelligence.
A dictionary will define the word grammar for you as (1) morphology (word forms and endings); (2) syntax (from the Greek ‘to arrange together’ – the ordering of words into clauses and sentences); and (3) phonology (speech sounds and their arrangements). But just as we often use grammar loosely to refer to socially correct usage, the linguists sometimes go to the opposite extreme, using overly narrow rather than overly broad definitions. They often use grammar to specify just a piece of the mental grammar – all the little helper words, like ‘near,’ ‘above,’ and ‘into,’ that aid in communicating such information as relative position. Whatever words like these are called, they too are quite important for our analysis of intelligence.
First of all, such grammatical items can express relative location (above, below, in, on, at, by, next to) and relative direction (to, from, through, left, right, up, down). Then there are the words for relative time (before, after, while, and the various indicators of tense) and relative number (many, few, some, the -s of plurality). The articles express a presumed familiarity or unfamiliarity (the for things the speaker thinks the hearer will recognize, a or an for things the speaker thinks the hearer won’t recognize) in a manner somewhat like pronouns. Other grammatical items in Bickerton’s list express relative possibility (can, may, might), relative contingency (unless, although, until, because), possession (of, the possessive version of -s, have), agency (by), purpose (for), necessity (must, have to), obligation (should, ought to), existence (be), nonexistence (no, none, not, un-), and so on. Some languages have verbal inflections that indicate whether you know something on the basis of personal experience or just at second hand.
So grammatical words help to position objects and events relative to each other on a mental map of relationships. Because relationships (‘bigger,’ ‘faster,’ and so forth) are what analogies usually compare (as in ‘bigger-is-faster’), this positioning-words aspect of grammar could also augment intelligence.
Syntax is a treelike structuring of relative relationships in your mental model of things which goes far beyond conventional word order or the aforementioned ‘positioning’ aspects of grammar. By means of syntax, a speaker can quickly convey a mental model to a listener of who did what to whom. These relationships are best represented by an upside-down tree structure – not the sentence diagraming of my high school days, but a modern version of diagraming known as an X-bar phrase structure. Since there are now some excellent popular books on this subject, I will omit explaining the diagrams here (whew!).
Treelike structure is most obvious when you consider subordinate clauses, as in the rhyme about the house that Jack built. (‘This is the farmer sowing the corn/ That kept the cock that crowed in the morn/ . . . That lay in the house that Jack built.’) Bickerton explains that such nesting or embedding is possible because
phrases are not, as they might appear to be, strung together serially, like beads on a string. Phrases are like Chinese boxes, stacked inside one another. The importance of this point can hardly be overestimated. Many people concerned with the origins of human language, or with the alleged language capacities of nonhuman species, have been led to propose grossly simplistic hypotheses about how language could have emerged, simply on the basis of a mistaken assumption. They assume that words are serially chained into phrases and phrases into sentences in pretty much the same way that steps are chained into walking. . . . Nothing could be further from the truth. . . . This can be seen by considering a phrase like the cow with the crumpled horn that Farmer Giles likes. Although no single word in this phrase is ambiguous, the phrase as a whole is, because we do not know whether it is the horn or the cow that Farmer Giles likes.
In addition to such ‘phrase structure’ (as this is called), there is ‘argument structure,’ which is particularly helpful in guessing the role of the various nouns in the sentence. If you see an intransitive verb, such as ‘sleep,’ you can be sure that one noun (or pronoun) will suffice to complete the thought – namely, the actor. This will be true in any language with a word for sleeping. Similarly, if a language has a verb meaning ‘beat,’ you can be sure that two nouns are involved, an actor and a recipient (and perhaps a third, for the instrument with which the beating is administered). A verb meaning ‘give’ calls for three nouns, as it also requires an item that is given to the recipient. So, any mental organization chart featuring ‘give’ will have three empty slots, which must be appropriately filled before you feel that you correctly ‘understand’ the sentence and can proceed to the next task. Sometimes the nouns are implicit, as in the exhortation ‘Give!’ where we fill in ‘you,’ ‘money,’ and ‘to me’ automatically.
As Bickerton notes, a sentence is like
a little play or story, one in which each of the characters has a specific role to perform. There is a finite and indeed very short list of these roles. Not all linguists are agreed as to exactly what they are, but most, if not all, would include the roles of Agent (JOHN cooked dinner), Patient or Theme (John cooked DINNER), Goal (I gave it TO MARY), Source (I bought it FROM FRED), Instrument (Bill cut it WITH A KNIFE), and Beneficiary (I bought it FOR YOU), as well as Time and Place.
No animal language in the wild has such structural features. At best, wild animal languages amount to a few dozen utterances and associated intensifiers (usually involving repetition, as in the circuits of the waggle dance or the repeats of a primate alarm cry), with combinations of utterances rarely utilized for new message types. With education, some animals have come to understand a consistent word order, so that they correctly respond to ‘Kanzi, touch the banana to the ball,’ in which word order is used to distinguish the actor from the acted upon.
Linguists, however, would like to place the language boundary well beyond such sentence comprehension: in looking at animal experiments, they want to see sentence production using a mental grammar; mere comprehension, they insist, is too easy. Though guessing at meaning often suffices for comprehension, the attempt to generate and speak a unique sentence quickly demonstrates whether or not you know the rules well enough to avoid ambiguities.
Yet that production test is more relevant to the scientist’s distinctions than those of the language-learner’s: after all. comprehension comes first in children. The original attempts to teach chimps the manual sign language of the deaf involved teaching the chimp how to produce the desired movements; comprehension of what the sign signified only came later, if at all. Now that the ape-language research has finally addressed the comprehension issue, it looks like more of a hurdle than anyone thought – but once an animal is past it, spontaneous production increases.
Linguists aren’t much interested in anything simpler than real rules, but ethologists and the comparative and developmental psychologists are. Sometimes, to give everyone a piece of the action, we talk about languages plural, ‘language’ in the sense of systematic communication, and Language with a capital L for the utterances of the advanced-syntax-using élite. All aid in the development of versatility and speed (and hence intelligence). While morphology and phonology also tell us something about cognitive processes, phrase structure, argument structure, and the relative-position words are of particular interest because of their architectural aspect – and that provides some insights about the mental structures available for the guessing-right type of intelligence.
Comprehension demands an active intellectual process of listening to another party while trying to figure out, from a short burst of sounds, the other’s meaning and intent – both of which are always imperfectly conveyed. Production, by contrast, is simple. We know what we think and what we wish to mean. We don’t have to figure out ‘what it is we mean,’ only how to say it. By contrast, when we listen to someone else, we not only have to determine what the other person is saying, but also what he or she means by what is said, without the insider’s knowledge that the speaker has.
Sue Savage-Rumbaugh, 1994
How much of language is innate in humans? Certainly the drive to learn new words via imitation is probably innate in a way that a drive to learn arithmetic is not. Other animals learn gestures by imitation, but preschool children seem to average ten new words every day – a feat that puts them in a whole different class of imitators. And they’re acquiring important social tools, not mere vocabulary: the right tool for the job, the British neuropsychologist Richard Gregory emphasizes, confers intelligence on its user – and words are social tools. So this drive alone may account for a major increment in intelligence over the apes.
There is also the drive of the preschooler to acquire the rules of combination we call mental grammar. This is not an intellectual task in the usual sense: even children of low-average intelligence seem to effortlessly acquire syntax by listening. Nor is acquisition of syntax a result of trial and error, because children seem to make fairly fast transitions into syntactic constructions. Learning clearly plays a role but some of the rigidities of grammar suggest innate wiring. As Derek Bickerton points out, our ways of expressing relationships (such as all those above/below words) are resistant to augmentation, whereas you can always add more nouns. Because of regularities across languages in the errors made by children just learning to speak, because of the way various aspects of grammar change together across languages (SVO uses prepositions such as ‘by bus,’ SOV tends to post-positions such as ‘bus by’), because of those adult Asian immigrants, and because of certain constructions that seem forbidden in any known language, linguists such as Noam Chomsky have surmised that something biological is involved – that the human brain comes wired for the treelike constructions needed for syntax, just as it is wired for walking upright:
Normal speech consists, in large part, of fragments, false starts, blends and other distortions of the underlying idealized forms. Nevertheless . . . what the child learns is the underlying [idealized form]. This is a remarkable fact. We must also bear in mind that the child constructs this [idealized form] without explicit instruction, that he acquires this knowledge at a time when he is not capable of complex intellectual achievements in many other domains, and that this achievement is relatively independent of intelligence.
There is, of course, a ‘language module’ in the brain – located just above the left ear in most of us – and Universal Grammar might be wired into it at birth. Monkeys lack this left lateral language area: their vocalizations (and the emotional utterances of humans) utilize a more primitive cortical speech area above the corpus callosum. Nobody knows yet whether apes have a lateral language area or similar arrangement.
If a young bonobo or chimpanzee had the two drives that young human children have – to seek out words and discover rules – in sufficient intensity and at the right time in brain development, would it self-organize a language cortex like ours and use it to crystallize a set of rules out of word mixtures? Or is that neural wiring innate in humans, there without the relevant experience and simply unused if the drives or opportunities are missing? Either, it seems to me, is consistent with the Chomskian claim. Universal Grammar might result from the ‘crystallization’ rules of the self-organization, arising just as ‘flashers’ and ‘gliders’ do from cellular automata.
And the way you experimentally distinguish between uniquely human innate wiring and input-driven crystallization is to push vocabulary and sentences on promising ape students, with clever motivation schemes attempting to substitute for the child’s untutored acquisitiveness. It is, I think, fortunate that the apes are borderline when it comes to having the linguists’ True Language, because by studying their struggles we might eventually glimpse the functional foundations of mental grammar. In the course of human evolution, the stepping-stones may have been paved over, overlaid by super-structures and streamlined beyond recognition.
Sometimes ontogeny recapitulates phylogeny (the baby’s attempts to stand up recapitulating the quadruped-to-biped phylogeny; the descent of the larynx in the baby’s first year partially recapitulating the ape-to-human changes). Yet, development can happen so rapidly that you fail to see the reenactment of evolutionary progress. If we could see the transition to fancier constructions in bonobos, however, we might be able to discover what sorts of learning augment syntax, what other tasks compete and so hinder language, and what brain areas ‘light up’ in comparison to those in humans. Besides the major implications for our view of what’s uniquely human, an understanding of ape linguistic foundations might help us teach the language-impaired, and might even reveal synergies that would aid language learning and better guessing. It is only through the efforts of skilled teachers of bonobos that we’re likely to answer questions about the stepping-stones.
Syntax is what you use, it would appear, to make those fancier mental models, the ones involving who did what to whom, why, when, and with what means. Or at least if you want to communicate such an elaborate understanding, you’ll have to translate your mental model of those relationships into the mental grammar of the language, then you order or inflect the words to help the listener reconstruct your mental model. It might, of course, be easier just to ‘think in syntax’ in the first place. In that sense, we’d expect the augmentation of syntax to result in a great augmentation of guessing-right intelligence.
The name of the game is to re-create your mental model in the listener’s mind. The recipient of your message will need to know the same mental grammar, in order to decode the string of words into approximately the same mental understanding. So syntax is about structuring relationships between items (words, usually) in your underlying mental model, not about the surface of things – such as SVO or inflections, which are mere clues. Your job as a listener is to figure out what sort of tree will provide an adequate fit to the string of words you hear. (Imagine being sent the numerical values for a spreadsheet and having to guess the spreadsheet formulas needed to interrelate them!)
The way this could work is that you try out a simple arrangement (actor, action, acted-upons, modifiers) and wind up with words left over. You try another tree, and discover that there are unfilled positions that can’t be left empty. You use those clues about tree structure provided by the speaker’s plurals and verbs – for example, you know that ‘give’ requires both a recipient and an item given. If there is no word (spoken or implied) to fill a required slot, then you scratch that tree and go on to yet another. Presumably you try a lot of different trees simultaneously rather than seriatim, because comprehension (finding a good enough interpretation of that word string) can operate at blinding speed.
In the end, several trees may fill properly with no words left over, so you have to judge which of the interpretations is most reasonable given the situation you’re in. Then you’re finished. That’s comprehension – at least in my (surely oversimplified) version of the linguists’ model.
Think in terms of a game of solitaire: you’re not finished until you have successfully uncovered all the face-down cards – while following the rules about descending order and alternate colors – and in some shuffles of the deck, it is impossible to succeed. You lose that round, shuffle the deck, and try again. For some word strings, no amount of rearranging will find a meaningful set of relationships – a story you can construct involving who did what to whom. If someone utters such an ambiguous word string to you, they’ve failed an important test of language ability.
For some sentences generated by a linguistically competent human, you have the opposite problem: you can construct multiple scenarios – alternative ways of understanding the word string. Generally, one of the candidates will satisfy the conventions of the language or the situation better than others, thus becoming the ‘meaning’ of the communication. Context creates default meanings for some items in the sentence, saving the speaker from producing a longer utterance. (Pronouns are such a shortcut.)
The kind of formal rules of compositional correctness you learned in high school are, in fact, violated all the time in the incomplete utterances of everyday speech. But everyday speech suffices, because the real test is in whether you convey your mental model of who did what to whom to your audience, and the context will usually allow the listener to fill in the missing pieces. Because a written message has to function without much of the context, and without such feedback as an enlightened or puzzled look on the listener’s face, we have to be more complete – indeed, more redundant – when writing than when speaking, making fuller use of syntax and grammatical rules.
Linguists would like to understand how sentences are generated and comprehended in a machinelike manner – what enables the blinding speed of sentence comprehension. I like to call this ‘language machine’ a lingua ex machina. That does, of course, invite comparison with the deus ex machina of classical drama – a platform wheeled on stage (the god machine), from which a god lectured the other actors, and more recently the term bestowed on any contrived resolution of a plot difficulty. Until our ‘playwrighting’ technique improves, our algorithms for understanding sentences will also seem contrived.
I’m going to propose how one such lingua ex machina could work, combining phrase structure and argument structure in an algorithmic way. Linguists will probably find it at least as contrived as other diagraming systems. But here’s a few paragraphs’ worth of Calvin’s Vacuum-Lifter Package-Carrying System, involving processes as simple as those of a shipping department or production line.
Let us say we have just heard or read a complete sentence, ‘The tall blond man with one black shoe gave the other one to her.’ How do we make a mental model of the action? We need to box up some of the pieces, and prepositional phrases are a good place to start. Our machine knows all the prepositions and takes the nouns adjacent to them (the following noun if the sentence is English, the preceding noun if it’s in Japanese) into the same box. I’ll use boxes with rounded corners to indicate the packaging of phrases – ‘with one black shoe’ and ‘to her.’ On occasion, nonlinguistic memories have to be brought to bear in order to box things up correctly, as in that ambiguous phrase ‘the cow with the crumpled horn that Farmer Giles likes.’ Knowing that Giles has a collection of horns over the fireplace could help you guess whether ‘that Farmer Giles likes’ should be boxed with ‘cow’ or with ‘crumpled horn.’
Verbs get special boxes, because of the special role they play. Had there been an -ly word (an adverb), or an auxiliary, such as ‘must,’ I would have boxed them up with the verb, even if they weren’t adjacent to it. Then we box up the noun phrases, incorporating any prepositional phrase boxes that modify them, so that we may have rounded boxes within rectangular boxes. If we have a nested phrase, it can function as a noun for purposes of the next boxing. Now we’ve got everything boxed up (there have to be at least two boxes but often there are more).
Next we’ve got to ‘lift’ them as a group and metaphorically carry this amalgamation away from the work space, understood at last. Will it get off the ground? There are a few different types of handles in my vacuum-lifter machine, and the one we must use depends on the verb we identified (in this case, the past tense of ‘give’). There is another vacuum sucker, for the noun phrase box containing the subject (I’ve drawn it as a little pyramid). You can’t have a sentence without both a subject and a verb, and if the subject is missing, air will be sucked in the opening, no vacuum will form, and the package lifter won’t lift. (That’s why I’ve used suckers here rather than hooks – to make a target obligatory.)
But, as I noted earlier, ‘give’ is peculiar, in that it requires two objects. (You can’t say, ‘I gave to her.’ Or ‘I gave it.’). Therefore this lift handle has two additional suction lines. I’ve also allowed it some nonvacuum lines – simple strings with hooks, which can carry as many optional noun phrases and prepositional phrases, however many that the verb allows.
Sometimes the suction tips and optional hooks need some guidance to find an appropriate target: for example, SVO might help the subject tip find the appropriate noun phrase – as might a case marker, such as ‘he.’ Other inflections help out, like gender or number agreement between verb and subject. The suction tips and the hooks could come with little labels on them for Beneficiary, Instrument, Negation, Obligation, Purpose, Possession, and so forth, mating only with words appropriate to those categories. Being able to lift the verb handle and carry all the packages, leaving none behind and with no unfilled suction tips, is what constitutes sentence recognition in this particular grammar machine. If a suction tip can’t find a home, no vacuum develops when you lift the handle, and your construction isn’t carried away. There’s no sense of completion.
As noted, each verb, once identified by the lingua ex machina, has a characteristic handle type: for instance, handles for intransitive verbs such as ‘sleep’ have only the one suction tip for the subject, but they have optional hooks, in case there are any extra phrases to be hauled along. ‘Sleep’ will support optional roles such as Time (‘after dinner’) and Place (‘on the sofa’) – but not Recipient.
There’s usually a vacuum tip for an Agent (though sometimes there isn’t an Agent – say, in sentences like ‘The ice melted’), perhaps other role-related suction tips, and some hooks for other possible roles in the verb’s storytelling repertoire.
And, of course, the same boxes-inside-boxes principle that allowed a prepositional phrase to serve as a noun can allow us to have sentences inside sentences, as in dependent clauses or ‘I think I saw him leave to go home.’
That’s the short version of my package-carrying system. If it seems worthy of Rube Goldberg, remember that he’s the patron saint of evolution.
I assume that, just as in a roomful of bingo players, many attempts at a solution are made in parallel, with multiple copies of the candidate sentence superimposed on different prototypical sentence scaffolds, and that most of these arrangements fail because of leftover words and unfilled suction tips. The version whose verb handle lifts everything shouts, ‘Bingo!’ and the deciphering game is over (unless, of course, there’s a tie).
Being able to lift everything is simply a test of a properly patterned sentence; note that, once lifted successfully, sequence and inflections no longer matter, because roles have been assigned. This lingua ex machina would lift certain kinds of nonsense – such as Chomsky’s famous example, ‘Colorless green ideas sleep furiously’ – but would, appropriately, fail to lift a nonsentence, such as ‘Colorless green ideas sleep them.’ (The ‘sleep’ verb handle has no hooks or suckers for leftover Objects.)
Though a sensible mental model of relationships may be the goal of communication, and ungrammatical sentences cannot be deciphered except through simple word association, grammatical patterns of words can nonetheless be generated that fit sentence expectations but have no reasonable mental model associated with them. The test of semantics is different than the test of grammar. Semantics is also the tie-breaker, deciding among multiple winners, in somewhat the same way as boxing matches without a knockout are decided on judges’ points. That’s also how we guess what Farmer Giles is likely to like, the cow or the horn.
While each sentence is a little story, we also build string-based conceptual structures far larger than sentences. They too have lots of obligatory and optional roles to fill. They come along in the wake of grammar, as the writer Kathryn Morton observes:
The first sign that a baby is going to be a human being and not a noisy pet comes when he begins naming the world and demanding the stories that connect its parts. Once he knows the first of these he will instruct his teddy bear, enforce his world view on victims in the sandlot, tell himself stories of what he is doing as he plays and forecast stories of what he will do when he grows up. He will keep track of the actions of others and relate deviations to the person in charge. He will want a story at bedtime.
Our plan-ahead abilities gradually develop from childhood narratives and are a major foundation for ethical choices, as we imagine a course of action, imagine its effects on others, and decide not to do it.
By borrowing the mental structures for syntax to judge other combinations of possible actions, we can extend our plan-ahead abilities and our intelligence. To some extent, this is done by talking silently to ourselves, making narratives out of what might happen next, and then applying syntax-like rules of combination to rate (a decision on points, again) a candidate scenario as dangerous nonsense, mere nonsense, possible, likely, or logical. But our intelligent guessing is not limited to language-like constructs; indeed, we may shout, ‘Eureka!’ when a set of mental relationships clicks into place, yet have trouble expressing this understanding verbally for weeks thereafter. What is it about human brains that allows us to be so good at guessing complicated relationships?
We do not realize how deeply our starting assumptions affect the way we go about looking for and interpreting the data we collect. We should recognize that nonhuman organisms need not meet every new definition of human language, tool use, mind, or consciousness in order to have versions of their own that are worthy of serious study. We have set ourselves too much apart, grasping for definitions that will distinguish man from all other life on the planet. We must rejoin the great stream of life from whence we arose and strive to see within it the seeds of all we are and all we may become.
Sue Savage-Rumbaugh, 1994
[We] can understand neither ourselves nor our world until we have fully understood what language is and what it has done for our species. For although language made our species and made the world we inhabit, the powers it unleashed drove us to understand and control our environment, rather than explore the mainspring of our own being. We have followed that path of control and domination until even the most daring among us have begun to fear where it may lead. Now the engine of our quest for power and knowledge should itself become the object that we seek to know.
Derek Bickerton, 1990
CHAPTER 6
EVOLUTION ON-THE-FLY
Foresight of phenomena and power over them depend on knowledge of their sequences, and not upon any notion we may have formed respecting their origin or inmost nature.
John Stuart Mill, Auguste Comte and Positivism, 1865
The problems are solved, not by giving new information, but by arranging what we have known since long.
Ludwig Wittgenstein, Philosophical Investigations, 1953
‘One thing follows another’ is a fairly simple concept, one that many animals can master. Indeed, it’s what most learning is all about; for Pavlov’s dogs, it was bell tends to be followed by food.
More than two things may be chained; many animals have elaborate song sequences, not to mention all those elaborate locomotion sequences, such as gaits. Acquiring vocabulary and understanding basic word order are, we just saw, relatively easy language tasks for both humans and bonobos.
If sequence is so elementary, why is planning ahead so rare in the animal kingdom, except for those trivial cases of foresight that mere melatonin can handle so well? What additional mental machinery is required in order to plan for a novel contingency? (Perhaps argument structure, as in those verb-lifting handles?) How do we do something we’ve never done before, with no exact memories to guide us? How do we even imagine such a thing?
We are always saying something we’ve never said before. The other novelty generator, operating just as frequently in our lives (though often subconsciously), is that ‘What happens next?’ predictor, mentioned in chapter 2 in the context of humor and the distressful effects of environmental incoherence.
Perhaps the mechanisms for foresight are similar to those used in the fancier aspects of mental grammar, the ones involving long-term dependencies, as when basic word order is replaced by the alternate forms for the who-what-when questions. Perhaps the trees used by phrase structure, or the obligatory roles of argument structure, are mental mechanisms that are useful for foresight in a more general way.
Mental grammar provides our most detailed set of insights into those mental structures that might be handy for intelligent guessing. This chapter will take a look at three more: chunking, sequencing, and darwinian processes.
Juggling a half-dozen things at the same time is one of those abilities measured by multiple-choice tests, particularly analogy questions (A is to B as C is to [D, E, F]). It also shows up in our ability to remember phone numbers for a long enough time to dial them. Many people can hang on to a seven-digit number between five and ten seconds, but will resort to writing it down if faced with an out-of-area number or an even longer international one.
The limitation, it turns out, is not the number of digits; it’s the number of chunks. I remember San Francisco’s area code, 415, as a single chunk, but the number 451 means nothing to me, so I would have to remember it as three chunks: 4, 5, and 1. Chunking refers to the process of collapsing 4, 1, and 5 into the entity 415. A ten-digit San Francisco phone number, such as 4153326106, is, to me, only eight chunks; our schemes for using nondialed separators when writing down numbers – as in (415)332–6106 or 415.332.6106 – are essentially aids to chunking. Since we are already familiar with many two-digit numbers as single words – for example, ‘nineteen’ – the Parisian 42–60–31–25 style of separators makes for more easily memorized eight-digit number strings.
How many chunks can you hang onto? That varies among people, but the typical range forms the title of a famous 1956 paper by the psychologist George Miller: ‘The Magical Number Seven, Plus or Minus Two.’ It’s as if the mind had room for only a limited number of items – at least, in the work space used for current problems. When you get close to your limit, you try to collapse several items into one chunk so as to make more room. Acronyms are a version of chunking, making one ‘word’ from many. Indeed, many new words are just substitutes for a longer phrase, as when someone invented ambivalence as a shortcut, to save a whole paragraph of explanation. A dictionary is a compendium of chunking over the centuries. The combination of chunking and rapid speech, so that much meaning can be accommodated within the brief span of short-term memory, has surely been important for holding as much information as possible in mind at the same time.
So one of the first lessons about working memory is that there’s seemingly a limited scratch pad, better suited to a half-dozen items than twice that number. This limitation probably has some implications for intelligence (certainly for IQ tests!), but the key feature of intelligent acts is creative divergent thinking, not memory per se. What we need is a process that will produce good guesses.
Language and intelligence are so powerful that we usually assume that more and more would be better and better. Evolutionary theorists, however, are fond of demonstrating that evolution is full of dead-end stabilities that can prevent such straightforward ‘progress’ and they like to point out evolution’s indirect routes involving multipurpose organs. Many organs are actually multipurpose and change their relative mix of functions over time. (When did that gas exchange organ in fish, known as the ‘swim bladder’ because of its role in neutralizing buoyancy, become a lung?) And, if the analogy to computer software is to be believed, it’s far easier for the brain to be multipurpose than it is for any other organ system. Some regions of the brain are surely multipurpose too.
So, in asking about how neural machinery for foresight or language got started, we must bear in mind that the underlying mechanisms might serve multiple functions, any one of which could be driven by natural selection and so incidentally benefit the others. They might be like what architects call core facilities, such as the rooms for the photocopy machines and the mailboxes. The mouth, for example, is a multipurpose core facility involved with drinking, tasting, ingesting, vocalization, and emotional expression; in some animals, also with breathing, cooling off, and fighting.
Bundling (paying for one thing, but getting something else ‘free’) is a familiar marketing strategy. What human abilities might come bundled together like the proverbial ‘free lunch’ that comes with the cost of the drinks? In particular, might syntax or planning come bundled with some other ability, simply because they can make spare-time use of a core facility?
I realize that a ‘free lunch’ explanation is going to offend the sensibilities of the more Calvinist of the strict adaptationists in evolutionary theory – the ones that think that every little feature has to pay its own way. But strict accounting isn’t always the name of the game. As noted earlier (enlarge one, enlarge them all), mammalian brain enlargements tend not to come piecemeal. And a free lunch is just another way of looking at what the original adaptationist himself emphasized. Charles Darwin reminded his readers, in a caution to his general emphasis on adaptations, that conversions of function were ‘so important.’
In the midst of converting function – swim bladder into lung, for example – there is likely a multifunctional period (indeed, the multifunctional period could last forever). During it, an anatomical feature formerly under natural selection for one function gives an enormous boost to some new function, far beyond whatever natural selection the new function has experienced so far. Lungs were ‘bootstrapped’ by earlier buoyancy considerations. What brain functions have bootstrapped others, and does it tell us anything about intelligence?
In considering transitions of organs, it is so important to bear in mind the probability of conversion from one function to another.
Charles Darwin, The Origin of Species, 1859
We certainly have a passion for stringing things together in structured ways, ones that go far beyond the sequences produced by other animals. Besides words into sentences, we combine notes into melodies, steps into dances, and elaborate narratives into games with procedural rules. Might structured strings be a core facility of the brain, useful for language, story-telling, planning ahead, games, and ethics? Might natural selection for any of these abilities augment the common neural machinery, so that improved grammar incidentally serves to expand plan-ahead abilities?
Some beyond-the-apes abilities – music, for example – are puzzling, because it is hard to imagine environments that would give the musically gifted an evolutionary advantage over the tone deaf. To some extent, music and dance are surely secondary uses of that very neural machinery that was shaped up by structured strings more exposed to natural selection, such as language.
What other beyond-the apes abilities were likely to have been under strong natural selection? As improbable as it initially seems, planning ballistic movements may have once promoted language, music, and intelligence. Apes have elementary forms of the rapid arm movements that we’re experts at – hammering, clubbing, and throwing – and one can imagine hunting and toolmaking scenarios that in some settings were important additions to the basic hominid gathering and scavenging strategies. If the same ‘structured string’ core facility is used for the mouth as is used for ballistic hand movements, then improvements in language might promote manual dexterity. It could work the other way, too: accurate throwing opens up the possibility of eating meat regularly, of being able to survive winter in the temperate zone – and of talking all the better as an incidental benefit, a ‘free lunch.’
Choosing between hand movements involves finding a candidate movement program – likely a characteristic firing pattern of cortical neurons – and then some more candidates. Little is yet known about how this transpires in the human brain, but a simple model involves multiple copies of each movement program, each competing for space in the brain. The program for an open palm might make copies more readily than the program for making a V-sign or a precision pincer grip.
Ballistic movements (so named because beyond a certain point there is no opportunity to modify the command) require a surprising amount of planning, compared to most movements. They also likely require lots of clones of the movement program.
For sudden limb movements lasting less than about an eighth of a second, feedback corrections are largely ineffective, because reaction times are so long. Nerves conduct too slowly, and decisions aren’t made quickly enough; feedback might help plan for next time, if the target hasn’t run away by then, but it’s no help in real time. For the last one-eighth second of clubbing, hammering, throwing, and kicking, the brain has to plan every detail of the movement and then spit it out, rather like punching a roll for a player piano and then letting it roll.
We need nearly complete advance planning for ballistic movements during ‘get set,’ with no reliance on feedback. Hammering requires planning the exact sequence of activation for dozens of muscles. In the case of throwing, the problem is difficult for an additional reason: there is a launch window – a range of times when the projectile can be released and still hit the target. Release occurs shortly after the velocity peaks, as the projectile sails out of the decelerating hand. Getting this peak velocity to occur at exactly the right time, at the appropriate angle from the horizontal, is the trick.
Given the launch-window problems, you can see why planning is so difficult for human ballistic movements. Launch windows depend on how far away the target is, and on how big it is. Let’s say that, eight tries out of ten, you can hit a rabbit-sized target from the length of one parallel parking space – which implies a launch window of 11 milliseconds. Hitting the same target from twice the distance with equal reliability means releasing within a launch window about eight times narrower, 1.4 msec. Neurons are not exactly atomic clocks when it comes to timing accuracy; there is a lot of jitter when impulses are produced, enough so that any one neuron would have trouble hitting the broad side of a barn if it had to time the ball’s release all by itself.
Happily, many noisy neurons are better than a few – so long as they’re all ‘doing their own thing’ and thereby making their own mistakes. Combining them can average out some of the noise. You can see this principle at work in the heart, making the heartbeat more regular. A fourfold increase in the number of pacemaker cells serves to cut the heartbeat jitter in half. To reduce ballistic release jitter eightfold requires averaging the output of 64 times as many noisy neurons as you needed to program the original throw. If you want to hit that same rabbit-sized target at three times the distance with that same eight-out-of-ten-times reliability, count on needing to recruit a lot of helpers: you will require 729 times as many neurons as the number sufficient for generating your standard short throw. It’s redundancy, but in a different sense from, say, the three ways every large airplane has of lowering the landing gear.
So now we have a third insight into relevant brain mechanisms for fancy sequences: besides those trees and handles of syntax, besides those limited scratch-pad memories that encourage chunking, we see that fancy sequences of activation such as the ballistic movements probably share cerebral real estate with other fancy sequences – and that some need hundredfold levels of redundancy when precision timing is important.
Lots of planning space is also needed when you are throwing at a nonstandard target distance – one for which you don’t have a stored movement plan (as you might for throwing darts or basketball free-throws). For nonstandard shots, you need to create an array of variants between two standard programs and pick the one that will come closest to hitting your target. Improvisation takes space. If, once you select the ‘best’ variant, all the other variants change to conform to it, then you would have the redundancy needed for staying inside the launch window. Imagine a roomful of soloists, all singing somewhat different melodies and then converging on the one that they could sing as a chorus. And then, for real precision, recruiting a lot of helpers, just as the expert choir recruits the audience in the Hallelujah Chorus.
A core facility for structured sequences could solve a lot of problems. Does one actually exist? If so, we might occasionally see some synergy or conflict between similar movements.
Charles Darwin was one of the first to suggest hand-to-mouth synergies in his 1872 book on the expression of the emotions: ‘Thus persons cutting anything with a pair of scissors may be seen to move their jaws simultaneously with the blades of the scissors. Children learning to write often twist about their tongues as their fingers move, in a ridiculous fashion.’
What kind of sequences are we talking about, anyway? Rhythmic movements per se are ubiquitous: chewing, breathing, locomotion, and so forth. They can be implemented by simple circuits at the level of the spinal cord. Like the simple onething-follows-another of learning, there is nothing distinctively cerebral about rhythm or other sequences. But novel sequences, that’s the rub. If there is a common sequencer for the fancier novel movements, where is it located in the brain?
Sequencing in itself doesn’t require a cerebral cortex. Much movement coordination in the brain is done at a subcortical level, in places known as the basal ganglia and the cerebellum. But novel movements tend to depend on the premotor and prefrontal cortex, in the rear two-thirds of the frontal lobes.
There are other regions of the cerebral cortex that are likely to be involved with sequential activities. The dorsolateral portions of the frontal lobe (dorso = top, lateral = side; if you had a pair of horns growing out of your forehead, these regions would lie beneath them) are crucial for delayed-response tasks. You show a monkey some food and allow him to watch where you hide it – but force him to wait twenty minutes before being allowed to go after it. Monkeys with damage to the dorsolateral frontal cortex will fail to retain that information. It may not be a failure of memory but a problem of formulating a lasting intention, perhaps even an ‘agenda.’
The great Russian neurologist Alexander Luria described a patient in bed with his arms under the covers. Luria asked him to raise his arm. He couldn’t seem to do that. But if Luria asked him to remove his arm from under the covers, he could do that. If Luria then asked him to raise his arm up and down in the air, the patient could do that, too. His difficulty was in planning the sequence – he got stuck on the condition of working around the obstacle of the confining bedcovers. Left prefrontal damage gives patients difficulty in unfolding a proper sequence of actions – or perhaps in planning them in the first place. Patients with damage to the left premotor cortex have trouble chaining the actions together into a fluent motion – what Luria called a kinetic melody.
Tumors or strokes in the bottom of the frontal lobe, just above the eyes, also affect sequences of activities, such as going shopping. One famous patient, an accountant, had a high IQ and did quite well on a battery of neuropsychological tests. Yet he had big problems in organizing his life: he was fired from a series of jobs, went bankrupt, and underwent two divorces in a two-year period as a result of impulsive marriages. Nonetheless, this man was often unable to make simple, rapid decisions – say, about what toothpaste to buy or what to wear. He would get stuck making endless comparisons and contrasts, often making no decision at all or a purely random one. If he wanted to go out for dinner, he had to consider the seating plan, the menu, the atmosphere, and the management of each possible restaurant. He might drive by them to see how busy they were, and even then would be unable to choose among them.
There are two major lines of evidence that suggest the lateral language area above the left ear also has a lot to do with nonlanguage sequencing. The Canadian neuropsychologist Doreen Kimura and her coworkers showed that left-lateral stroke patients with language difficulties (aphasia) also have considerable difficulty executing hand and arm movement sequences of a novel sort, a condition known as apraxia. (A fancy, though not novel, sequence would be taking your keys out of your pocket, finding the right one, inserting it into the lock, turning the key, and then pushing on the door.)
The Seattle neurosurgeon George Ojemann and his coworkers further showed, using electrical stimulation of the brain during epilepsy operations, that much of the left-lateral language specialization is involved with listening to sound sequences. These regions include the part of the frontal lobe adjacent to Broca’s Area, the top of the temporal lobe on either side of the primary auditory cortex, and some of the parietal lobe in back of the map of the skin surface. (In other words, they’re ‘perisylvian,’ bordering the Sylvian fissure.) The big surprise was that these exact same areas seem heavily involved in producing oral-facial movement sequences – even nonlanguage ones, such as mimicking a series of facial expressions.
One of the hazards of naming things in the brain is that we expect something called the language cortex to be devoted to language. But data such as Ojemann’s show that, at its core, the cortical specialization is far more generalized, concerned with novel sequences of various kinds: hand as well as mouth, sensation as well as movement, mimicry as well as narrative.
Not only can many species learn abstract symbols and a simple language, but some clearly can learn categories. Indeed, animals often overgeneralize, in the same way that a baby goes through a phase of calling all adult males ‘Daddy.’ Relationships can be learned, such as is-a or is-larger-than. A banana is a fruit, a banana is larger than a chestnut.
Closer to intelligence is the power of analogies, metaphors, similes, parables, and mental models. They involve the comparing of relationships, as when we make an imperfect analogy between is-bigger-than and is-faster-than, by inferring that bigger-is-faster.
We humans can mentally operate in a familiar domain (for example, filing a document in a file folder or throwing it in a wastebasket) and carry this relationship over to a less familiar domain (saving or deleting computer files, perhaps by means of moving icons on a screen). We can make a gesture in one mental domain and have it interpreted in another. These mappings all break down somewhere – and, in Robert Frost’s words, we have to know how far we can ride a metaphor, judge when it’s safe.
Consider the mapping from one domain to another that Umberto Eco creates here:
The fact is that the world is divided between users of the Macintosh computer and users of MS-DOS compatible computers. I am firmly of the opinion that the Macintosh is Catholic and that DOS is Protestant. Indeed, the Macintosh is counterreformist and has been influenced by the ‘ratio studiorum’ of the Jesuits. It is cheerful, friendly, conciliatory, it tells the faithful how they must proceed step by step to reach – if not the Kingdom of Heaven – the moment in which their document is printed. It is catechistic: the essence of revelation is dealt with via simple formulae and sumptuous icons. Everyone has a right to salvation.
DOS is Protestant, or even Calvinistic. It allows free interpretation of scripture, demands difficult personal decisions, imposes a subtle hermeneutics upon the user, and takes for granted the idea that not all can reach salvation. To make the system work you need to interpret the program yourself: a long way from the baroque community of revelers, the user is closed within the loneliness of his own inner torment.
You may object that, with the passage to Windows, the DOS universe has come to resemble more closely the counterreformist tolerance of the Macintosh. It’s true: Windows represents an Anglican-style schism, big ceremonies in the cathedral, but there is always the possibility of a return to DOS to change things in accordance with bizarre decisions. . . .
And machine code, which lies beneath both systems (or environments, if you prefer)? Ah, that is to do with the Old Testament, and is Talmudic and cabalistic.
Most mappings are simpler, as when objects are associated with a sequence of phonemes (as in naming). Chimpanzees, with some effort, can learn simple analogies, such as A is to B as C is to D. If the chimp could apply such mental manipulations to events in its everyday life, instead of using them only while at the testing apparatus, it would be a more capable ape. Humans, obviously, keep mapping into more and more abstract domains, notching stratified stability up a few more levels.
Safety is the big problem with trial combinations, ones that produce behaviors that have never been done before. Bigger isn’t always faster. Even simple reversals in order can yield dangerous novelty, as in ‘Look after you leap.’ In 1943, in his book The Nature of Explanation, the British psychologist Kenneth Craik proposed that
the nervous system is . . . a calculating machine capable of modeling or paralleling external events. . . . If the organism carries a ‘small-scale model’ of external reality and of its own possible actions within its head, it is able to try out various alternatives, conclude which is the best of them, react to future situations before they arise, utilise the knowledge of past events in dealing with the future, and in every way to react in a much fuller, safer and more competent manner to the emergencies which face it.
Humans can simulate future courses of action and weed out the nonesence off-line; as the philosopher Karl Popper has said, this ‘permits our hypotheses to die in our stead.’ Creativity – indeed, the whole high end of intelligence and consciousness – involves playing mental games that shape up quality.
What sort of mental machinery might it take to do something of the sort that Craik suggests?
The American psychologist William James was talking about mental processes operating in a darwinian manner in the 1870s, little more than a decade after Charles Darwin published On the Origin of Species. The notion of trial and error was developed by the Scottish psychologist Alexander Bain in 1855, but James was using evolutionary thinking in addition.
Not only might darwinism shape up a better brain in two million years without the guiding hand of a master potter, but another darwinian process, operating in the brain, might shape up a more intelligent solution to a problem on the milliseconds-to-minutes timescale of thought and action. The body’s immune response also appears to be a darwinian process, whereby antibodies that are better and better fits to the invading molecule are shaped up in a series of generations spanning several weeks.
Darwinian processes tend to start from the biological basic: reproduction. Copies are always happening. One theory of making up your mind is that you form some plans for movement – making an open hand, or a V-sign, or a precision pincer movement – and that these alternative movement plans reproductively compete with one another until one ‘wins.’ According to that theory, it takes a critical mass of command clones before any action is finally initiated.
Darwinism requires a lot more than just reproduction and competition, however. When I try to abstract the essential features of a darwinian process from what we know about species evolution and the immune response, it appears that a Darwin Machine must possess six essential properties, all of which must be present for the process to keep going:
It involves a pattern. Classically, this is the string of DNA bases called a gene. As Richard Dawkins pointed out in The Selfish Gene, the pattern could also be a cultural one such as a melody, and he usefully coined the term meme for such patterns. The pattern could also be the brain activity patterns associated with thinking a thought.
Copies are somehow made of this pattern. Cells divide. People hum or whistle a tune they’ve overheard. Indeed, the unit pattern (that’s the meme) is defined by what’s semi-reliably copied – for example, the gene’s DNA sequence is semi-reliably copied during meiosis, whereas whole chromosomes or organisms are not reliably copied at all.
Patterns occasionally change. Point mutations from cosmic rays may be the best-known alterations, but far more common are copying errors and (as in meiosis) shuffling the deck.
Copying competitions occur for occupation of a limited environmental space. For example, several variant patterns called bluegrass and crabgrass compete for my backyard.
The relative success of the variants is influenced by a multifaceted environment. For grass, the operative factors are nutrients, water, sunshine, how often the grass is cut, and so on. We sometimes say that the environment ‘selects,’ or that there is selective reproduction or selective survival. Charles Darwin called this biasing by the term natural selection.
The next generation is based on which variants survive to reproductive age and successfully find mates. The high mortality among juveniles makes their environment much more important than that of adults. This means that the surviving variants place their own reproductive bets from a shifted base, not from where the center of the variations was at conception (this is what Darwin called the inheritance principle). In this next generation, a spread around the currently successful is again created. Many new variants will be worse than the parental average, but some may be even better ‘fitted’ to the environment’s collection of features.
From all this, one gets that surprising darwinian drift toward patterns that almost seem designed for their environment. (There! I actually managed to work ‘intelligent design’ into this intelligence book; maybe there’s hope yet for ‘military intelligence.’)
Sex (which is shuffling genes using two decks) isn’t essential to the darwinian process, and neither is climate change – but they add spice and speed to it, whether it operates in milliseconds or millennia. A third factor accelerating the darwinian process is fragmentation and the isolation that follows: the darwinian process operates more quickly on islands than on continents. For some fancy darwinian processes requiring speed (and the timescale of thought and action certainly does), that might make fragmentation processes essential. A decelerating factor is a pocket of stability that requires considerable back-and-forth rocking in order to escape from it; most stable species are trapped in such stabilizing pockets.
People are always confusing particular parts, such as ‘natural selection,’ with the darwinian whole. But no one part by itself will suffice. Without all six essentials, the process will shortly grind to a halt.
People also associate the darwinian essentials exclusively with biology. But selective survival, for example, can be seen when flowing water carries away the sand and leaves the pebbles behind. Mistaking a part for the process (‘Darwinism is selective survival’) is why it has taken a century for scientists to realize that thought patterns may also need to be repeatedly copied – and that copies of thoughts may need to compete with copies of alternative ones on ‘islands’ during a series of mental ‘climate changes’ in order to rapidly evolve an intelligent guess.
In our search for suitable brain mechanisms for guessing intelligently, we now have (1) those nested boxes of syntax that underlie strings; (2) argument structure with all its clues about probable roles; (3) those relative position words such as near-into-above; (4) the limited size of scratch-pad memory and the consequent chunking tendencies; and (5) common core facilities for fancy sequences, with quite a lot of need for extra copies of the neural patterns used to produce ballistic movements. Our sixth clue, from darwinian processes, now appears to be a whole suite of features: distinctive patterns, copying them, establishing variants via errors (with most of the variants coming from the most successful), competition, and the biasing of copying competitions by a multifaceted environment. What’s more, it looks as if the multifaceted environment is partly remembered and partly current.
Fortunately, there is some overlap of darwinian considerations with those from the ballistic movements: darwinian backyard work spaces might utilize the ‘get set’ scratch pads, and darwinian copying could help produce the jitter-reducing movement command clones. What else might correspond? In particular, what are those patterns that we might need to clone, on the timescale of thought and action?
Thoughts are combinations of sensations and memories – or, looked at another way, thoughts are movements that haven’t happened yet (and maybe never will). They’re fleeting and mostly ephemeral. What does this tell us?
The brain produces movements by means of a barrage of nerve impulses going to the muscles, whether of the limbs or the larynx. Each muscle is activated at a somewhat different time, often only briefly; the whole sequence is timed as carefully as the finale of a fireworks display. A plan for a movement is like a sheet of music or a player-piano roll. In the latter case, the plan covers eighty-eight output channels and the times at which each key is struck, and, indeed, the ballistic movements involve almost as many muscles as the piano has notes. So a movement is a spatiotemporal pattern not unlike a musical refrain. It might repeat over and over, like the rhythms of locomotion, but it could also be more like a one-shot arpeggio, triggered by another temporal pattern.
Some spatiotemporal patterns in the brain probably qualify for the name cerebral code. Though individual neurons are more sensitive to some features of an input than others, no single neuron represents your grandmother’s face. Just as your sense of a color depends on the relative activity in three different cone pathways from the retina, and a taste can be represented by the relative amounts of activity in about four different types of tongue receptors, so any one item of memory is likely to involve a committee of neurons. A single neuron, like any one key on the piano, is likely to play different roles in different melodies (most often, of course, its role is to keep quiet – again, like a piano key).
A cerebral code is probably the spatiotemporal activity pattern in the brain which represents an object, an action, or an abstraction such as an idea – just as bar codes on product packages serve to represent without resembling. When we see a banana, various neurons are stirred by the sight: some of the neurons happen to specialize in the color yellow, others in the short straight lines tangent to the banana’s curve. Evoking a memory is simply reconstituting such a pattern of activity, according to the cell-assembly hypothesis put forward in 1949 by the Canadian psychologist Donald O. Hebb.
So, the banana committee is like a melody, if we imagine the neurons involved as unpacked along a musical scale. Some neurophysiologists think that the involved neurons all have to fire synchronously, as in a chord, but I think that a cerebral code is more like a short musical melody, comprised of chords and individual notes; we neurophysiologists just find it easier to interpret chords than we do scattered single notes. What we really need are the families of strange attractors associated with words, but that’s another book! (The Cerebral Code.)
Music is the effort we make to explain to ourselves how our brains work. We listen to Bach transfixed because this is listening to a human mind.
Lewis Thomas, The Medusa and the Snail, 1979
We know that long-term memories cannot be spatiotemporal patterns. For one thing, they survive even massive shutdowns of the electrical activity in the brain, as in seizures or coma. But we now have lots of examples of how to convert a spatial pattern into a spatiotemporal one: musical notation, player pianos, phonograph records – even bumps in a washboarded road waiting for a car to come along and recreate a bouncing spatiotemporal pattern.
This is what Donald Hebb called dual-trace memory: a short-term active version (spatiotemporal) and a long-term spatial-only version similar to a sheet of music or the grooves on a phonograph record.
Some of these ‘cerebral ruts’ are as permanent as those in the grooves of a phonograph record. The bumps and ruts are, essentially, the strengths of the various synapses that predispose the cerebral cortex to produce a repertoire of spatiotemporal patterns, much like the connection strengths in the spinal cord predispose it to produce the spatiotemporal patterns we know as walking, trotting, galloping, running, and so forth. But short-term memories can be either active spatiotemporal patterns (probably what is called ‘working memory’ in the psychology literature) or transient spatial-only patterns – temporary ruts that somewhat overwrite the permanent ruts but don’t vibrate (they merely fade in a matter of minutes). They’re simply the altered synaptic strengths (what is called ‘facilitation’ and ‘long-term potentiation’ in the neurophysiological literature), the bumps left behind by a repetition or two of the characteristic spatiotemporal pattern.
The truly persistent bumps and ruts are unique to the individual, even to each identical twin, as the American psychologist Israel Rosenfield explains:
Historians constantly rewrite history, reinterpreting (reorganizing) the records of the past. So, too, when the brain’s coherent responses become part of a memory, they are organized anew as part of the structure of consciousness. What makes them memories is that they become part of that structure and thus form part of the sense of self; my sense of self derives from a certainty that my experiences refer back to me, the individual who is having them. Hence the sense of the past, of history, of memory, is in part the creation of the self.
Copying is going to be needed over long distances in the brain. Like a fax machine, the brain must take a pattern and make a distant copy of it, perhaps on the other side of the brain. The pattern cannot be physically transported in the manner of a letter, so telecopying is likely to be important when the visual cortex wants to tell the language area that an apple has been seen. The need for copying suggests that the pattern we seek is the working memory, that active spatiotemporal pattern, since it is difficult to see how ‘ruts’ would otherwise copy themselves at a distance.
A darwinian model of mind and my analysis of the activity of throwing suggest that many clones might be needed locally, not just a few in distant places. Furthermore, in a darwinian process, an activated memory must somehow compete with other spatiotemporal patterns for occupation of a work space. And the other question we must answer is, What decides if one such ‘melody’ is better than another?
Suppose a spatiotemporal pattern, produced in one little part with the aid of some appropriate ‘ruts,’ manages to induce the same melody in another cortical area that lacks those ruts. But the pattern can nonetheless be performed there, thanks to the active copying process nearby, even if it might not sustain itself without the driving patterns, the same way a square dance might fizzle out without a caller. If an adjacent area has bumps and ruts that are ‘close enough,’ the melody might catch on better, and die out less readily, than some other imposed melody. So resonating with a passive memory could be the aspect of the multifaceted environment that biases a competition.
In this way, the permanent bumps and ruts bias the competition. But so do the fading ones that were made by spatiotemporal activity patterns in that same patch of cortex a few minutes earlier. So, too, do the current active inputs to the region from elsewhere – the ones that are (like most synaptic inputs) in themselves too weak to induce a melody or create ruts. Probably most important is the background of secretions from the four major diffuse projection systems, the ones associated with the serotonin, norepinephrine, dopamine, and acetylcholine neuromodulators. Other emotional biases surely come from the neocortical projections of the such subcortical brain sites as the amygdala. Thalamic and cingulate gyrus inputs may bias competitions elsewhere, in the name of shifting your attention from external to memorized environments. Thus the current real-time environment, memories of near-past and long-past environments, emotional state, and attention all change the resonance possibilities, all likely bias the competition that shapes up a thought. Yet they could do it without themselves forming up clones to compete for cortical territory.
The picture that emerges from such theoretical considerations is one of a quilt, some patches of which enlarge at the expense of their neighbors as one code copies more successfully than another. As you try to decide whether to pick an apple or a banana from the fruit bowl (so my theory goes), the cerebral code for apple may be having a cloning competition with the one for banana. When one code has enough active copies to trip the action circuits, you might reach for the apple.
But the banana codes need not vanish; they could linger in the background as subconscious thoughts, undergoing variations. When you unsuccessfully try to remember someone’s name, the candidate codes might continue copying for the next half hour, until suddenly Jane Smith’s name seems to ‘pop into your mind,’ because your variations on the spatiotemporal theme finally hit a resonance good enough to generate a critical mass of identical copies. Our conscious thought may be only the currently dominant pattern in the copying competition, with many other variants competing for dominance, one of which will win a moment later, when your thoughts seem to shift focus.
It may be that darwinian processes are only the frosting on the cognitive cake; it may be that much is routine or rulebound. But we often deal with novel situations in creative ways, as when you decide what to fix for dinner tonight. You survey what’s already in the refrigerator and on the kitchen shelves. You think about a few alternatives, keeping track of what else you might have to fetch from the grocery store. All this can flash through your mind within seconds – and that’s probably a darwinian process at work, as is speculating about what tomorrow might bring.
We build mental models that represent significant aspects of our physical and social world, and we manipulate elements of those models when we think, plan, and try to explain events of that world. The ability to construct and manipulate valid models of reality provides humans with our distinctive adaptive advantage; it must be considered one of the crowning achievements of the human intellect.
Gordon H. Bower and Daniel G. Morrow, 1990
Conflicts of representation are painful for a variety of reasons. On a very practical level, it is painful to have a model of reality that conflicts with those of the people around you. The people around you soon make you aware of that. But why should this conflict worry people, if a model is only a model, a best guess at reality that each of us makes? Because nobody thinks of it in that way. If the model is the only reality you can know, then that model is reality, and if there is only one reality, then the possessor of a different model must be wrong.
Derek Bickerton, 1990
CHAPTER 7
SHAPING UP AN INTELLIGENT ACT FROM HUMBLE ORIGINS
The schematicism by which our understanding deals with the phenomenal world . . . is a skill so deeply hidden in the human soul that we shall hardly guess the secret trick that Nature here employs.
Immanuel Kant, Kritik der reinen Vernunft, 1787
‘Why,’ said the Dodo, ‘the best way to explain it is to do it.’
Lewis Carroll, Alice’s Adventures in Wonderland, 1865
Is this chapter really necessary? Well, no – in the sense that many people could skip to the last chapter without realizing that something was missing.
It all depends on how satisfied you are with organization charts. Some people don’t want to know any more. ‘Skip the details,’ they say, ‘and just stick with the executive summary.’ But this chapter really isn’t about the details omitted from the last chapter – it’s written from a different perspective, bottom-up rather than inferred from principles.
Principles are, unfortunately, rather like organization charts – a sketchy, convenient fiction. Real organizations have a flow of information and decision making that isn’t captured by the boxes and labels. Charts fail to take account of people and how they talk to one another, fail to take account of the ‘institutional memory.’ They fail to take account of how experts can also be generalists, of how decisions taken at one level interact with those taken at another. Any schematic account of the brain will share the shortcomings of organization charts.
This account of intelligence has, so far, failed to take much account of neurons – the nerve cells of the brain – and how they talk to one another, how they remember past events, how they collectively make decisions on a local and regional scale. Some of that simply isn’t known yet, but it is certainly possible to sketch out a plausible account of copying competitions among the cerebral codes.
Whenever you are talking science, a good general rule is always to give a specific example – even if it is only a possible mechanism rather than a well-established one.
That’s what this chapter provides: an example of how our cerebral cortex might function as a Darwin Machine and, in the process, create that constantly shifting focus of consciousness, even those subconscious thoughts that every so often pop into the foreground, unbeckoned. It shows how we might achieve the off-line ability to simulate our future actions in the real world – an ability that is the essence of guessing-right intelligence.
The inability to imagine a mechanism that could produce mind is at the heart of many of the Janitor’s Dream and the mind-in-a-computer objections. This chapter describes the building blocks with which I can imagine how a thinking machine could be constructed. Your mileage may vary – but here’s your chance, just one chapter long, to see a bottom-up mechanistic example of how our mental lives might operate, both consciously and subconsciously, both for the novel and the routine.
Gray matter isn’t really gray, except in a dead brain; in a living brain, it has a rich blood supply. Think of those rivers that run reddish-grayish-brown after a thunderstorm, and you’ll have the right hue for the dynamic ‘gray matter.’
The white matter in the brain, however, is really white, a porcelain hue, because of the fat that insulates the long, stringy part of a neuron. This part, which is called the ‘axon,’ is analogous to a wire and carries the neuron’s output signal to near and distant targets. ‘Myelin’ is the proper name for its fatty insulation. White matter is simply wire bundles, going every which way, much as you would see in the basement of a telephone-central-office building. The bulk of the brain is insulated wires connecting the parts that do the hard work, which are far smaller.
At one end of the axon is the neuron’s cell body, the globular part of the cell containing the nucleus, with the DNA blueprints for the cell’s day-to-day operation and maintenance. There are lots of treelike branches, called dendrites, arising from the cell body. Because cell bodies and dendrites lack the white insulation, large collections of them look ‘gray.’ The far end of a neuron’s axon appears to be touching the dendrite of a downstream neuron – though, if you look carefully with an electron microscope, you’ll see a little gap between the two cells called a synapse. Into this no-man’s-land the upstream neuron releases a little neurotransmitter, which drifts across the gap and opens up some channels through the membrane of the downstream neuron. (Though there are some retrograde neurotransmitters in addition, a synapse is usually a one-way street, so it’s useful to refer to ‘upstream’ and ‘downstream’ neurons.)
Overall, a single neuron looks like a bush, or the root of some herb such as ginger. It is the typical unit of computation, summing up the influences of a few thousand inputs – most of them excitatory, and some of them inhibitory, like deposits and checks – and speaking, in a single voice, to several thousand hardwired listeners.
The message sent from this ‘checking account’ mostly concerns its ‘account balance’ and how fast that balance is increasing. No message is sent unless the balance exceeds some threshold. Big deposits generate big messages, like interest payments with a bonus. But, just as piano keys don’t produce any sound unless struck hard enough, cortical neurons are usually silent unless input conditions are surging – and then their output is proportional to how much they’re stimulated by that account balance. (Oversimplified binary models usually treat a neuron as more like a harpsichord key, with a threshold but no gradation in volume for harder hits.)
Though the messages from short neurons can be simpler, neurons with axons longer than about 0.5 mm always utilize a signal booster: the impulse, a brief up-and-down voltage change of a standard size (like the loudness of that harpsichord key). Amplified and fed into a loudspeaker, the impulse sounds like a click (and we talk of the neuron ‘firing’). To get around the standard-size limitation, impulses usually repeat at a rate proportional to the account balance, the same way that a few quick repetitions of a harpsichord note may imitate a hard-struck piano note. Sometimes – especially in the cerebral cortex – just a few inputs, out of thousands, can conspire to trigger an impulse.
The really interesting gray matter is that of the cerebral cortex, because that’s where most of the novel associations are thought to be made – where the sight of a comb, say, is matched up to the feel of a comb in your hand. The cerebral codes for sight and feel are different, but they become associated somehow in the cortex, along with those for hearing the sound /kōm/ or hearing the characteristic sounds that the teeth of a comb make when they’re plucked. You can, after all, identify a comb in any of these ways. It’s hypothesized that there are specialized places in the cortex, called ‘convergence zones for associative memories,’ where those different modalities come together.
On the production side, you have linked cerebral codes for pronouncing /kōm/ and for generating the movements that manipulate a comb through the hair on your head. So between the sensory version of the word ‘comb’ and the various movement manifestations, we expect to find a dozen different cortical codes associated with combs.
The cortical areas that do all this associating for us are a thin layer of icing on the cake of the white matter. The cerebral cortex is only about 2 mm thick, though it is deeply wrinkled. The neocortex (which is all of the cerebral cortex except for the hippocampus and some of the olfactory areas) has a surprisingly uniform packing density (with the exception of one layer of the primary visual cortex). If you made a grid atop the cortical surface, each square millimeter would have about 148,000 neurons beneath it – whether it was language cortex or motor cortex. But a sideways look, at the layers within that 2 mm depth, reveals some regional differences.
It’s the icing of this cake that contains the layers, not the cake itself. A better bakery analogy might be a flaky pie crust made of croissantlike layers. The deepest layers are like an out-box, their wires mostly heading out of the cortex, bound for distant subcortical structures, such as the thalamus or the spinal cord. The middle layer is an in-box, with wires arriving from the thalamus and other such places. The superficial layers are like an interoffice-box; they make ‘corticocortical’ connections with the superficial layers of other regions, both adjacent and distant. It’s their axons that go through the corpus callosum to the other side of the brain – but most of the interoffice mail is delivered locally, within several millimeters. Such axon branches run sideways, rather than detouring through the white matter like the longer ‘U-fiber’ branches.
Some regions have big in-boxes and small out-boxes, just like the ones to be found on the editorial-department desk that deals with letters to the editor. Superimposed on this stacked horizontal organization, moreover, is a fascinating set of vertical arrangements, similar to newspaper columns.
If we go around wiretapping the individual neurons in the cerebral cortex, we discover that neurons with similar interests tend to be vertically arrayed there, forming cylinders known as cortical columns, which cut through most of the layers. It’s almost like a club that self-organizes out of a crowd at a party, where people of similar interests tend to cluster together. We have naturally given names to these cortical clubs. Some of the names reflect their size, some their seeming specialties (so far as we know them).
The thin cylinders, or minicolumns, are only about 0.03 mm in diameter (that’s a very thin hair, closer to the threads of a spider web). The best-known examples of these are the visual cortex’s orientation columns, whose neurons seem to like visual objects with a line or border tilted at a particular angle. The neurons in one minicolumn will respond best to boundaries tilted at 35°, those in another will like horizontals or verticals, and so forth.
You can look in a microscope and see (well, it takes some doing, even after a century of progress in neuroanatomical technique) a group of cortical neurons bundled together like stalks of celery. There is a tall ‘apical dendrite’ that stretches up toward the cortical surface from the cell body (which is often triangular in appearance, hence the name ‘pyramidal neuron’). It is those apical dendrites of the pyramidal neurons that are bundled together, with 0.03 mm between adjacent bundles. There are about a hundred neurons in a minicolumn organized around one of those bundles, though the bundle at any one level might only have a dozen apical dendrites in it. Bundling is commonplace outside visual cortex, so minicolumns are likely a common element of cortical organization, just from the anatomy – but elsewhere we are ignorant of what the neurons of a minicolumn are ‘interested in.’
Other ‘interest groups’ tend to be much larger and comprised of more than a hundred minicolumns; these so-called macrocolumns are 0.4–1.0 mm across (that’s a thin pencil lead) and sometimes appear more like elongated curtain folds than like proper cylinders. Such macrocolumns seem to result from an organization of the inputs – for example, in visual cortex those axons carrying information from the left eye tend to alternate every 0.4 mm with those being relayed from the right eye. Inputs from other parts of the cortex itself tend to do the same thing; for example, looking at the cortical area just in front of the corpus callosum, you can see the inputs from the prefrontal cortex forming a macrocolumn, flanked on either side by macrocolumns formed by a clustering of parietal-lobe inputs.
The cortical neurons interested in color tend to cluster together (though not exclusively) in ‘blobs.’ Unlike macrocolumns, blobs don’t extend through all layers of the cortex; they’re found only in the superficial layers – up there with the interoffice mail. And they’re not exclusively comprised of color specialists: perhaps only 30 percent of the neurons in a blob are color sensitive. The distances between blobs are similar (if not identical) to the those of the macrocolumns.
Next level of organization? On the basis of layer thickness changing, there are fifty-two ‘Brodmann Areas’ in each human hemisphere. At a boundary between Areas, you’ll see the relative thickness of those interoffice-in-out stacked boxes change, as if the relative amounts of incoming, outgoing, and interoffice mail differed on adjacent ‘desks.’
Area 17 is better known as the primary visual cortex, but generally it is premature to put functional labels on these areas in the manner of departments on an organization chart (Area 19, for example, has a half-dozen functional subdivisions). A Brodmann Area averages 21 cm2 in unwrinkled area. If the visual cortex ratio holds elsewhere, that’s on the order of 10,000 macrocolumns and a million minicolumns in the average cortical area.
That factor of a hundred keeps recurring: a hundred neurons to a minicolumn, roughly a hundred minicolumns to a macrocolumn, a hundred times a hundred macrocolumns to a cortical area (which makes me wonder if we’re missing an intermediate ‘super-column’ or ‘mini-area’ organization on the scale of a hundred macrocolumns), and there are just over a hundred Brodmann Areas when you total those in both cerebral hemispheres.
Can we extend this hundredfold multiplier further? It does put us into the scale of social organizations: What’s a hundred brains? That suggests certain legislative bodies such as the US Senate. And the United Nations is representative of more than a hundred legislatures.
Permanent elements of brain organization, such as cortical areas or minicolumns, are nice to know about. But we also need to understand those temporary work spaces of the brain – something closer to scratch pads and buffers – that are likely superimposed on the more permanent forms of anatomical organization.
To deal with the novel, we are indeed going to need some empirical types of organization, like those hexagonal cells that form in the cooking oatmeal when you forget to stir it – forms that are used temporarily and then disappear. Occasionally these forms of organization are recalled to life if some aspect of them earlier formed enough ‘ruts’ in the landscape of interconnection strengths – in which case the empirical organization became a new memory or habit.
In particular, we need to know about the cerebral codes – those patterns that represent each of the words of our vocabulary, and so forth – and what creates them. At first, it appears that we are dealing with a four-dimensional pattern – the active neurons scattered through three-dimensional cortex, as they perform in time. But largely because the minicolumns seem to organize all the cortical layers around similar interests, most people working on cortex think of it as a two-dimensional sheet, rather like the retina (yes, the retina is 0.3 mm thick and is subdivided into a few layers, but the mapping is clearly for a two-dimensional image).
So we can try thinking of two dimensions, plus time, for cortex (which is, of course, the way we apprehend the images on a movie screen or computer terminal) – perhaps with transparent overlays when the different cortical layers do different things. Just imagine the human cortex flattened out on those four sheets of typing paper like pie crust, with little patches lighting up like message-board pixels. What patterns will we observe when that cortex is seeing a comb? When the word ‘comb’ is heard, or said? When the cortex is commanding a hand to comb the hair?
Memory recall may consist of the creation of a spatiotemporal sequence of neuron firings – probably a sequence similar to the firing sequence at the time of the input to memory, but shorn of some of the nonessential frills that promoted it. The recalled spatiotemporal pattern would be something like a message board in a stadium, with lots of little lights flashing on and off, creating an overall pattern. A somewhat more general version of such a Hebbian cell assembly would avoid anchoring the spatiotemporal pattern to particular cells, to make it more like the way the message board can scroll. The pattern continues to mean the same thing, even when it’s implemented by different lights.
Though we tend to focus on the lights that turn on, note that lights that stay off also contribute to the pattern; if they are turned on randomly – by seizures, for example – they fog the pattern. Something similar to this fogging seems to happen in concussions: while an injured football player is being helped off the field, he can often tell you what play he was running, but ten minutes later he can’t remember what happened to him. Injury slowly causes a lot of neurons to ‘light up’ and patterns therefore become obscured in the manner of bright fog – what mountain climbers call ‘white-outs.’ (Just remember: blackouts are sometimes from white-outs.)
What’s the most elementary pattern that means something? A major clue, it seems to me, is that pattern copying is needed, for various reasons.
Before DNA leapt to prominence, geneticists and molecular biologists were searching for a molecular structure that was capable of being reliably copied during cell division. One of the reasons that the double helix structure was so deeply satisfying when it was discovered in 1953 by Crick and Watson (and I write this while temporarily at the University of Cambridge, just across the courtyard from the building where they worked) was that it provided a way of making a copy, via the complementary pairs of DNA bases (C bonds with G, A pairs with T). Unzip the double helix into two separate halves and each DNA position on a half-zipper will soon be paired with another of its opposite type, just from all the loose ones floating around in the nucleotide soup. This gives you two identical double helices, where there was only one before. This copying principle paved the way for the understanding of the genetic code (how those DNA triplets ‘represented’ the amino acid string that folds up into a protein) a few years later.
Is there a similar copying mechanism for cerebral activity patterns, and might it help us identify the most relevant of the Hebbian cell assemblies? That’s the one we could properly call the cerebral code because it is the most elementary way of representing something (a particular connotation of a word, an imagined object, and so forth).
Copying hasn’t been observed in the brain yet – we don’t currently have tools of sufficient spatial and temporal resolution, though we’re close. But there are three reasons why I think it’s a safe bet.
The strongest argument for the existence of copying is the darwinian process itself, which is inherently a copying competition biased by a multifaceted environment. It’s so elementary a method for shaping up randomness into something fancy that it would be surprising if the brain didn’t use it.
Copying is also what’s needed for precision ballistic movements, such as throwing – those dozens-to-hundreds of clones of the movement-command patterns that are required to hit the launch window.
Then there’s that faux fax argument of the last chapter: communication within the brain requires the telecopying of patterns.
Since 1991, my favorite candidate for a local neural circuit that could make copies of spatiotemporal patterns has been the mutually reinforcing circuitry of the interoffice-mail layers. The wiring of those superficial layers of cerebral cortex is, in a word, peculiar. Indeed, to a neurophysiologist, almost alarming. I look at those circuits and wonder how runaway activity is reined in, why seizures and hallucinations aren’t frequent events. But those same circuits have some crystallization tendencies that ought to be particularly good at cloning spatiotemporal patterns.
Of the hundred neurons in a minicolumn, about thirty-nine are superficial pyramidal neurons (that is, their cell bodies reside in the superficial layers II and III). It’s their circuitry that is peculiar.
Like all other pyramidal neurons, they secrete an excitatory neurotransmitter, usually glutamate. There’s nothing peculiar about glutamate per se; it’s one of the amino acids, more typically used as a building block of peptides and proteins. Diffusing across the synapse, the glutamate opens up several types of ion channels through the membrane of the next cell’s dendrite. The first channel specializes in letting sodium ions through; that in turn raises the internal voltage of the downstream neuron.
A second downstream channel activated by glutamate is known as the NMDA channel, and it allows calcium ions into the downstream neuron along with some more sodium. NMDA channels are particularly interesting to neurophysiologists because they contribute to so-called long-term potentiation (LTP), a change in synaptic strength that endures for some minutes in neocortex. (Minutes, actually, are closer to the neurophysiological ‘short-term,’ but LTP sometimes lasts days in the hippocampus – which is an older and simpler version of cortex – and that’s where the ‘long-term’ name originated.)
LTP occurs when there is near synchrony (within dozens to hundreds of milliseconds) of several inputs to the downstream neuron; it simply turns up the ‘loudness control’ for those inputs for a few minutes. These are the ‘bumps and ruts’ that temporarily make it easier to recreate a particular spatiotemporal pattern. LTP is our best candidate for a short-term memory that can survive a distraction. It is also thought to contribute the scaffolding for the construction of truly long-lasting structural changes in synapses – the permanent bumps and ruts that aid in the recreation of long-unused spatiotemporal patterns.
The interoffice layers are where most of the NMDA channels are located, and where most of the neocortical LTP occurs. These superficial layers have two more peculiarities, both of them having to do with the connections that their pyramidal neurons make with one another. On average, a cortical neuron contacts fewer than 10 percent of all neurons within a radius of 0.3 mm. But roughly 70 percent of the excitatory synapses on any superficial-layer pyramidal neuron are derived from pyramidal neurons less than 0.3 mm away, so these neurons may be said to have an unusually strong propensity to excite one another. To a neurophysiologist, that raises all sorts of red flags: it’s a perfect setup for instability and wild oscillations, unless it’s carefully regulated.
There is also a peculiar patterning to these ‘recurrent excitatory’ connections – a patterning not seen in the lower cortical layers. The axon of a superficial pyramidal neuron travels sideways a characteristic distance without making any synapses with other neurons, and then it produces a tight terminal cluster. Like an express train, it skips intermediate stops. In the primary visual cortex, the distance from the cell body to the center of the terminal cluster is about 0.43 mm in primary visual cortex; next door in a secondary visual area, it’s 0.65 mm; in the sensory strip, it’s 0.73 mm; and in motor cortex of monkeys, it’s 0.85 mm. Let me, for convenience, just call this skip-spacing a generic 0.5 mm. The axon may then continue for an identical distance and sprout another terminal cluster, and this express train line may continue for some millimeters.
This skip-spacing is distinctly peculiar in the annals of cortical neuroanatomy. Its function is unknown, but it certainly does make you think that regions 0.5 mm apart might be doing the same thing on occasion – that there could be repeating patterns of activity, in the manner of recurring patterns within wallpaper.
The skip-spacing, you may have noticed, is the same half millimeter or so as the distance between macrocolumns. Color blobs, too, are about that far apart from one another. Yet there’s a difference.
A second superficial pyramidal neuron 0.2 mm from the first will itself have an axon with different express stops, still at 0.5 mm skips but each cluster landing 0.2 mm from those of the first. In my undergraduate days, the Chicago Transit Authority had exactly such a system of A trains and B trains, one taking the ‘even’ stops and the other the ‘odd’ numbered ones, with a few common stops for transferring between trains. Of course, any one subway stop is sometimes stretched out over more than a city block – and our superficial pyramidal neurons are also not located at a single point, as their dendritic tree spreads sideways from the cell body, often 0.1 mm or more.
Contrast this to macrocolumns. So far, they’ve been territories within which there is common source of input, as if you could draw a fence around a group of minicolumns on the basis of their all being on the same mailing list. And the blobs have an output target in common (secondary cortical areas specializing in color). So we’re not talking macrocolumns with our sideways-running excitatory axon branches, though perhaps the skip-spacing is a cause (or effect) of the macrocolumns at an adjacent level of organization. Imagine a forest where tree branches interdigitate, where each tree has a telephone line leaving it and contacting a distant tree, not only bypassing the intermediate ones, but leaping over the common-input fences subdividing the forest.
Sideways ‘recurrent’ connections are common in real neural networks; lateral inhibition was the topic of two Nobel Prizes (to Georg von Békésy in 1961 and H. Keffer Hartline in 1967). It tends to sharpen up fuzzy boundaries in a spatial pattern (while they may compensate for fuzzy optics, they can also produces a few side effects, such as some of the visual illusions). But our superficial pyramidal neurons are excitatory to one another, suggesting that their activity could feed on itself like a spreading brushfire, unless held in check by inhibitory neurons. What’s going on here? Is recurrent excitation why the cerebral cortex is so prone to epileptic seizures, when the inhibitory neurons are fatigued?
Furthermore, the standard skip-spacing means that a round-trip might be possible – a reverberating circuit, of the kind postulated by early neurophysiologists. Two neurons that are 0.5 mm apart may keep each other going. A neuron has a refractory period – a kind of ‘dead zone’ – after an impulse is produced: for a millisecond or so, it is almost impossible to initiate another impulse. The travel time over that 0.5 mm is also about a millisecond, and then the synaptic delay slows delivery by another half a millisecond – so if the connections between the two neurons were otherwise strong enough, you can imagine the second neuron’s impulse getting back to the first neuron about the time it has recovered its ability to generate another impulse. But usually connection strength between neurons isn’t strong enough, and usually such rapid firing cannot be kept up, even if it does get started. (In the heart, however, connection strengths between adjacent cells are indeed strong enough, and circus re-excitation is an important pathology when injury slows travel times.)
If the implication of cortex’s standard skip-spacing isn’t an impulse chasing its tail, then what is it? Probably synchronization.
If you sing in the chorus, you get in sync with the others by hearing them – usually hearing yourself coming in too late or starting too early. But you, of course, are also influencing them. Even if everyone is a little hard of hearing, everyone soon gets synchronized, thanks to all that feedback.
Your position in that chorus is very much like that of a superficial pyramidal neuron in the neocortex, getting excitatory inputs from neighbors on all sides. Networks like this have been extensively studied, even if the one in superficial neocortex has not; synchronization will occur even with only small amounts of feedback (which is why I postulated that you were hard of hearing, just then). Two identical pendulums will tend to synchronize if they are adjacent, just from the air and shelf vibrations they create. Menstrual cycles are said to synchronize in women’s dormitories. Though harmonic oscillators, such as the pendulums, take a while to get in sync, nonlinear systems, such as impulse production in neurons, can synchronize very quickly, even if the mutual connection strengths are relatively weak.
And what does this tendency to synchronize have to do with copying spatiotemporal patterns? Happily, it’s all a matter of simple geometry, the kind that the ancient Greeks discovered while staring at the tile mosaics of their bathhouse floors (and that many of us have rediscovered in wallpaper patterns).
Let us suppose that a ‘banana committee’ is forming among all the superficial pyramidal neurons scattered around the primary visual cortex that respond to one feature or another of the banana you’re looking at. The lines forming the outline of the banana are a particularly effective prod to those neurons that specialize in boundaries and their orientation. Then there are those blob neurons that like yellow.
Since they tend to excite one another, given that 0.5 mm skip-distance for their axon terminal clusters, there is going to be a tendency for them to synchronize – not that all impulses in the neuron I’ll call Yellow One will be synchronized with those in Yellow Two, but a certain percentage will occur within a few milliseconds of one another.
Suppose now that there is another superficial pyramidal neuron, 0.5 mm equidistant from both Yellow One and Yellow Two. Perhaps it only receives a weak yellow input, so that it isn’t actively firing away, signaling yellow. Now, however, Yellow Three is getting inputs from both One and Two. Furthermore, some of those inputs from One and Two – the synchronized ones – will arrive at Three’s dendrites together. (They both have the same 0.5 mm travel distance.) This is exactly what hi-fi buffs call ‘sitting in the hot spot,’ equidistant from both speakers at the apex of an equilateral triangle (move even slightly to either side and the stereo illusion collapses into the nearest speaker, leaving you with mono sound). At the cortical hot spot near Three, the two synaptic inputs summate, 2 + 2 = 4 (approximately). But the distance remaining to the impulse threshold may be 10, so Three still remains silent.
Not very interesting. But these are glutamate synapses in the superficial cortical layers, so they’ve got NMDA channels across the synapse to let both sodium and calcium into the downstream neuron. Again, not so important – by itself.
But I temporarily omitted telling you why neurophysiologists find NMDA channels so fascinating compared with other synaptic channels: they are sensitive not only to arriving glutamate but also to the preexisting voltage across the post-synaptic membrane. Raise that voltage and the next glutamate to arrive will cause a bigger effect, sometimes twice the standard amount. This is because many of the NMDA channels are normally sitting there plugged up: there’s a magnesium ion stuck in the middle of the tunnel through the membrane; increased voltage will pop it out of there – and that in turn allows formerly blocked sodium and calcium to flow into the dendrite on the next occasion when arriving glutamate opens up the gates.
The consequence of this is important: it means that synchronously arriving impulses are more effective than 2 + 2 would predict: the sum could be 6 or 8 instead (welcome to nonlinearity). Repeated near-synchronization of two inputs is even more effective, as they clean the magnesium plugs out of each other’s channels. Pretty soon, those repeatedly synchronous inputs from Yellow One and Yellow Two might be able to trigger an impulse from Yellow Three.
Standard-distance mutual re-excitation and NMDA synaptic strength augmentation have this interesting hand-and-glove fit, all because of the tendency to synchronize. Emergent properties often come from just such combinations of the seemingly unrelated.
We now have three active neurons, forming the corners of an equilateral triangle. But there might be a fourth, over on the other side of One and Two, also equidistant at 0.5 mm away. There isn’t very much data yet on how many axon branches there are from a single superficial pyramidal neuron – but looking down from the top, in one dye-filled and exhaustively reconstructed superficial pyramidal neuron, showed branches in many directions. So there ought to be a doughnutlike ring of excitation, about 0.5 mm away from the neuron. Two such rings, with centers 0.5 mm apart – as from Yellow One and Two – have two intersections, just as in that plane geometry exercise about bisecting a line.
So it wouldn’t be surprising if Yellow One and Yellow Two, once they got their act in sync, managed to recruit a Yellow Four as well as a Yellow Three. And there are other neurons at the hot spot of the pair formed by One and Three: perhaps a Yellow Five will join the chorus, if it already has enough other inputs to put the paired inputs within range of its threshold. As you can see, there is a tendency to form a triangular array of often-synchronized neurons that could extend for a few millimeters across the cortical surface.
Because one neuron can become surrounded by six others, all telling it to fire at a certain time, we have error correction: even if a neuron tries to do something different, it is forced back to the choral pattern that has become established by its insistent neighbors. That is essentially an error-correction procedure, just what the faux fax needs – if only the long corticocortical axon terminals did what the local ones do: fan out into patches about 0.5 mm apart rather than ending in a point.
And they do fan out in a patchy manner – in about the right ballpark.
The notion of ‘convergence zones’ for associative memories raises the issue of maintaining the identity of a spatiotemporal code during long-distance corticocortical messaging, such as through the corpus callosum from the left side to the right of the brain. Distortions of the spatiotemporal pattern by a lack of precise topographic mappings (axon terminations are always fanning out, not ending in a single point), or a dispersion in time (conduction velocities are not uniform), might be unimportant where the information flows in only the one direction – in that case, one arbitrary code is simply replaced by another arbitrary code in that pathway.
But because the connections between distant cortical regions are typically (six in every seven paths) reciprocal, any distortions of the original spatiotemporal firing pattern during forward transmission would need to be compensated in the reverse path, in order to maintain the characteristic spatiotemporal pattern as the local code for a sensory or motor schema. You could straighten out the distortion with an inverse transform, just as in decompressing a compressed file. Or you could fix it with the aforementioned error-correction mechanism. Or you could just live with different codes locally meaning the same thing, like names and nicknames – what’s called a degenerate code, as when six different DNA triplets all code for leucine. I used to think that either alternative was more likely than an error-correction scheme, but then I didn’t realize how simply error correction could emerge from the crystallization that ought to accompany recurrent excitation and synchrony-sensitive NMDA channels.
Imagine an optical fiber array connecting one cortical area with its homologous one on the other side. Real optical fiber bundles subdivide an image into dots, then faithfully pipe each dot a long distance so that, looking at the end of the fiber bundle, you see a pattern of lighted dots identical to that at the front end.
An axon is not like a light pipe because of all the ‘sprouts’ at each end. It doesn’t end in a point: a single axon fans out into many terminals, spreading over macrocolumn dimensions. Bundles of real axons also aren’t like a coherent fiberoptic bundle, where neighbors remain faithful neighbors; real axons can get intertwined with one another, so that a dot goes astray and ends up displaced at the other end. Real axons also vary somewhat in conduction velocity: impulses that started out together may arrive at different times, distorting the spatiotemporal pattern.
But the local error-correction property suggests that none of this might matter very much, at the far end of a corticocortical bundle. What’s being sent is a redundant spatiotemporal pattern, thanks to those triangular arrays on the originating end. Each point on the distant end might get an input from a dead-on-target axon, plus up to six inputs from neighbors 0.5 mm away back home; yes, some get lost, and some impulses arrive too late, but a receiving neuron preferentially pays attention to the repeatedly synchronous inputs, and perhaps only a few of them are needed to reproduce the firing pattern of the originating point, effectively ignoring the stragglers and the wanderers.
Once a small region of spatiotemporal pattern is re-formed again on the distant end, it can expand to clone a larger territory, just as explained earlier. So synchronized triangular arrays make it possible for sloppy wiring to send spatiotemporal patterns over long distances in the cortex – provided you start with a dozen or so spatial repeats of the spatiotemporal pattern and end with a sufficient territory of the same pattern on the far end.
How big might an array become? It might be confined to its original Brodmann area, if the skip-spacing changed at the boundary. For example, in primary visual cortex, the skip-spacing in monkeys is 0.43 mm, and next door in the secondary visual area it’s 0.65 mm; recruiting across the boundary might not work, but that’s an empirical question – we’ll just have to see. And recruiting more neurons into the triangular array requires candidates that are already mildly interested in the banana.
So the triangular array of Yellows might not be too much larger than the part of the visual cortex receiving the image of the yellow banana. The neurons sensitive to line orientation might have been doing the same thing, too: several getting in sync, recruiting a chorus of the predisposed, and so forming another 0.5 mm triangular array centered elsewhere. For each separately detected feature of the banana, there would be a different triangular array – and not necessarily extending the same distances across the cortex. Looking down on the flattened cortex (and assuming that a minicolumn lights up when an impulse fires), we would see a lot of flickering lights.
If we restricted our field of view to a 0.5 mm circle, we would be unlikely to see much synchrony, just one of the Yellows firing a few times a second, one of the Lines firing a dozen times a second, and so on. But if we broadened our field of view to several millimeters, we would see a half-dozen spots lighting up at once, then another group lighting up. Each specialty has its own triangular array; taken together, the various arrays constitute a Banana Committee.
Note that the original committee of Yellows and Lines might have been larger than 0.5 mm across, back before recruitment began to fill in things. Even if the original committee was scattered over a few millimeters, the triangular arrays serve to create a unit pattern that is much smaller (and potentially easier to recreate, when recalling the pattern). We have, in effect, compacted the code into a smaller space than it originally occupied, as well as making redundant copies. That has some interesting implications.
This is a spatiotemporal pattern having something to do with a banana’s representation, but is it the cerebral code for banana? I’d call that the smallest such pattern that didn’t omit anything important – the elementary pattern from which the triangular arrays of Lines and Yellow could be recreated.
If we zoom in, shrinking our field of view of the flickering minicolumns, what area will it cover at the point where we can no longer find synchronized minicolumns? Yes, it’s about 0.5 mm, but it isn’t a 0.5 mm circle – it is a hexagon that is 0.5 mm across between parallel faces. This is a simple matter of geometry: corresponding points (say, the upper right corners) of hexagonal tiles form triangular arrays. Anything larger than that hexagon will start including some redundant points that are represented by another of their triangular array (and we’d sometimes see two synchronized points in our restricted view of view).
The elementary pattern wouldn’t usually fill the hexagon (I imagine it as a dozen minicolumns active, out of a hundred or more in the hexagon – but the rest have to stay silent in order not to fog the pattern). We wouldn’t be able to see the boundaries outlined – so that we wouldn’t see a honeycomb when we looked down on the cortical surface while a territory was being cloned. Indeed, when wallpaper designers create a repeating pattern, they often make sure that the pattern unit’s boundary cannot be easily detected, so that the overall pattern will appear seamless. Though the triangular arrays do the recruitment and create the compact pattern, it is as if hexagons were being cloned.
The triangular synchronicity doesn’t necessarily last for very long – it’s an ephemeral form of organization, and it might be wiped out during certain phases of an EEG rhythm associated with decreases in cortical excitability. If we want to recreate a spatiotemporal pattern that has died out, we could get it started from within two adjacent hexagons – indeed, from any two adjacent hexagons that the extended banana mosaic covered originally. It wouldn’t have to be the original pairs. The memory trace – the essential bumps and ruts for resurrecting the spatiotemporal pattern – could be as small as the circuitry in two adjacent hexagons.
So repeated copying of the minimal pattern could colonize a region, in much the same way that a crystal grows or wallpaper repeats an elementary pattern. If the melody recurred enough times before it stopped, LTP might linger in such a way that the spatiotemporal pattern was easily restarted, at one place or another.
If the spatial pattern was relatively sparse, several cerebral codes (say, the ones for Apple and Tangerine) could be superimposed to give you a category, such as Fruit. If you tried superimposing several letters from a dot-matrix printer, you’d get a black mess. But if the matrix is sparsely filled, you can probably recover the individual members, because they each produce such distinctive spatiotemporal patterns. So this type of code could also be handy for forming categories that could be decomposed into examples, just as superimposed melodies can often be heard individually. Thanks to the telecopying aspect, you could form multimodal categories – such as all of the connotations of comb.
My friend Don Michael suggests that meditation might correspond to creating, via a mantra, a large mosaic of a nonsense code, one without significant resonances or associations. If you maintained it long enough to wipe the slate clean of cares and preoccupations, allowing those short-term ruts to fade, it might give you a fresh start in accessing long-term memory ruts without getting hung up on short-term concerns.
[Meditation’s] exquisite state of unconcerned immersion in oneself is not, unfortunately, of long duration. It is liable to be disturbed from inside. As though sprung from nowhere, moods, feelings, desires, worries and even thoughts incontinently rise up, in a meaningless jumble, and the more far-fetched and preposterous they are, and the less they have to do with that on which one has fixed one’s consciousness, the more tenaciously they hang on. . . . The only successful way of rendering this disturbance inoperative is to keep on breathing, quietly and unconcernedly, to enter into friendly relations with whatever appears on the scene, to accustom oneself to it, to look at it equably and at last grow weary of looking.
Eugen Herrigel, Zen in the Art of Archery, 1953
There are some attractive features to what emerges from this analysis of the superficial pyramidal neurons: Donald Hebb would have loved it, because it shows how some of the most puzzling features of short-and long-term memory might be explained with cell assemblies (the memory trace is stored in a distributed way, with no one site crucial for its recall, and so forth). The gestalt psychologists would have liked the way that it makes possible the comparison of figure and ground by the triangular arrays potentially extending beyond the object boundaries, a spatiotemporal pattern forming that represents the figure-ground combination, rather than just one or the other.
And I like to think that Charles Darwin and William James would have liked the idea that mental life involves copying competitions biased by a multifaceted environment. Sigmund Freud might have been intrigued with the mechanism it suggests for how subconscious associations could occasionally pop into the foreground of consciousness.
While I think that divergent thinking is the most important application of the neocortical Darwin Machine, let me first illustrate how it might work with a convergent thinking problem. Suppose that something whizzes past you and disappears under a chair. You thought it was round, and maybe orange or yellow, but it was moving very quickly, and now it’s out of sight, so you can’t get a second look. What was it? How do you guess, when the answer isn’t obvious? Your process first needs to find some candidates, then it needs to compare them for reasonableness.
Happily, cloning competitions can do that. There’s a tentative cerebral code for the object, formed by all the feature detectors that it activated: color, shape, motion, and maybe the sound of it bouncing on the floor. This spatiotemporal pattern starts making clones of itself, in a manner of speaking.
Whether it can set up a clone next door depends on the resonances next door, those bumps in the road provided by the pattern of synaptic strengths and by whatever else is going on in the adjacent cortex. If you’d seen such an object many times before, there might be a perfect resonance – but you haven’t. Still, the tentative cerebral code has specialty components of Round, Yellow, Fast. Tennis balls have such attributes, and you have a good tennis-ball resonance, so the adjacent area pops into the melody for Tennis Ball (a nice feature of attractors is that a near fit can be captured and transformed to the characteristic pattern). Cloning with poor resonances leads to dropouts of some components, so perhaps your Tangerine resonance captures a variant in another patch of cortex despite the color not being quite right.
What about cloning competitions? Here we have Unknown, Tennis Ball, and Tangerine cerebral codes cloning away. Perhaps Apple pops out as well: if you saw someone eating an apple a few minutes ago, there would be temporary ruts for Apple, because of the NMDA synapses that were strengthened in that pattern. But then Apple is overrun by the Tangerine pattern, which is cloning away. Over on the other side of Unknown’s current territory, Tennis Ball is doing quite well and eventually it overrides and replaces Unknown, even encroaching on Tangerine’s territory. At about this time, you say, ‘I think that was a tennis ball,’ because there were finally enough clones in the Tennis Ball chorus to get a coherent message through to your left-lateral language cortex, over the corticocortical pathways from the occipital lobe to the temporal lobe.
Something else happens now: a new spatiotemporal pattern starts cloning through the work space; this time, you see something very familiar (the chair) and a critical chorus of Chair is quickly established without any real competition, because the sensory spatiotemporal pattern hits instantly on a resonance before any variants have time to get going. The NMDA synapses used in the Tennis Ball and Tangerine patterns are still jazzed up, however, and for another five minutes it will be easier than usual to recreate either of these spatiotemporal patterns in the parts of the work space that they last occupied. Perhaps Tangerine continues to clone and make mistakes, hitting upon the Orange Fruit resonance – so that a minute later, you wonder if you were wrong about that tennis ball.
That’s how it could happen – how I imagine that our subconscious processes sometimes come up with someone’s name a half hour too late. The pattern resonances are not unlike how we imagine locomotion to work in the spinal cord: there’s a connectivity – all those synaptic strengths between the various neurons – and, given certain initial conditions, you can pop into the resonance for the spatiotemporal pattern that implements Walk. With other initial conditions, you pop instead into Jog, Lope, Run, or Hopscotch.
In the sensory cortex, you may pop into Orange or Tangerine even when you see fruit that is neither. As I mentioned in chapter 5, categories are why the Japanese have so much trouble with English L and R sounds: both are captured by their mental category for a particular Japanese phoneme. Reality is quickly replaced by mental models. As Henry David Thoreau said, ‘We hear and apprehend only what we already half know.’
The cortex is in the business of quickly learning new patterns, whether sensory or movement, and creating variations on them. The variations allow for competitions to determine what pattern best resonates with the connectivities, and these in turn are often biased by a number of sensory inputs and emotional drives.
Relationships, too, can be coded by spatiotemporal patterns – just as sensory or movement schemas are. Just combine codes to make a new arbitrary pattern, in the same way that a left-hand rhythm can be superimposed on a right-hand melody.
The lingua ex machina of chapter 5 offered some specific examples of what fancy relationships (as in a sentence) might involve – all those obligatory and optional roles. Those obligatory arguments of a verb such as give are about relationships, and cognitive dissonance results when an obligatory role goes unfilled (as, alas, advertising agencies have discovered; Give Him forces you to read the billboard again to see what you missed, and thereby remember the ad better).
So, is a sentence simply one big spatiotemporal pattern, cloning away in competition with other sentence codes? Yes, but not always. We don’t require copying competitions in order to make a decision; simple rating schemes ought to suffice, if nothing particularly new is involved. Recall the cormorant of chapter 2: rating schemes will do for its decision making, because the choices (swim, dive, dry wings, fly away, look around a little longer) are already well shaped by evolution over the generations. Copyable schemas aren’t everything, once you get close enough to standard meanings.
The superficial cortical layers in many primates have the standard skip-wiring that predicts the ephemeral triangular arrays. It is not known how often any animal uses this wiring for cloning wallpaperlike hexagonal patterns; perhaps it only happens briefly, during prenatal development – as a sort of test pattern that guides use-dependent connections – and never occurs again. Or perhaps some areas of cortex are committed to full-time specialization and never clone such ephemeral patterns, while other areas often support sideways copying and become erasable workspaces for darwinian shaping-up processes. Since clones of movement commands would be particularly useful for throwing – they can reduce timing jitter – perhaps there was some natural selection for big work spaces in the hominid evolution of throwing accuracy. They’re all empirical questions. Once we have improved the resolution of our recording techniques, we’ll have to see where hexagonal cloning lies on the spectrum of possibilities.
But something very close to such cloning competitions is needed to satisfy the essentials for the Darwin Machine – that’s the real reason why I’ve led the reader through this cortical maze. Here, at least, we have (1) a distinctive pattern; (2) copying; (3) variation; (4) possible competitions for work spaces; (5) multifaceted environments (both current and memorized) to bias the competition; and (6) a next generation more likely to have variants established from the clones with the biggest territories (big territories have more perimeter, which is where variants can escape error-correction tendencies and get started cloning their new pattern).
In a longer book about the neocortical Darwin Machine itself (The Cerebral Code), I’ll explain all about the spice and speed that you’d get from cerebral analogues of sex, islands, and climate change. And speed we need, if a darwinian process in the brain is to work fast enough to provide our guessing-right intelligence.
We keep trying to carve up the cerebral cortex into specialized ‘expert’ modules. It’s a good research strategy to look for specialization, but I don’t take it seriously as an overview of how the association cortex works. We need some erasable work spaces, and we need to be able to recruit helpers for difficult tasks. That suggests that any ‘expert’ modules are also generalists – as when a neurosurgeon acts as a paramedic in an emergency. One of the things I like about the ephemeral hexagonal mosaic is that it suggests a resolution of the expert-generalist dilemmas: even a cortical area with ‘expert’ long-term ruts could serve as a work space, using overlaid short-term ruts to bias competitions.
Such a mosaic also suggests a way that subconscious thoughts could meander and occasionally pop some relevant fact from the past into your stream of consciousness. Best of all, because variants themselves can clone their way to temporary success, the patchwork quilt is creative – it can be shaped up from humble beginnings into something of quality. Even higher forms of relationships, such as metaphor, seem likely to arise, because the cerebral codes are arbitrary and capable of forming new combinations. Who knows – perhaps by now you’ve even acquired a cerebral code for Umberto Eco’s Mac-PC analogy.
The synchronized triangular arrays with such interesting implications for darwinian copying competitions turn out to have implications for fancy language as well, potentially giving intelligence a boost from another direction.
0 notes
somerabbitholes · 3 years
Text
Essays
Here’s a (non-exhaustive) list of essays I like/find interesting/are food for thought; I’ve tried to sort them as much as possible. The starred (*) ones are those I especially love
also quick note: some of these links, especially the ones that are from books/anthologies redirect you to libgen or scihub, and if that doesn’t work for you, do message me; I’d be happy to send them across!
Literature + Writing
Godot Comes to Sarajevo - Susan Sontag
The Strangeness of Grief - V. S. Naipaul*
Memories of V. S. Naipaul - Paul Theroux*
A Rainy Day with Ruskin Bond - Mayank Austen Soofi
How Albert Camus Faced History - Adam Gopnik
Listen, Bro - Jo Livingstone
Rachel Cusk Gut-Renovates the Novel - Judith Thurman
Lost in Translation: What the First Line of “The Stranger” Should Be - Ryan Bloom
The Duke in His Domain - Truman Capote*
The Cult of Donna Tartt: Themes and Strategies in The Secret History - Ana Rita Catalão Guedes
Never Do That to a Book - Anne Fadiman*
Affecting Anger: Ideologies of Community Mobilisation in Early Hindi Novel - Rohan Chauhan*
Why I Write - George Orwell*
Rimbaud and Patti Smith: Style as Social Deviance - Carrie Jaurès Noland*
Art + Photography (+ Aesthetics)
Looking at War - Susan Sontag*
Love, sex, art, and death - Nan Goldin, David Wojnarowicz
Lyons, Szarkowski, and the Perception of Photography - Anne Wilkes Tucker
The Feminist Critique of Art History - Thalia Gouma-Peterson, Patricia Mathews
In Plato's Cave - Susan Sontag*
On reproduction of art (Chapter 1, Ways of Seeing) - John Berger*
On nudity and women in art (Chapter 3, Ways of Seeing) - John Berger*
Kalighat Paintings  - Sharmishtha Chaudhuri
Daydreams and Fragments: On How We Retrieve Images From the Past -  Maël Renouard
Arthur Rimbaud: the Aesthetics of Intoxication - Enid Rhodes Peschel
Cities
Tragic Fable of Mumbai Mills - Gyan Prakash
Whose Bandra is it? - Dustin Silgardo*
Timur's Registan: noblest public square in the world? - Srinath Perur
The first Starbucks coffee shop, Seattle - Colin Marshall*
Chhatrapati Shivaji Terminus, Mumbai's iconic railway station - Srinath Perur
From London to Mumbai and Back Again: Gentrification and Public Policy in Comparative Perspective -  Andrew Harris
The Limits of "White Town" in Colonial Calcutta - Swati Chattopadhyay
The Metropolis and Mental Life - Georg Simmel
Colonial Policy and the Culture of Immigration: Citing the Social History of Varanasi - Vinod Kumar, Shiv Narayan
A Caribbean Creole Capital: Kingston, Jamaica - Coln G. Clarke (from Colonial Cities by Robert Ross, Gerard J. Telkamp
The Colonial City and the Post-Colonial World - G. A. de Bruijne
The Nowhere City - Amos Elon*
The Vertical Flâneur: Narratorial Tradecraft in the Colonial Metropolis - Paul K. Saint-Amour
Philosophy
The trolley problem problem - James Wilson
A Brief History of Death - Nir Baram
Justice as Fairness: Political not Metaphysical - John Rawls*
Should Marxists be Interested in Exploitation? - John E. Roemer
The Discomfort You’re Feeling is Grief - Scott Berinato*
The Pandemic and the Crisis of Faith - Makarand Paranjape
If God Is Dead, Your Time is Everything - James Wood
Giving Up on God - Ronald Inglehart
The Limits of Consensual Decision - Douglas Rae*
The Science of "Muddling Through" - Charles Lindblom*
History
The Gruesome History of Eating Corpses as Medicine - Maria Dolan
The History of Loneliness - Jill Lepore*
From Tuskegee to Togo: the Problem of Freedom in the Empire of Cotton - Sven Beckert*
Time, Work-Discipline, and Industrial Capitalism - E. P. Thompson*
All By Myself - Martha Bailey*
The Geographical Pivot of History - H. J. Mackinder
The sea/ocean
Rim of Life - Manu Pillai
Exploring the Indian Ocean as a rich archive of history – above and below the water line - Isabel Hofmeyr, Charne Lavery
‘Piracy’, connectivity and seaborne power in the Middle Ages - Nikolas Jaspert (from The Sea in History)*
The Vikings and their age - Nils Blomkvist (from The Sea in History)*
Mercantile Networks, Port Cities, and “Pirate” States - Roxani Eleni Margariti
Phantom Peril in the Arctic - Robert David English, Morgan Grant Gardner*
Assorted ones on India
A departure from history: Kashmiri Pandits, 1990-2001 - Alexander Evans *
Writing Post-Orientalist Histories of the Third World - Gyan Prakash
Empire: How Colonial India Made Modern Britain - Aditya Mukherjee
Feminism and Nationalism in India, 1917-1947 - Aparna Basu
The Epic Riddle of Dating Ramayana, Mahabharata - Sunaina Kumar*
Caste and Politics: Identity Over System - Dipankar Gupta
Our worldview is Delhi based*
Sports (you’ll have to excuse the fact that it’s only cricket but what can i say, i’m indian)
'Massa Day Done:' Cricket as a Catalyst for West Indian Independence: 1950-1962 - John Newman*
Playing for power? rugby, Afrikaner nationalism and masculinity in South Africa, c.1900–70 - Albert Grundlingh
When Cricket Was a Symbol, Not Just a Sport - Baz Dreisinger
Cricket, caste, community, colonialism: the politics of a great game - Ramachandra Guha*
Cricket and Politics in Colonial India - Ramchandra Guha
MS Dhoni: A quiet radical who did it his way*
Music
Brega: Music and Conflict in Urban Brazil - Samuel M. Araújo
Color, Music and Conflict: A Study of Aggression in Trinidad with Reference to the Role of Traditional Music - J. D. Elder
The 1975 - ‘Notes On a Conditional Form’ review - Dan Stubbs*
Life Without Live - Rob Sheffield*
How Britney Spears Changed Pop - Rob Sheffield
Concert for Bangladesh
From “Help!” to “Helping out a Friend”: Imagining South Asia through the Beatles and the Concert for Bangladesh - Samantha Christiansen 
Gender
Clothing Behaviour as Non-verbal Resistance - Diana Crane
The Normalisation of Queer Theory - David M. Halperin
Menstruation and the Holocaust - Jo-Ann Owusu*
Women’s Suffrage the Democratic Peace - Allan Dafoe
Pink and Blue: Coloring Inside the Lines of Gender - Catherine Zuckerman*
Women’s health concerns are dismissed more, studied less - Zoanne Clack
Food
How Food-Obsessed Millennials Shape the Future of Food - Rachel A. Becker (as a non-food obsessed somewhat-millennial, this was interesting)
Colonialism's effect on how and what we eat - Coral Lee
Tracing Europe's influence on India's culinary heritage - Ruth Dsouza Prabhu
Chicken Kiev: the world’s most contested ready-meal*
From Russia with mayo: the story of a Soviet super-salad*
The Politics of Pancakes - Taylor Aucoin*
How Doughnuts Fuelled the American Dream*
Pav from the Nau
A Short History of the Vada Pav - Saira Menezes
Fantasy (mostly just harry potter and lord of the rings)
Purebloods and Mudbloods: Race, Species, and Power (from The Politics of Harry Potter)
Azkaban: Discipline, Punishment, and Human Rights (from The Politics of Harry Potter)*
Good and Evil in J. R. R. Tolkien's Lengendarium - Jyrki Korpua
The Fairy Story: J. R. R. Tolkien and C. S. Lewis - Colin Duriez (from Tree of Tales)*
Tolkien’s Augustinian Understanding of Good and Evil: Why The Lord of the Rings Is Not Manichean - Ralph Wood (from Tree of Tales)*
Travel
The Hidden Cost of Wildlife Tourism
Chronicles of a Writer’s 1950s Road Trip Across France - Kathleen Phelan
On the Early Women Pioneers of Trail Hiking - Gwenyth Loose
On the Mythologies of the Himalaya Mountains - Ed Douglas*
More random assorted ones
The cosmos from the wheelchair (The Economist obituaries)*
In El Salvador - Joan Didion
Scientists are unravelling the mystery of pain - Yudhijit Banerjee
Notes on Nationalism - George Orwell
Politics and the English Language - George Orwell*
What Do the Humanities Do in a Crisis? - Agnes Callard*
The Politics of Joker - Kyle Smith
Sushant Singh Rajput: The outsider - Uday Bhatia*
Credibility and Mystery - John Berger
happy reading :)
34K notes · View notes
Text
This character will quickly tug at your heart strings.
Meet Avaline, a young maidservant in the Royal Estate.
~~The girl nodded and coughed a little, the scar on her throat seeming to irritate her. I looked at Lita. “Could you go and get me some honey and milk? It’ll help to soothe her throat.” The woman nodded and left the room. I gently brought the girl over to my bed and sat her down. “Some honey will do you some good–” The girl suddenly grabbed my arm, her eyes frantic; I could feel her panic as if it were my own, my nerves jolting with a fight-or-flight response.
“You’re in danger.” She whispered, surprising me. With the severity of her scar, I would have thought that her vocal chords had been cut; “don’t trust the man in white. He hurt the other nice lady.” My eyes widened, wondering how she could remember Luxia when no one else seemed to.~~
Follow my socials:
kathleenilyons.writer (Instagram)
ArtKitty97 (Twitter)
and read the beta version
of Arboriah Lux
on Wattpad (Kathleen Lyons)
35 notes · View notes
Text
🌈 LGBTQIA+ Arthuriana in academia
it’s technically not pride month where i live, but aaaahhh i still love talking about this topic, so here’s some queer - or queer-adjacent - Arthurian articles i’ve stumbled across over the last few years 
general heads up that a) I haven’t read all of these, b) some of them definitely handle the topic better than others, and c) as you might expect these aren’t always safe-for-work!
(under the cut bc WOW this got long)  ❤️💛💚💙💜
Sir Gawain and the Green Knight
The Perverse Dynamics of "Sir Gawain and the Green Knight" by Gail Ashton
A Kiss Is Just a Kiss: Heterosexuality and Its Consolations in Sir Gawain and the Green Knight by Carolyn Dinshaw
Sodomy, Misogyny, and Displacement: Occluding Queer Desire in "Sir Gawain and the Green Knight" by David L. Boyd
Getting Medieval: Pulp Fiction, Gawain, Foucault by Carolyn Dinshaw (chapter in a book)
On Lesbian and Gay/Queer Medieval Studies by David Lorenzo Boyd
A Queer World: Feminine Subversions of Chivalric Homosocial Normativity by Jessica Pitts (someone’s masters thesis...im shameless sorry)
Lancelot stories
The Monstrous and the Queer (medieval studies blog, post by Jeffrey Cohen)
The Prose "Lancelot's" Galehot, Malory's Lavain, and the Queering of Late Medieval Literature by Gretchen Mieszkowski
Male Same-Sex Desire in the Romances of de Troyes by Basil A. Clark
'You Could Shame the Great Arthur Himself': A Queer Reading of Lancelot from BBC's "Merlin" with Respect to the Character in Malory, White, and Bradley by Joseph Brennan
Refashioning Courtly Love: Lancelot as Ladies’ Man or Lady/Man? by E. Jane Burns (...yeah, that title...i know)
Tristan stories
Bodies That Don’t Matter: Heterosexuality before Heterosexuality in Gottfriend’s Tristan by James A. Schultz
Between Knights: Triangular Desire and Sir Palomides in Sir Thomas Malory's "The Book of Sir Tristram de Lyones" by Olga Burakov Mongan
Grail stories
'A Mayde, and Last of Youre Blood': Galahad's Asexuality and its Significance in "Le Morte Darthur" by Megan Arkenberg
The Writable Lesbian and Lesbian Desire in Malory’s Morte DArthur by Kathleen Coyne Kelly
Chaste Subjects: Gender, Heroism, and Desire in the Grail Quest by Peggy Mccracken
Queer Fisher King: Castration as a Site of Queer Representation ("Perceval, Stabat Mater, The City of God") by Anna Roberts
Sodomy, Masculinity, and Law in Medieval Literature: France and England, 1050–1230 by William Burgwinkle (book - has a chapter that deals with one of the Perceval continuations!!)
Roman de Silence
The Third Path: Alternative Sex, Alternative Gender in "Le Roman de Silence" by Elizabeth A. Waters
The Importance of Being Gender 'Stable': Masculinity and Feminine Empowerment in "Le Roman de Silence" by Lorraine Kochanske Stock
Queering Gender and Naturalizing Class in the "Roman de Silence" by Robert L.A. Clark
Lesbian Desire in the Old French "Roman de Silence" by Kathleen M. Blumreich
General medieval
Eleanor Janega’s stuff as a starting point on medieval sexuality: That’s not what sodomy is, but OK, The History of Penis in Vagina as Default Sex, Considering bad motherfuckers: Hildegard of Bingen and Janelle Monáe, The Medieval Podcast – Medieval Sexuality 
Heterosexuality as a Threat to Medieval Studies by James A. Schultz
Crossing Disciplinary Boundaries: A Cross-Cultural Approach to Same-Sex Love Between Women by Sahar Amer (this is part of a larger book!!)
Befriending the Medieval Queer: A Pedagogy for Literature Classes by Richard E. Zeikowitz
Queering the Medieval by Glenn Burger and Steven F. Kruger (book)
Same-Sex Love and Desire Among Medieval Women by Francesca Canadé-Sautman and Pamela Sheingorn (weird stuff when u try google it...no idea what’s going on there...published 2090...written by NA...?????)
uuuhhmm only a handful of these are open-access im sorry 😔😔 but often u can still find part or all of them online if u dig (or feel free to hmu if u want me to try share a pdf, rip)
356 notes · View notes
anntickwittee · 4 years
Text
Media Coping Tool Box
TV Rewatch:
Binges:
Gilmore Girls (Netflix)
Chuck (Amazon prime)
Psych (Amazon prime)
Parks and Rec (Netflix)
Schitt’s Creek (Netflix)
Brooklyn Nine Nine (Hulu)
Superstore (Hulu)
The Good Place (Netflix)
The Marvelous Mrs. Maisel (Amazon Prime)
Watchmen (HBO)
Pose (Netflix)
Doctor Who 2020 (Spyfall and on)
One Day At A Time (Netflix)
I Love Lucy
New Cartoons:
Ducktales 2017 (disney+)
Infinity Train (cable streaming, Hulu)
She-Ra and the Princesses of Power (Netflix)
Tangled the Series (disney+)
Green Eggs and Ham (Netflix)
The Dragon Prince (Netflix)
Miraculous Ladybug (Netflix)
Carmen Sandiego (Netflix)
Older Cartoons:
Bob’s Burgers
Steven Universe 
Black Panther
Gravity Falls
Young Justice S1&2 
The Spectacular Spider-Man  
Phineas and Ferb 
Ben 10
American Dragon: Jake Long
Teen Titans
Avatar the Last Airbender 
Danny Phantom
Fillmore!
Kim Possible
Jackie Chan Adventures
House of Mouse
The Wild Thornberries
Hey Arnold
DC cartoons: Batman the Animated Series, Batman Beyond, Static Shock, Justice League and Justice League Unlimited
Scooby Doo Where are You
Looney Tunes
Anime:
Fullmetal Alchemist: Brotherhood
Inuyasha
Sailor Moon Crystal
Akatsuki No Yona
Ouran High School Host Club
Boku No Hero Academia
New Books: 
Dragon Pearl by Yoon Ha Lee
Race to the Sun by Rebecca Roanhorse
The Walking Land by Callie Bates
The Bird King by G. Willow Wilson
The Bear and the Nightingale by Katherine Arden
Novice Dragoneer by E.E. Knight
The Ruin of Kings by Jenn Lyons
The Unwilling by Kelly Braffet
Anne of Avonlea by L. M. Montgomery
Wicked by Gregory Maguire
Catherine the Great by Robert Massie
Fearless Girls, Wise Women and Belove Sisters by Kathleen Ragan
Tales of India 
African Myths, Native American Myths, Aztec Myths
Games:
Chess.com
Zen Koi
‎Monument Valley
Art:
Museum virtual tours
Live stream concerts
Theatre Recordings:
Into the Woods
She Loves Me
Rent 
Company
Newsies
Falsettos
Legally Blonde
South Pacific
A Chorus Line
Beauty and the Beast 
Whatever I can find on youtube
8 notes · View notes
papermoonloveslucy · 6 years
Text
LUCY & HENRY FONDA ~ Part One
1935-1968 
Tumblr media
Lucille Ball and Henry Fonda were more than just co-workers. When Lucy first got to Hollywood, the two actually briefly dated. Lucy remembers,
Tumblr media
"We worked long and hard, Ginger [Rogers] and I, in front of our mirrors. We used eye shadow, plenty of mascara, pancake [make-up], deep red lipstick, rouge, everything we'd been taught in the studio cosmetic department. Then we went out to Brentwood, that's where the boys lived. My date was Fonda. Ginger's date was [Jimmy] Stewart. Henry cooked the dinner, and after we ate, Ginger and the boys turned on the radio in the living room and Ginger tried to teach them ‘The Carioca.’ I was left doing the dishes. When I finished, we went out dancing at the Coconut Grove. Freddie Martin's orchestra. There we were, Ginger and I in our long organdy dresses, looking just as summery and smooth as we could. The date stretched into daybreak. We'd had a hilarious, wonderful evening that came to an end at Barney's Beanery. Well, it was dark and we went in and light when we came out. Hank and Jim took one look at us and said, 'What happened?' We said, 'What do you mean what happened?' And Jimmy Stewart said, 'Well, your nighttime makeup is on awful heavy for this time of the morning.' And Henry Fonda said, 'Yuk!'"
In 1975 Fonda told this story at “The Dean Martin Celebrity Roast” for Lucille Ball. Ginger Rogers was also in attendance. He added that "If I hadn't said, 'Yuk!', if I'd behaved myself, they might have named that studio Henrylu, not Desilu."
Tumblr media
Perhaps it is a good thing that Fonda and Ball never married as genealogists point out that they are related - 8th cousins. The pair acted in three feature films together and made numerous television appearances opposite one-another. Curiously, although he was sometimes mentioned, Fonda never guest-starred on a “Lucy” sitcom.  
Tumblr media
I Dream Too Much (1935)
Producer: Pandro S. Berman Director: John Cromwell Choreographer: Hermes Pan Screenplay: Elsie Finn (story), David G. Wittels (story), Edmund North Songs: Jerome Kern and Dorothy Fields 
Cast: Lily Pons (Annette Monard Street), Henry Fonda (Jonathan Street), Eric Blore (Roger Briggs), Osgood Perkins (Paul Darcy), Lucien Littlefield (Hubert Dilley), Lucille Ball (Gwendolyn Dilley)
Synopsis: Annette Monard Street (Lily Pons) is an aspiring singer, who falls in love with and marries Jonathan Street (Henry Fonda), a struggling young composer. Jonathan pushes her into a singing career, and she soon becomes a star. Meanwhile, Jonathan is unable to sell his music, and he finds himself jealous of his wife's success. Concerned about their relationship, Annette uses her influence to get Jonathan's work turned into a musical comedy. Once she achieves this, she then retires from public life in order to raise a family.
"Lucille replaced Betty Grable, an eighteen-year-old stock player... in the minor role of Gwendolyn Dilley, a bleached-blonde gum-chewer visiting Paris with her parents and little brother.” ~ Kathleen Brady, Lucille
Tumblr media
Gwendolyn Dilley (Lucille Ball): "Culture is making my feet hurt."
TRIVIA
At this point in her career, Lucy was a platinum blonde. She had dyed it from her natural mousy brown to get more attention from casting agents and producers. She did not begin coloring her hair its trademark red until the technicolor film Du Barry Was A Lady in 1943.
A brief clip of Lucy in the film is included in “Hollywood the Golden Years: The RKO Story: A Woman's Lot” (1987).  
Tumblr media
The Big Street (1942)
Producer: Damon Runyon Director: Irving Reis Screenplay: Leonard Spigelgass, based on the short story “Little Pinks” by Damon Runyon
Cast: Henry Fonda (Little Pinks), Lucille Ball (Gloria Lyons), Barton MacLane (Case Ables), Eugene Pallette (Nicely Nicely Johnson), Agnes Moorehead (Violette Shumberg), Sam Levene (Horsethief), Ray Collins (Professor B)
Uncredited actor Hans Conried played a waiter. On “I Love Lucy” he played Harry Martin in “Redecorating” (S2;E8) and Percy Livermore in “Lucy Hires an English Tutor” (S2;E13), both in 1952. He also did two episodes of “The Lucy Show,” both as her music tutor Dr. Gitterman in 1963.  
'Queen of the Extras' Bess Flowers made numerous uncredited background appearances on both “I Love Lucy” and “The Lucy Show.”  
Uncredited actor Gil Perkins (Mug) later turned up on a 1970 episode of “Here's Lucy” (S2;E21).  
TRIVIA
During filming, Lucy's new husband Desi Arnaz felt so insecure about leaving Lucy and Fonda alone together that he’d often pop by the set to keep an eye on them. His paranoia so exasperated director Irving Reis that he finally banned him from the set.
This was Lucille Ball's favorite of her nearly 80 films. She felt her performance was unjustly ignored by the Academy.
Tumblr media
Damon Runyon also created the source material for the hit Broadway musical Guys and Dolls (1950), which starred Robert Alda, who went on to make several appearances on “The Lucy Show.” The two stories share the character of Nicely Nicely Johnson. When the film version was made by MGM in 1955, Lucy and Desi were also under contract to the studio. A brief clip of the film was inserted into the middle of an episode of “I Love Lucy” called “Lucy and the Dummy” (S5;E3), although the clip was removed after its initial airing. Further, when Lucille Ball first came to Hollywood, before becoming a contract player at RKO, she worked for Sam Goldwyn as one of the Goldwyn Girls. In Guys and Dolls, the Hot Box Girls are played by the Goldwyn Girls.
Tumblr media
Gloria Lyons (Lucille Ball): “Love is something that gets you one room, two chins, and three kids.”
A brief clip from the film is seen in “Lucy and Desi: A Home Movie.”
Tumblr media
“The Good Years” (January 12, 1962)
Produced by: Leland Heyward Directed by: Franklin L. Schaffner
Cast: Lucille Ball, Henry Fonda, Mort Sahl, Margaret Hamilton (Narrator)
Characters included Teddy Roosevelt, Sandow the Bodybuilder, the Wright Brothers, J.P. Morgan, Lizzy Borden   
Tumblr media
TRIVIA
This CBS special was billed as 'Lucille Ball's return to television' after leaving Lucy Ricardo behind in April 1960. It would be several more months before the debut of “The Lucy Show” in Fall 1962.   
Tumblr media
Based on a best-selling book by Walter Lord first published in 1960 about the years leading up to World War One, the special was a hodge-podge of sketches and musical numbers about the time period 1900 through 1920.
Tumblr media
Mort Sahl: “Lucille Ball came into rehearsal. She had a later call and a lot of doubts about the script.”
Tumblr media
The 90-minute special was a critical failure and has largely been forgotten. There are few photographs and video copies are held at the Museum of Broadcasting. 
“All About People” (1967)
Director: Saul Rubin
Narrators: Lucille Ball, Henry Fonda, Jack Benny, George Burns, Carol Channing, Eydie Gorme, Charleton Heston, Eartha Kitt, Burt Lancaster, Edward G. Robinson
TRIVIA
This was a 30-minute black and white documentary made by the United Jewish Welfare Fund about its history. 
After marrying Gary Morton (nee Morton Goldapper), Lucille Ball was active in Jewish charities. On December 9, 1961, Lucy had appeared on the “Twelve Star Salute to the Federation of Jewish Philanthropies.” 
Burns, Benny, and Gorme, all later made appearances on “Here's Lucy.” Edward G. Robinson did a cameo on “The Lucy Show.”  
Although Ball and Fonda are both involved in the project, they likely recorded their narration separately. 
Tumblr media
Yours, Mine and Ours  (1968)
Producer: Robert F. Blumofe Director: Melville Shavelson Screenplay: Melville Shavelson and Mort Lachman, with story by Bob Carroll Jr. and Madelyn Davis (Lucy’s TV writers), based on the book Who Gets The Drumsticks? by Helen Eileen Beardsley
Cast: Lucille Ball (Helen North Beardsley), Henry Fonda (Frank Beardsley), Van Johnson (Darrel Harrison)
Nancy Howard (Nancy Beardsley) made three appearances on “Here's Lucy.” Tim Matheson (Mike Beardsley) made an appearance on a 1972 “Here's Lucy” playing Kim Carter’s boyfriend. 
Uncredited extras Leon Alton, Paul Bradley, Charles Cirillo, George Boyce, Paul King, Joseph LaCava, and Leoda Richards all made numerous background appearances on “The Lucy Show” and “Here's Lucy.”
Synopsis: A widower with ten children falls for a widow with eight, and they must decide about forming a huge, unconventional family.
TRIVIA
Jane Fonda claimed that her father was deeply in love with Lucy and that the two were "very close" during the filming of Yours, Mine and Ours but that Lucy wasn't in love with him.
After purchasing the rights to the book the film was based on, Lucille Ball became very close to the real Beardsleys and even treated the whole family to a vacation at Disneyland. 
In 1959, Lucille Ball and Desi Arnaz, still affiliated with MGM, were going to star as Frank and Helen Beardsley but the studio had trouble with the casting until the late 1960s. In addition, their marriage was then on the rocks, a situation which would have made working together on the optimistic comedy somewhat problematic.
Lucy's old friend John Wayne was initially considered to play Frank Beardsley. The role was cast with Fred MacMurray, but he was replaced by Henry Fonda.  
Tumblr media
Frank (Henry Fonda): “I don't quite understand. Am I being stupid?” Helen (Lucille Ball): “No, you're being a man. Which is sometimes the same thing.”
Lucille Ball co-produced the film under her company, Desilu Productions. When the film became a surprise smash hit grossing over $17 million on a $2.5 million investment, she hadn't anticipated the film's huge box-office success and failed to provide a tax shelter for her personal profits, resulting in most of her earnings going toward taxes.
Tumblr media
The success of the film led to Lucy being considered to play Mrs. Brady in “The Brady Bunch,” a TV sitcom with a similar story of a blended family. Lucy decided to do her own sitcom, “Here's Lucy,” instead.
In 1968, Van Johnson guest starred on “Here's Lucy” as both himself and an impostor look-alike in “Guess Who Owes Lucy $23.50” (HL S1;E11). The dialogue contained references to Yours, Mine and Ours and their co-star Henry Fonda.
Tumblr media
Van Johnson Impostor: “I loved working with that kooky redhead.” Lucy Carter: “Personally, I thought she was much too young for Henry Fonda.”
Johnson was in the cast of Too Many Girls, the film which introduced Lucy to Desi in 1940. Johnson also guest-starred on “I Love Lucy” in “The Dancing Star” (S4;E27) in 1955.
Click Here for Part Two: 1975 to 1979!
14 notes · View notes
visplay · 2 years
Text
Tumblr media
Chris: I try to run unusual pick films for myself from HBO Max on weekend mornings when I work on projects, and I had never seen this one before, noteworthy due to the strange swearing of Carrie Snodgrass’ character; I would only recommend this crime action film to Charles Bronson fans as Watch: On Subscription Service.
0 notes
coolthooks · 3 years
Text
Blue Cross Blue Shield of Illinois : Demand that Blue Cross Blue Shield of Illinois approve life-saving transplant for Kathleen Lyons
Blue Cross Blue Shield of Illinois : Demand that Blue Cross Blue Shield of Illinois approve life-saving transplant for Kathleen Lyons
Hey, I just signed the petition “Blue Cross Blue Shield of Illinois : Demand that Blue Cross Blue Shield of Illinois approve life-saving transplant for Kathleen Lyons” and wanted to see if you could help by adding your name. Our goal is to reach 25,000 signatures and we need more support. You can read more and sign the petition here: http://chng.it/CkcvSGkm8k Thanks! Tanya
View On WordPress
0 notes