Friday, June 15, 2007

Personal Empowerment in the Age of Corporate Schmuckism....

... or Schmoeism. Or, just, kind of, bland adult-ishness.

Anyhow-- just a note to myself (hence the italicizing). . . this is something I want to explore, a little. Much to the dismay of my vestigial, internalized 19 year old self, I am not anti-business, or anti-corporation, but I do think that the corporation, or even the small business, as an entity, can be a simultaneously dangerous and enlightening place. . .kind of sublime, Shakespearian, double-edged, massive. . . in the development of an individual. Kind of a modern dragon, locus of many Herculean labours? Maybe? I'm still learning the ropes.

What I want to think about is a little bit trickier than left vs. right, old school vs. new, hippie vs. mainstreamer, zombie vs. self-actualized individual; in fact, its not anything vs. anything. . . Part of the trick is to straddle the whole numbing vs. ideology in the first place and. . .not just the whole "dualism is bad" routine. It is and it isn't . . .

I actually have some thoughts on this, but I'm too distracted by work to get to them now, so maybe on the weekend.

Soon, anyhow.

Saturday, May 12, 2007

The Abandoned Blog

Man, I am so lame. It's bad enough I would start a blog, but then... not even to post on it? I wonder how many abandoned blogs there are? People just running out of things to say. . . or being abducted. . . killed for knowing too much? How many of the universe's mysteries were just about to be revealed when. . . bam! some brilliant blogger was nabbed on his way home from the library. . . or the bank. . . or his job in a call centre. . . or a big box store. . .

It's sad really. I have no such excuse. My problem is that I don't want to ramble on about tripe, and much of what I think about in the run of a day is just that. Politics. My job. Groceries. Pillow top mattresses.

The blogosphere is flooded with political blogs. . .literary blogs. . . angst-ridden blogs. . . what can I make of mine?

I called it Signastra. . . why? Because I am very interested in symbolism-- signs & semiotics. I'm interested in how you can crack open just about any word, object or image and extract layer upon layer of complex meaning. Everything is contextual, nothing exists in isolation, everything relates. . .there are no one-offs. . . Do I really believe this? Yes, I guess I do. The platypus aside.

Signastra. . . there's resonance there. It sounds like Sinatra. . .whom I have always adored in a naive, let's not look at the reality kind of way. How's that for paradox?

What does Sinatra symbolize for me? Glamour, exapansiveness, power, flash. And quirkiness. Teleprompters. . . nostalgia. . .an era. Not an era I would have liked to inhabit, mind you, but an era nonetheless. Cool. Style. Savvy. Yes. And he was Sicilian-- my ancestors were also. So, in a sense, it's also a personal signature of mine.

Next item: it is phonetically similar to the word "sinister" as well. . . Sinister, associated with the left hand, the occult. I love the occult. I love the thought of uncovering what is hidden. . .

And "astra", the stars. . . From a Cheryl Savageau poem "All I know is we come from the stars. . ." Yes. Je suis fasciné par les étoiles.

All of these reasons. . . and now I am completely adrift. . .without a center of gravity. The name means something, but it hasn't grown into anything substantive just yet.

In time.

Monday, March 26, 2007

Dream Language

Okay. What strange and difficult to decipher dreams I have been having lately... Some have been so bizarre that I think they actually bear posting here. . .

Last night, I dreamt a childhood sweetheart of mine had turned into a butterfly/moth, and he lived in a little paper pouch/box that had cellular service. . . This did not strike me as odd in the least. However, for some reason, this pouch/box fell into my possession, and we went out canoeing in open water. We ran into rough weather-- the box fell open--and as I was struggling to attach our canoe to a small yacht nearby, who could pull us safely to shore-- the moth/butterfly fell into the ocean... and I was quite upset, feeling I had caused my old friend to drown. But adreamvoice kept telling me that it was okay, that the type of butterfly he had chosen to be was immortal, in a sense, and that even if the butterfly form was injured, hurt or killed, the real person was unaffected-- he'd materialize somewhere on land, and just turn himself back into a butterfly (or was it a moth? This was never clear) like magick. Like magick. And this appeased me somewhat. Once safely back on land, I tried to find him to make sure he was safe, but couldn't get through on the phone. I told myself I'd just have to wait and see if I ever heard tell of him again. I wasn't sure why I'd been in possession of the butterfly version of him, in any event. . . .

Overall, the tone of the dream was one of puzzlement. I knew it was very bizarre to have a person become a butterfly and live in a tiny pouch/box, but I also understood it was something some people chose to do . . .

Oddddddd..... I'll definitely have to give this one some thought. I do remember learning that when we dream of childhood friends or old flames, it is symbolic of our own relationship to our former self. . . oooooh.....

Lots to think about here!!!

Until next time!!

Sunday, March 25, 2007

Openings & Closures


I've been distracted from this blog by many things lately-- a surfeit of choices urging life changes, and requiring me to make decisions: something I've never been great at. Every option has it's pros and cons, its potential benefits and fallouts.

I'm in a pleroma, of sorts-- all is possibility, nothing is concrete. No path through the woods has appeared. . . everything is hint and glimmer. By all accounts, this should be a highly creative time-- a dreaming time, a time when the future is being seeded by the collusion of seemingly random events and impulses.

I keep sort of vaguely waiting for certainty to set in--to wake up knowing what I have to do. It hasn't happened yet. Instead, I vascillate widly. In the morning I am convinced of one thing, and by evening, another holds sway.

It is absolutely a blessing to have options, choices, potentials to play with--and I wouldn't trade it for anything. But being in an undecided state can be paralyzing as well. There are so many factors to assess I could sit still for years just picking through the nuance.

Ultimately, it's going to come down to me making a firm decision and just pursuing that course of action with everything that I've got. Focus. Times of focus can be extremely creative, too-- in fact, when we narrow our field of vision, we can actually accomplish great things, contrary to that common artistic banality: "stay open, stay open."

The whole "stay open" platitude is misleading. Sometimes you really do have to shut out the rest of the world and follow your own, unilateral vision, however selfish or foolhardy this may seem. The world will bait you into doing it's bidding, if you let it. It's important to check ourselves every so often to make sure that it really is our own values we are living by and working for, our own goals we are pursuing. To do otherwise is to remain on autopilot, asleep at the wheel.

There is a lot to be said for stability, but there is also much to be said for flux.

I'm trying to keep my eyes open in the midst of flux, watching for an exit--a clearly marked road leading to the cottage hidden deep in the forest, so that I may hunker down into a little stability. For a spell. And then the flux will come on again. . . it always does. Inevitable as spring.

However, in the process of all this decision making and priority setting, it has begun to come clear to me that some things really are deeply innate with us, and others are situational.

One of the decisions I have to make involves whether or not to continue pursuing my career as a writer. . . .ooooof. Part of me always hinges on giving it up, which some would claim is a sure sign that I am not a real writer, after all. No real writer would ever debate giving it up.

Well, ha. Ha, ha, ha. If I weren't one, there would be nothing to give up in the first place, would there? So you see.....

But "things" keep happening that remind me of who I am, and of the path I should take if I am to remain true to myself.

For instance, one of the big issues on my artistic journey has been "poetry vs. practicality", or "creative thought vs. business".

Last night, at a gala I was attending, a very respected politician began quoting poetry as something that had guided him through the darkest hours of his life. I was there because of a work obligation (and work has been consuming so much of my time lately, poetry has fallen by the wayside...) It certainly jogged my thinking.... so much to consider....

I am going to do my level best to keep this blog interesting and up to date, but for now, I am totally in the pleroma. . . .sifting through the darkness, trying to imagine the appearance of a road. In the end, I will have to clear a path myself. But first I have to imagine a destination.

And then focus.

I promise that as that happens, this blog will get more intense, and more interesting as well.

Wednesday, March 14, 2007

Portrait


My friend Elizabeth completed this portrait of me at Christmas... it's so good I just had to post it. She's amazing. This is taken from a picture of me shot on New Year's Eve 1999-- what an image to memorialize like this!!! I'm so lucky to have this..... now I'm getting her to do one of my sister....

She's in the midst launching her own jewellery business now- I will definitely be linking much more to her work and to her site as things progress.... Thank you again, Liz!! Brava... you made me look awesome!!1

Monday, March 5, 2007

Darwin's God

I must be thinking in synch with the New York Times-- happily, this reporter synthesizes and researches much, much better than I do. . . .


By ROBIN MARANTZ HENIG
New York Times
Published: March 4, 2007

God has always been a puzzle for Scott Atran. When he was 10 years old, he scrawled a plaintive message on the wall of his bedroom in Baltimore. “God exists,” he wrote in black and orange paint, “or if he doesn’t, we’re in trouble.” Atran has been struggling with questions about religion ever since — why he himself no longer believes in God and why so many other people, everywhere in the world, apparently do.

Call it God; call it superstition; call it, as Atran does, “belief in hope beyond reason” — whatever you call it, there seems an inherent human drive to believe in something transcendent, unfathomable and otherworldly, something beyond the reach or understanding of science. “Why do we cross our fingers during turbulence, even the most atheistic among us?” asked Atran when we spoke at his Upper West Side pied-à-terre in January. Atran, who is 55, is an anthropologist at the National Center for Scientific Research in Paris, with joint appointments at the University of Michigan and the John Jay College of Criminal Justice in New York. His research interests include cognitive science and evolutionary biology, and sometimes he presents students with a wooden box that he pretends is an African relic. “If you have negative sentiments toward religion,” he tells them, “the box will destroy whatever you put inside it.” Many of his students say they doubt the existence of God, but in this demonstration they act as if they believe in something. Put your pencil into the magic box, he tells them, and the nonbelievers do so blithely. Put in your driver’s license, he says, and most do, but only after significant hesitation. And when he tells them to put in their hands, few will.

If they don’t believe in God, what exactly are they afraid of?
Atran first conducted the magic-box demonstration in the 1980s, when he was at Cambridge University studying the nature of religious belief. He had received a doctorate in anthropology from Columbia University and, in the course of his fieldwork, saw evidence of religion everywhere he looked — at archaeological digs in Israel, among the Mayans in Guatemala, in artifact drawers at the American Museum of Natural History in New York. Atran is Darwinian in his approach, which means he tries to explain behavior by how it might once have solved problems of survival and reproduction for our early ancestors. But it was not clear to him what evolutionary problems might have been solved by religious belief. Religion seemed to use up physical and mental resources without an obvious benefit for survival. Why, he wondered, was religion so pervasive, when it was something that seemed so costly from an evolutionary point of view?

The magic-box demonstration helped set Atran on a career studying why humans might have evolved to be religious, something few people were doing back in the ’80s. Today, the effort has gained momentum, as scientists search for an evolutionary explanation for why belief in God exists — not whether God exists, which is a matter for philosophers and theologians, but why the belief does.

This is different from the scientific assault on religion that has been garnering attention recently, in the form of best-selling books from scientific atheists who see religion as a scourge. In “The God Delusion,” published last year and still on best-seller lists, the Oxford evolutionary biologist Richard Dawkins concludes that religion is nothing more than a useless, and sometimes dangerous, evolutionary accident. “Religious behavior may be a misfiring, an unfortunate byproduct of an underlying psychological propensity which in other circumstances is, or once was, useful,” Dawkins wrote. He is joined by two other best-selling authors — Sam Harris, who wrote “The End of Faith,” and Daniel Dennett, a philosopher at Tufts University who wrote “Breaking the Spell.” The three men differ in their personal styles and whether they are engaged in a battle against religiosity, but their names are often mentioned together. They have been portrayed as an unholy trinity of neo-atheists, promoting their secular world view with a fervor that seems almost evangelical.

Lost in the hullabaloo over the neo-atheists is a quieter and potentially more illuminating debate. It is taking place not between science and religion but within science itself, specifically among the scientists studying the evolution of religion. These scholars tend to agree on one point: that religious belief is an outgrowth of brain architecture that evolved during early human history. What they disagree about is why a tendency to believe evolved, whether it was because belief itself was adaptive or because it was just an evolutionary byproduct, a mere consequence of some other adaptation in the evolution of the human brain.

Which is the better biological explanation for a belief in God — evolutionary adaptation or neurological accident? Is there something about the cognitive functioning of humans that makes us receptive to belief in a supernatural deity? And if scientists are able to explain God, what then? Is explaining religion the same thing as explaining it away? Are the nonbelievers right, and is religion at its core an empty undertaking, a misdirection, a vestigial artifact of a primitive mind? Or are the believers right, and does the fact that we have the mental capacities for discerning God suggest that it was God who put them there?
In short, are we hard-wired to believe in God? And if we are, how and why did that happen?

“All of our raptures and our drynesses, our longings and pantings, our questions and beliefs . . . are equally organically founded,” William James wrote in “The Varieties of Religious Experience.” James, who taught philosophy and experimental psychology at Harvard for more than 30 years, based his book on a 1901 lecture series in which he took some early tentative steps at breaching the science-religion divide.
In the century that followed, a polite convention generally separated science and religion, at least in much of the Western world. Science, as the old trope had it, was assigned the territory that describes how the heavens go; religion, how to go to heaven.

Anthropologists like Atran and psychologists as far back as James had been looking at the roots of religion, but the mutual hands-off policy really began to shift in the 1990s. Religion made incursions into the traditional domain of science with attempts to bring intelligent design into the biology classroom and to choke off human embryonic stem-cell research on religious grounds. Scientists responded with counterincursions. Experts from the hard sciences, like evolutionary biology and cognitive neuroscience, joined anthropologists and psychologists in the study of religion, making God an object of scientific inquiry.

The debate over why belief evolved is between byproduct theorists and adaptationists. You might think that the byproduct theorists would tend to be nonbelievers, looking for a way to explain religion as a fluke, while the adaptationists would be more likely to be believers who can intuit the emotional, spiritual and community advantages that accompany faith. Or you might think they would all be atheists, because what believer would want to subject his own devotion to rationalism’s cold, hard scrutiny? But a scientist’s personal religious view does not always predict which side he will take. And this is just one sign of how complex and surprising this debate has become.

Angels, demons, spirits, wizards, gods and witches have peppered folk religions since mankind first started telling stories. Charles Darwin noted this in “The Descent of Man.” “A belief in all-pervading spiritual agencies,” he wrote, “seems to be universal.” According to anthropologists, religions that share certain supernatural features — belief in a noncorporeal God or gods, belief in the afterlife, belief in the ability of prayer or ritual to change the course of human events — are found in virtually every culture on earth.

This is certainly true in the United States. About 6 in 10 Americans, according to a 2005 Harris Poll, believe in the devil and hell, and about 7 in 10 believe in angels, heaven and the existence of miracles and of life after death. A 2006 survey at Baylor University found that 92 percent of respondents believe in a personal God — that is, a God with a distinct set of character traits ranging from “distant” to “benevolent.”

When a trait is universal, evolutionary biologists look for a genetic explanation and wonder how that gene or genes might enhance survival or reproductive success. In many ways, it’s an exercise in post-hoc hypothesizing: what would have been the advantage, when the human species first evolved, for an individual who happened to have a mutation that led to, say, a smaller jaw, a bigger forehead, a better thumb? How about certain behavioral traits, like a tendency for risk-taking or for kindness?

Atran saw such questions as a puzzle when applied to religion. So many aspects of religious belief involve misattribution and misunderstanding of the real world. Wouldn’t this be a liability in the survival-of-the-fittest competition? To Atran, religious belief requires taking “what is materially false to be true” and “what is materially true to be false.” One example of this is the belief that even after someone dies and the body demonstrably disintegrates, that person will still exist, will still be able to laugh and cry, to feel pain and joy. This confusion “does not appear to be a reasonable evolutionary strategy,” Atran wrote in “In Gods We Trust: The Evolutionary Landscape of Religion” in 2002. “Imagine another animal that took injury for health or big for small or fast for slow or dead for alive. It’s unlikely that such a species could survive.” He began to look for a sideways explanation: if religious belief was not adaptive, perhaps it was associated with something else that was.

Atran intended to study mathematics when he entered Columbia as a precocious 17-year-old. But he was distracted by the radical politics of the late ’60s. One day in his freshman year, he found himself at an antiwar rally listening to Margaret Mead, then perhaps the most famous anthropologist in America. Atran, dressed in a flamboyant Uncle Sam suit, stood up and called her a sellout for saying the protesters should be writing to their congressmen instead of staging demonstrations. “Young man,” the unflappable Mead said, “why don’t you come see me in my office?”

Atran, equally unflappable, did go to see her — and ended up working for Mead, spending much of his time exploring the cabinets of curiosities in her tower office at the American Museum of Natural History. Soon he switched his major to anthropology.
Many of the museum specimens were religious, Atran says. So were the artifacts he dug up on archaeological excursions in Israel in the early ’70s. Wherever he turned, he encountered the passion of religious belief. Why, he wondered, did people work so hard against their preference for logical explanations to maintain two views of the world, the real and the unreal, the intuitive and the counterintuitive?

Maybe cognitive effort was precisely the point. Maybe it took less mental work than Atran realized to hold belief in God in one’s mind. Maybe, in fact, belief was the default position for the human mind, something that took no cognitive effort at all.
While still an undergraduate, Atran decided to explore these questions by organizing a conference on universal aspects of culture and inviting all his intellectual heroes: the linguist Noam Chomsky, the psychologist Jean Piaget, the anthropologists Claude Levi-Strauss and Gregory Bateson (who was also Margaret Mead’s ex-husband), the Nobel Prize-winning biologists Jacques Monod and Francois Jacob. It was 1974, and the only site he could find for the conference was at a location just outside Paris. Atran was a scraggly 22-year-old with a guitar who had learned his French from comic books. To his astonishment, everyone he invited agreed to come.

Atran is a sociable man with sharp hazel eyes, who sparks provocative conversations the way other men pick bar fights. As he traveled in the ’70s and ’80s, he accumulated friends who were thinking about the issues he was: how culture is transmitted among human groups and what evolutionary function it might serve. “I started looking at history, and I wondered why no society ever survived more than three generations without a religious foundation as its raison d’être,” he says. Soon he turned to an emerging subset of evolutionary theory — the evolution of human cognition.

Some cognitive scientists think of brain functioning in terms of modules, a series of interconnected machines, each one responsible for a particular mental trick. They do not tend to talk about a God module per se; they usually consider belief in God a consequence of other mental modules.

Religion, in this view, is “a family of cognitive phenomena that involves the extraordinary use of everyday cognitive processes,” Atran wrote in “In Gods We Trust.” “Religions do not exist apart from the individual minds that constitute them and the environments that constrain them, any more than biological species and varieties exist independently of the individual organisms that compose them and the environments that conform them.”
At around the time “In Gods We Trust” appeared five years ago, a handful of other scientists — Pascal Boyer, now at Washington University; Justin Barrett, now at Oxford; Paul Bloom at Yale — were addressing these same questions. In synchrony they were moving toward the byproduct theory.

Darwinians who study physical evolution distinguish between traits that are themselves adaptive, like having blood cells that can transport oxygen, and traits that are byproducts of adaptations, like the redness of blood. There is no survival advantage to blood’s being red instead of turquoise; it is just a byproduct of the trait that is adaptive, having blood that contains hemoglobin.

Something similar explains aspects of brain evolution, too, say the byproduct theorists. Which brings us to the idea of the spandrel.
Stephen Jay Gould, the famed evolutionary biologist at Harvard who died in 2002, and his colleague Richard Lewontin proposed “spandrel” to describe a trait that has no adaptive value of its own. They borrowed the term from architecture, where it originally referred to the V-shaped structure formed between two rounded arches. The structure is not there for any purpose; it is there because that is what happens when arches align.
In architecture, a spandrel can be neutral or it can be made functional. Building a staircase, for instance, creates a space underneath that is innocuous, just a blank sort of triangle. But if you put a closet there, the under-stairs space takes on a function, unrelated to the staircase’s but useful nonetheless. Either way, functional or nonfunctional, the space under the stairs is a spandrel, an unintended byproduct.
“Natural selection made the human brain big,” Gould wrote, “but most of our mental properties and potentials may be spandrels — that is, nonadaptive side consequences of building a device with such structural complexity.”

The possibility that God could be a spandrel offered Atran a new way of understanding the evolution of religion. But a spandrel of what, exactly?
Hardships of early human life favored the evolution of certain cognitive tools, among them the ability to infer the presence of organisms that might do harm, to come up with causal narratives for natural events and to recognize that other people have minds of their own with their own beliefs, desires and intentions. Psychologists call these tools, respectively, agent detection, causal reasoning and theory of mind.

Agent detection evolved because assuming the presence of an agent — which is jargon for any creature with volitional, independent behavior — is more adaptive than assuming its absence. If you are a caveman on the savannah, you are better off presuming that the motion you detect out of the corner of your eye is an agent and something to run from, even if you are wrong. If it turns out to have been just the rustling of leaves, you are still alive; if what you took to be leaves rustling was really a hyena about to pounce, you are dead.

A classic experiment from the 1940s by the psychologists Fritz Heider and Marianne Simmel suggested that imputing agency is so automatic that people may do it even for geometric shapes. For the experiment, subjects watched a film of triangles and circles moving around. When asked what they had been watching, the subjects used words like “chase” and “capture.” They did not just see the random movement of shapes on a screen; they saw pursuit, planning, escape.
So if there is motion just out of our line of sight, we presume it is caused by an agent, an animal or person with the ability to move independently. This usually operates in one direction only; lots of people mistake a rock for a bear, but almost no one mistakes a bear for a rock.
What does this mean for belief in the supernatural? It means our brains are primed for it, ready to presume the presence of agents even when such presence confounds logic. “The most central concepts in religions are related to agents,” Justin Barrett, a psychologist, wrote in his 2004 summary of the byproduct theory, “Why Would Anyone Believe in God?” Religious agents are often supernatural, he wrote, “people with superpowers, statues that can answer requests or disembodied minds that can act on us and the world.”

A second mental module that primes us for religion is causal reasoning. The human brain has evolved the capacity to impose a narrative, complete with chronology and cause-and-effect logic, on whatever it encounters, no matter how apparently random. “We automatically, and often unconsciously, look for an explanation of why things happen to us,” Barrett wrote, “and ‘stuff just happens’ is no explanation. Gods, by virtue of their strange physical properties and their mysterious superpowers, make fine candidates for causes of many of these unusual events.” The ancient Greeks believed thunder was the sound of Zeus’s thunderbolt. Similarly, a contemporary woman whose cancer treatment works despite 10-to-1 odds might look for a story to explain her survival. It fits better with her causal-reasoning tool for her recovery to be a miracle, or a reward for prayer, than for it to be just a lucky roll of the dice.
A third cognitive trick is a kind of social intuition known as theory of mind. It’s an odd phrase for something so automatic, since the word “theory” suggests formality and self-consciousness. Other terms have been used for the same concept, like intentional stance and social cognition. One good alternative is the term Atran uses: folkpsychology.

Folkpsychology, as Atran and his colleagues see it, is essential to getting along in the contemporary world, just as it has been since prehistoric times. It allows us to anticipate the actions of others and to lead others to believe what we want them to believe; it is at the heart of everything from marriage to office politics to poker. People without this trait, like those with severe autism, are impaired, unable to imagine themselves in other people’s heads.
The process begins with positing the existence of minds, our own and others’, that we cannot see or feel. This leaves us open, almost instinctively, to belief in the separation of the body (the visible) and the mind (the invisible). If you can posit minds in other people that you cannot verify empirically, suggests Paul Bloom, a psychologist and the author of “Descartes’ Baby,” published in 2004, it is a short step to positing minds that do not have to be anchored to a body. And from there, he said, it is another short step to positing an immaterial soul and a transcendent God.

The traditional psychological view has been that until about age 4, children think that minds are permeable and that everyone knows whatever the child himself knows. To a young child, everyone is infallible. All other people, especially Mother and Father, are thought to have the same sort of insight as an all-knowing God.

But at a certain point in development, this changes. (Some new research suggests this might occur as early as 15 months.) The “false-belief test” is a classic experiment that highlights the boundary. Children watch a puppet show with a simple plot: John comes onstage holding a marble, puts it in Box A and walks off. Mary comes onstage, opens Box A, takes out the marble, puts it in Box B and walks off. John comes back onstage. The children are asked, Where will John look for the marble?

Very young children, or autistic children of any age, say John will look in Box B, since they know that’s where the marble is. But older children give a more sophisticated answer. They know that John never saw Mary move the marble and that as far as he is concerned it is still where he put it, in Box A. Older children have developed a theory of mind; they understand that other people sometimes have false beliefs. Even though they know that the marble is in Box B, they respond that John will look for it in Box A.

The adaptive advantage of folkpsychology is obvious. According to Atran, our ancestors needed it to survive their harsh environment, since folkpsychology allowed them to “rapidly and economically” distinguish good guys from bad guys. But how did folkpsychology — an understanding of ordinary people’s ordinary minds — allow for a belief in supernatural, omniscient minds? And if the byproduct theorists are right and these beliefs were of little use in finding food or leaving more offspring, why did they persist?

Atran ascribes the persistence to evolutionary misdirection, which, he says, happens all the time: “Evolution always produces something that works for what it works for, and then there’s no control for however else it’s used.” On a sunny weekday morning, over breakfast at a French cafe on upper Broadway, he tried to think of an analogy and grinned when he came up with an old standby: women’s breasts. Because they are associated with female hormones, he explained, full breasts indicate a woman is fertile, and the evolution of the male brain’s preference for them was a clever mating strategy. But breasts are now used for purposes unrelated to reproduction, to sell anything from deodorant to beer. “A Martian anthropologist might look at this and say, ‘Oh, yes, so these breasts must have somehow evolved to sell hygienic stuff or food to human beings,’ ” Atran said. But the Martian would, of course, be wrong. Equally wrong would be to make the same mistake about religion, thinking it must have evolved to make people behave a certain way or feel a certain allegiance.

That is what most fascinated Atran. “Why is God in there?” he wondered.
The idea of an infallible God is comfortable and familiar, something children readily accept. You can see this in the experiment Justin Barrett conducted recently — a version of the traditional false-belief test but with a religious twist. Barrett showed young children a box with a picture of crackers on the outside. What do you think is inside this box? he asked, and the children said, “Crackers.” Next he opened it and showed them that the box was filled with rocks. Then he asked two follow-up questions: What would your mother say is inside this box? And what would God say?

As earlier theory-of-mind experiments already showed, 3- and 4-year-olds tended to think Mother was infallible, and since the children knew the right answer, they assumed she would know it, too. They usually responded that Mother would say the box contained rocks. But 5- and 6-year-olds had learned that Mother, like any other person, could hold a false belief in her mind, and they tended to respond that she would be fooled by the packaging and would say, “Crackers.”

And what would God say? No matter what their age, the children, who were all Protestants, told Barrett that God would answer, “Rocks.” This was true even for the older children, who, as Barrett understood it, had developed folkpsychology and had used it when predicting a wrong response for Mother. They had learned that, in certain situations, people could be fooled — but they had also learned that there is no fooling God.

The bottom line, according to byproduct theorists, is that children are born with a tendency to believe in omniscience, invisible minds, immaterial souls — and then they grow up in cultures that fill their minds, hard-wired for belief, with specifics. It is a little like language acquisition, Paul Bloom says, with the essential difference that language is a biological adaptation and religion, in his view, is not. We are born with an innate facility for language but the specific language we learn depends on the environment in which we are raised. In much the same way, he says, we are born with an innate tendency for belief, but the specifics of what we grow up believing — whether there is one God or many, whether the soul goes to heaven or occupies another animal after death — are culturally shaped.

Whatever the specifics, certain beliefs can be found in all religions. Those that prevail, according to the byproduct theorists, are those that fit most comfortably with our mental architecture. Psychologists have shown, for instance, that people attend to, and remember, things that are unfamiliar and strange, but not so strange as to be impossible to assimilate. Ideas about God or other supernatural agents tend to fit these criteria. They are what Pascal Boyer, an anthropologist and psychologist, called “minimally counterintuitive”: weird enough to get your attention and lodge in your memory but not so weird that you reject them altogether. A tree that talks is minimally counterintuitive, and you might believe it as a supernatural agent. A tree that talks and flies and time-travels is maximally counterintuitive, and you are more likely to reject it.

Atran, along with Ara Norenzayan of the University of British Columbia, studied the idea of minimally counterintuitive agents earlier this decade. They presented college students with lists of fantastical creatures and asked them to choose the ones that seemed most “religious.” The convincingly religious agents, the students said, were not the most outlandish — not the turtle that chatters and climbs or the squealing, flowering marble — but those that were just outlandish enough: giggling seaweed, a sobbing oak, a talking horse. Giggling seaweed meets the requirement of being minimally counterintuitive, Atran wrote. So does a God who has a human personality except that he knows everything or a God who has a mind but has no body.

It is not enough for an agent to be minimally counterintuitive for it to earn a spot in people’s belief systems. An emotional component is often needed, too, if belief is to take hold. “If your emotions are involved, then that’s the time when you’re most likely to believe whatever the religion tells you to believe,” Atran says. Religions stir up emotions through their rituals — swaying, singing, bowing in unison during group prayer, sometimes working people up to a state of physical arousal that can border on frenzy. And religions gain strength during the natural heightening of emotions that occurs in times of personal crisis, when the faithful often turn to shamans or priests. The most intense personal crisis, for which religion can offer powerfully comforting answers, is when someone comes face to face with mortality.

In John Updike’s celebrated early short story “Pigeon Feathers,” 14-year-old David spends a lot of time thinking about death. He suspects that adults are lying when they say his spirit will live on after he dies. He keeps catching them in inconsistencies when he asks where exactly his soul will spend eternity. “Don’t you see,” he cries to his mother, “if when we die there’s nothing, all your sun and fields and what not are all, ah, horror? It’s just an ocean of horror.”
The story ends with David’s tiny revelation and his boundless relief. The boy gets a gun for his 15th birthday, which he uses to shoot down some pigeons that have been nesting in his grandmother’s barn. Before he buries them, he studies the dead birds’ feathers. He is amazed by their swirls of color, “designs executed, it seemed, in a controlled rapture.” And suddenly the fears that have plagued him are lifted, and with a “slipping sensation along his nerves that seemed to give the air hands, he was robed in this certainty: that the God who had lavished such craft upon these worthless birds would not destroy His whole Creation by refusing to let David live forever.”

Fear of death is an undercurrent of belief. The spirits of dead ancestors, ghosts, immortal deities, heaven and hell, the everlasting soul: the notion of spiritual existence after death is at the heart of almost every religion. According to some adaptationists, this is part of religion’s role, to help humans deal with the grim certainty of death. Believing in God and the afterlife, they say, is how we make sense of the brevity of our time on earth, how we give meaning to this brutish and short existence. Religion can offer solace to the bereaved and comfort to the frightened.

But the spandrelists counter that saying these beliefs are consolation does not mean they offered an adaptive advantage to our ancestors. “The human mind does not produce adequate comforting delusions against all situations of stress or fear,” wrote Pascal Boyer, a leading byproduct theorist, in “Religion Explained,” which came out a year before Atran’s book. “Indeed, any organism that was prone to such delusions would not survive long.”
Whether or not it is adaptive, belief in the afterlife gains power in two ways: from the intensity with which people wish it to be true and from the confirmation it seems to get from the real world. This brings us back to folkpsychology. We try to make sense of other people partly by imagining what it is like to be them, an adaptive trait that allowed our ancestors to outwit potential enemies. But when we think about being dead, we run into a cognitive wall. How can we possibly think about not thinking? “Try to fill your consciousness with the representation of no-consciousness, and you will see the impossibility of it,” the Spanish philosopher Miguel de

Unamuno wrote in “Tragic Sense of Life.” “The effort to comprehend it causes the most tormenting dizziness. We cannot conceive of ourselves as not existing.”
Much easier, then, to imagine that the thinking somehow continues. This is what young children seem to do, as a study at the Florida Atlantic University demonstrated a few years ago. Jesse Bering and David Bjorklund, the psychologists who conducted the study, used finger puppets to act out the story of a mouse, hungry and lost, who is spotted by an alligator. “Well, it looks like Brown Mouse got eaten by Mr. Alligator,” the narrator says at the end. “Brown Mouse is not alive anymore.”

Afterward, Bering and Bjorklund asked their subjects, ages 4 to 12, what it meant for Brown Mouse to be “not alive anymore.” Is he still hungry? Is he still sleepy? Does he still want to go home? Most said the mouse no longer needed to eat or drink. But a large proportion, especially the younger ones, said that he still had thoughts, still loved his mother and still liked cheese. The children understood what it meant for the mouse’s body to cease to function, but many believed that something about the mouse was still alive.

“Our psychological architecture makes us think in particular ways,” says Bering, now at Queens University in Belfast, Northern Ireland. “In this study, it seems, the reason afterlife beliefs are so prevalent is that underlying them is our inability to simulate our nonexistence.”
It might be just as impossible to simulate the nonexistence of loved ones. A large part of any relationship takes place in our minds, Bering said, so it’s natural for it to continue much as before after the other person’s death. It is easy to forget that your sister is dead when you reach for the phone to call her, since your relationship was based so much on memory and imagined conversations even when she was alive. In addition, our agent-detection device sometimes confirms the sensation that the dead are still with us. The wind brushes our cheek, a spectral shape somehow looks familiar and our agent detection goes into overdrive. Dreams, too, have a way of confirming belief in the afterlife, with dead relatives appearing in dreams as if from beyond the grave, seeming very much alive.

Belief is our fallback position, according to Bering; it is our reflexive style of thought. “We have a basic psychological capacity that allows anyone to reason about unexpected natural events, to see deeper meaning where there is none,” he says. “It’s natural; it’s how our minds work.”
Intriguing as the spandrel logic might be, there is another way to think about the evolution of religion: that religion evolved because it offered survival advantages to our distant ancestors. This is where the action is in the science of God debate, with a coterie of adaptationists arguing on behalf of the primary benefits, in terms of survival advantages, of religious belief.

The trick in thinking about adaptation is that even if a trait offers no survival advantage today, it might have had one long ago. This is how Darwinians explain how certain physical characteristics persist even if they do not currently seem adaptive — by asking whether they might have helped our distant ancestors form social groups, feed themselves, find suitable mates or keep from getting killed. A facility for storing calories as fat, for instance, which is a detriment in today’s food-rich society, probably helped our ancestors survive cyclical famines.
So trying to explain the adaptiveness of religion means looking for how it might have helped early humans survive and reproduce. As some adaptationists see it, this could have worked on two levels, individual and group. Religion made people feel better, less tormented by thoughts about death, more focused on the future, more willing to take care of themselves. As William James put it, religion filled people with “a new zest which adds itself like a gift to life . . . an assurance of safety and a temper of peace and, in relation to others, a preponderance of loving affections.”

Such sentiments, some adaptationists say, made the faithful better at finding and storing food, for instance, and helped them attract better mates because of their reputations for morality, obedience and sober living. The advantage might have worked at the group level too, with religious groups outlasting others because they were more cohesive, more likely to contain individuals willing to make sacrifices for the group and more adept at sharing resources and preparing for warfare.

One of the most vocal adaptationists is David Sloan Wilson, an occasional thorn in the side of both Scott Atran and Richard Dawkins. Wilson, an evolutionary biologist at the State University of New York at Binghamton, focuses much of his argument at the group level. “Organisms are a product of natural selection,” he wrote in “Darwin’s Cathedral: Evolution, Religion, and the Nature of Society,” which came out in 2002, the same year as Atran’s book, and staked out the adaptationist view. “Through countless generations of variation and selection, [organisms] acquire properties that enable them to survive and reproduce in their environments. My purpose is to see if human groups in general, and religious groups in particular, qualify as organismic in this sense.”

Wilson’s father was Sloan Wilson, author of “The Man in the Gray Flannel Suit,” an emblem of mid-’50s suburban anomie that was turned into a film starring Gregory Peck. Sloan Wilson became a celebrity, with young women asking for his autograph, especially after his next novel, “A Summer Place,” became another blockbuster movie. The son grew up wanting to do something to make his famous father proud.
“I knew I couldn’t be a novelist,” said Wilson, who crackled with intensity during a telephone interview, “so I chose something as far as possible from literature — I chose science.” He is disarmingly honest about what motivated him: “I was very ambitious, and I wanted to make a mark.” He chose to study human evolution, he said, in part because he had some of his father’s literary leanings and the field required a novelist’s attention to human motivations, struggles and alliances — as well as a novelist’s flair for narrative.

Wilson eventually chose to study religion not because religion mattered to him personally — he was raised in a secular Protestant household and says he has long been an atheist — but because it was a lens through which to look at and revivify a branch of evolution that had fallen into disrepute. When Wilson was a graduate student at Michigan State University in the 1970s, Darwinians were critical of group selection, the idea that human groups can function as single organisms the way beehives or anthills do. So he decided to become the man who rescued this discredited idea. “I thought, Wow, defending group selection — now, that would be big,” he recalled. It wasn’t until the 1990s, he said, that he realized that “religion offered an opportunity to show that group selection was right after all.”

Dawkins once called Wilson’s defense of group selection “sheer, wanton, head-in-bag perversity.” Atran, too, has been dismissive of this approach, calling it “mind blind” for essentially ignoring the role of the brain’s mental machinery. The adaptationists “cannot in principle distinguish Marxism from monotheism, ideology from religious belief,” Atran wrote. “They cannot explain why people can be more steadfast in their commitment to admittedly counterfactual and counterintuitive beliefs — that Mary is both a mother and a virgin, and God is sentient but bodiless — than to the most politically, economically or scientifically persuasive account of the way things are or should be.”

Still, for all its controversial elements, the narrative Wilson devised about group selection and the evolution of religion is clear, perhaps a legacy of his novelist father. Begin, he says, with an imaginary flock of birds. Some birds serve as sentries, scanning the horizon for predators and calling out warnings. Having a sentry is good for the group but bad for the sentry, which is doubly harmed: by keeping watch, the sentry has less time to gather food, and by issuing a warning call, it is more likely to be spotted by the predator. So in the Darwinian struggle, the birds most likely to pass on their genes are the nonsentries. How, then, could the sentry gene survive for more than a generation or two?

To explain how a self-sacrificing gene can persist, Wilson looks to the level of the group. If there are 10 sentries in one group and none in the other, 3 or 4 of the sentries might be sacrificed. But the flock with sentries will probably outlast the flock that has no early-warning system, so the other 6 or 7 sentries will survive to pass on the genes. In other words, if the whole-group advantage outweighs the cost to any individual bird of being a sentry, then the sentry gene will prevail.

There are costs to any individual of being religious: the time and resources spent on rituals, the psychic energy devoted to following certain injunctions, the pain of some initiation rites. But in terms of intergroup struggle, according to Wilson, the costs can be outweighed by the benefits of being in a cohesive group that out-competes the others.

There is another element here too, unique to humans because it depends on language. A person’s behavior is observed not only by those in his immediate surroundings but also by anyone who can hear about it. There might be clear costs to taking on a role analogous to the sentry bird — a person who stands up to authority, for instance, risks losing his job, going to jail or getting beaten by the police — but in humans, these local costs might be outweighed by long-distance benefits. If a particular selfless trait enhances a person’s reputation, spread through the written and spoken word, it might give him an advantage in many of life’s challenges, like finding a mate. One way that reputation is enhanced is by being ostentatiously religious.
“The study of evolution is largely the study of trade-offs,” Wilson wrote in “Darwin’s Cathedral.” It might seem disadvantageous, in terms of foraging for sustenance and safety, for someone to favor religious over rationalistic explanations that would point to where the food and danger are.

But in some circumstances, he wrote, “a symbolic belief system that departs from factual reality fares better.” For the individual, it might be more adaptive to have “highly sophisticated mental modules for acquiring factual knowledge and for building symbolic belief systems” than to have only one or the other, according to Wilson. For the group, it might be that a mixture of hardheaded realists and symbolically minded visionaries is most adaptive and that “what seems to be an adversarial relationship” between theists and atheists within a community is really a division of cognitive labor that “keeps social groups as a whole on an even keel.”
Even if Wilson is right that religion enhances group fitness, the question remains: Where does God come in? Why is a religious group any different from groups for which a fitness argument is never even offered — a group of fraternity brothers, say, or Yankees fans?
Richard Sosis, an anthropologist with positions at the University of Connecticut and Hebrew University of Jerusalem, has suggested a partial answer. Like many adaptationists, Sosis focuses on the way religion might be adaptive at the individual level. But even adaptations that help an individual survive can sometimes play themselves out through the group. Consider religious rituals.

“Religious and secular rituals can both promote cooperation,” Sosis wrote in American Scientist in 2004. But religious rituals “generate greater belief and commitment” because they depend on belief rather than on proof. The rituals are “beyond the possibility of examination,” he wrote, and a commitment to them is therefore emotional rather than logical — a commitment that is, in Sosis’s view, deeper and more long-lasting.

Rituals are a way of signaling a sincere commitment to the religion’s core beliefs, thereby earning loyalty from others in the group. “By donning several layers of clothing and standing out in the midday sun,” Sosis wrote, “ultraorthodox Jewish men are signaling to others: ‘Hey! Look, I’m a haredi’ — or extremely pious — ‘Jew. If you are also a member of this group, you can trust me because why else would I be dressed like this?’ ” These “signaling” rituals can grant the individual a sense of belonging and grant the group some freedom from constant and costly monitoring to ensure that their members are loyal and committed. The rituals are harsh enough to weed out the infidels, and both the group and the individual believers benefit.
In 2003, Sosis and Bradley Ruffle of Ben Gurion University in Israel sought an explanation for why Israel’s religious communes did better on average than secular communes in the wake of the economic crash of most of the country’s kibbutzim. They based their study on a standard economic game that measures cooperation. Individuals from religious communes played the game more cooperatively, while those from secular communes tended to be more selfish. It was the men who attended synagogue daily, not the religious women or the less observant men, who showed the biggest differences. To Sosis, this suggested that what mattered most was the frequent public display of devotion. These rituals, he wrote, led to greater cooperation in the religious communes, which helped them maintain their communal structure during economic hard times.

In 1997, Stephen Jay Gould wrote an essay in Natural History that called for a truce between religion and science. “The net of science covers the empirical universe,” he wrote. “The net of religion extends over questions of moral meaning and value.” Gould was emphatic about keeping the domains separate, urging “respectful discourse” and “mutual humility.” He called the demarcation “nonoverlapping magisteria” from the Latin magister, meaning “canon.”
Richard Dawkins had a history of spirited arguments with Gould, with whom he disagreed about almost everything related to the timing and focus of evolution. But he reserved some of his most venomous words for nonoverlapping magisteria. “Gould carried the art of bending over backward to positively supine lengths,” he wrote in “The God Delusion.” “Why shouldn’t we comment on God, as scientists? . . . A universe with a creative superintendent would be a very different kind of universe from one without. Why is that not a scientific matter?”

The separation, other critics said, left untapped the potential richness of letting one worldview inform the other. “Even if Gould was right that there were two domains, what religion does and what science does,” says Daniel Dennett (who, despite his neo-atheist label, is not as bluntly antireligious as Dawkins and Harris are), “that doesn’t mean science can’t study what religion does. It just means science can’t do what religion does.”
The idea that religion can be studied as a natural phenomenon might seem to require an atheistic philosophy as a starting point. Not necessarily. Even some neo-atheists aren’t entirely opposed to religion. Sam Harris practices Buddhist-inspired meditation. Daniel Dennett holds an annual Christmas sing-along, complete with hymns and carols that are not only harmonically lush but explicitly pious.

And one prominent member of the byproduct camp, Justin Barrett, is an observant Christian who believes in “an all-knowing, all-powerful, perfectly good God who brought the universe into being,” as he wrote in an e-mail message. “I believe that the purpose for people is to love God and love each other.”
At first blush, Barrett’s faith might seem confusing. How does his view of God as a byproduct of our mental architecture coexist with his Christianity? Why doesn’t the byproduct theory turn him into a skeptic?

“Christian theology teaches that people were crafted by God to be in a loving relationship with him and other people,” Barrett wrote in his e-mail message. “Why wouldn’t God, then, design us in such a way as to find belief in divinity quite natural?” Having a scientific explanation for mental phenomena does not mean we should stop believing in them, he wrote. “Suppose science produces a convincing account for why I think my wife loves me — should I then stop believing that she does?”

What can be made of atheists, then? If the evolutionary view of religion is true, they have to work hard at being atheists, to resist slipping into intrinsic habits of mind that make it easier to believe than not to believe. Atran says he faces an emotional and intellectual struggle to live without God in a nonatheist world, and he suspects that is where his little superstitions come from, his passing thought about crossing his fingers during turbulence or knocking on wood just in case. It is like an atavistic theism erupting when his guard is down. The comforts and consolations of belief are alluring even to him, he says, and probably will become more so as he gets closer to the end of his life. He fights it because he is a scientist and holds the values of rationalism higher than the values of spiritualism.

This internal push and pull between the spiritual and the rational reflects what used to be called the “God of the gaps” view of religion. The presumption was that as science was able to answer more questions about the natural world, God would be invoked to answer fewer, and religion would eventually recede. Research about the evolution of religion suggests otherwise. No matter how much science can explain, it seems, the real gap that God fills is an emptiness that our big-brained mental architecture interprets as a yearning for the supernatural. The drive to satisfy that yearning, according to both adaptationists and byproduct theorists, might be an inevitable and eternal part of what Atran calls the tragedy of human cognition.

Robin Marantz Henig, a contributing writer, has written recently for the magazine about the neurobiology of lying and about obesity.

Sunday, March 4, 2007

Logotherapy: After Frankl

I've been working on this weird little sonnet. I can't imagine I'd ever get this to the point where I would attempt to publish it, but here it is--I'll discuss what I'm on about afterwards.

Logotherapy

We need the stars in our psychology;

given pale, inherited horizons,

kindling, enflamed by the eulogy

we seek mountains to pronounce illusion,

find books refuting eschatologies

before the crushing grief gets to the bones.

What’s gone is gone—a damned tautology

defeats us. Fucking time marches on, on

through revolutions and technologies,

boots going in and out of fashion,

young worlds cut, bleeding ideology—

and gone is gone. Fed by Auriga

we lock horns with the verdict: how far gone?


So, it's a sonnet--14 lines. Usually, the last two lines of a sonnet are comprised of a new cc rhyme, or conclude on aa. . . (if sonnet structure is alien to you, suffice it to say that I screwed around with the end rhyme, brought in Auriga, then reverted to the "on" sound, or the b rhyme again in line 14.)

It's not a successful poem by a far cry, but what I'm aiming for, in reverting to the b rhyme, is the idea that in our quest for meaning, as humans, we often fall back on the schools of thought which nourished us in the past--we cling to religion, or myth, or superstition, or even our childhood illusions.

Sometimes, this can be crippling to the imagination--we continue to rely on modes of thought that no longer serve us--but, at other times--it seems fair to suggest that these very faiths can help to pull us out of the swamp when nothing else will.

As a matter of fact, these belief systems can sometimes serve as a defense against even more harmful belief systems. Tom Robbins wrote that disbelief in magic can force a poor soul into believing in government and business, and I can attest that some of the more magical, intangible, and supernatural beliefs I held as a child still offer me some refuge against the harsh realities of adulthood.

For instance, I still look to the stars, and attempt to cull meaning from them, and the stories associated with them, and this perspective often allows me to overcome some of the more restrictive thinking necessitated by working in (with, for) the business community. Or at least not feel so trapped by it, enabling me to imagine, and sometimes really believe, that my life has greater purpose.

This is a very moderate example-- "logotherapy", or the school of psychiatry that examines the "will to meaning" was developed by Viktor Frankl, a Holocaust survivor and extrordinary thinker. To read more about logotherapy's assumptions, click here.

Logotherapy is concerned with "noetics."Noetic, a term derivitative of the Greek "nous", meaning, roughly, "intellect" or "consciousness. According to Wikipedia:

"The Institute of Noetic Sciences defines noetic as, roughly, 'relating to consciousness or intuition'. The mystical, cosmic and holistic connotations of the term are why the New Age movement has embraced the term noetic and used it to refer to the exploration of individual or cosmic consciousness."

These schools of thought are interesting to contrast with neuroscientific theories like Pinker's.

Is it possible to understand our own minds? Is religious faith a sort of hard wall which prevents people from questioning the nature of reality, and their own lives, any further?

I always revert back to prefering "the Great Mystery"--in this way, I suppose, my belief system mirrors Judaism's stance that human beings cannot know the will of God, while yet supposing that there is, in fact, a God.

Beyond that, however, I part ways with most religions--in fact, my supposition is more that, there is in fact "a something". Is that something spiritual? I don't know--it certainly would seem to be intelligent. And yet, I don't agree with teaching intelligent design to children when we human beings are still walking around with appendix and some children are still born with tails.

Essentially, at some point, the "spiritual" would converge with the merely human. My reasoning is as follows--if there is a higher spiritual plane capable of interacting with life on earth, this means that, by default, life on earth is connected to this higher plane--somehow they coexist. From this worldview, we are walking spirits who just happen to be caught in a dense vibration. . . It all comes back to "everything is one". . . one state of being passes into the next into the next, and these various levels are capable of informing and communicating with one another.

To what avail? We are were we are and we only know what we know--in order to find meaning, we have to consider the past, be very conscious of how we arrived at where we are, and begin to imagine what the purpose of this might be in the future.

It can become almost an adolescent's game of "if I hadn't gone to the mall that day, I never would have met Mathilda, and if not for meeting her, I'd have never applied to art school or bought a horse and if it wasn't for my horse, well, I'd never have spent that year in college. . . " ( note: the sentence "if it wasn't for my horse, I'd never have spent that year in college", is a famous Lewis Black-ism--one of those overheard snippets of conversation that it's so impossible to contextualize that you'll go mad if you try-- come to think of it, Lew's routine is actually in the noetic vein---when we are unable to make sense of our world, we begin to seriously lose it. He jokes that statements like that are where aneurysms come from--these words get inside our heads and then--pop--he may not be so far off....)

Anyhow: once we figure out how we got to where we are, we have to imagine where we might be going. . .planning the future is a big preoccupation of mine. . .finding meaning and purpose also used to be high up there, but my plans never come off quite as anticipated, so I often feel that if it isn't the gods determining my destiny, it surely is government and big business and my social status.

If Pinker and co. are right, our brains seem to be hellbent on making meaning--which many use to argue against the existence of an 'other-life' or spiritual realm beyond our ken. The argument is that since our consciousness is geared toward creating the illusion of meaning, in order to help us survive and strategize, why, then. . .surely our conceptualizations of God and the Divine are part and parcel of this great, selective delusion. . .

Ultimately it is impossible to know what lies beyond our own consciousness. . .and therein lies the conundra. We are getting closer to understanding what is "behind" it, but the beyond remains up for debate. . .and always will. . . .

and if there is beauty in that, and the possibility of peace, there is also considerable alienation and doubt . . . . . and room for many wars to be fought still. Surrendering to the mystery, to 'not knowing', seems the only chance we have for peace. . .

I'm rambling now, just wanted to point out the existence of Logotherapy, and recommend Frankl's book "Man's Search for Meaning", a book that, oddly, found me at a time when I really needed a jolt of meaning in my own life. . . . coincidence? I tend to doubt there's any such thing, but I do understand that the jury is out. . . .

Thursday, March 1, 2007

Daffodils

Well, it's March. The creative laze is passing. Finally, finally. Now I'm waiting for April.

March's flower is the daffodil--and I don't have any pictures of daffodils, so here are some day lilies from my garden last summer-- they won't bloom again til July, but they're yellow, and they're vaguely suggestive of daffodils. Kind of like jonquilian cognates. . .

Besides promoting cancer awareness, daffodils are symbols of the spring. . .and oddly enough, they are one of Aquarius' flowers--a month late. Not that anything associated with Aquarius is normal. It is the "sign of the future"--so maybe it is appropriate that it's flower be one that will only come into season after the sun has moved along.


Here's William Wordsworth's Daffodils:

I wandered lonely as a cloud
That floats on high o'er vales and hills,
When all at once I saw a crowd,
A host, of golden daffodils;
Beside the lake, beneath the trees,
Fluttering and dancing in the breeze.

Continuous as the stars that shine
And twinkle on the milky way,
They stretched in never-ending line
Along the margin of a bay:
Ten thousand saw I at a glance,
Tossing their heads in sprightly dance.

The waves beside them danced; but they
Out-did the sparkling waves in glee:
A poet could not but be gay,
In such a jocund company:
I gazed--and gazed--but little thought
What wealth the show to me had brought:

For oft, when on my couch I lie
In vacant or in pensive mood,
They flash upon that inward eye
Which is the bliss of solitude;
And then my heart with pleasure fills,
And dances with the daffodils.


Interesting that he likens them to stars. . . the link between stars and flowers is one that still fascinates me. Children doodle star-shapes meaning flowers, and every star sign is associated with certain flowers. Botanical life on earth comes from the stars as surely as we do . . . and many flowers seem possessed of undeniably starry signatures--almost like they're gesturing, twinkling, implying.







Wednesday, February 28, 2007

Creative Laze in February

That's what Kevin Smith calls it-- creative laze. . . .

And that's what I've been in for the latter half of cold, blues-triggering February. There's just something about this month. You feel it will never end. One of my favorite songs, ironically enough, is Dar Williams' "February". Is it legal for me to post her lyrics here? Well, I'm crediting her. It's a great tune:

I threw your keys in the water, I looked back,
They'd frozen halfway down in the ice.
They froze up so quickly, the keys and their owners,
Even after the anger, it all turned silent, and
The everyday turned solitary,
So we came to February.

First we forgot where we'd planted those bulbs last year,
Then we forgot that we'd planted at all,
Then we forgot what plants are altogether,
and I blamed you for my freezing and forgetting and
The nights were long and cold and scary,
Can we live through February?

You know I think Christmas was a long red glare,
Shot up like a warning, we gave presents without cards,
And then the snow,
And then the snow came, we were always out shoveling,
And we'd drop to sleep exhausted,
Then we'd wake up, and its snowing.

And February was so long that it lasted into March
And found us walking a path alone together.
You stopped and pointed and you said, "That's a crocus,"
And I said, "What's a crocus?" and you said, "It's a flower,"
I tried to remember, but I said, "What's a flower?"
You said, "I still love you."

The leaves were turning as we drove to the hardware store,
My new lover made me keys to the house,
And when we got home, well we just started chopping wood,
Because you never know how next year will be,
And we'll gather all our arms can carry,
I have lost to February.



I love the bit about "I tried to remember, but I said 'what's a flower'? I've added some of her albums to the Amazon store--she's woefully under-appreciated, and her music is a lot fun, often funny, sometimes "poignant" (that's an in-joke) and always inspiring. Highly recommended.

She's captured it, here, the feeling of February-- shortest month, but it can seem endless. I've often wondered if this is related to the beginning of Pisces, that formlessness, that dissolution. I often feel there's nothing to grab hold of by Feb 15 (and it ain't the Valentine Blues--I got my stuffed toy and my chocolates).

Here, in New Brunswick, the minus 30 cold snap just seemed to slog on and on. I found myself snapping at coworkers, complaining long and pointlessly to my significant other, and, in general, cursing much more than usual. Of course, this is the sort of frustration that fuels "plans for the future" (oooh, the portentive whisper). It is a dreaming time, I guess. That's what I keep reminding myself: "It's not quite time to wake up yet". Pisces is extremely connected to the "dream time"--- and I have had a thousand and one brilliant ideas this month. I just don't have the energy or the will to execute them quite yet. March 21st can't come fast enough, but then I'll be complaining there's no time available for daydreaming.

Can't win, can't win--a negative Piscean sentiment.

But it is now 11:14pm on February 28th. It's March in England!! It will be March here in 36 minutes. And the very word inspires energy and will: "March!" The weather has been warming up a little. . . it's almost time to think about seeds. Last October's gourds are now hardened and ready to glaze. . . . .a new cycle is just about to begin. But here in the Northern Climes, there is still a little time for dreaming yet. . . and I am indulging, even as the days lengthen, even as my energy returns. . .

For more on Pisces, here is the link to my entry on the Pisces Moon.

Happy March, everyone!! It's almost time.

Wednesday, February 7, 2007

Contemporary Heresy: Denying Mainstream Wisdom

Frank Furedi's excellent article on contemporary heresy, and the importance of being free to deny, or contest mainstream wisdom. He raises a very interesting point--whereas "denial" was once associated with critical thought, it is now associated with "refusal to acknowledge the truth"-- which is a very important distinction. Galileo denied that the Earth was the center of the universe and look what happened... If there is anything science has proven, it is that we never know as much as we think we do-- mainstream wisdom is a useful guide, but there will always be those visionaries who are able to see further, and deeper into certain issues than anybody else. These will invariably be those who think outside the box. Sure, sometimes denial is the result of quackery, but its important we keep our minds open and really listen for insights that go beyond mainstream, "common knowledge".

To read Furedi's article, click here.

Saturday, February 3, 2007

Climate Change




The UN's Intergovernmental Panel on Climate Change released their latest report this week, to much media frenzy and a flurry of debate, panic, fear-mongering, finger-pointing and anger. David Suzuki is on a cross-country tour of Canada. Executives continue to fly all over hell and creation, promoting their cities, their towns, pandering, pandering, selling, presenting, trying to promote their businesses, cities and towns as examples of sustainable economic ventures, locations or entities.

In my home province, the oil dictatorship has announced that it plans to build a second oil refinery.... the announcement of which came via a big, loud, balls-on media announcement (well, they do own every newspaper in the province)-- meanwhile, quietly, certain government officials look into tidal and wind power. Not much happens. There's just no infrastructure, and everybody wants money.....why hitch your wagon to a not-for-immediate-profit star? Not when you have the requisite 2.5 kids to feed, kids that need to look really happy and well-fed and clothed for your campaign.

In neighbouring Nova Scotia, bigger steps toward developing wind and tidal power have been taken, but then, that particular province is sicker from coal than we are from our nuclear power plant. Prince Edward Island is doing quite a lot. . . I just hope they can stay above water.

It is almost tempting to say "So what?" The damage, apparently, if not completely irreversible, will take 1000 years to reverse. Will humanity survive that long? Can it? Maybe a few lucky ones will be able to hide out underground for 10 centuries, subsisting on astronaut packets, tang, and some miraculous underground reservoir of water. Magickally free of contaminants. But what's to save even them from the inevitable earthquakes and ensuing cave-ins?

Or perhaps we can live on great, big floating cities in the North and South Poles, in order to escape the devastating heat. Or maybe it really is about time we head off into deep-space in search of other life-sustaining planets.

That's how it feels. It feels impossible. Everyday in the news it is "environmental catastrophe" and "imminent pandemic". I don't know which to be more worried about, and regardless of what I do, oil magnates will continue to do what they're doing, and executives will continue to fly and fly and fly, their jet-fuel contrails slashing the already toxic sky like murderer's knives.

Winter came abnormally, disturbingly late this year in Atlantic Canada, and I keep thinking about the summer of 2005, when the heat waves made going outside near impossible. While winter has finally arrived this season, and it is typically cold outside, the mercury even dropping to -30 last week, evidence of climate change is deeply disconcerting, particularly for coastal dwellers such as myself. How much are the waters going to rise? Scientists have predicted my hometown will be underwater in 100 years.

How is a person to process all of this, emotionally and intellectually?

Canada is taking a beating for its failure to uphold Kyoto, and the issue of the environment is shaping up to be the number one issue in our next federal election. Our new Liberal Leader, Stephane Dion, a former Environment Minister, and the owner of a dog named Kyoto is heading the charge against Stephen Harper, and taking a lot of flack for his less than stellar environmental policies in the process.

I enjoy politics as much as the next person, the theatre of it, the cloak and dagger-ism, the legacies made and lost. . . but at some point, some issues have to transcend politics. That our leader should be making the best policy decisions possible on the environment should be as much a given as that he or she should be able to read at a Grade 6 level.

This means, of course, that we need leaders and policy-makers who are not subject to greed, and therefore not apt to fold to the interests of big business in the name of donations, sponsorships and kickbacks.

Industry is the primary cause of what's happening environmentally, and carbon-credits is the closest we've come to finding a way to control their emissions.

Where does the answer lie? Surely with all of us, but recycling our bottles is just one step in the process. The people have to strong-arm not only the government, but big business, and short of all quitting our jobs and opting out of the culture of convenience and status, how can we do that?
The debate is about to ramp up, but it doesn't seem we have a lot of time to talk, so we'd better think fast and we'd better be blunt and we'd better pour our precious, human intelligence and energy into looking for real solutions, whether or not things are as bad as the prognosticators say. Because even if the situation isn't quite so dire...it soon will be. If you don't change direction, you end up where you're headed, as the old dicton goes.

So, how does civilization change course without being forced to by unimaginable catastrophe?

Ideas? Anyone?

Wednesday, January 31, 2007

The Mystery of Consciousness


Pinker is definitely one of the most interesting thinkers around today.....





Friday, Jan. 19, 2007
The Mystery of Consciousness
By Steven Pinker


The young women had survived the car crash, after a fashion. In the five months since parts of her brain had been crushed, she could open her eyes but didn't respond to sights, sounds or jabs. In the jargon of neurology, she was judged to be in a persistent vegetative state. In crueler everyday language, she was a vegetable.
So picture the astonishment of British and Belgian scientists as they scanned her brain using a kind of MRI that detects blood flow to active parts of the brain. When they recited sentences, the parts involved in language lit up. When they asked her to imagine visiting the rooms of her house, the parts involved in navigating space and recognizing places ramped up. And when they asked her to imagine playing tennis, the regions that trigger motion joined in. Indeed, her scans were barely different from those of healthy volunteers. The woman, it appears, had glimmerings of consciousness.
Try to comprehend what it is like to be that woman. Do you appreciate the words and caresses of your distraught family while racked with frustration at your inability to reassure them that they are getting through? Or do you drift in a haze, springing to life with a concrete thought when a voice prods you, only to slip back into blankness? If we could experience this existence, would we prefer it to death? And if these questions have answers, would they change our policies toward unresponsive patients--making the Terri Schiavo case look like child's play?
The report of this unusual case last September was just the latest shock from a bracing new field, the science of consciousness. Questions once confined to theological speculations and late-night dorm-room bull sessions are now at the forefront of cognitive neuroscience. With some problems, a modicum of consensus has taken shape. With others, the puzzlement is so deep that they may never be resolved. Some of our deepest convictions about what it means to be human have been shaken.


It shouldn't be surprising that research on consciousness is alternately exhilarating and disturbing. No other topic is like it. As René Descartes noted, our own consciousness is the most indubitable thing there is. The major religions locate it in a soul that survives the body's death to receive its just deserts or to meld into a global mind. For each of us, consciousness is life itself, the reason Woody Allen said, "I don't want to achieve immortality through my work. I want to achieve it by not dying." And the conviction that other people can suffer and flourish as each of us does is the essence of empathy and the foundation of morality.


To make scientific headway in a topic as tangled as consciousness, it helps to clear away some red herrings. Consciousness surely does not depend on language. Babies, many animals and patients robbed of speech by brain damage are not insensate robots; they have reactions like ours that indicate that someone's home. Nor can consciousness be equated with self-awareness. At times we have all lost ourselves in music, exercise or sensual pleasure, but that is different from being knocked out cold.


THE "EASY" AND "HARD" PROBLEMS
WHAT REMAINS IS NOT ONE PROBLEM ABOUT CONSCIOUSNESS BUT two, which the philosopher David Chalmers has dubbed the Easy Problem and the Hard Problem. Calling the first one easy is an in-joke: it is easy in the sense that curing cancer or sending someone to Mars is easy. That is, scientists more or less know what to look for, and with enough brainpower and funding, they would probably crack it in this century.
What exactly is the Easy Problem? It's the one that Freud made famous, the difference between conscious and unconscious thoughts. Some kinds of information in the brain--such as the surfaces in front of you, your daydreams, your plans for the day, your pleasures and peeves--are conscious. You can ponder them, discuss them and let them guide your behavior. Other kinds, like the control of your heart rate, the rules that order the words as you speak and the sequence of muscle contractions that allow you to hold a pencil, are unconscious. They must be in the brain somewhere because you couldn't walk and talk and see without them, but they are sealed off from your planning and reasoning circuits, and you can't say a thing about them.
The Easy Problem, then, is to distinguish conscious from unconscious mental computation, identify its correlates in the brain and explain why it evolved.


The Hard Problem, on the other hand, is why it feels like something to have a conscious process going on in one's head--why there is first-person, subjective experience. Not only does a green thing look different from a red thing, remind us of other green things and inspire us to say, "That's green" (the Easy Problem), but it also actually looks green: it produces an experience of sheer greenness that isn't reducible to anything else. As Louis Armstrong said in response to a request to define jazz, "When you got to ask what it is, you never get to know."
The Hard Problem is explaining how subjective experience arises from neural computation. The problem is hard because no one knows what a solution might look like or even whether it is a genuine scientific problem in the first place. And not surprisingly, everyone agrees that the hard problem (if it is a problem) remains a mystery.


Although neither problem has been solved, neuroscientists agree on many features of both of them, and the feature they find least controversial is the one that many people outside the field find the most shocking. Francis Crick called it "the astonishing hypothesis"--the idea that our thoughts, sensations, joys and aches consist entirely of physiological activity in the tissues of the brain. Consciousness does not reside in an ethereal soul that uses the brain like a PDA; consciousness is the activity of the brain.


THE BRAIN AS MACHINE
SCIENTISTS HAVE EXORCISED THE GHOST FROM THE MACHINE NOT because they are mechanistic killjoys but because they have amassed evidence that every aspect of consciousness can be tied to the brain. Using functional MRI, cognitive neuroscientists can almost read people's thoughts from the blood flow in their brains. They can tell, for instance, whether a person is thinking about a face or a place or whether a picture the person is looking at is of a bottle or a shoe.


And consciousness can be pushed around by physical manipulations. Electrical stimulation of the brain during surgery can cause a person to have hallucinations that are indistinguishable from reality, such as a song playing in the room or a childhood birthday party. Chemicals that affect the brain, from caffeine and alcohol to Prozac and LSD, can profoundly alter how people think, feel and see. Surgery that severs the corpus callosum, separating the two hemispheres (a treatment for epilepsy), spawns two consciousnesses within the same skull, as if the soul could be cleaved in two with a knife.


And when the physiological activity of the brain ceases, as far as anyone can tell the person's consciousness goes out of existence. Attempts to contact the souls of the dead (a pursuit of serious scientists a century ago) turned up only cheap magic tricks, and near death experiences are not the eyewitness reports of a soul parting company from the body but symptoms of oxygen starvation in the eyes and brain. In September, a team of Swiss neuroscientists reported that they could turn out-of-body experiences on and off by stimulating the part of the brain in which vision and bodily sensations converge.


THE ILLUSION OF CONTROL
ANOTHER STARTLING CONCLUSION FROM the science of consciousness is that the intuitive feeling we have that there's an executive "I" that sits in a control room of our brain, scanning the screens of the senses and pushing the buttons of the muscles, is an illusion. Consciousness turns out to consist of a maelstrom of events distributed across the brain. These events compete for attention, and as one process outshouts the others, the brain rationalizes the outcome after the fact and concocts the impression that a single self was in charge all along.


Take the famous cognitive-dissonance experiments. When an experimenter got people to endure electric shocks in a sham experiment on learning, those who were given a good rationale ("It will help scientists understand learning") rated the shocks as more painful than the ones given a feeble rationale ("We're curious.") Presumably, it's because the second group would have felt foolish to have suffered for no good reason. Yet when these people were asked why they agreed to be shocked, they offered bogus reasons of their own in all sincerity, like "I used to mess around with radios and got used to electric shocks."


It's not only decisions in sketchy circumstances that get rationalized but also the texture of our immediate experience. We all feel we are conscious of a rich and detailed world in front of our eyes. Yet outside the dead center of our gaze, vision is amazingly coarse. Just try holding your hand a few inches from your line of sight and counting your fingers. And if someone removed and reinserted an object every time you blinked (which experimenters can simulate by flashing two pictures in rapid sequence), you would be hard pressed to notice the change. Ordinarily, our eyes flit from place to place, alighting on whichever object needs our attention on a need-to-know basis. This fools us into thinking that wall-to-wall detail was there all along--an example of how we overestimate the scope and power of our own consciousness.


Our authorship of voluntary actions can also be an illusion, the result of noticing a correlation between what we decide and how our bodies move. The psychologist Dan Wegner studied the party game in which a subject is seated in front of a mirror while someone behind him extends his arms under the subject's armpits and moves his arms around, making it look as if the subject is moving his own arms. If the subject hears a tape telling the person behind him how to move (wave, touch the subject's nose and so on), he feels as if he is actually in command of the arms.
The brain's spin doctoring is displayed even more dramatically in neurological conditions in which the healthy parts of the brain explain away the foibles of the damaged parts (which are invisible to the self because they are part of the self). A patient who fails to experience a visceral click of recognition when he sees his wife but who acknowledges that she looks and acts just like her deduces that she is an amazingly well-trained impostor. A patient who believes he is at home and is shown the hospital elevator says without missing a beat, "You wouldn't believe what it cost us to have that installed."


Why does consciousness exist at all, at least in the Easy Problem sense in which some kinds of information are accessible and others hidden? One reason is information overload. Just as a person can be overwhelmed today by the gusher of data coming in from electronic media, decision circuits inside the brain would be swamped if every curlicue and muscle twitch that was registered somewhere in the brain were constantly being delivered to them. Instead, our working memory and spotlight of attention receive executive summaries of the events and states that are most relevant to updating an understanding of the world and figuring out what to do next. The cognitive psychologist Bernard Baars likens consciousness to a global blackboard on which brain processes post their results and monitor the results of the others.


BELIEVING OUR OWN LIES
A SECOND REASON THAT INFORMATION MAY BE SEALED OFF FROM consciousness is strategic. Evolutionary biologist Robert Trivers has noted that people have a motive to sell themselves as beneficent, rational, competent agents. The best propagandist is the one who believes his own lies, ensuring that he can't leak his deceit through nervous twitches or self-contradictions. So the brain might have been shaped to keep compromising data away from the conscious processes that govern our interaction with other people. At the same time, it keeps the data around in unconscious processes to prevent the person from getting too far out of touch with reality.


What about the brain itself? You might wonder how scientists could even begin to find the seat of awareness in the cacophony of a hundred billion jabbering neurons. The trick is to see what parts of the brain change when a person's consciousness flips from one experience to another. In one technique, called binocular rivalry, vertical stripes are presented to the left eye, horizontal stripes to the right. The eyes compete for consciousness, and the person sees vertical stripes for a few seconds, then horizontal stripes, and so on.


A low-tech way to experience the effect yourself is to look through a paper tube at a white wall with your right eye and hold your left hand in front of your left eye. After a few seconds, a white hole in your hand should appear, then disappear, then reappear.


Monkeys experience binocular rivalry. They can learn to press a button every time their perception flips, while their brains are impaled with electrodes that record any change in activity. Neuroscientist Nikos Logothetis found that the earliest way stations for visual input in the back of the brain barely budged as the monkeys' consciousness flipped from one state to another. Instead, it was a region that sits further down the information stream and that registers coherent shapes and objects that tracks the monkeys' awareness. Now this doesn't mean that this place on the underside of the brain is the TV screen of consciousness. What it means, according to a theory by Crick and his collaborator Christof Koch, is that consciousness resides only in the "higher" parts of the brain that are connected to circuits for emotion and decision making, just what one would expect from the blackboard metaphor.


WAVES OF BRAIN
CONSCIOUSNESS IN THE BRAIN CAN BE TRACKED NOT JUST IN SPACE but also in time. Neuroscientists have long known that consciousness depends on certain frequencies of oscillation in the electroencephalograph (EEG). These brain waves consist of loops of activation between the cortex (the wrinkled surface of the brain) and the thalamus (the cluster of hubs at the center that serve as input-output relay stations). Large, slow, regular waves signal a coma, anesthesia or a dreamless sleep; smaller, faster, spikier ones correspond to being awake and alert. These waves are not like the useless hum from a noisy appliance but may allow consciousness to do its job in the brain. They may bind the activity in far-flung regions (one for color, another for shape, a third for motion) into a coherent conscious experience, a bit like radio transmitters and receivers tuned to the same frequency. Sure enough, when two patterns compete for awareness in a binocular-rivalry display, the neurons representing the eye that is "winning" the competition oscillate in synchrony, while the ones representing the eye that is suppressed fall out of synch.


So neuroscientists are well on the way to identifying the neural correlates of consciousness, a part of the Easy Problem. But what about explaining how these events actually cause consciousness in the sense of inner experience--the Hard Problem?


TACKLING THE HARD PROBLEM
TO APPRECIATE THE HARDNESS OF THE HARD PROBLEM, CONSIDER how you could ever know whether you see colors the same way that I do. Sure, you and I both call grass green, but perhaps you see grass as having the color that I would describe, if I were in your shoes, as purple. Or ponder whether there could be a true zombie--a being who acts just like you or me but in whom there is no self actually feeling anything. This was the crux of a Star Trek plot in which officials wanted to reverse-engineer Lieut. Commander Data, and a furious debate erupted as to whether this was merely dismantling a machine or snuffing out a sentient life.
No one knows what to do with the Hard Problem. Some people may see it as an opening to sneak the soul back in, but this just relabels the mystery of "consciousness" as the mystery of "the soul"--a word game that provides no insight.


Many philosophers, like Daniel Dennett, deny that the Hard Problem exists at all. Speculating about zombies and inverted colors is a waste of time, they say, because nothing could ever settle the issue one way or another. Anything you could do to understand consciousness--like finding out what wavelengths make people see green or how similar they say it is to blue, or what emotions they associate with it--boils down to information processing in the brain and thus gets sucked back into the Easy Problem, leaving nothing else to explain. Most people react to this argument with incredulity because it seems to deny the ultimate undeniable fact: our own experience.


The most popular attitude to the Hard Problem among neuroscientists is that it remains unsolved for now but will eventually succumb to research that chips away at the Easy Problem. Others are skeptical about this cheery optimism because none of the inroads into the Easy Problem brings a solution to the Hard Problem even a bit closer. Identifying awareness with brain physiology, they say, is a kind of "meat chauvinism" that would dogmatically deny consciousness to Lieut. Commander Data just because he doesn't have the soft tissue of a human brain. Identifying it with information processing would go too far in the other direction and grant a simple consciousness to thermostats and calculators--a leap that most people find hard to stomach. Some mavericks, like the mathematician Roger Penrose, suggest the answer might someday be found in quantum mechanics. But to my ear, this amounts to the feeling that quantum mechanics sure is weird, and consciousness sure is weird, so maybe quantum mechanics can explain consciousness.


And then there is the theory put forward by philosopher Colin McGinn that our vertigo when pondering the Hard Problem is itself a quirk of our brains. The brain is a product of evolution, and just as animal brains have their limitations, we have ours. Our brains can't hold a hundred numbers in memory, can't visualize seven-dimensional space and perhaps can't intuitively grasp why neural information processing observed from the outside should give rise to subjective experience on the inside. This is where I place my bet, though I admit that the theory could be demolished when an unborn genius--a Darwin or Einstein of consciousness--comes up with a flabbergasting new idea that suddenly makes it all clear to us.


Whatever the solutions to the Easy and Hard problems turn out to be, few scientists doubt that they will locate consciousness in the activity of the brain. For many nonscientists, this is a terrifying prospect. Not only does it strangle the hope that we might survive the death of our bodies, but it also seems to undermine the notion that we are free agents responsible for our choices--not just in this lifetime but also in a life to come. In his millennial essay "Sorry, but Your Soul Just Died," Tom Wolfe worried that when science has killed the soul, "the lurid carnival that will ensue may make the phrase 'the total eclipse of all values' seem tame."


TOWARD A NEW MORALITY
MY OWN VIEW IS THAT THIS IS backward: the biology of consciousness offers a sounder basis for morality than the unprovable dogma of an immortal soul. It's not just that an understanding of the physiology of consciousness will reduce human suffering through new treatments for pain and depression. That understanding can also force us to recognize the interests of other beings--the core of morality.


As every student in Philosophy 101 learns, nothing can force me to believe that anyone except me is conscious. This power to deny that other people have feelings is not just an academic exercise but an all-too-common vice, as we see in the long history of human cruelty. Yet once we realize that our own consciousness is a product of our brains and that other people have brains like ours, a denial of other people's sentience becomes ludicrous. "Hath not a Jew eyes?" asked Shylock. Today the question is more pointed: Hath not a Jew--or an Arab, or an African, or a baby, or a dog--a cerebral cortex and a thalamus? The undeniable fact that we are all made of the same neural flesh makes it impossible to deny our common capacity to suffer.
And when you think about it, the doctrine of a life-to-come is not such an uplifting idea after all because it necessarily devalues life on earth. Just remember the most famous people in recent memory who acted in expectation of a reward in the hereafter: the conspirators who hijacked the airliners on 9/11.


Think, too, about why we sometimes remind ourselves that "life is short." It is an impetus to extend a gesture of affection to a loved one, to bury the hatchet in a pointless dispute, to use time productively rather than squander it. I would argue that nothing gives life more purpose than the realization that every moment of consciousness is a precious and fragile gift.


Steven Pinker is Johnstone Professor of Psychology at Harvard and the author of The Language Instinct, How the Mind Works and The Blank Slate.
Jennifer H's Facebook profile