Friday, March 04, 2011

The Threatening Scent of Fertile Women

I've finally had time to check out the abundant links to background articles in John Tierney's Feb. 21 article in the NYTimes. I had not realized that there was such firm evidence that men can unconsciously detect, via changes in womens' cues and behavior and perhaps via pheromones, when women are at the ovulation stage of their menstrual cycle.
...recent studies have found large changes in cues and behavior when a woman is at this stage of peak fertility. Lap dancers get much higher tips (unless they’re taking birth-control pills that suppress ovulation, in which case their tips remain lower). The pitch of a woman’s voice rises. Men rate her body odor as more attractive and respond with higher levels of testosterone.
I would recommend reading the whole article, which describes experiments supporting rather elaborate evolutionary psychology theories about maintaining breeding relationships while also enhancing genetic diversity, such as:
...the “good genes” evolutionary explanation for adultery: a quick fling with a good-looking guy can produce a child with better genes, who will therefore have a better chance of passing along the mother’s genes. But this sort of infidelity is risky if the woman’s unsexy long-term partner finds out and leaves her alone to raise the child. So it makes sense for her to limit her risks by being unfaithful only at those times she’s fertile....By that same evolutionary logic, it makes sense for her partner to be most worried when she’s fertile.

Thursday, March 03, 2011

Hugs follow a 3-second rule.

A fascinating bit from ScienceNow points to work by Emese Nagy:
How long does a hug lasts? The quick answer is about 3 seconds, according to Nagy's study of the post-competition embraces of Olympic athletes. The long answer is more profound. A hug lasts about as much time as many other human actions and neurological processes, which supports a hypothesis that we go through life perceiving the present in a series of 3-second windows...Crosscultural studies dating back to 1911 have shown that people tend to operate in 3-second bursts. Goodbye waves, musical phrases, and infants' bouts of babbling and gesturing all last about 3 seconds. Many basic physiological events, such as relaxed breathing and certain nervous system functions do, too. And several other species of mammals and birds follow the general rule in their body-movement patterns. A 1994 study of giraffes, okapis, roe deer, raccoons, pandas, and kangaroos living in zoos, for example, found that although the duration of the animals' every move, from chewing to defecating, varied considerably, the average was...3 seconds...The results reinforce an idea current among some psychologists that intervals of about 3 seconds are basic temporal units of life that define our perception of the present moment...the "feeling of nowness" tends to last 3 seconds.

Hope over experience - the persistence of optimism

Many studies have shown that people are excessively optimistic about marriage, work, sports, health, and life expectancy. Maasey et al. have asked: does such optimism persists as people acquire feedback from real-world experiences? And, second, is optimism actually caused by desire or hope? These are important question, for many of life’s most consequential decisions (e.g., about health, investments, or relationships) feature both strong preferences and the chance to revise beliefs in light of new information (e.g., medical exams, balance statements, or a second date). They asked National Football League (NFL) fans to predict game outcomes before each week of the 17-week NFL season. Studying football predictions offered four important benefits:
First, the 17-week season provided participants with quick, frequent, and unambiguous feedback over a significant (and nonarbitrary) duration of time, and thus it provided an ideal context for evaluating the effect of experience on optimism. Second, NFL fans’ preferences for their favorite teams are strong and often held with a degree of intensity unlikely to be generated by incentives offered in the laboratory. Third, a number of alternative explanations for the effects of desirability, such as those implicating team strength and familiarity, can be controlled methodologically and statistically. Finally, unlike predictions in other emotionally important domains, football predictions offer the benefit of objective benchmarks—both ex ante and ex post—against which the accuracy of predictions can be evaluated.
Their results:
We found that people are optimistic in their predictions—they judge preferred outcomes to be more likely than nonpreferred outcomes. We extended this observation in two important ways. First, we showed that optimism persists despite extensive experience—football fans are as optimistic after 4 months of feedback as they are after 4 weeks of feedback. Second, we found that desirability fueled this optimistic bias. Using four distinct tests, and a wide variety of control variables, we found that optimistic predictions were robust and uniquely related to the desirability of the outcome.

Wednesday, March 02, 2011

Happy people live longer.

It's a common assumption that happier people live longer, but the convincing data on this has not been easy to come by. Studies have offered widely different and competing findings - some finding no causation or reverse causation, others suggesting that unidentified, unobserved factors influence both happiness and longevity and health. Frey points to work by Diener and Chan showing that many kinds of studies, using different methods, conclude that happiness has a positive causal effect on longevity and physiological health. One of the studies noted by their survey is a meta-analysis based on 24 studies that estimates that happy people live 14% longer than persons who report that they are unhappy. In a survey of people living in industrial countries, happier people enjoy an increased longevity of between 7.5 and 10 years. Happier people are also less likely to commit suicide, and they are less often the victims of accidents...In longitudinal studies individuals are followed over many years, to identify whether the happier ones live longer. In the famous "nun study" researchers asked young women about their subjective happiness level before they entered a monastery. Those who perceived themselves to be happier died at a median age of 93.5 years. Those who considered themselves to be less happy died at a median age of 86.6 years.

Conscious attention influences unconscious cognition.

This work by Martens et al. demonstrates that our conscious attention to one task area (such as noting geometric shapes) can sensitize unconscious processes in that area such as subliminal priming (facilitatory effects elicited by masked stimuli that are not consciously perceived.) This shows that unconscious processing is crucially dependent on top-down attention, and contrasts with the classical view that unconscious cognition is characterized by the lack of top-down influences.
Are unconscious processes susceptible to attentional influences? In two subliminal priming experiments, we investigated whether task sets differentially modulate the sensitivity of unconscious processing pathways. We developed a novel procedure for masked semantic priming of words (Experiment 1) and masked visuomotor priming of geometrical shapes (Experiment 2). Before presentation of the masked prime, participants performed an induction task in which they attended to either semantic or perceptual object features designed to activate a semantic or perceptual task set, respectively. Behavioral and electrophysiological effects showed that the induction tasks differentially modulated subliminal priming: Semantic priming, which involves access to conceptual meaning, was found after the semantic induction task but not after the perceptual induction task. Visuomotor priming was observed after the perceptual induction task but not after the semantic induction task. These results demonstrate that unconscious cognition is influenced by attentional control. Unconscious processes in perceptual and semantic processing streams are coordinated congruently with higher-level action goals.

Tuesday, March 01, 2011

Go easy on yourself...

Being a self critical person who rarely gets off his own case,  I was considerably softened by a Sarah Parker-Pope article in this morning's Science section of the NYTimes on:
...a burgeoning new area of psychological research called self-compassion — how kindly people view themselves...The research suggests that giving ourselves a break and accepting our imperfections may be the first step toward better health. People who score high on tests of self-compassion have less depression and anxiety, and tend to be happier and more optimistic. Preliminary data suggest that self-compassion can even influence how much we eat and may help some people lose weight.
The article contains links to several original research papers.

Sharing the spoils develops early in humans.

From Tomasello and collaborators:
Egalitarian behavior is considered to be a species-typical component of human cooperation. Human adults tend to share resources equally, even if they have the opportunity to keep a larger portion for themselves. Recent experiments have suggested that this tendency emerges fairly late in human ontogeny, not before 6 or 7 years of age. Here we show that 3-year-old children share mostly equally with a peer after they have worked together actively to obtain rewards in a collaboration task, even when those rewards could easily be monopolized. These findings contrast with previous findings from a similar experiment with chimpanzees, who tended to monopolize resources whenever they could. The potentially species-unique tendency of humans to share equally emerges early in ontogeny, perhaps originating in collaborative interactions among peers.

Monday, February 28, 2011

Politics in Madison, WI

I spend the spring, summer, and fall months in Madison, WI., writing this blog from my university office there, and so have been closely following the theatrical union busting efforts of the new Republican majority.  With all the agonizing over long term debt incurred by the states, I thought this graphic from yesterday's NY Times Magazine, from an article on New Jersey governor Christie, was revealing (click on figure to enlarge): 

Improving your cognitive toolkit - V

Continuation of my abstracting of a few of the answers to the annual question at edge.org, "What scientific concept would improve everybody's cognitive toolkit?":

Daniel Dennett - Look for the Cycles
A good rule of thumb...when confronting the apparent magic of the world of life and mind is: look for the cycles that are doing all the hard work....it turns out that all the "magic" of cognition depends, just as life itself does, on cycles within cycles of recurrent, re-entrant, reflexive information-transformation processes from the biochemical scale within the neuron to the whole brain sleep cycle, waves of cerebral activity and recovery revealed by EEGs. Computer programmers have been exploring the space of possible computations for less than a century, but their harvest of invention and discovery so far includes millions of loops within loops within loops. The secret ingredient of improvement is always the same: practice, practice, practice.
Andy Clark - We are engines of prediction
The basic idea is simple, the brain is basically an engine of prediction. To perceive the world is to successfully predict our own sensory states. The brain uses stored knowledge about the structure of the world and the probabilities of one state or event following another to generate a prediction of what the current state is likely to be, given the previous one and this body of knowledge. Mismatches between the prediction and the received signal generate error signals that nuance the prediction or (in more extreme cases) drive learning and plasticity.
Some Implications:
First, the notion of good ('veridical') sensory contact with the world becomes a matter of applying the right expectations to the incoming signal...Second, the time course of perception becomes critical. Predictive coding models suggest that what emerges first is the general gist (including the general affective feel) of the scene, with the details becoming progressively filled in as the brain uses that larger context — time and task allowing — to generate finer and finer predictions of detail...Third, the line between perception and cognition becomes blurred. What we perceive (or think we perceive) is heavily determined by what we know, and what we know (or think we know) is constantly conditioned on what we perceive (or think we perceive)...Fourth, if we now consider that prediction errors can be suppressed not just by changing predictions but by changing the things predicted, we have a simple and powerful explanation for behavior and the way we manipulate and sample our environment. In this view, action is there to make predictions come true.

Friday, February 25, 2011

A potpuorri of almost-posts

As I scan journals' tables of contents and abstracts of articles looking for potential MindBlog posts, I accumulate a list of post candidates. Most never make it into a post, yet are very interesting chunks. My list now is 30 pages long, and I though I would begin to dispel that list by selecting a few to pass as a list of links, hoping that a few readers might find an item useful to them.

Writing about testing worries boosts exam performance.

Memory formation without the hippocampus.

Temporary suppression of fear memories during adolescence.

Solving the cocktail party problem.

Your Brain on Improv....to be creative, you have to have this weird dissociation in your frontal lobe. One area turns on, and a big area shuts off, so that you're not inhibited, so that you're willing to make mistakes, so that you're not constantly shutting down all of these new generative impulses.

Control your spotlight.

Retracted Autism Study.

Past adversity strengthens.

Courage.

Your brain on technology.

Krugman - The New Voodo.

Oliver Sachs - Change your brain.

Thursday, February 24, 2011

How stress damages or enhances different aspects of our vision

Richie Davidson, Alex Schackman and colleagues report a fascinating study that demonstrates that stress (painful, but not harmful, random electric shocks delivered to informed consent study participants) makes us more sensitive to our external surroundings as a way of learning where or what a threat may be, but interferes with our ability to do more complex thinking:
Stress can fundamentally alter neural responses to incoming information. Recent research suggests that stress and anxiety shift the balance of attention away from a task-directed mode, governed by prefrontal cortex, to a sensory-vigilance mode, governed by the amygdala and other threat-sensitive regions. A key untested prediction of this framework is that stress exerts dissociable effects on different stages of information processing. This study exploited the temporal resolution afforded by event-related potentials to disentangle the impact of stress on vigilance, indexed by early perceptual activity, from its impact on task-directed cognition, indexed by later postperceptual activity in humans. Results indicated that threat of shock amplified stress, measured using retrospective ratings and concurrent facial electromyography. Stress also double-dissociated early sensory-specific processing from later task-directed processing of emotionally neutral stimuli: stress amplified N1 (184–236 ms) and attenuated P3 (316–488 ms) activity. This demonstrates that stress can have strikingly different consequences at different processing stages. Consistent with recent suggestions, stress amplified earlier extrastriate activity in a manner consistent with vigilance for threat (N1), but disrupted later activity associated with the evaluation of task-relevant information (P3). These results provide a novel basis for understanding how stress can modulate information processing in everyday life and stress-sensitive disorders.

Wednesday, February 23, 2011

Uncertainty increases romantic attraction

Another interesting bit of work from Daniel Gilbert and collaborators:
This research qualifies a social psychological truism: that people like others who like them (the reciprocity principle). College women viewed the Facebook profiles of four male students who had previously seen their profiles. They were told that the men (a) liked them a lot, (b) liked them only an average amount, or (c) liked them either a lot or an average amount (uncertain condition). Comparison of the first two conditions yielded results consistent with the reciprocity principle. Participants were more attracted to men who liked them a lot than to men who liked them an average amount. Results for the uncertain condition, however, were consistent with research on the pleasures of uncertainty. Participants in the uncertain condition were most attracted to the men—even more attracted than were participants who were told that the men liked them a lot. Uncertain participants reported thinking about the men the most, and this increased their attraction toward the men.

Tuesday, February 22, 2011

Strategies for remembering.

Two recent articles offer different perspectives on retaining information. The interesting NYTimes magazine article by world class memory athlete Joshua Foer describes an ancient memory training tradition that flourished in ancient Greece before the invention of printing.  Brain imaging of  memory athletes who have engaged this training (who are of ordinary intelligence in virtually all other respects) shows more activation of brain areas known to be involved in spatial memory during memory tasks. The training derives from discoveries of Simonides of Ceos in the fifth century B.C.,
...who reasoned that just about anything could be imprinted upon our memories, and kept in good order, simply by constructing a building in the imagination and filling it with imagery of what needed to be recalled. This imagined edifice could then be walked through at any time in the future. Such a building would later come to be called a memory palace...a short Latin rhetoric textbook called “Rhetorica ad Herennium,” written sometime between 86 and 82 B.C. It is the only comprehensive discussion of the memory techniques attributed to Simonides to have survived into the Middle Ages. The techniques described in this book were widely practiced in the ancient and medieval worlds. Memory training was considered a centerpiece of classical education in the language arts, on par with grammar, logic and rhetoric...What distinguishes a great mnemonist...is the ability to create lavish images on the fly, to paint in the mind a scene so unlike any other it cannot be forgotten. And to do it quickly. Many competitive mnemonists argue that their skills are less a feat of memory than of creativity.
Both ancient and more recent memory enhancement techniques focus on the processing that occurs when we encode knowledge. A recent fairly simple experiment by Karpicke and Blunt suggests the effectiveness of another strategy that emphasizes practice.  From their introduction:
It is beyond question that activities that promote effective encoding, known as elaborative study tasks, are important for learning. However, research in cognitive science has challenged the assumption that retrieval is neutral and uninfluential in the learning process. Not only does retrieval produce learning, but a retrieval event may actually represent a more powerful learning activity than an encoding event. This research suggests a conceptualization of mind and learning that is different from one in which encoding places knowledge in memory and retrieval simply accesses that stored knowledge. Because each act of retrieval changes memory, the act of reconstructing knowledge must be considered essential to the process of learning.
Their abstract:
Educators rely heavily on learning activities that encourage elaborative studying, whereas activities that require students to practice retrieving and reconstructing knowledge are used less frequently. Here, we show that practicing retrieval produces greater gains in meaningful learning than elaborative studying with concept mapping. The advantage of retrieval practice generalized across texts identical to those commonly found in science education. The advantage of retrieval practice was observed with test questions that assessed comprehension and required students to make inferences. The advantage of retrieval practice occurred even when the criterial test involved creating concept maps. Our findings support the theory that retrieval practice enhances learning by retrieval-specific mechanisms rather than by elaborative study processes. Retrieval practice is an effective tool to promote conceptual learning about science.

Monday, February 21, 2011

Improving your cognitive toolkit - IV

Continuation of my abstracting of a few of the answers to the annual question at edge.org, "What scientific concept would improve everybody's cognitive toolkit?":

Charles Seife - Random events in the aggregate behave predictably.
Randomness is daunting; it sets limits where even the most sophisticated theories can not go, shielding elements of nature from even our most determined inquiries. Nevertheless, to say that something is random is not equivalent to saying that we can't understand it. Far from it...Randomness follows its own set of rules — rules that make the behavior of a random process understandable and predictable...These rules state that even though a single random event might be completely unpredictable, a collection of independent random events is extremely predictable — and the larger the number of events, the more predictable they become. The law of large numbers is a mathematical theorem that dictates that repeated, independent random events converge with pinpoint accuracy upon a predictable average behavior. Another powerful mathematical tool, the central limit theorem, tells you exactly how far off that average a given collection of events is likely to be. With these tools, no matter how chaotic, how strange a random behavior might be in the short run, we can turn that behavior into stable, accurate predictions in the long run.

The rules of randomness are so powerful that they have given physics some of its most sacrosanct and immutable laws. Though the atoms in a box full of gas are moving at random, their collective behavior is described by a simple set of deterministic equations. Even the laws of thermodynamics derive their power from the predictability of large numbers of random events; they are indisputable only because the rules of randomness are so absolute...Paradoxically, the unpredictable behavior of random events has given us the predictions that we are most confident in.
Emanuel Derman - Distrust Pragmamorphism
Anthropomorphism means attributing the characteristics of human beings to inanimate things or animals...I have invented the word pragmamorphism as a short-hand extension for the attribution of the properties of inanimate things to human beings...One of the meanings of the Greek word pragma is a material object...It's pragmamorphic to equate material correlates with human psychological states, to equate PET scans with emotion. It's also pragmamorphic to ignore human qualities you cannot measure.

We have discovered useful metrics for material objects -- length, temperature, pressure, volume, kinetic energy, etc. Pragmamorphism is a good word for the attempt to assign such one-dimensional thing-metrics to the mental qualities of humans...IQ, a length scale for intelligence, is a result of pragmamorphism. Intelligence is more diffuse than linear...The utility function in economics is similar. It's clear that people have preferences. But is it clear that there is a function that describes their preferences?
Nicholas Carr - Respect your cognitive load
...our brains can hold only about seven pieces of information simultaneously. Even that figure may be too high. Some brain researchers now believe that working memory has a maximum capacity of just three or four elements...The amount of information entering our consciousness at any instant is referred to as our cognitive load. When our cognitive load exceeds the capacity of our working memory, our intellectual abilities take a hit. Information zips into and out of our mind so quickly that we never gain a good mental grip on it...The information vanishes before we've had an opportunity to transfer it into our long-term memory and weave it into knowledge. We remember less, and our ability to think critically and conceptually weakens. An overloaded working memory also tends to increase our distractedness.

...it's important to remember that, when it comes to the way your brain works, information overload is not just a metaphor; it's a physical state. When you're engaged in a particularly important or complicated intellectual task, or when you simply want to savor an experience or a conversation, it's best to turn the information faucet down to a trickle.

Sunday, February 20, 2011

Empire at the End of Decadence

Since MindBlog seems to be doing a political weekend, I'm thinking I should, after yesterday's pointer to George Lakoff, also note the excellent summary by Charles Blow in Friday's NYTimes on the eroding American Empire. Here is the graphic (you should be able to read the small print if you click on the image.)

Friday, February 18, 2011

Suppressing emotions correlates with larger prefrontal cortical area

I'm a "keep calm and carry on" kind of guy, always suppressing my emotional reactivity in charged situations (which is probably why my colleagues said I was a good university department chair back when...). Kühn et al suggest that this might mean that the grey matter (nerve cell) volume of my dorsomedial (top, toward the middle) prefrontal cortex is a bit larger than that of those exploding around me:
There is a growing appreciation that individuals differ systematically in their use of particular emotion regulation strategies. Our aim was to examine the structural correlates of the habitual use of expressive suppression of emotions. Based on our previous research on the voluntary suppression of actions we expected this response-focused emotion regulation strategy to be associated with increased grey matter volume in the dorsomedial prefrontal cortex (dmPFC). On high-resolution MRI scans of 42 college-aged healthy adults we computed optimized voxel-based-morphometry (VBM) to explore the correlation between grey matter volume and inter-individual differences in the tendency to suppress the expression of emotions assessed by means of the Emotion Regulation Questionnaire (Gross & John, 2003). We found a positive correlation between the habitual use of expressive suppression as an emotion regulation strategy and grey matter volume in the dmPFC. No other brain area showed a significant positive or negative correlation with the Emotion Regulation Questionnaire scores. The association between the suppression of expression of emotions and volume in the dmPFC supports the behavioural stability and biological foundation of the concept of this particular emotion regulation strategy within an age-homogenous sample of adults.

Thursday, February 17, 2011

Placebo in 3 minutes

This sent by a loyal blog reader, pulled from this Australian site:

Blue-enriched light keeps us alert...

I think I've mentioned before that I used to put natural spectrum florescent lights (with more blue wavelengths) in my research laboratory, because I found it promoted my relaxation and alertness, and my graduate students and post-docs reported the same effect. Chellappa et al. offer yet another study that documents this effect, and relates it to suppression of melatonin levels by blue wavelengths. The observations can be made by simply comparing commercially available compact fluorescent lamps that provide correlated lamp colour temperature in kelvin (K), that indicate the relative proportion of warm versus cool colours in a light source. The authors found that 2-hour exposure to light in the evening with compact fluorescent light at 6500K (blue shifted) will attenuate the expression of endogenous melatonin levels, and also promote an augmentation of subjective and objective alertness levels when compared with lights at 2500K and at 3000K (more yellow). The light with more blue wavelengths has the overall effect of enhancing alertness and performance in cognitive tasks specifically associated with sustained attention.

Wednesday, February 16, 2011

Skeuomorphs - why innovation is also a throwback

Joshua Brustein writes a nice piece on how innovation usually doffs an old hat, maintaining superfluous marks of its evolution (skeuomorphs - from the Greek words for tool and form) to ease people's comfort with the transition to the new.
Digital cameras produce a reassuringly retro but artificial shutter snap when you push the button to take a photograph; cellphones have keyboards with layouts originally meant to keep typewriters from jamming; and blue jeans have pockets that are a throwback to a time when watches dangled from chains...Designers in all fields are regularly confronted with versions of this choice: whether to incorporate cues to keep people grounded in what has come before, or scrap convention completely. In transportation, for instance, the power of steam engines was initially described in relation to that of horses, a practice that has continued to the present day. Automobile designers have incorporated visual cues suggesting carriages; for example, adding nonfunctional spokes on wheels...This tension is palpable in many efforts to create new digital media experiences. The Daily, Rupert Murdoch’s publication designed specifically for tablet computers, incorporates video and interactivity into what is essentially a newspaper. At the same time, it is designed to show up on a reader’s digital doorstep once a day, a concept that seems as old-fashioned as pocket watches when compared with Web sites that are updated continually...Apple, probably the best symbol of the march into a new digital era, also encourages designers to incorporate analog references in its devices. On the iPad, users enter appointments into a calendar that is encased in an on-screen leather ledger, scrawl notes on what looks like a legal pad and advance through digital books by swiping their fingers across the screen, prompting an animation that actually looks like a page being turned.

Sleep enhances memories relevant to the future.

Here is a fascinating bit of work from Wilhelm et al., which possibly explains why in my first moments of starting to awaken, I notice that finger sequences of piano pieces I am studying to perform are playing in my head....
The brain encodes huge amounts of information, but only a small fraction is stored for a longer time. There is now compelling evidence that the long-term storage of memories preferentially occurs during sleep. However, the factors mediating the selectivity of sleep-associated memory consolidation are poorly understood. Here, we show that the mere expectancy that a memory will be used in a future test determines whether or not sleep significantly benefits consolidation of this memory. Human subjects learned declarative memories (word paired associates) before retention periods of sleep or wakefulness. Postlearning sleep compared with wakefulness produced a strong improvement at delayed retrieval only if the subjects had been informed about the retrieval test after the learning period. If they had not been informed, retrieval after retention sleep did not differ from that after the wake retention interval. Retention during the wake intervals was not affected by retrieval expectancy. Retrieval expectancy also enhanced sleep-associated consolidation of visuospatial (two-dimensional object location task) and procedural motor memories (finger sequence tapping). Subjects expecting the retrieval displayed a robust increase in slow oscillation activity and sleep spindle count during postlearning slow-wave sleep (SWS). Sleep-associated consolidation of declarative memory was strongly correlated to slow oscillation activity and spindle count, but only if the subjects expected the retrieval test. In conclusion, our work shows that sleep preferentially benefits consolidation of memories that are relevant for future behavior, presumably through a SWS-dependent reprocessing of these memories.

Tuesday, February 15, 2011

Foundations of religious belief

Judith Shulevitz reviews James Kugel's "In the Valley of the Shadow - on the Foundations of Religious Belief." The book rose from the author's experience of still being alive seven years after being told he would die of cancer within a few years. His points on the utility of religious belief (even it if is a cognitive error) remind me of last Friday's MindBlog post on the utility of the size-weight illusion in throwing. Here are a few clips from the review:
...the recent debates about religion — is it a force for good or for evil, intrinsically violent or intrinsically peaceful? — have on the whole been a bit “narrow.” Too many pundits, anthropologists and evolutionary biologists fail to imagine their way into the rich, elusive mental condition called “believing in God” or “being religious.” They dismiss it as a neurosis, a superstition or a mistake. An otherwise appealing evolutionary theory of religion, for instance, holds that God and the gods are ghostlike entities created by a “hyperactive agent detection device” in the brain — that is, a hair-trigger response to unusual stimuli that evolved to protect us from danger, but wound up making us mistakenly attribute intention and even divinity to things that have none.

Kugel asks whether it’s the skeptics who are being willfully blind to the ancient truths bundled into these apparent errors. Consider the band of prehistoric hunter-gatherers made aware of their fragility by the magnitude of what they were up against. “This little group was endlessly overshadowed by all that was outside of them, forever on the receiving end of whatever You — immanent in the great Outside all around — happened to be dishing out,” he reminds us. To call their brains “hyperactive” because they identify that “You” as a mindful agent, Kugel says, is “ludicrous.” The “great Outside” was nearly all-powerful: why shouldn’t it mean to make things happen? “On the contrary,” Kugel writes, “it would require some sort of extraordinarily twisted spirit to look up and not see You, Your hand gloved in cloud and sky, Your voice mingling with cricket song and crashing waves, doing all the things that impinged on the little band’s existence. You were practically everything, and You completely overwhelmed their own little reality.”

Believing in God, Kugel suggests — possibly being a tad ahistorical — originally meant aligning yourself with the force of the universe, of humbly opening yourself up to its grandeur, more than it meant asserting faith in a particular deity. Kugel reviews the literature on epilepsy and the “God spot,” the “verbal conceptual association area” where various lobes of the brain come together. When stimulated, as in epileptic seizures, it has been shown to lead to visions of God or at least a sense of what one researcher called “connection with an overwhelmingly powerful being.” You could say the God spot proves that religion is a matter of brain malfunction, Kugel observes. Or you could call the epileptic’s aura “a privileged moment, an opening of the mind to something it cannot normally perceive.”

To the religious — or at least to Kugel and his sources — religion is an experience more than a cosmology. “It is not God’s sovereignty over the entire universe that is at issue so much as his sovereignty over the cubic centimeter of space that sits just in front of our own noses,” he writes. “That is to say, religion is first of all about fitting into the world and fitting into one’s borders. There may indeed be something ‘mythic’ about it, but it pales before the mythic quality of our own clumsy, modern selves.”

Improving your cognitive toolkit - part III

CContinuation of my abstracting of a few of the answers to the annual question at edge.org, "What scientific concept would improve everybody's cognitive toolkit?":

Sean Carroll - The Pointless Universe
Things happen because the laws of nature say they will — because they are the consequences of the state of the universe and the path of its evolution. Life on Earth doesn't arise in fulfillment of a grand scheme, but rather as a byproduct of the increase of entropy in an environment very far from equilibrium. Our impressive brains don't develop because life is guided toward greater levels of complexity and intelligence, but from the mechanical interactions between genes, organisms, and their surroundings.

None of which is to say that life is devoid of purpose and meaning. Only that these are things we create, not things we discover out there in the fundamental architecture of the world. The world keeps happening, in accordance with its rules; it's up to us to make sense of it and give it value.

Rudy Rucker - The World is Unpredictable
The media cast about for the proximate causes of life's windfalls and disasters. The public demands blocks against the bad and pipelines to the good. Legislators propose new regulations, fruitlessly dousing last year's fires, forever betting on yesterday's winning horses...A little-known truth: Every aspect of the world is fundamentally unpredictable. Computer scientists have long since proved this.

At a personal level, even if the world is as deterministic as a computer program, you still can't predict what you're going to do. This is because your prediction method would involve a mental simulation of you that produces its results slower than you. You can't think faster than you think. You can't stand on your own shoulders...It's a waste to chase the pipedream of a magical tiny theory that allows us to make quick and detailed calculations about the future. We can't predict and we can't control. To accept this can be a source of liberation and inner peace. We're part of the unfolding world, surfing the chaotic waves.

Monday, February 14, 2011

Anniversary of Deric's MindBlog

I just realized that Feb. 6 marked the start of year 6 of this blog. We've now clocked ~2,500 Posts, and there are roughly 2,000 subscribers (The green line in the FeedBurner summary plot) to the blog's RSS feed.


Last year's birthday notice indulged in a writing identity crisis which I will spare you from this year.

The Net Delusion

I've been meaning to point to Lee Siegel's review of Evgeny Morozov's new book "The Net Delusion - The Dark Side of Internet Freedom." A few clips:
Contrary to the “cyberutopians,” as he calls them, who consider the Internet a powerful tool of political emancipation, Morozov convincingly argues that, in freedom’s name, the Internet more often than not constricts or even abolishes freedom...He quotes the political blogger Andrew Sullivan, who proclaimed after protesters took to the streets in Tehran that “the revolution will be Twittered.” The revolution never happened, and the futilely tweeting protesters were broken with an iron hand... What was broadcast on Twitter and elsewhere was repression of the revolution. The Iranian regime used the Web to identify photographs of protesters; to find out their personal information and whereabouts (through Facebook, naturally); to distribute propagandistic videos; and to text the population into counterrevolutionary paranoia.

As Morozov points out, don’t expect corporations like Google to liberate anyone anytime soon. Google did business in China for four years before economic conditions and censorship demands — not human rights concerns — forced it out. And it is telling that both Twitter and Facebook have refused to join the Global Network Initiative, a pact that Morozov describes as “an industrywide pledge . . . to behave in accordance with the laws and standards covering the right to freedom of expression and privacy embedded in internationally recognized documents like the Universal Declaration of Human Rights.”

Morozov urges the cyberutopians to open their eyes to the fact that the ­asocial pursuit of profit is what drives social media. “Not surprisingly,” he writes, “the dangerous fascination with solving previously intractable social problems with the help of technology allows vested interests to disguise what essentially amounts to advertising for their commercial products in the language of freedom and liberation.” In 2007, when he was at the State Department, Jared Cohen wrote with tragic wrongheadedness that “the Internet is a place where Iranian youth can . . . say anything they want as they operate free from the grips of the police-state apparatus.” Thanks to the exciting new technology, many of those freely texting Iranian youths are in prison or dead. Cohen himself now works for Google as the director of “Google Ideas.”

Friday, February 11, 2011

The size weight illusion is not an illusion for throwing.

Zhu and Bingham offer an interesting bit of work that suggests that human throwing and speaking abilities developed in a manner that is consistent with their evolutionary history (Accurate long distance throwing ability is unique to humans):
Long-distance throwing is uniquely human and enabled Homo sapiens to survive and even thrive during the ice ages. The precise motoric timing required relates throwing and speech abilities as dependent on the same uniquely human brain structures. Evidence from studies of brain evolution is consistent with this understanding of the evolution and success of H. sapiens. Recent theories of language development find readiness to develop language capabilities in perceptual biases that help generate ability to detect relevant higher order acoustic units that underlie speech. Might human throwing capabilities exhibit similar forms of readiness? Recently, human perception of optimal objects for long-distance throwing was found to exhibit a size–weight relation similar to the size–weight illusion; greater weights were picked for larger objects and were thrown the farthest. The size–weight illusion is: lift two objects of equal mass but different size, the larger is misperceived to be less heavy than the smaller. The illusion is reliable and robust. It persists when people know the masses are equal and handle objects properly. Children less than 2 years of age exhibit it. These findings suggest the illusion is intrinsic to humans. Here we show that perception of heaviness (including the illusion) and perception of optimal objects for throwing are equivalent. Thus, the illusion is functional, not a misperception: optimal objects for throwing are picked as having a particular heaviness. The best heaviness is learned while acquiring throwing skill. We suggest that the illusion is a perceptual bias that reflects readiness to acquire fully functional throwing ability. This unites human throwing and speaking abilities in development in a manner that is consistent with the evolutionary history.

Thursday, February 10, 2011

Preverbal infants mentally represent social dominance.

Interesting observations from Thomsen et al:
Human infants face the formidable challenge of learning the structure of their social environment. Previous research indicates that infants have early-developing representations of intentional agents, and of cooperative social interactions, that help meet that challenge. Here we report five studies with 144 infant participants showing that 10- to 13-month-old, but not 8-month-old, infants recognize when two novel agents have conflicting goals, and that they use the agents’ relative size to predict the outcome of the very first dominance contests between them. These results suggest that preverbal infants mentally represent social dominance and use a cue that covaries with it phylogenetically, and marks it metaphorically across human cultures and languages, to predict which of two agents is likely to prevail in a conflict of goals.

Wednesday, February 09, 2011

Bias within - politics of the professoriat

In the Tuesday Science section of the NY Times, Tierny does a fascinating article on social psychologists, the folks who do research on racial prejudice, homophobia, sexism, stereotype threat and unconscious bias against minorities. He discusses a talk given by Jonathan Haidt at their national conference. Haight:
...polled his audience at the San Antonio Convention Center, starting by asking how many considered themselves politically liberal. A sea of hands appeared, and Dr. Haidt estimated that liberals made up 80 percent of the 1,000 psychologists in the ballroom. When he asked for centrists and libertarians, he spotted fewer than three dozen hands. And then, when he asked for conservatives, he counted a grand total of three.

“This is a statistically impossible lack of diversity,” Dr. Haidt concluded, noting polls showing that 40 percent of Americans are conservative and 20 percent are liberal. In his speech and in an interview, Dr. Haidt argued that social psychologists are a “tribal-moral community” united by “sacred values” that hinder research and damage their credibility — and blind them to the hostile climate they’ve created for non-liberals.

“Anywhere in the world that social psychologists see women or minorities underrepresented by a factor of two or three, our minds jump to discrimination as the explanation,” said Dr. Haidt, who called himself a longtime liberal turned centrist. “But when we find out that conservatives are underrepresented among us by a factor of more than 100, suddenly everyone finds it quite easy to generate alternate explanations.”
And one further clip from Tierney's article (which you should read).
Moynihan was shunned by many of his colleagues at Harvard as racist,” Dr. Haidt said. “Open-minded inquiry into the problems of the black family was shut down for decades, precisely the decades in which it was most urgently needed. Only in the last few years have liberal sociologists begun to acknowledge that Moynihan was right all along.”

Similarly, Larry Summers, then president of Harvard, was ostracized in 2005 for wondering publicly whether the preponderance of male professors in some top math and science departments might be due partly to the larger variance in I.Q. scores among men (meaning there are more men at the very high and very low ends). “This was not a permissible hypothesis,” Dr. Haidt said. “It blamed the victims rather than the powerful. The outrage ultimately led to his resignation. We psychologists should have been outraged by the outrage. We should have defended his right to think freely.”

Instead, the taboo against discussing sex differences was reinforced, so universities and the National Science Foundation went on spending tens of millions of dollars on research and programs based on the assumption that female scientists faced discrimination and various forms of unconscious bias. But that assumption has been repeatedly contradicted, most recently in a study published Monday in the Proceedings of the National Academy of Sciences by two Cornell psychologists, Stephen J. Ceci and Wendy M. Williams. After reviewing two decades of research, they report that a woman in academic science typically fares as well as, if not better than, a comparable man when it comes to being interviewed, hired, promoted, financed and published.

“Thus,” they conclude, “the ongoing focus on sex discrimination in reviewing, interviewing and hiring represents costly, misplaced effort. Society is engaged in the present in solving problems of the past.” Instead of presuming discrimination in science or expecting the sexes to show equal interest in every discipline, the Cornell researchers say, universities should make it easier for women in any field to combine scholarship with family responsibilities.

Walking improves your memory.

Erickson et al. show that exercise training increases the size of our hippocampus and improves memory. They divided 120 sedentary healthy adults in their mid-60s into two groups. From a summary:
One group walked around a track three times a week, building up to 40 minutes at a stretch; the other did a variety of less aerobic exercises, including yoga and resistance training with bands. After a year, brain scans showed that among the walkers, the hippocampus had increased in volume by about 2 percent on average; in the others, it had declined by about 1.4 percent. Such a decline is normal in older adults.
Here is the abstract:
The hippocampus shrinks in late adulthood, leading to impaired memory and increased risk for dementia. Hippocampal and medial temporal lobe volumes are larger in higher-fit adults, and physical activity training increases hippocampal perfusion, but the extent to which aerobic exercise training can modify hippocampal volume in late adulthood remains unknown. Here we show, in a randomized controlled trial with 120 older adults, that aerobic exercise training increases the size of the anterior hippocampus, leading to improvements in spatial memory. Exercise training increased hippocampal volume by 2%, effectively reversing age-related loss in volume by 1 to 2 y. We also demonstrate that increased hippocampal volume is associated with greater serum levels of BDNF, a mediator of neurogenesis in the dentate gyrus. Hippocampal volume declined in the control group, but higher preintervention fitness partially attenuated the decline, suggesting that fitness protects against volume loss. Caudate nucleus and thalamus volumes were unaffected by the intervention. These theoretically important findings indicate that aerobic exercise training is effective at reversing hippocampal volume loss in late adulthood, which is accompanied by improved memory function.

Tuesday, February 08, 2011

Did Frédéric Chopin have temporal lobe epilepsy?

I'm working up the incredible Chopin Fantasy in F minor for a spring concert, and am always eager to learn more about this remarkable composer (the technical requirements of his music conform to the natural musculature of the hands and arms in a way that no previous composer had...Bach and Beethoven sometimes make very unnatural and contortionistic demands.) Chopin was viewed as a tortured artist because at several performances he suddenly stopped in the middle of a piece and left the stage:
"I was about to play the [Funeral] March when, suddenly, I saw emerging from the half-open case of my piano those cursed creatures that had appeared to me on a lugubrious night at the Carthusian monastery. I had to leave for a while in order to recover myself, and after that I continued playing without saying a word."
An article by by Sara Reardon points to a paper by radiologist Manuel Vásquez Caruncho of Xeral-Calde Hospital in Lugo, Spain and neurologist Francisco Brañas Fernández that
...draws heavily from descriptions of Chopin's behavior by his friends and pupils and from his own writings. Their vivid recollections report finding the composer late at night, "pale in front of the piano, with wild eyes and his hair on end," unable to recognize them for short periods. He spoke often of a "cohort of phantoms" that haunted him, of seeing his friends as the walking dead, and feeling "like steam."
Only a handful of neurological disorders produce the phantasmagoria that tormented Chopin, who didn't abuse drugs or alcohol. The visions he described, such as demons crawling out of his piano, are now known as Lilliputian hallucinations: detailed visions of people or objects that are much smaller than they are in life. The authors rule out schizophrenia and other common psychoses because Chopin's hallucinations were visual, not auditory, and because he lacked other telltale symptoms such as eye problems or migraines. His short hallucinatory episodes are a hallmark of temporal lobe epilepsy,

Monday, February 07, 2011

Improving your cognitive toolkit - part II

This posts continues my abstracting of some of my favorite responses to the Edge.org annual question "What scientific concept would improve everybody's cognitive toolkit?"

Martin Seligman - PERMA
Is global well being possible?...The elements of well being must be exclusive, measurable independently of each other, and ideally, exhaustive. I believe there are five such elements and they have a handy acronym, PERMA, a shorthand abstraction for the enabling conditions of life:

P Positive Emotion
E Engagement
R Positive Relationships
M Meaning and Purpose
A Accomplishment

There has been forward movement in the measurement of these over the last decade. Taken together PERMA forms a more comprehensive index of well being than "life satisfaction" and it allows for the combining of objective and subjective indicators. PERMA can index the well being of individuals, of corporations, and of cities. The United Kingdom has now undertaken the measurement of well being for the nation and as one criterion — in addition to Gross Domestic Product — of the success of its public policy.
Steven Pinker - Positive-Sum Games
...when people become consciously aware of the game-theoretic structure of their interaction (that is, whether it is positive-, negative-, or zero-sum), they can make choices that bring them valuable outcomes — like safety, harmony, and prosperity — without their having to become more virtuous, noble, or pure...Some examples. Squabbling colleagues or relatives agree to swallow their pride, take their losses, or lump it to enjoy the resulting comity rather than absorbing the costs of continuous bickering in hopes of prevailing in a battle of wills. Two parties in a negotiation split the difference in their initial bargaining positions to "get to yes."

Has an increasing awareness of the zero- or nonzero-sumness of interactions in the decades since 1950 (whether referred to in those terms or not) actually led to increased peace and prosperity in the world? It's not implausible. International trade and membership in international organizations has soared in the decades that game-theoretic thinking has infiltrated popular discourse. And perhaps not coincidentally, the developed world has seen both spectacular economic growth and a historically unprecedented decline in several forms of institutionalized violence, such as war between great powers, war between wealthy states, genocides, and deadly ethnic riots. Since the 1990s these gifts have started to accrue to the developing world as well, in part because they have switched their foundational ideologies from ones that glorify zero-sum class and national struggle to ones that glorify positive-sum market cooperation. (All these claims can be documented from the literature in international studies.)

The enriching and pacifying effects of participation in positive-sum games long antedate the contemporary awareness of the concept. The biologists John Maynard Smith and Eörs Szathmáry have argued that an evolutionary dynamic which creates positive-sum games drove the major transitions in the history of life: the emergence of genes, chromosomes, bacteria, cells with nuclei, organisms, sexually reproducing organisms, and animal societies. In each transition, biological agents entered into larger wholes in which they specialized, exchanged benefits, and developed safeguards to prevent one from exploiting the rest to the detriment of the whole. The journalist Robert Wright sketched a similar arc in his book Nonzero and extended it to the deep history of human societies. An explicit recognition among literate people of the shorthand abstraction "positive-sum game" and its relatives may be extending a process in the world of human choices that has been operating in the natural world for billions of years.

Friday, February 04, 2011

Skilled object recognition uses both our left and right hemispheres

Bilalić et al. make the interesting observation that skilled chess players, while no faster or better than amateurs at geometric object recognition (which mainly engages left hemisphere), are more rapid than amateurs at identifying chess positions, while at the same time engaging additional areas of their right hemisphere. This expanded use of brain areas requires extensive training. (When the subjects were shown the chess diagrams, the novices looked directly at the pieces to recognize them, while the experts looked on the middle of the boards and took everything in with their peripheral vision.) (Wan et al. report a similar study in Japan examining experts in shogi, a game similar to chess. It highlights further brain areas involved in expertise.) Here is the Bilalić et al. abstract (dorsal means along the upper part of the brain, ventral is lower):
Our object recognition abilities, a direct product of our experience with objects, are fine-tuned to perfection. Left temporal and lateral areas along the dorsal, action related stream, as well as left infero-temporal areas along the ventral, object related stream are engaged in object recognition. Here we show that expertise modulates the activity of dorsal areas in the recognition of man-made objects with clearly specified functions. Expert chess players were faster than chess novices in identifying chess objects and their functional relations. Experts' advantage was domain-specific as there were no differences between groups in a control task featuring geometrical shapes. The pattern of eye movements supported the notion that experts' extensive knowledge about domain objects and their functions enabled superior recognition even when experts were not directly fixating the objects of interest. Functional magnetic resonance imaging (fMRI) related exclusively the areas along the dorsal stream to chess specific object recognition. Besides the commonly involved left temporal and parietal lateral brain areas, we found that only in experts homologous areas on the right hemisphere were also engaged in chess specific object recognition. Based on these results, we discuss whether skilled object recognition does not only involve a more efficient version of the processes found in non-skilled recognition, but also qualitatively different cognitive processes which engage additional brain areas.

Thursday, February 03, 2011

The biology of morality

I have already pointed to a TED talk by Sam Harris, and thought I would pass on a few clips from a review of his related book, "The Moral Landscape - How Science Can Determine Human Values." On Harris:
...his dispensation is that “Faith, if it is ever right about anything, is right by accident.” In applying reason to questions of morality, Harris claims that we can define morality only as it relates to the well-being of conscious organisms and that such well-being is completely measurable using the methods of neurobiology. This suggests to him that any action can be clearly classified as moral (increasing well-being) or immoral (decreasing well-being) without ambiguity. However, it doesn't mean that there is only one answer to a question of morality. He contends that “the existence of multiple peaks on the moral landscape does not make them any less real or worthy of discovery. Nor would it make the difference between being on a peak and being stuck deep in a valley any less clear or consequential.” But Harris firmly disagrees with the moral relativist views that there is no clearly defined morality that cuts across different societies and that therefore all views of morality are equally meritorious. He writes, “Multiculturalism, moral relativism, political correctness, tolerance even of intolerance—these are the familiar consequences of separating facts and values on the left.” “My goal,” he states, “is to convince you that human knowledge and human values can no longer be kept apart.”

Harris isn't choosy when it comes to vilifying religions. He notes the willingness of many to ignore genocide or cases of sexual abuse within their churches while taking strong actions against individuals who perform abortions (or refuse to prohibit them). He also draws from history examples of undeniably immoral choices in the name of religion. Harris criticizes scientists for persisting in their faith and for failing to confront head-on a society that he thinks is mired in superstition.

Harris thinks too many scientists have compromised on principles. “Many of our secular critics worry that if we oblige people to choose between reason and faith, they will choose faith and cease to support scientific research.” Even the journal Nature upholds the idea of nonoverlapping magisteria of Gould. Harris complains, “It is one thing to be told that the pope is a peerless champion of reason and that his opposition to embryonic stem-cell research is both morally principled and completely uncontaminated by religious dogmatism; it is quite another to be told this by a Stanford physician who sits on the President's Council on Bioethics.”

One might conclude that although at one time the best way to define and enforce moral behavior was through revealed faith, as science and reason advance, we can chip away at the old edifice and build anew. Stories of a young-Earth creation now look rather untenable, but in the past they might have been the only way to instill awe and teach a new and meaningful moral code. Rather than nonoverlapping magisteria, the domains of science and religion are intermingling all the time. The Moral Landscape may represent a new beach-head in this quest.

Wednesday, February 02, 2011

Meanwhile...back in Wisconsin

As I sit at the keyboard in my Fort Lauderdale snowbird condo looking out the open patio door, gentle breeze, 72 degree Farenheit,


My younger partner, still in the working world, emails me an iPhone picture he just took of our rural snowbound Middleton, Wisconsin home.  The blizzard has shut down all commercial and educational facilities.

What would improve your cognitive toolkit?

My first MindBlog post in 2006 was a description of answers given to an annual question posed each year to prominent public intellectuals by Edge.org. The question for 2010 is "What scientific concept would improve everybody's cognitive toolkit?"
A "scientific concept" may come from philosophy, logic, economics, jurisprudence, or other analytic enterprises, as long as it is a rigorous conceptual tool that may be summed up succinctly (or "in a phrase") but has broad application to understanding the world...James Flynn has defined "shorthand abstractions" (or "SHA's") as concepts drawn from science that have become part of the language and make people smarter by providing widely applicable templates ("market", "placebo", "random sample," "naturalistic fallacy," are a few of his examples). His idea is that the abstraction is available as a single cognitive chunk which can be used as an element in thinking and debate.
I'm going to give brief sketches of a few responses that I found most interesting.   I try to edit  the author's point to a single declarative phrase, the 'single cognitive chunk'  requirement suggested above (I'm surprised that in most cases the authors didn't do this more effectively). I'll list a few in this post, and as I have time to continue reading through the 164 contributions, perhaps do some further posts...

Howard Gardner - Try to disprove your viewpoint.
"If American citizens, or, for that matter, citizens anywhere were motivated to decribe the conditions under which they would relinquish their beliefs, they would begin to think scientifically. And if they admitted that empirical evidence would not change their minds, then at least they'd have indicated that their views have a religious or an ideological, rather than a scientific basis.

Christian Keysers - Avoid the mirror fallacy
...our brain mirrors the states of the people we observe...When the person we see has the exact same body and brain as we do, mirroring would tell us what the other feels. Whenever the other person is different in some relevant way, however, mirroring will mislead us...The world is full of such fallacies: we feel dolphins are happy just because their face resembles ours while we smile or we attribute pain to robots in sci-fi movies.

George Lakoff - Be aware of the conceptual metaphors you are using.
All concepts are physical brain circuits deriving their meaning via neural cascades that terminate in linkage to the body. That is how embodied cognition arises...Primary metaphors are brain mappings linking disparate brain regions, each tied to the body in a different way. For example, More Is Up (as in "prices rose") links a region coordinating quantity to another coordinating verticality...Complex conceptual metaphors arise via neural bindings, both across metaphors and from a given metaphor to a conceptual frame circuit. Metaphorical reasoning arises when source domain inference structures are used for target domain reasoning via neural mappings... A central consequence is the huge range of concepts that use metaphor cannot be defined relative to the outside world, but are instead embodied via interactions of the body and brain with the world...Every time you think of paying moral debts, or getting bogged down on a project, or losing time, or being at a crossroads in a relationship, you are unconsciously activating a conceptual metaphor circuit in your brain, reasoning using it, and quite possibly making decisions and living your life on the basis of your metaphors. And that's just normal. There's no way around it!..But it can do harm if you are unaware of it.

Tuesday, February 01, 2011

Modern virtue - the religion of physical fittness

As I have morphed during my life from a devout teenage christian church organist into a crusty old materialistic atheist, I have found a new church in the cult of physical exercise and fitness. Virtue and badness can be simply measured by whether I have worked out today. Until I read this fascinating tribute to Jack LeLane, who recently died at the age of 96, I had not realized what a modern invention my church is, growing from the opening of his first gym in Oakland, CA. in 1936:
With “The Jack LaLanne Show,” he also had a hand in the spread — a contagion, really — of television programs exhorting viewers to rise up from their La-Z-Boys...An army of spandex missionaries was unleashed....What he left behind when he died last week...was not only a sweaty culture of relentless crunching and spinning but also the notion that fitness equals character, and that self-actualization begins with the self-discipline to get and stay in shape. In the post-LaLanne landscape, it’s not the eyes but the abdominals that are windows to the soul...A “new you” usually means a trimmer, tauter version, not someone who has learned to speak Mandarin or picked up woodworking skills...There’s a bullying strain to the modern fitness ethos, a blurred line between cheerleading and hectoring...When exercise comes wrapped in value judgments, does it wind up entangled in an anxiety that threatens the very resolve to get fit? As Mr. LaLanne was siring new methods for shaping up, he was fathering something else, too: a potent, and in some cases immobilizing, strain of contemporary guilt.

Monday, January 31, 2011

Modern conversation.

As relevant to this morning's previous post, I had to pass on today's Doonesbury cartoon, having noted during my kid's New Year's visit how our conversations transitioned seamlessly from periods of actual talking to tapping on our iPhones and back to talking again, no requirement that one excuse oneself from actual talk, the accepted procedure was just to suddenly divert attention and start tapping on the new prosthetic device. This clearly is how the world of 20- and 30-somethings now works.

Is our technology replacing our identity?

After publishing an optimistic book about the internet in 1995 ("Life on the Screen"), MIT social science professor Sherry Turkle has now written a darker tome, "Alone Together,"  worrying that we are moving more of our lives online and away from real physical human contacts. The first half of the book is about social robots. From Jonah Lehrer's review:
“Dependence on a robot presents itself as risk free,” Turkle writes. “But when one becomes accustomed to ‘companionship’ without demands, life with people may seem overwhelming.” A blind date can be a fraught proposition when there’s a robot at home that knows exactly what we need. And all she needs is a power outlet...The reason robots are such a slippery slope, according to Turkle, is that they take advantage of a deeply human instinct. When it comes to the perception of other minds, we are extremely gul­lible, bestowing agency on even the most inanimate of objects.
The second part of the book deals with Turkle's concern that the internet is becoming our way of being with other people, in a style that turns them into objects. (Why did I just text my friend instead of actually just calling and talking with him.)
...the online world is no longer a space of freedom and re­invention. Instead, we have been trapped by Facebook profiles and Google cache, in which verbs like “delete” and “erase” are mostly metaphorical...We aren’t “happy” anymore: we’re simply a semicolon followed by a parenthesis. Instead of talking on the phone, we send a text; instead of writing wistful letters, we edit our Tumblr blog...these obvious objections shouldn’t obscure the real mystery: If the Internet is such an alienating force, then why can’t we escape it? If Facebook is so insufferable, then why do hundreds of millions of people check their page every day?
My own experience is that my participation in social web sites has broadened my world of real world contacts and friends, as noted by Lehrer:
...despite our misgivings about the Internet, its effects on real-life relationships seem mostly positive, if minor. A 2007 study at Michigan State University involving 800 undergraduates, for instance, found that Facebook users had more social capital than abstainers, and that the site increased measures of “psychological well-being,” especially in those suffering from low self-esteem. Other studies have found that frequent blogging leads to increased levels of social support and integration and may serve as “the core of building intimate relationships.” One recurring theme to emerge from much of this research is that most people, at least so far, are primarily using the online world to enhance their offline relationships, not supplant them.

Friday, January 28, 2011

Our social brain - what are smiles for?

In this past Tuesday's NYTimes science section, Carl Zimmer offers a brief review of some of the many social functions served by our smiling at each other (signaling happiness, social bonding, embarrassment, dominance, etc.). He focuses on the work of Paula Niedenthal. I found this particular bit interesting:
In one study, she and her colleagues are testing the idea that mimicry lets people recognize authentic smiles. They showed pictures of smiling people to a group of students. Some of the smiles were genuine and others were fake. The students could readily tell the difference between them....Then Dr. Niedenthal and her colleagues asked the students to place a pencil between their lips. This simple action engaged muscles that could otherwise produce a smile. Unable to mimic the faces they saw, the students had a much harder time telling which smiles were real and which were fake.

The scientists then ran a variation on the experiment on another group of students. They showed the same faces to the second group, but had them imagine the smiling faces belonged to salesclerks in a shoe store. In some cases the salesclerks had just sold the students a pair of shoes — in which they might well have a genuine smile of satisfaction. In other trials, they imagined that the salesclerks were trying to sell them a pair of shoes — in which case they might be trying to woo the customer with a fake smile...In reality, the scientists use a combination of real and fake smiles for both groups of salesclerks. When the students were free to mimic the smiles, their judgments were not affected by what the salesclerk was doing...But if the students put a pencil in their mouth, they could no longer rely on their mimicry. Instead, they tended to believe that the salesclerks who were trying to sell them shoes were faking their smiles — even when their smiles were genuine. Likewise, they tended to say that the salesclerks who had finished the sale were smiling for real, even when they weren’t. In other words, they were forced to rely on the circumstances of the smile, rather than the smile itself.

Thursday, January 27, 2011

Our Guts 'R Us

It dents my tidy self image just a bit when I read articles like this one from Atarashi et al. (summarized by Barnes and Powrie). A very indispensible part of my 'self' is an astounding 1014 bacteria that reside in the large intestine alone, alongside various viruses, fungi, protozoa, and parasites, all of which can affect chronic disease progression. Lee and Mazmanian point out that:
Although microbes have been classically viewed as pathogens, it is now well established that the majority of host-bacterial interactions are symbiotic. During development and into adulthood, gut bacteria shape the tissues, cells, and molecular profile of our gastrointestinal immune system. This partnership, forged over many millennia of coevolution, is based on a molecular exchange involving bacterial signals that are recognized by host receptors to mediate beneficial outcomes for both microbes and humans....specific aspects of the adaptive immune system are influenced by intestinal commensal bacteria. Understanding the molecular mechanisms that mediate symbiosis between commensal bacteria and humans may redefine how we view the evolution of adaptive immunity and consequently how we approach the treatment of numerous immunologic disorders.
The Lee and  Mazmanian review contains this striking graphic:


Legend: The microbiome of various anatomical locations of the human body. Numerous bacterial species colonize the mouth, upper airways, skin, vagina, and intestinal tract of humans. The phylogenetic trees show the speciation of bacterial clades from common ancestors at each anatomical site. Although the communities in different regions of the body share similarities, they each have a unique site-specific “fingerprint” made of many distinct microbes. Each site has a very high level of diversity, as shown by the individual lines on the dendrograms. Data are from the NIH-funded Human Microbiome Project; circles represent bacterial species whose sequences are known.
The  Atarashi et al. article demonstrates that
...indigenous species of Clostridium bacteria, a large component of our mammalian microbiota, promote anti-inflammatory immune responses by expanding and activating regulatory T cells...Oral inoculation of Clostridium during the early life of conventionally reared mice results in resistance to colitis and systemic immunoglobulin E responses in adult mice, suggesting a new therapeutic approach to autoimmunity and allergy.
These results are of some interest to us humans!

Wednesday, January 26, 2011

Some factoids on exercise and health.

I thought I would list out four interesting chunks from the Wellness Blog at the NY Times:

It's well know that exercise can be an antidote to depression. This is particularly relevant to people living in Northern regions with very short periods of daylight. How much exercise do you need to avoid feeling gloomy? Not much, it turns out. According to a study published in this month’s British Journal of Sports Medicine: a mere 20 minutes a week of any physical activity, whether sports, walking, gardening or even housecleaning, will do it... Exactly how much physical activity is required to obtain its health benefits? A review of dozens of studies on the health effects of exercise did:
ultimately reach some conclusions about how much — or, really, how little — exercise we each should be doing. That minimum amount of exercise required to see a significant lowering of your risk of dying prematurely was, they concluded, 500 MET minutes of exercise a week... A single MET, or Metabolic Equivalent of Task, is the amount of energy a person uses at rest. Two METs represent twice the energy burned at rest; four METs, four times the energy used at rest; and so on. Walking at three miles per hour is a 3.3-MET activity, while running at 6 miles an hour is a 10-MET activity. The committee concluded that a person needs to accumulate a weekly minimum of 500 MET minutes of exercise, which does not mean 500 minutes of exercise. Instead, 150 minutes a week (two and a half hours) of a moderate, three- to five-MET activity, such as walking, works out to be about 500 MET minutes. Half as much time (an hour and 15 minutes per week) spent on a 6-plus MET activity like easy jogging seems, according to the committee, to have similar health effects.

Interestingly, they did not find that exercise beyond a certain point conferred significant additional health benefits. Instead, the “dose response” for exercise, the committee found, is “curvilinear.” In other words, people who are the least active to start with get the most health benefit from starting to exercise. People who already are fit don’t necessarily get a big additional health benefit from adding more workout time to their regimens.
Of particular interest to a worrywart like myself who is extremely physically fit, but concerned about the normal loss of muscle mass that occurs in my age group (65-75) over the next ten years,  a recent study suggests that visiting the gym only once a week is sufficient to hold on to muscle mass and previous strength gains.

It has now been shown that weight resistance training, as well as aerobic exercise, boosts levels of the brain growth factor BNDF and cognitive enhancement (in rats, that is, and most likely us too.)

A further curious counterintuitive note:  Exercising increases the urge to drink alcohol,  and drinkers are more likely to exercise more.

Tuesday, January 25, 2011

Why loneliness is hazardous to your health

Greg Miller does an interesting review article in Science Magazine that describes numerous recent studies on the effect of loneliness on our physiology, health, and longevity. The article focuses on the work of John Cacioppo, a Univ. of Chicago psychologist who is credited with founding the field of social neuroscience. Here is the summary, followed by a few clips and rephrasings from the article:
In a steady stream of recent papers, social psychologists have identified several potentially unhealthy changes in the cardiovascular, immune, and nervous systems of chronically lonely people. The findings could help explain why epidemiological studies have often found that socially isolated people have shorter life spans and increased risk of a host of health problems, including infections, heart disease, and depression. The work also adds a new wrinkle, suggesting that it's the subjective experience of loneliness that's harmful, not the actual number of social contacts a person has. An impressive network of collaborations with researchers in other disciplines is now pioneering a new science of loneliness.

...scores of studies have found that people who lack social support are more prone to a variety of ailments. An analysis of 148 of these studies, published in the July 2010 issue of PLoS Medicine, suggests that social isolation increases the risk of death about as much as smoking cigarettes and more than either physical inactivity or obesity...loneliness is a health risk on its own, apart from conditions such as depression or stress that are common fellow travelers. More specifically, it seems to be the subjective experience of loneliness that's important for people's well-being rather than any objective measure of social connectivity (the number of close contacts someone has, for example).

The UCLA Loneliness Scale is based on a questionnaire that tries to size up how people perceive their social situation, with questions about how often they feel a lack of companionship, feel they have no one to talk to, or feel out of tune with those around them....When people score high on this scale, they also tend to exhibit several physiological changes that effectively put the body in a state of alert... people exhibit higher vascular resistance, a tightening of the arteries that raises blood pressure, forces the heart to work harder, and contributes to wear and tear on vessels...They have elevated molecular markers of stress - cortisol and epinephrine are elevated in saliva and urine, respectively.

Lonely people exhibit increased activity for several genes encoding signaling molecules that promote inflammation and decreased activity for genes that normally put the brakes on inflammation. They also show diminished activity in genes that help mount a defense against viral invaders....This..jibes with epidemiologic findings that socially isolated people are more susceptible to viruses, from the common cold to HIV, and to cardiovascular disease, which has been linked to excess inflammation.

Those who are lonelier, as rated by the UCLA Loneliness Scale, exhibit less activation in the ventral striatum, a component of the brain's reward circuitry, when they view pictures of smiling faces. They also show increased reactivity in the threat-detecting amygdala to pictures of angry or fearful faces.

...loneliness is partly heritable..a recent study, published in the July 2010 issue of Behavioral Genetics, used an abbreviated version of the UCLA Loneliness Scale in a survey sent to 8683 twins and family members. In this group, genetics accounted for 37% of the variability in loneliness, somewhat lower than in some previous studies. Overall, the heritability of loneliness is comparable to that of depression, but less than that of traits such as high blood pressure and cholesterol levels.

The news isn't all bad...Even for hard cases, Cacioppo believes loneliness can be overcome. He and colleagues recently conducted a meta-analysis of 20 studies on interventions for loneliness. Simply providing social support doesn't seem to work, especially if people know they're being looked after...The most effective interventions were those that borrowed methods from cognitive behavioral therapy to shift people's attention and interpretation of social situations in a more positive direction.

Monday, January 24, 2011

The chemistry of enjoying music

Robert Zatorre is an imaginative and productive musician/scientist. His website offers a cornucopia of information on the brain and music. A recent study from his group published in Nature Neuroscience observes how the 'pleasure molecule' dopamine and different brain regions change activity during the anticipation of, and emotional response to, music (many other physiological variables are also measured). His website lets you listen to the musical stimuli used in the studies, some of the most emotional chunks of music ever written. Just listening to a few of them reduced me to a puddle. (Now I know where to go for a quick emotional pleasure boost!). Jonah Lehrer's review of this work is excellent, a much better summary than I would do.  Here is Zatorre's abstract:
Music, an abstract stimulus, can arouse feelings of euphoria and craving, similar to tangible rewards that involve the striatal dopaminergic system. Using the neurochemical specificity of [11C]raclopride positron emission tomography scanning, combined with psychophysiological measures of autonomic nervous system activity, we found endogenous dopamine release in the striatum at peak emotional arousal during music listening. To examine the time course of dopamine release, we used functional magnetic resonance imaging with the same stimuli and listeners, and found a functional dissociation: the caudate was more involved during the anticipation and the nucleus accumbens was more involved during the experience of peak emotional responses to music. These results indicate that intense pleasure in response to music can lead to dopamine release in the striatal system. Notably, the anticipation of an abstract reward can result in dopamine release in an anatomical pathway distinct from that associated with the peak pleasure itself. Our results help to explain why music is of such high value across all human societies.

Friday, January 21, 2011

Age defying or denying? Real life among the old old.

Studies or ideas about human aging are frequently the subjects of posts in this blog (given that I have a vested interest in the aging of yours truly...), and I thought this perspective offered by Susan Jacoby, author of the forthcoming “Never Say Die: The Myth and Marketing of the New Old Age,” as she turns 65, was a very sane one:
..people my age and younger still pretend that old age will yield to what has long been our generational credo — that we can transform ourselves endlessly, even undo reality, if only we live right. “Age-defying” is a modifier that figures prominently in advertisements for everything from vitamins and beauty products to services for the most frail among the “old old,” as demographers classify those over 85. You haven’t experienced cognitive dissonance until you receive a brochure encouraging you to spend thousands of dollars a year for long-term care insurance as you prepare to “defy” old age....“Deny” is the word the hucksters of longevity should be using. Nearly half of the old old — the fastest-growing segment of the over-65 population — will spend some time in a nursing home before they die, as a result of mental or physical disability.

Researchers who were part of a panel discussion titled “90 Is the New 50,” presented at the World Science Festival in 2008, spoke to a middle-aged, standing-room-only audience about imminent medical miracles...The trouble with expecting 90 to become the new 50 is it can stop rational discussion — on a societal as well as individual level — about how to make 90 a better 90. This fantasy is a lot like waiting for Prince Charming, in that it doesn’t distinguish between hope and reasonable expectation.

The risk of dementia, of which Alzheimer’s is the leading cause, doubles every five years after 65...Contrary to what the baby boom generation prefers to believe, there is almost no scientifically reliable evidence that “living right” — whether that means exercising, eating a nutritious diet or continuing to work hard — significantly delays or prevents Alzheimer’s. This was the undeniable and undefiable conclusion in April of a major scientific review sponsored by the National Institutes of Health...Good health habits and strenuous intellectual effort are beneficial in themselves, but they will not protect us from a silent, genetically influenced disaster that might already be unfolding in our brains. I do not have the slightest interest in those new brain scans or spinal fluid tests that can identify early-stage Alzheimer’s. What is the point of knowing that you’re doomed if there is no effective treatment or cure? As for imminent medical miracles, the most realistic hope is that any breakthrough will benefit the children or grandchildren of my generation, not me.
Jacoby on her own situation:
I would rather share the fate of my maternal forebears — old old age with an intact mind in a ravaged body — than the fate of my other grandmother. But the cosmos is indifferent to my preferences, and it is chilling to think about becoming helpless in a society that affords only the most minimal support for those who can no longer care for themselves. So I must plan, as best I can, for the unthinkable.

..I must find someone I trust to make medical decisions for me if I cannot make them myself. This is a difficult emotional task, and it does not surprise me, for all of the public debate about end-of-life care in recent years, that only 30 percent of Americans have living wills. Even fewer have actually appointed a legal representative, known as a health care proxy, to make life-and-death decisions...I can see that the “90 is the new 50” crowd might object to my thinking more about worst-case scenarios than best-case ones. But if the best-case scenario emerges and I become one of those exceptional “ageless” old people so lauded by the media, I won’t have a problem. I can also take it if fate hands me a passionate late-in-life love affair, a financial bonanza or the energy to write more books in the next 25 years than I have in the past 25.

Thursday, January 20, 2011

Distortions of mind perception in psychopathology

A really interesting perspective comes from Daniel Wegner and co-workers in the Psychology Dept. at Harvard. (Wegner's book, "The Illusion of Conscious Will" is in my cannon of must-read books on self and consciousness). From their introduction:
Originally thought to proceed along a single dimension, mind perception has been revealed in a factor-analytic study to occur along independent dimensions of experience (e.g., the capacity for pleasure, fear, hunger) and agency (e.g., the capacity for self-control, planning, memory. Adult humans are typically seen as capable of both experience and agency, whereas children and animals are seen as capable of mainly experience. Gods and robots are seen as capable of mainly agency, and the dead are seen as capable of neither.

We suggest that a number of disorders may be characterized as specific distortions of mind perception, atypical ascriptions of mental capacities to other entities....Successful interaction with the world requires knowing which entities have minds and which do not. Mind perception can therefore be distorted by overperception (perceiving a nonexistent mind) and underperception (failing to perceive an existent mind). Research suggests that both can be associated with adverse consequences for perceivers and targets, consequences that range from social faux pas to violence and death. For example, the overperception of mind in infants can lead to child abuse, but the underperception of mind in adults can lead to the denial of moral rights.
Their abstract:
It has long been known that psychopathology can influence social perception, but a 2D framework of mind perception provides the opportunity for an integrative understanding of some disorders. We examined the covariation of mind perception with three subclinical syndromes—autism-spectrum disorder, schizotypy, and psychopathy—and found that each presents a unique mind-perception profile. Autism-spectrum disorder involves reduced perception of agency in adult humans. Schizotypy involves increased perception of both agency and experience in entities generally thought to lack minds. Psychopathy involves reduced perception of experience in adult humans, children, and animals. Disorders are differentially linked with the over- or underperception of agency and experience in a way that helps explain their real-world consequences.

Smart Phone people...

A bit of humour from the TalkAndroid website (click on image to enlarge and read the captions):