Monday, November 10, 2008

Chill-out...

From the "Random Samples" section of the Oct. 31 Science Magazine:
Psychologists at the University of Hertfordshire, U.K., last week unveiled what they are billing as "the world's most relaxing room." The 160-square-meter space, bathed in green lights with an artificially lit blue sky, is furnished with soft mats and lavender-scented pillows "to create a relaxing environment with no sense of threat," explains the project's mastermind Richard Wiseman.

The design is based on research on the effects of light, scent, and music in relaxation. "Cold colors such as blue and green tend to be perceived as calming, whereas warm colors can be perceived as arousing," explains Birgitta Gatersleben, an environmental psychologist from the University of Surrey in Guildford, U.K. Lavender is said to reduce anxiety and induce sleep by lowering the levels of the stress hormone cortisol. The room also features specially composed music with a slow, steady beat and low-frequency tones.

So far, the room's 200 visitors have given it mixed reviews. "Some people absolutely love it and can't have enough of it," Wiseman says. "But people who thrive on and need stress to work absolutely hate it."

The project was designed to be easy to replicate in offices and other real-life environments. "I would like to see relaxation rooms in public spaces," Wiseman adds. "If we pay 20p to use a toilet in King's Cross train station, why not pay for 20 minutes of peace?"

Symbolic markers, cultural groups, and ingroup favoritism

A 'This week in Science' section of Science Magazine points to the work of Efferson et al. :
In human social interactions, it is not uncommon to draw inferences about hidden characteristics (attitudes, beliefs, or behavioral norms) on the basis of observable markers that may bear no fundamental connection to the underlying quantity but have become associated with specific groups over time. For instance, individuals sporting insignia of the Boston Red Sox or Manchester United may be classified as friends (or foes, if one should happen to be a New York Yankees or Chelsea fan). Although much research has been devoted to how a member of one cultural or ethnic group views other in-group and out-group members, less is known about the process by which symbolic markers come to be used as signals to define group membership. Efferson et al. have designed a laboratory-based economic game in which subjects were free to associate arbitrary markers with varying payoffs. Cultural groups (those in which members had adopted the same marker) and consequent ingroup favoritism developed only when the marker was both predictive of behavior in the game as well as changeable over time.
The abstract from Efferson et al.:
Cultural boundaries have often been the basis for discrimination, nationalism, religious wars, and genocide. Little is known, however, about how cultural groups form or the evolutionary forces behind group affiliation and ingroup favoritism. Hence, we examine these forces experimentally and show that arbitrary symbolic markers, though initially meaningless, evolve to play a key role in cultural group formation and ingroup favoritism because they enable a population of heterogeneous individuals to solve important coordination problems. This process requires that individuals differ in some critical but unobservable way and that their markers be freely and flexibly chosen. If these conditions are met, markers become accurate predictors of behavior. The resulting social environment includes strong incentives to bias interactions toward others with the same marker, and subjects accordingly show strong ingroup favoritism. When markers do not acquire meaning as accurate predictors of behavior, players show a markedly reduced taste for ingroup favoritism. Our results support the prominent evolutionary hypothesis that cultural processes can reshape the selective pressures facing individuals and so favor the evolution of behavioral traits not previously advantaged.

Friday, November 07, 2008

Values of early music training.

Forgeard et al. show that children who receive at least three years of instrumental music training outperform their control counterparts on auditory discrimination abilities and fine motor skills, as well as vocabulary and nonverbal reasoning skills. They did not confirm a suggestion from previous research that music training enhances heightened spatial skills, phonemic awareness, or mathematical abilities.

The creationists go to war over the brain.

A Zoology colleague of mine pointed out an article by Amander Gefter (PDF here) in The New Scientist on a group of "non-materialist neuroscientists" that is trying to resurrect Cartesian Dualism. It is particularly sad that one of these is Jeffrey Schwartz, who has done classic work showing how cognitive therapy can amelioate obsessive compulsive disorder.
Schwartz used scanning technology to look at the neural patterns thought to be responsible for OCD. Then he had patients use "mindful attention" to actively change their thought processes, and this showed up in the brain scans: patients could alter their patterns of neural firing at will. From such experiments, Schwartz and others argue that since the mind can change the brain, the mind must be something other than the brain, something non-material. In fact, these experiments are entirely consistent with mainstream neurology - the material brain is changing the material brain.

Clearly, while there is a genuine attempt to appropriate neuroscience, it will not influence US laws or education in the way that anti-evolution campaigns can because neuroscience is not taught as part of the core curriculum in state-funded schools. But as Andy Clark, professor of logic and metaphysics at the University of Edinburgh, UK, emphasises: "This is real and dangerous and coming our way." He and others worry because scientists have yet to crack the great mystery of how consciousness could emerge from firing neurons. "Progress in science is slow on many fronts," says John Searle, a philosopher at the University of California, Berkeley. "We don't yet have a cure for cancer, but that doesn't mean cancer has spiritual causes." And for Patricia Churchland, a philosopher of neuroscience at the University of California, San Diego, "it is an argument from ignorance. The fact something isn't currently explained doesn't mean it will never be explained or that we need to completely change not only our neuroscience but our physics." The attack on materialism proposes to do just that, but it all turns on definitions. "At one time it looked like all physical causation was push/pull Newtonianism," says Owen Flanagan, professor of philosophy and neurobiology at Duke University, North Carolina. "Now we have a new understanding of physics. What counts as material has changed. Some respectable philosophers think that we might have to posit sentience as a fundamental force of nature or use quantum gravity to understand consciousness. These stretch beyond the bounds of what we today call 'material', and we haven't discovered everything about nature yet. But what we do discover will be natural, not supernatural."

And as Clark observes: "This is an especially nasty mind-virus because it piggybacks on some otherwise reasonable thoughts and worries. Proponents make such potentially reasonable points as 'Oh look, we can change our brains just by changing our minds,' but then leap to the claim that mind must be distinct and not materially based. That doesn't follow at all. There's nothing odd about minds changing brains if mental states are brain states: that's just brains changing brains."

Thursday, November 06, 2008

A conference on "Happiness and its Causes."

There is a "Happiness and its Causes" conference in San Francisco Nov. 24-25, with very flashy marketing and quite a cast of stars: Paul Ekman, Robert Sapolsky, Anne Harrington and others.

Our brain's large scale functional architecture.

He et al. have published an important study that shows correlations between spontaneous fluctuations in slow cortical potentials recorded by electrocorticography and fMRI BOLD signals are maintained across wakefulness, slow-wave sleep, and rapid-eye-movement sleep. Balduzzi et al. note in their review of the work that the study
... makes it clear that both BOLD and ECoG fluctuations display a pattern of regional correlations, or functional connectivity, which closely reflects those regions' anatomical connectivity. Inverting a well known adagio, what wires together, fires together. Indeed, it seems that it could not be otherwise. If neurons are connected in a certain way, and if they are spontaneously active, functional connectivity is bound to reflect anatomical connectivity, just like traffic patterns must reflect the underlying roadmap.
The reviewers also give a nice description of alternative ideas about what the brain's spontaneous or background activity might be for:
The steady depolarization and firing of neurons, even when the brain is supposedly “at rest,” also called the “default mode” of activity, consumes approximately two-thirds of the brain's already disproportionate energy budget, so it better do something useful. For instance, spontaneous activity may be important for the brain's trillions of synapses, perhaps by keeping them exercised or consolidating and renormalizing their strength. Another notion is that spontaneous activity may be necessary to maintain a fluid state of readiness that allows the cortex to rapidly enter any of a number of available states or firing patterns—a kind of metastability. Theoretical work suggests that the repertoire of available states is maximal under moderate spontaneous activity, and shrinks dramatically with either complete inactivity or hyperactivity. But what kind of neural states? One possibility is that the cortex is like a sea undulating gently, and that evoked or task-related responses would be like small ripples on its surface. This possibility is consistent with fMRI studies, because spontaneous slow fluctuations in BOLD are as large or larger than those evoked by stimuli. Also, it would fit nicely with the trial-to-trial variability of behavioral responses. Another possibility is that there are distinct modes of neuronal activity, such as a READY mode and a GO mode (and possibly an inhibited, STOP mode). Spontaneous activity would then be the READY mode of neuronal firing signaling the absence of preferred stimuli (an ongoing, low-level buzz). By contrast, in the GO mode, neurons, or local populations of neurons, would signal the presence of a preferred stimulus by firing at much higher rates for short periods of time (a brief and loud shout). Unit recording studies have provided plenty of evidence that neurons respond strongly and distinctly to specific stimuli. In this view, the cortex would be more like a sea pierced by sharp islands. On the other hand, the slow hemodynamic response function underlying the BOLD signal may make fMRI partly blind to the distinction between slow, low-amplitude fluctuations in firing and fast, high-amplitude bursts of activity. If there are two modes of neural activity, it bears keeping in mind that neurons in the READY mode would be as necessary as neurons in the GO mode in specifying different cognitive states, just as the background is as necessary as the foreground.

The belt of enlightenment...

A map, via Andrew Sullivan's blog, of the counties, mainly in Arkansas through Appalachia, in which more people voted republican in 2008 than in 2004. Overlaps with maps of ignorance, racism and poverty. Not many MindBlog readers in this region!

Could you figure this out?

Get the peanut out of the tube:

Wednesday, November 05, 2008

Language conflict in the bilingual brain

An open access article by Van Heuven et al. in Cerebral Cortex:
The large majority of humankind is more or less fluent in 2 or even more languages. This raises the fundamental question how the language network in the brain is organized such that the correct target language is selected at a particular occasion. Here we present behavioral and functional magnetic resonance imaging data showing that bilingual processing leads to language conflict in the bilingual brain even when the bilinguals’ task only required target language knowledge. This finding demonstrates that the bilingual brain cannot avoid language conflict, because words from the target and nontarget languages become automatically activated during reading. Importantly, stimulus-based language conflict was found in brain regions in the LIPC associated with phonological and semantic processing, whereas response-based language conflict was only found in the pre-supplementary motor area/anterior cingulate cortex when language conflict leads to response conflicts.

Disputed definitions: paradigm shift

NatureNews has an interesting article on words whose definitions get scientists most worked up. Take 'paradigm shift,' for example:
Paradigm shift has a definite origin and originator: Thomas Kuhn, writing in his 1962 book The Structure of Scientific Revolutions, argued against the then prevalent view of science as an incremental endeavour marching ever truthwards. Instead, said Kuhn, most science is "normal science", which fills in the details of a generally accepted, shared conceptual framework. Troublesome anomalies build up, however, and eventually some new science comes along and overturns the previous consensus. Voilà, a paradigm shift. The classic example, Kuhn said, is the Copernican revolution, in which Ptolemaic theory was swept away by putting the Sun at the centre of the Solar System. Post-shift, all previous observations had to be reinterpreted.

Kuhn's theory about how science works was arguably a paradigm shift of its own, by changing the way that academics think about science. And scientists have been using the phrase ever since.

In a postscript to the second edition of his book, Kuhn explained that he used the word 'paradigm' in at least two ways (noting that one "sympathetic reader" had found 22 uses of the term). In its broad form, it encompasses the "entire constellation of beliefs, values, techniques and so on shared by the members of a given community". More specifically it refers to "the concrete puzzle-solutions" that are used as models for normal science post-shift.

Scientists who use the term today don't usually mean that their field has undergone a Copernican-scale revolution, to the undying annoyance of many who hew to Kuhn's narrower definition. But their usage might qualify under his broader one. And so usage becomes a matter of opinion and, perhaps, vanity.

The use of the term in titles and abstracts of leading journals jumped from 30 papers in 1991 to 124 in 1998, yet very few of these papers garnered more than 10 citations apiece1. Several scientists contacted for this article who had used paradigm shift said that, in retrospect, they were having second thoughts. In 2002, Stuart Calderwood, an oncologist at Harvard Medical School in Boston, Massachusetts, used it to describe the discovery that 'heat shock proteins', crucial to cell survival, could work outside the cell as well as in2. "If you work in a field for a long time and everything changes, it does seem like a revolution," he says. But now he says he may have misused the phrase because the discovery was adding to, rather than overturning, previous knowledge in the field.

Arvid Carlsson, of the University of Gothenburg in Sweden stands by his use of the phrase. "Until a certain time, the paradigm was that cells communicate almost entirely by electrical signals," says Carlsson. "In the 1960s and '70s, this changed. They do so predominantly by chemical signals. In my opinion, this is dramatic enough to deserve the term paradigm shift." Few would disagree: base assumptions were overturned in this case, and Carlsson's own work on the chemical neurotransmitter dopamine (which was instrumental in this particular shift) earned him the 2000 Nobel Prize in Physiology or Medicine

Unless a Nobel prize is in the offing, it might be wise for scientists to adopt the caution of contemporary historians of science and think twice before using a phrase with a complex meaning and a whiff of self promotion. "Scientists all want to be the scientists that generate a new revolution," says Kuhn's biographer, Alexander Bird, a philosopher at the University of Bristol, UK. "But if Kuhn is right, most science is normal science and most people can't perform that role."

Tuesday, November 04, 2008

A break from watching the election returns....

The psychology of voting - US readers: VOTE TODAY!

An interesting article in today's New York Times Science section by Benedict Carey, on the value of voting beyond politics.

Neuroeconomics - the neural circuitry of overbidding

Delgado et al. offer an interesting article in Science Magazine suggesting that fear of loss may mediate overbidding in auctions. Here is their abstract, followed by comments in a review by Maskin.
We take advantage of our knowledge of the neural circuitry of reward to investigate a puzzling economic phenomenon: Why do people overbid in auctions? Using functional magnetic resonance imaging (fMRI), we observed that the social competition inherent in an auction results in a more pronounced blood oxygen level–dependent (BOLD) response to loss in the striatum, with greater overbidding correlated with the magnitude of this response. Leveraging these neuroimaging results, we design a behavioral experiment that demonstrates that framing an experimental auction to emphasize loss increases overbidding. These results highlight a role for the contemplation of loss in understanding the tendency to bid "too high." Current economic theories suggest overbidding may result from either "joy of winning" or risk aversion. By combining neuroeconomic and behavioral economic techniques, we find that another factor, namely loss contemplation in a social context, may mediate overbidding in auctions.
Maskin's comments are based on followup experiments not mentioned in the abstract:
The fMRI data show that subjects experience a lower blood oxygen level in the striatum in response to losing an auction, but no significant change in reaction to winning one. The authors interpret this result as suggesting that subjects experience "fear of losing" and that this fear accounts for their overbidding. But actually modeling fear explicitly--making it precise--does not seem straightforward.

A natural modeling device would be simply to subtract something from the subject's payoff when she loses. However, such a modification would not accord with the authors' findings in their subsequent experiment. In the follow-up, there were two treatments: one in which a subject is initially given a bonus sum of money S but told that she has to return it if she loses the auction; the other in which the subject is promised that if she wins she will get S. The two treatments are, ex post, identical: In both cases, the subject ends up with the bonus if and only if she wins. However, in practice, subjects bid more in the former treatment than the latter. Such behavior sharply contradicts the "payment subtraction" hypothesis, under which behavior in the two treatments would be the same. Moreover, it seems difficult to find a natural alternative formulation of the "fear of losing" idea that explains the results simultaneously from both Delgado et al. experiments. Even so, there is a well-known principle that could account for the behavioral discrepancy between the two treatments in the follow-up experiment: the "endowment" effect. When a subject is given a bonus S at the outset, she may become possessive and so move more aggressively to retain it than she would act to obtain a contingent bonus at the end of the experiment.

As for why subjects overbid, perhaps the answer is that high-bid auctions are just too complex for a typical buyer to analyze completely systematically. The buyer will easily see that she has to shade her bid (bid strictly below v) to get a positive payoff. Still, she won't want to shade too much because shading reduces her probability of winning. A simple rule of thumb would be to shade just a little. But this leads immediately to overbidding, because risk-neutral equilibrium bidding entails a great deal of shading: A buyer will bid only one-half her valuation.

The behavioral revolution in economics

In light of the work by Delgado et al. mentioned in today's other posting, I thought it appropriate to pass on this NY Times Op-Ed piece by David Brooks, on the decline of economic models that presume that people are mostly engaged in rationally calculating and maximizing their self-interest. Rather, people put great energy in perceiving things that aren't true. Brooks emphasizes the work of Taleb:
Taleb believes that our brains evolved to suit a world much simpler than the one we now face. His writing is idiosyncratic, but he does touch on many of the perceptual biases that distort our thinking: our tendency to see data that confirm our prejudices more vividly than data that contradict them; our tendency to overvalue recent events when anticipating future possibilities; our tendency to spin concurring facts into a single causal narrative; our tendency to applaud our own supposed skill in circumstances when we’ve actually benefited from dumb luck.

And looking at the financial crisis, it is easy to see dozens of errors of perception. Traders misperceived the possibility of rare events. They got caught in social contagions and reinforced each other’s risk assessments. They failed to perceive how tightly linked global networks can transform small events into big disasters.

Monday, November 03, 2008

Explanatory Neurophilia - seduction without cause

Trout discusses the work of Weinberg (mentioned in a previous post) who has shown in a series of experiment how non-expert consumers of behavioral explanations assign greater standing to explanations that contain neuroscientific details, even if these details provide no additional explanatory value. He discusses the part that this ‘placebic’ information might play in producing a potentially misleading sense of intellectual fluency and, consequently, an unreliable sense of understanding. I'm passing the article on to you here.

Happiness, Eudaimonia, etc.

As I was doing some homework in preparation for a gig as a talking heading ‘expert’ on a web radio show called “Make Me Happy” I ran across this 2002 article, “Pleasure, Meaning & Eudaimonia” by Martin Seligman.
So the core thesis in Authentic Happiness is that there are three very different routes to happiness. First the Pleasant Life, consisting in having as many pleasures as possible and having the skills to amplify the pleasures. This is, of course, the only true kind of happiness on the Hollywood view. Second, the Good Life, which consists in knowing what your signature strengths are, and then recrafting your work, love, friendship, leisure and parenting to use those strengths to have more flow in life. Third, the Meaningful Life, which consists of using your signature strengths in the service of something that you believe is larger than you are.

Sunday, November 02, 2008

MindBlog as election worker...

I got a call from the Obama campaign asking if I would work as a poll watcher at an election polling site on Tuesday. I said yes, went to the training session yesterday (shown at left), and came away in awe of the micro-analytical detail and power of the Obama ground effort. Our main job is to be sure people make it through the long waiting lines, give out registration information and assistance, be alert for any signs of voter intimidation. Another class of poll worker, designated a "Houdini", records the identity of every voter, and at 30 min- 1hr intervals calls in the information to command central, which takes those names off the canvass and phone lists to reallocate resources in real time towards getting out the people who haven't voted.

What is the difference between a 40 and a 70 year old brain...

A piece by Dr. Robert Epstein relevant to the current presidential election.

Friday, October 31, 2008

A question to readers... MindBlog podcasts?

Blog reader Patrice responded to the previous posting in a comment asking about the possibly of my doing podcasts on some mindblog topics. Another friend has mentioned this, yet I haven't thought much about it because I've never really gotten into listening to podcasts myself. When I am driving I usually listen to music. However, I am an experienced university lecturer, and actually sort of miss the more spontaneous and improvisational energy of talking versus writing. So, I'm curious to know how many of you might actually find occasional podcasts useful. And if so, do you have opinions about optimal length and subjects that are most interesting to you? I would appreciate responses, either to this posting, or by email to mdbownds@wisc.edu.

Fan Mail

Although I feel totally immodest about doing this, I have to pass on this text of a email today from a reader of this blog in Tasmania, Australia. Whenever I question how worthwhile this effort is, a statement like this really warms me up:
Just a quick appreciation: I love your blog. I don’t know how you do it but just about every day there’s something that I really want to read. You appear to write a blog that is bang on some of my interests, in particular, building the full - crazy, intelligent, imaginative, smart, musical, superstitious, religious, even - human from biology. We appear to be living in a very exciting time where a lot of stuff that was in what I’d call the realm of superstition – is being dusted off and actually inspected. Maybe it’ll go on like this forever, but I can’t imagine that we won’t run out of things pretty soon at the current rate.

Psychology of the electorate

I point to two interesting articles in the New York Times. Dewan and Brown discuss how the work of psychologist Drew Westen (mentioned in three - previous - posts) has shaped the message delivery of democratic candidates so thoroughly that the rhetorical dominance enjoyed by republican for years has been completely reversed. Kristof deals again with the issue of unconscious bias in voters, and links in his article direct you to online tests for your own unconscious biases.

Emergent properties of human groups

Just as ants interact to form elaborate colonies and neurons interact to create structured thought, groups of people interact to create emergent organizations that the individuals may not understand or even perceive. I recently went to an interesting seminar on this topic in the Psychology Department at Univ. Wisc. given by Robert Goldstone at Indiana Univ. He has set up an interesting web based experiment to test agent based models of emergent properties of human groups. You can read about and also become a subject in the experiments here.

Thursday, October 30, 2008

Religion: Bound to believe?

I pass on clips from an article by Pascal Boyer that explains why a slew of cognitive traits shared by humans will always make atheism a hard sell.
In the past ten years, the evolutionary and cognitive study of religion has begun to mature. It does not try to identify the gene or genes for religious thinking. Nor does it simply dream up evolutionary scenarios that might have led to religion as we know it. It does much better than that. It puts forward new hypotheses and testable predictions. It asks what in the human make-up renders religion possible and successful. Religious thought and behaviour can be considered part of the natural human capacities, such as music, political systems, family relations or ethnic coalitions. Findings from cognitive psychology, neuroscience, cultural anthropology and archaeology promise to change our view of religion.

Unlike other social animals, humans are very good at establishing and maintaining relations with agents beyond their physical presence; social hierarchies and coalitions, for instance, include temporarily absent members. This goes even further. From childhood, humans form enduring, stable and important social relationships with fictional characters, imaginary friends, deceased relatives, unseen heroes and fantasized mates. Indeed, the extraordinary social skills of humans, compared with other primates, may be honed by constant practice with imagined or absent partners.

It is a small step from having this capacity to bond with non-physical agents to conceptualizing spirits, dead ancestors and gods, who are neither visible nor tangible, yet are socially involved. This may explain why, in most cultures, at least some of the superhuman agents that people believe in have moral concerns.

In addition, the neurophysiology of compulsive behaviour in humans and other animals is beginning to shed light on religious rituals. These behaviours include stereotyped, highly repetitive actions that participants feel they must do, even though most have no clear, observable results, such as striking the chest three times while repeating a set formula. Ritualized behaviour is also seen in patients with obsessive-compulsive disorders and in the routines of young children. In these contexts, rituals are generally associated with thoughts about pollution and purification, danger and protection, the required use of particular colours or numbers or the need to construct a safe and ordered environment.

So is religion an adaptation or a by-product of our evolution? Perhaps one day we will find compelling evidence that a capacity for religious thoughts, rather than 'religion' in the modern form of socio-political institutions, contributed to fitness in ancestral times. For the time being, the data support a more modest conclusion: religious thoughts seem to be an emergent property of our standard cognitive capacities.

Religious concepts and activities hijack our cognitive resources, as do music, visual art, cuisine, politics, economic institutions and fashion. This hijacking occurs simply because religion provides some form of what psychologists would call super stimuli. Just as visual art is more symmetrical and its colours more saturated than what is generally found in nature, religious agents are highly simplified versions of absent human agents, and religious rituals are highly stylized versions of precautionary procedures. Hijacking also occurs because religions facilitate the expression of certain behaviours. This is the case for commitment to a group, which is made all the more credible when it is phrased as the acceptance of bizarre or non-obvious beliefs.

Some form of religious thinking seems to be the path of least resistance for our cognitive systems. By contrast, disbelief is generally the result of deliberate, effortful work against our natural cognitive dispositions — hardly the easiest ideology to propagate.

Fair and Balanced - measuring media bias

One hears charges from both left and right about media bias, with FOX News frequently cited as the most extreme case. Tim Groeling has done interesting work to objectively measure the bias shown by television media, in a paper (PDF here) to be published in the December issue of Presidential Studies Quarterly. He simply collected the in-house presidential approval polling by ABC, CBS, NBC and FOX News and compared these with the actual broadcasts of such polls on evening news shows from 1997 to early 2008. As an example, CBS was 35 percent less likely to report a five-point drop in approval for Bill Clinton than a similar rise in approval and was 33 percent more likely to report a five-point drop than a rise for George W. Bush. FOX News was 67 percent less likely to report a rise in approval for Clinton than a decrease and 36 percent more likely to report an increase rather than a decrease for Bush.

Wednesday, October 29, 2008

Watching the amygdala signal the good and the bad

An interesting piece in Jour. of Neuoscience by Belova et al., in which recordings from single cells in the monkey amygdala indicate their division into those that track either positive state or negative states:
As an organism interacts with the world, how good or bad things are at the moment, the value of the current state of the organism, is an important parameter that is likely to be encoded in the brain. As the environment changes and new stimuli appear, estimates of state value must be updated to support appropriate responses and learning. Indeed, many models of reinforcement learning posit representations of state value. We examined how the brain mediates this process by recording amygdala neural activity while monkeys performed a trace-conditioning task requiring fixation. The presentation of different stimuli induced state transitions; these stimuli included unconditioned stimuli (USs) (liquid rewards and aversive air puffs), newly learned reinforcement-predictive visual stimuli [conditioned stimuli (CSs)], and familiar stimuli long associated with reinforcement [fixation point (FP)]. The FP had a positive value to monkeys, because they chose to foveate it to initiate trials. Different populations of amygdala neurons tracked the positive or negative value of the current state, regardless of whether state transitions were caused by the FP, CSs, or USs. Positive value-coding neurons increased their firing during the fixation interval and fired more strongly after rewarded CSs and rewards than after punished CSs and air puffs. Negative value-coding neurons did the opposite, decreasing their firing during the fixation interval and firing more strongly after punished CSs and air puffs than after rewarded CSs and rewards. This representation of state value could underlie how the amygdala helps coordinate cognitive, emotional, and behavioral responses depending on the value of one's state.

Happiness and the 'prosperoscope'

The random samples section of Science Magazine discusses the latest prosperity or 'happiness index' report of the Legatum Institute:
The results are in: Australia is the most prosperous country in the world; Yemen drags at the bottom of the list. But it's not just wealth that makes a country prosperous, according to the 2008 prosperity index, also known as the "happiness index," published last week by the Legatum Institute (LI) in Dubai (see www.prosperity.com). The institute based its rankings on surveys of economic competitiveness and comparative livability from 140 countries, including factors such as capital investment and the degree of social equality.

This year, for the first time, countries' environmental efforts counted toward their scores, says LI Senior Vice President William Inboden. The institute selected an objective measurement--the ratio of developed land to land remaining in its natural state in each country--and added questions about how respondents felt about their country's environmental policies. Depending on a country's wealth, the environmental measures could count for as much as 4% of a country's prosperity score. Although Australia was the most prosperous country overall, New Zealand topped the environmental measures. The most environmentally unhappy people were Ukrainians, who particularly dislike their air quality.
You might enjoy playing with the 'prosperiscope' on the site.
"The Legatum Prosperiscope™ is a powerful interactive online tool that allows users to customise analyses across 104 countries using 22 different factors. Users can also compare countries against each other to identify their relative strengths and weaknesses. Follow the simple steps on the right to start using the Prosperiscope."

Tuesday, October 28, 2008

Instinctual math

Natalie Angier does an interesting article on our instinctual intuitive math versus our more analytical learned number crunching, and the relationship between them. Apparently our evolutionarily endowed sense of approximation is related to how good we are at formal math. The article contains a neat interactive demonstration of our non verbal intuitive math abilities.
Brain imaging studies have traced the approximate number sense to a specific neural structure called the intraparietal sulcus, which also helps assess features like an object’s magnitude and distance. Symbolic math, by contrast, operates along a more widely distributed circuitry, activating many of the prefrontal regions of the brain that we associate with being human. Somewhere, local and global must be hooked up to a party line.

...open questions include how malleable our inborn number sense may be, whether it can be improved with training, and whether those improvements would pay off in a greater appetite and aptitude for math. If children start training with the flashing dot game at age 4, will they be supernumerate by middle school?

Views of the presidential candidates on science

Here is the PDF from Science Magazine.

Monday, October 27, 2008

Poulenc Novelette in C major for a Monday morning

Recorded 10/25/08 on the Steinway B in my home on Twin Valley Road in Middleton, Wisconsin.

Experiencing physical warmth promotes interpersonal warmth

Here is the abstract of the article by Williams and Bargh which has the title of this post:
"Warmth" is the most powerful personality trait in social judgment, and attachment theorists have stressed the importance of warm physical contact with caregivers during infancy for healthy relationships in adulthood. Intriguingly, recent research in humans points to the involvement of the insula in the processing of both physical temperature and interpersonal warmth (trust) information. Accordingly, we hypothesized that experiences of physical warmth (or coldness) would increase feelings of interpersonal warmth (or coldness), without the person's awareness of this influence. In study 1, participants who briefly held a cup of hot (versus iced) coffee judged a target person as having a "warmer" personality (generous, caring); in study 2, participants holding a hot (versus cold) therapeutic pad were more likely to choose a gift for a friend instead of for themselves.

Embodied cognition, a cold stare makes you feel cold.

I show below the abstract from Zhong and Leonardelli, followed by more explanation from an article by Benedict Carey.
Metaphors such as icy stare depict social exclusion using cold-related concepts; they are not to be taken literally and certainly do not imply reduced temperature. Two experiments, however, revealed that social exclusion literally feels cold. Experiment 1 found that participants who recalled a social exclusion experience gave lower estimates of room temperature than did participants who recalled an inclusion experience. In Experiment 2, social exclusion was directly induced through an on-line virtual interaction, and participants who were excluded reported greater desire for warm food and drink than did participants who were included. These findings are consistent with the embodied view of cognition and support the notion that social perception involves physical and perceptual content. The psychological experience of coldness not only aids understanding of social interaction, but also is an integral part of the experience of social exclusion.
More detail on the two experiments:
In one, they split 65 students into two groups, instructing those in one to recall a time when they felt socially rejected, and those in the other to summon a memory of social acceptance.

Many of the students were recent immigrants and had fresh memories of being isolated in the dorms, left behind while roommates went out, Dr. Zhong said.

The researchers then had each of the participants estimate the temperature in the lab room. The students who had recalled being excluded estimated the temperature to be, on average, 5 degrees Fahrenheit lower than the others.

In the second experiment, the researchers had 52 students come into the lab and play a computer game, one at a time. The students “threw” a ball back and forth with three other figures on the computer screen that — so the participants thought — represented other students playing from remote locations.

In fact a computer program was running the game, and it excluded half the study participants, throwing them the virtual ball a couple of times in the beginning, then ignoring them altogether. The other group of students in the study were included in the virtual game of catch.

After playing the game, the participants in this study then rated their preferences for a variety of foods and drinks, including hot soup, coffee, an apple and crackers. Those who had been isolated in the computer game showed a strong preference for the soup and coffee over the other items; the included students had no such preference.

Sunday, October 26, 2008

Drinking alone, with "friends" on Facebook

I'm struck by how very little I know about the ~10,000 visitors that Quantcast tells me come to Deric's MindBlog each month. Thus this essay in the "Lives" section of today's New York Times Magazine struck a chord with me.

Friday, October 24, 2008

Distortion of science by headline-grabbing

I'm always being concerned in scanning journals for possible postings in this blog that my eye is too readily caught by the flashy marketing phrase or popularizing twist, thus neglecting more boring, but possibly much more significant, work. An article in The Economist reinforces my concern by pointing to work of Young Ioannidis, and Al-Ubaydli in PLos Medicine. that uses economic commodity theory to show how the current scientific publishing system is biased towards favoring trumpeted results that are also more likely to be false.
In economic theory the winner’s curse refers to the idea that someone who places the winning bid in an auction may have paid too much. Consider, for example, bids to develop an oil field. Most of the offers are likely to cluster around the true value of the resource, so the highest bidder probably paid too much.

The same thing may be happening in scientific publishing, according to a new analysis. With so many scientific papers chasing so few pages in the most prestigious journals, the winners could be the ones most likely to oversell themselves—to trumpet dramatic or important results that later turn out to be false. This would produce a distorted picture of scientific knowledge, with less dramatic (but more accurate) results either relegated to obscure journals or left unpublished.

It starts with the nuts and bolts of scientific publishing. Hundreds of thousands of scientific researchers are hired, promoted and funded according not only to how much work they produce, but also to where it gets published. For many, the ultimate accolade is to appear in a journal like Nature or Science. Such publications boast that they are very selective, turning down the vast majority of papers that are submitted to them.

The assumption is that, as a result, such journals publish only the best scientific work. But Dr Ioannidis and his colleagues argue that the reputations of the journals are pumped up by an artificial scarcity of the kind that keeps diamonds expensive. And such a scarcity, they suggest, can make it more likely that the leading journals will publish dramatic, but what may ultimately turn out to be incorrect, research.

How we know our own minds...

For those of you who might be more heavily into philosophy of mind and introspective psychology, I want to pass along this PDF of a draft article that is to appear in Brain and Behavioral Sciences, "How we know our own minds: the relationship between mindreading and metacognition." Carruthers defends the idea that our knowledge of our own attitudes results from turning our mindreading capacities upon ourselves, not from introspection for propositional attitudes. Here is his abstract, showing the organization of his arguments:
Four different accounts of the relationship between third-person mindreading and first-person metacognition are compared and evaluated. While three of them endorse the existence of introspection for propositional attitudes, the fourth (defended here) claims that our knowledge of our own attitudes results from turning our mindreading capacities upon ourselves. Section 1 introduces the four accounts. Section 2 develops the “mindreading is prior” model in more detail, showing how it predicts introspection for perceptual and quasi-perceptual (e.g. imagistic) mental events while claiming that metacognitive access to our own attitudes always results from swift unconscious self-interpretation. It also considers the model’s relationship to the expression of attitudes in speech. Section 3 argues that the commonsense belief in the existence of introspection should be given no weight. Section 4 argues briefly that data from childhood development are of no help in resolving this debate. Section 5 considers the evolutionary claims to which the different accounts are committed, and argues that the three introspective views make predictions that aren’t borne out by the data. Section 6 examines the extensive evidence that people often confabulate when self-attributing attitudes. Section 7 considers “two systems” accounts of human thinking and reasoning, arguing that although there are inrospectable events within System 2, there are no introspectable attitudes. Section 8 examines alleged evidence of “unsymbolized thinking”. Section 9 considers the claim that schizophrenia exhibits a dissociation between mindreading and metacognition. Finally, Section 10 evaluates the claim that autism presents a dissociation in the opposite direction, of metacognition without mindreading.

Thursday, October 23, 2008

Different aspects of human intelligence correlate with cortical thickness versus neural activation.

Choi et al. report an interesting study in the Journal of Neuroscience. I'm passing on the abstract and a bit of explanation of crystallized versus fluid intelligence, but not the usual flashy fMRI graphics:
We hypothesized that individual differences in intelligence (Spearman's g) are supported by multiple brain regions, and in particular that fluid (gF) and crystallized (gC) components of intelligence are related to brain function and structure with a distinct profile of association across brain regions. In 225 healthy young adults scanned with structural and functional magnetic resonance imaging sequences, regions of interest (ROIs) were defined on the basis of a correlation between g and either brain structure or brain function. In these ROIs, gC was more strongly related to structure (cortical thickness) than function, whereas gF was more strongly related to function (blood oxygenation level-dependent signal during reasoning) than structure. We further validated this finding by generating a neurometric prediction model of intelligence quotient (IQ) that explained 50% of variance in IQ in an independent sample. The data compel a nuanced view of the neurobiology of intelligence, providing the most persuasive evidence to date for theories emphasizing multiple distributed brain regions differing in function.
As background:
gC, sometimes described as verbal ability, is more dependent on accumulated knowledge in long-term storage, including semantic memory. gF refers to reasoning ability, and is known to depend on working memory. Although gC and gF are typically correlated and can be considered subfactors of g (Jensen), they are conceptually and empirically separable. For instance, gC continues to increase over the lifespan, but gF peaks in early adulthood and then declines . Furthermore, at the neural level, lesion studies demonstrated that patients with anterior temporal damages perform poorly on tests of semantic knowledge, whereas prefrontal patients typically show profound deficits in solving diverse reasoning tasks.

Sleep accelerates improvement in working memory.

I've mentioned the n-back task for improving working memory and intelligence in several - previous - posts. Kuriyama et al. now use this test to show that post-training sleep significantly enhances this improvement:
Working memory (WM) performance, which is an important factor for determining problem-solving and reasoning ability, has been firmly believed to be constant. However, recent findings have demonstrated that WM performance has the potential to be improved by repetitive training. Although various skills are reported to be improved by sleep, the beneficial effect of sleep on WM performance has not been clarified. Here, we show that improvement in WM performance is facilitated by posttraining naturalistic sleep. A spatial variant of the n-back WM task was performed by 29 healthy young adults who were assigned randomly to three different experimental groups that had different time schedules of repetitive n-back WM task sessions, with or without intervening sleep. Intergroup and intersession comparisons of WM performance (accuracy and response time) profiles showed that n-back accuracy after posttraining sleep was significantly improved compared with that after the same period of wakefulness, independent of sleep timing, subject's vigilance level, or circadian influences. On the other hand, response time was not influenced by sleep or repetitive training schedules. The present study indicates that improvement in n-back accuracy, which could reflect WM capacity, essentially benefits from posttraining sleep.

Wednesday, October 22, 2008

Cognitive therapy versus medication for depression

DeRubeis et al. offer an interesting review article in Nature Reviews Neuroscience on treatment outcomes and neural mechanisms, from which I pass on part of the abstract and some summary graphs:
Studies have shown that cognitive therapy is as efficacious as antidepressant medication at treating depression, and it seems to reduce the risk of relapse even after its discontinuation. Cognitive therapy and antidepressant medication probably engage some similar neural mechanisms, as well as mechanisms that are distinctive to each.


Cognitive therapy and antidepressant medication have comparable effects. This graph shows the response of outpatients who had moderate-to-severe depression to cognitive therapy (CT), antidepressant medication (ADM) or placebo. Patients who were assigned to ADM or to CT showed a significantly higher response rate after 8 weeks of treatment than those who were assigned to placebo. After 16 weeks of treatment the response rates of ADM and CT were almost identical.


Less relapse after cognitive therapy compared with antidepressant medication. The second phase of the parent antidepressant medication (ADM) versus cognitive therapy (CT) study followed patients who had responded to ADM or to CT3. Patients who had responded to ADM were randomly assigned to either continue ADM treatment for one year (beige and red lines) or to change to placebo treatment for 1 year (green line). Patients who responded to CT were allowed three sessions of CT during the 1-year continuation period. In the follow-up period, none of the patients received any treatment. The figure shows that prior treatment with CT protected against relapse of depression at least as well as the continued provision of ADM, and better than ADM treatment that was subsequently discontinued. Note that the patient group that was given ADM in the continuation year contained a number of patients who did not adhere to the medication regimen. The red line indicates the response of the ADM-continuation group including these non-compliant patients, whereas the beige line shows the response of the patients in this group after the non-compliant patients had been removed from the analysis.

After a graphic showing changes in blood-oxygen-level-dependent (BOLD) signal in response to cognitive and emotional tasks associated with cognitive therapy, the authors offer a summary hypothetical time course of the changes to amygdala and prefrontal function that are associated with antidepressant medication and cognitive therapy.


a | During acute depression, amygdala activity is increased (red) and prefrontal activity is decreased (blue) relative to activity in these regions in healthy individuals. b | Cognitive therapy (CT) effectively exercises the prefrontal cortex (PFC), yielding increased inhibitory function of this region. c | Antidepressant medication (ADM) targets amygdala function more directly, decreasing its activity. d | After ADM or CT, amygdala function is decreased and prefrontal function is increased. The double-headed arrow between the amygdala and the PFC represents the bidirectional homeostatic influences that are believed to operate in healthy individuals.

Our somatosensory cortex embodies the facial expressions of others

The Editor's choice section of science describes an interesting bit of work by Pitcher et al. showing the embodyment of our social cognition:
Humans are especially interested in faces, as a means of sending signals--witness the sizeable arc of somatosensory cortex devoted to representation of one's own face--and as a substrate for social cognition. Pitcher et al. describe results supporting theories of embodied cognition and emotion, which posit cognition and emotion as being shaped by our bodily movements and perceptions. They used repetitive transcranial magnetic stimulation (rTMS) to interfere with neural activity in the face areas of the somatosensory cortex while people discriminated the emotional expressions of faces (happy, sad, surprised, fearful, angry, and disgusted) and found that accuracy dropped significantly, as it also did when the occipital face area was similarly stimulated. The temporal sequence of neural processing was then delineated using double-pulse TMS, showing that the occipital area acted in the time window from 60 to 100 ms after the face stimulus was shown, whereas the somatosensory area was active a bit later, between 100 and 170 ms.

Tuesday, October 21, 2008

Botnets

Here is a scary article.

Evolution of Religious Prosociality

Norenzayan and Shariff offer an interesting review article on empirical evidence for religious prosociality. Here is one clip and two figures from the article.
Agreement is emerging that selective pressures over the course of human evolution can explain the wide cross-cultural re-occurrence, historical persistence, and predictable cognitive structure of religious beliefs and behaviors. The tendency to detect agency in nature likely supplied the cognitive template that supports the pervasive belief in supernatural agents. These agents are widely believed to transcend physical, biological, and psychological limitations. However, other important details are subject to cultural variation. Although in many societies supernatural agents are not directly concerned with human morality, in many others, morally concerned agents use their supernatural powers to observe and, in some cases, to punish and reward human social interactions. Examples include the God of Abrahamic religions and Viracocha, the Incan supreme God, but also many morally concerned deities found in traditional societies, such as the adalo, ancestral spirits of the Kwaio Solomon islanders. These beliefs are likely to spread culturally to the extent that they facilitate ingroup cooperation. This could occur by conforming to individual psychology that favors reputation-sensitive prosocial tendencies, as the by-product account holds; by competition among social groups, as the cultural group selection account would suggest; or possibly by some combination of the two. Religious behaviors and rituals, if more costly to cooperating group members than to freeloaders, may have reliably signaled the presence of devotion and, therefore, cooperative intention toward ingroup members, in turn, buffering religious groups against defection from freeloaders and reinforcing cooperative norms. Religious prosociality, thus, may have softened the limitations that kinship-based and (direct or indirect) reciprocity-based altruism place on group size. In this way, the cultural spread of religious prosociality may have facilitated the rise of stable, large, cooperative communities of genetically unrelated individuals.


Figure - Implicit activation of God concepts, relative to a neutral prime, increased offers in the one-shot, anonymous Dictator Game. Priming secular concepts indicating moral authority had a similar effect. The results showed not only a quantitative increase in generosity, but also a qualitative shift in social norms. In the control group, the modal response was selfishness, a plurality of players pocketed all $10. In the God group, the mode shifted to fairness, a plurality of players split the money evenly (N = 75). It remains to be seen, however, whether these effects
would occur if the recipient was clearly marked as an outgroup member.


Figure - Life expectancy of religious versus secular communes. An analysis of 200 religious and secular communes in 19th-century America (29), for every year of their life course, religious communes were about four times as likely to survive than their secular counterparts. This difference remained after statistically controlling for type of commune movement, year founded, and year at risk of dissolution (the last control assesses major historical trends that may independently impact commune dissolution).

Redefining Depression as Mere Sadness

An article by Pies with the title of this post is worth reading. It deals with the criticism that modern psychiatric practice, in collusion with pill pushing pharmaceutical companies, has medicalized “normal sadness” brought on by external circumstances. (Added note: Pies has emailed this this link to a more detailed discussion posted on the PsychCentral website.) Here are some clips from the NYTimes article:
In their recent book “The Loss of Sadness” (Oxford, 2007), Allan V. Horwitz and Jerome C. Wakefield assert that for thousands of years, symptoms of sadness that were “with cause” were separated from those that were “without cause.” Only the latter were viewed as mental disorders.

With the advent of modern diagnostic criteria, these authors argue, doctors were directed to ignore the context of the patient’s complaints and focus only on symptoms — poor appetite, insomnia, low energy, hopelessness and so on. The current criteria for major depression, they say, largely fail to distinguish between “abnormal” reactions caused by “internal dysfunction” and “normal sadness” brought on by external circumstances. And they blame vested interests — doctors, researchers, pharmaceutical companies — for fostering this bloated concept of depression.

But while this increasingly popular thesis contains a kernel of truth, it conceals a bushel basket of conceptual and scientific problems.

For one thing, if modern diagnostic criteria were converting mere sadness into clinical depression, we would expect the number of new cases of depression to be skyrocketing compared with rates in a period like the 1950s to the 1970s. But several new studies in the United States and Canada find that the incidence of serious depression has held relatively steady in recent decades.

Second, it may seem easy to determine that someone with depressive complaints is reacting to a loss that touched off the depression. Experienced clinicians know this is rarely the case.

Third, and perhaps most troubling, is the implication that a recent major loss makes it more likely that the person’s depressive symptoms will follow a benign and limited course, and therefore do not need medical treatment. This has never been demonstrated, to my knowledge, in any well-designed studies. And what has been demonstrated, in a study by Dr. Sidney Zisook, is that antidepressants may help patients with major depressive symptoms occurring just after the death of a loved one.

Monday, October 20, 2008

The iPathology of your iBrain

Small and Vorgan offer an engaging article in Scientific American Mind on how daily exposure to television, computers, smart phones, video games, search engines and web browsers is rewiring our brains. A modern generation is rising with brains that are very different from the brains of those of us whose basic brain wiring was laid down in a time when direct social interactions were more the norm. One of the authors (Small) compared brain activities in computer- savvy versus computer-naive 50-60 year olds while searching for accurate information on a topic using Google, subtracting activity associated with just reading a text to determine activity specific to the searching function (which was the same in the two groups). In the baseline scanning session during searching on Google, the computer-savvy subjects engaged their dorsolateral prefrontal cortex while the Internet-naive subjects showed minimal activation in this region. After just five days of practice,the exact same neural circuitry in the front part of the brain became active in the Internet-naive subjects. Five hours on the Internet, and these participants had already rewired their brains.
Our high-tech revolution has plunged us into a state of “continuous partial attention,” which software executive Linda Stone, who coined the term in 1998, describes as continually staying busy—keeping tabs on everything while never truly focusing on anything. Continuous partial attention differs from multitasking, wherein we have a purpose for each task and we are trying to improve efficiency and productivity. Instead, when our minds partially attend, and do so continuously, we scan for an opportunity for any type of contact at every given moment. We virtually chat as our text messages flow, and we keep tabs on active buddy lists (friends and other screen names in an instant message program); everything, everywhere, is connected through our peripheral attention. Although having all our pals online from moment to moment seems intimate, we risk losing personal touch with our real-life relationships and may experience an artificial sense of intimacy as compared with when we shut down our devices and devote our attention to one
individual at a time.

When paying continuous partial attention, people may place their brain in a heightened state of stress. They no longer have time to reflect, contemplate or make thoughtful decisions. Instead they exist in a sense of constant crisis—on alert for a new contact or bit of exciting news or information at any moment. Once people get used to this state, they tend to thrive on the perpetual connectivity. It feeds their ego and sense of self-worth, and it becomes irresistible. Neuroimaging studies suggest that this sense of selfworth may protect the size of the hippocampus—the horseshoeshaped brain region in the medial (inward-facing) temporal lobe, which allows us to learn and remember new information. Psychiatry professor Sonia J. Lupien and her associates at McGill University studied hippocampal size in healthy younger and older adult volunteers. Measures of self esteem correlated significantly with hippocampal size, regardless of age. They also found that the more people felt in control of their lives, the larger the hippocampus. But at some point, the sense of control and self-worth we feel when we maintain continuous partial attention tends to break down—our brains were not built to sustain such monitoring for extended periods. Eventually the hours of unrelenting digital connectivity can create a unique type of brain strain. Many people who have been working on the Internet for several hours without a break report making frequent errors in their work. On signing off, they notice feeling spaced out, fatigued, irritable and distracted, as if they are in a “digital fog.” This new form of mental stress, what Small terms “techno-brain burnout,” is threatening to become an epidemic. Under this kind of stress, our brains instinctively signal the adrenal gland to secrete cortisol and adrenaline. In the short run, these stress hormones boost energy levels and augment memory, but over time they actually impair cognition, lead to depression, and alter the neural circuitry in the hippocampus, amygdala and prefrontal cortex—the brain regions that control mood and thought. Chronic and prolonged techno-brain burnout can even reshape the underlying brain structure.

While the brains of today’s digital natives are wiring up for rapid-fire cyber searches, however, the neural circuits that control the more traditional learning methods are neglected and gradually diminished. The pathways for human interaction and communication weaken as customary one-on-one people skills atrophy. Our U.C.L.A. research team and other scientists have shown that we can intentionally alter brain wiring and reinvigorate some of these dwindling neural pathways, even while the newly evolved technology circuits bring our brains to extraordinary levels of potential.

Bayesian estimation on the presidential election.

From the "Random Samples" section of the Oct. 17 Science Magazine:
In the winner-take-all world of politics, candidates know that even a modest lead in the polls can spell almost certain victory. Sheldon Jacobson, an operations research specialist at the University of Illinois, Urbana-Champaign, and colleagues, including a group of students, have attempted to quantify that insight for the current United States presidential election, putting their predictions for the Electoral College on a Web site, election08.cs.uiuc.edu. Using a statistical method known as Bayesian estimation, they combined an analysis of results from the 2004 Bush-versus-Kerry contest with current state-by-state polls for Obama versus McCain to produce probabilities for each candidate of carrying each state. They then converted the estimates into a probability distribution for the total number of Electoral College votes a candidate might receive. In Indiana, for example, polls as of 4 October gave McCain a slight 2.5% lead. But given that Bush carried Indiana in 2004 by 20.7%, a Bayesian calculation indicates McCain's chance of winning the state's 11 Electoral College votes at about 87%. Most states are now in the bag for one candidate or the other; only a handful are truly in Bayesian play. Current calculations give McCain no chance of victory. "However," Jacobson cautions, "if the polls move, then so will our forecasts."

Friday, October 17, 2008

The rise of the machines

Richard Dooloing writes on how a physicist, a wizard and a serial killer warned us of the current financial crisis.
We are living, we have long been told, in the Information Age. Yet now we are faced with the sickening suspicion that technology has run ahead of us. Man is a fire-stealing animal, and we can’t help building machines and machine intelligences, even if, from time to time, we use them not only to outsmart ourselves but to bring us right up to the doorstep of Doom.

We are still fearful, superstitious and all-too-human creatures. At times, we forget the magnitude of the havoc we can wreak by off-loading our minds onto super-intelligent machines, that is, until they run away from us, like mad sorcerers’ apprentices, and drag us up to the precipice for a look down into the abyss.

As the financial experts all over the world use machines to unwind Gordian knots of financial arrangements so complex that only machines can make — “derive” — and trade them, we have to wonder: Are we living in a bad sci-fi movie? Is the Matrix made of credit default swaps?

When Treasury Secretary Paulson (looking very much like a frightened primate) came to Congress seeking an emergency loan, Senator Jon Tester of Montana, a Democrat still living on his family homestead, asked him: “I’m a dirt farmer. Why do we have one week to determine that $700 billion has to be appropriated or this country’s financial system goes down the pipes?”

“Well, sir,” Mr. Paulson could well have responded, “the computers have demanded it.”

Men and women - different gene expression changes in brain on aging

Berchtold et al. pack quite a lot of interesting information into their abstract describing work on sexually dimorphic gene expression changes during human aging. I'm right in the middle of this quote: "Prominent change occurred in the sixth to seventh decades across cortical regions, suggesting that this period is a critical transition point in brain aging, particularly in males." Here is the abstract:
Gene expression profiles were assessed in the hippocampus, entorhinal cortex, superior-frontal gyrus, and postcentral gyrus across the lifespan of 55 cognitively intact individuals aged 20–99 years. Perspectives on global gene changes that are associated with brain aging emerged, revealing two overarching concepts. First, different regions of the forebrain exhibited substantially different gene profile changes with age. For example, comparing equally powered groups, 5,029 probe sets were significantly altered with age in the superior-frontal gyrus, compared with 1,110 in the entorhinal cortex. Prominent change occurred in the sixth to seventh decades across cortical regions, suggesting that this period is a critical transition point in brain aging, particularly in males. Second, clear gender differences in brain aging were evident, suggesting that the brain undergoes sexually dimorphic changes in gene expression not only in development but also in later life. Globally across all brain regions, males showed more gene change than females. Further, Gene Ontology analysis revealed that different categories of genes were predominantly affected in males vs. females. Notably, the male brain was characterized by global decreased catabolic and anabolic capacity with aging, with down-regulated genes heavily enriched in energy production and protein synthesis/transport categories. Increased immune activation was a prominent feature of aging in both sexes, with proportionally greater activation in the female brain. These data open opportunities to explore age-dependent changes in gene expression that set the balance between neurodegeneration and compensatory mechanisms in the brain and suggest that this balance is set differently in males and females, an intriguing idea.

Thursday, October 16, 2008

Embodyment and Art

I just received an email from Andrew Werth regarding the previous post. Andrew is a former software engineer turned artist whose paintings are about perception and embodiment. I thought I would pass on this link to his website, which lets one view paintings in his Embodyment Series.

Arguing for Embodied Consciousness

I thought I would pass along portions of a review in Science by Harold Fromm which has the title of this post, of Edward Slingerland's new book, "What Science Offers the Humanities - Integrating Body and Culture."
...his overall task is to address the befuddled dualism that still dominates most of our intellectual disciplines...Slingerland's central theme is that everything human has evolved in the interests of the materiality of the body. He identifies objectivist realism and postmodern relativity, both insufficiently attentive to the body, as the major epistemologies to be swept away, followed by the dualism of body and soul. For Slingerland, the presiding genii behind such a cleansing are George Lakoff and Mark Johnson, with heavier debts to Johnson [whose terse summary of embodiment in (1) appeared too late for Slingerland to reference]. They view all thought and human behavior as generated by the body and expressed as conceptual metaphors that translate physical categories (such as forward, backward, up, and down) into abstract categories (such as progress, benightedness, divinity, immorality). These body-driven metaphors, Slingerland writes, are a "set of limitations on human cognition, constraining human conceptions of entities, categories, causation, physics, psychology, biology, and other humanly relevant domains."

The supposedly objective world is not "some preexisting object out there in the world, with a set of invariant and observer-independent properties, simply waiting to be found the way one finds a lost sock under the bed." All we can ever see or understand is what our own bodily faculties permit via the current structure of the brain.

In opposition to objective realism, postmodern relativity regards language and culture as constituting the only "real" world possible for us. It posits an endless hall of mirrors with no access to outside--epitomized by Derrida's notorious remark that there is nothing (at least for humans) outside of texts (i.e., culture). This view, which dominated the humanities for several decades, is mercifully beginning to fade as the cognitive sciences have matured and are increasingly promulgated.

Even though the knowing human subject is itself just a thing and not an immaterial locus of reason, the universe it experiences is as real and functional for us as any "thing" could possibly be. We do get some things "right," even if we can never know the noumenal genesis behind our knowledge. And the very concept of noumena (things in themselves independent of any observer) now seems somewhat obsolete, given that the intuition of discrete, self-bounded "things" is as built-in to the human psyche as the Kantian intuitions of space and time, grounding all experience.

Our million billion synapses produce a "person" with the illusion of a self. Slingerland holds that "we are robots designed to be constitutionally incapable of experiencing ourselves and other conspecifics as robots." Our innate and overactive theory of mind (that other people, like ourselves, have "intentions") projects agency onto everything--in the past, even onto stones and trees. The "hard problem" for philosophy of consciousness (to use David Chalmers's phrase) remains: what are thoughts, cogitations, thinkers, qualia? Chalmers's solution, alas, swept away Cartesian dualism only to sneak his own magic spook, conscious experience (for him, on par with mass, charge, and space-time), in through the back door (2, 3).

Slingerland starts with Darwin and eventually follows Daniel Dennett so far as to agree that consciousness can be done full justice through third-person descriptions that require no mysterious, unaccounted-for, nonmaterial, first-person entity as substrate. Thus the famous "Mary," who intellectually knows everything there is to know about color despite having been sequestered for life in a color-free lab, will recognize red the first time she steps outside (4). And Thomas Nagel's famous bats don't know anything about bathood that we can't figure out for ourselves from observation (5). No first-person construct, no locus of consciousness, need be invoked.

The next step, if you want to go so far (the jury is out), is to eliminate consciousness altogether, because there's nothing for it to do that can't be done without it. And with it, you need a spook to keep the show on the road. Choose your insoluble problem: eliminate consciousness altogether as superfluous or explain it (if there's really a you who makes such choices). Slingerland prefers the first option.

His conclusion, which I can hardly do justice to here, is relatively satisfying. He notes that although we don't have great difficulty knowing that Earth revolves around the Sun while feeling that the Sun is rising and setting (Dennett's favorite example of folk psychology), "no cognitively undamaged human being can help acting like and at some level really feeling that he or she is free"--however nonsensical the notion of agencyless free will (i.e., "choices" without a self to make them). Still, once the corrosive acid of Darwinism [to use Dennett's figure from (6)] has resolved the body-mind dualism into body alone, some but not most of us are able "to view human beings simultaneously under two descriptions: as physical systems and as persons."

References

1. M. Johnson, The Meaning of the Body: Aesthetics of Human Understanding (Univ. of Chicago Press, Chicago, 2007).
2. D. J. Chalmers, J. Consciousness Stud. 2, 200 (1995).
3. D. J. Chalmers, The Conscious Mind: In Search of a Fundamental Theory (Oxford Univ. Press, Oxford, 1996).
4. F. Jackson, Philos. Q. 32, 127 (1982).
5. T. Nagel, Philos. Rev. 83, 435 (1974).
6. D. C. Dennett, Darwin's Dangerous Idea: Evolution and the Meaning of Life (Simon and Schuster, New York, 1995).

Your bladder and your brain.

Overactive bladder, usually caused by bladder obstruction in males, apparently affects ~17% of the population, towards whom those awful pharmaceutical television adds are directed. Signals arising from bladder or colonic pathology are processed by the cortex and can potentially be expressed as central symptoms (e.g., hyperarousal, attention disorders, anxiety) that occur alongside the visceral pathology. Rickenbacher et al. now show, in a rat model, that bladder obstruction not only botches up the bladder, but also brain regions involved in its regulation.
Neural circuits that allow for reciprocal communication between the brain and viscera are critical for coordinating behavior with visceral activity. At the same time, these circuits are positioned to convey signals from pathologic events occurring in viscera to the brain, thereby providing a structural basis for comorbid central and peripheral symptoms. In the pons, Barrington's nucleus and the norepinephrine (NE) nucleus, locus coeruleus (LC), are integral to a circuit that links the pelvic viscera with the forebrain and coordinates pelvic visceral activity with arousal and behavior. Here, we demonstrate that a prevalent bladder dysfunction, produced by partial obstruction in rat, has an enduring disruptive impact on cortical activity through this circuit. Within 2 weeks of partial bladder obstruction, the activity of LC neurons was tonically elevated. LC hyperactivity was associated with cortical electroencephalographic activation that was characterized by decreased low-frequency (1–3 Hz) activity and prominent theta oscillations (6–8 Hz) that persisted for 4 weeks. Selective lesion of the LC–NE system significantly attenuated the cortical effects. The findings underscore the potential for significant neurobehavioral consequences of bladder disorders, including hyperarousal, sleep disturbances, and disruption of sensorimotor integration, as a result of central noradrenergic hyperactivity. The results further imply that pharmacological manipulation of central NE function may alleviate central sequelae of these visceral disorders.

Wednesday, October 15, 2008

Applied neuroeconomics - the fear of loss

Bajaj does an interesting writeup of well-known crowd psychological dynamics behind recent "irrational" drops in the stock market. Fear is a more powerful force than greed. Our aversive reaction to losing $1000 is greater than our pleasure at earning the same amount...
fear now seems to rule, with investors often exhibiting a Wall Street version of the fight-or-flight mechanism — selling first, and asking questions later...some analysts are starting to suggest the markets are showing signs of “capitulation” — what happens when even the bullish holdouts, the unflagging optimists, throw up their hands and join the stampede out of the market...To some, signs of capitulation can be read as an indicator that the bottom may be near.
The opposite swing of the cycle is buying at the top of a bubble. I remember during my winter stays in Ft. Lauderdale in 2005 and 2006, every fourth person I chatted with seemed to be a realtor and dinner conversations were dominated by stories about fast profits on flipped condominiums.

Wordwatchers

An article by Wapner points to the work of James Pennebacker and his Wordwatchers blog that is tracking the candidates use of words during the 2008 election. The blog makes fascinating reading. Here is just one clip:
Predicting how they will govern. Most language dimensions that we study are probably better markers of how people will lead than who will vote for them. Some dimensions that are relevant include:

Cognitive complexity. A particularly reliable marker of cognitive complexity is the exclusive word dimension. Exclusive words such as but, except, without, exclude, signal that the speaker is making an effort to distinguish what is in a category and not in a category. Those who use more exclusive words make better grades in college, are more honest in lab studies, and have more nuanced understanding of events and people. Through the primaries until now, Obama has consistently been the highest in exclusive word use and McCain the lowest.

Categorical versus fluid thinking. Some people naturally approach problems by assigning them to categories. Categorical thinking involves the use of articles (a, an, the) and concrete nouns. Men, for example, use articles at much higher rates than women. Fluid thinking involves describing actions and changes, often in more abstract ways. A crude measure of fluid thinking is the use of verbs. Women use verbs more than men.

McCain and Obama could not be more different in their use of articles and verbs. McCain uses verbs at an extremely low rate and articles at a fairly high rate. Obama, on the other hand, is remarkably high in his use of verbs and low in his use of articles. These patterns suggest that McCain’s natural way of understanding the world is to first label the problem and find a way to put it into a pre-existing category. Obama is more likely to define the world as ongoing actions or processes.

Personal and socially connected. Individuals who think about and try to connect with others tend to use more personal pronouns (I, we, you, she, they) than those who are more socially detached. Bush was higher than Kerry or Gore. McCain has consistently been much higher than any other candidate in this election cycle. His use of 1st person singular (I, me, my) is particularly high which often signals an openness and honesty. Obama uses personal pronouns at moderate levels - similar to Hillary Clinton and most other primary candidates of both parties.

Restrained versus impulsive. People vary in the degree to which they act quickly or shoot from the hip versus stand back and consider their options. Over the last few years, some have argued that the use of negations (e.g., no, not, never) indicate a sign of inhibition or constraint. Low use of negations may be linked to impulsiveness. Bush was low in negations whereas Kerry was quite high. Across the election cycle, Obama has consistently been the highest user of negations - suggesting a restrained approach - where as McCain has been the lowest - a more impulsive way of dealing with the world.