Analyses of the relation between obesity and mortality typically evaluate risk with respect to weight recorded at a single point in time. As a consequence, there is generally no distinction made between nonobese individuals who were never obese and nonobese individuals who were formerly obese and lost weight. We introduce additional data on an individual’s maximum attained weight and investigate four models that represent different combinations of weight at survey and maximum weight. We use data from the 1988–2010 National Health and Nutrition Examination Survey, linked to death records through 2011, to estimate parameters of these models. We find that the most successful models use data on maximum weight, and the worst-performing model uses only data on weight at survey. We show that the disparity in predictive power between these models is related to exceptionally high mortality among those who have lost weight, with the normal-weight category being particularly susceptible to distortions arising from weight loss. These distortions make overweight and obesity appear less harmful by obscuring the benefits of remaining never obese. Because most previous studies are based on body mass index at survey, it is likely that the effects of excess weight on US mortality have been consistently underestimated.
This blog reports new ideas and work on mind, brain, behavior, psychology, and politics - as well as random curious stuff. (Try the Dynamic Views at top of right column.)
Tuesday, February 02, 2016
The burden of obesity revisited.
A previous mindblog post has noted the obesity paradox, generated by data suggesting that fat people may live longer. Stokes and Preston,however, note an important distinction that may evaporate the apparent paradox:
Monday, February 01, 2016
The 8 second attention span.
Wow, Egan notes that our average attention span has fallen to eight seconds, down from 12 in the year 2000. (That makes it gratifying that the average amount of time spent by someone on this website is over 4 minutes.)
...a quote from Satya Nadella, the chief executive officer of Microsoft... “The true scarce commodity of the near future will be human attention.”...Putting aside Microsoft’s self-interest in promoting quick-flash digital ads with what may be junk science, there seems little doubt that our devices have rewired our brains. We think in McNugget time. The trash flows, unfiltered, along with the relevant stuff, in an eternal stream. And the last hit of dopamine only accelerates the need for another one.
You see it in the press, the obsession with mindless listicles that have all the staying power of a Popsicle. You see it in our politics, with fear-mongering slogans replacing anything that requires sustained thought. And the collapse of a fact-based democracy, where, for example, 60 percent of Trump supporters believe Obama was born in another country, has to be a byproduct of the pick-and-choose news from the buffet line of our screens.Egan suggests that gardening and deep reading of biographies are useful antidotes.
Friday, January 29, 2016
Fourth Industrial Revolution?? Maybe not.....
Just after starting my second slog through Schwab's 'the fourth industrial revolution' noted in yesterday's post, I see Paul Krugman's review of 'The Rise and Fall of American Growth' by Robert Gordon. Gordon's magnum opus suggests that Schwab's futurism is overblown. (see also Thomas Edsall's excellent piece on the divide and debate between optimistic and pessimistic economists. Also, see this Slate article that chronicles how many times over the past 75 years the term "fourth industrial revolution" has been fetched up to describe a recent or coming advance.) Some clips from Klugman's review:
...[Gordon] has argued that the I.T. revolution is less important than any one of the five Great Inventions that powered economic growth from 1870 to 1970: electricity, urban sanitation, chemicals and pharmaceuticals, the internal combustion engine and modern communication.
What happened between 1870 and 1940, he argues, and I would agree, is what real transformation looks like. Any claims about current progress need to be compared with that baseline to see how they measure up.
And it’s hard not to agree with him that nothing that has happened since is remotely comparable. Urban life in America on the eve of World War II was already recognizably modern; you or I could walk into a 1940s apartment, with its indoor plumbing, gas range, electric lights, refrigerator and telephone, and we’d find it basically functional. We’d be annoyed at the lack of television and Internet — but not horrified or disgusted.
By contrast, urban Americans from 1940 walking into 1870-style accommodations — which they could still do in the rural South — were indeed horrified and disgusted. Life fundamentally improved between 1870 and 1940 in a way it hasn’t since.One of Gordon's arguments against the techno-optimists is that:
...genuinely major innovations normally bring about big changes in business practices, in what workplaces look like and how they function. And there were some changes along those lines between the mid-1990s and the mid-2000s — but not much since, which is evidence for Gordon’s claim that the main impact of the I.T. revolution has already happened.Techno-futurists would argue strongly against this, citing the rise of the sharing economy, entities like Airbnb and Uber, and changes in the workplace from hierarchical to distributed organization.
Gordon suggests that the future is all too likely to be marked by stagnant living standards for most Americans, because the effects of slowing technological progress will be reinforced by a set of “headwinds”: rising inequality, a plateau in education levels, an aging population and more.
It’s a shocking prediction for a society whose self-image, arguably its very identity, is bound up with the expectation of constant progress. And you have to wonder about the social and political consequences of another generation of stagnation or decline in working-class incomes.
Of course, Gordon could be wrong: Maybe we’re on the cusp of truly transformative change, say from artificial intelligence or radical progress in biology (which would bring their own risks). But he makes a powerful case. Perhaps the future isn’t what it used to be.
Blog Categories:
culture/politics,
future,
futures,
technology
Thursday, January 28, 2016
Predicting the Future: The Fourth Industrial Revolution
Klaus Schwab, the guy who runs the annual Davos Switzerland World Economic Forum of “very important people” in the world, has generated a book with the title of this year's meeting, “The Fourth Industrial Revolution,” about a future that is both terrifying and optimistic. Despite its gobbledegook, committee-speak, and bullet points, it provides a lot of fascinating information and is well worth a look. After doing a quick read-through, I’m starting a second more careful read and clicking through to many of the references (the kindle version is handy for this.)
The first industrial revolution spanned from about 1760 to around 1840. Triggered by the construction of railroads and the invention of the steam engine, it ushered in mechanical production. The second industrial revolution, which started in the late 19th century and into the early 20th century, made mass production possible, fostered by the advent of electricity and the assembly line. The third industrial revolution began in the 1960s. It is usually called the computer or digital revolution because it was catalysed by the development of semiconductors, mainframe computing (1960s), personal computing (1970s and 80s) and the internet (1990s). Mindful of the various definitions and academic arguments used to describe the first three industrial revolutions, I believe that today we are at the beginning of a fourth industrial revolution. It began at the turn of this century and builds on the digital revolution. It is characterized by a much more ubiquitous and mobile internet, by smaller and more powerful sensors that have become cheaper, and by artificial intelligence and machine learning.
I am well aware that some academics and professionals consider the developments that I am looking at as simply a part of the third industrial revolution. Three reasons, however, underpin my conviction that a fourth and distinct revolution is underway:
Velocity: Contrary to the previous industrial revolutions, this one is evolving at an exponential rather than linear pace. This is the result of the multifaceted, deeply interconnected world we live in and the fact that new technology begets newer and ever more capable technology.
Breadth and depth: It builds on the digital revolution and combines multiple technologies that are leading to unprecedented paradigm shifts in the economy, business, society, and individually. It is not only changing the “what” and the “how” of doing things but also “who” we are.
Systems Impact: It involves the transformation of entire systems, across (and within) countries, companies, industries and society as a whole.
Blog Categories:
culture/politics,
future,
futures,
technology
Wednesday, January 27, 2016
The ascendant candidacy of Donald J. Trump
You should check Steven Rattner's interesting commentary asking what we owe to those on the losing end of globalization:
Percent change in number of US employees in each industry, 2000 to 2015
Percent change in wages for each industry, 2009 to 2015:
Not surprisingly, the shifts shown in these graphs are ending badly politically. The 'losers' form a core of support for both Donald Trump and Ted Cruz.
Axelrod notes:
Mr. Trump has found an audience with Americans disgruntled by the rapid, disorderly change they associate with national decline and their own uncertain prospects. Policies be damned, who better to set things right than the defiant strong man who promises by sheer force of will to make America great again?
Open-seat presidential elections are shaped by perceptions of the style and personality of the outgoing incumbent. Voters rarely seek the replica of what they have. They almost always seek the remedy, the candidate who has the personal qualities the public finds lacking in the departing executive...who among the Republicans is more the antithesis of Mr. Obama than the trash-talking, authoritarian, give-no-quarter Mr. Trump?
Percent change in number of US employees in each industry, 2000 to 2015
Percent change in wages for each industry, 2009 to 2015:
Not surprisingly, the shifts shown in these graphs are ending badly politically. The 'losers' form a core of support for both Donald Trump and Ted Cruz.
Axelrod notes:
Mr. Trump has found an audience with Americans disgruntled by the rapid, disorderly change they associate with national decline and their own uncertain prospects. Policies be damned, who better to set things right than the defiant strong man who promises by sheer force of will to make America great again?
Open-seat presidential elections are shaped by perceptions of the style and personality of the outgoing incumbent. Voters rarely seek the replica of what they have. They almost always seek the remedy, the candidate who has the personal qualities the public finds lacking in the departing executive...who among the Republicans is more the antithesis of Mr. Obama than the trash-talking, authoritarian, give-no-quarter Mr. Trump?
Tuesday, January 26, 2016
Social relationships and biomarkers of longevity across our life span.
Yang et al. aggregate a massive amount of data from four large longitudinal surveys to show associations between social relationships physiological markers of health:
Two decades of research indicate causal associations between social relationships and mortality, but important questions remain as to how social relationships affect health, when effects emerge, and how long they last. Drawing on data from four nationally representative longitudinal samples of the US population, we implemented an innovative life course design to assess the prospective association of both structural and functional dimensions of social relationships (social integration, social support, and social strain) with objectively measured biomarkers of physical health (C-reactive protein, systolic and diastolic blood pressure, waist circumference, and body mass index) within each life stage, including adolescence and young, middle, and late adulthood, and compare such associations across life stages. We found that a higher degree of social integration was associated with lower risk of physiological dysregulation in a dose–response manner in both early and later life. Conversely, lack of social connections was associated with vastly elevated risk in specific life stages. For example, social isolation increased the risk of inflammation by the same magnitude as physical inactivity in adolescence, and the effect of social isolation on hypertension exceeded that of clinical risk factors such as diabetes in old age. Analyses of multiple dimensions of social relationships within multiple samples across the life course produced consistent and robust associations with health. Physiological impacts of structural and functional dimensions of social relationships emerge uniquely in adolescence and midlife and persist into old age.
Blog Categories:
aging,
culture/politics,
fear/anxiety/stress,
social cognition
Monday, January 25, 2016
The evolution of dance
From Laland et al.:
Evidence from multiple sources reveals a surprising link between imitation and dance. As in the classical correspondence problem central to imitation research, dance requires mapping across sensory modalities and the integration of visual and auditory inputs with motor outputs. Recent research in comparative psychology supports this association, in that entrainment to a musical beat is almost exclusively observed in animals capable of vocal or motor imitation. Dance has representational properties that rely on the dancers’ ability to imitate particular people, animals or events, as well as the audience’s ability to recognize these correspondences. Imitation also plays a central role in learning to dance and the acquisition of the long sequences of choreographed movements are dependent on social learning. These and other lines of evidence suggest that dancing may only be possible for humans because its performance exploits existing neural circuitry employed in imitation.
This painting by Edgar Degas not only depicts a ballet rehearsal but also illustrates the roles of imitation and synchronyClips from the body of the article, with another figure:
...there can be no doubt that, compared to other animals, humans are exceptional imitators. It may be no coincidence that a recent PET scan analysis of the neural basis of dance found that foot movement to music excited regions of neural circuitry (e.g. the right frontal operculum) previously associated with imitation. Dancing may only be possible because its performance exploits the neural circuitry employed in imitation. Such reasoning applies equally where individuals dance alone; unlike much human behavior, dancing inherently seems to require a brain capable of solving the correspondence problem.
Dance often tells a story, and this representational quality provides another link with imitation. For instance, in the astronomical dances of ancient Egypt, priests and priestesses, accompanied by harps and pipes, mimed significant events in the story of a god or imitated cosmic patterns, such as the rhythm of night and day. Africa, Asia, Australasia and Europe all possess long-standing traditions for masked dances, in which the performers portray the character associated with the mask and enact religious stories. Native Americans have many animal dances, such as the Buffalo dance, which was thought to lure buffalo herds close to the village, and the eagle dance, which is a tribute to these revered birds. This tradition continues to the present. In 2009, Rambert (formerly the Rambert Dance Company), a world leader in contemporary dance, marked the bicentenary of Charles Darwin’s birth and 150th anniversary of his seminal work On the Origin of Species by collaborating with one of us (N.C.) to produce Comedy of Change (Figure above), which evoked animal behaviour on stage with spellbinding accuracy. In all such instances, the creation and performance of the dance requires an ability on the part of the dancer to imitate the movements and sounds of particular people, animals, or events. Such dances re-introduce the correspondence problem, as the dancer, choreographer and audience must be able to connect the dancers’ movements to the target phenomenon they represent.
Blog Categories:
culture,
culture/politics,
mirror neurons,
social cognition
Friday, January 22, 2016
On teaching old dogs new tricks.
Metcalfe et al. find that older healthy adults not only are better than young adults at answering general-information questions in the first place, but also, when they do make a mistake, they are more likely than young adults to correct those errors. Correcting errors is, of course, the quintessential new-learning task: To correct mistakes, one needs to supplant entrenched responses with new ones. The fact that older adults display greater facility at error correction than young adults contravenes the view that aging necessarily produces cognitive rigidity and an inability to learn. Here is their abstract:
Although older adults rarely outperform young adults on learning tasks, in the study reported here they surpassed their younger counterparts not only by answering more semantic-memory general-information questions correctly, but also by better correcting their mistakes. While both young and older adults exhibited a hypercorrection effect, correcting their high-confidence errors more than their low-confidence errors, the effect was larger for young adults. Whereas older adults corrected high-confidence errors to the same extent as did young adults, they outdid the young in also correcting their low-confidence errors. Their event-related potentials point to an attentional explanation: Both groups showed a strong attention-related P3a in conjunction with high-confidence-error feedback, but the older adults also showed strong P3as to low-confidence-error feedback. Indeed, the older adults were able to rally their attentional resources to learn the true answers regardless of their original confidence in the errors and regardless of their familiarity with the answers.
Blog Categories:
aging,
brain plasticity,
memory/learning
Thursday, January 21, 2016
Several perspectives on the valuation of outgroups.
A recent issue of the Proceedings of the National Academy of Science has two relevant articles:
Keelah et al. show that Americans’ stereotypes about racial groups may actually reflect their stereotypes about these groups’ presumed home ecologies. Harsh and unpredictable (“desperate”) ecologies induce fast strategy behaviors such as impulsivity, whereas resource-sufficient and predictable (“hopeful”) ecologies induce slow strategy behaviors such as future focus.
Keelah et al. show that Americans’ stereotypes about racial groups may actually reflect their stereotypes about these groups’ presumed home ecologies. Harsh and unpredictable (“desperate”) ecologies induce fast strategy behaviors such as impulsivity, whereas resource-sufficient and predictable (“hopeful”) ecologies induce slow strategy behaviors such as future focus.
...when provided with information about a person’s race (but not ecology), individuals’ inferences about blacks track stereotypes of people from desperate ecologies, and individuals’ inferences about whites track stereotypes of people from hopeful ecologies. However, when provided with information about both the race and ecology of others, individuals’ inferences reflect the targets’ ecology rather than their race: black and white targets from desperate ecologies are stereotyped as equally fast life history strategists, whereas black and white targets from hopeful ecologies are stereotyped as equally slow life history strategists. These findings suggest that the content of several predominant race stereotypes may not reflect race, per se, but rather inferences about how one’s ecology influences behavior.And, Ginges et al. show that thinking from God's perspective decreases biased valuation of the life of a nonbeliever.
Religious belief is often thought to motivate violence because it is said to promote norms that encourage tribalism and the devaluing of the lives of nonbelievers. If true, this should be visible in the multigenerational violent conflict between Palestinians and Israelis which is marked by a religious divide. We conducted experiments with a representative sample of Muslim Palestinian youth (n = 555), examining whether thinking from the perspective of Allah (God), who is the ultimate arbitrator of religious belief, changes the relative value of Jewish Israelis’ lives (compared with Palestinian lives). Participants were presented with variants of the classic “trolley dilemma,” in the form of stories where a man can be killed to save the lives of five children who were either Jewish Israeli or Palestinian. They responded from their own perspective and from the perspective of Allah. We find that whereas a large proportion of participants were more likely to endorse saving Palestinian children than saving Jewish Israeli children, this proportion decreased when thinking from the perspective of Allah. This finding raises the possibility that beliefs about God can mitigate bias against other groups and reduce barriers to peace.Also, in the journal Psychological Science, Roets et al. consider the case of Singapore, which contradicts:
...numerous empirical studies that have consistently demonstrated the seemingly inextricable link between authoritarianism and negative attitudes about out-groups. Indeed, in the authoritarian mind, minorities are readily perceived as “bad, disruptive, immoral, and deviant” people who do not fit into society... However, what if authoritarians live in a society in which a very strong and established authority most explicitly endorses diversity and multiculturalism, thereby enforcing a social norm that is in direct opposition to authoritarians’ “natural” negative attitudes toward minorities? Over the past 50 years, the Singaporean government (run by the People’s Action Party) has been highly committed to regulating its ethnically diverse society and promoting multiculturalism through a variety of ingenious yet most consequential measures. A prime example is the imposition of strict ethnic quotas in public residential estates
They analyzed data from a questionnaire measuring authoritarianism that was completed by 249 Singaporean students (the target sample; and 245 Belgian students (the comparison group)...the Belgian control group showed the usual negative relationships between authoritarianism and multiculturalism and between authoritarianism and positive attitudes about out-groups, as found in all previous research. In the Singaporean sample, however, there were significant, positive relationships between authoritarianism and multiculturalism and between authoritarianism and positive attitudes about out-groups... [The] results demonstrate that when a strong authority explicitly and relentlessly endorses diversity and multiculturalism, such a perspective can be adopted even (and especially) by people who are intuitively most opposed to diversity.You might also note the comments of Aaron Wendland on the writings of Emmanuel Levinas, after World War II, on deep-seated and often irrational fear of the “other.”
Levinas’s antihistamine for our allergic reactions involves three things: an appeal to the “infinity” in human beings, a detailed description of face-to-face encounters and an account of a basic hospitality that constitutes humanity.
Blog Categories:
culture/politics,
religion,
social cognition
Wednesday, January 20, 2016
Paying attention to your body can increase resilience to stress
Resilience is the ability to rapidly return to normal, both physically and emotionally, after a stressful event. Reynolds points to work by Haase et al., who provide fMRI evidence that in high resilience individuals brain areas receiving signals from the body become more active during stress and supress signals to brain areas that intensify body arousal. Individuals with lower resilience show reduced attention to bodily signals but greater neural processing to aversive bodily perturbations. Here's the Haase et al. abstract:
Highlights
Highlights
• Low resilience individuals are less sensitive to body-relevant information. • Low resilience individuals show an exaggerated brain response to an aversive interoceptive stimulus. • This mismatch between attention to and processing of interoceptive afferents may result in poor adaptation in stressful situations.Abstract
This study examined neural processes of resilience during aversive interoceptive processing. Forty-six individuals were divided into three groups of resilience Low (LowRes), high (HighRes), and normal (NormRes), based on the Connor-Davidson Resilience Scale (2003). Participants then completed a task involving anticipation and experience of loaded breathing during functional magnetic resonance imaging (fMRI) recording. Compared to HighRes and NormRes groups, LowRes self-reported lower levels of interoceptive awareness and demonstrated higher insular and thalamic activation across anticipation and breathing load conditions. Thus, individuals with lower resilience show reduced attention to bodily signals but greater neural processing to aversive bodily perturbations. In low resilient individuals, this mismatch between attention to and processing of interoceptive afferents may result in poor adaptation in stressful situations.
Blog Categories:
brain plasticity,
fear/anxiety/stress
Tuesday, January 19, 2016
Biochemical individuality obsoletes many dietary recommendations and the glycemic index.
An important paper by a Weizmann Institute group has been languishing in my list of potential posts, and Murphy's summary of their work now prods me to pass on their bottom line: the glycemic index, used to rank foods according to their effects on blood sugar, is not really useful, because people differ strikingly in their indivual biochemistries, the way in which they extract energy from different foods. The Weizmann group devised "a machine-learning algorithm that integrates blood parameters, dietary habits, anthropometrics, physical activity, and gut microbiota measured...and showed that it accurately predicts personalized postprandial glycemic response to real-life meals." Companies like Nutrigenomix have begun to offer personalized nutrition assessment based on individual genomics. Here is the Weizmann Inst. info:
Highlights
Highlights
•High interpersonal variability in post-meal glucose observed in an 800-person cohort
•Using personal and microbiome features enables accurate glucose response prediction
•Prediction is accurate and superior to common practice in an independent cohort
•Short-term personalized dietary interventions successfully lower post-meal glucoseSummary
Elevated postprandial blood glucose levels constitute a global epidemic and a major risk factor for prediabetes and type II diabetes, but existing dietary methods for controlling them have limited efficacy. Here, we continuously monitored week-long glucose levels in an 800-person cohort, measured responses to 46,898 meals, and found high variability in the response to identical meals, suggesting that universal dietary recommendations may have limited utility. We devised a machine-learning algorithm that integrates blood parameters, dietary habits, anthropometrics, physical activity, and gut microbiota measured in this cohort and showed that it accurately predicts personalized postprandial glycemic response to real-life meals. We validated these predictions in an independent 100-person cohort. Finally, a blinded randomized controlled dietary intervention based on this algorithm resulted in significantly lower postprandial responses and consistent alterations to gut microbiota configuration. Together, our results suggest that personalized diets may successfully modify elevated postprandial blood glucose and its metabolic consequences.
Monday, January 18, 2016
Posthumanism - the quantification craze and the death of beauty
An eloquent recent essay, “Among the Disrupted” by Leon Wieseltier (pointed to by a Brooks Op-Ed piece), is worth your attention. It opens with a screed on the how journalism has degenerated into a “twittering cacophony of one-liners and promotional announcements” with its force of expression diminishing as its frequency grows, and how culture is being degraded by “the idolatry of metrics and quantification applied to things that cannot be captured by numbers.“ Below are a few clips...
Here is a core sentiment:
The notion that the nonmaterial dimensions of life must be explained in terms of the material dimensions, and that nonscientific understandings must be translated into scientific understandings if they are to qualify as knowledge, is increasingly popular inside and outside the university, where the humanities are disparaged as soft and impractical and insufficiently new. The contrary insistence that the glories of art and thought are not evolutionary adaptations, or that the mind is not the brain, or that love is not just biology’s bait for sex, now amounts to a kind of heresy. So, too, does the view that the strongest defense of the humanities lies not in the appeal to their utility — that literature majors may find good jobs, that theaters may economically revitalize neighborhoods — but rather in the appeal to their defiantly nonutilitarian character, so that individuals can know more than how things work, and develop their powers of discernment and judgment, their competence in matters of truth and goodness and beauty, to equip themselves adequately for the choices and the crucibles of private and public life.Actually, I think the sentiments in the above paragraph are misguided (is Wieseltier really a classical vitalist?), but that's possible grist for another post. Continuing....
...the worldview that is ascendant may be described as posthumanism.
…what is humanism?…The most common understanding of humanism is that it denotes a pedagogy and a worldview. ..The worldview takes many forms: a philosophical claim about the centrality of humankind to the universe, and about the irreducibility of the human difference to any aspect of our animality; a methodological claim about the most illuminating way to explain history and human affairs, and about the essential inability of the natural sciences to offer a satisfactory explanation; a moral claim about the priority, and the universal nature, of certain values, not least tolerance and compassion.
Here is a humanist proposition for the age of Google: The processing of information is not the highest aim to which the human spirit can aspire, and neither is competitiveness in a global economy. The character of our society cannot be determined by engineers.
…machines may be more neutral about their uses than the propagandists and the advertisers want us to believe. We can leave aside the ideology of digitality and its aggressions, and regard the devices as simply new means for old ends. Tradition “travels” in many ways. It has already flourished in many technologies — but only when its flourishing has been the objective. I will give an example from the humanities. The day is approaching when the dream of the democratization of knowledge — Borges’s fantasy of “the total library” — will be realized. Soon all the collections in all the libraries and all the archives in the world will be available to everyone with a screen. Who would not welcome such a vast enfranchisement? But universal accessibility is not the end of the story, it is the beginning. The humanistic methods that were practiced before digitalization will be even more urgent after digitalization, because we will need help in navigating the unprecedented welter. Searches for keywords will not provide contexts for keywords. Patterns that are revealed by searches will not identify their own causes and reasons. The new order will not relieve us of the old burdens, and the old pleasures, of erudition and interpretation.
The persistence of humanism through the centuries, in the face of formidable intellectual and social obstacles, has been owed to the truth of its representations of our complexly beating hearts, and to the guidance that it has offered, in its variegated and conflicting versions, for a soulful and sensitive existence…In a society rife with theories and practices that flatten and shrink and chill the human subject, the humanist is the dissenter. Never mind the platforms. Our solemn responsibility is for the substance.
Friday, January 15, 2016
Homosexuality as a discrete class.
Norris et al. contribute to previous work engaging the question of whether homosexuality has a taxonic structure of categories of individuals with distinct orientation, or whether sexual orientation lies on the sort of continuum suggested by Kinsey and others. Because individuals who report nonheterosexual identities, behavior, and attractions are more likely than heterosexual individuals to meet criteria for a psychiatric disorder their study utilized a National Epidemiologic Survey on Alcohol and Related Conditions. The survey was conducted through personal interviews with one randomly selected adult in each household. Their abstract:
Previous research on the latent structure of sexual orientation has returned conflicting results, with some studies finding a dimensional structure (i.e., ranging quantitatively along a spectrum) and others a taxonic structure (i.e., categories of individuals with distinct orientations). The current study used a sample (N = 33,525) from the National Epidemiologic Survey on Alcohol and Related Conditions (NESARC). A series of taxometric analyses were conducted using three indicators of sexual orientation: identity, behavior, and attraction. These analyses, performed separately for women and men, revealed low-base-rate same-sex-oriented taxa for men (base rate = 3.0% of those sampled) and women (base rate = 2.7%). Generally, taxon membership conferred an increased risk for psychiatric and substance-use disorders. Although taxa were present for men and women, women demonstrated greater sexual fluidity, such that any level of same-sex sexuality conferred taxon membership for men but not for women.
Another installment on anti-aging chemistry - pterostilbene
I thought I would report on my recent meandering into recent work on the putative anti-aging compounds Resveratrol and Pterostilbene, which are found in some fruits, nuts, and vegetables. (I reported my experience with resveratrol in a 2008 MindBlog post.) The meandering started with a glance at Weintraub's piece on health benefits of red wine versus grape juice. She quoted M.I.T. anti-aging researcher Leonard Guarente, who started the company Elysium, which sells Pterostilbene, a resveratrol cousin that is more easily absorbed after oral ingestion:
Poulose et al. have recently reviewed articles on the effects of pterostilbene and resveratrol on brain health and chemistry (motivated readers can obtain a PDF from me, see also the article by Mitteldorf and the website examine.com).
Even after the neutral to negative experience with resveratrol that I reported in the MindBlog post mentioned above, I've just ordered a bit of pterostilbene to see if I experience some of the cognitive effects claimed. I'll report on the experience.
Poulose et al. have recently reviewed articles on the effects of pterostilbene and resveratrol on brain health and chemistry (motivated readers can obtain a PDF from me, see also the article by Mitteldorf and the website examine.com).
Even after the neutral to negative experience with resveratrol that I reported in the MindBlog post mentioned above, I've just ordered a bit of pterostilbene to see if I experience some of the cognitive effects claimed. I'll report on the experience.
Thursday, January 14, 2016
How to be bad together - Punishing pro-social behavior.
I pass on part of a brief essay by Gloria Origgi noting the work of Gächter and Herrmann, who examined positive and negative reciprocity in different groups selected within Switzerland and Russia.:
There is a vast literature showing how direct and indirect reciprocity are important tools for dealing with human cooperation. Many experiments have shown that people use “altruistic punishment” to sustain cooperation, that is, they are willingly to pay without receiving anything back just in order to sanction those who don’t cooperate, and hence promote pro-social behavior.
Yet, Gächter and Herrmann showed in their surprising paper that in some cultures, when people were tested in cooperative games (such as the “public good game”), the people who cooperated were punished, rather than the free-riders.
In some societies, people prefer to act anti-socially and they take actions to make sure that the others do the same! This means that cooperation in societies is not always for the good: you can find cartels of anti-social people who don’t care at all for the common good and prefer to cooperate for keeping a status quo that suits them even if the collective outcome is a mediocre result.
As an Italian with first-hand experience in living in a country where, if you behave well, you are socially and legally sanctioned, this news was exciting, even inspiring … perhaps cooperation is not an inherent virtue of the human species. Perhaps, in many circumstances, we prefer to stay with those who share our selfishness and weaknesses and to avoid pro-social altruistic individuals. Perhaps it's not abnormal to live outside a circle of empathy.
So what’s the scientific “news that stays news”: Cooperation for the collective worse is as widespread as cooperation for a better society!
Wednesday, January 13, 2016
Purity, Disgust and Donald Trump
Following the thread of the past few posts (on the partisan divide and empathy), I want to pass on this must-read article by Thomas Edsall, who summarizes work of academic researchers probing the role of purity/disgust, order/chaos, anger and fear, in the electorate. Here is a great graphic from the article:
Blog Categories:
culture/politics,
morality,
social cognition
Tuesday, January 12, 2016
Our strongest prejudice - partisan hostility
Jonathan Haidt points to the fascinating work by political scientists Shanto Iyengar and Sean Westwood, titled “Fear and Loathing Across Party Lines: New Evidence on Group Polarization.”, who found - in four studies designed to reveal prejudice based on race, gender, religion, or political party or ideology - that cross-partisan prejudice was the largest. For white participants who identified with a party, the cross-partisan effect was about 50 percent larger than the cross-race effect. Haidt points out that This is extremely bad news for America because:
...rising cross-partisan hostility means that Americans increasingly see the other side not just as wrong but as evil, as a threat to the very existence of the nation, according to Pew Research. Americans can expect rising polarization, nastiness, paralysis, and governmental dysfunction for a long time to come...This is extremely bad news for science and universities because universities are usually associated with the left...we can expect increasing hostility from Republican legislators toward universities and the things they desire, including research funding and freedom from federal and state control...This is a warning for the rest of the world because some of the trends that have driven America to this point are occurring in many other countries, including: rising education and individualism (which make people more ideological), rising immigration and ethnic diversity (which reduces social capital and trust), and stagnant economic growth (which puts people into a zero-sum mindset).The situation is made worse by the "motive attribution asymmetry" that I have referenced in a previous post. Both sides of a political divide attribute their own aggressive behavior to love, but the opposite side's to hatred. Millions of Americans believe that their side is basically benevolent while the other side is evil and out to get them.
Monday, January 11, 2016
How learning shapes the empathic brain.
From Hein et al.:
Abstract
Abstract
Deficits in empathy enhance conflicts and human suffering. Thus, it is crucial to understand how empathy can be learned and how learning experiences shape empathy-related processes in the human brain. As a model of empathy deficits, we used the well-established suppression of empathy-related brain responses for the suffering of out-groups and tested whether and how out-group empathy is boosted by a learning intervention. During this intervention, participants received costly help equally often from an out-group member (experimental group) or an in-group member (control group). We show that receiving help from an out-group member elicits a classical learning signal (prediction error) in the anterior insular cortex. This signal in turn predicts a subsequent increase of empathy for a different out-group member (generalization). The enhancement of empathy-related insula responses by the neural prediction error signal was mediated by an establishment of positive emotions toward the out-group member. Finally, we show that surprisingly few positive learning experiences are sufficient to increase empathy. Our results specify the neural and psychological mechanisms through which learning interacts with empathy, and thus provide a neurobiological account for the plasticity of empathic reactions.
Blog Categories:
culture/politics,
mirror neurons,
morality
Friday, January 08, 2016
Virtual reality going mainstream will enhance our understanding of consciousness.
These clips from Thomas Metzinger are fascinating:
2016 will be the year in which VR finally breaks through at the mass consumer level. What is more, users will soon be enabled to toggle between virtual, augmented, and substitutional reality, experiencing virtual elements intermixed with their “actual” physical environment or an omnidirectional video feed giving them the illusion of being in a different location in space and/or time, while insight may not always be preserved. Oculus Rift, Zeiss VR One, Sony PlayStation VR, HTC Vive, Samsung’s Galaxy Gear VR or Microsoft’s HoloLens are just the very beginning...
The real news, however, may be that the general public will gradually acquire a new and intuitive understanding of what their very own conscious experience really is and what it always has been. VR is the representation of possible worlds and possible selves, with the aim of making them appear as real as possible—ideally, by creating a subjective sense of “presence” in the user. Interestingly, some of our best theories of the human mind and conscious experience describe it in a very similar way: Leading theoretical neurobiologists like Karl Friston and eminent philosophers like Jakob Hohwy and Andy Clark describe it as the constant creation of internal models of the world, virtual neural representations of reality which express probability density functions and work by continuously generating hypotheses about the hidden causes of sensory input, minimizing their prediction error. In 1995, Finnish philosopher Antti Revonsuo already pointed out how conscious experience exactly is a virtual model of the world, a dynamic internal simulation, which in standard situations cannot be experienced as a virtual model because it is phenomenally transparent—we “look through it” as if we were in direct and immediate contact with reality. What is historically new, and what creates not only novel psychological risks but also entirely new ethical and legal dimensions, is that one virtual reality gets ever more deeply embedded into another virtual reality: The conscious mind of human beings, which has evolved under very specific conditions and over millions of years, now gets causally coupled and informationally woven into technical systems for representing possible realities. Increasingly, consciousness is not only culturally and socially embedded, but also shaped by a specific technological niche that, over time, quickly acquires rapid, autonomous dynamics and ever new properties. This creates a complex convolution, a nested form of information flow in which the biological mind and its technological niche influence each other in ways we are just beginning to understand.
Blog Categories:
consciousness,
culture/politics,
future,
futures
Thursday, January 07, 2016
How our brains change during the day.
I am reminded what a rigid daily schedule my body keeps and expects every time I vary my routine slightly (changing a meal time, exercise, happy hour, bedtime)...my body doesn't like it, feels off kilter. Travel of the sort I've been doing over the past month is a huge disrupter. I'm also clear that any demanding and analytical thinking I might want to do should happen before noontime, because by 4 p.m. (a low blood sugar time for the body), my mind has become very lazy.
McClung and collaborators have done the fascinating experiment of looking at gene expression during the day in brain areas important in learning, memory, and emotion, in 146 young and old brains, finding differences on aging. The brains were from subjects who had died suddenly, as in a car accident. They built on the work of Akil and collaborators who earlier had shown the activity of ~1000 genes varies in a daily pattern that allows the time of death to be predicted within an hour. That pattern was disrupted in people with major depressive disorders. From McClung's group:
Significance
McClung and collaborators have done the fascinating experiment of looking at gene expression during the day in brain areas important in learning, memory, and emotion, in 146 young and old brains, finding differences on aging. The brains were from subjects who had died suddenly, as in a car accident. They built on the work of Akil and collaborators who earlier had shown the activity of ~1000 genes varies in a daily pattern that allows the time of death to be predicted within an hour. That pattern was disrupted in people with major depressive disorders. From McClung's group:
Significance
Circadian rhythms are important in nearly all processes in the brain. Changes in rhythms that come with aging are associated with sleep problems, problems with cognition, and nighttime agitation in elderly people. In this manuscript, we identified transcripts genome-wide that have a circadian rhythm in expression in human prefrontal cortex. Moreover, we describe how these rhythms are changed during normal human aging. Interestingly, we also identified a set of previously unidentified transcripts that become rhythmic only in older individuals. This may represent a compensatory clock that becomes active with the loss of canonical clock function. These studies can help us to develop therapies in the future for older people who suffer from cognitive problems associated with a loss of normal rhythmicity.Abstract
With aging, significant changes in circadian rhythms occur, including a shift in phase toward a “morning” chronotype and a loss of rhythmicity in circulating hormones. However, the effects of aging on molecular rhythms in the human brain have remained elusive. Here, we used a previously described time-of-death analysis to identify transcripts throughout the genome that have a significant circadian rhythm in expression in the human prefrontal cortex [Brodmann’s area 11 (BA11) and BA47]. Expression levels were determined by microarray analysis in 146 individuals. Rhythmicity in expression was found in ∼10% of detected transcripts (P less than 0.05). Using a metaanalysis across the two brain areas, we identified a core set of 235 genes (q less than 0.05) with significant circadian rhythms of expression. These 235 genes showed 92% concordance in the phase of expression between the two areas. In addition to the canonical core circadian genes, a number of other genes were found to exhibit rhythmic expression in the brain. Notably, we identified more than 1,000 genes (1,186 in BA11; 1,591 in BA47) that exhibited age-dependent rhythmicity or alterations in rhythmicity patterns with aging. Interestingly, a set of transcripts gained rhythmicity in older individuals, which may represent a compensatory mechanism due to a loss of canonical clock function. Thus, we confirm that rhythmic gene expression can be reliably measured in human brain and identified for the first time (to our knowledge) significant changes in molecular rhythms with aging that may contribute to altered cognition, sleep, and mood in later life.
Subscribe to:
Posts (Atom)