Since the origins of agriculture, the scale of human cooperation and societal complexity has dramatically expanded. This fact challenges standard evolutionary explanations of prosociality because well-studied mechanisms of cooperation based on genetic relatedness, reciprocity and partner choice falter as people increasingly engage in fleeting transactions with genetically unrelated strangers in large anonymous groups. To explain this rapid expansion of prosociality, researchers have proposed several mechanisms. Here we focus on one key hypothesis: cognitive representations of gods as increasingly knowledgeable and punitive, and who sanction violators of interpersonal social norms, foster and sustain the expansion of cooperation, trust and fairness towards co-religionist strangers. We tested this hypothesis using extensive ethnographic interviews and two behavioural games designed to measure impartial rule-following among people (n = 591, observations = 35,400) from eight diverse communities from around the world: (1) inland Tanna, Vanuatu; (2) coastal Tanna, Vanuatu; (3) Yasawa, Fiji; (4) Lovu, Fiji; (5) Pesqueiro, Brazil; (6) Pointe aux Piments, Mauritius; (7) the Tyva Republic (Siberia), Russia; and (8) Hadzaland, Tanzania. Participants reported adherence to a wide array of world religious traditions including Christianity, Hinduism and Buddhism, as well as notably diverse local traditions, including animism and ancestor worship. Holding a range of relevant variables constant, the higher participants rated their moralistic gods as punitive and knowledgeable about human thoughts and actions, the more coins they allocated to geographically distant co-religionist strangers relative to both themselves and local co-religionists. Our results support the hypothesis that beliefs in moralistic, punitive and knowing gods increase impartial behaviour towards distant co-religionists, and therefore can contribute to the expansion of prosociality.
This blog reports new ideas and work on mind, brain, behavior, psychology, and politics - as well as random curious stuff. (Try the Dynamic Views at top of right column.)
Wednesday, February 24, 2016
Moralistic gods enhance sociality.
Purzycki et al. combine laboratory experiments, cross-cultural fieldwork, and analysis of the historical record to propose that belief in judgmental deities was key to the cooperation needed to build and sustain large, complex societies.
Tuesday, February 23, 2016
Creative cognition and brain network dynamics.
Beaty et al. do a review that notes trends in recent neuroimaging studies suggesting that creative cognition involves increased cooperation of the default and executive control networks of our brain (Motivated readers can obtain the article from me.)
Several recent neuroimaging studies have found that creative cognition involves increased cooperation of the default and executive control networks, brain systems linked to self-generated thought and cognitive control.
Default–control network interactions occur during cognitive tasks that involve the generation and evaluation of creative ideas. This pattern of brain network connectivity has been reported across domain-general creative problem solving (e.g., divergent thinking) and domain-specific artistic performance (e.g., poetry composition, musical improvisation, and visual art production).
Default network activity during creative cognition appears to reflect the spontaneous generation of candidate ideas, or potentially useful information derived from long-term memory.
The control network may couple with the default network during idea generation or evaluation to constrain cognition to meet specific task goals.
Dorsolateral Prefrontal Cortex Connectivity During Musical Improvisation. The right dorsolateral prefrontal cortex (DLPFC; green) shows differential connectivity as a function of task goals during musical improvisation in professional pianists. (A) Functional connectivity associated with the goal of using specific sets of piano keys; brain maps show increased coupling between the right DLPFC and motor regions (yellow, e.g., dorsal pre-motor area and the pre-supplementary motor area). (B) Functional connectivity associated with the goal of expressing specific emotions; brain maps show increased coupling between the right DLPFC and default network regions [blue, e.g., medial prefrontal cortex (MPFC), posterior cingulate cortex (PCC), and bilateral inferior parietal lobule (IPL)].
Monday, February 22, 2016
A mindfulness meditation intervention enhances connectivity of brain executive and default modes and lowers inflammation markers.
Creswell et al. recruited 35 stressed-out adult job-seekers, getting half to participate in an intensive three-day mindfulness meditation retreat while the other half completed a three day relaxation retreat program without the mindfulness component. Brain scans and blood samples were obtained before and four months after the program. The result was that mindfulness meditation correlated with reduced blood levels of interleukin-6, a marker of stress and inflammation, and increased functional connectivity between the participants’ resting default mode network and areas in the dorsolateral prefrontal cortex important to attention and executive control. Neither of these changes were seen in participants who received only the relaxation training. The suggestion is that the brain changes cause the decrease in inflammatory markers. Here are some clips of context from their introduction:
Mindfulness meditation training programs, which train receptive attention and awareness to one’s present moment experience, have been shown to improve a broad range of stress-related psychiatric and physical health outcomes in initial randomized controlled trials...recent well-controlled studies indicate that mindfulness meditation training may reduce markers of inflammation (C Reactive Protein, Interleukin-6 (IL-6), neurogenic inflammation) in stressed individuals.
One possibility is that mindfulness meditation training alters resting state functional connectivity (rsFC) of brain networks implicated in mind wandering (the Default Mode Network, DMN) and executive control (the Executive Control Network, EC), which in turn improves emotion regulation, stress resilience, and stress-related health outcomes in at-risk patient populations.... a cross-sectional study (N=25) showed that advanced mindfulness meditation practitioners had increased functional connectivity of a key hub in the default mode network (DMN) (i.e., posterior cingulate cortex) with regions considered to be important in top down executive control (EC) (dorsolateral prefrontal cortex, dorsal ACC), both at rest and during a guided mindfulness meditation practice. This coupling of one’s DMN at rest with regions of the EC network may be important for emotion regulation and stress resilience effects, as greater activation and functional connectivity of EC regions, such as the dlPFC, is associated with reduced pain, negative affect, and stress.
Here we provide the first experimental test of whether an intensive 3-day mindfulness meditation training intervention (relative to a relaxation training intervention) alters DMN connectivity and circulating IL-6 in a high stress unemployed job-seeking community sample. IL-6 is an established clinical health biomarker that is elevated in high stress populations and is associated with elevated cardiovascular disease and mortality risk... unemployment is a well-known chronic stressor that can foster a loss of control, helplessness, and financial setbacks, and unemployment is associated with elevated inflammation. Building on initial cross-sectional evidence (17), we hypothesized that mindfulness meditation training would increase rsFC between the DMN and regions implicated in attention and executive control (dlPFC and dACC). Moreover, we tested whether mindfulness meditation training (relative to relaxation training) decreased circulating IL-6 at 4-month follow up, and whether pre-post intervention increases in DMN-dlPFC rsFC mediated IL-6 improvements at 4-month follow-up.
Blog Categories:
fear/anxiety/stress,
meditation,
mindfulness
Friday, February 19, 2016
Forming Beliefs: Why Valence Matters.
Sharot and Garrett do a review article that puts their work on how we see through rose colored glasses, mentioned in a previous MindBlog post, in perspective (Motivated readers can obtain a copy from me).
One of the most salient attributes of information is valence: whether a piece of news is good or bad. Contrary to classic learning theories, which implicitly assume beliefs are adjusted similarly regardless of valence, we review evidence suggesting that different rules and mechanisms underlie learning from desirable and undesirable information. For self-relevant beliefs this asymmetry generates a positive bias, with significant implications for individuals and society. We discuss the boundaries of this asymmetry, characterize the neural system supporting it, and describe how changes in this circuit are related to individual differences in behavior.
Thursday, February 18, 2016
A neural crossroads of psychiatric illnesses as a target for therapeutic brain stimulation
I want pass on this open source article by Downar et al. which contains some useful graphics for illustrating their point about a central role for anterior cingulate cortex and anterior insula in most psychiatric disorders. Here is their abstract:
Recent meta-analyses of structural and functional neuroimaging studies are converging on a collective core of brain regions affected across most psychiatric disorders, centered on the dorsal anterior cingulate cortex (dACC) and anterior insula. These nodes correspond well to an anterior cingulo-insular (aCIN) or ‘salience’ network, and stand at a crossroads within the functional architecture of the brain, acting as a switch to deploy other major functional networks according to motivational demands and environmental constraints. Therefore, disruption of these ‘linchpin’ areas may be disproportionately disabling, even when other networks remain intact. These regions may represent promising targets for a new generation of anatomically directed brain stimulation treatments. Here, we review the potential of the psychiatric core areas as targets for therapeutic brain stimulation in psychiatric disease.
Blog Categories:
brain plasticity,
fear/anxiety/stress
Wednesday, February 17, 2016
Finally...a brain area specialized for music has been found.
Norman-Haignere, Kanwisher, and McDermott have devised a new method to computationally dissect scans from functional magnetic resonance imaging of the brain to reveal an area within the major crevice, or sulcus, of the auditory cortex in the temporal lobe just above the ears that responds to music (any kind of music, drumming, whistling, pop songs, rap, anything melodic or rhythmic) independent of general properties of sound like pitch, spectrotemporal modulation, and frequency. They also found an area for speech not explainable by standard acoustic features.
It is possible that music sensitivity is more ancient and fundamental to the human brain than speech perception, with speech having evolved from music.
It is possible that music sensitivity is more ancient and fundamental to the human brain than speech perception, with speech having evolved from music.
The organization of human auditory cortex remains unresolved, due in part to the small stimulus sets common to fMRI studies and the overlap of neural populations within voxels. To address these challenges, we measured fMRI responses to 165 natural sounds and inferred canonical response profiles (“components”) whose weighted combinations explained voxel responses throughout auditory cortex. This analysis revealed six components, each with interpretable response characteristics despite being unconstrained by prior functional hypotheses. Four components embodied selectivity for particular acoustic features (frequency, spectrotemporal modulation, pitch). Two others exhibited pronounced selectivity for music and speech, respectively, and were not explainable by standard acoustic features. Anatomically, music and speech selectivity concentrated in distinct regions of non-primary auditory cortex. However, music selectivity was weak in raw voxel responses, and its detection required a decomposition method. Voxel decomposition identifies primary dimensions of response variation across natural sounds, revealing distinct cortical pathways for music and speech.
Tuesday, February 16, 2016
A 50 year misunderstanding of how we decide to initiate action - our intuition is valid
A commonly accepted assumption (that underlies all of the essays at dericbownds.net and many mindblog posts) is that a line of experiments starting in the early 1980s with Benjamin Libet’s work has demonstrated that our brains initiate an action 300 sec or more before our conscious ‘urge’ to move. It is as if we are ‘late’ to action because separate pathways are initiating the actual movement and our delayed ‘intention’ to move. This counterintuitive paradox has generated vigorous controversy for many years, debate of its implications regarding ‘free will’, etc.
Schurger et al. point out that the ‘readiness potential’ (RP, a slow build-up of scalp electrical potential preceding the onset of subjectively spontaneous voluntary movements ) discovered 50 years ago was presumed to reflect the consequence of decision process in the brain (‘the electro-physiological sign of planning, preparation, and initiation of volitional acts’). A new generation of experiments is now suggesting that brain activity preceding spontaneous voluntary movements (SVMs) "may reflect the ebb and flow of background neuronal noise, rather than the outcome of a specific neural event corresponding to a ‘decision’ to initiate movement... [Several studies] have converged in showing that bounded-integration processes, which involve the accumulation of noisy evidence until a decision threshold is reached, offer a coherent and plausible explanation for the apparent pre-movement build-up of neuronal activity."
Schurger et al. point out that the ‘readiness potential’ (RP, a slow build-up of scalp electrical potential preceding the onset of subjectively spontaneous voluntary movements ) discovered 50 years ago was presumed to reflect the consequence of decision process in the brain (‘the electro-physiological sign of planning, preparation, and initiation of volitional acts’). A new generation of experiments is now suggesting that brain activity preceding spontaneous voluntary movements (SVMs) "may reflect the ebb and flow of background neuronal noise, rather than the outcome of a specific neural event corresponding to a ‘decision’ to initiate movement... [Several studies] have converged in showing that bounded-integration processes, which involve the accumulation of noisy evidence until a decision threshold is reached, offer a coherent and plausible explanation for the apparent pre-movement build-up of neuronal activity."
SVMs rely on the same neural decision mechanism used for perceptual decisions – integration to bound – except that in this case there is no specific external sensory evidence to integrate. In particular, when actions are initiated spontaneously rather than in response to a sensory cue, the process of integration to bound is dominated by ongoing stochastic fluctuations in neural activity that influence the precise moment at which the decision threshold is reached. In this context, time locking to movement onset means time locking to crests in these temporally autocorrelated background fluctuations, which results in the appearance of a slow, nonlinear build-up in the average. This in turn gives the natural but erroneous impression of a goal-directed brain process corresponding to the ‘cerebral initiation of a spontaneous voluntary act’The Philosophical Implications:
Many have found Libet et al.’s results so striking because they appear to clash with our commonsense view of action initiation. However, the novel interpretation of the RP that the stochastic decision model provides actually suggests a close correspondence between the two. When one forms an intention to act, one is significantly disposed to act but not yet fully committed. The commitment comes when one finally decides to act. The stochastic decision model reveals a remarkably similar picture on the neuronal level, with the decision to act being a threshold-crossing neural event that is preceded by a neural tendency toward this event.
In addition, dropping the problematic theoretical assumption that a decision to act cannot occur without being conscious also helps to dispel the apparent air of ‘paradox’ surrounding these findings. As with other types of mental occurrence, the decision to initiate an action can occur before one is aware of it. So we can identify the neural initiating event with a decision that we may become aware of only a brief instant later. All this leaves our commonsense picture largely intact.
Finally, distinguishing between the decision to act and the earlier forming of an intention fits well with the distinction between a prediction and a forecast. If our concern is merely forecasting, what is relevant is the less-committed event of an intention's forming, which we identify with the neural tendency. If our concern is prediction, we should focus on the later event of deciding, which we identify with the crossing of the threshold.Added Note: Those interested in this vein of work might check Schultze-Kraft et al.:
In humans, spontaneous movements are often preceded by early brain signals. One such signal is the readiness potential (RP) that gradually arises within the last second preceding a movement. An important question is whether people are able to cancel movements after the elicitation of such RPs, and if so until which point in time. Here, subjects played a game where they tried to press a button to earn points in a challenge with a brain–computer interface (BCI) that had been trained to detect their RPs in real time and to emit stop signals. Our data suggest that subjects can still veto a movement even after the onset of the RP. Cancellation of movements was possible if stop signals occurred earlier than 200 ms before movement onset, thus constituting a point of no return.
Monday, February 15, 2016
Our eye movements are coupled to our heartbeats.
A fascinating finding by Ohl et al. that the darting about of our eyes (saccades) during visual search is coupled to our heart rate (the R-R interval), proving a powerful influence of body on visuomotor functioning.
ABSTRACT
SIGNIFICANCE STATEMENT
ABSTRACT
During visual fixation, the eye generates microsaccades and slower components of fixational eye movements that are part of the visual processing strategy in humans. Here, we show that ongoing heartbeat is coupled to temporal rate variations in the generation of microsaccades. Using coregistration of eye recording and ECG in humans, we tested the hypothesis that microsaccade onsets are coupled to the relative phase of the R-R intervals in heartbeats. We observed significantly more microsaccades during the early phase after the R peak in the ECG. This form of coupling between heartbeat and eye movements was substantiated by the additional finding of a coupling between heart phase and motion activity in slow fixational eye movements; i.e., retinal image slip caused by physiological drift. Our findings therefore demonstrate a coupling of the oculomotor system and ongoing heartbeat, which provides further evidence for bodily influences on visuomotor functioning.
SIGNIFICANCE STATEMENT
In the present study, we show that microsaccades are coupled to heartbeat. Moreover, we revealed a strong modulation of slow eye movements around the R peak in the ECG. These results suggest that heartbeat as a basic physiological signal is related to statistical modulations of fixational eye movements, in particular, the generation of microsaccades. Therefore, our findings add a new perspective on the principles underlying the generation of fixational eye movements. Importantly, our study highlights the need to record eye movements when studying the influence of heartbeat in neuroscience to avoid misinterpretation of eye-movement-related artifacts as heart-evoked modulations of neural processing.
Friday, February 12, 2016
What makes political leaders influential?
Brinke et al. examine whether politicians who ascend to influence have been driven by Aristotelian or Machiavellian values (i.e., virtue versus vice). Is political influence was to be found in virtuous practices such as temperance, courage, kindness, and humility; or, does it require force, fraud, manipulation, and strategic violence? Hmmmm....where would we place Ted Cruz and Donald Trump with regard to these qualities?
What qualities make a political leader more influential or less influential? Philosophers, political scientists, and psychologists have puzzled over this question, positing two opposing routes to political power—one driven by human virtues, such as courage and wisdom, and the other driven by vices, such as Machiavellianism and psychopathy. By coding nonverbal behaviors displayed in political speeches, we assessed the virtues and vices of 151 U.S. senators. We found that virtuous senators became more influential after they assumed leadership roles, whereas senators who displayed behaviors consistent with vices—particularly psychopathy—became no more influential or even less influential after they assumed leadership roles. Our results inform a long-standing debate about the role of morality and ethics in leadership and have important implications for electing effective government officials. Citizens would be wise to consider a candidate’s virtue in casting their votes, which might increase the likelihood that elected officials will have genuine concern for their constituents and simultaneously promote cooperation and progress in government.
Thursday, February 11, 2016
Linking sustained attention to brain connectivity.
Rosenberg et al. find that strengths of a specific set of brain connections - even when estimated from resting state data collected when subjects are not carrying out any explicit task - can be used to predict a subject's attention ability with high accuracy. A large set of connections is involved during successful attention, and a different large set correlates with lack of attention.
Although attention plays a ubiquitous role in perception and cognition, researchers lack a simple way to measure a person's overall attentional abilities. Because behavioral measures are diverse and difficult to standardize, we pursued a neuromarker of an important aspect of attention, sustained attention, using functional magnetic resonance imaging. To this end, we identified functional brain networks whose strength during a sustained attention task predicted individual differences in performance. Models based on these networks generalized to previously unseen individuals, even predicting performance from resting-state connectivity alone. Furthermore, these same models predicted a clinical measure of attention—symptoms of attention deficit hyperactivity disorder—from resting-state connectivity in an independent sample of children and adolescents. These results demonstrate that whole-brain functional network strength provides a broadly applicable neuromarker of sustained attention.
Functional connections predicting gradCPT performance and ADHD-RS scores. (gradCPT is a test of sustained attention and inhibition that produces a range of behavior across healthy participants, ADHD-RS is a clinical measure of attention deficit hyperactivity disorder.)
Wednesday, February 10, 2016
Everybody's a critic. And that's how it should be.
An article with the title of this post by NYTimes movie critic A.O. Scott is well worth a read. Here are his noble closing sentiments:
...criticism remains an indispensable activity. The making of art — popular or fine, abstruse or accessible, sacred or profane — is one of the glories of our species. We are uniquely endowed with the capacity to fashion representations of the world and our experience in it, to tell stories and draw pictures, to organize sound into music and movement into dance. Just as miraculously, we have the ability, even the obligation, to judge what we have made, to argue about why we are moved, mystified, delighted or bored by any of it. At least potentially, we are all artists. And because we have the ability to recognize and respond to the creativity of others, we are all, at least potentially, critics, too.
...It’s the mission of art to free our minds, and the task of criticism to figure out what to do with that freedom. That everyone is a critic means that we are each capable of thinking against our own prejudices, of balancing skepticism with open-mindedness, of sharpening our dulled and glutted senses and battling the intellectual inertia that surrounds us. We need to put our remarkable minds to use and to pay our own experience the honor of taking it seriously.
The real culture war (the one that never ends) is between the human intellect and its equally human enemies: sloth, cliché, pretension, cant. Between creativity and conformity, between the comforts of the familiar and the shock of the new. To be a critic is to be a soldier in this fight, a defender of the life of art and a champion of the art of living.
It’s not just a job, in other words.
Tuesday, February 09, 2016
Fantasies of the Future
Following up on my Jan. 28 post on Schwab's book "The Fourth Industrial Revolution" I thought I would pass on some speculations in the appendix of the book on the technology shifts enabled by digital connectivity and software technologies that might "fundamentally change society by 2025" (Really? Gimme a break...).
Wearable internet
Wearable internet
Examples: A baby tech-enabled onesie which tracks babies' breathing, body movements, sleep patterns and quality and transmits that information in real time to a smartphone app. A Ralph Lauren PoloTech shirt with silver fibers woven directly into the fabric that read heart rate and breathing depth and balance, as well as other key metrics, which are streamed to computer or smartphone via a detachable, Bluetooth-enabled black box. A sensor that collects data about multiple chemicals in body sweat.Implantable Technologies - Pacemakers and cochlear implants represent a beginning.
Smart tattoos and other unique chips could help with identification and location. Implanted devices will likely also help to communicate thoughts normally expressed verbally through a “built-in” smart phone, and potentially unexpressed thoughts or moods by reading brainwaves and other signals. (See "Top ten wearables soon to be in your body.")Vision as the New Interface
Glasses are already on the market today (not just produced by Google) that can: – Allow you to freely manipulate a 3D object, enabling it to be moulded like clay – Provide all the extended live information you need when you see something, in the same way the brain functions – Prompt you with an overlay menu of the restaurant you pass by – Project picture or video on any piece of paper. (see 10 Forthcoming Augmented Reality & Smart Glasses You Can Buy.)The internet of and for Things
It is economically feasible to connect literally anything to the internet. Intelligent sensors are already available at very competitive prices. All things will be smart and connected to the internet, enabling greater communication and new data-driven services based on increased analytics capabilities...The Ford GT has 10 million lines of computer code in it. And, see The connected home.I am fatiguing,...the list continues with mention of smart cities, driverless cars, big data for decisions, artificial intelligence for decision making (and the decimation of current white-collar jobs), the sharing economy, 3D printing for manufacturing and health, personalized medicine, designer humans, neurotechnologies, brain technologies.....etc.
Blog Categories:
culture/politics,
future,
futures,
human evolution,
technology
Monday, February 08, 2016
Our digital presence.
Here is an interesting tally: Active users of social media sites compared with the populations of the world's largest countries. if social media sites were countries, Facebook would be the world’s largest country with more active accounts than there are people in China. Twitter would rank 4th with twice the “population” of the USA and Instagram would round out the Top 10.
There are ~7.4 billion people in the world, ~43% are connected to the internet; ~4 billion, mostly in the developing world, lack internet access. Most pundits expect that by 2025, digital access will have spread to 80% of all people.
There are ~7.4 billion people in the world, ~43% are connected to the internet; ~4 billion, mostly in the developing world, lack internet access. Most pundits expect that by 2025, digital access will have spread to 80% of all people.
Blog Categories:
culture/politics,
future,
futures,
human evolution,
language
Friday, February 05, 2016
Consciousness as the product of carefully balanced chaos.
Tagliafuochi and collaborators have provided more evidence that our experience of consciousness and reality might result from a delicate balance or critical level of connectivity between brain networks in which the brain explores the maximum number of unique pathways to generate meaning. Consciousness slips away if there is too much or too little connectivity. Their abstract:
Loss of cortical integration and changes in the dynamics of electrophysiological brain signals characterize the transition from wakefulness towards unconsciousness. In this study, we arrive at a basic model explaining these observations based on the theory of phase transitions in complex systems. We studied the link between spatial and temporal correlations of large-scale brain activity recorded with functional magnetic resonance imaging during wakefulness, propofol-induced sedation and loss of consciousness and during the subsequent recovery. We observed that during unconsciousness activity in frontothalamic regions exhibited a reduction of long-range temporal correlations and a departure of functional connectivity from anatomical constraints. A model of a system exhibiting a phase transition reproduced our findings, as well as the diminished sensitivity of the cortex to external perturbations during unconsciousness. This framework unifies different observations about brain activity during unconsciousness and predicts that the principles we identified are universal and independent from its causes.
Thursday, February 04, 2016
Inequality and facing the future.
A recent piece by Arianna Huffington (The fourth industrial revolution meets the sleep revolution) suggests to me one aspect of yet another driver of future inequality beyond the declining share of income going to labor compared with capital. Two clear castes of people are emerging, those who can adapt psychologically to the mind-numbing complexity of the emerging digital environment by optimizing their minds and bodies (meditation, sleep, exercise, diet, etc.) and those of lower socioeconomic status who start off disadvantaged (see yesterday’s post) and find that their only mental refuge is some form of fundamentalism, a closing rather than opening of their minds (cf. Donald Trump and Ted Cruz supporters.) Given the chaos and disruption being visited on traditional political and economic arrangements by the fusion of digital, biological, and physical advances - the internet of things meeting the smart factory meeting synthetic biology - how are all humans going to be able share in an understanding and shaping of these changes in a way that keeps human beings at the center?
Wednesday, February 03, 2016
The neuroscience of poverty.
This open source review article by Alla Katsnelson is sobering, and worth a read. The major foci in the brain that appear to show disparities in poor children are the hippocampus and frontal lobe. I pass on this graphic illustrating the decline in total brain gray matter (nerve cell) volume in young children of middle and low socioeconomic status individuals.
Tuesday, February 02, 2016
The burden of obesity revisited.
A previous mindblog post has noted the obesity paradox, generated by data suggesting that fat people may live longer. Stokes and Preston,however, note an important distinction that may evaporate the apparent paradox:
Analyses of the relation between obesity and mortality typically evaluate risk with respect to weight recorded at a single point in time. As a consequence, there is generally no distinction made between nonobese individuals who were never obese and nonobese individuals who were formerly obese and lost weight. We introduce additional data on an individual’s maximum attained weight and investigate four models that represent different combinations of weight at survey and maximum weight. We use data from the 1988–2010 National Health and Nutrition Examination Survey, linked to death records through 2011, to estimate parameters of these models. We find that the most successful models use data on maximum weight, and the worst-performing model uses only data on weight at survey. We show that the disparity in predictive power between these models is related to exceptionally high mortality among those who have lost weight, with the normal-weight category being particularly susceptible to distortions arising from weight loss. These distortions make overweight and obesity appear less harmful by obscuring the benefits of remaining never obese. Because most previous studies are based on body mass index at survey, it is likely that the effects of excess weight on US mortality have been consistently underestimated.
Monday, February 01, 2016
The 8 second attention span.
Wow, Egan notes that our average attention span has fallen to eight seconds, down from 12 in the year 2000. (That makes it gratifying that the average amount of time spent by someone on this website is over 4 minutes.)
...a quote from Satya Nadella, the chief executive officer of Microsoft... “The true scarce commodity of the near future will be human attention.”...Putting aside Microsoft’s self-interest in promoting quick-flash digital ads with what may be junk science, there seems little doubt that our devices have rewired our brains. We think in McNugget time. The trash flows, unfiltered, along with the relevant stuff, in an eternal stream. And the last hit of dopamine only accelerates the need for another one.
You see it in the press, the obsession with mindless listicles that have all the staying power of a Popsicle. You see it in our politics, with fear-mongering slogans replacing anything that requires sustained thought. And the collapse of a fact-based democracy, where, for example, 60 percent of Trump supporters believe Obama was born in another country, has to be a byproduct of the pick-and-choose news from the buffet line of our screens.Egan suggests that gardening and deep reading of biographies are useful antidotes.
Friday, January 29, 2016
Fourth Industrial Revolution?? Maybe not.....
Just after starting my second slog through Schwab's 'the fourth industrial revolution' noted in yesterday's post, I see Paul Krugman's review of 'The Rise and Fall of American Growth' by Robert Gordon. Gordon's magnum opus suggests that Schwab's futurism is overblown. (see also Thomas Edsall's excellent piece on the divide and debate between optimistic and pessimistic economists. Also, see this Slate article that chronicles how many times over the past 75 years the term "fourth industrial revolution" has been fetched up to describe a recent or coming advance.) Some clips from Klugman's review:
...[Gordon] has argued that the I.T. revolution is less important than any one of the five Great Inventions that powered economic growth from 1870 to 1970: electricity, urban sanitation, chemicals and pharmaceuticals, the internal combustion engine and modern communication.
What happened between 1870 and 1940, he argues, and I would agree, is what real transformation looks like. Any claims about current progress need to be compared with that baseline to see how they measure up.
And it’s hard not to agree with him that nothing that has happened since is remotely comparable. Urban life in America on the eve of World War II was already recognizably modern; you or I could walk into a 1940s apartment, with its indoor plumbing, gas range, electric lights, refrigerator and telephone, and we’d find it basically functional. We’d be annoyed at the lack of television and Internet — but not horrified or disgusted.
By contrast, urban Americans from 1940 walking into 1870-style accommodations — which they could still do in the rural South — were indeed horrified and disgusted. Life fundamentally improved between 1870 and 1940 in a way it hasn’t since.One of Gordon's arguments against the techno-optimists is that:
...genuinely major innovations normally bring about big changes in business practices, in what workplaces look like and how they function. And there were some changes along those lines between the mid-1990s and the mid-2000s — but not much since, which is evidence for Gordon’s claim that the main impact of the I.T. revolution has already happened.Techno-futurists would argue strongly against this, citing the rise of the sharing economy, entities like Airbnb and Uber, and changes in the workplace from hierarchical to distributed organization.
Gordon suggests that the future is all too likely to be marked by stagnant living standards for most Americans, because the effects of slowing technological progress will be reinforced by a set of “headwinds”: rising inequality, a plateau in education levels, an aging population and more.
It’s a shocking prediction for a society whose self-image, arguably its very identity, is bound up with the expectation of constant progress. And you have to wonder about the social and political consequences of another generation of stagnation or decline in working-class incomes.
Of course, Gordon could be wrong: Maybe we’re on the cusp of truly transformative change, say from artificial intelligence or radical progress in biology (which would bring their own risks). But he makes a powerful case. Perhaps the future isn’t what it used to be.
Blog Categories:
culture/politics,
future,
futures,
technology
Thursday, January 28, 2016
Predicting the Future: The Fourth Industrial Revolution
Klaus Schwab, the guy who runs the annual Davos Switzerland World Economic Forum of “very important people” in the world, has generated a book with the title of this year's meeting, “The Fourth Industrial Revolution,” about a future that is both terrifying and optimistic. Despite its gobbledegook, committee-speak, and bullet points, it provides a lot of fascinating information and is well worth a look. After doing a quick read-through, I’m starting a second more careful read and clicking through to many of the references (the kindle version is handy for this.)
The first industrial revolution spanned from about 1760 to around 1840. Triggered by the construction of railroads and the invention of the steam engine, it ushered in mechanical production. The second industrial revolution, which started in the late 19th century and into the early 20th century, made mass production possible, fostered by the advent of electricity and the assembly line. The third industrial revolution began in the 1960s. It is usually called the computer or digital revolution because it was catalysed by the development of semiconductors, mainframe computing (1960s), personal computing (1970s and 80s) and the internet (1990s). Mindful of the various definitions and academic arguments used to describe the first three industrial revolutions, I believe that today we are at the beginning of a fourth industrial revolution. It began at the turn of this century and builds on the digital revolution. It is characterized by a much more ubiquitous and mobile internet, by smaller and more powerful sensors that have become cheaper, and by artificial intelligence and machine learning.
I am well aware that some academics and professionals consider the developments that I am looking at as simply a part of the third industrial revolution. Three reasons, however, underpin my conviction that a fourth and distinct revolution is underway:
Velocity: Contrary to the previous industrial revolutions, this one is evolving at an exponential rather than linear pace. This is the result of the multifaceted, deeply interconnected world we live in and the fact that new technology begets newer and ever more capable technology.
Breadth and depth: It builds on the digital revolution and combines multiple technologies that are leading to unprecedented paradigm shifts in the economy, business, society, and individually. It is not only changing the “what” and the “how” of doing things but also “who” we are.
Systems Impact: It involves the transformation of entire systems, across (and within) countries, companies, industries and society as a whole.
Blog Categories:
culture/politics,
future,
futures,
technology
Subscribe to:
Posts (Atom)