Showing posts with label predictive processing. Show all posts
Showing posts with label predictive processing. Show all posts

Monday, April 27, 2026

When Execution Gets Cheap, What Remains Scarce? - The last biological moat.

I have enjoyed reading a recent essay by Aneesh Sathe, which I recommend you read in full. I will pass on here the equally amazing job that Claude Sonnet 4.6 did in summarizing its main points, then responding to my editing and emphasis suggestions, and also adding some additional ideas on the relevant neurobiological substrate: "the last biological moat" - what is left for the human mind to do. Here is the result, which begins with the title of Sathe's essay: 

The Lightening of Intent: When Execution Gets Cheap, What Remains Scarce?

Aneesh Sathe's essay "The Lightening of Intent" is one of the more intellectually disciplined takes on AI's civilizational meaning I've encountered — grounded in energy economics, manuscript history, and complexity theory rather than hype. It builds explicitly on Venkatesh Rao's "World Machines" framework, published just this week, and the two pieces read best together.


The World Machines Backdrop

Rao's framework organizes civilizational history into overlapping "machines" — planetarities, each with a nominal lifespan of about a millennium, cycling through Dawn, Day, and Dusk phases. Currently, the Modernity Machine is entering its Dusk stage, the Divergence Machine has reached its Day stage, and the Liveness Machine has just been born into its Dawn.

The Liveness Machine is only being born now because real AI has emerged. The most leveraged use of energy, whether renewable or not, will be to power AI. And AI will animate a planet-scale Liveness Machine — whether it is a grimdark or solarpunk version is yet to be determined.

Sathe's essay fills in the economic and physical mechanisms underneath that historical arc.


The Core Argument

The cost of putting an idea into the world has fallen by roughly five orders of magnitude over the last millennium. The bottleneck has reversed: arranging atoms used to be the hard part; now, having ideas is. Soon, it will be intents.

The Codex Amiatinus — the oldest complete Latin Bible — is Sathe's anchor image. It weighed about seventy-five pounds, required close to one thousand calfskins, cost years of scribal labor from sixty monks, and the life of the abbot who carried it toward Rome in 716 CE.  Today, a blog post costs nothing and reaches more readers in an afternoon.


The Numbers Worth Noting

Manuscript-to-print transition:

  • Pre-print Europe held fewer than five million manuscripts; the sixteenth century produced two hundred million printed books, the eighteenth a billion.
  • Gutenberg produced a hundred and eighty Bibles in the time a scriptorium managed one. Book prices fell 2.4 percent per year for over a century; each new printer in a city dropped prices by another quarter.
  • The doubling time for European book production collapsed from roughly 104 years before 1450 to 43 years after.

Energy rate density (Chaisson's framework): This quantity — free energy flow per unit mass in ergs per second per gram — rises monotonically with complexity: galaxies ≈ 0.5; stars ≈ 2; planets ≈ 75; plants ≈ 900; animals ≈ 20,000; the human brain ≈ 150,000; modern human society in aggregate ≈ 500,000 — the most energy-dense phenomenon known.  AI will push this higher still.

Per-capita energy consumption: It has risen from about two thousand kilocalories per day in the Paleolithic — all of it food — to two hundred and thirty thousand in the modern United States.

Energy return on investment (EROI):

  • Modern agriculture requires 13.3 calories of fossil-fuel input per calorie of food consumed.
  • Fossil fuels at the useful-energy stage return only about 3.5 calories per calorie invested; road transport, 1.6 to 1. The estimated minimum EROI for a complex society is about 5 to 1.
  • Solar PV costs have fallen from $106 per watt in 1976 to under $0.10 today — a 1,300-fold decline in under fifty years — with an estimated useful-stage energy return of 25 to 30:1, seven to nine times higher than fossil fuels.

Data accumulation: The internet holds something on the order of two hundred zettabytes by 2026, mostly text and image, mostly read by machines. Roughly ninety percent of all data ever created has been generated in the last two years.


Key Conceptual Moves

The substrate-spark distinction. Sathe draws an analogy to the prebiotic ocean: the pre-life ocean held amino acids and nucleotides for hundreds of millions of years before anything used the accumulation. The chemistry was not the difference; what mattered was that something started to act on it. Data without intent is a soup of records that accumulates and forgets.  LLMs are the first time the substrate has been wired to a borrowed spark of human intent — which maps closely to what Rao calls the Liveness Machine's defining property: AI is oozy, like a primordial soup that harbors intensely reactive chemistry.

Atoms downstream. The HTTP standard, written as a specification in the early 1990s in some weeks, has restructured several trillion dollars of physical economic activity over thirty years. The atoms moved themselves.  The direction of causation between ideas and matter has inverted.

The auteur mode. A bench scientist in 2026 submits a query to a generative model and receives a thousand candidate molecules in twenty minutes; her job is no longer to generate, it is to pick.  Taste, selection, and direction become the scarce inputs. Rao frames this as "execution pull" — AI drawing us out from vita contemplativa regimes into vita activa regimes.

Intents red in tooth and claw. As the substrate becomes more responsive, intent becomes the competitive variable. The first generation of intent-collisions is three to five years out; the shape of the era will be determined in that interval.  Rao places this on a longer timescale: divergence will dominate in the short term (2–5 years) but liveness effects will compound more steadily and dominate in the long term (beyond 5 years).

The energy caveat. The whole argument rides on an energy transition. If the solar transition holds, the Liveness era inherits a re-powered version of the Modernity Machine's infrastructure, sustained on incoming sunlight rather than deposited carbon. If the transition does not hold, the substrate degrades faster than the intent-driven economy can mature, and the lightening of intent ends as a brief anomaly. Both outcomes are within reach.


Why This Matters 

Sathe and Rao together make a tightly nested argument: civilization is a thermodynamic system that keeps burning hotter; each energy-surplus step builds infrastructure that amplifies individual intention; AI is the latest and sharpest such amplifier; and the emerging bottleneck is not execution but what you actually want. For those of us who have spent careers thinking about the neural substrates of agency and intention, the question has an obvious next layer: what, neurobiologically, is the capacity that remains scarce when everything else gets cheap? Sitting with confusion long enough for clarity to emerge — Sathe's phrase — sounds a lot like what the prefrontal cortex does when it holds competing representations in working memory and waits for resolution. That may be the last purely biological moat.


Sathe's companion essay, "The Viscous Frontier", takes up how to act in this regime — with attention as your constraint and no canonical direction pulling. Rao's full World Machines archive is at Contraptions.

The Last Biological Moat: Intention as Prediction Error Suppression

Sathe's claim that sitting with confusion long enough for clarity to emerge remains irreducibly human invites a neuroscientific gloss. In Friston's active inference framework, intentional action is not the initiation of a motor command but the suppression of prediction error about a desired future state. The brain generates a model of how the world should be — the goal — and then acts to make sensory input conform to that model, minimizing the divergence between predicted and actual states. What Sathe calls "formulating a direction" is, in these terms, the construction and stabilization of a prior over future states: the brain committing, against competing attractors, to one preferred trajectory through state space. This is metabolically and computationally expensive precisely because it requires holding an unresolved representation in working memory — prefrontal cortex sustaining an active prior — while suppressing the pull of more immediately rewarding or more habitual alternatives. The "confusion" phase is not inefficiency; it is the system sampling the landscape before locking the prior. AI systems, by contrast, have no intrinsic priors about what they want the world to be. They are extraordinarily powerful at executing on a prior once supplied, but the prior itself — the intent — must come from outside the model. This is why Sathe's bottleneck and Friston's framework converge on the same point: what remains scarce, and stubbornly biological, is the capacity to generate a stable, motivationally loaded model of a preferred future and hold it long enough to act. Everything downstream of that — the scribal labor, the printing press, the HTTP spec, the generative model — is infrastructure for carrying the prior into the world. The infrastructure keeps getting cheaper and more powerful. The prior still has to come from somewhere.

 

Monday, April 20, 2026

What a self is.

Reading Michael Pollan’s account of his meeting with Anil Seth in his recent book "A World Appears" has prompted me to write down for my own use what I take a “self” to be. This post archives that summary and shares it with interested MindBlog readers.

So, here’s the summary:*

The self can be understood, to use Seth's phrase,  as a "controlled hallucination." Our brains build this construct to regulate the body using interoceptive signals—internal data about our heart rate, breathing, and chemistry—to maintain stability (homeostasis) in the face of constant disruption. From these signals arise feelings and emotions that drive us to act, biasing behavior toward preserving coherence and pushing back against the entropy that would otherwise dissolve it. Our illusion of having agency, of being able to do things that matter, is one of our most necessary and powerful emotions.

This hallucination is not just about the present moment. It stitches together a historical self from memory and prior experience (the brain’s ‘priors”), then projects that self forward in time, generating a predicted future to act into. We are, in this sense, always living slightly ahead of ourselves. The self is not a fixed entity but an ongoing process: a predictive framework that links memory, expectation, and action.

The self is also a stage — a theater or structural model that evolved to support the regulation of the neuroendocrine machinery underlying our social emotions and feelings such as fear, status, and affiliation - feelings that tie us to others and to our place in the world.  The theater of selfhood enables these processes to operate coherently across time and context.

And here is what I find most striking: once this elaborate scaffolding is in place, it sometimes becomes possible to step outside of it. To temporarily set aside the past and future timeline, the narrative, the predictions — and let awareness rest in the present moment alone. In that open, unhurried awareness, thoughts, feelings, and actions can be observed as they arise, like wisps of vapor emerging from some deeper source.

The self, it turns out, may be most clearly seen from just outside it.


*A note on how I arrived at the above text:    I wrote a paragraph of my ideas, and then presented the prompt “Please do an edit or redraft of this MindBlog post draft to make it more comprehensible to readers:” to four LLMs (ChatGPT, Gemini, Claude, and DeepSeek.) I then curated the four versions to select useful improvements of my text and did further editing myself to make the final product.
 

Friday, August 22, 2025

Predictability and the pleasure of music

Mas-Herrero et al. do an interesting study on how predictive processes shape individual musical preferences .

Significance

Using a novel decision-making task, we show that musical pleasure relies on a delicate balance between predictability and uncertainty, consistent with learning theories. In simple terms, music that is not overly expected nor too chaotic is most enjoyable—but the ideal mix of predictability depends on how much the melody keeps you guessing. Very predictable tunes can be delightful with small twists, while a melody full of surprises may need bigger unexpected moments to hit the sweet spot. Computational models incorporating this balance accurately predicted the types of music people like and the pleasure they derive from real compositions. These results reveal fundamental mechanisms driving musical pleasure and offer valuable insights for the music industry and music-based therapies.

Abstract

Current models suggest that musical pleasure is tied to the intrinsic reward of learning, as it relies on predictive processes that challenge our minds. According to predictive coding, optimal learning, which maximizes epistemic value, depends on balancing predictability and uncertainty, implying that musical pleasure should also reflect this equilibrium. We tested this idea in two independent large samples using a novel decision-making paradigm, where participants indicated preferences for melodies varying in surprise and entropy. Consistent with prior research, we found an inverted U-shaped relationship between predictability and preference. Moreover, our results revealed an interaction between predictability and entropy, with smaller surprises preferred in low-entropy melodies and larger surprises favored in high-entropy music, consistent with predictive coding principles. Computational models incorporating this interaction predicted individuals’ genre preferences and pleasure responses to real compositions, highlighting its applicability to real-world music experiences. These findings advance our understanding of the cognitive mechanisms driving music preferences and the role of predictive processes in affective responses.

 

Thursday, March 27, 2025

Deric and DeepSeek R1 on the enteroceptive self in a small world - taking refuge in troubled times

This post on an exchange with DeepSeek R1 is a followup on my previous post describing the Chat GPT 4o response to some of my idea fragments - further convincing me that current LLMs'  power to integrate and expand on concepts expressed in language are superior to mine.   

I presented exactly the same prompt to DeepSeek R1 that I had given to ChatGPT 4o: and the results astound me:

prompt:   

I’m curious to see how you might respond to reading the following notes on testing different topic ideas to write about and develop.  The issue is not one of whether the ideas are novel, but whether I might devise, with your help,  a different permutation, perspective, or take on them that I (and others) might find useful.  Can you can suggest some elaboration or further development of the ideas I put forward.

Here are the notes:

First Possible Title:  The interoceptive self as a basis for actions, thoughts and feelings.  

Subtitle: Moving the self’s experienced center of gravity from “I” to “it”

It is a question of who and/or what is running the show

Which brings it back to the old upstairs/downstairs chicken or egg circular causality issue.  

So, ‘it all depends’ (cf Sapolsky’s massive tome) 

And, there is no ‘there’ in terms of understanding, that isn’t in my 1999 “The Biology of Mind”  book.  

Second Possible Title: Defining a smaller self world as a human animal’s refuge in troubled times. 

Our predecessors were small mouse like creatures that survived the asteroid collision of  ~66 million years ago by burrowing into small protective cracks and burrows.  I wonder about robustly engaging this strategy, being one of the peasants or ‘sheeple’ challenged by the rapid disintegration of the geopolitical world order of the past 50 years

And then what?  How to rebuild a new world?… perhaps  Venkatest Rao’s cozy world, rising from the ooze of the permaweird,  protocols, etc.  

-------------

DeepSeek R1's response:

1. The Interoceptive Self as a Basis for Actions, Thoughts, and Feelings

Subtitle Refinement: From "I" to "It": How Bodily Awareness Rewires Agency and Identity

Key Angles to Develop:

  • Neuroscientific Foundations:

    • Link interoception (sensing internal states) to Antonio Damasio’s somatic marker hypothesis and Lisa Feldman Barrett’s predictive processing theory. Emphasize how the brain constructs "selfhood" by predicting and regulating bodily needs.

    • Contrast the "interoceptive self" with the Cartesian "I"—e.g., how gut microbiomes or heart-rate variability influence decisions, challenging the myth of a disembodied mind.

  • Philosophical Implications:

    • Explore parallels with Buddhist anattā (non-self) or Spinoza’s monism, where the self is a process, not an entity.

    • Debate free will: If the "it" (body/brain systems) runs the show, what does this mean for moral responsibility?

  • Mental Health and Society:

    • Discuss interoceptive dysregulation in anxiety/depression and therapies like biofeedback or mindfulness.

    • Argue for a cultural shift: Redefining success as alignment with bodily rhythms (e.g., circadian health, stress cycles) rather than productivity.

Provocative Questions:

  • Is the "self" merely the brain’s user interface for bodily homeostasis?

  • Could embracing the "it" dissolve existential angst by grounding identity in biology?


2. Defining a Smaller Self-World as a Human Animal’s Refuge in Troubled Times

Subtitle Refinement: From Survival to Regeneration: Lessons from Evolutionary Scarcity for Modern Collapse

Key Angles to Develop:

  • Evolutionary and Historical Precedents:

    • Compare the asteroid-surviving mammals to modern "digital burrowers" (e.g., remote workers, crypto-anarchists, homesteaders).

    • Analyze historical collapses (Bronze Age, Roman Empire) and how niche communities preserved knowledge/values.

  • The "Cozy World" and Permaweird:

    • Expand on Venkatesh Rao’s cozy futurism: How decentralized tech (DAOs, Web3) and subcultures (e.g., solarpunk) create insulated yet adaptive communities.

    • Critique escapism: When does "smallness" become complacency vs. a staging ground for renewal?

  • Rebuilding Strategies:

    • Propose a modular resilience framework: Micro-communities that share resources (e.g., mutual aid networks) but avoid overcentralization.

    • Highlight regenerative practices: Permaculture, open-source protocols, or trauma-informed governance models.

Provocative Questions:

  • Is retreat a radical act of self-preservation or a betrayal of collective responsibility?

  • Can "small worlds" scale without losing their adaptive edge?


Cross-Cutting Themes:

  • The Body-Politics Nexus:

    • Tie interoceptive awareness to societal resilience: A dysregulated body cannot engage in collective action. How might somatic practices (e.g., breathwork) fuel post-collapse rebuilding?

  • Technology as Double-Edged Sword:

    • Critique apps that commodify interoception (e.g., mindfulness apps) versus tools that foster bodily autonomy (e.g., DIY biohacking).

    • Explore how AI could either erode the "interoceptive self" (via attention economy) or enhance it (via personalized health insights).

  • Ethics of Adaptation:

    • Contrast adaptive passivity (e.g., burrowing) with adaptive activism (e.g., prefigurative politics). When is each strategy viable?


Practical Applications:

  1. Workbook/Journal Prompts:

    • For Title 1: "Map a decision to your bodily state: Did hunger/fatigue shape it?"

    • For Title 2: "Design your ‘cozy protocol’—what three practices would sustain you in collapse?"

  2. Case Studies:

    • Iceland’s post-2008 recovery (small-world resilience) vs. Silicon Valley’s "move fast and break things" ethos.

    • Communities using interoceptive practices (e.g., trauma-sensitive yoga in prisons) to rebuild agency.

  3. Futures Framework:

    • Scenarios for 2050: A world where "interoceptive literacy" is taught in schools, or where decentralized "tribes" negotiate resource-sharing via blockchain.


Counterarguments to Address:

  • For Title 1: Does biologizing the self risk nihilism? Rebuttal: Agency emerges from understanding bodily constraints, not denying them.

  • For Title 2: Does retreat perpetuate inequality? Rebuttal: Focus on redistribution of sovereignty (e.g., land trusts, UBI) as part of "small-world" ethics.

By weaving these threads, your work could bridge neuroscience, philosophy, and futurism—offering readers not just analysis, but practices for navigating an uncertain world. Would a deeper dive into any of these angles spark your interest?

 

 

Wednesday, January 08, 2025

Predictive coding for social perception

Rittershofer et al. introduce a special issue of Neuroscience and Biobehavioral Reviews titled "Neurobiology of Social Interactions Across Species."  Here is the first part of that introduction:

Perception cannot rely solely on bottom-up processes, whereby patterns of receptor stimulation are passed up the hierarchy to generate a corresponding awareness. Such bottom-up processes would always generate experiences that are out-of-date and saturated by noise. Predictive processes are thought to play a key role in overcoming these problems, allowing us to generate best guesses concerning the likely sensorium, and highlighting quickly when the world is not as we expect. Action provides a crucial predictive source and a mechanism for us to resolve uncertainty and surprise, but further complicates our understanding due to further predictive cues and continuous change of sensory input. Another agent who can also change the world and who we seek to understand adds another layer of complexity yet. How can we understand the predictive mechanisms supporting social interaction and understanding, with such a multitude of moving and interacting components? In this special issue, Keysers et al. (2024) outline how predictive coding can be applied to understanding the actions and emotions of others, with Mayo and Shamay-Tsoory (2024) discussing how these mutual predictions might shape social learning. They suggest that such social learning might be supported by interbrain synchronization and Antonelli et al. (2024) discuss the critical role of emotion in shaping these multibrain dynamics.
 
While it is clearly crucial that we understand the nature of the mechanisms underlying social interactions, we wish to highlight the challenges of this complexity for scientific progress. Particularly, how to find ways to properly test, refute, and improve our models, when the assumed supporting mechanisms are so complex.
 
How predictions shape neural processing is thought to differ across space and time, even for processing of the simplest (non-social; static) elements of our environment. Keysers et al. (2024) highlight the assumed neural interactions across cortical layers, such that predictions are passed down the hierarchy to hypothesis units in deep (and perhaps superficial) cortical layers, input arrives in middle layers, and error signals are calculated and represented in superficial layers. This idea is supported by recent 7 T MRI work from our lab demonstrating increased decoding of predicted Gabor orientations in deep layers of primary visual cortex, with an advantage for unpredicted orientations in superficial layers (Thomas et al., 2024). Recent evidence suggests opposing influences at the temporal level as well (McDermott et al., 2024). This electroencephalography (EEG) study found that early perceptual processing is biased towards what we expect (< 200 ms; optimizing veridicality) with the advantage flipping in later timeranges (> 200 ms; optimizing informativeness – in line with the opposing process account proposed in Press et al., 2020). Building testable mechanistic accounts of these interactions across time and space – even for the simple perception of deterministic sequences of Gabor patches – represents a continued puzzle for future work.
 
In the social domain, the stimuli are by their nature highly complex and dynamic (Keysers et al., 2024). Therefore, these above interactions across space and time must be continuously updated. Despite this complexity, there is some evidence cited by Keysers et al. (2024) inline with the above laminar conclusions in simpler environments. Specifically, there is increased deep-layer information about observed actions in parietal cortex when presented in a predictable order, mediated via feedback connections (from premotor cortex). Social domains also yield multiple sources of prediction about the self and other (Mayo and Shamay-Tsoory, 2024) and we must determine how we weight the precision, or reliability, of these different sources, as well as how we render information about the self and other separable. Is this achieved by different cell populations coding information about the self and other (Mayo and Shamay-Tsoory, 2024)? Or could mechanisms similar to those proposed to distinguish products of imagination from reality (similarly internal vs external sources), also help in determining the information source in social situations?
 
Social predictions might be supported by interbrain synchronization (measured via hyperscanning), as discussed by Mayo and Shamay-Tsoory (2024); focus on social learning) and Antonelli et al. (2024); focus on emotion). We propose that one key challenge for this approach is determining the role played by different event-related inputs and responses in the effects: Interpretation of hyperscanning data is plagued by the problem that brains will be “in synch” if two individuals are either perceiving the same events or producing the same behaviour. The brain’s responses to moving our arm or looking at a face are remarkably similar across individuals, such that if two of us perceive or produce the same event our neural response will be matched. Fluctuations in synchronisation according to, e.g., dominance of individuals or levels of excitement on stage, could be determined by fluctuations in whether we attend to, or produce, the same events. It is crucial to understand the fascinating influence of these effects on synchronisation.