A theory of reality as more than the sum of its parts
In his 1890 opus, The Principles of Psychology, William James invoked Romeo and Juliet to illustrate what makes conscious beings so different from the particles that make them up.
“Romeo wants Juliet as the filings want the magnet; and if no obstacles intervene he moves towards her by as straight a line as they,” James wrote. “But Romeo and Juliet, if a wall be built between them, do not remain idiotically pressing their faces against its opposite sides like the magnet and the filings. … Romeo soon finds a circuitous way, by scaling the wall or otherwise, of touching Juliet’s lips directly.”
Erik Hoel, a 29-year-old theoretical neuroscientist and writer, quoted the passage in a recent essay in which he laid out his new mathematical explanation of how consciousness and agency arise. The existence of agents — beings with intentions and goal-oriented behavior — has long seemed profoundly at odds with the reductionist assumption that all behavior arises from mechanistic interactions between particles. Agency doesn’t exist among the atoms, and so reductionism suggests agents don’t exist at all: that Romeo’s desires and psychological states are not the real causes of his actions, but merely approximate the unknowably complicated causes and effects between the atoms in his brain and surroundings.
Hoel’s theory, called “causal emergence,” roundly rejects this reductionist assumption.
“Causal emergence is a way of claiming that your agent description is really real,” said Hoel, a postdoctoral researcher at Columbia University who first proposed the idea with Larissa Albantakis and Giulio Tononi of the University of Wisconsin, Madison. “If you just say something like, ‘Oh, my atoms made me do it’ — well, that might not be true. And it might be provably not true.”
Using the mathematical language of information theory, Hoel and his collaborators claim to show that new causes — things that produce effects — can emerge at macroscopic scales. They say coarse-grained macroscopic states of a physical system (such as the psychological state of a brain) can have more causal power over the system’s future than a more detailed, fine-grained description of the system possibly could. Macroscopic states, such as desires or beliefs, “are not just shorthand for the real causes,” explained Simon DeDeo, an information theorist and cognitive scientist at Carnegie Mellon University and the Santa Fe Institute who is not involved in the work, “but it’s actually a description of the real causes, and a more fine-grained description would actually miss those causes.” [Continue reading…]