Nature is not homogenous. Most of the universe is complex and composed of various subsystems—self-contained systems within a larger whole. Microscopic cells and their surroundings, for example, can be divided into many different subsystems: the ribosome, the cell wall, and the intracellular medium surrounding the cell.
The Second Law of Thermodynamics tells us that the average entropy of a closed system in contact with a heat bath—roughly speaking, its "disorder"— always increases over time. Puddles never refreeze back into the compact shape of an ice cube and eggs never unbreak themselves. But the Second Law doesn't say anything about what happens if the closed system is instead composed of interacting subsystems.
New research by SFI Professor David Wolpert published in the New Journal of Physics considers how a set of interacting subsystems affects the second law for that system.
"Many systems can be viewed as though they were subsystems. So what? Why actually analyze them as such, rather than as just one overall monolithic system, which we already have the results for," Wolpert asks rhetorically.
The reason, he says, is that if you consider something as many interacting subsystems, you arrive at a "stronger version of the second law," which has a nonzero lower bound for entropy production that results from the way the subsystems are connected. In other words, systems made up of interacting subsystems have a higher floor for entropy production than a single, uniform system.
All entropy that is produced is heat that needs to be dissipated, and so is energy that needs to be consumed. So a better understanding of how subsystem networks affect entropy production could be very important for understanding the energetics of complex systems, such as cells or organisms or even machinery
Wolpert's work builds off another of his recent papers which also investigated the thermodynamics of subsystems. In both cases, Wolpert uses graphical tools for describing interacting subsystems.
For example, the following figure shows the probabilistic connections between three subsystems—the ribosome, cell wall, and intracellular medium.
Like a little factory, the ribosome produces proteins that exit the cell and enter the intracellular medium. Receptors on the cell wall can detect proteins in the intracellular medium. The ribosome directly influences the intracellular medium but only indirectly influences the cell wall receptors. Somewhat more mathematically: A affects B and B affects C, but A doesn't directly affect C.
Why would such a subsystem network have consequences for entropy production?
"Those restrictions—in and of themselves—result in a strengthened version of the second law where you know that the entropy has to be growing faster than would be the case without those restrictions," Wolpert says.
A must use B as an intermediary, so it is restricted from acting directly on C. That restriction is what leads to a higher floor on entropy production.
Plenty of questions remain. The current result doesn't consider the strength of the connections between A, B, and C—only whether they exist. Nor does it tell us what happens when new subsystems with certain dependencies are added to the network. To answer these and more, Wolpert is working with collaborators around the world to investigate subsystems and entropy production. "These results are only preliminary," he says.
More information: David H Wolpert. Minimal entropy production rate of interacting systems, New Journal of Physics (2020). DOI: 10.1088/1367-2630/abc5c6
Journal information: New Journal of Physics
Provided by Santa Fe Institute