|home >> research >> Maxwell's Demon...|
|Maxwell's demon is a fictitious, miniscule being imagined by Maxwell as able to reverse the second law of thermodynamics by manipulating individual molecules. In a tradition of work initiated by Szilard in the 1920s, it has become standard to predict the failure of the demon on information theoretic grounds through a connection supposed to obtain between information processing and entropy dissipation. In a study with John Earman, we have suggested that this account of the demon's failure is either based on question begging or groundless supposition.||With John Earman, "Exorcist XIV: The Wrath of Maxwell's
Demon." Studies in the History and Philosophy of Modern Physics, Part I
"From Maxwell to Szilard" 29(1998), pp.435-471; Part II: "From Szilard
to Landauer and Beyond," 30(1999), pp.1-40. Download.
|The present orthodoxy holds that Maxwell's demon must fail to reverse
the second law of thermodynamics because of a hidden entropy cost in
the erasure of information. The analysis is based on Landauer's
principle, which asserts that the erasure of n bits of information is
accompanied by the passage of least k ln n of entropy to the
surroundings. I argue that Landauer's principle is based on the
formation of illicit canonical ensembles in statistical physics that
give the illusion of the necessity of this entropy cost. I also urge
that, even if the principle were correct, the literarure seeks to
establish that it must defeat all Maxwell demons by the inadquate means
of merely displaying a few suggestive examples.
||"Eaters of the Lotus: Landauer's Principle and the Return of Maxwell's Demon." 36 (2005), pp. 375-411. Download|
|Landauer's Principle asserts that there is an unavoidable cost in thermodynamic entropy creation when data is erased. It is usually derived from incorrect assumptions, most notably, that erasure must compress the phase space of a memory device or that thermodynamic entropy arises from the probabilistic uncertainty of random data. Recent work seeks to prove Landauer’s Principle without using these assumptions. I show that the processes assumed in the proof, and in the thermodynamics of computation more generally, can be combined to produce devices that both violate the second law and erase data without entropy cost, indicating an inconsistency in the theoretical system. Worse, the standard repertoire of processes selectively neglects thermal fluctuations. Concrete proposals for how we might measure dissipationlessly and expand single molecule gases reversibly are shown to be fatally disrupted by fluctuations.||"Waiting for Landauer," Studies in History and Philosophy of Modern Physics, 42(2011), pp. 184-198. Download.
See also Goodies pages:
When a Good Theory meets a Bad Idealization: The Failure of the Thermodynamics of Computation.
No Go Result for the Thermodynamics of Computation
|Entropy creation in excess of that tracked by Landauer's principle is needed to overcome fluctuations in molecular scale computation. This paper is a short account of the "no go" result reported in "Waiting for Landauer."||"The End of the Thermodynamics of Computation: A No Go Result." Prepared for PSA 2012: Philosophy of Science Biennial Meeting. To appear in Philosophy of Science. Download.|