Entropy vs Disgregation - What's the difference?
entropy | disgregation |
(thermodynamics, countable)
# strictly thermodynamic entropy . A measure of the amount of energy in a physical system that cannot be used to do work.
# A measure of the disorder present in a system.
# The capacity factor for thermal energy that is hidden with respect to temperature [http://arxiv.org/pdf/physics/0004055].
# The dispersal of energy; how much energy is spread out in a process, or how widely spread out it becomes, at a specific temperature. [http://www.entropysite.com/students_approach.html]
(statistics, information theory, countable) A measure of the amount of information and noise present in a signal. Originally a tongue-in-cheek coinage, has fallen into disuse to avoid confusion with thermodynamic entropy.
(uncountable) The tendency of a system that is left to itself to descend into chaos.
Separation; scattering.
(thermodynamics) Entropy, defined as the magnitude of the separation of the particles of a system. Introduced by the German physicist Rudolf Clausius in 1862.
:::If we examine the conditions under which heat can be transformed into ergon, and, conversely, ergon into heat, we find, in the first place, that the commonest and simplest process is the following. The heat which exists in material bodies tends to alter their condition. It tends to expand them, to render solid bodies liquid and gaseous, and, as we may likewise add, to resolve chemical compounds into their elements. In all these cases the effect of the heat consists in loosening or completely dissolving the connexion which exists between the molecules or atoms, and in separating to the greatest possible distance such molecules as are already completely disconnected from each other.
:::In order to be able to express this action shortly, I have introduced a magnitude which denotes the extent to which this separation and parting of its smallest particles, which it is the tendency of heat to effect, has already been carried in the case of any body. This magnitude I call the Disgregation of the body. The disgregation of a body is consequently, among the three states of aggregation, least in the solid state, greater in the liquid state, and greatest of all in the gaseous state. In the last condition it can still be increased by the molecules separating further from each other—that is, by the gas expanding to a larger volume. In like manner, the decomposition of a chemically compound body into its elements is in general accompanied by an increase of disgregation.
:::By help of this conception the effect of heat can be simply expressed by saying that heat tends to increase the disgregation of bodies''.Clausius, R. ? On the Second Fundamental Theorem of the Mechanical Theory of Heat; a Lecture delivered before the Forty-first Meeting of the German Scientific Association, at Frankfurt on the Maine, September 23, 1867 ? ''The Philosophical Magazine and Journal of Science , June 1868, pp. 407–408
As nouns the difference between entropy and disgregation
is that entropy is (thermodynamics|countable) while disgregation is separation; scattering.entropy
English
Noun
(wikipedia entropy)- The thermodynamic free energy is the amount of work that a thermodynamic system can perform; it is the internal energy of a system minus the amount of energy that cannot be used to perform work. That unusable energy is given by the entropy''' of a system multiplied by the temperature of the system.''[http://en.wikipedia.org/wiki/Thermodynamic_free_energy] ''(Note that, for both Gibbs and Helmholtz free energies, temperature is assumed to be fixed, so '''entropy is effectively directly proportional to useless energy.)
- Ludwig Boltzmann defined entropy''' as being directly proportional to the natural logarithm of the number of microstates yielding an equivalent thermodynamic macrostate (with the eponymous constant of proportionality). Assuming (by the fundamental postulate of statistical mechanics), that all microstates are equally probable, this means, on the one hand, that macrostates with higher '''entropy''' are more probable, and on the other hand, that for such macrostates, the quantity of information required to describe a particular one of its microstates will be higher. That is, the Shannon '''entropy''' of a macrostate would be directly proportional to the logarithm of the number of equivalent microstates (making it up). In other words, thermodynamic and informational entropies are rather compatible, which shouldn't be surprising since Claude Shannon derived the notation 'H' for information '''entropy from Boltzmann's H-theorem.