Entropy (classical thermodynamics)



Conjugate variables
of thermodynamics
PressureVolume
(Strain)
TemperatureEntropy
Particle no.

In chemical potential tend to even out as time goes by, unless there is an outside influence which works to maintain the differences. The end point of this evening-out process is called equilibrium, and as the system moves toward equilibrium, the entropy of the system increases, becoming a maximum at equilibrium.

There are two equivalent definitions of entropy. The first definition is the thermodynamic definition. It was developed in the early 1850s by thermodynamicist Boltzmann's constant.

This article is concerned with the thermodynamic definition of entropy. Although thermodynamic entropy is a self-contained subject, it should be understood in parallel with the statistical definition. When the thermodynamic definition becomes most difficult to understand, the statistical definition brings a simple explanation, and where the link between the statistical theory and experiment becomes extended, the thermodynamic theory delivers a straightforward answer.

Clausius defined the change in entropy ds of a thermodynamic system, during a reversible process, as

dS = \frac{\delta Q}{T} \!

where

δQ is a small amount of heat introduced to the system,
T is a constant absolute temperature

Note that the small amount δQ of energy transferred by heating is denoted by δQ rather than dQ, because Q is not a state function while the entropy is.

Clausius gave the quantity S the name "entropy", from the Greek word τρoπή, "transformation". Since this definition involves only differences in entropy, the entropy itself is only defined up to an arbitrary additive constant.

When a process is irreversible, the above definition must be replaced by the statement that the entropy change is equal to the amount of energy required to return the system to its original state by a reversible transformation at a constant temperature, divided by that temperature. This is explained in more detail below.

Introduction

In a thermodynamic system, a "universe" consisting of "surroundings" and "systems" and made up of quantities of matter, its pressure differences, density differences, and temperature differences all tend to equalize over time. In the ice melting example, the difference in temperature between a warm room (the surroundings) and cold glass of ice and water (the system and not part of the room), begins to be equalized as portions of the heat energy from the warm surroundings become spread out to the cooler system of ice and water.   Over time the temperature of the glass and its contents and the temperature of the room become equal. The entropy of the room has decreased and some of its energy has been dispersed to the ice and water. However, as calculated in the example, the entropy of the system of ice and water has increased more than the entropy of the surrounding room has decreased. In an thermodynamic system is a measure of how far the equalization has progressed.

A special case of entropy increase, the entropy of mixing, occurs when two or more different substances are mixed. If the substances are at the same temperature and pressure, there will be no net exchange of heat or work - the entropy increase will be entirely due to the mixing of the different substances.[1]

From a macroscopic perspective, in heat (TR is the temperature of the system's external surroundings). Otherwise the process will not go forward.

In 1862, Clausius stated what he calls the “theorem respecting the equivalence-values of the transformations” or what is now known as the second law of thermodynamics, as such:

The algebraic sum of all the transformations occurring in a cyclical process can only be positive, or, as an extreme case, equal to nothing.

Quantitatively, Clausius states the mathematical expression for this theorem is as follows. Let δQ be an element of the heat given up by the body to any reservoir of heat during its own changes, heat which it may absorb from a reservoir being here reckoned as negative, and T the absolute temperature of the body at the moment of giving up this heat, then the equation:

\int \frac{\delta Q}{T} = 0

must be true for every reversible cyclical process, and the relation:

\int \frac{\delta Q}{T} \ge 0

must hold good for every cyclical process which is in any way possible. This is the essential formulation of the second law and one of the original forms of the concept of entropy. It can be seen that the dimensions of entropy are energy divided by temperature, which is the same as the dimensions of chemical process or as when a system changes state.

Heat engines

Clausius' identification of S as a significant quantity was motivated by the study of reversible and irreversible thermodynamic transformations. A thermodynamic transformation is a change in a system's thermodynamic properties, such as pressure waves are created, so the gas is not in equilibrium, and the transformation is irreversible.

A conservation of energy, the net energy lost by the environment is equal to the work done by the engine.

If every transformation in the cycle is reversible, the cycle is reversible, and it can be run in reverse, so that the energy transfers occur in the opposite direction and the amount of work done switches sign.

Definition of temperature

In thermodynamics, temperatures do not change when energy flows into or out of them. A reversible cycle exchanges heat with the two heat reservoirs:

\frac{Q_1}{T_1} = \frac{Q_2}{T_2} \!      

where

Q1 is the amount of heat that the cycle absorbs from the first reservoir,
T1 is the temperature of the first reservoir and
Q2 is the amount of heat delivered to the second,
T2 is the temperature of the second reservoir.
Proof: Introduce an additional heat reservoir at an arbitrary temperature T0, as well as N cycles with the following property: the j-th such cycle operates between the T0 reservoir and the Tj reservoir, transferring energy δQj to the latter. From the above definition of temperature, the energy extracted from the T0 reservoir by the j-th cycle is
\delta Q_{0,j} = T_0 \frac{\delta Q_j}{T_j} \,\!

Now consider one cycle of the heat engine, accompanied by one cycle of each of the smaller cycles. At the end of this process, each of the N reservoirs have zero net energy loss (since the energy extracted by the engine is replaced by the smaller cycles), and the heat engine has done an amount of work equal to the energy extracted from the T0 reservoir,

W = \sum_{j=1}^N \delta Q_{0,j} = T_0 \sum_{j=1}^N \frac{\delta Q_j}{T_j} \,\!

If this quantity is positive, this process would be a perpetual motion machine of the second kind, which is impossible. Thus,

\sum_{i=1}^N \frac{\delta Q_i}{T_i} \le 0 \,\!

Now repeat the above argument for the reverse cycle. The result is

\sum_{i=1}^N \frac{\delta Q_i}{T_i} = 0 \,\! (reversible cycles)

Now consider a reversible cycle in which the engine exchanges heats \delta Q_1,\delta Q_2,\cdots,\delta Q_N with a sequence of N heat reservoirs with temperatures T1, ..., TN. We can show (see the box on the right) that:

\sum_{i=1}^N \frac{\delta Q_i}{T_i} = 0 \!.

where

δQi is the amount of heat exchanged from the engine to the reservoir,
Ti is the temperature of each heat reservoir, and
i is an integer representing each different heat reservoir.

Since the cycle is reversible, the engine is always infinitesimally close to equilibrium, so its temperature is equal to any reservoir with which it is contact. In the limiting case of a reversible cycle consisting of a continuous sequence of transformations,

\oint \frac{\delta Q}{T} = 0 \,\! (reversible cycles)

where the integral is taken over the entire cycle, and T is the temperature of the system at each point in the cycle. This is a particular case of the Clausius theorem.

Entropy as a state function

We can now deduce an important fact about the entropy change during any thermodynamic transformation, not just a cycle. First, consider a reversible transformation that brings a system from an equilibrium state A to another equilibrium state B. If we follow this with any reversible transformation which returns that system to state A, our above result says that the net entropy change is zero. This implies that the entropy change in the first transformation depends only on the initial and final states.

This allows us to define the entropy of any equilibrium state of a system. Choose a reference state R and call its entropy SR. The entropy of any equilibrium state X is

S_X = S_R + \int_R^X \frac{\delta Q}{T} \,\!

Since the integral is independent of the particular transformation taken, this equation is well-defined.

Entropy change in irreversible transformations

We now consider irreversible transformations. It can be shown that the entropy change during any transformation between two equilibrium states is

\Delta S \ge \int \frac{\delta Q}{T} \,\!

where the equality holds if the transformation is reversible.

Notice that if δQ = 0, then ΔS ≥ 0. This is the Second Law of Thermodynamics, discussed earlier in the article.

Suppose a system is thermally and mechanically isolated from the environment. For example, consider an insulating rigid box divided by a movable partition into two volumes, each filled with gas. If the pressure of one gas is higher, it will expand by moving the partition, thus performing work on the other gas. Also, if the gases are at different temperatures, heat can flow from one gas to the other provided the partition is an imperfect insulator. Our above result indicates that the entropy of the system as a whole will increase during these process (it could in principle remain constant, but this is unlikely.) Typically, there exists a maximum amount of entropy the system may possess under the circumstances. This entropy corresponds to a state of stable equilibrium, since a transformation to any other equilibrium state would cause the entropy to decrease, which is forbidden. Once the system reaches this maximum-entropy state, no part of the system can perform work on any other part. It is in this sense that entropy is a measure of the energy in a system that "cannot be used to do work".

Measuring entropy

In real experiments, it is quite difficult to measure the entropy of a system. The techniques for doing so are based on the thermodynamic definition of the entropy, and require extremely careful calorimetry.

For simplicity, we will examine a mechanical system, whose thermodynamic state may be specified by its volume V and pressure P. In order to measure the entropy of a specific state, we must first measure the heat capacity at constant volume and at constant pressure (denoted CV and CP respectively), for a successive set of states intermediate between a reference state and the desired state. The heat capacities are related to the entropy S and the temperature T by

C_X = T \left(\frac{\partial S}{\partial T}\right)_X \,\!

where the X subscript refers to either constant volume or constant pressure. This may be integrated numerically to obtain a change in entropy:

\Delta S = \int \frac{C_X}{T} dT \,\!

We can thus obtain the entropy of any state (P,V) with respect to a reference state (P0,V0). The exact formula depends on our choice of intermediate states. For example, if the reference state has the same pressure as the final state,

S(P,V) = S(P, V_0) + \int^{T(P,V)}_{T(P,V_0)} \frac{C_P(P,V(T,P))}{T} dT \,\!

In addition, if the path between the reference and final states lies across any latent heat associated with the transition must be taken into account.

The entropy of the reference state must be determined independently. Ideally, one chooses a reference state at an extremely high temperature, at which the system exists as a gas. The entropy in such a state would be that of a classical ideal gas plus contributions from molecular rotations and vibrations, which may be determined ice by the latter method, assuming no entropy at zero temperature, falls short of the value obtained with a high-temperature reference state by 3.41 J/(mol·K). This is due to the "zero-point" entropy of ice mentioned earlier.

See also

References

  1. ^ See, e.g., Notes for a “Conversation About Entropy” for a brief discussion of both thermodynamic and "configurational" ("positional") entropy in chemistry.

Further reading

  • Goldstein, Martin, and Inge F., 1993. The Refrigerator and the Universe. Harvard Univ. Press. A gentle introduction at a lower level than this entry.
 
This article is licensed under the GNU Free Documentation License. It uses material from the Wikipedia article "Entropy_(classical_thermodynamics)". A list of authors is available in Wikipedia.