Statistical Mechanics Notes 1
Statistical Mechanics Notes 1
Contents
Conservation of Information, Energy, Entropy, and Temperature 1.......................................................... 1 1.1 Introduction ....................................................................................................................................... 1 conservation of distinction, or conservation of information. .......................................................... 3 1.2 The 1st Law of Thermodynamics: Energy is conserved ....................................................................... 3 1.3 Entropy is an information-theoretic concept. ..................................................................................... 3 Entropy is measured in bits .................................................................................................................. 4 The general definition of entropy ......................................................................................................... 4 1.4 Temperature ...................................................................................................................................... 5 1.5 The 2nd Second Law of Thermodynamics: Entropy always increases ................................................... 5 2 The Bolzmann Distribution .................................................................................................................... 6 2.1 Solving for the Lagrange Multipliers ................................................................................................... 8 2.2 Helmholtz Free Energy ....................................................................................................................... 8 3 Fluctuations .......................................................................................................................................... 9 Note on Boltzmann constant k = 1 ..................................................................................................... 10 4 Control Parameters............................................................................................................................. 10 4.1 The Adiabatic Theorem and the First Law ......................................................................................... 10 Adiabatic Invariants Entropy .............................................................................................................. 10 4.2 Processes at Fixed Temperature....................................................................................................... 11 Part II Some Simple Applications ............................................................................................................ 12 5 Ideal Gas ............................................................................................................................................. 12
Statistical mechanics is a useful tool in many areas of science in which large a number of variables have to be dealt with using statistical methods. My son, who studies neural networks, uses it. I have no doubt that some of the financial wizards at AIG and Lehman Brothers used it. Saying that Stat Mech is the theory of gasses is rather like saying calculus is the theory of planetary orbits. Stat Mech is really a particular type of probability theory. Coin flipping is a good place to start. The probabilities for heads (H) and tails (T) are both equal to 1/2. Why do I say that? One answer is that the symmetry between H and T means their probabilities are equal. Here is another example. Lets take a die (as in dice) and color the six faces red, yellow, blue, green, orange, and purple (R, Y, B, G,O, P). The obvious cubic symmetry of the die dictates that the probabilities are equal to 1/6. But what if we dont have a symmetry to rely on? How do we assign a priori probabilities? Suppose for example instead of the coloring scheme that I indicated above, I chose to color the purple face red. Then there would be only five colors. Would the probability of throwing a given color be 1/5? After all, if I just write (R, Y, B, G,O), the 5 names are just as symmetric as the original 6 names. Nonsense, you say: the real symmetry is among the 6 faces, and that is so. But what if there really is no obvious symmetry at all, for example if the die is weighted in some unfair way? In that case we would have to rely on a bunch of details such as the precise way the die was thrown by the hand that threw it, the wind, maybe even the surface that the die lands of (can it bounce?). As is often the case, we have to think of the system in question as part of a bigger system. But what about the bigger system? How do we assign its probabilities. Here is another idea that involves some dynamics. Suppose there is a law of motion (in this example time is discrete) that takes a configuration and in the next instant replaces is by another unique configuration. For example R B, B Y, Y G, G O, O P, P R. I can then ask what fraction of the time does that die spend in each configuration? The answer is 1/6. In fact there are many possible laws for which the answer will be the same. For example, R B, B G, G P, P O, O Y, Y R or R Y, Y P, P G, G O, O B, B R. But what about the law R B, B G, G R, P O, O Y, Y P. In this case there are two trajectories through the space of states. If we are oN one of them we dont jump to the other. So the probability will depend on the relative probability for beginning on the two cycles. In this last case there is a conserved quantity. Suppose we assign the number 1 to R,B,G and 0 to O,Y,P. Lets call this quantity the Zilch. Obviously Zilch is conserved. Whenever we have a conserved quantity like Zilch, either we have to specify its value. Thats ok, there are conserved quantities in nature, the most important in statistical mechanics being energy. But once we specify all the conserved Zilches, we can proceed as usual, and say that all the states on a trajectory of fixed Zilch-numbers are equally probable. That sounds good, but there are lots of counterexamples. Here is one. R R, B R, G R, P O R, Y R. No matter where you begin, you go to red in the next instant. There are no R,
conserved quantities but obviously the probability after a short time is completely unequal; only red is possible.
The entropy defined by 1.1 roughly measures the log of the number of states that have nonnegligible probability in the distribution P(i). In other words exp S is the width of the distribution. Note that it is zero if and only if P = 1 for a single state and vanishes for all others. Homework Problems: 1. Suppose you have a set of N coins. Each coin independently has a probability i/3 to be heads and 2/3 to be tails. What is the total entropy? 2. A variable q has a uniformly spaced spectrum of values with very small spacing S. The sum over states can be accurately approximated by: Suppose that the probability distribution for q is proportional to entropy?
What is the
1.4 Temperature
The average energy associated with a probabilty distribution (I will just call it E) is given by 1.3 where Ei is the energy of the state i. Now suppose we have a one-parameter family of probability distributions labeled by the average energy P(i; E). For each value of E P(i; E) satisfies the usual requirements of a probability distribution. Later we will think of it as the thermal equilibrium distribution for given average E. But for now it is just a one-parameter family. At each value of E we can compute the entropy so that S becomes a function of E, S(E) Consider the amount of energy that is needed to increase the entropy by one bit (by log 2). It is given by 1.4 We call the quantity dE/dS the temperature, T. 1.5 Slogan: Apart from a factor of log 2, the temperature is the amount of energy needed to increase the entropy by one bit. For example, if you erase a bit of information from your computer you are really transferring it from the computer out into the atmosphere, where it shows up as some heat. How much heat? The answer is T Iog 2. Except in very unusual circumstances, the temperature is always positive, i,e,. entropy is a monotonically increasing function of energy.
From the 2nd law we can prove that heat always flows from hot to cold. Consider two isolated systems , A and B, at different temperatures. Let them have energies, temperatures, and entropies EA, EB, TA, TB, SA, SB Without loss of generality we can assume that TB> TA Now bring them into contact so that energy (heat) flows between them. Suppose a small quantity of energy is exchanged. The total change in energy must be zero. Therefore 1.6 Since they must tend to equilibrium, if the entropy is not maximum it must increase. Hence 1.7 We can use 1.6 to eliminate SB from 1.7. We find 1.8 Since the system B is initially the hotter of the two, (TB TA) is positive. Therefore dSA and also TAdSA are positive. Equation 1.6 then tells us that TBdSB is negative. Equivalently, energy flows from hot to cold as equilibrium is established. The final equilibrium configuration in which energy has stopped flowing, must have TA = TB. In other words, temperature must be uniform in a system in thermal equilibrium.
2.3
# of distinct configurations of
The important point is that when N and n become large, subject to the constraints 2.1 and 2.2, then the quantity in 2.3 becomes very sharply peaked around some set of occupation numbers. Before we compute the occupation numbers that maximize 2.3, let me introduce some changes of notation. let P(i) to be the fraction of copies in state i. E be the average energy of a copy let Then 2.1 and 2.2 take a form identical to 1.2 and 1.3, namely 2.1a 2.2a Now we will assume that N and n are very large and use Stirlings approximation ( ) 2.3. But first let us take its logarithm (Maximizing a positive quantity is the same as maximizing its log.) We want to maximize this subject to 2.1 and 2.2. Substituting n/N = P and Etotal = NE we find that this is equivalent to maximizing subject to and In other words the probability distribution for thermal equilibrium maximizes the entropy subject to the constraint of a given average energy. In order for find the P(i) we use the method of Lagrange multiplier, to Implement the two constraints. The two multipliers are called and . Thus we maximize 2.4 At the end we choose , so that the constraints are satisfied. Differentiating with respect to P(i) and setting the result to zero gives 2.5 This is the Boltzmann distribution. Let us define Then 2.5 has the familiar form 2.6 Boltzmann distribution
and we obtain
The Lagrange multiplier is, of course, the inverse temperature. We will demonstrate this, but first let us derive another familiar thermodynamic formula.
The quantity A = -T log Z is called the Helmholtz free energy (reminder: T = 1/ .) Thus we find or 2.9 The Helmholtz free energy satisfies
Using dE = TdS we find 2.10 2.3 Why is T = 1/ ? We have proposed two definitions of temperature. The first is equation 1.5 (T=dE/dS)the second is the inverse of the Lagrange multiplier (T=1/ ) We would like to see that they are really the same. Consider a small change in the energy of a system,
Now use 2.9 in the form E = ST T logZ to get Using 2.9 again, and the definition of A, we find that the second and third term cancel, leaving 2.11 This is of course equivalent to 1.5. Thus the two definitions of temperature are the same.
3 Fluctuations
So far we have been deriving classical thermodynamics from statistical mechanics. We go beyond thermodynamics when we consider fluctuations of quantities about their averages. Such fluctuations are observablefor example Einsteins theory of the Brownian motion. In this section I will illustrate by considering the fluctuations of the energy of a system in contact with a heat bath. Given a probability distribution P(x) the fluctuation in x (called x)is defined by 3.1 which is also equal to 3.2 where means average (langle rangle) For any function f(x) the average is defined by sum is replaced by integral in the obvious way. . If x is continuous then the
Let us consider the fluctuation of energy of a system in equilibrium. We use the following: and The first identity is the usual identification of average energy in terms of the derivative of Z. The second identity is derived the same way as the first, noting that each derivative acting on e(- E) brings down a factor of -E. Thus is given by
Now note that (dE/dT) the specific heat of the system. Call it C. The final identity is 3.3 Thus we find that the fluctuation of the energy is proportional to the specific heat.
4 Control Parameters
So far we have considered closed systems characterized by constant values of the parameters. These parameters include all the parameters in the Lagrangian such as the masses of particles, the values of external electric and magnetic fields, and the shape and volume of the containers that enclose the system. Some of these parameters such as the volume of the system and the external fields may be controllable from the outside, for example by moving pistons to change the volume. We will call such macroscopic control variables Xm. For simplicity we will consider the case of only one X although the principles are the same for several of them. If you want to think of a specific example, X can represent the volume of the system.
In fact the levels will not cross over or disappear, if the change is adiabatc The implication is that the probability function P(i) is constant for each level, even as the energy along the way gradually varies. y Obvionsly if P(i) is constant, so is the entropy. That is why entropy is an adiabatic invariant. y y Now consider the change in energy of the system during the adiabatic process. That energy change is by definition, the work done on the system by changing X. The most familiar example is the work done in slowly compressing a gas in a insulated container. If the change in X is small (call it dX) we may assume the work done is small (call it dW). We can express the above idea in equations. In the general case of several control parameters this becomes Let us define the conjugate variables, Yn, to the Xn, by the formula 4.1 and it follows that 4.2 The most familiar example of (X, Y) is volume and pressure (V,P).
Let us suppose that an infinitesimal adiabatic change is followed by a second process in which energy is added to the system in the form of heatin other words a second process in which the control parameters are constant but the entropy changes. For this second process dE=TdS so that the combined effect of the work (adiabatic process) and the added heat give a change in energy, 4.3 First Law of Thermodynamics This relation is called the First Law of Thermodynamics, but it is really an expression of energy conservation.
Homework Problem: Prove the following identity: This identity is general but in the case where S, T, E have their usual thermodynamic meaning, we can use
to get
The x integrals in 5.2 are trivial since the integrand does not depend on x. It just gives V3N When combined with the 1/N! it gives
Since N/V is the particle density, call it p these factors combine to give
Notice how the N and V dependence nicely combine to give an expression which only depends on the density which we will keep fixed as the nnmber of particles tends to infinity. The momentum integral in 5.2 is a gaussian integral over 3N variables. In fact it is the 3N power of the one dimensional integral The final result for the partition function is 5.3 If we want to explicitly exhibit the dependence on volume we replace p by N/V. 5.4 Homework Problem: y Given the partition function in 5.4, compute the A, E, S, P as functions of the temperature. Derive the energy per particle and the ideal gas law P = pT. y What is the average speed (magnitude of velocity) of a particle? y On a PV diagram (pressure on one axis, volume on the other) what are the curves of constant temperature (isotherms) and constant entropy (adiabats)?