Dear Jim, it depends on how many different states - boxes - you have for those particles. If you still have 2 states, just like in the Wikipedia example, then for fermions, the probability is 0 everywhere - it's impossible to put more than 2 fermions to 2 states. For distinguishable particles, each particle has 50% odds to be in state 0 and 50% odds to be in state 1. So for $N$ distinguishable particles, there is $(N {\rm\, \, choose\, \,} k)/2^N$ chance for $k$ of them to be in state 0 and the remaining $N-k$ of them to be in state 1. Note that the odds sum to one, if you sum over $k$ from $0$ to $N$. For $N$ bosons, all separations to state 0 ($k$ particles there) and state 1 ($N-k$ particles there) are possible, and each of those separations only corresponds to one multi-particle state. So the odds are $1/(N+1)$ for each possibility - a direct generalization of the table that has $1/3$ in all columns for the $N=2$ bosons. So yes, this is probably what you mean by "equally distributed". The Fermi-Dirac and Bose-Einstein distributions - and both of them, with the same chances - can only be reduced to the Maxwell-Boltzmann distribution if the number of particles is high (it's not really essential) but, more importantly, the number of states in which they can be found is even much higher. Whenever there is a significant chance that the particles want to share a state etc., the classical Maxwell-Boltzmann approximation is inapplicable. So the toy model with two states 0,1 clearly doesn't allow you to approximate the Bose-Einstein or Fermi-Dirac distrtibutions by the Maxwell-Boltzmann one. This is particularly clear for the fermions because the Fermi-Dirac probabilities can't even be calculated because there is no way to arrange more than two fermions to two states. The Maxwell-Boltzmann distribution is applicable when the density of particles (number of particles divided by the number of states) is low. That's why the Maxwell-Boltzmann $\exp(-E/kT)$ is applicable as a replacement for high values of energy $E$ - because this is where a small number of particles will be found, so their arrangement won't be sensitive on their being bosons (collectivists who love siblings in the same state, more than the love between the distinguishable particles) or their being fermions (individualists who absolutely hate to share the state with others). For states at low $E$, that's where the concentration of particles per state is expected to be high, which is why you get the totally new quantum behaviors - such as the Bose-Einstein condensate in the ground state and/or the Fermi liquid - which obviously can't be described by the distinguishable statistics well. LECTURE 13 Maxwell–Boltzmann, Fermi, and Bose Statistics Suppose we have a gas of N identical point particles in a box of volume V. When we say “gas”, we mean that the particles are not interacting with one another. Suppose we know the single particle states in this gas. We would like to know what are the possible states of the system as a whole. There are 3 possible cases. Which one is appropriate depends on whether we use Maxwell–Boltzmann, Fermi or Bose statistics. Let’s consider a very simple case in which we have 2 particles in the box and the box has 2 single particle states. How many distinct ways can we put the particles into the 2 states?Maxwell–Boltzmann Statistics: This is sometimes called the classical case. In this case the particles are distinguishable so let’s label them A and B. Let’s call the 2 single particle states 1 and 2. For Maxwell–Boltzmann statistics any number of particles can be in any state. So let’s enumerate the states of the system: Single Particle State 1 2 ----------------------------------------------------------------------- AB AB A B B A We get a total of 4 states of the system as a whole. Half of the states have the particles bunched in the same state and half have them in separate states. Bose–Einstein Statistics: This is a quantum mechanical case. This means that the particles are indistinguishable. Both particles are labelled A. Recall that bosons have integer spin: 0, 1, 2, etc. For Bose statistics any number of particles can be in one state. So let’s again enumerate the states of the system: Single Particle State 1 2 ----------------------------------------------------------------------- AA AA A A We get a total of 3 states of the system as a whole. 2/3 of the states have the particles bunched in the same state and 1/3 of the states have them in separate states. Fermi Statistics: This is another quantum mechanical case. Again the particles are indistinguishable. Both particles are labelled A. Recall that fermions have half–integer spin: 1/2, 3/2, etc. According to the Pauli exclusion principle, no more than one particle can be in any one single particle state. So let’s again enumerate the states of the system: Single Particle State 1 2 ----------------------------------------------------------------------- A A We get a total of 1 state of the system as a whole. None of the states have the particles bunched up; the Pauli exclusion principle forbids that. 100% of the states have the particles in separate states. This simple example shows how the type of statistics influences the possible states of the system. Distribution Functions We can formalize this somewhat. We consider a gas of N identical particles in a volume V in equilibrium at the temperature T. We shall use the following notation:
Since the particles in the gas are not interacting or are interacting weakly, we can describe the state R of the system as having n1 particles in state r = 1, n2 particles in state r = 2, etc. The total energy of the state is
Since the total number of particles is N, then we must have
The partition function is given by
Here the sum is over all the possible states R of the whole gas, i.e., essentially over all the various possible values of the numbers n1, n2, n3, ... Now we want to find the mean number ns of particles in a state s. Since
is the probability of finding the gas in a particular state where there are n1 particles in state 1, n2 particles in state 2, etc., one can write for the mean number of particles in a state s:
We can rewrite this as
or
So to calculate the mean number of particles in a given single–particle state s, we just have to calculate the partition function Z and take the appropriate derivative. We want to calculate ns for Maxwell–Boltzmann, Bose and Fermi statistics. Maxwell–Boltzmann Statistics Let us begin by considering the classical case of Maxwell–Boltzmann statistics. In this case the particles are distinguishable but identical, so each particle has the same set of single particle energy levels. As a result we can write the partition function as
where the single particle partition function is
Then
Now we can calculate the mean occupation number
or
This is called the “Maxwell–Boltzmann distribution.” It is the same as our previous result when we applied the canonical distribution to N independent single particles in a classical system. The sum over r is a sum over single particle states. Alternative Derivation of Maxwell–Boltzmann Partition Function We can write the partition function of the gas as
Here we are summing over all possible states R of the gas, i.e., over all values
subject to the restriction
Since this is a classical system, the particles have to be considered distinguishable. Thus any permutation of two particles in different states must be counted as a distinct state of the whole gas even though the numbers n1,n2,n3,... are left unchanged. This was the case in our simple example. It is not enough to specify how many particles are in each single–particle state, but it is necessary to specify which particular particle is in which state. For a given set of values n1,n2,n3,..., there are
possible ways in which the particles can be put into the given single particle states with n1 particles in state 1, n2 particles in state 2, etc. This is the number of distinct states since the particles are distinguishable. Hence Because of (15), this last expression is just a multinomial expansion. So we can write
or
where ∑ re-βεr is just the partition function for a single particle. This is what we got before. Bose–Einstein and Photon Statistics Here the particles are to be considered as indistinguishable, so that the state of the gas can be specified by merely listing the number of particles in each single particle state: n1, n2, n3, .... Since there is no limit to the number of particles that can occupy a state, ns can equal 0,1,2,3,... for each state s. For photons the total number of particles is not fixed since photons can readily be emitted or absorbed by the walls of the container. Let’s calculate ns for the case of photon statistics. The partition function is given by
where the summation is over all values nr = 0, 1, 2, 3,... for each r, without any further restriction. We can rewrite (20) as
or
But each sum is a geometric series whose first term is 1 and where the ratio between successive terms is exp(-βεr). Thus it can be easily summed:
Hence eq. (22) becomes
or
So if we plug this into eqn. (7), we get
or
This is called the “Planck distribution.” We’ll come back to this a bit later when we talk about black body radiation. Photons are bosons, but their total number is not conserved because they can be absorbed and emitted. Other types of bosons, however, do have their total number conserved. One example is 4He atoms. A 4He atom is a boson because if you add the spin of the proton, neutron, and 2 electrons, you always will get an integer. If the number of bosons is conserved, then ns must satisfy the condition
where N is the total number of bosons in the system. In order to satisfy this condition, one slightly modifies the Planck distribution. The result is known as the Bose–Einstein distribution
where μ is the chemical potential. μ is adjusted so that eq. (28) is satisfied. Physically μ is the change in the energy of the system when one particle is added. Eqn. (29) is called the Bose–Einstein distribution function or the Bose distribution function for short. Often one writes this as a function of energy:
n(ε) is also called the Bose-Einstein distribution.
Notice that if ε < μ, then n(ε) < 0 which doesn’t make much sense. The Bose distribution only makes sense for ε > μ. We can explicitly derive (29). In order to satisfy the condition (28), one multiplies the partition function by a fudge factor exp(-αN). α is then adjusted to satisfy eqn. (28). α is an example of what is called a Lagrange multiplier. We use instead of Z because we have an extra factor of exp(-αN). is the “grand partition function” that we met when we discussed the grand canonical ensemble. This is just a product of simple geometric series. Hence
or
Recall that when we discussed the grand canonical ensemble and the grand partition function, we set
where μ is the chemical potential. We are basically assuming that we have a system of weakly interacting bosons in contact with both an energy and particle number reservoir, and that the temperature T and the chemical potential μ are fixed by the reservoir. So we can rewrite (34) to obtain the grand partition function:
We can now calculate the average value of N: μ is fixed by setting N = N where N is the total number of bosons in the system. Finally we will calculate the average number of bosons in state s: or
Note that we can recover the result for photons by setting μ = 0. Also note that
We will return to the Bose–Einstein distribution when we discuss black body radiation. Fermi–Dirac Statistics Recall that fermions have half–integer spin statistics and that at most one fermion can occupy each single particle state. This means that ns = 0 or 1. We can easily get some idea of what ns is by considering the very simple case of a system with just one single particle state. In this case
In this case the sums just have 2 terms. The denominator is
The numerator is
So we have
or
For a real system we have many single particle states and many particles. The expression for ns in this case must satisfy the condition that the number of particles is a constant:
The correct formula which satisfies this condition (46) is
Often one writes this as a function of energy:
f(ε) is called the Fermi distribution function. μ is adjusted to satisfy the constraint (46). As in the Bose–Einstein case, μ is called the chemical potential. This is basically the same as the Fermi energy. Notice that f(ε = μ) = 1∕2. This is always true of the Fermi distribution.
We can formally derive the Fermi distribution in much the same way as we derived the Bose distribution. We once again consider a system of weakly interacting fermions in contact with both an energy and a particle number reservoir. The grand partition function is given by
So the mean number of particles in the system is
or
The mean number of fermions in state s is
or
Note that
We will return to this when we discuss metals. Summary For future reference the two expressions for the average number of particles in the sth state for bosons and fermions are:
A more succinct way to write our results for the quantum statistics of ideal gases is
where the upper sign refers to Fermi statistics and the lower sign refers to Bose statistics. If the gas consists of a fixed number of particles, μ is determined by
In general the number N of particles is much smaller than the total number of single particle states s. Classical Limit Let us consider 2 limiting cases. Consider the low density limit where N is very small. The relation (60) can then only be satisfied if each term in the sum over all states is sufficiently small, i.e., if ns ≪ 1 or exp[β(εs - μ)] ≫ 1 for all states s.The other case to consider is the high temperature limit. Since β = 1∕kBT, the high temperature limit corresponds to small β. Now if β were 0, we would have
which is a disaster for both the Fermi–Dirac and Bose–Einstein cases. But β = 0 means that T = ∞. Let’s assume that the temperature is high but not infinite, so that β is small but not 0. At high temperatures, lots of high energy states are occupied. By “high energy,” I mean that εs ≫ μ. In order to satisfy the fixed N constraint of eqn. (60), it is necessary to have
such that
for all states s. (Remember that there are many more states s than particles N.) This is the same condition that came up in the low density case. We call the limit of sufficiently low concentration or sufficiently high temperature where (62) or (63) are satisfied the “classical limit.” In this limit ns reduces to
Plugging this into (60), we get
or
Thus
Hence we see that in the classical limit of sufficiently low density or sufficiently high temperature, the Fermi–Dirac and Bose–Einstein distribution laws reduce to the Maxwell–Boltzmann distribution. Relation of Z and In deriving the Bose–Einstein and Fermi–Dirac distributions, we used the grand canonical partition function. We can use to obtain an excellent approximation to the canonical partition function. We can writewhere Z(N′) is the canonical partition function for N′ particles. Since Z(N′) is a rapidly increasing function of N′ and eβμN′ is a rapidly decreasing function of N′ (for μ < 0), the grand partition function is sharply peaked at N′ = N. So we can write where the width of the peak is Δ*N′. Thus since ln is negligible. Or
This is the relation between the partition function Z and the grand partition function . Chemical Potential Since eβμN′Z(N′) is a sharply peaked function at N′ = N, we can use this to derive an expression for the chemical potential μ. Consider ln . By definition the maximum of this is given by
or
This is useful for calculating the chemical potential μ. Other Conventions for the Grand Canonical Ensemble In some books, the fugacity y is defined by
(Some books use z or λ to denote the fugacity.) The grand partition function is given in terms of the fugacity by
with Z0 ≡ 1. In the grand canonical ensemble, pressure is defined by
where V is volume. Some books define a thermodynamic potential Ω(V,T,y) by
(Do not confuse the thermodynamic potential with the number of microstates of a system, even though both are sometimes denoted by Ω.) In terms of the grand partition function, the mean number of particles N is
and the mean energy E is
(One can compare this to Eq. (70) in Lecture 9
to see the importance of noting what is kept constant and what isn’t in taking derivatives.) The first law of thermodynamics then becomes
or, in differential form,
Gibbs’ Paradox Revisited Now back to eq. (71). Plugging in for bosons from (34), we have
Similarly for fermions eq. (50) yields
We can combine these two expressions:
At high temperatures e-β(εr-μ) is small and we can expand the logarithm to obtain
where in the last step we used the high temperature limit of the Bose and Fermi distributions: Taking the logarithm of both sides of (87) yields
or
Plugging this into (86), we get The first two terms are Stirling’s approximation to N!:
The last term in (90) is the Maxwell–Boltzmann partition function.
So (90) becomes
or
This was the resolution to the Gibbs paradox. Without quantum mechanics we had to put in the factor N! by hand. But now we see that the indistinguishability of the particles comes out naturally. |