Next: About this document ...
LECTURE 2
BASIC CONCEPTS
States of a System
Let's consider how we specify the state of a system with N particles
in both classical mechanics and quantum mechanics.
In classical mechanics if we have a single particle in one dimension
then we can describe the system completely by specifying the position
coordinate
and the momentum coordinate
. We can represent this
graphically by labeling one axis with
and one axis with
:
=2.0 true in
We call the space spanned by
and
``phase space.'' It's the
space in which the point
exists.
If our system has N particles and exists in 3D, then we must provide
and
for all N particles. Since
and
are each 3 dimensional vectors, if we want to represent the
system as a point in phase space, we need 6N axes. In other words
phase space is 6N dimensional. The coordinates of the point representing
the system in phase space are
.
Spatial coordinates and momenta are continuous variables. To obtain a
countable number of states, we divide phase space into little boxes
or cells. For our one particle in 1D example, the volume of one of these
cells is
 |
(1) |
where
is some small constant having the dimensions of angular
momentum. The state of the system can then be specified by stating that
its coordinate lies in some interval between
and
and
between
and
.
For our N particle system,
spatial coordinates
and
momentum coordinates are required to specify the system. So the
volume of a cell in phase space is
 |
(2) |
Each cell in phase space corresponding to a state of the system can be
labeled with some number. The state of a system is provided by specifying
the number of the cell in phase space within which the system is located.
One microscopic state or microstate of the system
of N particles is defined by specifying
all the coordinates and momenta of each particle. An N particle system
at any instant of time is specified by only one point in a 6N
dimensional phase space and the corresponding microstate by
the numerical label of the cell in which this point is located.
As the system evolves in time, the coordinates of the particles
change, and the N particle system follows some trajectory in this 6N
dimensional phase space.
A macroscopic state or macrostate of the system is determined
by only a few macroscopic parameters such as temperature, energy,
pressure, magnetization, etc. Note that a macrostate contains much
less information about a system than a microstate. So a given macrostate
can correspond to any one of a large number of microstates.
How would quantum mechanics be used to describe a system? Any system of
N interacting particles can be described by a wavefunction
 |
(3) |
where the
are the appropriate ``coordinates'' for the N particles.
The coordinates include both spin and space coordinates for each particle
in the system. A particular state (or a particular wavefunction) is then
specified by providing the values of a set of quantum numbers
.
This set of quantum numbers can be regarded as labelling this state.
Different values of the quantum numbers correspond to different states.
For simplicity let's just label the states by some index
, where
1, 2, 3, ... The index
then labels the different microstates.
In quantum mechanics,
is replaced by Planck's constant
.
Thus both classical and quantum mechanics lead to a countable number of
microstates for an N particle system.
Ensemble
Statistical mechanics is based on probability considerations, and
averages over appropriate quantities. Thus one approach to statistical
mechanics is a consideration of a large number of identically prepared
systems, all subject to the same initial conditions and the same
set
of external parameters
such as the total energy, particle number, and volume. This hypothetical
collection of identical systems is called an ensemble. The
systems in the ensemble will, in general, be in different states and will,
therefore, also be characterized by different macroscopic parameters
(e.g., by different values of pressure or magnetic moment). (These
macroscopic parameters are not the ones in set
, but
they may be conjugate to them. For example pressure
is
conjugate to volume
because
is the work done by pressure
in changing the volume by
.) We can
calculate the probability of the occurrence of a particular value
of such an external parameter, i.e., we can determine the fraction of
cases in the
ensemble when the parameter assumes this particular value. For example
we can calculate the average pressure. Another
way to say this: any variable
or property that we are attempting to calculate will then be obtained
from an averging procedure over all members of the ensemble.
The aim of theory will be to predict the probability of occurrence in the
ensemble of various values of such a parameter on the basis of some
basic postulates.
This concept, and the term enemble, were introduced by J. W. Gibbs, an
American physicist around the turn of the 20th century.
Another basic approach to statistical mechanics, proposed by Boltzmann
and Maxwell, is known as the ergodic hypothesis. According to
this view, the macroscopic properties of a system represent averages
taken over the microstates traversed by a single system in the course
of time. It is supposed that systems traverse all the possible
microstates fast enough that the time averages are identical with
the averages taken over a large collection of identical and independent
systems, i.e., an ensemble. This is the idea behind Monte Carlo simulations.
Basic Postulates of Statistical Mechanics
To make any progress, we need some basic postulates about the relative
probability of finding a system in any of its accessible states. Usually
one only has partial information about a system. We don't know every
single thing about every particle. The states which are compatible with
the information that we have about the system are called ``accessible
states.'' The accessible states don't violate or contradict any
information that we have about the system. Now consider a thermally isolated
system. It cannot exchange energy with the system so its total energy
is fixed or conserved. We would like to make some statements about the
system in equilibrium. When the system is in equilibrium, things are not
changing in time. The macroscopic parameters are time-independent.
The probability of finding the system in any one state is independent
of time, i.e., the representative ensemble is the same irrespective of time.
This leads to the fundamental postulate of statistical mechanics:
An isolated system in equilibrium is equally likely to be in any of
its accessible states. In other words if phase space is subdivided into
small cells of equal size, then an isolated system in equilibrium is
equally likely to be in any of its accessible cells.
Certainly this seems reasonable. There is no reason for one microstate to
be preferred over another, as long as each microstate is consistent with
the macroscopic parameters. There are also more rigorous reasons to accept
this postulate. It's a consequence of Liouville's theorem (see Appendix
13 of Reif) that if a representative ensemble of such isolated systems
are distributed uniformly over their accessible states at any one time,
then they will remain uniformly distributed over these states forever.
One can think of the accessible states as the ``options'' that a system
has available to it. Lots of accessible states means lots of possible
microstates that the system can be in.
What if an isolated system is not equally likely to be found in any of
the states accessible to it? Then it is not in equilibrium. But it will
approach equilibrium. The system will make transitions between
all its various accessible states as a result of interactions
between its constituent particles. Once the system is equally likely to
be in any of its accessible states, it will be in equilibrium, and it
will stay that way forever (at least as long as it is isolated). The
idea that a nonequilibrium system will approach equilibrium is a
consequence of the H theorem (Appendix 12 in Reif). If we think in
terms of an ensemble of systems distributed over the points in phase space
in some arbitrary way, then the ensemble will evolve slowly in time
until phase space is uniformly occupied. The characteristic time
associated with attaining equilibrium is called the ``relaxation time.''
The magnitude of the relaxation time depends on the details of the system.
The relaxation time can range from less than a microsecond to longer
than the age of the universe (e.g., glass). Indeed the glass transition is
a good example of a system falling out of equilibrium because
the experimenter cannot wait long enough for the system to equilibrate.
Calculating the rate of
relaxation toward equilibrium is quite difficult, but once equilibrium
is reached and things become time-independent, the calculations become
quite straightforward. For example, many of the properties of the
early universe have been calculated using the assumption that things were
in equilibrium.
Probability calculations
From this basic postulate, how do we calculate various quantities of
interest? Let us consider a system of total energy between
and
. Let
be the total number of microstates
that satisfy this condition. Suppose that
is the number of
states contained within
that are characterized by the parameter
having the value
. For example,
might be the magnetic moment
of the system or the pressure exerted by the system. Since all states
are equally likely, we have for the probability
that the
parameter
of the system assumes the value
 |
(4) |
To calculate the mean value of the parameter
of the system, we
simply take the average over the systems in the ensemble; i.e.,
Here the sum over
denotes a sum over all possible values which the
parameter
can assume. Note that to calculate the average value of
the parameter
, we simply need to count states. However, this
may be highly nontrivial.
Density of States
Density of states is a useful concept. A macroscopic system, like a cup
of coffee or a block of copper, has a great many degrees of freedom.
Let
be the energy of the system. Suppose we divide up the energy
scale into small regions, each of size
, where
is much larger than the spacing between energy levels but macroscopically
small. Let
be the number of states whose energy lies between
and
. Then
must be proportional to
and we can write
 |
(6) |
where
is the ``density of states''. (Your book writes it as
.) The density of states is a characteristic property of the system
which measures the number of states per unit energy range. For example
one could have the number of states per eV. The density of states is
an important concept in systems with many particles. For example in a simple
metal where electrons conduct electric current, the density of electron states
near the Fermi energy determines how good a conductor the metal is.
If the density of states is high, then the metal is a good conductor because
the electrons near the Fermi energy will have a lot of empty states to choose
from when they hop. If
is small, then the metal is a poor conductor
because the electrons will not have many empty states to hop to. The density
of states is often useful in converting sums into integrals over energy:
 |
(7) |
Interaction Between Macroscopic Systems
Macroscopic systems are described by specifying some macroscopically
measurable independent parameters like the volume V or the applied
external electric and magnetic field. Now
consider two macroscopic systems A and A
which can interact with one
another so that they can exchange energy. Their total energy
remains constant since the combined system A
consisting of A and A
is isolated. They can interact in two ways:
mechanically and/or thermally. If they interact thermally, they exchange
heat but the energy levels of the systems do not change though
their occupation might. If they interact mechanically, the external
parameters (like volume) are changed and some of the energy levels are
shifted.
Thermal Interaction
Let's consider the thermal interaction. In a purely thermal
interaction, energy is transferred from one system to the other.
If we have an ensemble of interacting systems (A+A
), the
amount of energy transferred to each system A is not exactly the same
for the different members of the ensemble. We can
however talk in terms of the change in the mean energy of each
of the systems. This is called ``heat.'' More precisely, the change
of the mean energy of system A is called the
``heat
absorbed'' by this system; i.e.,
 |
(8) |
Heat is energy transfer.
The heat can be positive or negative.
is the heat given off by
a system;
is the heat absorbed by the system. Since the total energy
is unchanged
 |
(9) |
or
 |
(10) |
or
 |
(11) |
This is just conservation of energy.
Mechanical Interaction
Now suppose that the systems A and A
cannot interact thermally, i.e.,
they are themally isolated. However they can interact mechanically.
For example A
could do work on A. Consider a cylinder separated into two
parts by a movable piston. Let A be the gas in one part and A
be the
gas in the other part. Suppose A
expands, moves the piston and
compresses the gas in A.
This changes the energy of A by heating up the gas.
As before we think of an ensemble of identical systems and speak in
terms of the change in the mean energy. If the change in the mean
energy due to the change of the external parameters is denoted by
, then the macroscopic work done on the
system is defined as
 |
(12) |
The macroscopic work
done by the system is the negative of
:
 |
(13) |
Conservation of energy dictates that
 |
(14) |
or
 |
(15) |
Doing work on a system changes the positions of the energy levels
and the occupation of different states.
Generalized Force
In introductory physics we defined work as the force on an object times
the distance it moves (``force times distance''). Now we have a system
with
particles. How do we define work? ``Pressure times volume.''
We mentioned earlier that when something changes the volume by applying
pressure, the mean energy changes and work has been done on the system.
Pressure has units of force per unit area. The gas pushes on a wall and
produces pressure which is the force per unit area on the wall. Notice
that
has the same units as energy/volume. In fact the definition
of pressure is
 |
(16) |
(We should keep the entropy fixed in this derivative.)
We can make this more formal. When we say ``macroscopic work,'' we mean more
than
or
where
is a force. Let the energy of some
microstate
depend on external parameters
.
 |
(17) |
Then when the parameters are changed by infinitesimal amounts, the
corresponding change in energy is
 |
(18) |
The work
done by the system when it remains in this
particular state
is
then defined as
 |
(19) |
where
 |
(20) |
This is called the ``generalized force'' conjugate to the
external parameter
in the state
. Note that if
denotes a distance, then
simply is an ordinary force.
Consider an ensemble of similar systems. If the external parameters
are changed quasi-statically so that the system remains in equilibrium
at all times, then we can calculate the mean value averaged over all
accessible states
 |
(21) |
where
 |
(22) |
is the mean generalized force conjugate to
.
Note that
The macroscopic work
resulting from a finite quasi-static change of
external parameters can then be obtained by integration.
Examples
- Force times distance
 |
(24) |
where
is the linear dimension and
is the ordinary force in the
direction.
- Pressure times volume
 |
(25) |
where
is the average pressure and
is volume. We wrote
down the expression for
pressure before, but now we can be more precise. The pressure is the
generalized force associated with changes in volume.
 |
(26) |
or
 |
(27) |
where
is the macroscopic energy and
is the volume.
Your book talks about quasi-static processes in which the process occurs
so slowly that the system can be regarded as being in equilibrium throughout.
For example, the piston can be moved so slowly that the gas is always
arbitrarily close to equilibrium as its volume is being changed.
In this case the mean pressure has a well-defined meaning. If the
volume is changed by an infinitesimal amount
, then the work done
is
 |
(28) |
If the volume is changed from an initial volume
to a final volume
, then the macroscopic amount of work done is given by
 |
(29) |
This integral depends on the path taken from the initial to the final
volume. It is not path independent. So
is not an exact differential.
(Recall in electromagnetism, the potential difference is path
independent.)
is not the difference of 2 numbers referring to 2
neighboring macrostates; rather it is characteristic of the process
of going from state
to state
. Similarly the infinitesimal
amount of heat
absorbed by the system in some process is also
not an exact differential and in general, will depend how the process
occurs.
General Interaction between 2 Systems
In general two systems interact both thermally and mechanically.
Let
be the heat absorbed by the system and let
be
the work done by the system. Then the change in the mean
energy
is given by
 |
(30) |
This is the first law of thermodynamics.
If we write
 |
(31) |
then we can view the heat
as the mean energy change not due to a change in
the external parameters. For infinitesimal changes, we can write
 |
(32) |
Note that
is an exact differential. The change in
the mean energy is independent of the path taken between the initial
and final states. The energy is characteristic of the state, not of
the process in getting to that state.
For example, suppose we push a cart over a bumpy road to the top of
a hill. Let us suppose there are 2 roads to the top of the hill.
How much work we do and how much is lost to friction and heat depends
on which road we take and how long the road is. However, at the end
of our journey at the top of the hill, the (potential) energy is
independent of the road we chose. This is why
and
are inexact
differentials but
is an exact differential.
Note that if
,
is an exact differential. So if
, then
. On the other hand, if
,
and
is an exact differential. So if
,
.
Exact and Inexact Differentials
We must now make a small digression to remind ourselves of the difference
between exact and inexact differentials. Consider any function of
two independent variables
. Then the differential
is
defined by
and
 |
(34) |
Note that the integral of an exact differential depends only on the endpoints
(initial and final points) and not on the path of integration.
However, not every function is an exact differential. Consider
 |
(35) |
It is not guaranteed that there will exist a function
such
that
 |
(36) |
That is, it is not always true that
 |
(37) |
is independent of the path between the endpoints. The integral may depend
on the path of integration. As an example, consider
 |
(38) |
It is easy to show that
 |
(39) |
and
 |
(40) |
=2.0 true in
Note however that if
 |
(41) |
then
is an exact differential with
 |
(42) |
and
 |
(43) |
independent of path. The factor
is called an integrating factor for
.
Next: About this document ...
Clare Yu
2009-03-30