Next: About this document ...
LECTURE 18
The Ising Model
(References: Kerson Huang, Statistical Mechanics, Wiley and Sons (1963) and
Colin Thompson, Mathematical Statistical Mechanics, Princeton Univ. Press
(1972)).
One of the simplest and most famous models of an interacting system is
the Ising model. The Ising model was first proposed in Ising's Ph.D thesis
and appears in his 1925 paper based on his thesis (E. Ising, Z. Phys.
31, 253 (1925).) In his paper Ising gives credit to his advisor
Wilhelm Lenz for inventing the model, but everyone calls it the Ising model.
The model was originally proposed as a model for ferromagnetism. Ising
was very disappointed that the model did not exhibit ferromagnetism
in one dimension, and he gave arguments as to why the model would not
exhibit ferromagnetism in two and three dimensions. We now know that
the model does have a ferromagnetic transition in two and higher dimensions.
A major breakthrough came in 1941 when Kramers and Wannier gave a
matrix formulation of the problem. In 1944 Lars Onsager gave a complete
solution of the problem in zero magnetic field. This was the first
nontrivial demonstration of the existence of a phase transition from the
partition function alone.
Consider a lattice of
sites with a spin
on each site. Each spin
can take one of two possible values:
for spin up and
for spin
down. There are a total of
possible configurations of the system.
A configuration is specified by the orientations of the spins on all
sites:
.
is the spin on the
th lattice site.
The interaction energy is defined to be
 |
(1) |
where the subscript
represents the Ising model. A factor of 2 has
been absorbed into
and we set
in the last term.
means nearest neighbor pairs of spins. So
is the same as
.
is
the exchange constant; it sets the energy scale. For simplicity,
one sets
equal to a constant
. If
, then the spins
want to be aligned parallel to one another, and we say that the
interaction is ferromagnetic. If
, then the spins want to
be antiparallel to one another, and we say that the interaction is
antiferromagnetic. If
is a random number and can either be
positive or negative, then we have what is called a spin glass.
For simplicity we will set
and study the ferromagnetic
Ising model. The last term represents the coupling of the spins to
an external magnetic field
. The spins are assumed to lie along the
z-axis as is the magnetic field
. The spins lower
their energy by aligning parallel to the field. I put
to indicate
the possibility that the field could vary from spin to spin. If the
field
is random, this is called the random field Ising model.
We will assume a constant uniform magnetic field so that
.
So the interaction energy becomes
 |
(2) |
The partition function is given by
 |
(3) |
One Dimensional Ising Model and Transfer Matrices
Let us consider the one-dimensional Ising model where
spins are on a
chain. We will impose periodic boundary conditions so the spins are on a ring.
Each spin only interacts with its neighbors on either side and with the
external magnetic field
. Then we can write
 |
(4) |
The periodic boundary condition means that
 |
(5) |
The partition function is
![\begin{displaymath}
Z=\sum_{s_1=-1}^{+1}\sum_{s_2=-1}^{+1}...\sum_{s_N=-1}^{+1}
...
...ft[\beta \sum_{i=1}^{N}\left(JS_i S_{i+1}+BS_{i}\right)\right]
\end{displaymath}](img27.png) |
(6) |
Kramers and Wannier (Phys. Rev. 60, 252 (1941))
showed that the partition function can be expressed in terms of matrices:
![\begin{displaymath}
Z=\sum_{s_1=-1}^{+1}\sum_{s_2=-1}^{+1}...\sum_{s_N=-1}^{+1}
...
...i S_{i+1}+\frac{1}{2}B\left(S_{i}+S_{i+1}\right)\right)\right]
\end{displaymath}](img28.png) |
(7) |
This is a product of
matrices. To see this, let the matrix
be
defined such that its matrix elements are given by
![\begin{displaymath}
\langle S\vert P\vert S^{\prime}\rangle=\exp\left\{\beta\left[JSS^{\prime}
+\frac{1}{2}B(S+S^{\prime})\right]\right\}
\end{displaymath}](img31.png) |
(8) |
where
and
may independently take on the values
.
Here is a list of all the matrix elements:
Thus an explicit representation for
is
 |
(10) |
With these definitions, we can write the partition function in the form
where
and
are the two eigenvalues of
with
. The fact that
is the trace of the
th power of a matrix is a consequence of the periodic boundary condition
Eq. (5). The eigenvalue equation is
 |
(12) |
Solving this quadratic equation for
gives
![\begin{displaymath}
\lambda_{\pm}=e^{\beta J}\left[\cosh(\beta B)
\pm\sqrt{\cosh^{2}(\beta B)-2e^{-2\beta J}\sinh(2\beta J)}\right]
\end{displaymath}](img53.png) |
(13) |
When
,
Now back to the general case with
.
Notice that
where equality is in
the case of
. In the thermodynamic
limit (
), only the larger eigenvalue
is
relevant. To see this, we use
and write the Helmholtz free energy per spin:
So the Helmholtz free energy per spin is
The magnetization per spin is
At zero field (
), the magnetization is zero for all temperatures.
This means that there is no spontaneous magnetization and the one-dimensional
Ising model never exhibits ferromagnetism. The reason is that at any
temperature the average configuration is determined by two opposite
and competing tendencies: The tendency towards a complete alignment
of spins to minimize the energy, and the tendency towards randomization
to maximize the entropy. The over-all tendency is to minimize the free
energy
. For the one-dimensional model the tendency for alignment
always loses out, because there are not enough nearest neigbors. However,
in higher dimensions, there are enough nearest neighbors and a ferromagnetic
transition can occur.
The method of transfer matrices can be generalized to two and higher
dimensions, though the matrices become much larger. For example,
in two dimensions on an
square lattice, the matrices are
. In 1944, Onsager solved the two dimensional Ising model
exactly for the zero field case, and found a finite temperature
ferromagnetic phase transition. This is famous and is
known as the Onsager solution of the 2D Ising model.
No one has found an exact solution for the three dimensional Ising model.
Applications of the Ising Model
The Ising model can be mapped into a number of other models. Two of the
better known applications are the lattice gas and the binary alloy.
Lattice Gas
The term lattice gas was first coined by Yang and Lee in 1952, though
the interpretation of the model as a gas was known earlier. A lattice
gas is defined as follows. Consider a lattice of
sites (
= volume)
and a collection of
particles, where
. The particles are placed on the
vertices of the lattice such that not more than one particle can
occupy a given site, and only particles on nearest-neighbor lattice
sites interact. The interaciton potential between two lattice sites
and
is given by
with
 |
(19) |
where
is the lattice spacing.
The occupation
of a lattice site
is given by
 |
(20) |
The interaction energy is
 |
(21) |
We can map this into the Ising model by letting spin-up denote an occupied
site and letting spin-down denote an unoccupied site. Mathematically, we
write
 |
(22) |
So
means site
is occupied and
means site
is unoccupied.
One can then map the lattice gas model into Ising model. For example, by
comparing the partition functions, it turns out that
.
Binary Alloy
A binary alloy is a solid consisting of 2 different types of atoms. For example,
brass is a body-centered cubic lattice made up of Zn and Cu atoms.
At
, the lattice is completely ordered and a copper atom is
surrounded by zinc atoms and vice-versa. However, at non-zero temperatures
the zinc and copper atoms can exchange places. Above a critical temperature
of
K, the Zn and Cu atoms are thoroughly mixed so that the
probability of finding a Zn atom on given site is 1/2. Similarly
the probability of finding a Cu atom on given site is 1/2.
To model a binary alloy, one starts with a lattice of
sites, and two
different types of atoms, A and B. Each site has only one atom so
that
. The occupation of each site is
 |
(23) |
There are interaction energies between nearest neighbor sites:
,
and
.
One can map the binary alloy model into the lattice gas model and the Ising model.
Generalizations to Other Spin Models
One can generalize the Ising model in a number of ways, though the Ising spins
refer to entities with two possible values. Classically other types of spins are
spins which can rotate in the
plane and have a fixed
length (usually
). They have two components:
and
.
Heisenberg spins are fixed length spins that can point anywhere on a
unit sphere. Heisenberg spins have three components:
,
, and
.
Quantum mechanically, the values of
are discretized. So Ising spins
correspond to
.
where
are the Pauli spin matrices.
We have already mentioned
that the Ising model can be considered in higher dimensions.
Another variation of the Ising model is to consider other types
of lattices. For example in two dimensions, we could have a triangular
lattice. In three dimensions there are a wide variety of lattice
structures that could be considered. Or one could throw away the lattice
and have randomly placed sites to make a type of spin glass.
Other variations center around the form of the interaction.
For example, we can allow the nearest neighbor interactions to be antiferromagnetic.
Or we can allow the interactions to extend over a longer range to include
next nearest-neighbors or even farther, e.g., infinite range. The interaction
is contained in the exchange constant
. So one could have
something like
 |
(24) |
where
is the Coulomb interaction and
is similar to a dipolar
interaction.
Another interaction is the RKKY interaction. RKKY stands for
Ruderman-Kittel-Kasuya-Yosida. The RKKY interaction has the form
where
is the Fermi wavevector.
This interaction is found in metals with magnetic atoms. The interaction
is mediated by the conduction electrons.
Notice that the interaction oscillates and decays as a power law.
Frustration and Spin Glasses
Magnetic impurities randomly placed in a metal, e.g., Mn impurities in
copper, will interact with one another via the RKKY interaction. Because
of the oscillations, the interactions will be random. This spin system
is called a spin glass. For simplicity, the RKKY interaction is replaced by
a random
in the spin Hamiltonian:
 |
(26) |
Typically,
is chosen from a distribution
centered at
.
For example,
where
is a positive constant, and there is
an equal probability of choosing the plus or minus sign. Another possibility
is to have
be a Gaussian distribution centered at
. Obviously
the ground state of a such a system will be disordered, but a phase
transition from a paramagnetic phase at high temperatures to a
frozen spin configuration is possible.
=1.0 true in
One concept that is associated with spin glasses is ``frustration.''
The idea is best illustrated by considering an equilateral triangle with
a spin on each vertex. Suppose the interaction between nearest neighbor
spins is antiferromagnetic so that a spin wants to point opposite from its
neighbors. There is no way all the spins on the triangle can be satisfied.
This is an example of frustration. Frustration occurs when there is
no spin configuration where all the spins have their lowest possible
interaction energy. In a spin glass, there is a great deal of
frustration. As a result there is no clear ground state configuration.
One can associate an energy landscape with the energies of different
spin configurations. Valleys correspond to low energy configurations
and mountains to high energy configurations. The landscape exists in
the space of spin configurations. So to go from one valley to another,
the system must climb out of the first valley by going through some
high energy configurations and then descend into the second valley
by passing through configurations with decreasing energy.
=3.0 true in
Spin glasses are often used to model interacting systems with randomness.
They were originally proposed to explain metals with magnetic impurities
and were thought to be a simple model of a glass. Spin glass models have a wide
range of applications, e.g., they have been used to model the brain and
were the basis of neural networks and models of memory.
Monte Carlo Simulations
Reference: D. P. Landau and K. Binder, A Guide to Monte Carlo Simulations
in Statistical Physics, Cambridge Univ. Press (2000).
As one can see, analytic solutions to spin systems can be difficult to obtain.
So one often resorts to computer simulations, of which Monte Carlo is
one of the most popular. Monte Carlo simulations are used widely in
physics, e.g., condensed matter physics, astrophysics, high energy physics, etc.
Typically in Monte Carlo simulations,
one evolves the system in time. The idea is to visit a large number
of configurations in order to do statistical sampling. For a spin system
we would want to obtain
average values of thermodynamic quantities such as magnetization,
energy, etc. More generally, Monte Carlo is an
approach to computer simulations in which an event
occurs with a
certain probability
where
. In practice,
during each time step, a random number
is generated with uniform
probability between 0 and 1. If
, event A occurs; if
, event
does not occur. Monte Carlo is also able to
handle cases where multiple outcomes are possible. For example,
suppose there are three possibilities so that either event
can
occur with probability
or event
can occur with probability
or neither occurs. Then if
,
occurs; if
,
occurs; and if
,
neither occurs.
The Metropolis algorithm (Metropolis et al., J. Chem. Phys.
21, 1087 (1953)) is a classic Monte Carlo method.
Typically, configurations are generated from a previous state using
a transition probability which depends on the energy difference
between the initial and final states. For relaxational
models, such as the (stochastic) Ising model, the probability obeys
a master equation of the form:
![\begin{displaymath}
\frac{\partial P_n(t)}{\partial t}=\sum_{n\neq m}
\left[-P_n(t)W_{n\rightarrow m}+P_m(t)W_{m\rightarrow n}\right]
\end{displaymath}](img134.png) |
(27) |
where
is the probability of the system being in state
at time
, and
is the transition rate from state
to state
. This is an example of a master equation. Master equations are
found in a wide variety of contexts, including
physics, biology (signaling networks), economics, etc.
In equilibrium
and the two terms on the
right hand side must be equal. The result is known as `detailed balance':
 |
(28) |
Loosely speaking, this says that the flux going one way has to equal the
flux going the other way. Classically, the probability is the Boltzmann
probability:
 |
(29) |
The problem with this is that we do not know what the denominator
is.
We can get around this by generating a Markov chain of states, i.e.,
generate each new state directly from the preceding state. If we produce
the
th state from the
th state, the relative probability is given by
the ratio
where
.
Any transition rate which satisfies detailed balance is acceptable. Historically
the first
choice of a transition rate used in statistical physics was the Metropolis form:
 |
(31) |
Here is a recipe on how to implement the Metropolis algorithm on a spin system:
- Choose an initial state.
- Choose a site
.
- Calculate the energy change
which results if the spin at
site
is overturned.
- If
, flip the spin. If
, then
- Generate a random number
such that
.
- If
, flip the spin.
- Go to the next site and go to (3).
The random number
is chosen from a uniform distribution. The states are
generated with a Boltzmann probability. The desired average of some quantity
is given by
. In the simulation, this
just becomes the arithmetic average over the entire sample of states visited.
If a spin flip is rejected, the old state is counted again for the sake
of averaging. Every spin in the system is given a chance to flip. One pass
through the lattice is called a ``Monte Carlo step/site'' (MCS). This is the
unit of time in the simulation.
For purposes of a spin model, it is easier to calculate
if we
write
where the local magnetic field
due to the other spins is
 |
(33) |
To find the change in energy, we would do a trial flip of
and easily
calculate the new energy if we know the local field
. If we accept the
flip, then we have to update the local fields of the neighboring spins.
As we said before, Monte Carlo simulations are used widely in physics as
well as other fields such as chemistry, biology, engineering, finance, etc.
Next: About this document ...
Clare Yu
2009-03-30