Part 2 Basics of Statistical Physics1

Embed Size (px)

Citation preview

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    1/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof1

    Part 2: Basics of Statistical PhysicsProbabilities Simple probabilities

    averages

    Composite probabilitiesIndependent eventssimple and complexMany diceProbability distributions

    Statistical MechanicsHamiltonian descriptionAverages from derivativesone and many

    Structural InvarianceIntensive and Extensive

    GaussianStatistical Variables

    Integrals and Probabilities Statistical Distributions Averages

    Gaussian random variableApproximate Gaussian integrals

    Calculation of Averages and Fluctuations The Result

    Going Slowlysums and averages in classical mechanicsmore sums and averageshomework

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    2/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Simple probabilities (reprise)

    2

    total probability =1 --> =1

    ii.2

    cubic dice 6 sides: fair dice --> all probabilities are equal -->

    r=1--> z=6--> for all values of=1/6

    mutually exclusive events described by =1,2,3,......

    probability of getting a side with number is =N/N ii.1number of times turns up=N; total number of events N

    N= N

    relative probability: relative chance that will turn up =r, e.g. fair dice have r =constant

    from r to z= r

    normalize (=fix up size) : =r/z

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    3/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Averages (reprise)

    3

    average number on a throw = < >=

    = 3.5

    Do we understand what this formula means?? How would we describe a loaded die? Anaverage from a loaded die? If I told you that =2 was twice as likely as all the other values,and these others were all equally likely, what would be the relative probability? What wouldwe have for the average throw on the die?

    general rule: To calculate the average of anyfunction f() that gives the probability that what

    will come out will be , you use the formula < f() >= f()

    ii.3

    =1/6

    =? = ? This last quantity is called the variance of

    average number on a throw= =

    (

    N

    )/N

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    4/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Composite Probabilities

    4

    and are two different kinds of events might describe the temperature on January 1, computed as N /N

    might describe the precipitation on December 31, with probabilities

    Both kinds of events are complete

    The prime indicates that the two probabilities are quite different from one another.

    =1

    =1

    Let be the probability that both will happen. The technical term for

    this is ajointprobability. The joint probability satisfies

    ,

    ,,

    =1

    (|) is the probability that event occurs if that we know that event has or willoccur. This quantity is called a conditional probability. It obeys(|) = ,/

    Something must happen, implies that

    ( | )=

    1

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    5/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Independent Events

    5

    Physically two events are independent if the outcome of onedoes not affect the outcome of the other. It is a mutual relation,ifis independent of then is independent of.

    This can then be stated in terms of conditional probabilities. If (|) isindependent of then we say and are statistically independent. After alittle algebraic manipulation, it follows that the joint probability ,obeys

    , =

    equivalently, two events are statistically independent, if the number of times bothshow up is expressed in terms of the number of times each one individually shows upas

    This can be generalized to the statement that a series of m different events arestatistically independent if the joint probabilities of the outcomes of all these events issimply the product of all the m individual probabilities.

    N, = NN

    /N

    The word uncorrelated is also used to describe statistically independent quantities.

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    6/24

    Simple and Complex

    definition: simple outcome: can happen only one way: like 2 coming up when a die in thrown

    definition: complex outcome: can happen several ways: like 7 coming up when two dice arethrown.

    One should calculate probability of complex outcome as a sum of probabilities of simpleoutcomes.

    If the simple outcomes are equally likely, probability of complex outcome is the number ofdifferent simple outcomes times the probability of a single simple outcome. There is lots ofcounting in statistical mechanics. The number of ways that something can happen is oftendenoted by the symbol W. Entropy is given by

    Entropy S=k ln W , where k=kB is Boltzmanns constant.

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    7/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Many Dice

    7

    Given two dice independent of one another: one fair and the other one unfair. Howdoes one describe the probabilities of the outcome of throwing both dice together?

    what is the chance, rolling two fair dice, that we shall roll an eleven, a seven?

    Given two fair dicewhat is the average sum and product of what turns up

    Now we roll one hundred dice all at the same time. What is the average of the sum of thedice-values. How can one define a root mean square fluctuation in this value? How big isit?

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    8/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Probability Distributions

    8

    So far we have talked about discrete outcomes. A die may take on one of six possiblevalues. But measured things are often continuous. For example, in one dimension, the

    probability that a quantum particle will be found between x and x+dx is given in terms ofthe wave function, . In this context, the squared wave function appears as|(x)|2dx

    (x)a probability density. In general, we shall use the notation for a probability density, sayingthat is the probability for finding a particle between x and x+dx. The generalproperties of such probability densities are simple. They are positive. Since the totalprobability of some x must be equal to one they satisfy the normalization condition

    (x) dx

    +

    (x) dx = 1

    For example, in classical statistical mechanics, the probability density for finding aparticle with x-component of momentum equal to p is

    2

    m1/2

    exp[

    p2

    /(2m)]

    This is called a Gaussian probability distribution, i.e. one that is based on exp(-x2). Suchdistributions are very important in theoretical physics.

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    9/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Statistical Mechanics: Hamiltonian Description (reprise)

    9

    The first idea in statistical mechanics is that the relative probability of a configurationlabeled by , is given in terms of the Hamiltonian H or equivalently the energy E() in

    that the relative probability of such a configuration is exp(-H/T)=exp[-E()/T] where Tis

    the temperature measured in energy units. (This is usually written kTwhere k is theBoltzmann constant.) The general formulation of statistical mechanics is then that

    = eE()/T / Z

    Z= e

    E()/T

    with

    For example a single atomic spin in a magnetic field has a Hamiltonian which can be writtenas H=-B, with B thez component of the magnetic field and being thez component ofthe spin. In this model, this component takes on the values =1. (Most often, one putssubscripts z on the spin and magnetic field.)

    The relative probabilities for positive and negative values of the spin are respectively eh and

    e-hwith h=B/(kT) so that the partition function is

    Z= eh + e-h = 2 cosh h

    and the probabilities are

    (=1)= eh/zand (=-1)= e-h/Zso that we can calculate

    = (=1)(1)+(=-1)(-1)=(eh + e-h )/z=tanh h

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    10/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof10

    Averages from Derivatives (reprise)

    10

    z= exp(h) = 2coshh

    homework: Read Chapters 1 and 2 in textbook.

    show that -d(ln Z) /d = = and d2(ln Z) /d2 =< (-)2> and

    d(lnz) / dh= exp(h)

    / z=< >= tanhh

    note how the second derivative gives the mean squared fluctuations.

    All derivatives of the log of the partition function are thermodynamic

    functions of some kinds. As I shall say below, we expect simple behavior

    from the log of Z but not Z itself. The derivatives described above arerespectively called the magnetization, M= and the magneticsusceptibility, , = dM/dH. The analogous first derivative with respect to isminusthe energy. The next derivative with respect to is proportional tothe specific heat, or heat capacity, another traditional thermodynamicquantity.

    d2(lnz) / (dh)2 = ( < >)2 exp(h)

    / z=< ( < >)2 >

    =1 < >2=1 (tanh h)2

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    11/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    One and Many

    11

    Imagine a material with many atoms, each with its own spin. The systemhas a Hamiltonian which is a sum of the Hamiltonia of the different atoms

    H= h

    =1

    N

    and a probability distribution

    = exp(H) / Z= (1/ Z) exp(h)=1

    N

    which is a product of pieces which belong to the different atoms. Thedifferent pieces are then statistically independentof one another. Notethat the partition function is

    Z=

    =1

    N

    exp(h) = (2coshh

    =1

    )N = zN

    so that the entire probability is a product of N pieces connected with the N atoms

    The appearance of a product structure depends only upon having a Hamiltonianwhich is a sum of terms referring to individual parts of the system

    Hamiltonian is sum stat mech probability is product statistical independence

    {} = [exp(h

    ) / z]

    ii.4

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    12/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Structural invariance

    12

    Note how the very same structure which applies to one atom exp(-H)/Z

    carries over equally to many atoms.

    This structural invariance is characteristic of the mathematical basisof physical theories. Newtons gravitational theory seemed naturalbecause the same law which applied to one apple equally applies toan entire planet composed of apples.

    This same thing works for electromagnetism.

    A wave function is the same sort of thing for one electron or many.The structure of space and time has a similar invariance property.Remember that a journey of a thousand miles starts with but a singlestep. The similarity between a single step and a longer distance is akind of structural invariance. This invariance of space is called a scaleinvariance. It is quite important in all theories of space and time.

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    13/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Intensive and Extensive

    13

    Imagine a situation in which we have many independent identical subsystems, callediid by the mathematicians for independent and identically distributed. Then

    Z( )=z()N as in equation ii.4. In this kind of system, the free energy, the entropy,and the average energy will be precisely N times that for the single subsystem.

    Quantities that are a sum or integral over the entire system are called extensive.One which define local properties or are ratios of extensive quantities are calledintensive. The former are almost always linear in N, the latter are almost alwaysindependent ofN.

    The mean square fluctuation in energy is extensive. Why?? The root mean square

    fluctuation in energy, []1/2, is not extensive. Why?? In the limiting case inwhich the number of subsystems (particles) is very large, which is the relative size ofthe average energy and the root mean square fluctuation in the energy?

    I have just told you the reason why classical thermodynamics does not consider

    fluctuations in things. Again, what is this reason?

    How big is N, say for the particles in this room?

    In terms of real MKS numbers, roughly what is the average kinetic energy of theseparticles and what is the size of its fluctuation?

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    14/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Gaussian Statistical Variables

    14

    A Gaussian random variable, X, is one whichhas a probability distribution which is theexponential of a quadratic in X.

    The sum of two statistically independent Gaussian

    variables is also Gaussian. How does the variance addup?

    A Gaussian variable is an extreme example of astructurally stable quantity.

    Central Limit Theorem: A sum of a large number ofindividually quite small random variables need not besmall, but that sum is, to a good approximation, aGaussian variable, given only that the variance of theindividual variables is bounded.

    Carl Friedrich Gauss (1777 1855)

    A Gaussian distribution has a lot of structurally invariant properties.

    (x)= [/(2)]1/2 exp[(x)2/2]

    1/ is the variance of this distribution.

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    15/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Gaussian integrals and Gaussian probability distributions

    15

    I= dx exp(ax2 / 2+bx+ c)

    Gaussian integrals are of the form

    with a, b, and cbeing real numbers, complex numbers, or matrices.They are very, very useful in all branches of theoretical physics.

    There is a canonical form for Gaussian probability distributions, namely

    (X= x) = ( 2)1/2 exp[(x < X>)2 / 2]

    For Gaussian probability distributions, there is a very important result:

    We define the probability that the random variableXwill take on the value

    betweenxandx+dxas (X=x)dxormoresimplyas(x)dx

    Notice how the that appears in the numerator of the probability distribution reappears in thedenominator of the average.

    produced by completeing the square. Here 1/is the variance and is the average of therandom variable, X.

    prove this< exp(iqX) >= exp(iq< X>)exp[q

    2 / (2)] ii.5

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    16/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Gaussian Distributions

    16

    According to Ludwig Boltzmann (1844 1906) andJames Clerk Maxwell (1831-1879) theprobability distribution for a particle in a weakly interacting gas as is given by

    Here, the potential holds the particles in a box of volume , so that U is zero inside a boxof this volume and infinite outside of it. As usual, we go after thermodynamic properties bycalculating the partition function,

    (p,r) = (1/ z)exp(H)H= [px

    2

    + py2

    + pz2

    ] / 2m+U(r)

    z= [ dpexp(p2 / (2m ))]3 = (2m/ )3/2

    In the usual way, we find that the average energy is 3/(2) = (3/2)kT. The classical result is theaverage energy contains a term 1/2kT for each quadratic degree of freedom. Thus a harmonicoscillator has =kT.

    Hint for theorists: Calculations of Z (or of its quantum equivalent, the vacuum energy) areimportant. Once you can get this quantity , you are prepared to find out most other thingsabout the system.

    ii.6

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    17/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    The usual way

    17

    Z= d3p d3r exp(-[p2/(2m)+U(r)])=(2m/)3/2

    Let = [p2

    /(2m)+U(r)]

    ln Z/ =-(1/Z) d3p d3r exp(-)= -=3/(2)=(3/2)kT

    (The usual way)2

    2 ln Z /2= -/= ???

    Gaussian Averages

    Let one particle be confined to a box of volume . Let U(r) be zero inside the boxand + infinity outside. Then, in three dimensions

    Let N particles be confined to a box of volume . What is the RMS fluctuation in thepressure?

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    18/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Rapidly Varying Gaussian random variable

    18

    Later on we shall make use of a time-dependent gaussian random variable, (t). Inits usual use, (t) is a very rapidly varying quantity, with a time-integral whichbehaves like a Gaussian random variable. Specifically, it is defined to have twoproperties:

    < (t)>=0

    X(t) = du (u) is a Gaussian random variable with variance |s-t|.

    Here defines the strength of the oscillating random variable.

    st

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    19/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Approximate Gaussian Integrals

    19

    It is often necessary to calculate integrals like

    in the limit as M goes to infinity. Then the exponential varies over a wide range and theintegral appears very difficult. But, in the end its easy. The main contribution will come atthe maximum value of f in the interval [a,b]. Assume there is a unique maximum and thesecond derivative exists there. For definiteness say that the maximum occurs atx=0, witha

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    20/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    The usual way: in generalLets start from a Hamiltonian on a Lattice

    W{}=-H{} = hr

    r

    r + Krs

    This Hamiltonian defines what is called the Ising model. The first sumis a sum over all lattice sites, r. The second sum is a sum over

    nearest neighbors. The field, hr, which depends upon r, multiplies aspin variable which is different, of course, different on each site. Thenotation {} says that these things depend upon many spin variables.You can assume that the spin variables take on the value +1 or -1 ifyou want, but the argument is very general and the result does notdepend upon what might be. Start from Z =Tr exp W{}where Tr means a summation over all possible values of all the spinvariables. It is a repeat of the argument that we have given before to

    say that

    = ln Z/ hr

    The partial derivative means that we hold all other hs constant. The second derivative isgiven by

    - = ]> =2

    ln Z/h

    s h

    r

    =

    /h

    r

    Calculation of Averages and Fluctuations: Reprise

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    21/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    To see this we go slowly

    21

    / hr = Tr[s exp(W{})]/Tr exp(W{}) /hr

    = Tr[s r exp(W{})]/Z

    -Tr[s exp(W{})] Tr[r exp(W{})]/Z2

    =< s r > - < s > < r >

    This derivative can act in two places, one here and one here. In bothcases the derivative bring down a factor ofr . Since the denominator

    is called Z, we have that this expression is equal to

    which then simplifies to

    This expression is called the spin-spin correlation function. It expresseshow fully the fluctuations in the values of the spins in the two places arecorrelated. In general this kind of correlation falls off with distance, buthow it falls off tells us a lot about what is going on.

    =

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    22/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Sums and Averages in Classical Mechanics

    22

    The probability distribution for a single particle in a weakly interacting gas as is given by

    Here, the potential holds the particles in a box of volume , so that U is zero inside a boxof this volume and infinite outside of it.The partition function, is

    (p,r) = (1/ z)exp(H)

    H= [px2

    + py2

    + pz2

    ] / 2m+U(r)

    z= [ dpexp(p2 / (2m ))]3 = (2m/ )3/2

    The average of any function ofp and r is given by

    = dp dr(p,r) g(p,r)

    Since there are N particles in the system N dp dr(p,r) is the number of particles which haveposition and momentum within dp drabout the phase space pointp,r. The quantity

    N (p,r)=f(p,r) is called the distribution function. The totalamount of the quantity represented byg(p,r) is given in terms of the distribution function as

    total amount ofg = dp drf(p,r) g(p,r)

    Example: We calculated the average energy < p2/(2m) >=3 k T/2= dp dr(p,r)p2/(2m)The total energy in the system is dp drf(p,r)p2/(2m)= 3N k T/2.

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    23/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    More sums and averages

    23

    The normalization condition for the probability is dp dr(p,r) ==1

    The normalization for the distribution function is dp dr f(p,r) =N

    The pressure, P, is defined as the total momentum transfer to a wall per unit of area and unit oftime. Call these dA and dt. Since a low density gas is the same all over, the number hitting isthe number within the distance px/m dt of the area, for px >0, and hence the number within the

    volume px/m dt dA which is dp f(p,r) px/m dt dA with the integral covering all ps with thecondition that px >0. If a particle hits the wall and bounces back it transfers momentum 2px.

    Therefore the total momentum transferred is dp f(p,r) px/m dt dA 2px once again with the

    condition that px >0. An integral over all momenta would give a result twice as large. In theend we get that the pressure is

    P= dp f(p,r) px2/mwhich is then NkT as we knew it would be.

    The partition function is the sum over all variables of exp(-H). For large N, it can beinterpreted as W exp(-) , where W is the number of configuration which enter.Boltzmann got W in terms of the entropy as ln W=S/k. We put together previous results andfindW exp(-3N/2)= zN = N (2 m kT)3N/2 so that S/k= N [ln +3 ( ln (2 e m kT))/2 ]

  • 7/29/2019 Part 2 Basics of Statistical Physics1

    24/24

    Perimeter Institute Lecture Notes on Statistical Physics part 2: Basics of Statistical Physics Version 1.7 9/13/09 LeoKadanof

    Homework

    24

    Work out the value of the pressure for a classical relativistic gas with H=|p|c.

    Do this both by using kinetic theory, as just above, and also by differentiatingthe partition function.

    Statistical Mechanics started with Boltzmanns calculation off(p,r,t) for a gas of N particles in

    a non-equilibrium situation. He calculated a result for the entropy as a logarithm of the

    number of configurations, or in terms of an integral roughly of the form dp f(p,r) ln f(p,r).From our knowledge of equilibrium statistical mechanics how might we guess that this integralwas related to the entropy?

    The length of a random walk ofN steps is of the form of a sum ofN iid variables jeach of which takes on the values 0 and 1 with equal probability. What is the

    probability that a walker who takes 1000 steps will end up at a point at least 30 stepsaway from her starting point? (A relative accuracy of two decimal digits will suffice.)

    The Hamiltonian for N particles in a fluid is H=p2 /(2m) +