41
Basic numerical and statistical methods Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics: errors and goodness-of-fit 3. Data fitting by simple functions, solution of system of linear equations. 4. Data fitting by nonlinear models, methods for minimization. 5. Fourier transform, basic ideas.

Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

  • Upload
    others

  • View
    8

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Basic numerical and statistical methodsVladimir Volkov

Institute of Crystallography, Moscow, Russia

Plan of the talk:

1. Introduction: data intepretation

2. Statistics: errors and goodness-of-fit

3. Data fitting by simple functions, solution of system of linear equations.

4. Data fitting by nonlinear models, methods forminimization.

5. Fourier transform, basic ideas.

Page 2: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Main goal for end data interpretation

Y

X

rρρρρto build a mathematical description of a physical model,

calculate physical responses Y(x) from the model that could be measured in an experiment and

compare them with experimental data (what is goodness-of-fit?)

Then, correct the model parameters by using the comparison results (methods for automatic correction?)

Case study: structure refinement by molecular tectonics.

Page 3: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Case study: PDC in the crystal and in solution

3 nm

experiment crystal

solutionχ = 1.68χ = 0.94

Left: crystallographic model (interaction area, 13.44 nm2) Right: result of rigid body refinement (r.m.s., 0.58 nm)Svergun, D.I., Petoukhov, M.V., Koch, M.H.J. & König, S. (2000) J. Biol. Chem. 275, 297-303.

Page 4: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Good fit:Differences are less than above. In addition, they do not have systematic behaviour. This means, that the model may be adequate.

Merit function: analysis of residuals.

Goodness-of-fitPoor fit:Differences between experimental data and the model are large and systematic. The goal is to find model parameters that provide minimum difference.

0 1 2 3 4 5 6 7-0.2

0.0

0.2

0.4

0.6

X

Y Experiment Model Difference

0 1 2 3 4 5 6 7-0.1

0.0

0.1

0.2

0.3

0.4

0.5

0.6

X

Y Experiment Model, best fit Difference

Page 5: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Case of residuals: sequence of random numbers Y (errors)

0 200 400 600 800 1000-4-3-2-10123Sequence of normal random numbers

Value

N of value

-4

-3

-2

-1

0

1

2

3

0 500 1000 1500number of occurences

Middle box value

of Y

If sum over all occurences = 1 (normalization) then this plot relates to the probability density function: P{Y=y}.

y

Page 6: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Number of impulses counted by a detector: Poisson distribution.

0

20

40

60

80

0 1000 2000 3000 4000 5000 6000

Box value (number of occurences)

0 200 400 600 800 1000

0

10

20

30

40

50

60

70

80 Poisson random numbers with mean 50

N of value

0 200 400 600 800 10000

2

4

6

8

10

12

14

16 Poisson random numbers with mean 5Value

0

5

10

15

0 5000 10000 15000

Box value (number of occurences)

Middle box value

Asymmetric at low mean values

Almost symmetric, close to a normal distribution at large mean values

Page 7: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Statistical characterization of a data set

Most probable (expected) value of a random variable yi:

Mean value : , from

Mean value from density distribution histogram:

Measure of variancemean linear deviation:

variance (empirical):

standard deviation: σσσσ,

(measure of variability about mean value: )

�=

=N

iiy

Ny

1

1 yNyyN

i

N

ii ==��

== 11

��=j

bj

bj jj nnby

�=

−=N

jj yy

Nd

1

1

( )�=

−−

=N

jj yy

N 1

221

σ±y

-4

-3

-2

-1

0

1

2

3

0 500 1000 1500number of occurences

Middle box value

y

σσσσ

bj

jbn

Page 8: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

-4 -2 0 2 4 6 8 10 12 140.0

0.2

0.4

0.6

0.8

1.0

1.2

1.4

x

a

eh

h

σσσσ σσσσ

Poisson random variable (discrete), which is the number of successes that will be observed in repeated intervals of time, each of which may result in the occurenceor nonoccurence of a success.

where y is the fixed value, t is the observation time interval, Lt is the mean number of successes per interval of measurement t.

mean value (expected value) E{Y} = Lt,variance = Lt (thus, we measure values )The larger is t, the less is the relative error.

ty

ey

LtyYP −==!

}{

LtLtY ±=

Normal random variable (continuous).De Moivre-Laplace's theorem:

Most of statistical estimates uses this fact.Common formula:

mean value (expected value) E{Y} = a,variance = σ σ σ σ 2.

dxedLtLtYc d

c

x

N �−

∞→=��

���

� <−< 22

21limπ

22

2)(

21),( 2 σ

πσσ ⋅

−−=

ax

eaN

(probability that randomly obtained Y equals to the prescribed value y)

Page 9: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Measure of degree of adjacency of thwo data setsa1, a2, …, aN and b1, b2, …, bN

Our goal: to get a single characteristic number.We can not use summmation over all deviations because they may have opposite signs and compensate each other giving a rezult near zero.But we may use the sum over absolute values of deviations:

It works, but the use of this formula leads to significant mathematical difficulties: this function is not differentiable.Squaring the deviations proves to be a much more attractive solution:

�=

−=N

jjj ba

Nd

1

1

( )�=

−−

=N

jjj ba

Nd

1

221

1

Page 10: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Weighting of the summation terms:equalization of influence of errors

An example: level of Poisson noise is proportional to square root of the intensity. To reduce the influence of data points with large associated errors, all the elements are divided by their corresponding deviates. These deviates must be determined independently.

A: Curve to be fitted to

B’’: curve minimizing

E: Errors in Arelatively small large

B’: curve minimizing ( )�

=−

−=

N

jjj BA

Nd

1

221

1

�=

��

��

� −−

=N

j j

jj BAN

d1

22

11

σ

AB’

B’’E

0

I

Page 11: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Tangential gradient g

gmin=0

Chi-square: least squares fitting of a set of N data points Aj with errors σσσσj.

Main idea of a fitting procedure: At the point of the minimum, the derivatives (gradients)

because we have here zero slope of the surface of the minimizing function.

We use here the term “chi-square”, because sum of squares of random deviates have chi square distribution. Expected minimum value of chi-square is about unity.

02

=∂∂

kpχ

pk

��

��

��

� −−−

= �=

N

j j

jj BAMN 1

22

11min

σχ

Calculated from a model described by a set of M

parameters pk.

Page 12: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

σ±y

We consider the problem of fitting a set of N data points (xi; yi) to a straight-line model withparameters a and b :

y(x) = y(x; a; b) = a+bx

This problem is often called linear regression. We assume that the uncertaintyσσσσιιιι associated with each measurement yi is known, and that the xi’s (values ofthe dependent variable) are known exactly.

To measure how well the model agrees with the data, we use the chi-square merit function, which in this case is

At its minimum, derivatives with respect to a; b vanish.

�=

−−−=

∂∂=

N

j j

jij bxayxb 1

2

2 )(20

σχ

�=

−−−=

∂∂=

N

j j

ji bxaya 1

2

220

σχ

Fitting data by a straight line.

�=

��

��

� −−−

=N

j j

jj bxayN 1

22

11

σχ

x

Page 13: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

xyxxx

yxSbSaSSbSaS

=+=+

�= 2j

jjxy

yxS

σ�= 2

2

j

jxx

xS

σ

These conditions can be rewritten in a convenient form if we define the following sums:

With these definitions, we have here the system of linear equations:

This system may be solved in any common way giving

Fitting data by a straight line.

�= 21

jS

σ�= 2

j

jx

xS

σ�= 2

j

jy

yS

σ

Now, we must estimate the probable uncertainties in the estimates of a and b, since obviously the measurement errors in the data must introduce some uncertainty in the determination of those parameters. If the data are independent, then each contributes its own bit of uncertainty to the parameters.

2)( xxx SSS −=∆

∆−

= xyxyxx SSSSa

∆−

= yxxy SSSSb

Sourceof possible

computationalunstability

Page 14: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Consideration of propagation of errors shows that the variance in the value of any function will be

For the straight line,

Thus, we have the solution for coefficients of a straight line: Because we minimize sum of squares, it is possible to obtain once more Important estimate - the goodness-of-fit of the data to the model. This estimate indicates whether the parameters a and b in the model have any meaning at all! The probability Q that a value of chi-square as poor as the value we use to derive the solution,

should occur by chance is

where is incomplete gamma function. Using known tables of

statistics one may find that if Q > 0.1 then the solution is reliable, if 0.001<Q<0.1 then a and b may be acceptable. If Q < 0.001 then the model and/or estimation procedure can rightly be called into question.

Fitting data by a straight line. Propagation of errors in solution.

2

1

22��

��

∂∂= �

= j

N

jji y

fσσ

∆= xx

aS2σ

∆= S

b2σ

ba ba σσ ±± ,

�=

��

��

� −−−

=N

j j

jj bxayN 1

22

11

σχ �

��

� −Γ=2

,2

2 2χNQ inc

( )qpQ inc ,Γ=

i: index of a model parameter,

j: index of an experimental point

Page 15: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Fitting data by polynomialsm

mxaxaxaaxF ++++= ...)( 2210

�=

��

��

� −−−

=N

j j

jj xFymN 1

22 )(

11

σχ

mja j

,...,0,02

==∂∂χ

Polynom :

What we minimize:

Condition of minimum:

Systemof mlinear equations:

�=

N

j ja

1201

σ jN

j jxa �

=1211

σmj

N

j jm xa �

=121

σ jN

j jy�

=121

σ

jN

j jxa �

=1201

σ jN

j jxa �

=1211

σ1

121 +

=�

mj

N

j jm xa

σjj

N

j jyx�

=121

σ

mj

N

j jm xa �

=121

σ1

1211 +

=�

mj

N

j jxa

σm

jN

j jm xa 2

121

�= σ j

mj

N

j jyx�

=121

σ

+…+ +

+…+ +

+…+ + =

=

=

…….. …….. …….. ……..

Page 16: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Systems of linear equations - matrix notation

=C BAm

mMatrix notation: C A = B,

solution: A = C-1B

�=

N

j ja

1201

σj

N

j jxa �

=1211

σmj

N

j jm xa �

=121

σj

N

j jy�

=121

σ

jN

j jxa �

=1201

σ jN

j jxa �

=1211

σ1

121 +

=�

mj

N

j jm xa

σjj

N

j jyx�

=121

σ

mj

N

j jm xa �

=121

σ1

1211 +

=�

mj

N

j jxa

σm

jN

j jm xa 2

121

�= σ j

mj

N

j jyx�

=121

σ

+…+ +

+…+ +

+…+ + =

=

=

…….. …….. …….. ……..

Errors in solution: 122 −= jja Cj

χσ

Page 17: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Solution of systems of linear equations

=b1a1

m

mc11

cm1 cmm

c1m

am bm

Method of Gauss-Jordan elimination (triangulation)

1) C11a1 + C12a2 + C13a3 = b1

2) C21a1 + C22a2 + C23a3 = b2

3) C31a1 + C32a2 + C33a3 = b3multiply 1) by C21/C11, subtract from 2);multiply 1) by C31/C11, subtract from 3);

1) C11a1 + C12a2 + C13a3 = b1

2) C*22a2 + C*

23a3 = b*2

3) C*32a2 + C*

33a3 = b*3

1) C11a1 + C12a2 + C13a3 = b1

2) C*22a2 + C*

23a3 = b*2

3) C**33a3 = b**

3

multiply 2) by C*32 /C*22, subtract from 3);

Direct move:

Backward substitution:a3 = b**

3 / C**33 from 3);

a2 = (b*2 – C*

23a3) / C*22 from 2);

a1 = (b1 - C13a3 - C12a2) / C11 from 1)

“Pivoting”:Avoiding deletion by small coefficients by rearrangement of rows.

Page 18: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Solution of systems of linear equationsOverestimated systems: least squares solution.

This system does not have an exact solution.

Case study: fitting data by polynomials(N equations):

imimii yxaxaxaa =++++ ...2

210

=

D

BA

M= m+1

N

1 x1 x12 …x1

m

1 x2 x22 …x2

m

. . . . . . . . .

1 xn xn2 …xn

m

If we look at the matrix C, we find that

C = DTD (matrix of normal equations).

=DT

DC

Solution by least squares:C A = DT B,A=(C)-1 DT B

yi

11

Page 19: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Solution of systems of linear equations

Singular value decomposition (SVD)

.=.D VTU S

VT V

left

right

UUT =

=

1 01

10 1

1 01

10 1

Identity matrix (E)

Orthonormalmatrices

Diagonal matrix of singular values

U-1 = UT;V-1 = VT

Page 20: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Solution of systems of linear equations

Solution by least squares using normal equations:

Solution by least squares using SVD:D = U S VT;UT U S VT A = UT B; because UTU = E, and E play

role of a unity in multiplications:S VT A = UT B; S-1S = E, then

VT A = S-1 UT B; critical operation - inversion of siiA = V S-1 UT B because V VT = E

DTD A = DT B,A=(DTD)-1 DT B

)()1(1

)(1

)( 11m

mi

M

i i

iSSS

VVVBU

A ±±±���

����

� ⋅=�

=�

2

1

2�=

���

����

�=

M

i i

jia Sj

Errors

Page 21: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Minimization of functionsYou are given a single function f(x) that depends on one or more independent variables.

You want to find the value of those variables {x} where f(x) takes on a minimum value. You can then calculate what value of f(x) is achieved at the minimum.An extremum (minimum point) can be either global (truly the lowest function value) or local (the lowest in a finite neighborhood and not on the boundary of that neighborhood).

f(x)

x1

x2

Page 22: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Minimization of functions

One-dimensional search:Golden section approach

Successive bracketing of a minimum. The minimum is originally bracketed by points 1,3,2. The function is evaluated at 4, which replaces 2; then at 5, which replaces 1; then at 6, which replaces 4. The rule at each stage is to keep a centerpoint that is lower than the two outside points. After the steps shown, the minimum is bracketed by points 5,3,6.

Golden proprtion:

b / (a+b) = a / b ~ 0.618…

The main idea: at each step, calculate only one function value with reducing the interval of uncertainty.

2

1

3

4

5 6

a bc d

Page 23: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Minimization of functions

One-dimensional search:

Golden section with parabolic interpolation (Brent's Method)It is often much more effectivet than the successive golden section.

At any particular stage, it is keeping track of 5 function points (not necessarily all distinct), a, b, u, v, w and x, defined as follows:the minimum is bracketed between a and b;x is the point with the very least function value found so far (or the most recent one in case of a tie); w is the point with the second least function value;v is the previous value of w; u is the point at which the function was evaluated most recently.

a b

2

1

453

At the each particular step we first try to find the minimum of a

parabola passing throug the three best points and use golden section

if the minimum point does not reduce the interval better.

Page 24: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Minimization of functions

Multidimensional search.1. Direct methods (only function values are used): random search, downhill simplex method due to Nelder and Mead2. Gradient methods: conjugate gradients, variable metrix3. Nonlinear Least Squares: algorithm NL2SOL.3. Global minimization

Page 25: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Downhill simplex method due to Nelder and MeadPossible outcomes for a step in the downhill simplex method in a three-dimensional case f(x1,x2,x3): the simplex at the beginning of the step, here a tetrahedron, is shown, top. The simplex at the end of the step can be any one of (a) a reflection away from the high point, (b) a reflection and expansion away from the high point, (c) a contraction along one dimension from the high point, or (d) a contraction along all dimensions towards the low point. An appropriate sequence of such steps will always converge to a minimum of the function.

Sample page from NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press 1992.

Minimization of functions

Page 26: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Sample page from NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press 1992.

Minimization of functionsSteepest descent method

(most reasonable but with worst behaviour!):

Successive minimizations along coordinate directions in a long, narrow “valley” (shown as contour lines). Unless the valley is optimally oriented, this method is extremely inefficient, taking many tiny steps to get to the minimum, crossing and re-crossing the principal axis.

Gradient vector

Descent vector

Next gradient

Contour lines, view from the top to F(x)

Page 27: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Minimization of functionsConjugate gradient approach

xAxxgPx ⋅⋅+⋅−≈+∂∂

∂+∂∂+= �� 2

1...21)()(

,

2cxx

xxfx

xfff ji

ji jjii

i

-g :antigradient

vector

Function value at the current

point P

Taylor series approximation of any function:

The matrix A whose components are the second partial derivative matrix of the function is called theHessian matrix of the function at P.

Conjugate conditions for d (vectors of search steps)

jiwith ≠=⋅⋅ 0ji dAd

Quadratcapproximation

���

+==

1-kkkk

00dg-d

g-dβRule for generating directions

{ }��

���

+=+=

+

++

kk

kk1kdxdxx

k1k

1kαα

α

αminLocal search along dk

)g(gd)g(gg

AddAdg

1-kk1-k

1-kkk

1-k1-k

1-kk−

−==kβRule for conjugate gradient multiplier

Page 28: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Minimization of functionsConjuage gradient approach

Gradient vector g0 Descent

vector d0

Next gradient g1

01dβ

Antigradient direction(steepest descent) -g1

Conjugate directiond1 = -g1 + ββββ1d0

Page 29: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Minimization of functionsNewton method: A is known

01-

0min gAxx −=

xAxxgx ⋅⋅+⋅−=21)( cf

01- gA−

Gradient vector g0

No linear search is required for

quadratic functions.

Page 30: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

initial step

Newton step

Accumulation of A

step to min along dk

difference gradient

search path

Minimization of functionsQuasi-Newton methods

01-

0min gAxx −=

xAxxgx ⋅⋅+⋅−=21)( cf

�=

×=M

1k kkkk1-

ypppA

=x

kkk1-kkk1-kk

k1-k

kk

kkk ωωpAp

pAppA

ppyyU ×⋅⋅+−×= kϕ

00 gd −=

kkk gAd −=

1-k1-kk UAA +=

kk1k dxx kα+=+

1-kkk ggy −=

1k1-kkk dxxp −−=−= 1kα

Common formula for quasi-Newton corrections:

Page 31: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Search paths for different minimization methodsNelder-Mead

(each 5-th step)

100 iterations/260 function evaluations

Steepest descent (no success)

Conjugate directions

30/217

Quasi-Newton20/140

Page 32: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Algorithm NL2SOL for nonlinear least squares problems

Consider vector function r(x) which values are elements of discepancy between two data sets: yj and F(xj), j=1,…,M. Here xi are the model parameters, i=1,…,N.

j

jji

sFyxr

σ)(

)(−

=

Chi-square is computed as rr ⋅=21)(xf

j

ixr

∂∂=JJacobian matrix

rJrr T ⋅=��

��

∂∂=��

��

⋅∂∂=

∂∂= �

=

M

iirxxx

xfg1

221

21)(

SJJT +=��

��

∂∂+

��

��

��

���

∂∂

��

��

��

���

∂∂=

∂�=

2

2

12

2 )(

j

ii

T

j

i

j

iM

ij

j

xrr

xr

xr

x

xf

Gradient at the point x

Hessian matrix for f(x)

NL2SOL uses Quasi-Newton scheme where S is accumulated by quasi-Newton formulae and J is computed directly at each step. The approximation to A in this case is much better. This is the most powerful known algorithm for least-squares problems (TENSOLVE is better?).

Page 33: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Global minimization: Simulated annealing

Main idea of the simulated annealing: let the value of f(x) is the objective function. The system state is the point x. The control parameter T is something like a temperature, with an annealing schedule by which it is gradually reduced. And there must be a generator of random changes in the parameter values, a procedure for taking a random step from xi to xi+1 . Offered a succession of options, a simulating annealing method is assumed to jump to the new point with probability P=exp[(fi-fi+1)/kT]. Notice that if fi>fi+1,this probability is greater than unity; in such cases the change is arbitrarily assigned a probability P=1, i.e., the system always took such an option. This general scheme, of always taking a downhill step while sometimes taking an uphill step, has come to be known as the Metropolisalgorithm. In other words, the system sometimes goes uphill as well asdownhill; but the lower the temperature, the less likely is any significant uphill excursion. The schedule of the algorithm is in succesive decreasing the annealing temperature during the search. Finally, it takes only successful steps.Simulated annealing tries “random” steps; but in a long, narrow valley, almost all random steps are uphill and this algorithm is able to accept them thus continuing the minimization.

Page 34: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Simulated annealing approachbased on the downhill simplexmethod due to Nelder and MeadThe implementation of the Metropolis simulated annealing is here as follows: we ADD a [logarithmically] distributed random number, proportional to the temperature T, to the STORED vertex function value each time we look at it, and we SUBTRACT a random number from the function value calculated for every NEW point that is tried as a replacement candidate. This addition forces the simplex to randomly fluctuate keeping its size proportional to the current temperature. The subtraction forces the procedure always to accept a true downhill step, whereas the addition randomly leads to accepting an uphill one. Sometimes, it may help the program to fling out from a local minimum. The annealing itself is in decreasing the temperature during the search, T->0. At T=0 the method is exactly the downhill simplex scheme of Nelder and Mead.

Sample page from NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press 1992.

Minimization of functions:global search

Page 35: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Fourier transformA physical process can be described either in the time domain, by the values of some quantity h as a function of time t, e.g., h(t), or else in the frequency domain,where the process is specified by giving its amplitude H (generally a complex number indicating phase also) as a function of frequency f, that is H(f), with -1 < f <1. For many purposes it is useful to think of h(t) and H(f) as being two different representations of the same function. One goes back and forth between these two representations by means of the Fourier transform equations,

�∞

∞−= dtethfH iftπ2)()(

�∞

∞−= dfefHth iftπ2)()(

fπω 2=

�∞

∞−= dtethH tiωω )()(

�∞

∞−= ωω

πω deHth ti)(

21)(

If we use angular frequency, in radians:

This is a linear operation

Page 36: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

�∞

∞−= dtethH tiωω )()(

h(t) may be considered as coefficients of orthogonal functions which sum over all values of t reproduces the function Therefore, we could find these coefficient by a system of linear equations writing the equation coefficients

for each value of t:

tie ω

)(ωH

kjtie ω

)()( jti

kk

Heth kj ωω =�∞

−∞=This is the basis for different indirect methods of Fourier calculations. The Fourier transform may be represented as decomposition of a function to a sum of other orthogonal functions, such as sin, cos, etc.

h(t1) x

= + h(t2) x

+ h(t3) x

+ …

Page 37: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Case study: scattering from monodisperse systems

Scattering intensity from a dilute solution of chaotically oriented identical particles is proportional to the spherically averaged scattering from a single particle

p(r) is distance distribution function, D is the maximum particle size.

drsr

srrpsID

�=0

sin)(4)( π

Page 38: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

Case study: computation of the density profiles

♦ The electron density profile of a centrosymmetric bilayer is computed using one-dimensional Fourier transformation as

where N is the number of peaks and sl and S(l) are the position and the area of the l-th peak, respectively. The Lorentz factor sl reflects random orientation of the lamella microcrystals in PODS. The most plausible combination of signs of the amplitudes A(s,l) is selected by a visual inspection of the density profiles.

)cos(])([)cos(),()(11

srlSssrlsAsrN

ll

N

ll ππρ ±== ��

==

Page 39: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

0 2 4 61

2

3

l=2 l=3 l=5 l=6

l=1

lg I, relative

s, nm-1

Internal structure of PODS

Peak fitting for the initial PODS sample

♦ PODS samples display up to five Bragg peaks at sl=2πl/d, where d=5.24±0.03 nm is the bilayer thickness.

♦ The long-range order dimension L=37±3 nmprovides the average size of the POSD lamella crystallite (about seven bilayers)

Page 40: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

r, nm

0 1 2 3 4 5

ρ, relative

-6

-4

-2

0

2

4

6

8acetone, self-reductionethanol, self-reductionwater/CPC, NaBH4water/CPC, UV-irradiationPODS+CPCInitial PODS

Density profiles of PODS bilayers♦ Compound: AuCl3

♦ Small particles grow in the central (hydrophilic) part of the bilayer

Page 41: Vladimir Volkov - EMBL Hamburg · 2009. 7. 24. · Vladimir Volkov Institute of Crystallography, Moscow, Russia Plan of the talk: 1. Introduction: data intepretation 2. Statistics:

LiteratureR.J.Larsen, M.L.Marx. An Introduction to Mathematical Statistics and its Applications. Prentice-Hall, Inc. Englewood Cliffs, New Jersey 07632, 1981.Gilbert Streng. Linear Algebra and its Applications.Massachusetts Institute of Technology, Acad.Press, New York-London, 1976.P.E.Gill, W.Murray, M.H.Wright. Practical Optimization. SystemsOptimization Laboratory, Dept. of Operations Research, Stanford University.Acad.Press, London-San Francisco, 1981.NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press, 1992 (now available via Internet at www.nrc.com).