51
Advances in Convex Optimization: Theory, Algorithms, and Applications Stephen Boyd Electrical Engineering Department Stanford University (joint work with Lieven Vandenberghe, UCLA) ISIT 02 ISIT 02 Lausanne 7/3/02

Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

  • Upload
    others

  • View
    16

  • Download
    1

Embed Size (px)

Citation preview

Page 1: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Advances in Convex Optimization:Theory, Algorithms, and Applications

Stephen Boyd

Electrical Engineering DepartmentStanford University

(joint work with Lieven Vandenberghe, UCLA)

ISIT 02

ISIT 02 Lausanne 7/3/02

Page 2: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Two problems

polytope P described by linear inequalities, aTi x ≤ bi, i = 1, . . . , L

a1

P

Problem 1a: find minimum volume ellipsoid ⊇ P

Problem 1b: find maximum volume ellipsoid ⊆ P

are these (computationally) difficult? or easy?

ISIT 02 Lausanne 7/3/02 1

Page 3: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

problem 1a is very difficult

• in practice

• in theory (NP-hard)

problem 1b is very easy

• in practice (readily solved on small computer)

• in theory (polynomial complexity)

ISIT 02 Lausanne 7/3/02 2

Page 4: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Two more problems

find capacity of discrete memoryless channel, subject to constraints oninput distribution

Problem 2a: find channel capacity, subject to:no more than 30% of the probability is concentrated on any 10% of theinput symbols

Problem 2b: find channel capacity, subject to:at least 30% of the probability is concentrated on 10% of the input symbols

are problems 2a and 2b (computationally) difficult? or easy?

ISIT 02 Lausanne 7/3/02 3

Page 5: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

problem 2a is very easy in practice & theory

problem 2b is very difficult1

1I’m almost sure

ISIT 02 Lausanne 7/3/02 4

Page 6: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Moral

very difficult and very easy problems can look quite similar

. . . unless you’re trained to recognize the difference

ISIT 02 Lausanne 7/3/02 5

Page 7: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Outline

• what’s new in convex optimization

• some new standard problem classes

• generalized inequalities and semidefinite programming

• interior-point algorithms and complexity analysis

ISIT 02 Lausanne 7/3/02 6

Page 8: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Convex optimization problems

minimize f0(x)

subject to f1(x) ≤ 0, . . . , fL(x) ≤ 0, Ax = b

• x ∈ Rn is optimization variable

• fi : Rn → R are convex, i.e., for all x, y, 0 ≤ λ ≤ 1,

fi(λx + (1 − λ)y) ≤ λfi(x) + (1 − λ)fi(y)

examples:

• linear & (convex) quadratic programs

• problem 1b & 2a (if formulated properly)

ISIT 02 Lausanne 7/3/02 7

Page 9: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Convex analysis & optimization

nice properties of convex optimization problems known since 1960s

• local solutions are global

• duality theory, optimality conditions

• simple solution methods like alternating projections

convex analysis well developed by 1970s (Rockafellar)

• separating & supporting hyperplanes

• subgradient calculus

ISIT 02 Lausanne 7/3/02 8

Page 10: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

What’s new (since 1990 or so)

• powerful primal-dual interior-point methodsextremely efficient, handle nonlinear large scale problems

• polynomial-time complexity results for interior-point methodsbased on self-concordance analysis of Newton’s method

• extension to generalized inequalitiessemidefinite & maxdet programming

• new standard problem classesgeneralizations of LP, with theory, algorithms, software

• lots of applicationscontrol, combinatorial optimization, signal processing,circuit design, . . .

ISIT 02 Lausanne 7/3/02 9

Page 11: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Recent history

• (1984–97) interior-point methods for LP

– (1984) Karmarkar’s interior-point LP method– theory (Ye, Renegar, Kojima, Todd, Monteiro, Roos, . . . )– practice (Wright, Mehrotra, Vanderbei, Shanno, Lustig, . . . )

• (1988) Nesterov & Nemirovsky’s self-concordance analysis

• (1989–) semidefinite programming in control(Boyd, El Ghaoui, Balakrishnan, Feron, Scherer, . . . )

• (1990–) semidefinite programming in combinatorial optimization(Alizadeh, Goemans, Williamson, Lovasz & Schrijver, Parrilo, . . . )

• (1994) interior-point methods for nonlinear convex problems(Nesterov & Nemirovsky, Overton, Todd, Ye, Sturm, . . . )

• (1997–) robust optimization (Ben Tal, Nemirovsky, El Ghaoui, . . . )

ISIT 02 Lausanne 7/3/02 10

Page 12: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Some new standard (convex) problem classes

• second-order cone programming (SOCP)

• semidefinite programming (SDP), maxdet programming

• (convex form) geometric programming (GP)

for these new problem classes we have

• complete duality theory, similar to LP

• good algorithms, and robust, reliable software

• wide variety of new applications

ISIT 02 Lausanne 7/3/02 11

Page 13: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Second-order cone programming

second-order cone program (SOCP) has form

minimize cT0 x

subject to ‖Aix + bi‖2 ≤ cTi x + di, i = 1, . . . , m

Fx = g

• variable is x ∈ Rn

• includes LP as special case (Ai = 0, bi = 0), QP (ci = 0)

• nondifferentiable when Aix + bi = 0

• new IP methods can solve (almost) as fast as LPs

ISIT 02 Lausanne 7/3/02 12

Page 14: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Robust linear programming

robust linear program:

minimize cTx

subject to aTi x ≤ bi for all ai ∈ Ei

• ellipsoid Ei = { ai + Fip | ‖p‖2 ≤ 1 } describes uncertainty inconstraint vectors ai

• x must satisfy constraints for all possible values of ai

• can extend to uncertain c & bi, correlated uncertainties . . .

ISIT 02 Lausanne 7/3/02 13

Page 15: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Robust LP as SOCP

robust LP is

minimize cTx

subject to aTi x + sup{(Fip)Tx | ‖p‖2 ≤ 1} ≤ bi

which is the same as

minimize cTx

subject to aTi x + ‖FT

i x‖2 ≤ bi

• an SOCP (hence, readily solved)

• term ‖FTi x‖2 is extra margin required to accommodate uncertainty in ai

ISIT 02 Lausanne 7/3/02 14

Page 16: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Stochastic robust linear programming

minimize cTx

subject to Prob(aTi x ≤ bi) ≥ η, i = 1, . . . ,m

where ai ∼ N (ai,Σi), η ≥ 1/2 (c and bi are fixed)i.e., each constraint must hold with probability at least η

equivalent to SOCP

minimize cTx

subject to aTi x + Φ−1(η)‖Σ1/2

i x‖2 ≤ 1, i = 1, . . . ,m

where Φ is CDF of N (0, 1) random variable

ISIT 02 Lausanne 7/3/02 15

Page 17: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Geometric programming

log-sum-exp function:

lse(x) = log (ex1 + · · · + exn)

. . . a smooth convex approximation of the max function

geometric program (GP), with variable x ∈ Rn:

minimize lse(A0x + b0)subject to lse(Aix + bi) ≤ 0, i = 1, . . . ,m

where Ai ∈ Rmi×n, bi ∈ Rmi

new IP methods can solve large scale GPs (almost) as fast as LPs

ISIT 02 Lausanne 7/3/02 16

Page 18: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Dual geometric program

dual of geometric program is an unnormalized entropy problem

maximize∑m

i=0

(bTi νi + entr(νi)

)subject to νi � 0, i = 0, . . . ,m, 1Tν0 = 1,∑m

i=0 ATi νi = 0

• dual variables are νi ∈ Rmi, i = 0, . . . , m

• (unnormalized) entropy is

entr(ν) = −n∑

i=1

νi logνi

1Tν

• GP is closely related to problems involving entropy, KL divergence

ISIT 02 Lausanne 7/3/02 17

Page 19: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Example: DMC capacity problem

x ∈ Rn is distribution of input; y ∈ Rm is distribution of outputP ∈ Rm×n gives conditional probabilities: y = Px

primal channel capacity problem:

maximize −cTx + entr(y)

subject to x � 0, 1Tx = 1, y = Px

where cj = −∑mi=1 pij log pij

dual channel capacity problem is a simple GP:

minimize lse(u)

subject to c + P Tu � 0

ISIT 02 Lausanne 7/3/02 18

Page 20: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Generalized inequalities

with proper convex cone K ⊆ Rk we associate generalized inequality

x �K y ⇐⇒ y − x ∈ K

convex optimization problem with generalized inequalities:

minimize f0(x)

subject to f1(x) �K1 0, . . . , fL(x) �KL0, Ax = b

fi : Rn → Rki are Ki-convex: for all x, y, 0 ≤ λ ≤ 1,

fi(λx + (1 − λ)y) �Kiλfi(x) + (1 − λ)fi(y)

ISIT 02 Lausanne 7/3/02 19

Page 21: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Semidefinite program

semidefinite program (SDP):

minimize cTx

subject to A0 + x1A1 + · · · + xnAn � 0, Cx = d

• Ai = ATi ∈ Rm×m

• inequality is matrix inequality, i.e., K is positive semidefinite cone

• single constraint, which is affine (hence, matrix convex)

ISIT 02 Lausanne 7/3/02 20

Page 22: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Maxdet problem

extension of SDP: maxdet problem

minimize cTx − log det+(G0 + x1G1 + · · · + xmGm)subject to A0 + x1A1 + · · · + xnAn � 0, Cx = d

• x ∈ Rn is variable

• Ai = ATi ∈ Rm×m, Gi = GT

i ∈ Rp×p

• det+(Z) ={

detZ if Z � 00 otherwise

ISIT 02 Lausanne 7/3/02 21

Page 23: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Semidefinite & maxdet programming

• nearly complete duality theory, similar to LP

• interior-point algorithms that are efficient in theory & practice

• applications in many areas:

– control theory– combinatorial optimization & graph theory– structural optimization– statistics– signal processing– circuit design– geometrical problems– algebraic geometry

ISIT 02 Lausanne 7/3/02 22

Page 24: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Chebyshev bounds

generalized Chebyshev inequalities: lower bounds on

Prob(X ∈ C)

• X ∈ Rn is a random variable with EX = a, EXXT = S

• C is an open polyhedron C = {x | aTi x < bi, i = 1, . . . ,m}

cf. classical Chebyshev inequality on R

Prob(X < 1) ≥ 11 + σ2

if EX = 0, EX2 = σ2

ISIT 02 Lausanne 7/3/02 23

Page 25: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Chebyshev bounds via SDP

minimize 1 − ∑mi=1 λi

subject to aTi zi ≥ biλi, i = 1, . . . ,m

∑mi=1

[Zi zi

zTi λi

]�

[S aaT 1

][

Zi zi

zTi λi

]� 0, i = 1, . . . , m

• an SDP with variables Zi = ZTi ∈ Rn×n, zi ∈ Rn, and λi ∈ R

• optimal value is a (sharp) lower bound on Prob(X ∈ C)

• can construct a distribution with EX = a, EXXT = S that attainsthe lower bound

ISIT 02 Lausanne 7/3/02 24

Page 26: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Detection example

x = s + v

• x ∈ Rn: received signal

• s: transmitted signal s ∈ {s1, s2, . . . , sN} (one of N possible symbols)

• v: noise with E v = 0, E vvT = I (but otherwise unknown distribution)

detection problem: given observed value of x, estimate s

ISIT 02 Lausanne 7/3/02 25

Page 27: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

example (n = 2, N = 7)

s1

s2

s3

s4

s5s6

s7

• detector selects symbol sk closest to received signal x

• correct detection if sk + v lies in the Voronoi region around sk

ISIT 02 Lausanne 7/3/02 26

Page 28: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

example: bound on probability of correct detection of s1 is 0.205

s1

s2

s3

s4

s5s6

s7

solid circles: distribution with probability of correct detection 0.205

ISIT 02 Lausanne 7/3/02 27

Page 29: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Boolean least-squares

x ∈ {−1, 1}n is transmitted; we receive y = Ax + v, v ∼ N (0, I)

ML estimate of x found by solving boolean least-squares problem

minimize ‖Ax − y‖2

subject to x2i = 1, i = 1, . . . , n

• could check all 2n possible values of x . . .

• an NP-hard problem

• many heuristics for approximate solution

ISIT 02 Lausanne 7/3/02 28

Page 30: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Boolean least-squares as matrix problem

‖Ax − y‖2 = xTATAx − 2yTATx + yTy

= TrATAX − 2yTATx + yTy

where X = xxT

hence can express BLS as

minimize TrATAX − 2yTATx + yTysubject to Xii = 1, X � xxT , rank(X) = 1

. . . still a very hard problem

ISIT 02 Lausanne 7/3/02 29

Page 31: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

SDP relaxation for BLS

ignore rank one constraint, and use

X � xxT ⇐⇒[

X xxT 1

]� 0

to obtain SDP relaxation (with variables X , x)

minimize TrATAX − 2yTATx + yTy

subject to Xii = 1,[

X xxT 1

]� 0

• optimal value of SDP gives lower bound for BLS

• if optimal matrix is rank one, we’re done

ISIT 02 Lausanne 7/3/02 30

Page 32: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Stochastic interpretation and heuristic

• suppose X , x are optimal for SDP relaxation

• generate z from normal distribution N (x, X − xxT )

• take x = sgn(z) as approximate solution of BLS(can repeat many times and take best one)

ISIT 02 Lausanne 7/3/02 31

Page 33: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Interior-point methods

• handle linear and nonlinear convex problems (Nesterov & Nemirovsky)

• based on Newton’s method applied to ‘barrier’ functions that trap x ininterior of feasible region (hence the name IP)

• worst-case complexity theory: # Newton steps ∼ √problem size

• in practice: # Newton steps between 5 & 50 (!)

• 1000s variables, 10000s constraints feasible on PC; far larger if structureis exploited

ISIT 02 Lausanne 7/3/02 32

Page 34: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Log barrier

for convex problem

minimize f0(x)subject to fi(x) ≤ 0, i = 1, . . . , m

we define logarithmic barrier as

φ(x) = −m∑

i=1

log(−fi(x))

• φ is convex, smooth on interior of feasible set

• φ → ∞ as x approaches boundary of feasible set

ISIT 02 Lausanne 7/3/02 33

Page 35: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Central path

central path is curve

x?(t) = argminx

(tf0(x) + φ(x)) , t ≥ 0

• x?(t) is strictly feasible, i.e., fi(x) < 0

• x?(t) can be computed by, e.g., Newton’s method

• intuition suggests x?(t) converges to optimal as t → ∞• using duality can prove x?(t) is m/t-suboptimal

ISIT 02 Lausanne 7/3/02 34

Page 36: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Example: central path for LP

x?(t) = argminx

(tcTx − ∑6

i=1 log(bi − aTi x)

)

c

x?(0)

xopt x?(10)

ISIT 02 Lausanne 7/3/02 35

Page 37: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Barrier method

a.k.a. path-following method

given strictly feasible x, t > 0, µ > 1

repeat

1. compute x := x?(t)

(using Newton’s method, starting from x)

2. exit if m/t < tol

3. t := µt

duality gap reduced by µ each outer iteration

ISIT 02 Lausanne 7/3/02 36

Page 38: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Trade-off in choice of µ

large µ means

• fast duality gap reduction (fewer outer iterations), but

• many Newton steps to compute x?(t+)(more Newton steps per outer iteration)

total effort measured by total number of Newton steps

ISIT 02 Lausanne 7/3/02 37

Page 39: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Typical example

GP with n = 50 variables,m = 100 constraints, mi = 5

• wide range of µ works well

• very typical behavior(even for large m, n)

0 20 40 60 80 100 120

10−6

10−4

10−2

100

102

Newton iterations

dual

ity

gap

µ = 2µ = 50µ = 150

ISIT 02 Lausanne 7/3/02 38

Page 40: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Effect of µ

0 20 40 60 80 100 120 140 160 180 2000

20

40

60

80

100

120

140

µ

New

ton

iter

atio

ns

barrier method works well for µ in large range

ISIT 02 Lausanne 7/3/02 39

Page 41: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Typical effort versus problem dimensions

• LPs with n = 2m vbles, mconstraints

• 100 instances for each of20 problem sizes

• avg & std dev shown

101

102

103

15

20

25

30

35

m

New

ton

iter

atio

ns

ISIT 02 Lausanne 7/3/02 40

Page 42: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Other interior-point methods

more sophisticated IP algorithms

• primal-dual, incomplete centering, infeasible start

• use same ideas, e.g., central path, log barrier

• readily available (commercial and noncommercial packages)

typical performance: 10 – 50 Newton steps (!)— over wide range of problem dimensions, problem type, and problem data

ISIT 02 Lausanne 7/3/02 41

Page 43: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Complexity analysis of Newton’s method

• classical result: if |f ′′′| small, Newton’s method converges fast

• classical analysis is local, and coordinate dependent

• need analysis that is global, and, like Newton’s method, coordinateinvariant

ISIT 02 Lausanne 7/3/02 42

Page 44: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Self-concordance

self-concordant function f (Nesterov & Nemirovsky, 1988): whenrestricted to any line,

|f ′′′(t)| ≤ 2f ′′(t)3/2

• f SC ⇐⇒ f(z) = f(Tz) SC, for T nonsingular(i.e., SC is coordinate invariant)

• a large number of common convex functions are SC

x log x − log x, log detX−1, − log(y2 − xTx), . . .

ISIT 02 Lausanne 7/3/02 43

Page 45: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Complexity analysis of Newton’s method for

self-concordant functions

for self-concordant function f , with minimum value f?,

• theorem: #Newton steps to minimize f , starting from x:

#steps ≤ 11(f(x) − f?) + 5

• empirically: #steps ≈ 0.6(f(x) − f?) + 5

note absence of unknown constants, problem dimension, etc.

ISIT 02 Lausanne 7/3/02 44

Page 46: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

0 5 10 15 20 25 30 350

5

10

15

20

25

f(x(0)) − f?

#iter

atio

ns

ISIT 02 Lausanne 7/3/02 45

Page 47: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Complexity of path-following algorithm

• to compute x?(µt) starting from x?(t),

#steps ≤ 11m(µ − 1 − log µ) + 5

using N&N’s self-concordance theory, duality to bound f?

• number of outer steps to reduce duality gap by factor α: dlog α/ log µe

• total number of Newton steps bounded by product,

⌈log α

log µ

⌉(11m(µ − 1 − log µ) + 5)

. . . captures trade-off in choice of µ

ISIT 02 Lausanne 7/3/02 46

Page 48: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Complexity analysis conclusions

• for any choice of µ, #steps is O(m log 1/ε), where ε is final accuracy

• to optimize complexity bound, can take µ = 1 + 1/√

m, which yields#steps O(

√m log 1/ε)

• in any case, IP methods work extremely well in practice

ISIT 02 Lausanne 7/3/02 47

Page 49: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Conclusions

since 1985, lots of advances in theory & practice of convex optimization

• complexity analysis

• semidefinite programming, other new problem classes

• efficient interior-point methods & software

• lots of applications

ISIT 02 Lausanne 7/3/02 48

Page 50: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Some references

• Semidefinite Programming, SIAM Review 1996

• Determinant Maximization with Linear Matrix Inequality Constraints,SIMAX 1998

• Applications of Second-order Cone Programming, LAA 1999

• Linear Matrix Inequalities in System and Control Theory, SIAM 1994

• Interior-point Polynomial Algorithms in Convex Programming, SIAM1994, Nesterov & Nemirovsky

• Lectures on Modern Convex Optimization, SIAM 2001, Ben Tal &Nemirovsky

ISIT 02 Lausanne 7/3/02 49

Page 51: Advances in Convex Optimization: Theory, Algorithms, and ...boyd/papers/pdf/isit02.pdf · Convex analysis & optimization nice properties of convex optimization problems known since

Shameless promotion

Convex Optimization, Boyd & Vandenberghe

• to be published 2003

• pretty good draft available at Stanford EE364 (UCLA EE236B) classweb site as course reader

ISIT 02 Lausanne 7/3/02 50