46
Randomized Algorithms for Matrix Computations Ilse Ipsen Students: John Holodnak, Kevin Penner, Thomas Wentworth Research supported in part by NSF CISE CCF, DARPA XData

Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

  • Upload
    vodat

  • View
    221

  • Download
    1

Embed Size (px)

Citation preview

Page 1: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Randomized Algorithms for Matrix Computations

Ilse Ipsen

Students: John Holodnak, Kevin Penner, Thomas Wentworth

Research supported in part by NSF CISE CCF, DARPA XData

Page 2: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Randomized Algorithms

Solve a deterministic problem by statistical sampling

Monte Carlo MethodsVon Neumann & Ulam, Los Alamos, 1946

http://en.wikipedia.org/wiki/Monte_Carlo_integration

Simulated Annealing: global optimization

Page 3: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Application of Randomized Algorithms

Astronomy: Tamas Budavari (Johns Hopkins)

Classification of galaxiesSingular Value Decomposition/PCA, subset selection

Nuclear Engineering: Hany Abdel-Khalik (NCSU)

Model reduction in nuclear fissionLow-rank approximation

Population Genomics: Petros Drineas (RPI)

Find SNPs that predict population associationSVD/PCA, subset selection

Software for Machine Learning Applications:Matrix multiplication, linear systems, least squares,SVD/PCA, low rank approximation, matrix functions

Michael Mahoney (Stanford), Ken Clarkson and David Woodruf (IBM Almaden)

Haim Avron and Christos Boutsidis (IBM Watson)

Page 4: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

This Talk

Least squares/regression problems

Traditional, deterministic algorithms

A randomized algorithm

Sampling

A probabilistic bound

Coherence

Leverage scores

Summary

Page 5: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Least Squares/Regression Problems:Traditional, Deterministic Algorithms

Page 6: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Least Squares for Dense Matrices

Given:

Real m × n matrix A with rank(A) = nReal m × 1 vector b

Want: minx ‖Ax − b‖2 (two-norm)

Unique solution x = A†b, A† = (ATA)−1AT

Direct method:QR factorization A = QR QTQ = In, R is △

Solve Rx = QTb

Operation count: O(mn2)

Page 7: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

QR Factorization of Tall & Skinny Matrix

!

!

!

!

!

!

!

!

!

!

!

!

!

!

! !

!!!!!!!! !!!!!!!! ! !!!!!!!!!!!!!!!!!!!!!!!!!! !!!!!!!!!!!!!

!

!

!

!

!

!

!!!!!!!!!! !!!!!!!!!!!!!!!!!!!!!!!!"!!!!!!!!!!!!!!!!#!!!!!!!!!!!!!$!

Page 8: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Least Squares for Dense Matrices

Given:

Real m × n matrix A with rank(A) = nReal m × 1 vector b

Want: minx ‖Ax − b‖2 (two-norm)

Unique solution x = A†b, A† = (ATA)−1AT

Direct method:QR factorization A = QR QTQ = In, R is △

Solve Rx = QTb

Operation count: O(mn2)

But: Too expensive when A is large & sparseQR factorization produces fill-in

Page 9: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Least Squares for Sparse Matrices:

LSQR [Paige & Saunders 1982]

Krylov space method:Iterative method based on matrix vector multiplications

Iteration k :

xk ∈ span{

(ATA)k ATb, . . . , (ATA)ATb, ATb}

Residuals decrease (in exact arithmetic)

‖b − Axk‖2 ≤ ‖b − Axk−1‖2

Convergence fast if condition number κ(A) small

κ(A) = ‖A‖2‖A†‖2 ≥ 1

Page 10: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Accelerating Convergence

Find preconditioner P with

κ(AP) smallLinear system Px = rhs is easy to solve

Preconditioned least squares problem

Solve miny ‖AP y − b‖2 by LSQR {fast convergence}

Retrieve solution to original problem: x = Py

The ideal preconditioner

QR factorization A = QR QTQ = In, R is △

Preconditioner P = R−1 ⇒ AR−1 = QSolve miny ‖Qy − b‖2 by LSQRRetrieve solution: Solve Rx = y {△ system}

LSQR converges in 1 iteration: κ(AR−1) = κ(Q) = 1

Construction of preconditioner is way too expensive!

Page 11: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

A Randomized Preconditionerfor Least Squares Problems

Page 12: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

A Cheaper Preconditioner

Ideal preconditioner

QR factorization A = QR QTQ = In, R is △

Preconditioner R−1

AR−1 = Q has orthonormal columnsPerfect condition number κ(AR−1) = κ(Q) = 1

Randomized preconditionerQR factorization from only a few rows of A

Sample c ≥ n rows of A: SAQR factorization SA = QsRs QT

s Qs = In, Rs is △

Randomized preconditioner R−1s

Page 13: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

QR Factorizations of Original and Sampled Matrices

! !

!

!

!

!

!

!

!

!

!

!

!

!

!

!

!

!

! !

"!!!!!!!!!!!#!!!!!!!!!!!!!!!!$!!!!!!!!!!!!!!!!!!%!!!!!!!!!!!!!

!

!

!

!

!

!

!

!

!

!

!

!

!!!!!!!!!!!!!!!!!!!!

!!!!!!!! !!!!!!!"!!!!!!!!!!!!!!! !!!! !!!!!!!!!!!!!!!! !!!!!!!!!!!!!!!!

!!!!!!!!!!!!!!!!!!!!

Page 14: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Blendenpik [Avron, Maymounkov & Toledo 2010]

Solve minz ‖Az − b‖2

A is m × n, rank(A) = n and m ≫ n

{Construct preconditioner}Sample c ≥ n rows of A → SA {fewer rows}

QR factorization SA = QsRs

{Solve preconditioned problem}

Solve miny ‖AR−1s y − b‖2 with LSQR

Solve Rs z = y {△ system}

Hope:

AR−1s has almost orthonormal columns

Condition number almost perfect: κ(AR−1s ) ≈ 1

Page 15: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

From Sampling to Condition Numbers

[Avron, Maymounkov & Toledo 2010]

ComputedQR decomposition of sampled matrix: SA = QsRs

ConceptualQR decomposition of full matrix: A = QR

The Idea

1 Sampling rows of A ≡ Sampling rows of Q

2 Condition number of preconditioned matrix:

κ(AR−1s ) = κ(SQ)

We analyze κ(SQ)Sampled matrices with orthonormal columns

Page 16: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Sampling:

Uniform sampling of c out of m rowsfrom matrices with orthonormal columns

Page 17: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Uniform Sampling with Replacement (Exactly(c))

[Drineas, Kannan & Mahoney 2006]

for t = 1 : c doSample kt from {1, . . . ,m} with probability 1/m

independently and with replacementend for

Sampling matrix S =√

mc

eTk1...eTkc

S is c ×m, and samples exactly c rows

Expected value E(STS) = Im

S can sample a row more than once

Page 18: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Issues

How to implement sampling?

How large is the condition number of the sampled matrices?

How many sampled matrices are rank deficient?

Isn’t sampling without replacement a lot better?

Page 19: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

How to Implement Uniform Sampling

Sample kt from {1, . . . ,m} with probability 1/mindependently and with replacement

[Devroye 1986]

η = rand {uniform [0, 1] random variable}

kt = ⌊1 +m η⌋

Matlab: kt = randi(m)

Page 20: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Condition Numbers and Rank Deficiency

Sampling c rows from m × n matrix Q with QTQ = Inm = 104, n = 5 (30 runs for each value of c)

Sampled matrices SQ from three strategies:

Sampling with replacement (Exactly(c))Sampling without replacementBernoulli sampling

Plots:

1 Condition number of SQκ(SQ) = ‖SQ‖2 ‖(SQ)†‖2 (if SQ has full column rank)

2 Percentage of matrices SQ that are rank deficient(κ(SQ) ≥ 1016)

Page 21: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Comparison of Sampling StrategiesSampling with replacement (Exactly(c))

Sampling without replacement

Bernoulli sampling

Page 22: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Comparison of Sampling Strategies

Sampled matrices SQ from three strategies:

Sampling with replacement (Exactly(c))Sampling without replacementBernoulli sampling

Comparison: Summary

Little difference among the sampling strategiesIf SQ has full rank then κ(SQ) ≤ 10

Advantages of sampling with replacement

Fast: need to generate/inspect only c valuesEasy to implementReplacement does not affect accuracy

(for small amounts of sampling)

Page 23: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

A Probabilistic Bound forThe Condition Number of the Sampled Matrices

Page 24: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Setup for the Bound

Given: m × n matrix Q with orthonormal columns (QTQ = In)

Sampling: c ×m matrix

S =

m

c

eTk1...eTkc

Unbiased estimator: E[

QTSTSQ]

= QTQ = In

Sum of c random matrices: QTSTSQ = X1 + · · · + Xc

Xt =mcQT ekte

TktQ, 1 ≤ t ≤ c

Page 25: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Bernstein-Type Concentration Inequality [Recht 2011]

Yt independent random n × n matrices with E[Yt ] = 0

‖Yt‖2 ≤ τ almost surely

ρt ≡ max{‖E[YtYTt ]‖2, ‖E[Y

Tt Yt ]‖2}

Desired error 0 < ǫ < 1

Failure probability δ = 2n exp(

−32

ǫ2

3∑

t ρt+τ ǫ

)

With probability at least 1− δ

t

Yt

2

≤ ǫ {Deviation from mean}

Page 26: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Applying the Concentration Inequality

Sampled matrix:

QTSTSQ = X1 + · · ·+ Xc , Xt =mcQT ekt e

TktQ

Zero mean version:

QTSTSQ − In = Y1 + · · ·+ Yc , Yt = Xt −1cIn

By construction: E[Yt ] = 0

‖Yt‖2 ≤mcµ, E[Y 2

t ] ≤mc2

µ

Largest row norm squared: µ = max1≤j≤m ‖eTj Q‖22

With probability at least 1− δ, ‖(SQ)T (SQ)− In‖2 ≤ ǫ

Page 27: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Condition Number Bound

m × n matrix Q with orthonormal columns

Largest row norm squared: µ = max1≤j≤m ‖eTj Q‖22Number of rows to be sampled: c ≥ n

0 < ǫ < 1

Failure probability

δ = 2n exp

(

−c

m µ

ǫ2

3 + ǫ

)

With probability at least 1− δ:

κ(SQ) ≤

1 + ǫ

1− ǫ

Page 28: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Tightness of Condition Number Bound

Input: m × n matrix Q with QTQ = In withm = 104, n = 5, µ = 1.5 n/m

1 Exact condition number from sampling with replacement

Little sampling: n ≤ c ≤ 1000A lot of sampling: 1000 ≤ c ≤ m

2 Condition number bound√

1+ǫ

1−ǫ

where success probability 1− δ ≡ .99

ǫ ≡1

2c

(

ℓ+√

12cℓ+ ℓ2)

ℓ ≡ 23 (m µ− 1) ln(2n/ δ)

Page 29: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Little sampling (n ≤ c ≤ 1000)

0 100 200 300 400 500 600 700 800 900 10001

10

c

κ(S

Q)

Bound holds for c ≥ 93 ≈ 2(mµ − 1) ln(2n/δ)/ǫ2

Page 30: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

A lot of sampling (1000 ≤ c ≤ m)

1000 2000 3000 4000 5000 6000 7000 8000 9000 100001

1.2

1.3

1.4

1.5

1.1

c

κ(S

Q)

Bound predicts correct magnitude of condition number

Page 31: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Condition Number Bound

m × n matrix Q with orthonormal columns

Largest row norm squared: µ = max1≤j≤m ‖eTj Q‖22Number of rows to be sampled: c ≥ n

0 < ǫ < 1

Failure probability

δ = 2n exp

(

−c

m µ

ǫ2

3 + ǫ

)

With probability at least 1− δ:

κ(SQ) ≤

1 + ǫ

1− ǫ

The only distinction among different m × n matrices Q withorthonormal columns is µ

Page 32: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Conclusions from the Bound

Correct magnitude for condition number of sampled matrix,even for small matrix dimensions

Lower bound on number of sampled rows

c = O (m µ ln n)

Important ingredient µ = max1≤j≤m ‖eTj Q‖22

Page 33: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Coherence

µ = max1≤j≤m ‖eTj Q‖22

Largest squared row norm ofmatrix with orthonormal columns

Page 34: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Properties of Coherence

Coherence of m × n matrix Q with QTQ = In

µ = max1≤j≤m

‖eTj Q‖22

n/m ≤ µ(Q) ≤ 1

Maximal coherence: µ(Q) = 1At least one column of Q is a canonical vector

Minimal coherence: µ(Q) = n/mColumns of Q are columns of a Hadamard matrix

Coherence measures “correlation with standard basis”

Page 35: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Coherence in General

Donoho & Huo 2001Mutual coherence of two bases

Candes, Romberg & Tao 2006

Candes & Recht 2009Matrix completion: Recovering a low-rank matrixby sampling its entries

Mori & Talwalkar 2010, 2011

Estimation of coherence

Avron, Maymounkov & Toledo 2010

Randomized preconditioners for least squares

Drineas, Magdon-Ismail, Mahoney & Woodruff 2011

Fast approximation of coherence

Page 36: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Different Definitions

Coherence of subspaceQ is subspace of Rm of dimension nP orthogonal projector onto Q

µ0(Q) =m

nmax

1≤j≤m‖eTj P‖22 (1 ≤ µ0 ≤

mn)

Coherence of full rank matrixA is m × n with rank(A) = nColumns of Q are orthonormal basis for R(A)

µ(A) = max1≤j≤m

‖eTj Q‖22 ( nm

≤ µ ≤ 1)

Reflects difficulty of recovering the matrix from sampling

Page 37: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Effect of Coherence on Sampling

Input: m × n matrix Q with QTQ = Inm = 104, n = 5

Sampling with replacement

1 Low coherence: µ = 7.5 · 10−4 = 1.5 n/m

2 Higher coherence: µ = 7.5 · 10−2 = 150 n/m

Page 38: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Low Coherence

0 100 200 300 400 500 600 700 800 900 1000

1.5

2

3

4

c

κ(S

Q)

0 100 200 300 400 500 600 700 800 900 10000

25

50

75

100

c

Fai

lure

Per

cen

tag

e

Only a single failure (for c = 5)

Page 39: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Higher Coherence

200 300 400 500 600 700 800 900 1000

1.5

2

3

4

5

c

κ(S

Q)

0 100 200 300 400 500 600 700 800 900 10000

25

50

75

100

90

c

Fai

lure

Per

cen

tag

e

Very high failure rate when sampling at most 10% of rows

Page 40: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Improving on Coherence:

Leverage Scores

Page 41: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Leverage Scores

Idea: Use all row norms

Q is m × n with orthonormal columns

Leverage scores = squared row norms

ℓk = ‖eTk Q‖22, 1 ≤ k ≤ m

Coherence µ = maxk ℓk

Low coherence ≈ uniform leverage scores

Leverage scores of full column rank matrix A:Columns of Q are orthonormal basis for R(A)

ℓk(A) = ‖eTk Q‖22, 1 ≤ k ≤ m

Page 42: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Statistical Leverage Scores

Hoaglin & Welsch 1978Chatterjee & Hadi 1986

Identify potential outliers in minx ‖Ax − b‖2

Hb: Projection of b onto R(A) where H = A(ATA)−1AT

Leverage score: Hkk ∼ influence of kth data point on LS fit

QR decomposition: A = QR

Hkk = ‖eTk Q‖22 = ℓk(A)

Application to randomized algorithms: Mahoney & al. 2006–2012

Page 43: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Leverage Score Bound

m × n matrix Q with orthonormal columns

Leverage scores ℓj = ‖eTj Q‖22, µ = max1≤j≤m ℓj

L = diag(

ℓ1 . . . ℓm)

Sampling with replacement

0 < ǫ < 1

Failure probability

δ = 2n exp

(

−32

c ǫ2

m (3 ‖QTLQ‖2 + µ ǫ

)

With probability at least 1− δ: κ(SQ) ≤√

1+ǫ

1−ǫ

Page 44: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Improvement with Leverage Score Bound

Low coherence: µ = 1.5n/m, small amounts of sampling

0 100 200 300 400 500 600 700 800 900 10001

1.5

2

c

κ(S

Q)

Leverage score bound vs. Coherence bound(m = 104, n = 4, δ = .01)

Page 45: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Summary

Least squares problems minz ‖Az − b‖2A is m × n, rank(A) = n and m ≫ n

Solution by iterative Krylov space method LSQR

Convergence depends on condition number κ(A) = ‖A‖2‖A†‖2

Acceleration of convergence: Precondition the matrix

Randomized preconditionerκ(preconditioned matrix) = κ(sampled orthonormal matrix)

Sampling with replacement

Bounds for condition number of sampled orthonormal matricesExplicit and non-asymptotic, predictive even for small matrix dimensions

Sampling performance depends on coherence

Tighter bounds: Replace coherence by leverage scores

Page 46: Randomized Algorithms for Matrix Computationsipsen/ps/slides_chicago.pdf · Research supported in part by NSF CISE CCF, DARPA XData. Randomized Algorithms Solve a deterministic problem

Existing Work

Randomized algorithms for least squares

Drineas, Mahoney & Muthukrishnan 2006

Drineas, Mahoney, Muthukrishnan & Sarlos 2006

Rokhlin & Tygert 2008

Boutsidis & Drineas 2009

Blendenpik: Avron, Maymounkov & Toledo 2010

LSRN: Meng, Saunders & Mahoney 2011

Survey papers for randomized algorithms

Halko, Martinsson & Tropp 2011

Mahoney 2011

Graph theory

Preconditioning for graph Laplacians, graph sparsificationSpielman & Teng 2006, Koutis, Miller & Peng 2012, ....

Effective resistance = leverage scoresDrineas & Mahoney 2010