23
Sparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail Rensselaer Polytechnic Institute (Joint work with C. Boutsidis and P. Drineas) June 20, 2013.

Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

  • Upload
    others

  • View
    20

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparsity in Machine Learning

Sparsifiers

SVDLinear Regression

K-means

M. Magdon-IsmailRensselaer Polytechnic Institute

(Joint work with C. Boutsidis and P. Drineas)

June 20, 2013.

Page 2: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Out-of-Sample Prediction

f = −1

f = +1

f = ?

• A pattern exists (f)

•We don’t know it

•We have data to learn it (D)

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 2 /23

Page 3: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Data

ndatapoints

— xt

1 —

— xt

2 —...

— xt

n —

=

d dimensions

x11 x12 . . . x1d

x21 x22 . . . x2d... ... ... ...

xn1 xn2 . . . xnd

yt

1

yt

2...

yt

n

X ∈ Rn×d Y ∈ Rn×ω

viewers × movie ratings

credit applicants × credit features y = ±1 (approve or not)

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 3 /23

Page 4: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Data

X ∈ R231×174 Y ∈ R231×166

number of principal components

%reconstructionerror

10 20 30 40 50 600

5

10

15

number of principal components

%reconstructionerror

10 20 30 40 50 600

5

10

15

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 4 /23

Page 5: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparsity

Represent your solution using only a few . . .

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 5 /23

Page 6: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparsity

Represent your solution using only a few . . .

Example: linear regression

[ ]

=

Xw = y

y is an optimal linear combination of the columns in X.

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 6 /23

Page 7: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparsity

Represent your solution using only a few . . .

Example: linear regression

[ ]

=

Xw = y

y is an optimal linear combination of only a few columns in X.

(sparse regression; regularization (||w ||0 ≤ k); feature subset selection; . . . )

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 7 /23

Page 8: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparsity is Good

Sparse solutions generalize to out-of-sample better.

Sparse solutions are easier to interpret.

Computations are more efficient.

Problem: sparsity is a combinatorial constraint.

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 8 /23

Page 9: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Singular Value Decomposition (SVD)

X =[Uk Ud−k

][Σk 0

0 Σd−k

] [Vt

k

Vt

d−k

]

O(nd2)

U Σ Vt

(n× d) (d× d) (d× d)

Xk = UkΣkVt

k

= XVkVt

k

Xk is the best rank-k approximation to X.

Reconstruction of X using only a few features.

X X20 X40 X60

Vk is an orthonormal basis for the best k-dimensional subspace of the row space of X.

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 9 /23

Page 10: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Vk and Sparsity

Principal Components Analysis (PCA):

Z = XVk

(n× k)

Feature subset selection: Important “dimensions” of Vt

k are important for X

×s1 ×s2 ×s3 ×s4 ×s5

−→

Vt

k Vt

k ∈ Rk×r

The sampled r columns are “good” if

I = Vt

kVk ≈ Vt

kVk.

Sampling schemes: Largest norm (Jollife, 1972);Randomized norm sampling (Rudelson, 1999; RudelsonVershynin, 2007);Greedy (Batson et al, 2009; BDM, 2011).

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 10 /23

Page 11: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Approximate SVD

X = XVkVt

k︸ ︷︷ ︸

Xk

+E

Let Vk be an approximate Vk

X = XVkVt

k + E

Vk is good if

|| E || ≤ (1 + ǫ)||E ||.

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 11 /23

Page 12: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Approximate SVD

1: Z = XR R ∼ N (d× r), Z ∈ Rn×r

2: Q = qr.factorize(Z)

3: Vk ← svdk(QtX)

Theorem. Let r =⌈k(1 + 1

ǫ)⌉and E = X− XVkV

t

k . Then,

E [||E ||] ≤ (1 + ǫ)||X− Xk ||

running time is O(ndk) = o(svd)

[BDM, FOCS 2011]

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 12 /23

Page 13: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Approximate SVD

k = 20 k = 40 k = 60

Exact SVD

Approx. SVD

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 13 /23

Page 14: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparse PCA

A principal component is a “dense” combination of the feature dimensions.

A sparse principal component is a combination of a few feature dimensions.

Want Vk to have a few non-zeros in each column

Vk =

• •

• •

• •

• •

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 14 /23

Page 15: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparse PCA

1: Choose a few columns C of X; C ∈ Rn×r.

2: Find the best rank-k approximation of X in the span of C, XC,k.

3: Compute the SVDk of XC,k:

XC,k = UC,kΣC,kVt

C,k.

4:

Z = XC,kVC,k = UC,kΣC,k.

Each feature in Z is a mixture of only the few original r feature dimensions in C.

||X− ZZ†X || ≤ ||X− ZVt

C,k || = ||X− XC,k ||.

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 15 /23

Page 16: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparse PCA

1: Choose a few columns C of X; C ∈ Rn×r.

2: Find the best rank-k approximation of X in the span of C, XC,k.

3: Compute the SVDk of

XC,k = UC,kΣC,kVt

C,k.

4:

Z = XC,kVC,k.

Each feature in Z is a mixture of only the few original r feature dimensions in C.

||X− ZZ†X || ≤ ||X− ZVt

C,k || = ||X− XC,k || ≤(1 +O(2k

r))||X− Xk ||.

[BDM, FOCS 2011]

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 16 /23

Page 17: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Sparse PCA

k = 20 k = 40 k = 60

Dense PCA

Sparse PCA, r = 2k

Theorem. One can construct, in o(svd), sparse features that are as good as exact

dense PCA-features.

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 17 /23

Page 18: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Feature Subset Selection: K-Means

Choose a few features

Cluster the data using these features

PCA - dense features.

Sparse features: feature subset selection.

Compare the clusterings on all the dimensions.

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 18 /23

Page 19: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Feature Subset Selection: K-Means

Full PCA, k = 20 Sparse, r = 2k

3 clusters

4 Clusters

Theorem. There is a subset of features of size O(#clusters) which produces nearly

the optimal partition (within a constant factor). One can quickly produce features with

a log-approximation factor.

[BDM,2013]

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 19 /23

Page 20: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Feature Subset Selection: Regression

=

X w = Y

Y

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 20 /23

Page 21: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

Feature Subset Selection: Regression

PCA Sparse, r = 2k

k = 20

k = 40

Theorem. There are O(k) pure features which performs as well regressing on pcak

features (to within small additive error).

[BDM,2013]

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 21 /23

Page 22: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

The Proofs

All the algorithms use the sparsifier of Vt

k in [BDM,FOCS2011].

1. Choose columns of Vt

k to preserve its singular values.

2. Ensure that the selected columns preserve the structural properties of the objective

with respect to the columns of X that are sampled.

(In all cases, the objective is a squared (Frobenius) error.)

c© Magdon-Ismail. Sparsity in Machine Learning: Slide 22 /23

Page 23: Sparsity in Machine Learning - cs.rpi.edumagdon/ps/conference/ICML-NLA2013.pdfSparsity in Machine Learning Sparsifiers SVD Linear Regression K-means M. Magdon-Ismail RensselaerPolytechnicInstitute

THANKS!

Focussed on columns of Vt

k to “sparsify” dimensions.

Can quickly approximate Vk.

Can efficiently use it to obtain

sparse PCA

small subset of features for k-means, which results in near optimal clustering.

small subset of features for regression, which results regression comparable to pcak.

Sparse solutions: easy to interpret; better generalizers; faster computations.

Using Uk instead of Vt

k one can “sparsify” data points to get coresets. [BDM,2013]