42
© Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September 28, 2016 Reading: class handouts

Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

  • Upload
    others

  • View
    10

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 11

Machine Learning

10-701, Fall 2016

Advanced topics in Max-Margin Learning

Eric Xing

Lecture 7, September 28, 2016Reading: class handouts

Page 2: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 22

Recap: the SVM problem We solve the following constrained opt problem:

This is a quadratic programming problem. A global maximum of i can always be found.

The solution:

How to predict:

m

iiii yw

1x

m

jij

Tijiji

m

ii yy

11 21

,)()(max xx J

.0

,,1 0, s.t.

1

m

iii

i

y

mi

Page 3: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 33

The SMO algorithm Consider solving the unconstrained opt problem:

We’ve already see three opt algorithms! Coordinate ascent Gradient ascent Newton-Raphson

Coordinate ascend:

Page 4: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 44

Coordinate ascend

Page 5: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 55

Sequential minimal optimization Constrained optimization:

Question: can we do coordinate along one direction at a time (i.e., hold all [-i] fixed, and update i?)

m

jij

Tijiji

m

ii yy

11 21

,)()(max xx J

.0

,,1 ,0 s.t.

1

m

iii

i

y

miC

Page 6: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 66

The SMO algorithm

Repeat till convergence

1. Select some pair i and j to update next (using a heuristic that tries to pick the two that will allow us to make the biggest progress towards the global maximum).

2. Re-optimize J() with respect to i and j, while holding all the other k 's (k i; j) fixed.

Will this procedure converge?

Page 7: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 77

Convergence of SMO

Let’s hold 3 ,…, m fixed and reopt J w.r.t. 1 and 2

m

jij

Tijiji

m

ii yy

1,1)(

21)(max xx J

.

,, ,0 s.t.

m

iii

i

y

kiC

10

1

KKT:

Page 8: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 88

Convergence of SMO The constraints:

The objective:

Constrained opt:

Page 9: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 99

Advanced topics in Max-Margin Learning

Kernel

Point rule or average rule

Can we predict vec(y)?

m

jij

Tijiji

m

ii yy

11 21

,)()(max xx J

Page 10: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1010

Outline

The Kernel trick

Maximum entropy discrimination

Structured SVM, aka, Maximum Margin Markov Networks

Page 11: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1111

(1) Non-linear Decision Boundary So far, we have only considered large-margin classifier with a

linear decision boundary How to generalize it to become nonlinear? Key idea: transform xi to a higher dimensional space to “make

life easier” Input space: the space the point xi are located Feature space: the space of (xi) after transformation

Why transform? Linear operation in the feature space is equivalent to non-linear operation in input

space Classification can become easier with a proper transformation. In the XOR

problem, for example, adding a new feature of x1x2 make the problem linearly separable (homework)

Page 12: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 12

Non-linear Decision Boundary

12

Page 13: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1313

Transforming the Data

( )

( )

( )( )( )

( )

( )( )

(.) ( )

( )

( )( )( )

( )

( )

( )( ) ( )

Feature spaceInput spaceNote: feature space is of higher dimension than the input space in practice

Page 14: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1414

The Kernel Trick Recall the SVM optimization problem

The data points only appear as inner product As long as we can calculate the inner product in the feature

space, we do not need the mapping explicitly Many common geometric operations (angles, distances) can

be expressed by inner products Define the kernel function K by

m

jij

Tijiji

m

ii yy

11 21

,)()(max xx J

.0

,,1 ,0 s.t.

1

m

iii

i

y

miC

)()(),( jT

ijiK xxxx

Page 15: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1515

An Example for feature mapping and kernels Consider an input x=[x1,x2] Suppose (.) is given as follows

An inner product in the feature space is

So, if we define the kernel function as follows, there is no need to carry out (.) explicitly

2122

2121

2

1 2221 xxxxxxxx

,,,,,

''

,2

1

2

1

xx

xx

21 ')',( xxxx TK

Page 16: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1616

More examples of kernel functions Linear kernel (we've seen it)

Polynomial kernel (we just saw an example)

where p = 2, 3, … To get the feature vectors we concatenate all pth order polynomial terms of the components of x (weighted appropriately)

Radial basis kernel

In this case the feature space consists of functions and results in a non-parametric classifier.

')',( xxxx TK

pTK ')',( xxxx 1

2

21 'exp)',( xxxxK

Page 17: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1717

The essence of kernel Feature mapping, but “without paying a cost”

E.g., polynomial kernel

How many dimensions we’ve got in the new space? How many operations it takes to compute K()?

Kernel design, any principle? K(x,z) can be thought of as a similarity function between x and z This intuition can be well reflected in the following “Gaussian” function

(Similarly one can easily come up with other K() in the same spirit)

Is this necessarily lead to a “legal” kernel?(in the above particular case, K() is a legal one, do you know how many dimension (x) is?

Page 18: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1818

Kernel matrix Suppose for now that K is indeed a valid kernel corresponding

to some feature mapping , then for x1, …, xm, we can compute an mm matrix , where

This is called a kernel matrix!

Now, if a kernel function is indeed a valid kernel, and its elements are dot-product in the transformed feature space, it must satisfy: Symmetry K=KT

proof

Positive –semidefiniteproof?

Page 19: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 1919

Mercer kernel

Page 20: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2020

SVM examples

Page 21: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2121

Examples for Non Linear SVMs –Gaussian Kernel

Page 22: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2222

(2) Model averaging Inputs x, class y = +1, -1 data D = { (x1,y1), …. (xm,ym) }

Point Rule:

learn fopt(x) discriminant functionfrom F = {f} family of discriminants

classify y = sign fopt(x)

E.g., SVM

Page 23: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2323

Model averaging There exist many f with near optimal performance

Instead of choosing fopt, average over all f in F

Q(f) = weight of f

How to specify:F = { f } family of discriminant functions?

How to learn Q(f) distribution over F?

Page 24: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2424

Bayesian learning:

Bayes Predictor (model averaging):

What p0?

Recall Bayesian Inference

Bayes Learner

Recall in SVM:

Page 25: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2525

How to score distributions?

Entropy Entropy H(X) of a random variable X

H(X) is the expected number of bits needed to encode a randomly drawn value of X (under most efficient code)

Why?

Information theory:Most efficient code assigns -log2P(X=i) bits to encode the message X=I, So, expected number of bits to code one random X is:

Page 26: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2626

Given data set , find

solution QME correctly classifies D among all admissible Q, QME has max entropy max entropy "minimum assumption" about f

Maximum Entropy Discrimination

Page 27: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2727

Introducing Priors Prior Q0( f )

Minimum Relative Entropy Discrimination

Convex problem: QMRE unique solution MER "minimum additional assumption" over Q0 about f

p

Page 28: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2828

Convex problem: QME unique

Theorem:

i 0 Lagrange multipliers

finding QM : start with i 0 and follow gradient of unsatisfied constraints

Solution: Q ME as a projection

uniformQ0

QME

admissible Q

=0

ME

Page 29: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 2929

Solution to MED Theorem (Solution to MED):

– Posterior Distribution:

– Dual Optimization Problem:

Algorithm: to computer t , t = 1,...T

start with t = 0 (uniform distribution)

iterative ascent on J() until convergence

Page 30: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3030

Examples: SVMs Theorem

For f(x) =wTx + b, Q0(w) = Normal( 0, I ), Q0(b) = non-informative prior,the Lagrange multipliers are obtained by maximizing J() subject to 0t C and t tyt = 0, where

Separable D SVM recovered exactly Inseparable D SVM recovered with different

misclassification penalty

Page 31: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3131

SVM extensions

MRE Gaussian

Linear SVM

Max Likelihood Gaussian

Example: Leptograpsus Crabs (5 inputs, Ttrain=80, Ttest=120)

Page 32: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3232

(3) Structured Prediction

“Do you want sugar in it?” <verb pron verb noun prep pron>

Unstructured prediction

Structured prediction Part of speech tagging

Image segmentation

Page 33: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3333

OCR example

braceSequential structure

x y

a-z a-z a-z a-z a-zy

x

Page 34: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3434

Inputs: a set of training samples , where

and

Outputs: a predictive function :

Examples: SVM:

Logistic Regression:

where

Classical Classification Models

Page 35: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3535

Assumptions:

Linear combination of features

Sum of partial scores: index p represents a part in the structure

Random fields or Markov network features:

Structured Models

space of feasible outputsdiscriminant function

Page 36: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3636

Discriminative Learning Strategies Max Conditional Likelihood

We predict based on:

And we learn based on:

Max Margin: We predict based on:

And we learn based on:

c

ccc fwZ

p ),(exp),(

1)|(maxarg|* yxxw

xyxy wy

i c

iiccii

iiii fwZ

p ),(exp),(

1)|(maxarg,|* yxxw

xyxyw ww

),(maxarg),(maxarg|* yxwyxxyy

ffw T

ycccc

),(),(minmaxarg,|

,

*iii

T

iii ff

ixyxywxyw

yyw

Page 37: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3737

E.g. Max-Margin Markov Networks Convex Optimization Problem:

Feasible subspace of weights:

Predictive Function:

Page 38: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3838

OCR Example

We want:argmaxword wT f( , word) = “brace”

Equivalently:wT f( ,“brace”) > wT f( ,“aaaaa”)wT f( ,“brace”) > wT f( ,“aaaab”)…wT f( ,“brace”) > wT f( ,“zzzzz”)

a lot!

Page 39: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 3939

Brute force enumeration of constraints:

The constraints are exponential in the size of the structure

Alternative: min-max formulation add only the most violated constraint

Handles more general loss functions Only polynomial # of constraints needed Several algorithms exist …

Min-max Formulation

Page 40: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 4040

Results: Handwriting Recognition

Length: ~8 charsLetter: 16x8 pixels 10-fold Train/Test5000/50000 letters600/6000 words

Models:Multiclass-SVMs*M3 nets

*Crammer & Singer 01

0

5

10

15

20

25

30

MC–SVMs M^3 nets

Test

err

or (a

vera

ge p

er-c

hara

cter

) rawpixels

quadratickernel

cubickernel

33% error reduction over multiclass SVMs

better

Page 41: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 4141

Discriminative Learning Paradigms

?

SVM SVM b r a c e

M3N

MED MED

M3N

MED-MN= SMED + Bayesian M3N

See [Zhu and Xing, 2008]

Page 42: Machine Learningmgormley/courses/10701-f16/... · © Eric Xing @ CMU, 2006-2016 1 Machine Learning 10-701, Fall 2016 Advanced topics in Max-Margin Learning Eric Xing Lecture 7, September

© Eric Xing @ CMU, 2006-2016 4242

Summary Maximum margin nonlinear separator

Kernel trick Project into linearly separatable space (possibly high or infinite dimensional) No need to know the explicit projection function

Max-entropy discrimination Average rule for prediction, Average taken over a posterior distribution of w who defines the separation

hyperplane P(w) is obtained by max-entropy or min-KL principle, subject to expected

marginal constraints on the training examples

Max-margin Markov network Multi-variate, rather than uni-variate output Y Variable in the outputs are not independent of each other (structured input/output) Margin constraint over every possible configuration of Y (exponentially many!)