26
Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Embed Size (px)

Citation preview

Page 1: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Posterior Regularization for Structured Latent Variable Models

Li ZhonghuaI2R SMT Reading Group

Page 2: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Outline

• Motivation and Introduction• Posterior Regularization• Application• Implementation• Some Related Frameworks

Page 3: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Motivation and Introduction

Prior Knowledge

We posses a wealth of prior knowledge about most NLP tasks.

Page 4: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Motivation and Introduction --Prior Knowledge

Page 5: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Motivation and Introduction --Prior Knowledge

Page 6: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Motivation and Introduction

Leveraging Prior Knowledge Possible approaches and their limitations

Page 7: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Motivation and Introduction --Limited Approach

Bayesian Approach : Encode prior knowledge with a prior on parameters

Limitation: Our prior knowledge is not about parameters!Parameters are difficult to interpret; hard to get desired effect.

Page 8: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Augmenting Model : Encode prior knowledge with additional variables and dependencies.

Motivation and Introduction --Limited Approach

limitation: may make exact inference intractable

Page 9: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Posterior Regularization

• A declarative language for specifying prior knowledge

-- Constraint Features & Expectations

• Methods for learning with knowledge in this language

-- EM style learning algorithm

Page 10: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Posterior Regularization

Page 11: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Posterior Regularization

Original Objective :

Page 12: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Posterior RegularizationEM style learning algorithm

Page 13: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Posterior Regularization

Computing the Posterior Regularizer

Page 14: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Application

Statistical Word AlignmentsIBM Model 1 and HMM

Page 15: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Application

One feature for each source word m, that counts how many times it is aligned to a target word in the alignment y.

Page 16: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Application

Define feature for each target-source position pair i,j . The feature takes the value zero in expectation if a word pair i ,j is aligned with equal probability in both directions.

Page 17: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Application

Learning Tractable Word Alignment Models with Complex Constraints CL10

Page 18: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Application

• Six language pairs• both types of constraints improve over the

HMM in terms of both precision and recall• improve over the HMM by 10% to 15%• S-HMM performs slightly better than B-HMM• S-HMM performs better than B-HMM in 10

out of 12 cases• improve over IBM M4 9 times out of 12

Page 19: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Application

Page 20: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Implementation

• http://code.google.com/p/pr-toolkit/

Page 21: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Some Related Frameworks

Page 22: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Some Related Frameworks

Page 23: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Some Related Frameworks

Page 24: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Some Related Frameworks

Page 25: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Some Related Frameworks

Page 26: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

more info: http://sideinfo.wikkii.com many of my slides get from there

Thanks!