20
COLING 2016 Citation count: 1

COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

COLING 2016Citation count: 1

Page 2: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Motivation

• Incomplete questions do not make sense to conversation system

Page 3: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Methodology

Page 4: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Challenge

• How to handle OOV in small dataset (7k)

Page 5: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Syntactic Sequence Model

Page 6: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Problem

• Fails to utilize the meaning of each word, only the syntactic information

Page 7: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Semantic Sequence Model

• Assign each OOV a category number by k-means algorithm

• Use word embedding as clustering features

Page 8: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Dataset

• Collected from Amazon Mechanical Turk

• 7220 conversations (Q1 A1 Q2 R1)

• 134K/65K words for input/output sequence text

Page 9: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Experiment

Page 10: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

SIGIR 2017Citation count: 2

Page 11: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Methodology

Page 12: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Offline �Question Template Library

• 6420 → 5451

Page 13: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Problem

• Template library size is still too large (5k)

• Make online inference slow

Page 14: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Candidate �Question Template Selector

Page 15: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Re-ranking with Language Model

• Template selector seq2seq score:

• Language model score:

• Total score:

Page 16: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Same Dataset

• Collected from Amazon Mechanical Turk

• 7220 conversations (Q1 A1 Q2 R1)

• 134K/65K words for input/output sequence text

Page 17: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Experiment

Page 18: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Window Size of Pre�fix Tree

Page 19: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Siri Experiment (only dataset released)

Page 20: COLING 2016 Citation count: 1 · •7220 conversations (Q1 A1 Q2 R1) •134K/65K words for input/output sequence text. Experiment. SIGIR 2017 Citation count: 2. Methodology. Offline

Conclusion

• Syntactic information is important

• Template method works well on the dataset

• Use language model to further enhance performance