Probabilistic Semantic Similarity Measurements for Noisy Short Texts Using Wikipedia Entities Masumi...

Preview:

Citation preview

1

Probabilistic Semantic Similarity Measurements

for Noisy Short Texts Using Wikipedia Entities

Masumi Shirakawa1, Kotaro Nakayama2, Takahiro Hara1, Shojiro Nishio1

1Osaka University, Osaka, Japan2University of Tokyo, Tokyo, Japan

Challenge in short text analysis

Statistics are not always enough.

2

A year and a half after Google pulled its popular search engine out of mainland China

Baidu and Microsoft did not disclose terms of the agreement

Challenge in short text analysis

Statistics are not always enough.

3

A year and a half after Google pulled its popular search engine out of mainland China

Baidu and Microsoft did not disclose terms of the agreement

Search enginesand China

They are talking about...

Challenge in short text analysis

Statistics are not always enough.

4

A year and a half after Google pulled its popular search engine out of mainland China

Baidu and Microsoft did not disclose terms of the agreement

How do machines know that the two sentences mention about the similar

topic?

Search enginesand China

They are talking about...

Reasonable solution

Use external knowledge.

5

A year and a half after Google pulled its popular search engine out of mainland China

Baidu and Microsoft did not disclose terms of the agreement

Wikipedia Thesaurus [Nakayama06]

Related work

ESA: Explicit Semantic Analysis [Gabrilovich07]Add Wikipedia articles (entities) to a text as its semantic representation.

1. Get search ranking of Wikipedia for each term (i.e. Wiki articles and scores).2. Simply sum up the scores for aggregation.

6

Apple Inc.

Apple Inc.

pricing

pearApple sells

a new product

Apple

product

Key termextraction

Related entityfinding

Aggregation

iPhone

iPad

iPhonepear

Input: Tt c Output:

ranked list of csells

new

business

Problems in real world noisy short texts“Noisy” means semantically noisy in this work. (We do not handle informal or casual surface forms, or misspells)

Term ambiguity• Apple (fruit) should not be related with Microsoft.

Fluctuation of term dominance• A term is not always important in texts.

7

We explore more effective aggregation method.

We propose Extended naïve Bayes to aggregate related entities

Apple Inc.

Apple Inc.

Probabilistic method

8

𝑃 (𝑡∈𝑇 ) 𝑃 (𝑒|𝑡 ) 𝑃 (𝑐|𝑒)∏𝑘

𝑃 (𝑐|𝑡𝑘 )

𝑃 (𝑐 )𝐾−1

𝑃 (𝑐∨𝑇 )=∏𝑘=1

𝐾

(𝑃 (𝑡𝑘∈𝑇 )𝑃 (𝑐|𝑡𝑘)+(1−𝑃 (𝑡𝑘∈𝑇 ) )𝑃 (𝑐))𝑃 (𝑐 )𝐾− 1

From text T to related articles c

𝑃 (𝑐|𝑡 )

[Milne08]

[Mihalcea07] [Song1

1]

pricing

pearApple sells

a new product

Apple

product

Key termextraction

Related entityfinding

Aggregation

iPhone

iPad

iPhoneiPad

Input: T t c Output: ranked list of c

[Nakayama06]

When input is multiple terms

Apply naïve Bayes [Song11] to multiple terms to obtain related entity c using each probability P (c |tk ).

𝑃 (𝑐|𝑡1 ,…, 𝑡𝐾 )=𝑃 (𝑡 1 ,…, 𝑡𝐾|𝑐 )𝑃 (𝑐 )

𝑃 (𝑡1 ,…, 𝑡𝐾 )=𝑃 (𝑐 )∏

𝑘

𝑃 ( 𝑡𝑘|𝑐 )

𝑃 (𝑡 1 ,… ,𝑡𝐾 )=∏𝑘

𝑃 (𝑐|𝑡𝑘 )

𝑃 (𝑐 )𝐾−1

𝑡1

𝑡 2

𝑡𝐾

Apple

product

new

iPhone

Compute for each related entity c

𝑐

𝑃 (𝑐∨𝑡1)

𝑃 (𝑐∨𝑡2)

𝑃 (𝑐∨𝑡𝐾)

“iPhone”

9

When input is multiple terms

Apply naïve Bayes [Song11] to multiple terms to obtain related entity c using each probability P (c |tk ).

𝑃 (𝑐|𝑡1 ,…, 𝑡𝐾 )=𝑃 (𝑡 1 ,…, 𝑡𝐾|𝑐 )𝑃 (𝑐 )

𝑃 (𝑡1 ,…, 𝑡𝐾 )=𝑃 (𝑐 )∏

𝑘

𝑃 ( 𝑡𝑘|𝑐 )

𝑃 (𝑡 1 ,… ,𝑡𝐾 )=∏𝑘

𝑃 (𝑐|𝑡𝑘 )

𝑃 (𝑐 )𝐾−1

𝑡1

𝑡 2

𝑡𝐾

Apple

product

new

iPhone

Compute for each related entity c

𝑐

𝑃 (𝑐∨𝑡1)

𝑃 (𝑐∨𝑡2)

𝑃 (𝑐∨𝑡𝐾)

“iPhone”

10

By using naïve Bayes, entities that are related

to multiple terms can be boosted.

When input is text

Not “multiple terms” but “text,” i.e., we don’t know which terms are key terms.

We developed extended naïve Bayes to solve this problem.Cannot observewhich are key

terms

𝑡1

𝑡 2

𝑡𝐾

Apple

product

new

iPhone 𝑐

𝑃 (𝑐∨𝑡1)

𝑃 (𝑐∨𝑡2)

𝑃 (𝑐∨𝑡𝐾)

“iPhone”

11

Extended naïve Bayes𝑇 ′={𝑡1 }

𝑇 ′={𝑡1 , 𝑡2 }

𝑇 ′={𝑡1 ,⋯ ,𝑡𝐾 }

𝑇

Appleproduc

t

new

… Apple

Appleproduc

t

Appleproduc

t

new

Candidates of key

term

Apply naïve Bayesto each state T’

Probability that the set of key terms T is a

state T’ : …

12

Extended naïve Bayes𝑇 ′={𝑡1 }

𝑇 ′={𝑡1 , 𝑡2 }

𝑇 ′={𝑡1 ,⋯ ,𝑡𝐾 }

𝑇

Appleproduc

t

new

… Apple

Appleproduc

t

Appleproduc

t

new

Candidates of key

term

Apply naïve Bayesto each state T’

Probability that the set of key terms T is a

state T’ : ∑𝑇 ′

𝑃 (𝑐∨𝑇 ′)𝑃 (𝑇=𝑇 ′)=∏𝑘

(𝑃 ( 𝑡𝑘∈𝑇 )𝑃 (𝑐|𝑡𝑘 )+(1−𝑃 ( 𝑡𝑘∈𝑇 ))𝑃 (𝑐 ))𝑃 (𝑐 )𝐾−1

13

Extended naïve Bayes𝑇 ′={𝑡1 }

𝑇 ′={𝑡1 , 𝑡2 }

𝑇 ′={𝑡1 ,⋯ ,𝑡𝐾 }

𝑇

Appleproduc

t

new

… Apple

Appleproduc

t

Appleproduc

t

new

Candidates of key

term

Apply naïve Bayesto each state T’

Probability that the set of key terms T is a

state T’ : ∑𝑇 ′

𝑃 (𝑐∨𝑇 ′)𝑃 (𝑇=𝑇 ′)=∏𝑘

(𝑃 ( 𝑡𝑘∈𝑇 )𝑃 (𝑐|𝑡𝑘 )+(1−𝑃 ( 𝑡𝑘∈𝑇 ))𝑃 (𝑐 ))𝑃 (𝑐 )𝐾−1

14

Term dominance is incorporated into naïve Bayes

Experiments on short text sim datasets[Datasets] Four datasets derived from word similarity datasets using dictionary

[Comparative methods] Original ESA [Gabrilovich07], ESA with 16 parameter settings

[Metrics] Spearman’s rank correlation coefficient

15

ESA with well-adjusted parameter is superior to our method for “clean”

texts.

Tweet clustering

K-means clustering using the vector of related entities for measuring distance

[Dataset] 12,385 tweets including 13 topics

[Comparative methods] Bag-of-words (BOW), ESA with the same parameter, ESA with well-adjusted parameter

[Metric] Average of Normalized Mutual Information (NMI), 20 runs

16

#MacBook (1,251) #Silverlight (221)#VMWare (890)#MySQL (1,241) #Ubuntu (988) #Chrome (1,018)#NFL (1,044) #NHL (1,045)#NBA (1,085)#MLB (752) #MLS (981) #UFC (991)#NASCAR (878)

Results

10 20 50 100 200 500 1000 20000

0.1

0.2

0.3

0.4

0.5

0.6

0.429

0.421

0.524

0.567

BOW ESA-sameESA-adjusted Our method

Number of related entities

NM

I sc

ore p-value <

0.01

17

Results

10 20 50 100 200 500 1000 20000

0.1

0.2

0.3

0.4

0.5

0.6

0.429

0.421

0.524

0.567

BOW ESA-sameESA-adjusted Our method

Number of related entities

NM

I sc

ore p-value <

0.01

18

Our method outperformed ESA withwell-adjusted parameter for noisy short

texts.

Conclusion

We proposed extended naïve Bayes to derive related Wikipediaentities given a real world noisy short text.

[Future work]Tackle multilingual short textsDevelop applications of the method

19

Recommended