statistical nlp winter 2008

36
Statistical NLP Winter 2008 Lecture 5: word-sense disambiguation Roger Levy Thanks to Jason Eisner and Dan Klein

Upload: meagan

Post on 23-Jan-2016

31 views

Category:

Documents


0 download

DESCRIPTION

Statistical NLP Winter 2008. Lecture 5: word-sense disambiguation. Roger Levy Thanks to Jason Eisner and Dan Klein. A Concordance for “party” from www.webcorp.org.uk. A Concordance for “party” from www.webcorp.org.uk. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Statistical NLP Winter 2008

Statistical NLPWinter 2008

Lecture 5: word-sense disambiguation

Roger Levy

Thanks to Jason Eisner and Dan Klein

Page 2: Statistical NLP Winter 2008

A Concordance for “party” from www.webcorp.org.uk

Page 3: Statistical NLP Winter 2008

A Concordance for “party” from www.webcorp.org.uk

• thing. She was talking at a party thrown at Daphne's restaurant in • have turned it into the hot dinner-party topic. The comedy is the• selection for the World Cup party, which will be announced on May 1 • in the 1983 general election for a party which, when it could not bear to • to attack the Scottish National Party, who look set to seize Perth and • that had been passed to a second party who made a financial decision• the by-pass there will be a street party. "Then," he says, "we are going • number-crunchers within the Labour party, there now seems little doubt• political tradition and the same party. They are both relatively Anglophilic • he told Tony Blair's modernised party they must not retreat into "warm • "Oh no, I'm just here for the party," they said. "I think it's terrible • A future obliges each party to the contract to fulfil it by• be signed by or on behalf of each party to the contract." Mr David N

Page 4: Statistical NLP Winter 2008

What Good are Word Senses?

• thing. She was talking at a party thrown at Daphne's restaurant in • have turned it into the hot dinner-party topic. The comedy is the• selection for the World Cup party, which will be announced on May 1 • in the 1983 general election for a party which, when it could not bear to • to attack the Scottish National Party, who look set to seize Perth and • that had been passed to a second party who made a financial decision• the by-pass there will be a street party. "Then," he says, "we are going • number-crunchers within the Labour party, there now seems little doubt• political tradition and the same party. They are both relatively Anglophilic • he told Tony Blair's modernised party they must not retreat into "warm • "Oh no, I'm just here for the party," they said. "I think it's terrible • A future obliges each party to the contract to fulfil it by• be signed by or on behalf of each party to the contract." Mr David N

Page 5: Statistical NLP Winter 2008

What Good are Word Senses?

• thing. She was talking at a party thrown at Daphne's restaurant in • have turned it into the hot dinner-party topic. The comedy is the• selection for the World Cup party, which will be announced on May 1 • the by-pass there will be a street party. "Then," he says, "we are going • "Oh no, I'm just here for the party," they said. "I think it's terrible

• in the 1983 general election for a party which, when it could not bear to • to attack the Scottish National Party, who look set to seize Perth and • number-crunchers within the Labour party, there now seems little doubt• political tradition and the same party. They are both relatively Anglophilic • he told Tony Blair's modernised party they must not retreat into "warm

• that had been passed to a second party who made a financial decision• A future obliges each party to the contract to fulfil it by• be signed by or on behalf of each party to the contract." Mr David N

Page 6: Statistical NLP Winter 2008

What Good are Word Senses?

• Replace word w with sense s • Splits w into senses: distinguishes this token of w from

tokens with sense t• Groups w with other words: groups this token of w with

tokens of x that also have sense s

Page 7: Statistical NLP Winter 2008

Word senses create groupings

• number-crunchers within the Labour party, there now seems little doubt• political tradition and the same party. They are both relatively Anglophilic • he told Tony Blair's modernised party they must not retreat into "warm • thing. She was talking at a party thrown at Daphne's restaurant in • have turned it into the hot dinner-party topic. The comedy is the• selection for the World Cup party, which will be announced on May 1 • the by-pass there will be a street party. "Then," he says, "we are going • "Oh no, I'm just here for the party," they said. "I think it's terrible

• an appearance at the annual awards bash , but feels in no fit state to• -known families at a fundraising bash on Thursday night for Learning• Who was paying for the bash? The only clue was the name Asprey, • Mail, always hosted the annual bash for the Scottish Labour front-• popular. Their method is to bash sense into criminals with a short, • just cut off people's heads and bash their brains out over the floor,

Page 8: Statistical NLP Winter 2008

Word senses create groupings

• number-crunchers within the Labour party, there now seems little doubt• political tradition and the same party. They are both relatively Anglophilic • he told Tony Blair's modernised party they must not retreat into "warm

• thing. She was talking at a party thrown at Daphne's restaurant in • have turned it into the hot dinner-party topic. The comedy is the• selection for the World Cup party, which will be announced on May 1 • the by-pass there will be a street party. "Then," he says, "we are going • "Oh no, I'm just here for the party," they said. "I think it's terrible • an appearance at the annual awards bash, but feels in no fit state to• -known families at a fundraising bash on Thursday night for Learning• Who was paying for the bash? The only clue was the name Asprey, • Mail, always hosted the annual bash for the Scottish Labour front-

• popular. Their method is to bash sense into criminals with a short, • just cut off people's heads and bash their brains out over the floor,

Page 9: Statistical NLP Winter 2008

What Good are Word Senses?

• Semantics / Text understanding• Many systems• Axioms about TRANSFER apply to (some tokens of) throw• Axioms about BUILDING apply to (some tokens of) bank

• Machine translation• bass has Spanish translation of lubina or bajo depending on sense

• Info retrieval / Question answering / Text categ.• Query or pattern might not match document exactly

• Backoff for just about anything• what word comes next? (speech recognition, language ID, …)

• trigrams are sparse but tri-meanings might not be

• lexicalized probabilistic grammars (coming up later in the course)

• Speaker’s real intention is senses; words are a noisy channel

Page 10: Statistical NLP Winter 2008

Cues to Word Sense

• Adjacent words (or their senses)

• Grammatically related words (subject, object, …)

• Other nearby words• Topic of document• Sense of other tokens of the word in the same

document

Page 11: Statistical NLP Winter 2008

More details about word Senses

• Words have multiple distinct meanings, or senses:• Plant: living plant, manufacturing plant, …• Title: name of a work, ownership document, form of address,

material at the start of a film, …

• Many levels of sense distinctions• Homonymy: totally unrelated meanings (river bank, money bank)• Polysemy: related meanings (star in sky, star on tv)• Systematic polysemy: productive meaning extensions

(organizations to their buildings) or metaphor• Sense distinctions can be extremely subtle (or not)

• Granularity of senses needed depends a lot on the task

Page 12: Statistical NLP Winter 2008

Word Sense Disambiguation

• Example: living plant vs. manufacturing plant

• How do we tell these senses apart?• “context”

• Maybe it’s just text categorization• Each word sense represents a topic• Run the naive-bayes classifier from last class?

• Bag-of-words classification works ok for noun senses• 90% on classic, shockingly easy examples (line, interest, star)• 80% on senseval-1 nouns• 70% on senseval-1 verbs

The manufacturing plant which had previously sustained the town’s economy shut down after an extended labor strike.

Page 13: Statistical NLP Winter 2008

Verb WSD

• Why are verbs harder?• Verbal senses less topical• More sensitive to structure, argument choice

• Verb Example: “Serve”• [function] The tree stump serves as a table• [enable] The scandal served to increase his popularity• [dish] We serve meals for the homeless• [enlist] He served his country• [jail] He served six years for embezzlement• [tennis] It was Agassi's turn to serve• [legal] He was served by the sheriff

Page 14: Statistical NLP Winter 2008

Various Approaches to WSD

• Unsupervised learning• Bootstrapping (Yarowsky 95)• Clustering

• Indirect supervision• From thesauri• From WordNet• From parallel corpora

• Supervised learning• Most systems do some kind of supervised learning• Many competing classification technologies perform about the same (it’s

all about the knowledge sources you tap)• Problem: training data available for only a few words

Page 15: Statistical NLP Winter 2008

Resources

• WordNet• Hand-build (but large) hierarchy of word senses• Basically a hierarchical thesaurus

• SensEval• A WSD competition, of which there have been 3 iterations• Training / test sets for a wide range of words, difficulties, and parts-of-

speech• Bake-off where lots of labs tried lots of competing approaches

• SemCor• A big chunk of the Brown corpus annotated with WordNet senses

• OtherResources• The Open Mind Word Expert• Parallel texts• Flat thesauri

Page 16: Statistical NLP Winter 2008

Knowledge Sources

• So what do we need to model to handle “serve”?• There are distant topical cues

• …. point … court ………………… serve ……… game …

• We could just use a Naïve Bayes classifier (same as texcat!)

∏=i

in cwPcPwwwcP )|()(),,,( 21 K

c

w1 w2 wn. . .

Page 17: Statistical NLP Winter 2008

Weighted Windows with NB

• Distance conditioning• Some words are important only when they are nearby• …. as …. point … court ………………… serve ……… game …• …. ………………………………………… serve as……………..

• Distance weighting• Nearby words should get a larger vote• … court …… serve as……… game …… point

'

1 0 1 '( , ,..., , , , ) ( ) ( | , ( ))k

k k ii k

P c w w w w w P c P w c bin i− − + +=−

= ∏K

'( )

1 0 1 '( , ,..., , , , ) ( ) ( | )k

boost ik k i

i k

P c w w w w w P c P w c− − + +=−

= ∏K

boos

t

relative position i

Page 18: Statistical NLP Winter 2008

Better Features

• There are smarter features:• Argument selectional preference:

• serve NP[meals] vs. serve NP[papers] vs. serve NP[country]• Subcategorization:

• [function] serve PP[as]• [enable] serve VP[to]• [tennis] serve <intransitive>• [food] serve NP {PP[to]}

• Can capture poorly (but robustly) with local windows• … but we can also use a parser and get these features explicitly

• Other constraints (Yarowsky 95)• One-sense-per-discourse (only true for broad topical distinctions)• One-sense-per-collocation (pretty reliable when it kicks in: manufacturing

plant, flowering plant)

Page 19: Statistical NLP Winter 2008

Complex Features with NB?

• Example:

• So we have a decision to make based on a set of cues:• context:jail, context:county, context:feeding, …• local-context:jail, local-context:meals• subcat:NP, direct-object-head:meals

• Not clear how build a generative derivation for these because the features cannot be independent:• Choose topic, then decide on having a transitive usage, then pick

“meals” to be the object’s head, then generate other words?• How about the words that appear in multiple features?• Hard to make this work (though maybe possible)• No real reason to try

Washington County jail served 11,166 meals last month - a figure that translates to feeding some 120 people three times daily for 31 days.

Page 20: Statistical NLP Winter 2008

A Discriminative Approach

• View WSD as a discrimination task (regression, really)

• Have to estimate multinomial (over senses) where there are a huge number of things to condition on• History is too complex to think about this as a smoothing / back-

off problem

• Many feature-based classification techniques out there• We tend to need ones that output distributions over classes

P(sense | context:jail, context:county, context:feeding, … local-context:jail, local-context:meals subcat:NP, direct-object-head:meals, ….)

Page 21: Statistical NLP Winter 2008

Feature Representations

• Features are indicator functions fi which count the occurrences of certain patterns in the input

• We map each input to a vector of feature predicate counts

Washington County jail served 11,166 meals last month - a figure that translates to feeding some 120 people three times daily for 31 days.

context:jail = 1context:county = 1 context:feeding = 1context:game = 0…local-context:jail = 1local-context:meals = 1…subcat:NP = 1subcat:PP = 0…object-head:meals = 1object-head:ball = 0

{ ( )}if dd

Page 22: Statistical NLP Winter 2008

Linear Classifiers

• For a pair (c,d), we take a weighted vote for each class:

• There are many ways to set these weights• Perceptron: find a currently misclassified example, and nudge weights in the

direction of a correct classification• Other discriminative methods usually work in the same way: try out various

weights until you maximize some objective• We’ll look at an elegant probabilistic framework for weight-setting: maximum

entropy

( | ) exp ( ) ( )i ii

vote c d c f dλ= ∑Feature Food Jail Tennis

context:jail -0.5 * 1 +1.2 * 1 -0.8 * 1

subcat:NP +1.0 * 1 +1.0 * 1 -0.3 * 1

object-head:meals +2.0 * 1 -1.5 * 1 -1.5 * 1

object-head:years = 0 -1.8 * 0 +2.1 * 0 -1.1 * 0

TOTAL +3.5 +0.7 -2.6

Page 23: Statistical NLP Winter 2008

Maximum-Entropy Classifiers

• Exponential (log-linear, maxent, logistic, Gibbs) models:• Turn the votes into a probability distribution:

• For any weight vector {λi}, we get a conditional probability model P(c | d,λ).

• We want to choose parameters {λi} that maximize the conditional (log) likelihood of the data:

'

exp ( ') ( )i ic i

c f dλ∑ ∑=),|( λdcP

exp ( ) ( )i ii

c f dλ∑ Makes votes positive.

Normalizes votes.

∑∑∈∈

==),(),(),(),(

log),|(log),|(logDCdcDCdc

dcPDCP λλ

'

exp ( ) ( )i ic i

c f dλ∑ ∑

exp ( ) ( )i ii

c f dλ∑

Page 24: Statistical NLP Winter 2008

Building a Maxent Model

• How to define features:• Features are patterns in the input which we think the weighted vote

should depend on• Usually features added incrementally to target errors• If we’re careful, adding some mediocre features into the mix won’t hurt

(but won’t help either)

• How to learn model weights?• Maxent just one method• Use a numerical optimization package • Given a current weight vector, need to calculate (repeatedly):

• Conditional likelihood of the data• Derivative of that likelihood wrt each feature weight

Page 25: Statistical NLP Winter 2008

The Likelihood Value

• The (log) conditional likelihood is a function of the iid data (C,D) and the parameters λ:

• If there aren’t many values of c, it’s easy to calculate:

• We can separate this into two components:

∑∏∈∈

==),(),(),(),(

),|(log),|(log),|(logDCdcDCdc

dcPdcPDCP λλλ

∑∈

=),(),(

log),|(logDCdc

DCP λ

'

exp ( ) ( )i ic i

c f dλ∑ ∑exp ( ) ( )i i

i

c f dλ∑

( , ) ( , ) '

log exp ( ') ( )i ic d C D c i

c f dλ∈∑ ∑ ∑

( , ) ( , )

log exp ( ) ( )i ic d C D i

c f dλ∈∑ ∑ −=),|(log λDCP

)(λN )(λM=),|(log λDCP −

Page 26: Statistical NLP Winter 2008

The Derivative I: Numerator

( ) ( )

( )

i k i kk i

i

c f d

c

λ

λ

∂=

∑∑

:

( ) ( )

( )k

i i ki

k c c i

c f d

c

λ

λ=

∂=

∑∑

:

( )k

ik c c

f d=

= ∑

log exp ( ) ( )( )

( ) ( )

i k i kk i

i i

c f dN

c c

λλ

λ λ

∂∂

=∂ ∂

∑ ∑

Derivative of the numerator is the empirical count(fi, c)

E.g.: we actually saw the word “dish” with the “food” sense 3 times (maybe twice in one example and once in another).

Page 27: Statistical NLP Winter 2008

The Derivative II: Denominator

'

log exp ( ') ( )( )

( ) ( )

i i kk c i

i i

c f dM

c c

λλ

λ λ

∂∂

=∂ ∂

∑ ∑ ∑

'

''

exp ( ') ( )1

exp ( '') ( ) ( )

i i kc i

k i i k ic i

c f d

c f d c

λ

λ λ

∂=

∑ ∑∑∑ ∑

'''

exp ( ') ( ) ( ') ( )1

exp ( '') ( ) 1 ( )

i i k i i ki i

k ci i k ic i

c f d c f d

c f d c

λ λ

λ λ

∂=

∑ ∑∑ ∑∑ ∑

( | , ) ( )k i kk

P c d f dλ=∑ = predicted count(fi, λ)

Page 28: Statistical NLP Winter 2008

The Derivative III

• The optimum parameters are the ones for which each feature’s predicted expectation equals its empirical expectation. The optimum distribution is:• Always unique (but parameters may not be unique)• Always exists (if features counts are from actual data).

log ( | , )

( )i

P C D

c

λλ

∂=

∂actual count( , )if c ),(countpredicted λif−

d

c P(c | λ)meal, jail, …

jail, term, …

food

prison .8

.4

The λ(prison) weight for the “context-word:jail” feature: actual = 1 empirical = 1.2

Page 29: Statistical NLP Winter 2008

Summary

• We have a function to optimize:

• We know the function’s derivatives:

• Ready to feed it into a numerical optimization package…

• What did any of that have to do with entropy?

∑∈

=),(),(

log),|(logDCdc

DCP λ

'

exp ( ') ( )i ic i

c f dλ∑ ∑exp ( ) ( )i i

i

c f dλ∑

log ( | , ) / ( )iP C D cλ λ∂ ∂ =actual count( , )if c ),(countpredicted λif−

Page 30: Statistical NLP Winter 2008

Smoothing: Issues of Scale

• Lots of features:• NLP maxent models can have over 1M features.• Even storing a single array of parameter values can have a

substantial memory cost.

• Lots of sparsity:• Overfitting very easy – need smoothing!• Many features seen in training will never occur again at test time.

• Optimization problems:• Feature weights can be infinite, and iterative solvers can take a

long time to get to those infinities.

Page 31: Statistical NLP Winter 2008

Smoothing: Issues

• Assume the following empirical distribution:

• Features: {Heads}, {Tails}• We’ll have the following model distribution:

• Really, only one degree of freedom (λ = λH-λT)

Heads Tails

h t

TH

H

HEADS λλ

λ

eee

p+

=TH

T

TAILS λλ

λ

eee

p+

=

0HEADS TTTH

TH

ee

e

eeee

eep

+=

+= −−

λ

λ

λλλλ

λλ

0

0

TAILS ee

ep

+= λ

λ

Page 32: Statistical NLP Winter 2008

Smoothing: Issues

• The data likelihood in this model is:

TAILSHEADS loglog)|,(log ptphthP +=λ

)1(log)()|,(log λλλ ehththP ++−=

Heads Tails

2 2

Heads Tails

3 1

Heads Tails

4 0

λ λ λ

log P log P log P

Page 33: Statistical NLP Winter 2008

Smoothing: Early Stopping

• In the 4/0 case, there were two problems:• The optimal value of λ was , which is a long trip for

an optimization procedure.• The learned distribution is just as spiked as the

empirical one – no smoothing.

• One way to solve both issues is to just stop the optimization early, after a few iterations.• The value of λ will be finite (but presumably big).• The optimization won’t take forever (clearly).• Commonly used in early maxent work.• Disadvantage: you can’t reason about the outcome

of the model independent of the algorithm used to fit it

Heads Tails

4 0

Heads Tails

1 0

Input

Output

λ

Page 34: Statistical NLP Winter 2008

Smoothing: Priors (MAP)

• What if we had a prior expectation that parameter values wouldn’t be very large?

• We could then balance evidence suggesting large parameters (or infinite) against our prior.

• The evidence would never totally defeat the prior, and parameters would be smoothed (and kept finite!).

• We can do this explicitly by changing the optimization objective to maximum posterior likelihood:

• This is alternatively known as penalization in the non-Bayesian statistical literature, and the use of a prior in the Bayesian literature

),|(log)(log)|,(log λλλ DCPPDCP +=

Posterior Prior Evidence

Page 35: Statistical NLP Winter 2008

Smoothing: Priors

• Gaussian, or quadratic, priors:• Intuition: parameters shouldn’t be large.• Formalization: prior expectation that each

parameter will be distributed according to a gaussian with mean and variance 2.

• Penalizes parameters for drifting to far from their mean prior value (usually =0).

• 22=1 works surprisingly well (better to set using held-out data, though)

⎟⎟⎠

⎞⎜⎜⎝

⎛ −−=

2

2

2

)(exp

2

1)(

i

ii

i

iPσ

μλ

πσλ

22

=1

22 = 10

22 =

Page 36: Statistical NLP Winter 2008

Smoothing: Priors

• If we use gaussian priors:• Trade off some expectation-matching for smaller parameters.• When multiple features can be recruited to explain a data point, the more

common ones generally receive more weight.• Accuracy generally goes up!

• Change the objective:

• Change the derivative:

)(log λP−),|(log)|,(log λλ DCPDCP =

∑∈

=),(),(

),|()|,(logDCdc

dcPDCP λλ ki i

ii +−

−∑ 2

2

2)(

λ

),(predicted),(actual/)|,(log λλλ iii fCfDCP −=∂∂ 2/)( λ ii −−

22

=1

22 = 10

22 =