graphical models - iit bombay

28
Graphical Models Vikas Kedia August 18, 2005

Upload: others

Post on 18-Dec-2021

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Graphical Models - IIT Bombay

Graphical Models

Vikas Kedia

August 18, 2005

Page 2: Graphical Models - IIT Bombay

Outline

I What are Graphical Models?

I Inferencing Algorithms

I Conditional Random Fields

I Application Example

Page 3: Graphical Models - IIT Bombay

Graphical Models

Page 4: Graphical Models - IIT Bombay

Introduction

I Traditional Classification: each instance is labeled individuallyI In many tasks this model is inadequate

I POS tagging: tags of neighboring words important cluesI Web Page Classification: classes of linked pages useful

I Collective Classification: classes/labels of all the instancesinferred collectively

I Graphical Models a formalism for collective classification

Page 5: Graphical Models - IIT Bombay

Graphical Models

I Relations represented as a graph

Vertices Labels/Observations eg: features of web page(observed), class of page (hidden)

Edges Dependencies eg: edge between linked webpages

Page 6: Graphical Models - IIT Bombay

Markov Property

I Probability distribution defined over values of all nodes ingraph

I Local Markov Property : Given the values of its neighbours,value of the node is conditionally independent of values ofother nodes

p(Yv |Yw ,w 6= v) = p(Yv |Yw ,w ∼ v)

I Global Markov Property

p(YV ) =

∏C φC (YVC

)∑YV

∏C φC (YVC

)

φC Potential functions over labelsof nodes in clique C

Page 7: Graphical Models - IIT Bombay

Example Graph

p(XV ) =φ(x1, x2)φ(x1, x3)φ(x2, x4)φ(x3, x5)φ(x2, x5, x6)

Z

Page 8: Graphical Models - IIT Bombay

Inferencing Algorithm

Page 9: Graphical Models - IIT Bombay

Two Inferencing Tasks

I Finding the most likely value of the variables

maxx1,x2,x3,x4,x5,x6

p(XV )

I Finding the marginal probabilities

p(x1) =∑

x2,x3,x4,x5,x6

p(XV )

Page 10: Graphical Models - IIT Bombay

Naive Approach

I Enumerate all possible combinations of values to all thevariables

I Exponential number of possibilities, r6 where r is thecardinality of each variable

I Clearly intractable for large graphs

I Insight: multiplication distributes over both max and sumoperator

Page 11: Graphical Models - IIT Bombay

Example

p(x1) =

Px2,x3,x4,x5,x6

φ(x1, x2)φ(x1, x3)φ(x2, x4)φ(x3, x5)φ(x2, x5, x6)

Z

=

Px2

φ(x1, x2)P

x3φ(x1, x3)

Px4

φ(x2, x4)P

x5φ(x3, x5)

Px6

φ(x2, x5, x6)

Z

=

Px2

φ(x1, x2)P

x3φ(x1, x3)

Px4

φ(x2, x4)P

x5φ(x3, x5)m6(x2, x5)

Z

=

Px2

φ(x1, x2)P

x3φ(x1, x3)m5(x2, x3)

Px4

φ(x2, x4)

Z

=

Px2

φ(x1, x2)m4(x2)P

x3φ(x1, x3)m5(x2, x3)

Z

=

Px2

φ(x1, x2)m4(x2)m3(x1, x2)

Z

=m2(x1)

Z

Page 12: Graphical Models - IIT Bombay

Some Observations

I No more than 3 variables occur together in any summand

I Complexity is therefore r3

I The order in which variables were chosen is elimination order

I Complexity would depend on the elimination order

Page 13: Graphical Models - IIT Bombay

Graph Theory Problem

I Variable that is summed over can be removed from the graph

I Intermediate function created is function of all the variablesconnected to the variable being summed over

I Therefore create a clique of all those variables

I Repeat till all the nodes are removed

I Largest clique created corresponds to the complexity

Page 14: Graphical Models - IIT Bombay
Page 15: Graphical Models - IIT Bombay

Treewidth

I Different elimination order give rise to different max clique size

I Treewidth is the minimum over all such max clique size

I To minimise complexity chose elimination order which givesrise to treewidth

I Unfortunately this problem is NP-Hard

Page 16: Graphical Models - IIT Bombay

Elimination Order in Specific Cases

I For specific types of graphs optimal elimination order is easyto see

I Example: for chains just keep on removing vertices from oneend

I Gives rise to the Viterbi Algorithm

I Columns of the table correspond to the intermediate functions

Page 17: Graphical Models - IIT Bombay

Elimination Order for Trees

I Eliminate all children of a node before a node is eliminated

Page 18: Graphical Models - IIT Bombay

Conditional Random Fields

Page 19: Graphical Models - IIT Bombay

Limitation of HMM

I In Hidden Markov Models we assume that generation of atoken depends only on the current state

I This restriction might be too limiting, we might want toinclude arbitrary features of data

I For example: we might want to look at some tokens on bothsides of the current token

I Including such featuers in HMM increase the complexity ofinferencing

Page 20: Graphical Models - IIT Bombay

Conditional Random Fields

I CRF introduced to overcome this limitation

I Nodes in the graph correspond only to labels

I Model globally conditioned on the observation data

I Potential functions therefore can be over entire data sequence

p(YV |X ) =

∏C φC (YVC

,X )∑YV

∏C φC (YVC

,X )

Page 21: Graphical Models - IIT Bombay

Linear CRF

I Potential functions are assumed to be exponential

I Parameters are tied across cliques of same typeI eg: For a chain CRF

p(y |x) =exp(

∑e∈E ,k λk fk(ye , x) +

∑v∈V ,k µkgk(yv , x))

Z (x)

I λks and µks common for all edge and singleton cliques resp.

I Intuitively λks are similar to state transition probabilities and µk togeneration probabilities

Page 22: Graphical Models - IIT Bombay

Estimation and Inference

I λks and µks need to be learned from labelled training data

I Parameters are estimated using Maximum Likelihoodhypothesis

I Log likelihood of data is maximised using numerical methods

L =∑

j

(∑

e∈E ,k

λk fk(ye , x) +∑

v∈V ,k

µkgk(yv , x)− log Z (x))

I Gradient of log likelihood involves expected counts of feature values,calculated using dynamic programming

Page 23: Graphical Models - IIT Bombay

Graphical Models in Reconciliation

Page 24: Graphical Models - IIT Bombay

Reconciliation

I Reconciliation means finding duplicate records

I Traditionally based on syntactic similarity between pair ofrecords

I Information flows from similarity of attributes to similarity ofrecords

I But similarity between records also implies that the attributesare same

I eg similar citations means that the journal names in two alsorefer to same journal

I This bi-directional flow can be used for collective reconciliation

Page 25: Graphical Models - IIT Bombay

Example

Record Title Author Venueb1 Record Linkage using CRFs Linda Stewart KDD-2003b2 Record Linkage using CRFs Linda Stewart 9th SIGKDDb3 Learning Boolean Formulas Bill Johnson KDD-2003b4 Learning of Boolean Expressions William Johnson 9th SIGKDD

Table: Duplicate Citations[3]

I b1=b2 means that KDD-2003 is same as 9th SIGKDD

I This will help in inferring similarity between b3 and b4

Page 26: Graphical Models - IIT Bombay

Collective Model

I Binary nodes for each pair of records

I Nodes for all possible pairs of values for each attribute calledevidence nodes

I Value of evidence nodes is similarity measure and is observed

I Binary information nodes corresponding to each evidence node

I Information node represent whether the pair of attributevalues are same

Page 27: Graphical Models - IIT Bombay

Cliques in the Model

I Singleton cliques for information and record nodes

I Edges connecting record nodes to the correspondinginformation nodes

I Edge connecting information nodes to the correspondingevidence nodes

I Inferencing done using graph partitioning

Page 28: Graphical Models - IIT Bombay

References I

M. I. Jordan.Graphical models.In Statistical Science (Special Issue on Bayesian Statistics),pages 140–155, 2004.

John Lafferty, Andrew McCallum, and Fernando Pereira.Conditional random fields: Probabilistic models for segmentingand labeling sequence data.In Proc. 18th International Conf. on Machine Learning, pages282–289. Morgan Kaufmann, San Francisco, CA, 2001.

Parag and P. Domingos.Multi-relational record linkage.In Proceedings of the KDD-2004 Workshop onMulti-Relational Data Mining, pages 31–48, 2004.