Similarity and clustering Motivation • Problem: Query word could be ambiguous: – Eg: Query“Star” retrieves documents about astronomy, plants, animals etc. – Solution: Visualisation •

Download Report

Transcript Similarity and clustering Motivation • Problem: Query word could be ambiguous: – Eg: Query“Star” retrieves documents about astronomy, plants, animals etc. – Solution: Visualisation •

Similarity and clustering

Motivation

• Problem: – Query word could be ambiguous: Eg: Query “Star” retrieves documents about astronomy, plants, animals etc.

– Solution: Visualisation • Clustering document responses to queries along lines of different topics.

• Problem 2: – Solution: Manual construction of topic hierarchies and taxonomies • Preliminary clustering of large samples of web documents.

• Problem 3: Speeding up similarity search Clustering – Solution: • Restrict the search for documents similar to a query to most representative cluster(s).

2

Example

Scatter/Gather, a text clustering system, can separate salient topics in the response to keyword queries. (Image courtesy of Hearst)

Clustering 3

• • • •

Clustering

Task : Evolve measures of similarity to cluster a collection of documents/terms into groups within which similarity within a cluster is larger than across clusters.

Cluster Hypothesis: G iven a `suitable‘ clustering of a collection, if the user is interested in document/term

d/t

, he is likely to be interested in other members of the cluster to which

d/t

belongs.

Similarity measures – Represent documents by TFIDF vectors – Distance between document vectors – Cosine of angle between document vectors Issues – Large number of noisy dimensions 4

• •

Top-down clustering k

-Means: Repeat… – Choose k arbitrary ‘centroids’ – Assign each document to nearest centroid – Recompute centroids Expectation maximization (EM): – Pick k arbitrary ‘distributions’ – Repeat: • Find probability that document from distribution f for all d and d f is generated • Estimate distribution parameters from weighted contribution of documents Clustering 5

• • •

Choosing `k’

Mostly problem driven Could be ‘data driven’ only when either – Data is not sparse – Measurement dimensions are not too noisy Interactive – Data analyst interprets results of structure discovery Clustering 6

• •

Choosing ‘k’ : Approaches

Hypothesis testing: – Null Hypothesis (H o ): Underlying density is a mixture of ‘k’ distributions – Require regularity conditions on the mixture likelihood function (Smith’85) Bayesian Estimation – Estimate posterior distribution on k, given data and prior on k.

– Difficulty: Computational complexity of integration – Autoclass algorithm of (Cheeseman’98) uses approximations – (Diebolt’94) suggests sampling techniques Clustering 7

• •

Choosing ‘k’ : Approaches

Penalised Likelihood – To account for the fact that L k (D) is a non decreasing function of k.

– Penalise the number of parameters – Examples : Bayesian Information Criterion (BIC), Minimum Description Length(MDL), MML.

– Assumption: Penalised criteria are asymptotically optimal (Titterington 1985) Cross Validation Likelihood – Find ML estimate on part of training data – Choose k that maximises average of the M cross validated average likelihoods on held-out data D test – Cross Validation techniques: Monte Carlo Cross Validation (MCCV), v-fold cross validation (vCV) Clustering 8

Similarity and clustering

Motivation

• Problem: – Query word could be ambiguous: Eg: Query “Star” retrieves documents about astronomy, plants, animals etc.

– Solution: Visualisation • Clustering document responses to queries along lines of different topics.

• Problem 2: – Solution: Manual construction of topic hierarchies and taxonomies • Preliminary clustering of large samples of web documents.

• Problem 3: Speeding up similarity search Clustering – Solution: • Restrict the search for documents similar to a query to most representative cluster(s).

10

Example

Scatter/Gather, a text clustering system, can separate salient topics in the response to keyword queries. (Image courtesy of Hearst)

Clustering 11

• • •

Clustering

Task : which Evolve measures of similarity to similarity cluster a collection of documents/terms into groups within within a cluster is larger than across clusters.

Cluster Hypothesis: Given a `suitable‘ clustering of a collection, if the user is interested in document/term d/t , he is likely to be interested in other members of the cluster to which d/t belongs.

Collaborative filtering:

Clustering of two/more objects which have bipartite relationship Clustering 12

• • • • Clustering (contd)

Two important paradigms:

– Bottom-up agglomerative clustering – Top-down partitioning

Visualisation techniques:

Embedding of corpus in a low-dimensional space

Characterising the entities:

– Internally : Vector space model, probabilistic models – Externally: Measure of similarity/dissimilarity between pairs

Learning:

Supplement stock algorithms with experience with data Clustering 13

Clustering: Parameters • • • • Similarity measure: (eg: cosine similarity)  (

d

1 ,

d

2 ) Distance measure:  (

d

1 ,

d

2 ) (eg: eucledian distance) Number “k” of clusters Issues – Large number of noisy dimensions – Notion of noise is application dependent Clustering 14

• • •

Clustering: Formal specification

Partitioning Approaches – Bottom-up clustering – Top-down clustering Geometric Embedding Approaches – Self-organization map – Multidimensional scaling – Latent semantic indexing Generative models and probabilistic approaches – Single topic per document – Documents correspond to mixtures of multiple topics Clustering 15

• • • • •

Partitioning Approaches

Partition document collection into k clusters Choices: {

D

1 ,

D

2 .....

D k

} – Minimize intra-cluster distance   

i d

1 ,

d

2 

D i

– Maximize intra-cluster semblance (   

i d

1

d

1 ,

d

2 ,

d

2 

D i

) (

d

1 ,

d

2 ) If cluster representations are available – Minimize – Maximize   

i d D i

 

i d

D i

  (

d

(

d

, ,

D i D i

) )

D i

Soft clustering – d assigned to with `confidence’

z

clustering

d

,

i

maximize   

i d D i D z d

,

i i

 (

d

,

D i

)   

i d D i

Two ways to get partitions and

z d

,

i

 (

d z

,

d D i

) ,

i

bottom-up top-down clustering Clustering 16

Bottom-up clustering(HAC)

• • • • • Initially G is a collection of singleton groups, each with one document

d

Repeat – Find  ,  in G s (  ) – Merge group  with max similarity measure, with group  For each  keep track of best  Use above info to plot the hierarchical merging process (DENDOGRAM) To get desired number of clusters: cut across any level of the dendogram Clustering 17

Dendogram

A dendogram presents the progressive, hierarchy-forming merging process pictorially.

Clustering 18

• •

Similarity measure

Typically

s

(  ) decreases with increasing number of merges Self-Similarity – Average pair wise similarity between documents in 

s

(  )  1 

C

2

d

1 ,

d

   2

s

(

d

1 ,

d

2 ) –

s

(

d

1 ,

d

2 ) = inter-document similarity measure (say cosine of tfidf vectors) – Other criteria: Maximium/Minimum pair wise similarity between documents in the clusters Clustering 19

Un-normalized group profile:

Computation

 

d

 

p

  Can show:

s s

       (   ),   (   ) 1    ˆ

p

   ,

 

  (  

    ),    (       )     1      ,  2    ,

   

O ( n 2 log n ) algorithm with n 2 space

Clustering 20

Clustering

Similarity

 , 

s

(  ,  )   inner product

g

(

c

(  )),

g

(

c

(  ))

g

(

c

(  )) 

g

(

c

(  )) Normalized document profile:

p

(  ) 

g

(

c

(  ))

g

(

c

(  )) Profile for document group  :

p

(  )         

p

(  )

p

(  ) 21

• •

Switch to top-down

Bottom-up – Requires quadratic time and space Top-down or – Internal representation for documents as well as clusters – Partition documents into `k’ clusters – 2 variants • “Hard” (0/1) assignment of documents to clusters • “soft” : documents belong to clusters, with fractional scores – Termination • when assignment of documents to clusters ceases to change much OR • When cluster centroids move negligibly over successive iterations move-to-nearest Clustering 22

• •

Top-down clustering Hard k

-Means: Repeat… – Choose k arbitrary ‘centroids’ – Assign each document to nearest centroid – Recompute centroids

Soft

k-Means : – Don’t break close ties between document assignments to clusters – Don’t make documents contribute to a single cluster which wins narrowly 

d d

 

c

c

   

c

  exp( exp(   |

d

|

d

  

c

  | 2 ) | 2 )   

c

Clustering 23

• • • • •

Seeding `k’ clusters

  Run bottom-up group average clustering algorithm to reduce to groups or clusters :

O

(

knlogn k

) time Iterate assign-to-nearest

O

(1) times – Move each document to nearest cluster – Recompute cluster centroids Total time taken is

O

(

kn

) Non-deterministic behavior Clustering 24

• • •

Choosing `k’

Mostly problem driven Could be ‘data driven’ only when either – Data is not sparse – Measurement dimensions are not too noisy Interactive – Data analyst interprets results of structure discovery Clustering 25

• •

Choosing ‘k’ : Approaches

Hypothesis testing: – Null Hypothesis (H o ): Underlying density is a mixture of ‘k’ distributions – Require regularity conditions on the mixture likelihood function (Smith’85) Bayesian Estimation – Estimate posterior distribution on k, given data and prior on k.

– Difficulty: Computational complexity of integration – Autoclass algorithm of (Cheeseman’98) uses approximations – (Diebolt’94) suggests sampling techniques Clustering 26

• •

Choosing ‘k’ : Approaches

Penalised Likelihood – To account for the fact that L k (D) is a non decreasing function of k.

– Penalise the number of parameters – Examples : Bayesian Information Criterion (BIC), Minimum Description Length(MDL), MML.

– Assumption: Penalised criteria are asymptotically optimal (Titterington 1985) Cross Validation Likelihood – Find ML estimate on part of training data – Choose k that maximises average of the M cross validated average likelihoods on held-out data D test – Cross Validation techniques: Monte Carlo Cross Validation (MCCV), v-fold cross validation (vCV) Clustering 27

• • • • •

Visualisation techniques

Goal: Embedding of corpus in a low dimensional space Hierarchical Agglomerative Clustering (HAC) – lends itself easily to visualisaton Self-Organization map (SOM) – A close cousin of k-means Multidimensional scaling (MDS) – minimize the distortion of interpoint distances in the low-dimensional embedding as compared to the dissimilarity given in the input data.

Latent Semantic Indexing (LSI) – Linear transformations to reduce number of dimensions Clustering 28

Self-Organization Map (SOM)

• • • • • Like soft k-means – Determine association between clusters and documents iteratively refine 

c

 Unlike k-means – Embed the clusters in a low-dimensional space right from the beginning – Large number of clusters can be initialised even if eventually many are to remain devoid of documents Each cluster can be a slot in a square/hexagonal grid. The grid structure defines the neighborhood each cluster c N(c) for Also involves a proximity function between clusters and

c h

(

c

,  ) Clustering 29

SOM : Update Rule

Like Neural network – Data item neurons

N d

d (

c d

activates neuron (closest cluster) as well as the neighborhood ) – Eg Gaussian neighborhood function d (

t h

(

c

1 )    (

t

) || (

t

c

)

h

 , 

c

d

|| 2 )(

d

)  )  2 (

t

) learning rate parameter  (

t

Clustering 30

SOM : Example I

SOM computed from over a million documents taken from 80 Usenet newsgroups. Light areas have a high density of documents.

Clustering 31

SOM: Example II

Another example of SOM at work: the sites listed in the Open Directory have beenorganized within a map of Antarctica at http://antarcti.ca/.

Clustering 32

• • • •

Multidimensional Scaling(MDS)

Goal – “Distance preserving” low dimensional embedding of documents Symmetric inter-document distances

d ij

– Given apriori or computed from internal representation Coarse-grained user feedback

d ij

i and j .

– With increasing feedback, prior distances are overridden Objective : Minimize the stress of embedding

stress

i

, 

j

(

d ij

^ 

i

, 

j d ij

2

d ij

) 2 Clustering 33

• • •

MDS: issues

Stress not easy to optimize Iterative hill climbing 1. Points (documents) assigned random coordinates by external heuristic 2. Points moved by small distance in direction of locally decreasing stress For

n

documents

O

(

n O

(

n

) Clustering 34

• • • •

Fast Map [Faloutsos ’95]

No internal representation of documents available Goal – find a projection from an ‘n’ dimensional space to a space with a smaller number ` k‘ ’ of dimensions.

Iterative projection of documents along lines of maximum spread Each 1D projection preserves distance information Clustering 35

• • •

Best line

Pivots for a line: two points ( that determine it

a

and

b

) Avoid exhaustive checking by picking pivots that are far apart First coordinates of point on “best line” (

a

,

b

)

x

1

x

1  2

d a

,

x

 2

d a

,

b

d b

, 2

x

2

d a

,

b

Clustering 36

Iterative projection For i = 1 to k

1.Find a next (i th ) “best” line  A “best” line is one which gives maximum variance of the point-set in the direction of the line 2.Project points on the line 3.Project points on the “hyperspace” orthogonal to the above line Clustering 37

Projection

Purpose – To correct inter-point distances

x

,

y x

1

y

1 accounted for by the first pivot line.

d x

' ' ,

y

' 

d

2

x

,

y

 (

x

1 

y

1 ) 2 • • Project recursively upto 1-D space Time: O(nk) time Clustering 38

• • •

Issues

Detecting noise dimensions – Bottom-up dimension composition too slow – Definition of noise depends on application Running time – Distance computation dominates – Random projections – Sublinear time w/o losing small clusters Integrating semi-structured information – Hyperlinks, tags embed similarity clues – A link is worth a  ?

 words Clustering 39

• Expectation maximization (EM): – Pick k arbitrary ‘distributions’ – Repeat: • Find probability that document d from distribution f for all d and f is generated • Estimate distribution parameters from weighted contribution of documents Clustering 40

• • • • • •

Extended similarity

Where can I fix my scooter ?

A great garage to repair 2-wheeler is at … your auto and car co-occur often Documents having related words are related Useful for search and clustering Two basic approaches – Hand-made thesaurus (WordNet) – Co-occurrence and associations … auto …car … auto …car … car … auto car  auto … auto …  … car … Clustering 41

Clustering

Latent semantic indexing

Documents A d t Term SVD k car U auto Document D r k-dim vector V d 42

• • •

Collaborative recommendation

People=record, movies=features People and features to be clustered – Mutual reinforcement of similarity Need advanced models From

Clustering methods in collaborative filtering,

by Ungar and Foster Clustering 43

• • • • •

A model for collaboration

People and movies belong to unknown classes P k P l = probability a random person is in class = probability a random movie is in class P kl = probability of a class class l movie k person liking a l k Gibbs sampling: iterate – Pick a person or movie at random and assign to a class with probability proportional to P k or P l – Estimate new parameters Clustering 44

Aspect Model

• • • • • Metric data

vs

Dyadic data

vs

Proximity data

vs

Ranked preference data.

Dyadic data : domain with two finite sets of objects Observations : Of dyads

X

and

Y

Unsupervised learning from dyadic data Two sets of objects

X

 {

x

1 ....

x i

,

x n

},

Y

 {

y

1 ....

y i

,

y n

} Clustering 45

Aspect Model (contd)

Two main tasks – Probabilistic modeling: over

X

Y

– structure discovery: • identifying clusters and data hierarchies.

Clustering 46

Aspect Model

Statistical models – Empirical co-occurrence frequencies • Sufficient statistics – Data spareseness: • Empirical frequencies either 0 or significantly corrupted by sampling noise – Solution • Smoothing – Back-of method [Katz’87] – Model interpolation with held-out data [JM’80, Jel’85] – Similarity-based smoothing techniques [ES’92] • Model-based statistical approach: a principled approach to deal with data sparseness Clustering 47

• •

Aspect Model

Model-based statistical approach: a principled approach to deal with data sparseness – Finite Mixture Models [TSM’85] – Latent class [And’97] – Specification of a joint probability distribution for latent and observable variables [Hoffmann’98] Unifies – statistical modeling • Probabilistic modeling by marginalization – structure detection (exploratory data analysis) • Posterior probabilities by baye’s rule on latent space of structures Clustering 48

Aspect Model

S

 (

x n

,

y n

) 1 

n

N

: Realisation of an underlying sequence of random variables

S

 (

X n

,

Y n

) 1 

n

N

: • • 2 assumptions – All co-occurrences in sample S are iid –

X n

,

Y n

are independent given

A n

P(c) are the mixture components Clustering 49

Increasing Degree of Restriction On Latent space

Aspect Model: Latent classes

A n A

(

X n

,

Y n

) 1 

n

N

 {

a

1 ,....

a K

} {

C C

(

X n

),

Y n

} 1 

n

N

 {

c

1 ,...

c K

} Clustering {

C C

(

X n

),

D

(

Y n

)} 1 

n

N

 {

c

1 ,...

c K

}

D

 {

d

1 ,..

d L

} {

C C

(

X n

),

Y n

} 1 

n

N

 {

c

1 ,...

c K

} 50

Aspect Model

Symmetric

P

(

S

,

a

) 

N

n

 1

P

(

x n

,

y n

,

a n

) 

n N

  1

P

(

a n

)

P

(

x n

|

a n

)

P

(

y n

|

a n

) Asymmetric

P

(

S

)  

x

X

y

Y P

(

x

,

y

)

n

(

x

,

y

)  

x

X



y a A

[

P

(

a

)

P

(

x

|

a

)

P

(

y

|

a

)]

n

(

x

,

y

)

P

(

S

,

a

) 

N

n

 1

P

(

x n

,

y n

,

a n

) 

N

n

 1

P

(

a n

)

P

(

x n

|

a n

)

P

(

y n

|

a n

)

P

(

S

)  

x

X

y

Y P

(

x

,

y

)

n

(

x

,

y

)  

x

X P

(

x

) 

y

Y a

A

[

P

(

a

|

x

)

P

(

y

|

a

)]

n

(

x

,

y

) Clustering 51

Clustering

vs

Aspect

Clustering model – constrained aspect model

P

(

a

|

x

,

c

) 

P

(

A n

a

|

X n

x

,

C

(

x

) 

c

}  

ac

• For flat:

c k

a k

 

ac

Clustering • For hierarchical

a k

c k

– Notation • P(.) : are the parameters • P{.}: are posteriors  

ac

.

P

(

a

|

x

,

c

) – Group structure on object spaces as against partition the observations 52

Hierarchical Clustering model

One-sided clustering Hierarchical clustering 

P

(

S

)  

x

X

y

Y P

(

x

, 

x

X P

(

x

) 

y

Y a c C y

)

n

(

x

,

y

) 

P

(

c

) [

P

(

a

| 

x

X P

(

x

) 

y

Y a

A

[

P

(

a x

,

c

)

P

(

y

|

a

)]

n

(

x

,

y

)  |

x

)

P

(

y

c

C P

( |

a

)]

n

(

x

,

y

)

c

) 

x

X

[

P

(

x

) ]

n

(

x

) 

y

Y

[

P

(

y

|

a

)]

n

(

x

,

y

) Clustering 

P

(

S

)  

x

X P

( 

x

X

y

Y P

(

x

,

x

) 

y

Y a c C y

)

n

(

x

,

y

) 

P

(

c

) [

P

(

a

| 

x

X P

(

x

) 

y

Y a

A

[

P

(

a x

,

c

)

P

(

y

|

a

)]

n

(

x

,

y

) |

x

)

P

(

y

|

a

)]

n

(

x

,

y

) 53

Comparison of E’s

•Aspect model

P

{

A n

a

|

X n

 •One-sided aspect model

P

{

C

(

x

)

x

,

Y

c n

| 

S x

•Hierarchical aspect model

y

;  ,  } }  

c

' 

C a

'  

A P P

( (

a

)

a

'

P

( )

P x

(

P

(

c

)

[

P P

(

c

'

y

Y

)

y

Y

(

x

| [

P

|

y a

) ( |

a

'

y P c

| ) (

P

)]

n c y

' ( | ( )]

x y

,

n a

)

y

( | )

x a

' ,

y

) )

P

{

C

(

x

) 

c

|

S

,  } 

P c

 ' 

C

(

c

)

P



y

Y

(

c

'

a

A

) 

y

Y

[

P

( [

P

(

y y

| |

a

)

a

)

P

(

a P

(

a

| |

x

,

x

,

c

'

c

)]

n

(

x

,

y

) )]

n

(

x

,

y

)

P

{

A n

a

|

X n

x

,

Y n

y

,

C

(

x

) 

c

;  } 

a

'  

A P

(

a P

(

a

| |'

x

,

c

)

P

(

y x

,

c

)

P

(

y

| |

a

)

a

' ) 54 Clustering

• 5.

Tempered EM(TEM)

1.

Additively (on the log scale) discount the likelihood part in Baye’s formula: 2.

3.

Set and perform EM until the performance on held--out data deteriorates (early stopping).   1 As long as the performance on held-out data improves continue TEM iterations at this value of  4.

Stop on i.e., stop when decreasing does not yield further improvements, otherwise goto step (2)  Perform some final iterations using both, training and heldout data.

P

{

A n

a

|

X n

x

,

Y n

y

;  } 

a

'  

A P

(

P

(

a

)[

a

' )[

P

(

P

(

x x

| |

a

)

P

(

y a

' )

P

( |

y a

)]  |

a

' )]  Clustering 55

1.

P

(

x

|

a

)  Aspect 

n

:

x n

x P

(

a N

n

 1

P

(

a

| |

x n x n

, ,

y y n n

;  ;  ' ' ) )  2.

Assymetric

P

(

x

) 

n

(

x

)

N

y n

(

x

,

y

)

P

(

a

|

x

 ,'

y n

(

x

' ,

y

)

P

(

a

|

x

,

x

' ,

M-Steps

y

;  ' )

y

;  ' )

P

(

y P

(

a

| |

a

)

x

)  

n

:

y

n

y n N

  1

n

:

x

n

x N

n

 1

P

(

a P

(

a P

(

a P

(

a

| | | |

x n x n x n

,

x n

, , ,

y n y n

;  ; 

y n

;  ' ) ' ) ' )

y n

;  ' )   

x n

(

x

,

y

)

P

(

a

|

x

, 

y

' 

y n

(

n

(

x

,

x

,

y y

' )

P

(

a

)

P

(

a

| |

x

 ,'

y n

(

x

' ,

y

)

P

(

a

|

x

,

y

;  ' )

x

,

y

' ;  ' )

x

,

y

;  ' )

x

' ,

y

;  ' ) 3.

Hierarchical x-clustering

P

(

x

) 

n

(

x

)

N P

(

y

|

a

) 

n

: 

y n

y P

{

a

|

x n

,

y n

;  ' } 

n N

  1

P

{

a

|

x n

,

y n

;  ' } 

x n

(

x

,

y

)

P

{

a

|

x

,

y

;  ' }

x

, 

y

'

n

(

x

,

y

' )

P

{

a

|

x

,

y

' ;  ' } 4.

One-sided x-clustering

P

(

x

) 

n

(

x

)

N P

(

y

|

c

)  

x

x n

(

n

(

x

,

y

)

P

{

C x

)

P

{

C

( (

x

)

x

)  

c c

| |

S x

; 

S x

;  ' } ' } Clustering 56

Example Model

[Hofmann and Popat CIKM 2001] Hierarchy of document categories Clustering 57

Clustering

Example Application

58

• • •

Topic Hierarchies

To overcome sparseness problem in topic hierarchies with large number of classes Sparseness Problem: Small number of positive examples Topic hierarchies to reduce variance in parameter estimation  Automatically differentiate     Make use of term distributions estimated for more general, coarser text aspects to provide better, smoothed estimates of class conditional term distributions Convex combination of term distributions in a Hierarchical Mixture Model

P

(

w

|

c

) 

P

(

a

|

c

)

P

(

w

|

a

)

a

 

c

c.

 refers to all inner nodes a above the terminal class node Clustering 59

Topic Hierarchies (Hierarchical X-clustering)

P

(

y

| X = document, Y = word

a

) 

n

:

y

n

y N

n

 1

P

{

a P

{

a

| |

x n

,

x n

,

y n

;  ' }

y n

;  ' }  

x

, 

y

'

x n

(

x

,

n

(

x

,

y

)

P

{

a y

' )

P

{

a

|

x

,

y

;  ' } |

x

,

y

' ;  ' } 

c

(

x

 ) 

a n

(

c

(

x

),

y

)

P

{

a

|

c

(

x

),

y

;  ' }

c

(

x

 ) 

a

,

y

'

n

(

c

(

x

),

y

' )

P

{

a

|

c

(

x

),

y

' ;  ' }

P

{

a

|

x

,

y

,

c

(

x

);  } 

P

{

a

|

y

,

c

(

x

);  } 

a

 ' 

P A

(

a P

(

a

| |'

x

,

c

)

P

(

y x

,

c

)

P

(

y

|

a

) |

a

' ) 

P

(

a a

 ' 

c P

(

a

|' |

c

)

P

(

y

|

a

)

c

(

x

))

P

(

y

|

a

' )

P

{

a

|

c

(

x

);  }  

y n

(

y

,

c

)

P

(

a a

 ' 

c P

(

a

|' |

y

,

c

(

x

))

y

,

c

(

x

))

P

(

x

) 

n

(

x

)

N P

{

C

(

x

) 

c

|

S

,  } 

P

(

c

 ' 

C c

) 

y

Y P

(

c

'

a

A

) 

y

Y

[

P

( [

P

(

y y

| |

a

)

P

(

a a

)

P

(

a

| |

x

,

x

,

c

(

x

))]

n

(

x

,

y

)

c

' (

x

))]

n

(

x

,

y

) 

c P

(

c

)  ' 

C



y

Y a

P

(

c

' )

c

y

Y

[ [

P

(

P

(

y y

| |

a

)

a

)

P

(

P

(

a a

| |

c

'

c

( (

x

))]

n x

))]

n

( (

x x

, ,

y y

) ) Clustering 60

Document Classification Exercise

• Modification of Naïve Bayes

P

(

w

|

c

) 

a

 

c P

(

a

|

c

)

P

(

w

|

a

)

P

(

c

|

x

)  

c

'

P

(

c

)  

x P

(

c

'

y i

)  

x y i P

(

P

(

y i y i

|

c

) |

c

' ) Clustering 61

• •

Mixture vs Shrinkage

Shrinkage [McCallum Rosenfeld AAAI’98] nodes in the hierarchy represent coarser views of the data which are obtained by simple pooling scheme of term counts : Interior Mixture : Interior nodes represent abstraction levels with their corresponding specific vocabulary – Predefined hierarchy [Hofmann and Popat CIKM 2001] – Creation of hierarchical model from [Hofmann IJCAI’99] unlabeled data Clustering 62

.

Mixture Density Networks(MDN)

[Bishop CM ’94 Mixture Density Networks] • • • broad and flexible class of distributions that are capable of modeling completely general continuous distributions superimpose simple component densities with well known properties to generate or approximate more complex distributions Two modules: – Mixture models: Output has a distribution given as mixture of distributions – Neural Network: Outputs determine parameters of the mixture model Clustering 63

MDN: Example

Clustering A conditional mixture density network with Gaussian component densities 64

• •

MDN

Parameter Estimation : – Using Generalized EM (GEM) algo to speed up.

Inference – Even for a linear mixture, closed form solution not possible – Use of Monte Carlo Simulations substitute as a Clustering 65

• • • •

Document model

Vocabulary represented by

f

(

w i

,  )

V

, term

c

(

)

w

occurs in document  i  , document 

f

(

w i

,

) 

w i

V

is the number of times 

w

i Most

f

’s are zeroes for a single document Monotone component-wise damping function

g

such as log or square-root

g

(

c

(

))  

g

(

f

(

w i

,

)) 

w i

V

Clustering 66