ACM SIGIR 2005 Tutorial
Download
Report
Transcript ACM SIGIR 2005 Tutorial
Probabilistic Topic Models
ChengXiang Zhai
Department of Computer Science
Graduate School of Library & Information Science
Institute for Genomic Biology
Department of Statistics
University of Illinois, Urbana-Champaign
http://www.cs.illinois.edu/homes/czhai
[email protected]
Outline
1. General Idea of Topic Models
2. Basic Topic Models
-
We are here
Probabilistic Latent Semantic Analysis (PLSA)
Latent Dirichlet Allocation (LDA)
Applications of Basic Topic Models to Text Mining
3. Advanced Topic Models
-
Capturing Topic Structures
Contextualized Topic Models
Supervised Topic Models
4. Summary
2
Document as a Sample of Mixed Topics
Topic 1
[ Criticism of government response to the hurricane
primarily consisted of criticism of its response to the
approach of the storm and its aftermath, specifically in the
delayed response ] to the [ flooding of New Orleans. …
80% of the 1.3 million residents of the greater New Orleans
metropolitan area evacuated ] …[ Over seventy countries
pledged monetary donations or other assistance]. …
government 0.3
response 0.2
...
Topic 2
…
Topic k
Background k
city 0.2
new 0.1
orleans 0.05
...
donate 0.1
relief 0.05
help 0.02
...
is 0.05
the 0.04
a 0.03
...
•
•
How can we discover these topic
word distributions?
Many applications would be enabled
by discovering such topics
– Summarize themes/aspects
– Facilitate navigation/browsing
– Retrieve documents
– Segment documents
– Many other text mining tasks
3
Simplest Case: 1 topic + 1 “background”
Assume words in d are from
two distributions:
1 topic + 1 background
General Background
(rather than just one)
English Text
d
B
B
the 0.03
a 0.02
is 0.015
we 0.01
...
food 0.003
computer 0.00001
…
text 0.000006
…
Text mining
paper
d
How can we “get rid of” the
common words from the topic
to make it more discriminative?
Background LM: p(w|B)
the 0.031
a 0.018
…
text 0.04
mining 0.035
association 0.03
clustering 0.005
computer 0.0009
…
food 0.000001
…
Document LM: p(w|d)
4
The Simplest Case:
One Topic + One Background Model
Assume p(w|B) and are known
= assumed percentage of background words in d
Topic
choice
P(Topic)
Background words
P(w|B)
w
Document d
1-
Topic words
P(w| )
w
p( w) p( w | B ) (1 ) p( w | )
log p(d | ) c( w, d ) log[p( w | B ) (1 ) p( w | )]
wV
Maximum Likelihood
ˆ arg max log p ( d | )
5
Understanding a Mixture Model
Known
Background
p(w|B)
the 0.2
a 0.1
we 0.01
to 0.02
…
text 0.0001
mining 0.00005
…
Unknown
query topic
p(w|)=?
…
“Text mining”
…
text =?
mining =?
association =?
word =?
Suppose each model would be selected with
equal probability =0.5
The probability of observing word “text”:
p(“text”|B) + (1- )p(“text”| )
=0.5*0.0001 + 0.5* p(“text”| )
The probability of observing word “the”:
p(“the”|B) + (1- )p(“the”| )
=0.5*0.2 + 0.5* p(“the”| )
The probability of observing “the” & “text”
(likelihood)
[0.5*0.0001 + 0.5* p(“text”| )]
[0.5*0.2 + 0.5* p(“the”| )]
How to set p(“the”|) and p(“text”|) so as to maximize this likelihood?
assume p(“the”| )+p(“text”| )=constant
give p(“text”|) a higher probability than p(“the”|) (why?)
B and are competing for explaining words in document d!
6
Simplest Case Continued:
How to Estimate ?
Known
Background
p(w|B)
the 0.2
a 0.1
we 0.01
to 0.02
…
text 0.0001
mining 0.00005
…
=0.7
Observed
words
ML
Estimator
Unknown
query topic
p(w|)=?
“Text mining”
…
text =?
mining =?
association =?
word =?
…
=0.3
Suppose we know
the identity/label of each word ...
7
Can We Guess the Identity?
Identity (“hidden”) variable: zi {1 (background), 0(topic)}
zi
the
paper
presents
a
text
mining
algorithm
the
paper
...
1
1
1
1
0
0
0
1
0
...
Suppose the parameters are all known,
what’s a reasonable guess of zi?
- depends on (why?)
- depends on p(w|B) and p(w|) (how?)
p( zi 1 | wi )
p ( zi 1) p ( w | zi 1)
p( zi 1) p ( w | zi 1) p ( zi 0) p ( w | zi 0)
p ( w | B )
p( w | B ) (1 ) p current ( w | )
c( wi , d )(1 p ( zi 1 | wi ))
p new ( wi | )
w'V c(w' , d )(1 p( zi 1 | w' ))
E-step
M-step
Initially, set p(w| ) to some random values, then iterate …
8
An Example of EM Computation
p ( n ) ( zi 1 | wi )
p
( n 1)
( wi | )
Expectation-Step:
p ( wi | B )
p ( wi | B ) (1 ) p ( n ) ( wi | ) Augmenting data by guessing hidden variables
c( wi , d )(1 p ( n ) ( zi 1 | wi ))
c(w j , d )(1 p ( n) ( z j 1 | w j ))
w j vocabulary
Maximization-Step
With the “augmented data”, estimate parameters
using maximum likelihood
Assume =0.5
Word
#
P(w||B)
The
4 0.5
Paper
2 0.3
Text
4 0.1
Mining
2 0.1
Log-Likelihood
Iteration 1
P(z=1)
P(w|)
0.67
0.25
0.55
0.25
0.29
0.25
0.29
0.25
-16.96
Iteration 2
P(z=1)
P(w|)
0.71
0.20
0.68
0.14
0.19
0.44
0.31
0.22
-16.13
Iteration 3
P(z=1)
P(w|)
0.74
0.18
0.75
0.10
0.17
0.50
0.31
0.22
-16.02
9
Outline
1. General Idea of Topic Models
We are here
2. Basic Topic Models
-
Probabilistic Latent Semantic Analysis (PLSA)
Latent Dirichlet Allocation (LDA)
Applications of Basic Topic Models to Text Mining
3. Advanced Topic Models
-
Capturing Topic Structures
Contextualized Topic Models
Supervised Topic Models
4. Summary
10
Discover Multiple Topics in a Collection
k
pd ( w) B p ( w | B ) (1 B ) d , j p( w | j )
Percentage of background words
Coverage of topic j in doc d
j 1
k
log p (d ) c( w, d ) log[B p ( w | B ) (1 B ) d , j p( w | j )]
wV
j 1
Prob. of word w in topic j
k
log p (C | ) c( w, d ) log[B p( w | B ) (1 B ) d , j p( w | j )]
d C wV
j 1
Parameters: =(B, {d,j}, { j})
Topic coverage
in document d
Topic 1
warning 0.3
?
system 0.2..
?
Topic 2
aid 0.1
?
?
donation 0.05
support 0.02
? ..
1
statistics 0.2
?
loss 0.1
?
dead 0.05
? ..
k
Background B
is 0.05
?
?
the 0.04
a 0.03
? ..
“Generating” word w
in doc d in the collection
d,1
2
…
Topic k
Can be estimated using ML Estimator
d,2
1 - B
d, k
W
B
B
11
Probabilistic Latent Semantic
Analysis/Indexing (PLSA/PLSI) [Hofmann 99a, 99b]
•
•
•
•
Mix k multinomial distributions to generate a document
Each document has a potentially different set of mixing
weights which captures the topic coverage
When generating words in a document, each word may be
generated using a DIFFERENT multinomial distribution
(this is in contrast with the document clustering model
where, once a multinomial distribution is chosen, all the
words in a document would be generated using the same
multinomial distribution)
By fitting the model to text data, we can estimate (1) the
topic coverage in each document, and (2) word distribution
for each topic, thus achieving “topic mining”
12
How to Estimate Multiple Topics?
(Expectation Maximization)
Known
Background
p(w | B)
Unknown
topic model
p(w|1)=?
“Text mining”
Unknown
topic model
p(w|2)=?
“information
retrieval”
the 0.2
a 0.1
we 0.01
to 0.02
…
…
text =?
mining =?
association =?
word =?
…
…
E-Step:
Predict topic labels
using Bayes Rule
Observed Words
M-Step:
Max. Likelihood
Estimator
based on
“fractional
counts”
…
information =?
retrieval =?
query =?
document =?
…
13
Parameter Estimation
E-Step:
Word w in doc d is generated
- from cluster j
- from background
Application of Bayes rule
d( n, )j p ( n ) ( w | j )
p( zd ,w j )
p( zd ,w B)
M-Step:
Re-estimate
- mixing weights
- topic LM
Sum over all docs
in the collection
d( n, j 1)
(n) (n)
(w | j ' )
d
, j' p
j '1
k
B p ( w | B )
B p ( w | B ) (1 B ) j 1 d( n, )j p ( n ) ( w | j )
k
wV
j'
p ( n 1) ( w | j )
c( w, d )(1 p ( z d , w B )) p ( z d , w j )
wV
c( w, d )(1 p ( z d , w B )) p ( z d , w j ' )
c(w, d )(1 p( z B)) p( z j )
c(w' , d )(1 p( z B)) p( z j )
d C
w 'V
d C
d ,w
d ,w
d , w'
d , w'
Fractional counts contributing to
- using cluster j in generating d
- generating w from cluster j
14
How the Algorithm Works
c(w, d)
d1
d2
c(w,d)(1 - p(zd,w = B))p(zd,w=j)
c(w,d)p(zd,w = B)
πd1,1
( P(θ1|d1) )
aid 7
price 5
oil 6
Topic 1
( P(θ2|d1) )
πd2,1
πd2,2
( P(θ1|d2) )
( P(θ2|d2) )
Initial value
Topic 2
aid
price
Topic coverage
Initial value
aid 8
price 7
oil 5
P(w| θ)
πd1,2
Iteration
Iteration
Initializing
Iteration
1:
2:
1: EM
2:
πStep:
3,split
4,re-estimate
P(w|
5,word
…θj) counts
with
πd, j
d,Step:
j and
with
and P(w|
different
θj)Until
random
bytopics
adding
converging
(by
values
and
computing
normalizing
z’ s)
the splitted word counts
Initial value
oil
15
PLSA with Prior Knowledge
•
Users have some domain knowledge in mind, e.g.,
– We expect to see “retrieval models” as a topic in IR literature
– We want to see aspects such as “battery” and “memory” for
opinions about a laptop
– One topic should be fixed to model background words
(infinitely strong prior!)
•
We can easily incorporate such knowledge as priors of
PLSA model
16
Adding Prior :
Maximum a Posteriori (MAP) Estimation
* arg max p() p( Data | )
Most likely
Topic 1
Topic 2
warning 0.3
system 0.2..
aid 0.1
donation 0.05
support 0.02 ..
…
Topic k
statistics 0.2
loss 0.1
dead 0.05 ..
Background B
is 0.05
the 0.04
a 0.03 ..
Prior can be placed on as well
(more about this later)
Topic coverage
in document d
d,1
1
2
“Generating” word w
in doc d in the collection
d,2
1 - B
d, k
k
W
B
B
Parameters:
B=noise-level (manually set)
’s and ’s are estimated with Maximum A Posteriori (MAP)
17
Adding Prior as Pseudo Counts
Observed Doc(s)
Known
Background
p(w | B)
Unknown
topic model
p(w|1)=?
“Text mining”
Unknown
topic model
p(w|2)=?
“information
retrieval”
the 0.2
a 0.1
we 0.01
to 0.02
…
…
text =?
mining =?
association =?
word =?
…
…
…
information =?
retrieval =?
query =?
document =?
…
Suppose,
we know
the identity
of each
word ...
text
MAP
Estimator
Pseudo Doc
Size = μ
mining
18
Maximum A Posterior (MAP) Estimation
d( n, )j p ( n ) ( w | j )
p( zd ,w j )
p( zd ,w B)
( n 1)
d, j
k
j '1
d( n, )j ' p ( n ) ( w | j ' )
B p ( w | B )
B p ( w | B ) (1 B ) j 1 d( n, )j p ( n ) ( w | j )
k
wV
j'
p ( n 1) ( w | j )
c( w, d )(1 p ( z d , w B)) p ( z d , w j )
wV
c( w, d )(1 p ( z d , w B)) p ( z d , w j ' )
Pseudo counts of w from prior ’
c(w, d )(1 p( z B)) p( z j ) +p(w|’j)
c(w' , d )(1 p( z B)) p( z j ) +
d C
w 'V
d C
d ,w
d ,w
d , w'
d , w'
Sum of all pseudo counts
What if =0? What if =+?
A consequence of using conjugate prior is that the prior can be converted into
“pseudo data” which can then be “merged” with the actual data for parameter estimation
19
A General Introduction to EM
Data: X (observed) + H(hidden) Parameter:
“Incomplete” likelihood: L( )= log p(X| )
“Complete” likelihood: Lc( )= log p(X,H| )
EM tries to iteratively maximize the incomplete likelihood:
Starting with an initial guess (0),
1. E-step: compute the expectation of the complete likelihood
Q( ; ( n1) ) E ( n1) [ Lc ( ) | X ] p( H hi | X , ( n1) )log P( X , hi )
hi
2. M-step: compute (n) by maximizing the Q-function
( n) arg max Q( ; ( n1) ) arg max p( H hi | X , ( n1) )log P( X , hi )
hi
20
Convergence Guarantee
Goal: maximizing “Incomplete” likelihood: L( )= log p(X| )
I.e., choosing (n), so that L((n))-L((n-1))0
Note that, since p(X,H| ) =p(H|X, ) P(X| ) , L() =Lc() -log p(H|X, )
L((n))-L((n-1)) = Lc((n))-Lc( (n-1))+log [p(H|X, (n-1) )/p(H|X, (n))]
Taking expectation w.r.t. p(H|X, (n-1)),
L((n))-L((n-1)) = Q((n); (n-1))-Q( (n-1); (n-1)) + D(p(H|X, (n-1))||p(H|X, (n)))
Doesn’t contain H
EM chooses (n) to maximize Q
Therefore,
KL-divergence, always non-negative
L((n)) L((n-1))!
21
EM as Hill-Climbing:
converging to a local maximum
Likelihood p(X| )
L()= L((n-1)) + Q(; (n-1)) -Q( (n-1); (n-1) ) + D(p(H|X, (n-1) )||p(H|X, ))
L((n-1)) + Q(; (n-1)) -Q( (n-1); (n-1) )
next guess
current guess
Lower bound
(Q function)
E-step = computing the lower bound
M-step = maximizing the lower bound
22
Deficiency of PLSA
• Not a generative model
– Can’t compute probability of a new document
– Heuristic workaround is possible, though
• Many parameters high complexity of models
– Many local maxima
– Prone to overfitting
• Not necessary a problem for text mining (only
interested in fitting the “training” documents)
23
Latent Dirichlet Allocation (LDA) [Blei et al. 02]
• Make PLSA a generative model by imposing a
Dirichlet prior on the model parameters
– LDA = Bayesian version of PLSA
– Parameters are regularized
• Can achieve the same goal as PLSA for text
mining purposes
– Topic coverage and topic word distributions can be
inferred using Bayesian inference
24
LDA = Imposing Prior on PLSA
PLSA:
Topic coverage d,j is specific to
the “training documents”, thus
can’t be used to generate a new
document
d,1
1
2
LDA:
k
Topic coverage distribution {d,j }
for any document is sampled
from a Dirichlet distribution,
allowing for generating a new doc
d,2
“Generating” word w
in doc d in the collection
W
d, k
{d,j } are regularized
p( d ) Dirichlet( )
In addition, the topic word
distributions {j } are also drawn
from another Dirichlet prior
{d,j } are free for tuning
Topic coverage
in document d
Magnitudes of and
determine the variances of the prior,
thus also the strength of prior
(larger and stronger prior)
p(i ) Dirichlet( )
25
Equations for PLSA vs. LDA
PLSA
Core assumption
in all topic models
k
pd ( w | { j }, { d , j }) d , j p ( w | j )
j 1
k
log p (d | { j }, { d , j }) c( w, d ) log[ d , j p ( w | j )]
wV
j 1
log p (C | { j }, { d , j }) log p (d | { j }, { d , j })
d C
PLSA component
LDA
k
pd ( w | { j }, { d , j }) d , j p( w | j )
j 1
k
log p(d | , { j }) c( w, d ) log[ d , j p( w | j )] p( d | )d d
wV
j 1
k
log p(C | , ) log p(d | , { j }) p( j | )d1...d k
d C
j 1
Added by LDA
26
Parameter Estimation & Inferences in LDA
k
pd ( w | { j }, { d , j }) d , j p( w | j )
j 1
k
log p(d | , { j }) c( w, d ) log[ d , j p( w | j )] p( d | )d d
wV
j 1
k
log p(C | , ) log p(d | , { j }) p( j | )d1...d k
d C
j 1
Parameter estimation can be done in the same say as in PLSA:
Maximum Likelihood Estimator:
ˆ ˆ
( , ) arg max log p(C | , )
,
However, { j },{ d , j } must now be computed using posterior inference:
p
(
C
|
{
},
,
)
p
({
}
|
,
)
p
(
C
|
{
},
)
p
({
}
|
)
j
j
j
j
p ({ j } | C , , )
p (C | , )
p (C | , )
p (C | d , , ) p ( d | , ) p (C | d , , ) p ( d | )
Computationally intractable,
p ( d | C , , )
p (C | , )
p (C | , )
must resort to approximate inference!
27
LDA as a graph model
[Blei et al. 03a]
distribution over topics
for each document
(same as d on the previous slides)
Dirichlet priors
(d)
(d) Dirichlet()
distribution over words
for each topic
topic assignment
for each word
(j)
(same as j on the previous slides)
(j) Dirichlet()
zi Discrete( (d) )
zi
T
word generated from
assigned topic
wi Discrete( (zi) )
wi
Nd
D
Most approximate inference algorithms aim to infer p( zi | w, , )
from which other interesting variables can be easily computed
28
Approximate Inferences for LDA
• Many different ways; each has its pros & cons
• Deterministic approximation
– variational EM [Blei et al. 03a]
– expectation propagation [Minka & Lafferty 02]
• Markov chain Monte Carlo
– full Gibbs sampler [Pritchard et al. 00]
– collapsed Gibbs sampler [Griffiths & Steyvers 04]
Most efficient, and quite popular, but can only work with conjugate prior
29
The collapsed Gibbs sampler
[Griffiths & Steyvers 04]
• Using conjugacy of Dirichlet and multinomial
distributions, integrate out continuous parameters
P(z )
P(z | ) p()d
D
d 1
DT
P(w | z )
(d )
(
n
j j )
( )T
(T )
( n (jd ) )
j
T
P(w | z, ) p()d
j 1
TW
( j)
(
n
w w )
( )W
(W )
( nw( j ) )
w
• Defines a distribution on discrete ensembles z
P(w | z ) P(z )
P(z | w )
P(w | z ) P(z )
z
30
The collapsed Gibbs sampler
[Griffiths & Steyvers 04]
• Sample each zi conditioned on z-i
P ( zi | w , z i )
n
n
n
W n
T
( zi )
wi
( zi )
( di )
j
( di )
• This is nicer than your average Gibbs sampler:
– memory: counts can be cached in two sparse
matrices
– optimization: no special functions, simple arithmetic
– the distributions on and are analytic given z and
w, and can later be found for each sample
31
Gibbs sampling in LDA
iteration
1
i
wi
di
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
32
Gibbs sampling in LDA
iteration
1
2
i
wi
di
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
?
33
Gibbs sampling in LDA
iteration
1
2
i
wi
di
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
?
words in di assigned with topic j
Count of instances where wi is
assigned with topic j
Count of all words
assigned with topic j
words in di assigned with any topic
34
Gibbs sampling in LDA
iteration
1
2
i
wi
di
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
?
What’s the most likely topic for wi in di?
How likely would di choose topic j?
How likely would topic j
generate word wi ?
35
Gibbs sampling in LDA
iteration
1
2
i
wi
di
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
2
?
36
Gibbs sampling in LDA
iteration
1
2
i
wi
di
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
2
1
?
37
Gibbs sampling in LDA
iteration
1
2
i
wi
di
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
2
1
1
?
38
Gibbs sampling in LDA
iteration
1
2
i
wi
di
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
2
1
1
2
?
39
Gibbs sampling in LDA
1
2
iteration
…
1000
i
wi
di
zi
zi
zi
1
2
3
4
5
6
7
8
9
10
11
12
.
.
.
50
MATHEMATICS
KNOWLEDGE
RESEARCH
WORK
MATHEMATICS
RESEARCH
WORK
SCIENTIFIC
MATHEMATICS
WORK
SCIENTIFIC
KNOWLEDGE
.
.
.
JOY
1
1
1
1
1
1
1
1
1
1
2
2
.
.
.
5
2
2
1
2
1
2
2
1
2
1
1
1
.
.
.
2
2
1
1
2
2
2
2
1
2
2
1
2
.
.
.
1
2
2
2
1
2
2
2
1
2
2
2
2
.
.
.
1
…
40
Applications of Topic Models for Text Mining:
Illustration with 2 Topics
p(d | 1 2 ) [ p ( w | 1 ) (1 ) p ( w | 2 )]c ( w,d )
Likelihood:
wV
log p(d | 1 2 ) c( w, d ) log[p( w | 1 ) (1 ) p ( w | 2 )]
wV
Application Scenarios:
-p(w|1) & p(w|2) are known; estimate
The doc is about text mining and food nutrition,
how much percent is about text mining?
-p(w|1) & are known; estimate p(w|2)
30% of the doc is about text mining, what’s the
rest about?
-p(w|1) is known; estimate & p(w|2)
The doc is about text mining, is it also about some
other topic, and if so to what extent?
- is known; estimate p(w|1)& p(w|2)
30% of the doc is about one topic and 70% is about
another, what are these two topics?
-Estimate , p(w|1), p(w|2)
The doc is about two subtopics, find out what these two subtopics
are and to what extent the doc covers each.
41
Use PLSA/LDA for Text Mining
•
Both PLSA and LDA would be able to generate
– Topic coverage in each document: p(d = j)
– Word distribution for each topic: p(w|j)
– Topic assignment at the word level for each document
– The number of topics must be given in advance
•
These probabilities can be used in many different ways
– j naturally serves as a word cluster
– d,j can be used for document clustering
j* arg max d , j
j
– Contextual text mining: Make these parameters conditioned
on
context, e.g.,
• p(j |time), from which we can compute/plot p(time| j )
• p(j |location), from which we can compute/plot p(loc| j )
42
Sample Topics from TDT Corpus [Hofmann 99b]
43
How to Help Users Interpret a Topic Model?
[Mei et al. 07b]
• Use top words
term
0.16
– automatic, but hard to make sense relevance 0.08
weight
0.07
feedback
0.04
Term, relevance,
independence 0.03
weight, feedback
model
0.03
frequent
0.02
• Human generated labels
probabilistic 0.02
– Make sense, but cannot scale up
document
0.02
…
Retrieval Models
insulin
foraging
foragers
collected
grains
loads
collection
nectar
…
?
Question: Can we automatically generate
understandable labels for topics?
44
Automatic Labeling of Topics [Mei et al. 07b]
Statistical
topic models
term
0.1599
relevance
0.0752
weight
0.0660
feedback
0.0372
independence 0.0311
model
0.0310
frequent
0.0233
probabilistic 0.0188
document
0.0173
…
Collection (Context)
Relevance Score
term
0.1599
relevance
0.0752
weight
0.0660
feedback
0.0372
independence 0.0311
model
0.0310
frequent
0.0233
probabilistic 0.0188
document
0.0173
…
term
0.1599
relevance
0.0752
weight
0.0660
feedback
0.0372
independence 0.0311
model
0.0310
frequent
0.0233
probabilistic 0.0188
document
0.0173
…
Multinomial topic models
Coverage; Discrimination
Re-ranking
1
NLP Chunker
Ngram stat.
2
clustering algorithm;
database system, clustering algorithm, distance measure;
r tree, functional dependency, iceberg …
cube, concurrency control,
Ranked List
index structure …
Candidate label pool
of Labels
45
Relevance: the Zero-Order Score
• Intuition: prefer phrases well covering top words
p(“clustering”|) = 0.4 Clustering
√
Good Label (l1):
“clustering
algorithm”
p(“dimensional”|) = 0.3
dimensional
Latent
Topic
algorithm
p(clustering a lg orithm | )
p(clustering a lg orithm)
…
birch
shape
p(“shape”|) = 0.01
…
p(w|)
body
>
p(body shape| )
p(body shape)
Bad Label (l2):
“body shape”
p(“body”|) = 0.001
46
Relevance: the First-Order Score
• Intuition: prefer phrases with similar context (distribution)
Clustering
Clustering
Clustering
dimension
dimension dimension
Topic
P(w|)
partition
partition
algorithm
algorithm
…
hash
…
Good Label (l1):
…
“clustering
algorithm” algorithm
Bad Label (l2):
“hash join”
key
SIGMOD
Proceedings
hash
hash
p(w | clustering algorithm )
Score (l, )
p(w | ) PMI (w, l | C )
w
p(w | hash join)
D( | clustering algorithm) < D( | hash join)
47
Results: Sample Topic Labels
sampling 0.06
estimation 0.04
approximate 0.04
histograms 0.03
selectivity
0.03
histogram 0.02
answers
0.02
accurate
0.02
selectivity
estimation …
the, of, a, and,
to, data, > 0.02
…
clustering 0.02
clustering algorithm time
0.01
clustering structure
clusters
0.01
…
databases 0.01
large
0.01
large data, data
quality, high data, performance 0.01
0.005
data application, … quality
north
0.02
case
0.01
trial
0.01
iran
0.01
documents 0.01
walsh
0.009
reagan
0.009
charges 0.007
r tree
b tree …
indexing
methods
iran contra
…
tree
trees
spatial
b
r
disk
array
cache
0.09
0.08
0.08
0.05
0.04
0.02
0.01
0.01
48
Results: Contextual-Sensitive Labeling
sampling
estimation
approximation
histogram
selectivity
histograms
…
dependencies
functional
cube
multivalued
iceberg
buc
…
Context: Database
(SIGMOD Proceedings)
Context: IR
(SIGIR Proceedings)
selectivity estimation;
random sampling;
approximate answers;
distributed retrieval;
parameter estimation;
mixture models;
multivalue dependency
functional dependency
Iceberg cube
term dependency;
independence
assumption;
49
Using PLSA to Discover Temporal
Topic Trends [Mei & Zhai 05]
Normalized Strength of Theme
0.02
Biology Data
0.018
Web Information
0.016
Time Series
0.014
Classification
Association Rule
0.012
Clustering
0.01
Bussiness
0.008
0.006
0.004
0.002
0
1999
2000
2001
2002
2003
2004
gene 0.0173
expressions 0.0096
probability 0.0081
microarray 0.0038
…
marketing 0.0087
customer 0.0086
model 0.0079
business 0.0048
…
rules 0.0142
association 0.0064
support 0.0053
…
Time (year)
50
Construct Theme Evolution Graph [Mei & Zhai 05]
1999
2000
2001
2002
SVM 0.007
criteria 0.007
classifica –
tion
0.006
linear 0.005
…
decision 0.006
tree
0.006
classifier 0.005
class
0.005
Bayes
0.005
…
web 0.009
classifica –
tion 0.007
features0.006
topic 0.005
…
2003
mixture 0.005
random 0.006
cluster 0.006
clustering 0.005
variables 0.005
…
…
…
…
Classifica
- tion
text
unlabeled
document
labeled
learning
…
0.015
0.013
0.012
0.008
0.008
0.007
…
Informa
- tion 0.012
web
0.010
social 0.008
retrieval 0.007
distance 0.005
networks 0.004
…
2004
T
topic 0.010
mixture 0.008
LDA 0.006
semantic
0.005
…
51
Use PLSA to Integrate Opinions [Lu & Zhai 08]
Output
Topic: iPod
Expert review
with aspects
Text collection
of ordinary
opinions, e.g.
Weblogs
Design
Battery
Price..
Extra Aspects Review Aspects
Input
Similar
opinions
Design
Battery
Price
Supplementary
opinions
cute… tiny… ..thicker..
last many
die out
hrs
soon
could afford still
it
expensive
iTunes
warranty
… easy to use…
…better to extend..
Integrated Summary
52
Methods
• Semi-Supervised Probabilistic Latent
Semantic Analysis (PLSA)
– The aspects extracted from expert reviews
serve as clues to define a conjugate prior on
topics
– Maximum a Posteriori (MAP) estimation
– Repeated applications of PLSA to integrate
and align opinions in blog articles to expert
review
53
Results: Product (iPhone)
• Opinion Integration with review aspects
Review article
Similar opinions
You can make
N/A
emergency calls, but
you can't use any
other functions…
Confirm the
Activation
opinions from the
review will Feature
rated battery life of 8 iPhone
hours talk time, 24
Up to 8 Hours of Talk
hours of music
Time, 6 Hours of
playback, 7 hours of Internet Use, 7 Hours
video playback, and 6 of Video Playback or
hours on Internet use. 24 Hours of Audio
Playback
Battery
Supplementary opinions
… methods for unlocking the
iPhone have emerged on the
Unlock/hack
Internet in the past few weeks,
iPhone they involve tinkering
although
with the iPhone hardware…
Playing relatively high bitrate
VGA H.264 videos, our iPhone
lasted almost exactly 9 freaking
hours of continuous playback
with cell and WiFi on (but
Bluetooth off).
Additional info
under real usage
54
Results: Product (iPhone)
• Opinions on extra aspects
support
Supplementary opinions on extra aspects
15
You may have heard of iASign … an iPhone Dev Wiki tool that
Another way to
allows you to activate your phone without going through the
activate iPhone
iTunes rigamarole.
13
Cisco has owned the trademark on the name "iPhone" since
2000, when it acquired InfoGeariPhone
Technology
Corp., which
trademark
originally registered the name. originally owned by
13
Cisco
With the imminent availability of Apple's
uber cool iPhone, a
look at 10 things current smartphones like the Nokia N95 have
choiceand
for that the iPhone can't currently
been able to Adobetter
for a while
smart phones?
match...
55
Results: Product (iPhone)
• Support statistics for review aspects
People care about
price
Controversy: activation
requires contract with
AT&T
People comment a lot
about the unique wi-fi
feature
56
Comparison of Task Performance
of PLSA and LDA [Lu et al. 11]
• Three text mining tasks considered
– Topic model for text clustering
– Topic model for text categorization (topic model is used to obtain
low-dimensional representation)
– Topic model for smoothing language model for retrieval
• Conclusions
– PLSA and LDA generally have similar task performance for
clustering and retrieval
– LDA works better than PLSA when used to generate lowdimensional representation (PLSA suffers from overfitting)
– Task performance of LDA is very sensitive to setting of
hyperparameters
– Multiple local maxima problem of PLSA didn’t seem to affect task
performance much
57
Outline
1. General Idea of Topic Models
2. Basic Topic Models
-
Probabilistic Latent Semantic Analysis (PLSA)
Latent Dirichlet Allocation (LDA)
Applications of Basic Topic Models to Text Mining
3. Advanced Topic Models
-
We are here
Capturing Topic Structures
Contextualized Topic Models
Supervised Topic Models
4. Summary
58
Overview of Advanced Topic Models
• There are MANY variants and extensions of the
basic PLSA/LDA topic models!
• Selected major lines to cover in this tutorial
– Capturing Topic Structures
– Contextualized Topic Models
– Supervised Topic Models
59
Capturing Topic Structure:
Learning topic hierarchies
Topic
0
Topic
1.1
Topic
2.1
Topic
1.2
Topic
2.2
Topic
2.3
• Fixed hierarchies: [Hofmann 99c]
• Learning hierarchies: [Blei et al 03b]
60
Learning topic hierarchies
Topic
0
The topics in each
document form a path
from root to leaf
Topic
1.1
Topic
2.1
Topic
1.2
Topic
2.2
Topic
2.3
• Fixed hierarchies: [Hofmann 99c]
• Learning hierarchies:[Blei et al. 03b]
61
Twelve Years of NIPS [Blei et al. 03b]
62
Capturing Topic Structures:
Correlated Topic Model (CTM) [Blei & Lafferty 05]
63
Sample Result of CTM
64
Outline
1. Background
-
Text Mining (TM)
Statistical Language Models
2. Basic Topic Models
-
Probabilistic Latent Semantic Analysis (PLSA)
Latent Dirichlet Allocation (LDA)
Applications of Basic Topic Models to Text Mining
3. Advanced Topic Models
-
Capturing Topic Structures
Contextualized Topic Models
Supervised Topic Models
We are here
4. Summary
65
Contextual Topic Mining
• Documents are often associated with context (metadata)
– Direct context: time, location, source, authors,…
– Indirect context: events, policies, …
• Many applications require “contextual text analysis”:
– Discovering topics from text in a context-sensitive way
– Analyzing variations of topics over different contexts
– Revealing interesting patterns (e.g., topic evolution,
topic variations, topic communities)
66
Example: Comparing News Articles
Vietnam War
CNN
Afghan War
Fox
Before 9/11 During Iraq war
US blog
European blog
Iraq War
Blog
Current
Others
Common Themes
“Vietnam” specific
“Afghan” specific
“Iraq” specific
United nations
…
…
…
Death of people
…
…
…
…
…
…
…
What’s in common? What’s unique?
67
More Contextual Analysis Questions
• What positive/negative aspects did people say
about X (e.g., a person, an event)? Trends?
• How does an opinion/topic evolve over time?
• What are emerging research topics in computer
science? What topics are fading away?
• How can we mine topics from literature to
characterize the expertise of a researcher?
• How can we characterize the content exchanges
on a social network?
•…
68
Contextual Probabilistic
Latent Semantics Analysis [Mei & Zhai 06b]
Themes
Choose a theme
View1 View2 View3
Criticism
of government
Draw a word from
i
response togovernment
the hurricane
government 0.3
primarily
consisted of
response 0.2..
Document
government
response
criticism
of its response
context:
to … The total shut-in oil
Time = July
production
from 2005
the Gulf
Location
=
Texas
of Mexico
donate…
Author =24%
xxxof the
approximately
help
aid
annual production
and
Occup.
= Sociologist
theGroup
shut-in=gas
Age
45+
production … Over
… Orleans
seventy countries
new
pledged monetary
donations or other
assistance. …
donate 0.1
relief 0.05
help 0.02 ..
donation
city 0.2
new 0.1
orleans 0.05 ..
New
Orleans
Texas
July sociolo
2005
gist
Theme
coverages:
Choose a view
1
2
3
4
Texas
July 2005
1
2
3
4
……
document
1
2
3
4
Choose a
Coverage
69
Comparing News Articles [Zhai et al. 04]
Iraq War (30 articles) vs. Afghan War (26 articles)
The common theme indicates that “United Nations” is involved in both wars
Cluster 1
Common
Theme
Iraq
Theme
Afghan
Theme
united
nations
…
0.042
0.04
n
0.03
Weapons 0.024
Inspections 0.023
…
Northern 0.04
alliance
0.04
kabul
0.03
taleban
0.025
aid
0.02
…
Cluster 2
Cluster 3
killed
0.035
month
0.032
deaths
0.023
…
troops
0.016
hoon
0.015
sanches 0.012
…
taleban
0.026
rumsfeld 0.02
hotel
0.012
front
0.011
…
…
…
…
Collection-specific themes indicate different roles of “United Nations” in the two wars
70
Spatiotemporal Patterns in Blog Articles
[Mei et al. 06a]
•
•
Query= “Hurricane Katrina”
Topics in the results:
Government Response
bush 0.071
president 0.061
federal 0.051
government 0.047
fema 0.047
administrate 0.023
response 0.020
brown 0.019
blame 0.017
governor 0.014
•
New Orleans
city 0.063
orleans 0.054
new 0.034
louisiana 0.023
flood 0.022
evacuate 0.021
storm 0.017
resident 0.016
center 0.016
rescue 0.012
Oil Price
price 0.077
oil 0.064
gas 0.045
increase 0.020
product 0.020
fuel 0.018
company 0.018
energy 0.017
market 0.016
gasoline 0.012
Praying and Blessing
god 0.141
pray 0.047
prayer 0.041
love 0.030
life 0.025
bless 0.025
lord 0.017
jesus 0.016
will 0.013
faith 0.012
Aid and Donation
donate 0.120
relief 0.076
red 0.070
cross 0.065
help 0.050
victim 0.036
organize 0.022
effort 0.020
fund 0.019
volunteer 0.019
Personal
i 0.405
my 0.116
me 0.060
am 0.029
think 0.015
feel 0.012
know 0.011
something 0.007
guess 0.007
myself 0.006
Spatiotemporal patterns
71
Theme Life Cycles (“Hurricane Katrina”)
Oil Price
New Orleans
price 0.0772
oil 0.0643
gas 0.0454
increase 0.0210
product 0.0203
fuel 0.0188
company 0.0182
…
city 0.0634
orleans 0.0541
new 0.0342
louisiana 0.0235
flood 0.0227
evacuate 0.0211
storm 0.0177
…
72
Theme Snapshots (“Hurricane Katrina”)
Week2: The discussion moves towards the north and
west
Week1: The theme is the strongest along the Gulf of
Week3: The theme distributes more uniformly over the
Mexico
states
Week4: The theme is again strong along the east coast and the Gulf of
Mexico
Week5: The theme fades out in most states
73
Multi-Faceted Sentiment Summary [Mei et al. 07a]
(query=“Da Vinci Code”)
Facet 1:
Movie
Facet 2:
Book
Neutral
Positive
Negative
... Ron Howards selection of
Tom Hanks to play Robert
Langdon.
Tom Hanks stars in the
movie,who can be mad at
that?
But the movie might get
delayed, and even killed off if
he loses.
Directed by: Ron Howard
Writing credits: Akiva
Goldsman ...
Tom Hanks, who is my
favorite movie star act the
leading role.
protesting ... will lose your faith
by ... watching the movie.
After watching the movie I
went online and some
research on ...
Anybody is interested in
it?
... so sick of people making
such a big deal about a
FICTION book and movie.
I remembered when i first
read the book, I finished the
book in two days.
Awesome book.
... so sick of people making
such a big deal about a
FICTION book and movie.
I’m reading “Da Vinci Code”
now.
So still a good book to
past time.
This controversy book cause
lots conflict in west society.
…
74
Separate Theme Sentiment Dynamics
“book”
“religious beliefs”
75
Event Impact Analysis: IR Research
[Mei & Zhai 06b]
Theme:
retrieval
models
term
0.1599
relevance
0.0752
weight
0.0660
feedback
0.0372
independence 0.0311
model
0.0310
frequent
0.0233
probabilistic 0.0188
document
0.0173
…
xml
0.0678
email
0.0197
model
0.0191
collect
0.0187
judgment 0.0102
rank
0.0097
subtopic 0.0079
…
vector
0.0514
concept 0.0298
extend
0.0297
model
0.0291
space
0.0236
boolean 0.0151
function 0.0123
feedback 0.0077
…
1992
SIGIR papers
Publication of the paper “A language
modeling approach to information
retrieval”
Starting of the TREC
conferences
probabilist 0.0778
model
0.0432
logic
0.0404
ir
0.0338
boolean 0.0281
algebra 0.0200
estimate 0.0119
weight
0.0111
…
year
1998
model
language
estimate
parameter
distribution
probable
smooth
markov
likelihood
…
0.1687
0.0753
0.0520
0.0281
0.0268
0.0205
0.0198
0.0137
0.0059
76
The Author-Topic model
[Rosen-Zvi et al. 04]
each author has a distribution
over topics
the author of each word is chosen
uniformly at random
(a) Dirichlet()
(a)
A
xi Uniform(A (d) )
xi
(j) Dirichlet()
(j)
zi
zi Discrete( (xi) )
wi
wi Discrete( (zi) )
T
Nd
D
77
Four example topics from NIPS
TOPIC 19
TOPIC 24
TOPIC 29
WORD
PROB.
WORD
PROB.
WORD
LIKELIHOOD
0.0539
RECOGNITION
0.0400
MIXTURE
0.0509
CHARACTER
0.0336
POLICY
EM
0.0470
CHARACTERS
0.0250
DENSITY
0.0398
TANGENT
GAUSSIAN
0.0349
ESTIMATION
0.0314
DIGITS
LOG
0.0263
MAXIMUM
TOPIC 87
PROB.
WORD
PROB.
KERNEL
0.0683
0.0371
SUPPORT
0.0377
ACTION
0.0332
VECTOR
0.0257
0.0241
OPTIMAL
0.0208
KERNELS
0.0217
HANDWRITTEN 0.0169
ACTIONS
0.0208
SET
0.0205
0.0159
FUNCTION
0.0178
SVM
0.0204
IMAGE
0.0157
REWARD
0.0165
SPACE
0.0188
0.0254
DISTANCE
0.0153
SUTTON
0.0164
MACHINES
0.0168
PARAMETERS
0.0209
DIGIT
0.0149
AGENT
0.0136
ESTIMATE
0.0204
HAND
0.0126
DECISION
0.0118
MARGIN
0.0151
AUTHOR
PROB.
AUTHOR
PROB.
AUTHOR
PROB.
AUTHOR
PROB.
Tresp_V
0.0333
Simard_P
0.0694
Singh_S
0.1412
Smola_A
0.1033
Singer_Y
0.0281
Martin_G
0.0394
Barto_A
0.0471
Scholkopf_B
0.0730
Jebara_T
0.0207
LeCun_Y
0.0359
Sutton_R
0.0430
Burges_C
0.0489
Ghahramani_Z
0.0196
Denker_J
0.0278
Dayan_P
0.0324
Vapnik_V
0.0431
Ueda_N
0.0170
Henderson_D
0.0256
Parr_R
0.0314
Chapelle_O
0.0210
Jordan_M
0.0150
Revow_M
0.0229
Dietterich_T
0.0231
Cristianini_N
0.0185
Roweis_S
0.0123
Platt_J
0.0226
Tsitsiklis_J
0.0194
Ratsch_G
0.0172
Schuster_M
0.0104
Keeler_J
0.0192
Randlov_J
0.0167
Laskov_P
0.0169
Xu_L
0.0098
Rashid_M
0.0182
Bradtke_S
0.0161
Tipping_M
0.0153
Saul_L
0.0094
Sackinger_E
0.0132
Schwartz_A
0.0142
Sollich_P
0.0141
REINFORCEMENT 0.0411
REGRESSION 0.0155
78
Dirichlet-multinomial Regression (DMR)
[Mimno & McCallum 08]
Allows arbitrary features to be used to influence choice of topics
79
Supervised LDA [Blei & McAuliffe 07]
80
Sample Results of Supervised LDA
81
Latent Aspect Rating Analysis [Wang et al. 11]
• Given a set of review articles about a topic with overall
•
ratings (ratings as “supervision signals”)
Output
– Major aspects commented on in the reviews
– Ratings on each aspect
•
– Relative weights placed on different aspects by reviewers
Many applications
– Opinion-based entity ranking
– Aspect-level opinion summarization
– Reviewer preference analysis
– Personalized recommendation of products
– …
82
An Example of LARA
How to infer aspect ratings?
How to infer aspect weights?
Value
Location
Service
….. Value
Location
Service
…..
83
A Unified Generative Model for LARA
Entity
Aspects
Review
Aspect Rating Aspect Weight
Location
location
amazing
walk
anywhere
Room
room
dirty
appointed
smelly
Service
terrible
front-desk
smile
unhelpful
Excellent location in walking
distance to Tiananmen Square and
shopping streets. That’s the best
part of this hotel! The rooms are
getting really old. Bathroom was
nasty. The fixtures were falling off,
lots of cracks and everything
looked dirty. I don’t think it worth
the price. Service was the most
disappointing part, especially the
door men. this is not how you treat
guests, this is not hospitality.
0.86
0.04
0.10
84
Latent Aspect Rating Analysis Model
[Wang et al. 11]
• Unified framework
Excellent location in walking
distance to Tiananmen Square and
shopping streets. That’s the best
part of this hotel! The rooms are
getting really old. Bathroom was
nasty. The fixtures were falling off,
lots of cracks and everything
looked dirty. I don’t think it worth
the price. Service was the most
disappointing part, especially the
door men. this is not how you treat
guests, this is not hospitality.
Rating prediction module Aspect modeling module
85
Aspect Identification
• Amazon reviews: no guidance
battery life accessory service file format
volume video
86
Network Supervised Topic Modeling [Mei et al. 08]
•
Probabilistic topic modeling as an optimization problem
(e.g., PLSA/LDA: Maximum Likelihood):
O(Collection| Model) log(P(Collection| Model))
•
Regularized objective function with network constrains
– Topic distribution are smoothed over adjacent vertices
O(Collection, Network | Model)
log(P(Collection| Model)) Regularizer(Model, Network)
ModelParams arg maxO(Collection[, Network] | Model)
params
•
Flexibility in selecting topic models and regularizers
87
Instantiation: NetPLSA
• Basic Assumption: Neighbors have similar topic
distribution
topic distribution of a document
PLSA
k
O(C , G ) (1 ) ( c( w, d ) log p( j | d ) p( w | j ))
d
1
(
2
tradeoff
j 1
w
u ,v E
w(u, v) ( p( j | u ) p( j
Graph Harmonic Regularizer,
Generalization of [Zhu ’03],
1
2
f
j 1...k
T
j
k
f j , where f j ,u p( j | u )
j 1
difference
| v))2 ) of topic
distributio
n
importance
(weight) of
an edge
88
Topical Communities with PLSA
Topic 1
Topic 2
Topic 3
Topic 4
term
0.02
peer
0.02
visual
0.02
interface
0.02
question
0.02
patterns
0.01
analog
0.02
towards
0.02
protein
0.01
mining
0.01
neurons
0.02
browsing
0.02
training
0.01
clusters
0.01
vlsi
0.01
xml
0.01
weighting
0.01
stream
0.01
motion
0.01
generation
0.01
multiple
0.01
frequent
0.01
chip
0.01
design
0.01
recognition 0.01
e
0.01
natural
0.01
engine
0.01
relations
0.01
page
0.01
cortex
0.01
service
0.01
library
0.01
gene
0.01
spike
0.01
social
0.01
?? ? ?
Noisy
community
assignment
89
Topical Communities with NetPLSA
Topic 1
retrieval
Topic 2
Topic 3
Topic 4
mining
0.11
neural
0.06
web
0.05
information 0.05
data
0.06
learning
0.02
services
0.03
document
0.03
discovery
0.03
networks
0.02
semantic
0.03
query
0.03
databases
0.02
recognition 0.02
services
0.03
text
0.03
rules
0.02
analog
0.01
peer
0.02
search
0.03
association 0.02
vlsi
0.01
ontologies
0.02
evaluation
0.02
patterns
0.02
neurons
0.01
rdf
user
0.02
frequent
0.01
gaussian
0.01
management 0.01
relevance
0.02
streams
0.01
network
0.01
ontology
0.13
Information
Retrieval
Data mining
Web
Coherent
community
assignment
0.02
0.01
Machine
learning
90
Outline
1. General Idea of Topic Models
2. Basic Topic Models
-
Probabilistic Latent Semantic Analysis (PLSA)
Latent Dirichlet Allocation (LDA)
Applications of Basic Topic Models to Text Mining
3. Advanced Topic Models
-
Capturing Topic Structures
Contextualized Topic Models
Supervised Topic Models
4. Summary
We are here
91
Summary
•
Statistical Topic Models (STMs) are a new family of
language models, especially useful for
– Discovering latent topics in text
– Analyzing latent structures and patterns of topics
– Extensible for joint modeling and analysis of text and
associated non-textual data
•
•
•
PLSA & LDA are two basic topic models that tend to
function similarly, with LDA better as a generative model
Many different models have been proposed with probably
many more to come
Many demonstrated applications in multiple domains and
many more to come
92
Summary (cont.)
•
However, all topic models suffer from the problem of
multiple local maxima
– Make it hard/impossible to reproduce research results
– Make it hard/impossible to interpret results in real applications
•
Complex models can’t scale up to handle large amounts of
text data
– Collapsed Gibbs sampling is efficient, but only working for
conjugate priors
– Variational EM needs to be derived in a model-specific way
– Parallel algorithms are promising
•
Many challenges remain….
93
Challenges and Future Directions
• Challenge 1: How can we quantitatively evaluate
the benefit of topic models for text mining?
– Currently, most quantitative evaluation is based on
perplexity which doesn’t reflect the actual utility of a
topic model for text mining
– Need to separately evaluate the quality of both topic
word distributions and topic coverage
– Need to consider multiple aspects of a topic (e.g.,
coherent?, meaningful?) and define appropriate
measures
– Need to compare topic models with alternative
approaches to solving the same text mining problem
(e.g., traditional IR methods, non-negative matrix
factorization)
– Need to create standard test collections
94
• Challenge 2: How can we help users interpret a
topic?
– Most of the time, a topic is manually labeled in a
research paper; this is insufficient for real
applications
– Automatic labeling can help, but the utility still
needs to evaluated
– Need to generate a summary for a topic to enable a
user to navigate into text documents to better
understand a topic
– Need to facilitate post-processing of discovered
topics (e.g., ranking, comparison)
95
Challenges and Future Directions (cont.)
• Challenge 3: How can we address the problem of
multiple local maxima?
– All topic models have the problem of multiple local
maxima, causing problems with reproducing results
– Need to compute the variance of a discovered topic
– Need to define and report the confidence interval for a
topic
• Challenge 4: How can we develop efficient
estimation/inference algorithms for sophisticated
models?
– How can we leverage a user’s knowledge to speed up
inferences for topic models?
– Need to develop parallel estimation/inference algorithms
96
Challenges and Future Directions (cont.)
• Challenge 5: How can we incorporate linguistic knowledge
into topic models?
– Most current topic models are purely statistical
– Some progress has been made to incorporate linguistic
knowledge (e.g., [Griffiths et al. 04, Wallach 08])
– More needs to be done
•
Challenge 6: How can we incorporate domain knowledge
and preferences from an analyst into a topic model to
support complex text mining tasks?
– Current models are mostly pre-specified with little
flexibility for an analyst to “steer” the analysis process
– Need to develop a general analysis framework to enable
an analyst to use multiple topic models together to
perform complex text mining tasks
97
References
[Blei et al. 02] D. Blei, A. Ng, and M. Jordan. Latent dirichlet allocation. In T G Dietterich, S. Becker, and Z. Ghahramani,
editors, Advances in Neural Information Processing Systems 14, Cambridge, MA, 2002. MIT Press.
[Blei et al. 03a] David M. Blei, Andrew Y. Ng, Michael I. Jordan: Latent Dirichlet Allocation. Journal of Machine Learning
Research 3: 993-1022 (2003)
[Griffiths et al. 04] Thomas L. Griffiths, Mark Steyvers, David M. Blei, Joshua B. Tenenbaum: Integrating Topics and
Syntax. NIPS 2004
[Blei et al. 03b] David M. Blei, Thomas L. Griffiths, Michael I. Jordan, Joshua B. Tenenbaum: Hierarchical Topic Models
and the Nested Chinese Restaurant Process. NIPS 2003
[Teh et al. 04] Yee Whye Teh, Michael I. Jordan, Matthew J. Beal, David M. Blei: Sharing Clusters among Related Groups:
Hierarchical Dirichlet Processes. NIPS 2004
[Blei & Lafferty 05] David M. Blei, John D. Lafferty: Correlated Topic Models. NIPS 2005
[Blei & McAuliffe 07] David M. Blei, Jon D. McAuliffe: Supervised Topic Models. NIPS 2007
[Hofmann 99a] T. Hofmann. Probabilistic latent semantic indexing. In Proceedings on the 22nd annual international ACMSIGIR 1999, pages 50-57.
[Hofmann 99b] Thomas Hofmann: Probabilistic Latent Semantic Analysis. UAI 1999: 289-296
[Hofmann 99c] Thomas Hofmann: The Cluster-Abstraction Model: Unsupervised Learning of Topic Hierarchies from Text
Data. IJCAI 1999: 682-687
[Jelinek 98] F. Jelinek, Statistical Methods for Speech Recognition, Cambirdge: MIT Press, 1998.
[Lu & Zhai 08] Yue Lu, Chengxiang Zhai: Opinion integration through semi-supervised topic modeling. WWW 2008: 121130
[Lu et al. 11] Yue Lu, Qiaozhu Mei, ChengXiang Zhai: Investigating task performance of probabilistic topic models: an
empirical study of PLSA and LDA. Inf. Retr. 14(2): 178-203 (2011)
[Mei et al. 05] Qiaozhu Mei, ChengXiang Zhai: Discovering evolutionary theme patterns from text: an exploration of
temporal text mining. KDD 2005: 198-207
[Mei et al. 06a] Qiaozhu Mei, Chao Liu, Hang Su, ChengXiang Zhai: A probabilistic approach to spatiotemporal theme
pattern mining on weblogs. WWW 2006: 533-542
98
References
]Mei & Zhai 06b] Qiaozhu Mei, ChengXiang Zhai: A mixture model for contextual text mining. KDD 2006: 649-655
[Met et al. 07a] Qiaozhu Mei, Xu Ling, Matthew Wondra, Hang Su, ChengXiang Zhai: Topic sentiment mixture: modeling
facets and opinions in weblogs. WWW 2007: 171-180
[Mei et al. 07b] Qiaozhu Mei, Xuehua Shen, ChengXiang Zhai: Automatic labeling of multinomial topic models. KDD
2007: 490-499
[Mei et al. 08] Qiaozhu Mei, Deng Cai, Duo Zhang, ChengXiang Zhai: Topic modeling with network regularization. WWW
2008: 101-110
[Mimno & McCallum 08[ David M. Mimno, Andrew McCallum: Topic Models Conditioned on Arbitrary Features with
Dirichlet-multinomial Regression. UAI 2008: 411-418
[Minka & Lafferty 03] T. Minka and J. Lafferty, Expectation-propagation for the generative aspect model, In Proceedings
of the UAI 2002, pages 352--359.
[Pritchard et al. 00] J. K. Pritchard, M. Stephens, P. Donnelly, Inference of population structure using multilocus
genotype data,Genetics. 2000 Jun;155(2):945-59.
[Rosen-Zvi et al. 04] Michal Rosen-Zvi, Thomas L. Griffiths, Mark Steyvers, Padhraic Smyth: The Author-Topic Model for
Authors and Documents. UAI 2004: 487-494
[Wnag et al. 10] Hongning Wang, Yue Lu, Chengxiang Zhai: Latent aspect rating analysis on review text data: a rating
regression approach. KDD 2010: 783-792
[Wang et al. 11] Hongning Wang, Yue Lu, ChengXiang Zhai: Latent aspect rating analysis without aspect keyword
supervision. KDD 2011: 618-626
[Zhai et al. 04] ChengXiang Zhai, Atulya Velivelli, Bei Yu: A cross-collection mixture model for comparative text mining.
KDD 2004: 743-748
99