Gaussian Mixture Model
Download
Report
Transcript Gaussian Mixture Model
Gaussian Mixture Model
Mixture Modeling
A formalism for modeling a
probability density function
as a sum of parameterized
functions.
Normal parameters
Number of hidden
components
Class weights
Px , m Px m , m
M
m 1
Observations
Class weight,
class prior
probability,
multinomial
Multivariate
Normal
Normal = Gaussian
Gaussian Mixture
Gaussian Mixture
Let
Data Likelihood
Zi1
Z
Zi i2
ZiM
where Zim 1 if X i is fromgroup m, and 0 otherwise.
Under the assumption that the pairs (Zi,Xi) are mutually independent, their
joint density may be written
N
M
i 1
m
pZ, X | m px i | m
Zim
Data Log Likelihood
N
M
i 1
m
pZ, X | m px i | m
Zim
The complete-data log likelihood is thus
N
M
logpZ, X | Z im log m px i | m
i 1 m 1
EM Algorithm
E-Step: 估计Zim
M-Step: 估计πm, µm,∑m
EM Algorithm: E-Step
EM Algorithm: M-Step
Bayesian Ying-Yang Learning
• Proposed by Prof. Lei XU
Reference: Jinwen Ma, Jianfeng Liu. The BYY annealing learning algorithm for
Gaussian mixture with automated model selection, Pattern Recognition, 2007,
40:2029-2037.
Fig. 6. The segmentation result on the color image of house. (a) The
original color image of house; (b) the segmented image via the BYY
annealing learning algorithm (after 21 iterations).
Fig. 8. The segmentation result on the color image of jellies.
(a) The original color image of jellies;
(b) the segmented image via the BYY annealing learning algorithm
(after 22 iterations).