Contents

Details

Given a set
X
of
n
feature vectors
x
1
= (
x
11
,…,
x
1
p
), ...,
x
n
= (
x
n
1
,…,
x
np
) of dimension
p
, the problem is to find a maximum-likelihood estimate of the parameters of the underlying distribution when the data is incomplete or has missing values.

Expectation-Maximization (EM) Algorithm in the General Form

Let
X
be the observed data which has log-likelihood
l
(
θ
;
X
) depending on the parameters
θ
. Let
X
m
be the latent or missing data, so that
T
=(
X
,
X
m
) is the complete data with log-likelihood
l
0
(
θ
;
X
). The algorithm for solving the problem in its general form is the following EM algorithm ([Dempster77], [Hastie2009]):
1. Choose initial values of the parameters
θ
(0)
2. Expectation step
: in the
j
-th step, compute
Q
(
θ
',
θ
(
j
)
) =
E
(
l
0
(
θ
';
T
)|
X
,
θ
(
j
)
) as a function of the dummy argument
θ
'
3. Maximization step
: in the
j
-th step, calculate the new estimate
θ
(
j
+1)
by maximizing
Q
(
θ
',
θ
(
j
)
) over
θ
'
4. Repeat steps 2 and 3 until convergence

EM algorithm for the Gaussian Mixture Model

Gaussian Mixture Model (GMM) is a mixture of
k
p
-dimensional multivariate Gaussian distributions represented as
where
Σ
k
i
= 1
α
i
= 1 and
α
i
≥ 0.
The
pd
(
x
|
θ
i
) is the probability density function with parameters
θ
i
= (
m
i
,
Σ
i
), where
m
i
is the vector of means, and
Σ
i
is the variance-covariance matrix. The probability density function for a
p
-dimensional multivariate Gaussian distribution is defined as follows:
Let
z
ij
=
I
{
x
i
belongs to j mixture component
} be the indicator function and
θ
=(
α
1
, ...,
α
k
;
θ
1
, ...,
θ
k
).
Computation
The EM algorithm for GMM includes the following steps:
Define the weights as follows:
for
i
=1, ...,
n
and
j
=1, …,
k
.
1. Choose initial values of the parameters
2. Expectation step
: in the
j
-th step, compute the matrix
W
= (
w
ij
)
nxk
with the weights
w
ij
3. Maximization step
: in the
j
-th step, for all
r
=1, ...,
k
compute:
1. The mixture weights
where
is the "
amount
" of the feature vectors that are assigned to the
r
-th mixture component
2. Mean estimates
3. Covariance estimate
of size
p
x
p
with
4. Repeat steps 2 and 3 until any of these conditions is met:
• where the likelihood function is
• The number of iterations exceeds the predefined level.
Initialization
The EM algorithm for GMM requires initialized vector of weights, vectors of means, and variance-covariance [Biernacki2003, Maitra2009].
The EM initialization algorithm for GMM includes the following steps:
1. Perform
nTrials
starts of the EM algorithm with
nIterations
iterations and start values:
• Initial means -
k
different random observations from the input data set
• Initial weights - the values of 1/
k
• Initial covariance matrices - the covariance of the input data
2. Regard the result of the best EM algorithm in terms of the likelihood function values as the result of initialization

Product and Performance Information

1

Intel's compilers may or may not optimize to the same degree for non-Intel microprocessors for optimizations that are not unique to Intel microprocessors. These optimizations include SSE2, SSE3, and SSSE3 instruction sets and other optimizations. Intel does not guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel. Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference Guides for more information regarding the specific instruction sets covered by this notice.

Notice revision #20110804