Unsupervised Learning Marco Chiarandini Department of Mathematics - - PowerPoint PPT Presentation

unsupervised learning
SMART_READER_LITE
LIVE PREVIEW

Unsupervised Learning Marco Chiarandini Department of Mathematics - - PowerPoint PPT Presentation

DM825 Introduction to Machine Learning Lecture 13 Unsupervised Learning Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark k -means Outline EM Algorithm 1. k -means 2. Expectation Maximization


slide-1
SLIDE 1

DM825 Introduction to Machine Learning Lecture 13

Unsupervised Learning

Marco Chiarandini

Department of Mathematics & Computer Science University of Southern Denmark

slide-2
SLIDE 2

k-means EM Algorithm

Outline

  • 1. k-means
  • 2. Expectation Maximization Algorithm

2

slide-3
SLIDE 3

k-means EM Algorithm

Outline

  • 1. k-means
  • 2. Expectation Maximization Algorithm

3

slide-4
SLIDE 4

k-means EM Algorithm

k-means

Given { x1, . . . , xm} and no yi we want to cluster the data Initialize cluster centroids randomly µ1, . . . , µk ∈ Rn repeat for i = 1 . . . m do ci ← arg minl xi − µl 2; // assign for l = 1 . . . k do µl ←

m

  • i=1

I{ci=l}xi

m

  • i=1

I{ci=l} ;

// move until convergence ; k is a parameter Optimization of the distortion function J( c, µ) = m

i=1 xi − µci 2

k-means ≡ coordinate descent on J: solve in c, µ by changing one variable while keeping the others fixed. Each probability solved optimally. J( c, µ) is non convex hence local optimality issues Convergence guaranteed by decreasing J.

4

slide-5
SLIDE 5

k-means EM Algorithm 5

slide-6
SLIDE 6

k-means EM Algorithm

In R

✞ ☎

k <- kmeans(train[,1:2], 3) > k$centers x y 1 8.0123 1.0406 2 1.5735 -0.7285 3 2.1856 7.5940 plot(train[,1:2], type=’n’) text(train[,1:2], as.character(k$cluster))

✝ ✆

−5 5 10 −15 −10 −5 5 10 15 x y A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B BB B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B BB B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C CC C C C C C C C C C C C C C C CC C C C C C C C C CC C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C CC C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C CC C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C CC C C C C C C CC C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C CC C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C CC C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C −5 5 10 −15 −10 −5 5 10 15 x y 1 1 1 3 2 3 1 2 1 3 1 3 3 3 1 3 3 3 1 3 3 2 1 3 1 1 3 3 2 1 3 3 3 3 1 2 1 1 2 1 3 2 1 2 1 3 2 1 3 2 1 3 3 1 1 1 3 3 1 2 1 2 1 1 3 2 2 3 3 1 3 3 2 3 3 3 1 2 2 2 3 2 3 1 3 1 2 2 3 3 3 3 2 1 3 2 1 3 2 1 1 1 3 2 2 2 2 1 1 3 3 2 1 3 2 2 1 3 3 3 1 3 3 2 1 3 2 3 3 1 1 2 2 2 2 1 3 3 1 2 3 1 3 1 1 3 2 3 3 1 1 1 3 3 1 2 1 2 2 1 3 3 3 3 1 1 3 1 3 1 3 3 3 1 1 3 3 2 2 3 1 1 1 1 3 3 3 3 3 2 3 1 3 2 2 1 3 3 3 1 3 1 3 2 2 1 3 1 3 1 2 3 1 3 2 3 2 3 2 3 1 1 3 2 2 1 3 1 3 1 2 2 3 2 2 2 1 3 1 3 1 3 1 3 3 3 1 3 2 3 1 2 2 3 1 3 3 1 3 2 2 3 1 1 1 3 2 3 2 3 3 2 3 1 1 2 2 3 3 3 1 3 1 3 3 3 3 1 1 3 3 3 3 3 2 1 3 2 2 2 1 3 3 3 3 1 1 3 3 1 1 1 3 1 3 1 2 3 3 3 1 2 3 2 3 3 3 1 2 2 3 1 1 2 3 2 3 2 2 3 2 1 3 1 3 3 3 1 3 3 3 3 1 1 3 2 1 1 1 1 3 3 3 2 1 3 3 1 1 3 3 2 1 1 3 3 1 1 2 2 1 3 3 3 3 3 3 3 2 1 3 1 1 1 1 1 2 3 2 3 3 3 3 3 3 2 1 1 2 3 1 3 1 2 1 3 3 3 1 1 2 1 3 3 3 1 3 2 2 2 2 3 1 1 2 3 3 1 3 3 2 3 3 1 1 2 1 2 1 1 2 3 1 1 3 3 1 2 1 1 2 1 3 3 1 1 3 1 1 3 3 3 3 2 2 3 1 1 1 2 1 1 2 1 3 1 3 3 3 2 2 1 1 3 3 2 2 3 3 1 1 1 2 3 2 3 1 1 2 1 2 1 1 2 2 3 3 3 2 3 2 3 1 2 2 3 3 3 1 2 3 3 3 2 3 3 1 3 1 2 1 1 3 3 3 1 1 1 3 1 1 3 2 3 1 3 2 2 1 2 2 1 3 2 3 1 1 3 2 1 1 1 2 3 3 3 3 2 2 3 1 1 1 1 1 3 2 1 3 3 2 2 2 3 1 1 2 1 2 3 2 1 1 3 1 3 3 1 1 3 1 2 1 3 1 1 3 1 1 3 3 1 1 2 3 3 1 3 2 1 3 1 2 3 3 2 1 1 1 3 2 1 3 2 3 1 3 3 1 2 11 2 3 2 3 1 3 1 1 3 2 3 2 1 3 2 3 2 1 3 3 3 2 3 2 1 1 3 3 2 1 2 2 3 2 1 3 3 3 1 3 3 22 3 1 1 3 1 3 2 2 3 3 3 1 3 1 3 1 3 2 3 2 3 1 1 3 1 2 3 2 2 1 3 3 3 2 1 3 2 1 1 1 3 3 3 1 3 3 1 3 2 3 1 3 2 3 3 3 2 1 1 1 3 1 1 2 1 1 1 1 3 3 1 1 2 1 3 2 1 3 3 3 1 3 2 3 3 2 3 1 2 1 1 3 3 3 1 2 1 2 2 1 2 1 2 3 3 1 3 1 2 1 3 1 3 3 1 3 2 1 3 1 1 3 3 2 3 2 3 2 11 3 3 3 3 3 3 2 3 1 1 2 3 3 1 2 1 3 1 1 2 3 1 1 1 3 2 1 3 3 3 3 3 1 3 1 3 3 3 2 3 1 3 3 2 2 1 2 3 3 3 3 2 2 1 3 3 3 3 3 2 2 3 1 2 1 1 2 3 1 1 2 1 2 2 1 2 3 3 3 3 3 3 2 3 1 3 1 1 1 3 3 1 3 1 3 1 2 1 3 1 3 3 1 3 2 2 2 3 1 1 3 1 2 1 3 2 3 3 3 3 2 3 3 1 3 3 3 3 3 2 1 3 1 1 3 1 2 3 3 3 2 2 3 1 3 1 2 3 3 3 2 2 3 3 1 3 1 3 2 2 2 3 1 2 3 2 3 3 2 3 3 3 1 2 3 3 1 2 2 1 1 2 3 1 3 3 1 1 3 1 3 2 2 2 3 1 3 2 1 3 3 1 1 3 3 3 2 3 3 1 1 1 3 1 3 3 1 3 1 1 3 3 3 1 1 2 3 1 1 1 1 3 2 1 3 1 2 1 1 3 3 2 1 1 2 1 3 1 3 2 1 3 2 2 3 2 1 3 3 3 3 2 3 1 3 3 2 1 1 1 2 1 3 1 3 1 1 3 2 2 3 2 3 3 2 2 2 2 2 3 1 3 1 3 1 1 2 2 1 3 3 1 1 2 2 3 2 1 1 1 3 2 1 3 3 2 2 3 1 3 1 2 1 3 3 2 3 2 3 1 1 3 3 1 1 1 1 2 1 3 1 1 3 2 1 1 3 1 2 2 1 2 1 3 3 1 1 3 3 2 3 3 2 2 2 3 3 2 3 1 3 1 2 1 3 3 3 3 2 2 2 3 1 2 3 3 3 3 2 1 3 1 1 2 3 3 2 2 2 3 1 3 1 3 3 3 1 3 3 3 3 1 1 3 2 3 3 3 3 1 3 3 1 2 2 1 3 1 2 3 2 1 2 3 3 2 3 1 2 3 2 1 2 3 3 1 1 3 2 3 3 1 3 1 3 1 3 3 1 3 2 3 2 3 2 3 2 1 2 2 1 3 2 3 2 3 3 2 1 3 3 1 3 1 3 3 1 1 2 3 3 2 3 2 3 3 3 2 1 1 1 2 3 2 2 3 3 1 2 3 1 2 2 1 1 3 1 2 3 3 3 3 1 3 2 1 1 3 3 1 1 1 1 3 1 3 3 3 1 3 1 1 2 1 1 2 3 3 2 2 1 11 1 3 1 3 2 3 1 1 2 1 2 2 2 1 2 1 2 2 1 2 3 3 1 1 1 3 3 2 1 1 3 3 3 1 1 3 1 3 2 2 1 2 3 3 3 1 2 2 3 1 2 3 1 1 3 3 3 3 2 1 3 2 3 1 3 3 3 1 3 3 3 2 3 2 2 3 3 3 2 1 1 1 2 1 2 2 3 3 1 3 3 1 1 2 3 1 1 1 1 1 3 3 3 1 2 2 2 2 3 2 1 3 1 2 3 3 3 1 1 1 2 3 3 1 1 2 2 2 3 3 3 3 3 1 1 1 1 3 1 1 1 3 3 1 1 1 1 33 1 3 2 3 3 3 2 1 3 3 1 3 1 3 3 2 1 1 2 3 3 3 1 2 1 3 1 3 3 2 3 1 3 3 3 3 1 1 1 1 3 3 1 2 3 1 3 2 2 1 3 1 2 3 1 3 2 3 1 1 2 1 1 1 1 1 1 2 2 1 3 3 3 2 3 1 3 2 1 1 3 2 3 1 3 1 2 3 3 3 1 3 2 1 2 1 1 1 1 1 3 3 3 3 3 1 3 2 1 2 3 1 1 1 1 1 1 1 2 3 2 2 2 3 3 2 2 3 3 2 1 2 3 1 2 3 1 1 3 3 1 3 1 3 3 1 3 3 2 2 3 1 3 3 1 2 3 1 1 1 1 2 3 3 2 3 3 2 1 3 2 2 3 3 1 1 1 3 2 1 1 1 3 3 2 3 3 1 3 3 3 1 1 1 3 2 1 3 1 3 1 3 3 2 3 1 2 1 1 1 1 3 1 3 1 3 1 1 1 1 3 3 3 3 3 3 3 3 3 1 3 1 3 1 1 3 2 3 3 3 1 3 3 1 3 2 2 1 1 3 1 3 2 3 3 1 3 2 2 1 3 3 1 2 2 1 3 3 3 2 3 3 2 2 3 2 3 2 3 3 1 1 1 3 3 3 1 3 1 1 1 1 3 2 1 1 3 3 3 2 1 3 1 3 3 1 1 3 3 3 2 2 1 3 2 2 3 1 3 3 1 3 3 3 3 3 2 3 1 2 3 3 3 3 2 2 1 1 2 3 1 1 3 3 2 1 2 1 1 3 2 3 3 3 3 1 3 2 1 3 3 3 1 3 2 1 2 3 1 1 3 1 3 1 1 1 3 3 2 3 3 2 1 3 2 1 3 1 1 1 1 3 1 3 1 2 3 2 3 1 3 2 3 3 3 3 3 1 3 1 2 1 3 3 3 3 3 1 2 1 1 1 3 2 1 3 2 1 1 3 3 1 3 3 3 3 1 3 1 2 2 3 3 1 2 1 1 2 3 2 2 1 3 1 1 3 3 1 1 3 2 3 2 3 1 1 1 2 1 3 3 3 1 2 3 3 3 2 2 2 3 3 1 1 1 1 3 2 2 3 1 1 3 3 3 1 1 1 1 3 3 2 1 3 3 1 1 3 1 1 1 2 1 3 2 1 2 2 2 3 2 3 2 3 1 2 3 3 1 3 1 1 2 3 3 2 1 1 2 2 1 3 1 3 3 3 3 3 3 1 1 3 2 3 1 2 3 3 3 1 2 2 3 2 2 2 2 2 1 1 1 1 3

6

slide-7
SLIDE 7

k-means EM Algorithm

Outline

  • 1. k-means
  • 2. Expectation Maximization Algorithm

7

slide-8
SLIDE 8

k-means EM Algorithm

Mixture Models

We can simplfy complicated distributions p( x) by introducing latent variables. Then: p( x) =

  • z

p( x, z) =

  • z

p( x | z)p( z) p( x | z) may be more tractable to express.

8

slide-9
SLIDE 9

k-means EM Algorithm

Expectation Maximization Algorithm

Given { x1, . . . , xm} and no yi we want to cluster the points. we wish to model the joint prob. distribution p(xi, zi) = p(xi | zi)p(zi) zi are latent random variables

◮ zi ∼ Multinomial(

φ), φl ≥ 0, k

l=1 φl = 1 (p(zi = l) = φl) ◮ xi | zi = l ∼ N(µl, Σl)

Estimation of φ, µ, σ (learning) ℓ(φ, µ, Σ) =

m

  • i=1

log p(xi, φ, µ, Σ) =

m

  • i=1

log

k

  • zi=l

p(xi | zi, µ, Σ)p(zi, φ)

9

slide-10
SLIDE 10

k-means EM Algorithm

If zi known (supervised learning): Gaussian discriminant analysis generalized to k > 2 and different variance ℓ(φ, µ, Σ) =

m

  • i=1

log p(xi | zi, µ, Σ) + log p(zi, φ) φl = 1 m

m

  • i=1

I{zi = l} µl = m

i=1 I{zi = l}xi

m

i=1 I{zi = l}

Σl = m

i=1 I{zi = l}(xi −

µi)(xi − µi)T m

i=1 I{zi = l}

10

slide-11
SLIDE 11

If zi not known (unsupervised learning): repeat for i=1. . . m, l=1. . . k do wj ← p(zi = l | xi, φ, µ, Σ); // (E-step) for l=1. . . k do φl = 1 m

m

  • i=1

wi

l

µl = m

i=1 wi lxi

m

i=1 wi l

(M-step) Σl = m

i=1 wi l(xi −

µi)(xi − µi)T m

i=1 wi l

until convergence ; wj ← p(zi = l | xi, φ, µ, Σ) = p(xi = l | zi = l, φ, µ, Σ)p(zi = l, φ) k

l=1 p(xi = l | zi = l, φ, µ, Σ)p(zi = l, φ)

slide-12
SLIDE 12

k-means EM Algorithm

Analysis of EM algorithm

Definition (Convex functions) f : R → R is convex ⇐ ⇒ f ′′ ≥ 0 ∀x ∈ R f : Rn → R is convex ⇐ ⇒ H ≥ 0 ∀x ∈ Rn Theorem (Jensen’s inequality) f convex, x random variable ⇒ E[f(x)] ≥ f(E[x]) (if f strictly convex = ⇒ E[f(x)] = f(E[x]) iff x = E[x], ie. x = c)

12

slide-13
SLIDE 13

k-means EM Algorithm

We wish to fit the parameters of a model p(x, z) ℓ( θ) =

m

  • i=1

log p(xi, θ) =

m

  • i=1

log

i

  • z

p(xi, zi, θ) zi not observed opt problem not easy EM does max likelihood estimation:

◮ E-step construct lower bound for ℓ(

θ)

◮ M-step optimize the LB

Qj distribution over zi (Qi(z) ≥ 0),

z Qi(z) = 1)

ℓ(θ) =

  • i

log

  • zi

p(xi, zi, θ) =

  • i

log

  • zi

Qi(zi)p(xi, zi, θ) Qi(zi) Jensen’s ineq. for concave functions ≥

  • i
  • zi

Qi(zi) log p(xi, zi, θ) Qi(zi) (*)

13

slide-14
SLIDE 14

(*) gives a LB for ℓ(θ)∀Qi. Which Qi should we choose? Given some parameters θ, try to make qi highest possible. It holds with equality, i.e.: p(xi, zi, θ) Qi(zi) = c Qi(zi) ∝ p(xi, zi, θ) Qi(zi) = p(xi, zi, θ)

  • zi p(xi, zi, θ)

= p(xi, zi, θ) p(xi, θ) = = p(zi | xi, θ) Then maximize (*) wrt θ repeat for each i do Qi(zi) ← p(zi | xi, θ) ; // E-step θ ← arg maxθ

  • i
  • zi Qi(zi) log p(xi,zi,θ)

Qi(zi) ;

// M-step until convergence ;

slide-15
SLIDE 15

k-means EM Algorithm

Convergence: we want to show that ℓ(θt) ≤ ℓ(θt+1) Qt

i(zi) = p(zi | xi, θt)

ℓ(θt) =

  • i
  • zi

Qt

i(zi) log p(xi, zi, θt)

Qt

i(zi)

ℓ(θt+1) ≥

  • i
  • zi

Qt

i(zi) log p(xi, zi, θt+1)

Qt

i(zi)

because Jensen ∀θ ≥

  • i
  • zi

Qt

i(zi) log p(xi, zi, θt)

Qt

i(zi)

because θt+1 max’s ℓ(θ) = ℓ(θt) Thus monotonic convergence. Stop if improvement smaller than a tollerance. EM-algorithm as a coordinate descent on J(Q, θ) =

  • i
  • zi

Qi(zi) log p(xi, zi, θ) Qi(zi)

15

slide-16
SLIDE 16

k-means EM Algorithm

Mixture of Gaussian revisited E-step: wl

i = Qi(zi = l) = p(zi = l | xi, φ, µ, Σ)

(prob. of zi taking l under Qi(zi = l)) M-step: maximize w.r.t. φ, µ, Σ: ℓ(θ) =

  • i
  • zi

Qi(zi) log p(xi, zi, θt) Qt

i(zi)

=

  • i

k

  • l=1

Qi(zi) log p(xi | zi = l, θt)p(zi = l, φ) Qt

i(zi)

=

  • i

k

  • l=1

wi

l log 1 2πn/2Σ1/2 exp(−1/2(xi − µl)Σ−1 l

(xi − µl)) wi

l

16