Correlated T
- pic Models
Authors: Blei and LaffertY, 2006 Reviewer: Casey Hanson
Correlated T opic Models Authors: Blei and LaffertY, 2006 - - PowerPoint PPT Presentation
Correlated T opic Models Authors: Blei and LaffertY, 2006 Reviewer: Casey Hanson Recap Latent Dirichlet Allocation set of documents . = set of topics . = set of all words. || words in each doc.
Authors: Blei and LaffertY, 2006 Reviewer: Casey Hanson
ππ ~ πΈππ (π½)
π,π β‘ ππ’β word in document π. π π,π~ Multi(πΆππ,π)
1:πΈ,1:π, π½, π)
1:πΈ,1:π, π½, π)
πΎ π π π(. . ), is intractable
π π¦ = 1 π 2π πβ π¦βπ 2
2π2
1 π π=1 π
π¦π
1 π π=1 π
π¦π β π 2
π π = π
π¦ π1 β¦ ππ =
1 2π π/2 det Ξ£ πβ1
2 πβπ πΞ£β1(πβπ)
Example: 2D Case
πΉ[π¦1] πΉ[π¦2] = π1 π2
πΉ π¦1 β π1 2 πΉ π¦1 β π1 π¦2 β π2 πΉ π¦1 β π1 π¦2 β π2 πΉ π¦2 β π2 2
ππ1
2
ππ1,π2ππ1ππ2 ππ1,π2ππ1ππ2 ππ2
2
π¦1 β π1 π¦2 β π2 = π=1
π ππ,1βπ1 ππ,2βπ2 π
to correct assumption 1.
π π¦ π = β π¦ β ππ π β π π¦ βπ΅ π
π π¦|π = π π¦ ππ¦(1 β π)πβπ¦, π¦ β 0,1,2, β¦ , π
π 1βπ
β π¦ =
π π¦ , π΅ π = π log 1 β π, π π¦ = π¦
π π¦ = π π¦ π
π¦β log π 1βπ +πβ log 1βπ
Natural Parameterization
π1 ππ π2 ππ 1
π1 ππ log π2 ππ 1
πππβ ππ π=1 πππ
π1 ππ log π2 ππ β¦ .0], πβ²π = log ππ ππ
ππβ²πβ ππ 1+ π=1
πβ1 ππ ππβ²
π1 ππ log π2 ππ β¦ 0]
π ~ πͺ
πβ1 π, π
ππ ππ
π = 0 0 π, π = [1 0; 0 1]
π = β0.9, β0.9 , Ξ£ = [1 0; 0 1] π = β0.9, β0.9 , Ξ£ = [1 β 0.9; β0.9 1]
π = β0.9, β0.9 , Ξ£ = [1 0.4; 0.4 1]
π,π| ππ,π, πΎ1:πΏ ~ Categorical πΎππ
set, given you trained on the previous 9 sets, for both LDA and CTM.
CTM shows a much higher log likelihood as the number of topics increases.
number of equally likely words
word resolution.
words in a document, how likely is the rest of the words in the document according to your model?
topic probabilities.
drawn, from a Dirichlet to a logistic normal function.
than LDA.