csci 590 : machine learning lower dimensional latent...
TRANSCRIPT
![Page 1: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/1.jpg)
Models
CSCI 590 : Machine Learning
Lower Dimensional Latent Semantic Space
LDA,PLSI,NMF
Halid Ziya Yerebakan
March 23 2015
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 2: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/2.jpg)
iupui
Models
Document-Term Matrices
Figure: Document-term matrix
Document is vector offrequencies of the words.
Mostly sparse matrices.
Bag of words assumption,order information ignored.
Other contexts : Imagepixel , haplotypes , movierating.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 3: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/3.jpg)
iupui
Models
Latent Semantic Space
Figure: Latent Semantic Space
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 4: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/4.jpg)
Example
Figure: LDA on TREC AP Corpus [1]
![Page 5: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/5.jpg)
iupui
Models
De�nitions
De�nition
Topic : Topic is a distribution over words.
Figure: Topic distribution
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 6: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/6.jpg)
iupui
Models
De�nitions
De�nition
Topic proportions : Mixture weights of topics for a document.Documents have multiple topics.
Figure: Topic proportionsHalid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 7: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/7.jpg)
iupui
Models
PLSILDANMFImplementations
Outline
1 ModelsPLSILDANMFImplementations
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 8: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/8.jpg)
iupui
Models
PLSILDANMFImplementations
PLSI References
Hofman, T. (1999, August). Probabilistic latent semanticindexing. In Proceedings of the 22nd annual international ACMSIGIR conference on Research and development in informationretrieval (pp. 50-57). ACM.
Hofmann, T. (1999, July). Probabilistic latent semanticanalysis. In Proceedings of the Fifteenth conference onUncertainty in arti�cial intelligence (pp. 289-296). MorganKaufmann Publishers Inc..
Hofmann, T. (2001). Unsupervised learning by probabilisticlatent semantic analysis. Machine learning, 42(1-2), 177-196.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 9: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/9.jpg)
iupui
Models
PLSILDANMFImplementations
PLSI Neighbors
Figure: PLSI Concept Graph
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 10: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/10.jpg)
iupui
Models
PLSILDANMFImplementations
Graphical Generative Model
Figure: PLSI Model
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 11: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/11.jpg)
iupui
Models
PLSILDANMFImplementations
Generative Model
Figure: PLSI Model
Generative Process
1 Select document with probabilityP(d)
2 Pick a latent class (topic) z withprobability P(z|d)
3 Generate a word w withprobability P(w|z)
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 12: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/12.jpg)
Results
Topics
Figure: Few Topics in PLSI[1]
![Page 13: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/13.jpg)
IR Performance
Figure: PLSI IR performance[1]
![Page 14: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/14.jpg)
iupui
Models
PLSILDANMFImplementations
Advantages - Disadvantages
Advantages
De�nes proper probability distributions on words
More realistic document model, interpretable topics.
Disadvantages
K is �xed
Local Maximum , Over�tting
PLSI is not nested like LSA ( n dimensional solution includesn-1 dimensional solution).
Slow training speed compared to LSA
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 15: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/15.jpg)
iupui
Models
PLSILDANMFImplementations
Advantages - Disadvantages
Advantages
De�nes proper probability distributions on words
More realistic document model, interpretable topics.
Disadvantages
K is �xed
Local Maximum , Over�tting
PLSI is not nested like LSA ( n dimensional solution includesn-1 dimensional solution).
Slow training speed compared to LSA
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 16: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/16.jpg)
iupui
Models
PLSILDANMFImplementations
Outline
1 ModelsPLSILDANMFImplementations
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 17: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/17.jpg)
iupui
Models
PLSILDANMFImplementations
LDA References
Blei, D. M., Ng, A. Y., & Jordan, M. I. (2003). Latent dirichletallocation. the Journal of machine Learning research, 3,993-1022.
Blei, D. M., Ng, A. Y., & Jordan, M. I. (2001). Latent dirichletallocation. In Advances in neural information processingsystems.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 18: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/18.jpg)
iupui
Models
PLSILDANMFImplementations
LDA Concept Graph
Figure: LDA Concept Graph
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 19: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/19.jpg)
iupui
Models
PLSILDANMFImplementations
LDA Model
Figure: LDA Model
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 20: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/20.jpg)
iupui
Models
PLSILDANMFImplementations
LDA Model
1 For each document2 Chose N∼Poisson(.)
1 Chose θ ∼ Dir(α)2 For each word wn
1 Chose topic zn ∼ Multinomial(θ)
2 Chose a word from wn ∼ p(wn|βzn )
P(θ, z,w|α.β) = P(θ|α)N∏
n=1
P(zn|θ)P(wn|zn, β)
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 21: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/21.jpg)
iupui
Models
PLSILDANMFImplementations
LDA Geometry
[1]
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 22: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/22.jpg)
Experiments
Figure: LDA on TREC AP Corpus [1]
![Page 23: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/23.jpg)
iupui
Models
PLSILDANMFImplementations
Advantages-Disadvantages
Advantages
De�nes a proper generative model on documents.
It can be incorporated in more complex models
Disadvantages
K is �xed
IID generative assumption of topics.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 24: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/24.jpg)
iupui
Models
PLSILDANMFImplementations
Advantages-Disadvantages
Advantages
De�nes a proper generative model on documents.
It can be incorporated in more complex models
Disadvantages
K is �xed
IID generative assumption of topics.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 25: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/25.jpg)
iupui
Models
PLSILDANMFImplementations
Outline
1 ModelsPLSILDANMFImplementations
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 26: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/26.jpg)
iupui
Models
PLSILDANMFImplementations
References
Lee, D. D., & Seung, H. S. (1999). Learning the parts ofobjects by non-negative matrix factorization. Nature,401(6755), 788-791.
Seung, D., & Lee, L. (2001). Algorithms for non-negativematrix factorization. Advances in neural information processingsystems, 13, 556-562.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 27: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/27.jpg)
iupui
Models
PLSILDANMFImplementations
Concept Graph
Figure: NMF Concept Graph
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 28: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/28.jpg)
iupui
Models
PLSILDANMFImplementations
Non Negative Matrix Factorization
Figure: NMF
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 29: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/29.jpg)
iupui
Models
PLSILDANMFImplementations
Motivation
Matrix Factorization
Viµ ≈ (WH)iµ =∑r
a=1WiaHaµ
Designed for
Part based representation , no cancellation.
Compressed form of data (n+m)r<nm
Intuitive notions of basis vectors
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 30: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/30.jpg)
iupui
Models
PLSILDANMFImplementations
Motivation
Matrix Factorization
Viµ ≈ (WH)iµ =∑r
a=1WiaHaµ
Designed for
Part based representation , no cancellation.
Compressed form of data (n+m)r<nm
Intuitive notions of basis vectors
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 31: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/31.jpg)
iupui
Models
PLSILDANMFImplementations
Algorithm
Objective Function
F =n∑
i=1
m∑µ=1
[Viµlog(WH)iµ − (WH)iµ]
s.t Wia ≥ 0,Haµ ≥ 0
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 32: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/32.jpg)
iupui
Models
PLSILDANMFImplementations
Algorithm
Update Equations
Wia ← Wia
∑µ
Viµ
(WH)iµHaµ
Wia ←Wia∑j Wja
Haµ ← Haµ
∑i
WiaViµ
(WH)iµ
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 33: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/33.jpg)
iupui
Models
PLSILDANMFImplementations
Face Images
Figure: NMF,VQ,PCA on face images [1]
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 34: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/34.jpg)
iupui
Models
PLSILDANMFImplementations
Advantages-Disadvantages
Advantages
Part based representations , intuitive
Sparse encoding
Disadvantages
Di�erent viewpoints or articulated objects cannot be learned.
No learning of syntactic relation between parts.
K is �xed.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 35: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/35.jpg)
iupui
Models
PLSILDANMFImplementations
Advantages-Disadvantages
Advantages
Part based representations , intuitive
Sparse encoding
Disadvantages
Di�erent viewpoints or articulated objects cannot be learned.
No learning of syntactic relation between parts.
K is �xed.
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 36: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/36.jpg)
iupui
Models
PLSILDANMFImplementations
Outline
1 ModelsPLSILDANMFImplementations
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space
![Page 37: CSCI 590 : Machine Learning Lower Dimensional Latent ...mdundar/CSCIMachineLearning/Lecture16b.pdf · Halid Ziya erebakYan March 23 2015 Halid Ziya erebakYan CSCI 590 : Machine Learning](https://reader034.vdocuments.site/reader034/viewer/2022050408/5f84ddbaf986b566600cc492/html5/thumbnails/37.jpg)
iupui
Models
PLSILDANMFImplementations
Code
http://www.nltk.org/
Topic Model Toolbox
Mallet
Matlab nnmf
Blei's implementations
Halid Ziya Yerebakan CSCI 590 : Machine Learning Lower Dimensional Latent Semantic Space