一 模型含义
With plate notation, the dependencies among the many variables can be captured concisely. The boxes are “plates” representing replicates. The outer plate represents documents, while the inner plate represents the repeated choice of topics and words within a document. M denotes the number of documents, N the number of words in a document. Thus:
- α is the parameter of the Dirichlet prior on the per-document topic distributions.
- β is the parameter of the Dirichlet prior on the per-topic word distribution.
-
is the topic distribution for document i,
-
is the word distribution for topic k,
-
is the topic for the jth word in document i, and
-
is the specific word.
The are the only observable variables, and the other variables are latent variables. Mostly, the basic LDA model will be extended to a smoothed version to gain better results. The plate notation is shown on the right, where K denotes the number of topics considered in the model and:
-
is a K*V ( V is the dimension of the vocabulary) Markov matrix each row of which denotes the word distribution of a topic.
The generative process behind is that documents are represented as random mixtures over latent topics, where each topic is characterized by a distribution over words. LDA assumes the following generative process for each document in a corpus D :
1. Choose , where
and
is the Dirichlet distribution for parameter
2. Choose , where
3. For each of the words , where
-
(a) Choose a topic
-
(b) Choose a word
.
(Note that the Multinomial distribution here refers to the Multinomial with only one trial. It is formally equivalent to the categorical distribution.)
The lengths are treated as independent of all the other data generating variables (
and
). The subscript is often dropped, as in the plate diagrams shown here.
二 数学定义
A formal description of smoothed LDA is as follows:
Variable | Type | Meaning |
---|---|---|
![]() | integer | number of topics (e.g. 50) |
![]() | integer | number of words in the vocabulary (e.g. 50,000 or 1,000,000) |
![]() | integer | number of documents |
![]() | integer | number of words in document d |
![]() | integer | total number of words in all documents; sum of all ![]() ![]() |
![]() | positive real | prior weight of topic k in a document; usually the same for all topics; normally a number less than 1, e.g. 0.1, to prefer sparse topic distributions, i.e. few topics per document |
![]() | K-dimension vector of positive reals | collection of all ![]() |
![]() | positive real | prior weight of word w in a topic; usually the same for all words; normally a number much less than 1, e.g. 0.001, to strongly prefer sparse word distributions, i.e. few words per topic |
![]() | V-dimension vector of positive reals | collection of all ![]() |
![]() | probability (real number between 0 and 1) | probability of word w occurring in topic k |
![]() | V-dimension vector of probabilities, which must sum to 1 | distribution of words in topic k |
![]() | probability (real number between 0 and 1) | probability of topic k occurring in document d for a given word |
![]() | K-dimension vector of probabilities, which must sum to 1 | distribution of topics in document d |
![]() | integer between 1 and K | identity of topic of word w in document d |
![]() | N-dimension vector of integers between 1 and K | identity of topic of all words in all documents |
![]() | integer between 1 and V | identity of word w in document d |
![]() | N-dimension vector of integers between 1 and V | identity of all words in all documents |
We can then mathematically describe the random variables as follows: