Latent Class Analysis
Latent Class Analysis assumes that people belong to different groups (i.e., classes) that are due to nonobservable characteristics. In a latent class model, we aim to estimate two kinds of probabilities:
- The probability that a person belongs to a particular class.
- The conditional probabilities of item responses if a person belongs to a given class.
The likelihood
Suppose that a sample of people respond to items and is a vector that contains the scores to each item . Let denote the number of latent classes and let be class . Then, the likelihood of the response pattern , if it was observed times in the sample and every person responds independently of each other, can be written as
Assuming local independence (responses within a person are independent conditional on class), we have where denotes the score on item . Hence, and the log-likelihood becomes The term inside the parenthesis is the probability of a single pattern, . Assuming independence between people with different response patterns, the log-likelihood of the whole sample is the sum of log-likelihoods of each response pattern.
To simplify the computation of the logarithm likelihood and related derivatives, let , so that
First-order derivatives
For a fixed pattern , define Then For a specific item and class ,
Notice that this last expression is just the posterior, , weighted by .
Second-order derivatives
For classes ,
For items and classes ,
For the mixed second derivative,
Collecting these terms gives the Hessian in block form:
Models for the conditional likelihoods
The conditional probabilities need to be parameterized with a likelihood. We consider a multinomial likelihood for categorical items and a Gaussian likelihood for continuous items.
Multinomial
For categorical items, let be the probability of scoring category on item if a subject belongs to class . Then where is such that . With this parameterization, and
Consequently, the Hessian for each conditional parameter has the following block form: where is a vector of zeroes with a 1 in the position corresponding to the parameter .
Notice that each conditional parameter has a Hessian matrix.
Gaussian
For continuous items, let denote the normal density. Let and be the mean and standard deviation for item in class . Then
First derivatives:
Second-order derivatives:
Consequently, the Hessian for each conditional parameter has the following block form:
Notice that each conditional parameter has a Hessian matrix.
Model for the latent class probabilities
Probabilities of class membership are parameterized with the softmax transformation:
where is the log-scale parameter associated with class .
The jacobian of this transformation is given by
Finally, the Hessian for each probability is
where is a vector of zeroes with a in position .
Model for the conditional probabilities of the multinomial model
Probabilities of conditional responses are parameterized with the softmax transformation:
where is the log-scale parameter associated with response to item in class .
The jacobian of this transformation is given by
Finally, the Hessian for each probability is
where is a vector of zeroes with a in position .
Constant priors
For latent class probabilities
For conditional likelihoods
Multinomial
For the conditional probabilities modeled with a multinomial likelihood, we add the following term to the log-likelihood for each class :
Where is the proportion of times category was selected in item .
The first-order derivatives are
The second-order derivatives are
Gaussian
For the conditional probabilities modeled with a gaussian likelihood, we add the following term to the log-likelihood for each class :
where is the variance of item .
The first-order derivatives are
The second-order derivatives are