Journal article
Transactions of the Association for Computational Linguistics, 2022
APA
Click to copy
Sridhar, D., Daumé, H., & Blei, D. (2022). Heterogeneous Supervised Topic Models. Transactions of the Association for Computational Linguistics.
Chicago/Turabian
Click to copy
Sridhar, Dhanya, Hal Daumé, and D. Blei. “Heterogeneous Supervised Topic Models.” Transactions of the Association for Computational Linguistics (2022).
MLA
Click to copy
Sridhar, Dhanya, et al. “Heterogeneous Supervised Topic Models.” Transactions of the Association for Computational Linguistics, 2022.
BibTeX Click to copy
@article{dhanya2022a,
title = {Heterogeneous Supervised Topic Models},
year = {2022},
journal = {Transactions of the Association for Computational Linguistics},
author = {Sridhar, Dhanya and Daumé, Hal and Blei, D.}
}
Abstract Researchers in the social sciences are often interested in the relationship between text and an outcome of interest, where the goal is to both uncover latent patterns in the text and predict outcomes for unseen texts. To this end, this paper develops the heterogeneous supervised topic model (HSTM), a probabilistic approach to text analysis and prediction. HSTMs posit a joint model of text and outcomes to find heterogeneous patterns that help with both text analysis and prediction. The main benefit of HSTMs is that they capture heterogeneity in the relationship between text and the outcome across latent topics. To fit HSTMs, we develop a variational inference algorithm based on the auto-encoding variational Bayes framework. We study the performance of HSTMs on eight datasets and find that they consistently outperform related methods, including fine-tuned black-box models. Finally, we apply HSTMs to analyze news articles labeled with pro- or anti-tone. We find evidence of differing language used to signal a pro- and anti-tone.