Transactions of the Association for Computational Linguistics • Vol 13
LLM Reading Tea Leaves: Automatically Evaluating Topic Models with Large Language Models
January 2025 • Xiaohao Yang, He Zhao, Dinh Phung, Wray Buntine, Lan Du
Abstract Topic modeling has been a widely used tool for unsupervised text analysis. However, comprehensive evaluations of a topic model remain challenging. Existing evaluation methods are either less comparable across different models (e.g., perplexity) or focus on only one specific aspect of a model (e.g., topic quality or document representation quality) at a time, which is insufficient to reflect the overall model performance. In this paper, we propose WALM (Word Agreement with Language Model), a new evaluation…