Evaluating topic model interpretability from a primary care physician perspective

Comput Methods Programs Biomed. 2016 Feb:124:67-75. doi: 10.1016/j.cmpb.2015.10.014. Epub 2015 Oct 30.

Abstract

Background and objective: Probabilistic topic models provide an unsupervised method for analyzing unstructured text. These models discover semantically coherent combinations of words (topics) that could be integrated in a clinical automatic summarization system for primary care physicians performing chart review. However, the human interpretability of topics discovered from clinical reports is unknown. Our objective is to assess the coherence of topics and their ability to represent the contents of clinical reports from a primary care physician's point of view.

Methods: Three latent Dirichlet allocation models (50 topics, 100 topics, and 150 topics) were fit to a large collection of clinical reports. Topics were manually evaluated by primary care physicians and graduate students. Wilcoxon Signed-Rank Tests for Paired Samples were used to evaluate differences between different topic models, while differences in performance between students and primary care physicians (PCPs) were tested using Mann-Whitney U tests for each of the tasks.

Results: While the 150-topic model produced the best log likelihood, participants were most accurate at identifying words that did not belong in topics learned by the 100-topic model, suggesting that 100 topics provides better relative granularity of discovered semantic themes for the data set used in this study. Models were comparable in their ability to represent the contents of documents. Primary care physicians significantly outperformed students in both tasks.

Conclusion: This work establishes a baseline of interpretability for topic models trained with clinical reports, and provides insights on the appropriateness of using topic models for informatics applications. Our results indicate that PCPs find discovered topics more coherent and representative of clinical reports relative to students, warranting further research into their use for automatic summarization.

Keywords: Clinical reports; Primary care; Topic modeling.

Publication types

  • Evaluation Study
  • Research Support, N.I.H., Extramural

MeSH terms

  • Attitude of Health Personnel*
  • Comprehension*
  • Electronic Health Records / classification*
  • Electronic Health Records / organization & administration
  • Information Storage and Retrieval / methods
  • Meaningful Use / statistics & numerical data
  • Medical Record Linkage / methods
  • Models, Statistical
  • Natural Language Processing
  • Physicians, Primary Care / statistics & numerical data*
  • Semantics
  • Terminology as Topic*
  • Vocabulary, Controlled*