Coding latent concepts: a human and LLM-coordinated content analysis procedure

Jia Fan*, Yushi Ai, Xiaofan Liu, Yilin Deng, Yongning Li

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

2 Citations (Scopus)

Abstract

Measuring complex and latent concepts at a large-scale poses significant challenges for communication researchers. While computational and crowdsourced methods offer solutions, they often require high professional thresholds or incur significant costs. The recent advent of large language models has revolutionized content analysis. This paper employs a human and LLM-coordinated analysis procedure to measure complex and latent concepts in 1,000 public comments, exemplified by the multi-dimension concept of “deliberativeness.” We showcase the collaboration between humans and LLMs in completing complex coding tasks by designing and refining a codebook for human use and corresponding prompts for LLMs. Surprisingly, we find that fine-tuned GPT-3.5-turbo-1106 with smaller datasets can surpass GPT-4o-2024-05-13’s performance and match manual content analyses. This paper provides communication researchers with an efficient and cost-effective reference for measuring latent concepts. © 2024 Eastern Communication Association

Original languageEnglish
Pages (from-to)324-334
Number of pages11
JournalCommunication Research Reports
Volume41
Issue number5
DOIs
Publication statusOnline published - 3 Oct 2024

Bibliographical note

Publisher Copyright:
© 2024 Eastern Communication Association.

Research Keywords

  • complex and latent concepts
  • Content analysis
  • large language model

Fingerprint

Dive into the research topics of 'Coding latent concepts: a human and LLM-coordinated content analysis procedure'. Together they form a unique fingerprint.

Cite this