Optimizing Social Media Annotation of HPV Vaccine Skepticism and Misinformation Using Large Language Models: An Experimental Evaluation of In-Context Learning and Fine-Tuning Stance Detection Across Multiple Models

📅 2024-11-22
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the low annotation efficiency and high labor cost associated with stance detection in HPV vaccine–related tweets on social media. We systematically compare in-context learning (ICL) against supervised fine-tuning across mainstream large language models (LLMs), including GPT-4, Mistral, and Llama3. Through ablation studies, we find—novelly for this task—that ICL consistently outperforms fine-tuning. The optimal configuration employs six hierarchically sampled, high-quality exemplars paired with comprehensive contextual prompts, achieving an F1 score of 0.82—4.7 percentage points higher than the fine-tuning baseline. We further uncover differential sensitivity of LLMs to ICL prompt structure. Our results demonstrate strong cross-model generalizability of the ICL approach and establish a reusable, low-cost, high-accuracy framework for annotating stances toward vaccine-related misinformation.

Technology Category

Application Category

📝 Abstract
This paper leverages large-language models (LLMs) to experimentally determine optimal strategies for scaling up social media content annotation for stance detection on HPV vaccine-related tweets. We examine both conventional fine-tuning and emergent in-context learning methods, systematically varying strategies of prompt engineering across widely used LLMs and their variants (e.g., GPT4, Mistral, and Llama3, etc.). Specifically, we varied prompt template design, shot sampling methods, and shot quantity to detect stance on HPV vaccination. Our findings reveal that 1) in general, in-context learning outperforms fine-tuning in stance detection for HPV vaccine social media content; 2) increasing shot quantity does not necessarily enhance performance across models; and 3) different LLMs and their variants present differing sensitivity to in-context learning conditions. We uncovered that the optimal in-context learning configuration for stance detection on HPV vaccine tweets involves six stratified shots paired with detailed contextual prompts. This study highlights the potential and provides an applicable approach for applying LLMs to research on social media stance and skepticism detection.
Problem

Research questions and friction points this paper is trying to address.

Optimizing stance detection on HPV vaccine tweets using LLMs
Comparing in-context learning and fine-tuning for misinformation detection
Evaluating prompt engineering strategies across multiple LLM variants
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses LLMs for HPV vaccine stance detection
Compares in-context learning and fine-tuning
Optimizes prompt design and shot quantity
🔎 Similar Papers
No similar papers found.
L
Luhang Sun
School of Journalism and Mass Communication, University of Wisconsin–Madison
V
Varsha Pendyala
Department of Electrical & Computer Engineering, University of Wisconsin–Madison
Y
Yun-Shiuan Chuang
Department of Computer Science, University of Wisconsin–Madison
S
Shanglin Yang
Department of Computer Science, University of Wisconsin–Madison
J
Jonathan Feldman
School of Interactive Computing, Georgia Institute of Technology
Andrew Zhao
Andrew Zhao
Tsinghua University
Reinforcement LearningLanguage AgentReasoning
Munmun De Choudhury
Munmun De Choudhury
Georgia Institute of Technology
Computational Social ScienceSocial ComputingMental HealthLanguage
Sijia Yang
Sijia Yang
Associate Professor, School of Journalism and Mass Communication, University of Wisconsin-Madison
Computational Communication ScienceMessage EffectsHealth CommunicationMoral psychology
D
Dhavan Shah
School of Journalism and Mass Communication, University of Wisconsin–Madison