Login (DCU Staff Only)
Login (DCU Staff Only)

DORAS | DCU Research Repository

Explore open access research and scholarly works from DCU

Advanced Search

Tell Me Why: Explainable Public Health Fact-Checking with Large Language Models

Zarharan, Majid, Wullschleger, Pascal, Kia, Babak Behkam, Pilehvar, Mohammad Taher and Foster, Jennifer orcid logoORCID: 0000-0002-7789-4853 (2024) Tell Me Why: Explainable Public Health Fact-Checking with Large Language Models. In: 4th Workshop on Trustworthy NLP (TrustNLP 2024).

Abstract
This paper presents a comprehensive analysis of explainable fact-checking through a series of experiments, focusing on the ability of large language models to verify public health claims and provide explanations or justifications for their veracity assessments. We examine the effectiveness of zero/few-shot prompting and parameter-efficient fine-tuning across various open and closed-source models, examining their performance in both isolated and joint tasks of veracity prediction and explanation generation. Importantly, we employ a dual evaluation approach comprising previously established automatic metrics and a novel set of criteria through human evaluation. Our automatic evaluation indicates that, within the zero-shot scenario, GPT-4 emerges as the standout performer, but in few-shot and parameter-efficient fine-tuning contexts, open source models demonstrate their capacity to not only bridge the performance gap but, in some instances, surpass GPT-4. Human evaluation reveals yet more nuance as well as indicating potential problems with the gold explanations.
Metadata
Item Type:Conference or Workshop Item (Paper)
Event Type:Workshop
Refereed:Yes
Subjects:Computer Science > Artificial intelligence
Computer Science > Machine learning
DCU Faculties and Centres:DCU Faculties and Schools > Faculty of Engineering and Computing
DCU Faculties and Schools > Faculty of Engineering and Computing > School of Computing
Research Institutes and Centres > ADAPT
Published in: Proceedings of the 4th Workshop on Trustworthy NLP (TrustNLP 2024). . Association for Computational Linguistics.
Publisher:Association for Computational Linguistics
Official URL:https://aclanthology.org/2024.trustnlp-1.21.pdf
Funders:Science Foundation Ireland
ID Code:30561
Deposited On:10 Dec 2024 10:48 by Jennifer Foster . Last Modified 10 Dec 2024 10:48
Documents

Full text available as:

[thumbnail of 2024.trustnlp-1.21 (1).pdf]
Preview
PDF - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
Creative Commons: Attribution 4.0
554kB
Metrics

Altmetric Badge

Dimensions Badge

Downloads

Downloads

Downloads per month over past year

Archive Staff Only: edit this record