Using LLMs and Websearch in Order to Perform Fact Checking on Texts Generated by LLMs

Publikation: Beitrag in Buch/Bericht/TagungsbandKonferenzbeitragBegutachtung

Abstract

Finding out if a given text contains any false information is not an easy task. On large corpora of data, such as the tremendous amount of texts generated by LLMs, fact checking is a prohibitively expensive task. To address this challenge, we propose a novel approach that combines fact checking by LLMs with web search. This method can be applied not only to single sentences by applying a true, false or unknown label. For whole text paragraphs, a 0..1 score representing the truthfulness from the sentence labels is calculated. The process begins by extracting claims from the text, which is done using GPT3. Then, these claims are validated, with Google search results being used to supplement the GPT3 results. When validating our method against a corpus of 122 LLM-generated text samples, we achieve an accuracy of 0.79. To compare our work to other approaches, we also applied our fact checking to the FEVER dataset, achieving an accuracy of 0.78. Which is similar than the current best accuracy of 0.79 on the FEVER dataset. This demonstrates the potential of our proposed approach for automated fact checking.
OriginalspracheEnglisch
TitelComputer Aided Systems Theory – EUROCAST 2024 - 19th International Conference, 2024, Revised Selected Papers
Redakteure/-innenAlexis Quesada-Arencibia, Michael Affenzeller, Roberto Moreno-Díaz
Seiten326-332
Seitenumfang7
Band15173
DOIs
PublikationsstatusVeröffentlicht - 25 Apr. 2025

Publikationsreihe

NameLecture Notes in Computer Science
Band15173 LNCS
ISSN (Print)0302-9743
ISSN (elektronisch)1611-3349

Zitieren