Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Truth-O-Meter: Collaborating with LLM in Fighting its Hallucinations

Version 1 : Received: 24 July 2023 / Approved: 25 July 2023 / Online: 25 July 2023 (13:09:30 CEST)

How to cite: Galitsky, B.A. Truth-O-Meter: Collaborating with LLM in Fighting its Hallucinations. Preprints 2023, 2023071723. https://doi.org/10.20944/preprints202307.1723.v1 Galitsky, B.A. Truth-O-Meter: Collaborating with LLM in Fighting its Hallucinations. Preprints 2023, 2023071723. https://doi.org/10.20944/preprints202307.1723.v1

Abstract

A text obtained by a Large Language Model (LLM) such as GPT4 usually has issues in terms of incorrectness and hallucinations. We build a fact-checking system 'Truth-O-Meter' which identifies wrong facts, comparing the generation results with the web and other sources of information, and suggests corrections. Text mining and web mining techniques are leveraged to identify correct corresponding sentences; also, the syntactic and semantic generalization procedure adopted to the content improvement task. To handle inconsistent sources while fact-checking, we rely on an argumentation analysis in the form of defeasible logic programming. We compare our fact checking engine with competitive approach based on reinforcement learning on top of LLM or token-based hallucination detection. It is observed that LLM content can be substantially improved for factual correctness and meaningfulness.

Keywords

Large Language Model; hallucination; fact-checking; multiple inconsistent sources

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.