SM

S. Mukherjee

5 records found

Explainable Fact-Checking with Large Language Models

How Prompt Style Variation affects Accuracy and Faithfulness in Claim Justifications

Large Language Models (LLMs) such as GPT-4 and LLaMA have demonstrated promising performance in fact-checking tasks, particularly in labeling the veracity of claims. However, the real-world utility of such fact-checking systems depends not only on label accuracy but also on the f ...

Explainable Fact-Checking with LLMs

How do different LLMs compare in their rationales?

Large Language Models (LLMs) are becoming more commonplace in today's society. However their adoption rate, especially in the fact checking field, is being slowed down by the distrust in their thinking process and the rationales leading to the results. In crucial moments the just ...

Evaluating Faithfulness of LLM Generated Explanations for Claims: Are Current Metrics Effective?

Analysing the Capabilities of Evaluation Metrics to Represent the Difference Between Generated and Expert-written Explanations

Large Language Models (LLMs) are increasingly used to generate fact-checking explanations, but evaluating how faithful these justifications are remains a major challenge. In this paper, we examine how well four popular automatic metrics—G-Eval, UniEval, FactCC, and QAGs—capture f ...
Large Language Models (LLMs) are increasingly transforming how scientists approach research, with emerging tools supporting ideation, experimentation, and publication in attempts to expedite the research process. This work focuses on the foundational first step: generating novel, ...
We investigate the application of Retrieval-Augmented Generation (RAG) for enhancing the analysis of corporate sustainability disclosures. We introduce CorSus, a novel dataset for evaluating RAG models in answering corporate sustainability-focused claims, using data from the Tran ...