FACTS Grounding: A new benchmark for evaluating the factuality of large language models
Responsibility & Safety
Published
17 December 2024
Authors
Our comprehensive benchmark and online leaderboard offer a much-needed measure of how accurately LLMs ground their responses in provided source material and avoid hallucinationsLarge language models (LLMs) are transforming how we access information, yet their grip on factual accuracy remains imperfect. They can “hallucinate” false information, particularly when given complex inputs. In turn, this can erode trust in LLMs and limit their a...
Read more at deepmind.google