Research#llm🏛️ OfficialAnalyzed: Jan 3, 2026 05:54

FACTS Grounding: A new benchmark for evaluating the factuality of large language models

Published:Dec 17, 2024 15:29
1 min read
DeepMind

Analysis

This article announces a new benchmark, FACTS Grounding, developed by DeepMind, designed to assess the accuracy of Large Language Models (LLMs) in grounding their responses in provided source material and avoiding hallucinations. The article highlights the importance of this benchmark by stating it offers a much-needed measure of LLM factuality.

Reference

Our comprehensive benchmark and online leaderboard offer a much-needed measure of how accurately LLMs ground their responses in provided source material and avoid hallucinations