Search:
Match:
1 results
Research#LLM🔬 ResearchAnalyzed: Jan 10, 2026 14:23

CoreEval: Enhancing LLM Reliability Through Contamination-Resilient Datasets

Published:Nov 24, 2025 08:44
1 min read
ArXiv

Analysis

This ArXiv paper introduces CoreEval, a method for creating datasets robust to contamination, crucial for reliable Large Language Model (LLM) evaluation. The work's focus on contamination resilience is a vital contribution to ensuring the validity of LLM performance assessments and mitigating biases.
Reference

CoreEval automatically builds contamination-resilient datasets.