Search:
Match:
1 results

Context Reduction in Language Model Probabilities

Published:Dec 29, 2025 18:12
1 min read
ArXiv

Analysis

This paper investigates the minimal context required to observe probabilistic reduction in language models, a phenomenon relevant to cognitive science. It challenges the assumption that whole utterances are necessary, suggesting that n-gram representations are sufficient. This has implications for understanding how language models relate to human cognitive processes and could lead to more efficient model analysis.
Reference

n-gram representations suffice as cognitive units of planning.