Gemini's Memory Issues: User Reports Limited Context Retention
Published:Dec 29, 2025 05:44
•1 min read
•r/Bard
Analysis
This news item, sourced from a Reddit post, highlights a potential issue with Google's Gemini AI model regarding its ability to retain context in long conversations. A user reports that Gemini only remembered the last 14,000 tokens of a 117,000-token chat, a significant limitation. This raises concerns about the model's suitability for tasks requiring extensive context, such as summarizing long documents or engaging in extended dialogues. The user's uncertainty about whether this is a bug or a typical limitation underscores the need for clearer documentation from Google regarding Gemini's context window and memory management capabilities. Further investigation and user reports are needed to determine the prevalence and severity of this issue.
Key Takeaways
- •Gemini may have limitations in retaining context in long conversations.
- •The reported context window is significantly smaller than the total conversation length.
- •Users should be aware of potential memory limitations when using Gemini for tasks requiring extensive context.
Reference
“Until I asked Gemini (a 3 Pro Gem) to summarize our conversation so far, and they only remembered the last 14k tokens. Out of our entire 117k chat.”