Speeding Up Generative AI: Anthropic and OpenAI's Fast Mode Innovations

infrastructure#llm👥 Community|Analyzed: Feb 15, 2026 13:02
Published: Feb 15, 2026 09:27
1 min read
Hacker News

Analysis

Exciting developments in Large Language Model (LLM) inference speeds are here! Both Anthropic and OpenAI have unveiled "fast mode" options, promising substantial boosts in token processing. This innovation could revolutionize how we interact with Generative AI models.
Reference / Citation
View Original
"Anthropic’s offers up to 2.5x tokens per second (so around 170, up from Opus 4.6’s 65). OpenAI’s offers more than 1000 tokens per second (up from GPT-5.3-Codex’s 65 tokens per second, so 15x)."
H
Hacker NewsFeb 15, 2026 09:27
* Cited for critical analysis under Article 32.