Research#llm🏛️ OfficialAnalyzed: Jan 3, 2026 15:45

Generative modeling with sparse transformers

Published:Apr 23, 2019 07:00
1 min read
OpenAI News

Analysis

This article announces a new deep neural network, the Sparse Transformer, developed by OpenAI. The key innovation is an improvement to the attention mechanism, allowing it to process significantly longer sequences (30x) than previous models. This suggests advancements in handling complex patterns in data like text, images, and sound.

Reference

We’ve developed the Sparse Transformer, a deep neural network which sets new records at predicting what comes next in a sequence—whether text, images, or sound. It uses an algorithmic improvement of the attention mechanism to extract patterns from sequences 30x longer than possible previously.