Search:
Match:
3 results

DeepSeek's mHC: Improving Residual Connections

Published:Jan 2, 2026 15:44
1 min read
r/LocalLLaMA

Analysis

The article highlights DeepSeek's innovation in addressing the limitations of the standard residual connection in deep learning models. By introducing Manifold-Constrained Hyper-Connections (mHC), DeepSeek tackles the instability issues associated with previous attempts to make residual connections more flexible. The core of their solution lies in constraining the learnable matrices to be double stochastic, ensuring signal stability and preventing gradient explosion. The results demonstrate significant improvements in stability and performance compared to baseline models.
Reference

DeepSeek solved the instability by constraining the learnable matrices to be "Double Stochastic" (all elements ≧ 0, rows/cols sum to 1). Mathematically, this forces the operation to act as a weighted average (convex combination). It guarantees that signals are never amplified beyond control, regardless of network depth.

DeepSeek's mHC: Improving the Untouchable Backbone of Deep Learning

Published:Jan 2, 2026 15:40
1 min read
r/singularity

Analysis

The article highlights DeepSeek's innovation in addressing the limitations of residual connections in deep learning models. By introducing Manifold-Constrained Hyper-Connections (mHC), they've tackled the instability issues associated with flexible information routing, leading to significant improvements in stability and performance. The core of their solution lies in constraining the learnable matrices to be double stochastic, ensuring signals are not amplified uncontrollably. This represents a notable advancement in model architecture.
Reference

DeepSeek solved the instability by constraining the learnable matrices to be "Double Stochastic" (all elements ≧ 0, rows/cols sum to 1).

Analysis

This paper addresses the instability and scalability issues of Hyper-Connections (HC), a recent advancement in neural network architecture. HC, while improving performance, loses the identity mapping property of residual connections, leading to training difficulties. mHC proposes a solution by projecting the HC space onto a manifold, restoring the identity mapping and improving efficiency. This is significant because it offers a practical way to improve and scale HC-based models, potentially impacting the design of future foundational models.
Reference

mHC restores the identity mapping property while incorporating rigorous infrastructure optimization to ensure efficiency.