Search:
Match:
1 results

Analysis

This paper addresses the redundancy in deep neural networks, where high-dimensional widths are used despite the low intrinsic dimension of the solution space. The authors propose a constructive approach to bypass the optimization bottleneck by decoupling the solution geometry from the ambient search space. This is significant because it could lead to more efficient and compact models without sacrificing performance, potentially enabling 'Train Big, Deploy Small' scenarios.
Reference

The classification head can be compressed by even huge factors of 16 with negligible performance degradation.