Search:
Match:
1 results

Analysis

This paper addresses the computational bottleneck of Transformer models in large-scale wireless communication, specifically power allocation. The proposed hybrid architecture offers a promising solution by combining a binary tree for feature compression and a Transformer for global representation, leading to improved scalability and efficiency. The focus on cell-free massive MIMO systems and the demonstration of near-optimal performance with reduced inference time are significant contributions.
Reference

The model achieves logarithmic depth and linear total complexity, enabling efficient inference across large and variable user sets without retraining or architectural changes.