Search:
Match:
1 results

Analysis

This paper addresses the challenge of designing multimodal deep neural networks (DNNs) using Neural Architecture Search (NAS) when labeled data is scarce. It proposes a self-supervised learning (SSL) approach to overcome this limitation, enabling architecture search and model pretraining from unlabeled data. This is significant because it reduces the reliance on expensive labeled data, making NAS more accessible for complex multimodal tasks.
Reference

The proposed method applies SSL comprehensively for both the architecture search and model pretraining processes.