BandiK: Efficient Multi-Task Learning with Multi-Bandits

Published:Dec 31, 2025 08:25
1 min read
ArXiv

Analysis

This paper addresses the challenge of efficient auxiliary task selection in multi-task learning, a crucial aspect of knowledge transfer, especially relevant in the context of foundation models. The core contribution is BandiK, a novel method using a multi-bandit framework to overcome the computational and combinatorial challenges of identifying beneficial auxiliary task sets. The paper's significance lies in its potential to improve the efficiency and effectiveness of multi-task learning, leading to better knowledge transfer and potentially improved performance in downstream tasks.

Reference

BandiK employs a Multi-Armed Bandit (MAB) framework for each task, where the arms correspond to the performance of candidate auxiliary sets realized as multiple output neural networks over train-test data set splits.