Continual Learning for LLMs: Merge Before Forgetting with LoRA
Analysis
Key Takeaways
- •Proposes a novel continual learning method for LLMs using LoRA.
- •Employs orthogonal initialization and time-aware scaling for merging LoRAs.
- •Aims to improve memory efficiency and reduce task interference.
- •Maintains constant memory complexity with respect to the number of tasks.
“The method leverages orthogonal basis extraction from previously learned LoRA to initialize the learning of new tasks, further exploits the intrinsic asymmetry property of LoRA components by using a time-aware scaling mechanism to balance new and old knowledge during continual merging.”