AnyMS: Training-Free Multi-Subject Customization with Layout Guidance

Published:Dec 29, 2025 15:26
1 min read
ArXiv

Analysis

This paper introduces AnyMS, a novel training-free framework for multi-subject image synthesis. It addresses the challenges of text alignment, subject identity preservation, and layout control by using a bottom-up dual-level attention decoupling mechanism. The key innovation is the ability to achieve high-quality results without requiring additional training, making it more scalable and efficient than existing methods. The use of pre-trained image adapters further enhances its practicality.

Reference

AnyMS leverages a bottom-up dual-level attention decoupling mechanism to harmonize the integration of text prompt, subject images, and layout constraints.