Search:
Match:
1 results

Analysis

This paper addresses limitations in video-to-audio generation by introducing a new task, EchoFoley, focused on fine-grained control over sound effects in videos. It proposes a novel framework, EchoVidia, and a new dataset, EchoFoley-6k, to improve controllability and perceptual quality compared to existing methods. The focus on event-level control and hierarchical semantics is a significant contribution to the field.
Reference

EchoVidia surpasses recent VT2A models by 40.7% in controllability and 12.5% in perceptual quality.