Research#llm🔬 ResearchAnalyzed: Dec 25, 2025 02:58

Learning to Refocus with Video Diffusion Models

Published:Dec 24, 2025 05:00
1 min read
ArXiv Vision

Analysis

This paper introduces a novel approach to post-capture refocusing using video diffusion models. The method generates a realistic focal stack from a single defocused image, enabling interactive refocusing. A key contribution is the release of a large-scale focal stack dataset acquired under real-world smartphone conditions. The method demonstrates superior performance compared to existing approaches in perceptual quality and robustness. The availability of code and data enhances reproducibility and facilitates further research in this area. The research has significant potential for improving focus-editing capabilities in everyday photography and opens avenues for advanced image manipulation techniques. The use of video diffusion models for this task is innovative and promising.

Reference

From a single defocused image, our approach generates a perceptually accurate focal stack, represented as a video sequence, enabling interactive refocusing.