Search:
Match:
3 results

Hierarchical VQ-VAE for Low-Resolution Video Compression

Published:Dec 31, 2025 01:07
1 min read
ArXiv

Analysis

This paper addresses the growing need for efficient video compression, particularly for edge devices and content delivery networks. It proposes a novel Multi-Scale Vector Quantized Variational Autoencoder (MS-VQ-VAE) that generates compact, high-fidelity latent representations of low-resolution video. The use of a hierarchical latent structure and perceptual loss is key to achieving good compression while maintaining perceptual quality. The lightweight nature of the model makes it suitable for resource-constrained environments.
Reference

The model achieves 25.96 dB PSNR and 0.8375 SSIM on the test set, demonstrating its effectiveness in compressing low-resolution video while maintaining good perceptual quality.

Analysis

This paper addresses the challenge of compressing multispectral solar imagery for space missions, where bandwidth is limited. It introduces a novel learned image compression framework that leverages graph learning techniques to model both inter-band spectral relationships and spatial redundancy. The use of Inter-Spectral Windowed Graph Embedding (iSWGE) and Windowed Spatial Graph Attention and Convolutional Block Attention (WSGA-C) modules is a key innovation. The results demonstrate significant improvements in spectral fidelity and reconstruction quality compared to existing methods, making it relevant for space-based solar observations.
Reference

The approach achieves a 20.15% reduction in Mean Spectral Information Divergence (MSID), up to 1.09% PSNR improvement, and a 1.62% log transformed MS-SSIM gain over strong learned baselines.

Analysis

This paper addresses the critical need for real-time, high-resolution video prediction in autonomous UAVs, a domain where latency is paramount. The authors introduce RAPTOR, a novel architecture designed to overcome the limitations of existing methods that struggle with speed and resolution. The core innovation, Efficient Video Attention (EVA), allows for efficient spatiotemporal modeling, enabling real-time performance on edge hardware. The paper's significance lies in its potential to improve the safety and performance of UAVs in complex environments by enabling them to anticipate future events.
Reference

RAPTOR is the first predictor to exceed 30 FPS on a Jetson AGX Orin for $512^2$ video, setting a new state-of-the-art on UAVid, KTH, and a custom high-resolution dataset in PSNR, SSIM, and LPIPS. Critically, RAPTOR boosts the mission success rate in a real-world UAV navigation task by 18%.