DarkEQA: Benchmarking VLMs for Low-Light Embodied Question Answering

Paper#llm🔬 Research|Analyzed: Jan 3, 2026 06:16
Published: Dec 31, 2025 17:31
1 min read
ArXiv

Analysis

This paper addresses a critical gap in the evaluation of Vision-Language Models (VLMs) for embodied agents. Existing benchmarks often overlook the performance of VLMs under low-light conditions, which are crucial for real-world, 24/7 operation. DarkEQA provides a novel benchmark to assess VLM robustness in these challenging environments, focusing on perceptual primitives and using a physically-realistic simulation of low-light degradation. This allows for a more accurate understanding of VLM limitations and potential improvements.
Reference / Citation
View Original
"DarkEQA isolates the perception bottleneck by evaluating question answering from egocentric observations under controlled degradations, enabling attributable robustness analysis."
A
ArXivDec 31, 2025 17:31
* Cited for critical analysis under Article 32.