Search:
Match:
2 results
Paper#llm🔬 ResearchAnalyzed: Jan 3, 2026 19:27

HiSciBench: A Hierarchical Benchmark for Scientific Intelligence

Published:Dec 28, 2025 12:08
1 min read
ArXiv

Analysis

This paper introduces HiSciBench, a novel benchmark designed to evaluate large language models (LLMs) and multimodal models on scientific reasoning. It addresses the limitations of existing benchmarks by providing a hierarchical and multi-disciplinary framework that mirrors the complete scientific workflow, from basic literacy to scientific discovery. The benchmark's comprehensive nature, including multimodal inputs and cross-lingual evaluation, allows for a detailed diagnosis of model capabilities across different stages of scientific reasoning. The evaluation of leading models reveals significant performance gaps, highlighting the challenges in achieving true scientific intelligence and providing actionable insights for future model development. The public release of the benchmark will facilitate further research in this area.
Reference

While models achieve up to 69% accuracy on basic literacy tasks, performance declines sharply to 25% on discovery-level challenges.

Research#VLM🔬 ResearchAnalyzed: Jan 10, 2026 07:38

VisRes Bench: Evaluating Visual Reasoning in VLMs

Published:Dec 24, 2025 14:18
1 min read
ArXiv

Analysis

This research introduces VisRes Bench, a benchmark for evaluating the visual reasoning capabilities of Vision-Language Models (VLMs). The study's focus on benchmarking is a crucial step in advancing VLM development and understanding their limitations.
Reference

VisRes Bench is a benchmark for evaluating the visual reasoning capabilities of VLMs.