Reasoning-Answer Misalignment in Multilingual LLMs

Published:Dec 27, 2025 21:55
1 min read
ArXiv

Analysis

This paper addresses a crucial gap in evaluating multilingual LLMs. It highlights that high accuracy doesn't guarantee sound reasoning, especially in non-Latin scripts. The human-validated framework and error taxonomy are valuable contributions, emphasizing the need for reasoning-aware evaluation.

Reference

Reasoning traces in non-Latin scripts show at least twice as much misalignment between their reasoning and conclusions than those in Latin scripts.