Search:
Match:
1 results

Analysis

This paper addresses a crucial gap in evaluating multilingual LLMs. It highlights that high accuracy doesn't guarantee sound reasoning, especially in non-Latin scripts. The human-validated framework and error taxonomy are valuable contributions, emphasizing the need for reasoning-aware evaluation.
Reference

Reasoning traces in non-Latin scripts show at least twice as much misalignment between their reasoning and conclusions than those in Latin scripts.