Researchers have developed CNSL-bench, a new benchmark designed to evaluate the sign language understanding capabilities of multimodal large language models (MLLMs). This benchmark is grounded in the official Chinese National Sign Language Dictionary and includes aligned textual descriptions, images, and videos, covering diverse articulatory forms. Evaluations of 21 MLLMs using CNSL-bench revealed that current models significantly underperform human capabilities, with notable disparities across different input modalities and articulation types. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
IMPACT Establishes a new evaluation standard for MLLMs in sign language, highlighting current performance gaps compared to human understanding.
RANK_REASON Academic paper introducing a new benchmark for evaluating MLLMs on sign language understanding.