PulseAugur
LIVE 06:30:13
research · [1 source] ·
0
research

New benchmark UNIKIE-BENCH evaluates large multimodal models for document information extraction

Researchers have introduced UNIKIE-BENCH, a new benchmark designed to systematically evaluate the performance of Large Multimodal Models (LMMs) in extracting key information from visual documents. The benchmark features two tracks: one for constrained-category KIE with predefined schemas and another for open-category KIE. Experiments using 15 state-of-the-art LMMs highlighted significant performance drops when dealing with varied schemas, long-tail information, and complex layouts, indicating ongoing challenges in accuracy and reasoning for LMMs in this domain. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT Provides a standardized evaluation framework for LMMs in document information extraction, highlighting current limitations.

RANK_REASON This is a research paper introducing a new benchmark for evaluating LMMs.

Read on arXiv cs.CV →

COVERAGE [1]

  1. arXiv cs.CV TIER_1 · Yifan Ji, Zhipeng Xu, Zhenghao Liu, Zulong Chen, Qian Zhang, Zhibo Yang, Junyang Lin, Yu Gu, Ge Yu, Maosong Sun ·

    UNIKIE-BENCH: Benchmarking Large Multimodal Models for Key Information Extraction in Visual Documents

    arXiv:2602.07038v2 Announce Type: replace Abstract: Key Information Extraction (KIE) from real-world documents remains challenging due to substantial variations in layout structures, visual quality, and task-specific information requirements. Recent Large Multimodal Models (LMMs)…