PulseAugur
LIVE 14:40:58
research · [2 sources] ·
0
research

AppTek releases multi-accent ASR benchmark for conversational AI

Researchers have introduced the AppTek Call-Center Dialogues corpus, a new benchmark designed to evaluate English Automatic Speech Recognition (ASR) systems. This dataset features spontaneous, role-played conversations from fourteen different English accents across sixteen service scenarios. The corpus aims to address the limitations of existing datasets, which often use segmented or read speech and lack diverse accent annotations, crucial for developing robust conversational AI. AI

Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →

IMPACT Provides a new benchmark for evaluating ASR robustness across diverse English accents, crucial for improving conversational AI.

RANK_REASON This is a research paper introducing a new dataset and benchmark for ASR systems.

Read on arXiv cs.CL →

COVERAGE [2]

  1. arXiv cs.CL TIER_1 · Eugen Beck, Sarah Beranek, Uma Moothiringote, Daniel Mann, Wilfried Michel, Katie Nguyen, Taylor Tragemann ·

    AppTek Call-Center Dialogues: A Multi-Accent Long-Form Benchmark for English ASR

    arXiv:2604.27543v1 Announce Type: new Abstract: Evaluating English ASR systems for conversational AI applications remains difficult, as many publicly available corpora are either pre-segmented into short segments, consist of read or prepared speech, or lack explicit dialect annot…

  2. arXiv cs.CL TIER_1 · Taylor Tragemann ·

    AppTek Call-Center Dialogues: A Multi-Accent Long-Form Benchmark for English ASR

    Evaluating English ASR systems for conversational AI applications remains difficult, as many publicly available corpora are either pre-segmented into short segments, consist of read or prepared speech, or lack explicit dialect annotations to evaluate robustness for a diverse user…