PulseAugur
LIVE 10:45:11
research · [2 sources] ·
0
research

New HATS dataset integrates human perception for ASR evaluation

Researchers have introduced HATS, a new French dataset designed to evaluate Automatic Speech Recognition (ASR) systems by incorporating human perception. The dataset was created by having 143 individuals compare and select the better transcription from two options generated by different ASR systems. This effort aims to move beyond traditional metrics like Word Error Rate (WER), which are considered insufficient for assessing ASR quality from a human user's perspective. AI

Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →

IMPACT Introduces a new dataset for evaluating ASR systems, potentially leading to more human-aligned transcription quality assessments.

RANK_REASON The cluster describes a new academic paper introducing a novel dataset for ASR evaluation.

Read on arXiv cs.CL →

COVERAGE [2]

  1. arXiv cs.CL TIER_1 · Thibault Ba\~neras Roux, Jane Wottawa, Mickael Rouvier, Teva Merlin, Richard Dufour ·

    HATS: An Open data set Integrating Human Perception Applied to the Evaluation of Automatic Speech Recognition Metrics

    arXiv:2604.27542v1 Announce Type: new Abstract: Conventionally, Automatic Speech Recognition (ASR) systems are evaluated on their ability to correctly recognize each word contained in a speech signal. In this context, the word error rate (WER) metric is the reference for evaluati…

  2. arXiv cs.CL TIER_1 · Richard Dufour ·

    HATS: An Open data set Integrating Human Perception Applied to the Evaluation of Automatic Speech Recognition Metrics

    Conventionally, Automatic Speech Recognition (ASR) systems are evaluated on their ability to correctly recognize each word contained in a speech signal. In this context, the word error rate (WER) metric is the reference for evaluating speech transcripts. Several studies have show…