Detail publikačního výsledku

Unifying Global and Near-Context Biasing in a Single Trie Pass

THORBECKE, I.; VILLATORO-TELLO, E.; ZULUAGA, J.; KUMAR, S.; BURDISSO, S.; RANGAPPA, P.; CAROFILIS, A.; MADIKERI, S.; MOTLÍČEK, P.; PANDIA, K.; HACIOGLU, K.; STOLCKE, A.

Originální název

Unifying Global and Near-Context Biasing in a Single Trie Pass

Anglický název

Unifying Global and Near-Context Biasing in a Single Trie Pass

Druh

Stať ve sborníku v databázi WoS či Scopus

Originální abstrakt

Despite the success of end-to-end automatic speech recognition (ASR) models, challenges persist in recognizing rare, out-of-vocabulary wordsincluding named entities (NE)-and in adapting to new domains using only text data. This work presents a practical approach to address these challenges through an unexplored combination of an NE bias list and a word-level n-gram language model (LM). This solution balances simplicity and effectiveness, improving entities' recognition while maintaining or even enhancing overall ASR performance. We efficiently integrate this enriched biasing method into a transducer-based ASR system, enabling context adaptation with almost no computational overhead. We present our results on three datasets spanning four languages and compare them to state-of-the-art biasing strategies We demonstrate that the proposed combination of keyword biasing and n-gram LM improves entity recognition by up to 32% relative and reduces overall WER by up to a 12% relative.

Anglický abstrakt

Despite the success of end-to-end automatic speech recognition (ASR) models, challenges persist in recognizing rare, out-of-vocabulary wordsincluding named entities (NE)-and in adapting to new domains using only text data. This work presents a practical approach to address these challenges through an unexplored combination of an NE bias list and a word-level n-gram language model (LM). This solution balances simplicity and effectiveness, improving entities' recognition while maintaining or even enhancing overall ASR performance. We efficiently integrate this enriched biasing method into a transducer-based ASR system, enabling context adaptation with almost no computational overhead. We present our results on three datasets spanning four languages and compare them to state-of-the-art biasing strategies We demonstrate that the proposed combination of keyword biasing and n-gram LM improves entity recognition by up to 32% relative and reduces overall WER by up to a 12% relative.

Klíčová slova

Contextualisation and adaptation of ASR, real-time ASR, Aho-Corasick algorithm, Transformer-Transducer

Klíčová slova v angličtině

Contextualisation and adaptation of ASR, real-time ASR, Aho-Corasick algorithm, Transformer-Transducer

Autoři

THORBECKE, I.; VILLATORO-TELLO, E.; ZULUAGA, J.; KUMAR, S.; BURDISSO, S.; RANGAPPA, P.; CAROFILIS, A.; MADIKERI, S.; MOTLÍČEK, P.; PANDIA, K.; HACIOGLU, K.; STOLCKE, A.

Rok RIV

2026

Vydáno

25.08.2026

Nakladatel

Springer Nature

Místo

CHAM

ISBN

978-3-032-02547-0

Kniha

Lecture Notes in Artificial Intelligence

Periodikum

Lecture Notes in Computer Science

Svazek

16029

Stát

Švýcarská konfederace

Strany od

170

Strany do

181

Strany počet

12

URL

BibTex

@inproceedings{BUT201441,
  author="{} and  {} and  {} and  {} and  {} and  {} and  {} and  {} and Petr {Motlíček} and  {} and  {} and  {}",
  title="Unifying Global and Near-Context Biasing in a Single Trie Pass",
  booktitle="Lecture Notes in Artificial Intelligence",
  year="2026",
  journal="Lecture Notes in Computer Science",
  volume="16029",
  pages="170--181",
  publisher="Springer Nature",
  address="CHAM",
  doi="10.1007/978-3-032-02548-7\{_}15",
  isbn="978-3-032-02547-0",
  url="https://www.fit.vut.cz/research/group/speech/public/publi/2025/Iuliia_TSD2025_2025_co-author_Motlicek.pdf"
}