Transformer-based language modeling and decoding for conversational speech recognition
Abstract
A transformer-based language model is integrated into conversational speech recognition for efficient decoding and re-scoring of lattice structures, capturing longer-range history and parallel computation benefits.
We propose a way to use a transformer-based language model in conversational speech recognition. Specifically, we focus on decoding efficiently in a weighted finite-state transducer framework. We showcase an approach to lattice re-scoring that allows for longer range history captured by a transfomer-based language model and takes advantage of a transformer's ability to avoid computing sequentially.
Get this paper in your agent:
hf papers read 2001.01140 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper