Papers
Topics
Authors
Recent
Search
2000 character limit reached

Exploring Energy-based Language Models with Different Architectures and Training Methods for Speech Recognition

Published 22 May 2023 in cs.CL | (2305.12676v3)

Abstract: Energy-based LLMs (ELMs) parameterize an unnormalized distribution for natural sentences and are radically different from popular autoregressive LLMs (ALMs). As an important application, ELMs have been successfully used as a means for calculating sentence scores in speech recognition, but they all use less-modern CNN or LSTM networks. The recent progress in Transformer networks and large pretrained models such as BERT and GPT2 opens new possibility to further advancing ELMs. In this paper, we explore different architectures of energy functions and different training methods to investigate the capabilities of ELMs in rescoring for speech recognition, all using large pretrained models as backbones.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.