Papers
Topics
Authors
Recent
Search
2000 character limit reached

Direct Output Connection for a High-Rank Language Model

Published 30 Aug 2018 in cs.CL | (1808.10143v2)

Abstract: This paper proposes a state-of-the-art recurrent neural network (RNN) LLM that combines probability distributions computed not only from a final RNN layer but also from middle layers. Our proposed method raises the expressive power of a LLM based on the matrix factorization interpretation of language modeling introduced by Yang et al. (2018). The proposed method improves the current state-of-the-art LLM and achieves the best score on the Penn Treebank and WikiText-2, which are the standard benchmark datasets. Moreover, we indicate our proposed method contributes to two application tasks: machine translation and headline generation. Our code is publicly available at: https://github.com/nttcslab-nlp/doc_lm.

Citations (35)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

GitHub