Papers
Topics
Authors
Recent
Search
2000 character limit reached

Unsupervised Improvement of Factual Knowledge in Language Models

Published 4 Apr 2023 in cs.CL | (2304.01597v1)

Abstract: Masked language modeling (MLM) plays a key role in pretraining LLMs. But the MLM objective is often dominated by high-frequency words that are sub-optimal for learning factual knowledge. In this work, we propose an approach for influencing MLM pretraining in a way that can improve LLM performance on a variety of knowledge-intensive tasks. We force the LLM to prioritize informative words in a fully unsupervised way. Experiments demonstrate that the proposed approach can significantly improve the performance of pretrained LLMs on tasks such as factual recall, question answering, sentiment analysis, and natural language inference in a closed-book setting.

Citations (6)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.