Automatic Question-Answer Generation for Long-Tail Knowledge
Abstract: Pretrained LLMs have gained significant attention for addressing open-domain Question Answering (QA). While they exhibit high accuracy in answering questions related to common knowledge, LLMs encounter difficulties in learning about uncommon long-tail knowledge (tail entities). Since manually constructing QA datasets demands substantial human resources, the types of existing QA datasets are limited, leaving us with a scarcity of datasets to study the performance of LLMs on tail entities. In this paper, we propose an automatic approach to generate specialized QA datasets for tail entities and present the associated research challenges. We conduct extensive experiments by employing pretrained LLMs on our newly generated long-tail QA datasets, comparing their performance with and without external resources including Wikipedia and Wikidata knowledge graphs.
- Semantic parsing on freebase from question-answer pairs. In Proceedings of the 2013 conference on empirical methods in natural language processing. 1533–1544.
- Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
- Searchqa: A new q&a dataset augmented with context from a search engine. arXiv preprint arXiv:1704.05179 (2017).
- Aaron Gokaslan and Vanya Cohen. 2019. Openwebtext corpus.
- Survey of hallucination in natural language generation. Comput. Surveys 55, 12 (2023), 1–38.
- Triviaqa: A large scale distantly supervised challenge dataset for reading comprehension. arXiv preprint arXiv:1705.03551 (2017).
- Large language models struggle to learn long-tail knowledge. arXiv preprint arXiv:2211.08411 (2022).
- Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906 (2020).
- Are pretrained language models symbolic reasoners over knowledge? arXiv preprint arXiv:2006.10413 (2020).
- Natural questions: a benchmark for question answering research. Transactions of the Association for Computational Linguistics 7 (2019), 453–466.
- The bigscience roots corpus: A 1.6 tb composite multilingual dataset. Advances in Neural Information Processing Systems 35 (2022), 31809–31826.
- Latent retrieval for weakly supervised open domain question answering. arXiv preprint arXiv:1906.00300 (2019).
- The structure and performance of an open-domain question answering system. In Proceedings of the 38th annual meeting of the Association for Computational Linguistics. 563–570.
- Language models as knowledge bases? arXiv preprint arXiv:1909.01066 (2019).
- Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research 21, 1 (2020), 5485–5551.
- Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084 (2019).
- SLING: A framework for frame semantic parsing. arXiv preprint arXiv:1710.07032 (2017).
- How much knowledge can you pack into the parameters of a language model? arXiv preprint arXiv:2002.08910 (2020).
- Jointlk: Joint reasoning with language models and knowledge graphs for commonsense question answering. arXiv preprint arXiv:2112.02732 (2021).
- Memorisation versus generalisation in pre-trained language models. arXiv preprint arXiv:2105.00828 (2021).
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023).
- Ellen M Voorhees et al. 1999. The trec-8 question answering track report.. In Trec, Vol. 99. 77–82.
- Denny Vrandečić and Markus Krötzsch. 2014. Wikidata: a free collaborative knowledgebase. Commun. ACM 57, 10 (2014), 78–85.
- Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022).
- Greaselm: Graph reasoning enhanced language models for question answering. arXiv preprint arXiv:2201.08860 (2022).
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.