Papers
Topics
Authors
Recent
Search
2000 character limit reached

Enhancing Language Models for Financial Relation Extraction with Named Entities and Part-of-Speech

Published 2 May 2024 in cs.CL, cs.IR, and cs.LG | (2405.06665v1)

Abstract: The Financial Relation Extraction (FinRE) task involves identifying the entities and their relation, given a piece of financial statement/text. To solve this FinRE problem, we propose a simple but effective strategy that improves the performance of pre-trained LLMs by augmenting them with Named Entity Recognition (NER) and Part-Of-Speech (POS), as well as different approaches to combine these information. Experiments on a financial relations dataset show promising results and highlights the benefits of incorporating NER and POS in existing models. Our dataset and codes are available at https://github.com/kwanhui/FinRelExtract.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Unsupervised cross-lingual representation learning at scale. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp.  8440–8451, 2020.
  2. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp.  4171–4186, 2019.
  3. A survey on personalized itinerary recommendation: From optimisation to deep learning. Applied Soft Computing, pp.  111200, 2024.
  4. Kpi-bert: A joint named entity recognition and relation extraction model for financial reports. In Proceedings of the 26th International Conference on Pattern Recognition (ICPR), pp.  606–612. IEEE, 2022.
  5. A french corpus and annotation schema for named entity recognition and relation extraction of financial news. In Proceedings of the Twelfth Language Resources and Evaluation Conference, pp.  2293–2299, 2020.
  6. Spanbert: Improving pre-training by representing and predicting spans. Transactions of the association for computational linguistics, 8:64–77, 2020.
  7. Refind: Relation extraction financial dataset. arXiv preprint arXiv:2305.18322, 2023.
  8. Albert: A lite bert for self-supervised learning of language representations. In Proceedings of the International Conference on Learning Representations, 2019.
  9. A transformer-based framework for poi-level social post geolocation. In European Conference on Information Retrieval, pp.  588–604. Springer, 2023.
  10. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
  11. Deep neural approaches to relation triplets extraction: A comprehensive survey. Cognitive Computation, 13:1215–1232, 2021.
  12. A simple approach to financial relation classification with pre-trained language models. In Proceedings of the 4th Workshop on Knowledge Discovery from Unstructured Data in Financial Services (KDF), 2023.
  13. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108, 2019.
  14. Finred: A dataset for relation extraction in financial domain. In Companion Proceedings of the Web Conference 2022, pp.  595–597, 2022.
  15. imetre: Incorporating markers of entity types for relation extraction. arXiv preprint arXiv:2307.00132, 2023.
  16. Deep neural network-based relation extraction: an overview. Neural Computing and Applications, pp.  1–21, 2022.
  17. Enriching pre-trained language model with entity information for relation classification. In Proceedings of the 28th ACM international conference on information and knowledge management, pp.  2361–2364, 2019.
  18. Finbert: A pretrained language model for financial communications. arXiv preprint arXiv:2006.08097, 2020.
  19. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems, 32, 2019.
  20. Relation classification via recurrent neural network. arXiv preprint arXiv:1508.01006, 2015.
  21. Position-aware attention and supervised data improve slot filling. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp.  35–45, 2017.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 0 likes about this paper.