Papers
Topics
Authors
Recent
Search
2000 character limit reached

JuriBERT: A Masked-Language Model Adaptation for French Legal Text

Published 4 Oct 2021 in cs.CL | (2110.01485v2)

Abstract: LLMs have proven to be very useful when adapted to specific domains. Nonetheless, little research has been done on the adaptation of domain-specific BERT models in the French language. In this paper, we focus on creating a LLM adapted to French legal text with the goal of helping law professionals. We conclude that some specific tasks do not benefit from generic LLMs pre-trained on large amounts of data. We explore the use of smaller architectures in domain-specific sub-languages and their benefits for French legal text. We prove that domain-specific pre-trained models can perform better than their equivalent generalised ones in the legal domain. Finally, we release JuriBERT, a new set of BERT models adapted to the French legal domain.

Citations (29)

Summary

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.