2021
DOI: 10.1007/s10506-021-09301-8
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised law article mining based on deep pre-trained language representation models with application to the Italian civil code

Abstract: Modeling law search and retrieval as prediction problems has recently emerged as a predominant approach in law intelligence. Focusing on the law article retrieval task, we present a deep learning framework named LamBERTa, which is designed for civil-law codes, and specifically trained on the Italian civil code. To our knowledge, this is the first study proposing an advanced approach to law article prediction for the Italian legal system based on a BERT (Bidirectional Encoder Representations from Transformers) … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 28 publications
(9 citation statements)
references
References 47 publications
0
9
0
Order By: Relevance
“…In [ 14 ], an approach for classifying articles of law for the Italian legal system based on a learning framework based on Bidirectional Encoder Representations from Transformers (BERT) was presented. Other examples are presented in [ 15 ], which performed experiments to detect through the analysis of judgments of the European Court of Human Rights whether the case was judged as a violation of someone’s right or not, and [ 16 ] used Glove word vectors generated for the Portuguese Language and Convolutional Neural Network (CNN) to perform the classification of types of documents in judicial proceedings of the Court of Justice of Minas Gerais (TJMG).…”
Section: State Of the Artmentioning
confidence: 99%
“…In [ 14 ], an approach for classifying articles of law for the Italian legal system based on a learning framework based on Bidirectional Encoder Representations from Transformers (BERT) was presented. Other examples are presented in [ 15 ], which performed experiments to detect through the analysis of judgments of the European Court of Human Rights whether the case was judged as a violation of someone’s right or not, and [ 16 ] used Glove word vectors generated for the Portuguese Language and Convolutional Neural Network (CNN) to perform the classification of types of documents in judicial proceedings of the Court of Justice of Minas Gerais (TJMG).…”
Section: State Of the Artmentioning
confidence: 99%
“…Clearly, advances in NLP are needed. Research on this is already being done, (Branting et al 2021), 26 including the first legal applications of currently fashionable language models like BERT (Chalkidis et al 2020;Zheng et al 2021b;Tagarelli and Simeri 2021). Moreover, given the arguably limited generality of the symbolic legal-reasoning models underlying SMILE-IBP and VJAP, further work on developing such models is also needed.…”
Section: Discussionmentioning
confidence: 99%
“…Moreover, more recent studies follow semi-supervised techniques to extract relevant concepts from judicial texts (Thomas & Sangeetha, 2021). Other works apply Naive Bayes and other supervised methods (Medvedeva et al, 2020;Hausladen et al, 2020) and transformer-based language models and other Deep Neural Network (dnn) architectures, such as that by Tagarelli & Simeri (2021), a fine-tuned version of Italian bert 1 (Devlin et al, 2019) for law article classification; and the Convolutional Neural Network (cnn) for multi-label Chinese legal text classification by Qiu et al (2020). Unfortunately, most pre-trained language models' embeddings are not adapted to the legal domain (Beltagy et al, 2019), so they tend to underperform compared to traditional ml algorithms like rf, as noted by Chen et al (2022).…”
Section: Related Workmentioning
confidence: 99%