2020
DOI: 10.1016/j.wpi.2020.101965
|View full text |Cite
|
Sign up to set email alerts
|

Patent classification by fine-tuning BERT language model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
75
0
3

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 143 publications
(79 citation statements)
references
References 8 publications
1
75
0
3
Order By: Relevance
“…Li et al [33] proposed a DeepPatent that combines the convolutional neural network (CNN) model with the word embedding model for classifying patents. Lee and Hsiang [34] fine-tuned a bidirectional encoder representations from transformers (BERT) model to classify patents and compared the fine-tuned model with the previously mentioned model, DeepPatent, and the result shows that the precision is 9% higher. Jun [35] proposed a method for technical integration and analysis using boosting (an ML algorithm that can be used to reduce bias in supervised learning) and ensemble learning.…”
Section: Patent Miningmentioning
confidence: 99%
“…Li et al [33] proposed a DeepPatent that combines the convolutional neural network (CNN) model with the word embedding model for classifying patents. Lee and Hsiang [34] fine-tuned a bidirectional encoder representations from transformers (BERT) model to classify patents and compared the fine-tuned model with the previously mentioned model, DeepPatent, and the result shows that the precision is 9% higher. Jun [35] proposed a method for technical integration and analysis using boosting (an ML algorithm that can be used to reduce bias in supervised learning) and ensemble learning.…”
Section: Patent Miningmentioning
confidence: 99%
“…Gururangan et al (2020) showed that continuing the training of a model with additional domain-adaptive and task-adaptive pretraining with unlabeled data leads to performance gains for both high-and low-resource settings for numerous English domains and tasks. This is also displayed in the number of domain-adapted language models (Alsentzer et al, 2019;Adhikari et al, 2019;Lee and Hsiang, 2020;Jain and Ganesamoorty, 2020, (i.a. )), most notably BioBERT that was pre-trained on biomedical PubMED articles and SciBERT (Beltagy et al, 2019) for scientific texts.…”
Section: Domain-specific Pre-trainingmentioning
confidence: 94%
“…BERT is pre-trained on two tasks: the masked language model and next sentence prediction (Devlin et al (2018)). Applying BERT means fine-tuning the pre-trained BERT to a task like patent classiffcation (Sun et al (2019) and Lee & Hsiang (2020a)). Patents are classified according to standards as international patent classiffcation (IPC, Note 3) and cooperative patent classiffcation (CPC, Note 4) by patent offices according to technical features characterizing the invention.…”
Section: Related Literaturementioning
confidence: 99%