2019
DOI: 10.1109/access.2019.2953990
|View full text |Cite
|
Sign up to set email alerts
|

Improving BERT-Based Text Classification With Auxiliary Sentence and Domain Knowledge

Abstract: General language model BERT pre-trained on cross-domain text corpus, BookCorpus and Wikipedia, achieves excellent performance on a couple of natural language processing tasks through the way of fine-tuning in the downstream tasks. But it still lacks of task-specific knowledge and domain-related knowledge for further improving the performance of BERT model and more detailed fine-tuning strategy analyses are necessary. To address these problem, a BERT-based text classification model BERT4TC is proposed via const… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
37
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 96 publications
(38 citation statements)
references
References 4 publications
1
37
0
Order By: Relevance
“…They have experienced that transfer learning models can perform better than other state-of-the-art methods in NLP. BERT is trained on BookCorpus, text corpus, and Wikipedia which can give overwhelming results in some areas of natural language processing but it still needs to be improved [ 32 ]. It somewhere lacks domain-related and task-related knowledge.…”
Section: Literature Reviewmentioning
confidence: 99%
“…They have experienced that transfer learning models can perform better than other state-of-the-art methods in NLP. BERT is trained on BookCorpus, text corpus, and Wikipedia which can give overwhelming results in some areas of natural language processing but it still needs to be improved [ 32 ]. It somewhere lacks domain-related and task-related knowledge.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Sun et al [41] make detailed experiments on BERT and suggest several techniques to improve the results on text classification task. Yu et al [50] propose a BERT-based model for text classification to utilize more task-specific knowledge and achieve better results on multi-classification task. Yeung [49] inserts legal domain vocabulary to BERT, reports no improvement and explains their findings by the high overlap between vocabularies.…”
Section: Evaluation and Limitations Of Plmsmentioning
confidence: 99%
“…Likewise, the precision represents the ratio of correctly predicted positive labels through the proposed approach to the total predicted positive labels. Consequently, the F-measure is calculated as the harmonic mean of precision and recall to show the cumulative effect of both measures as shown in equation (12).…”
Section: Accuracy =mentioning
confidence: 99%
“…The BERT model is designed to pre-train deep bidirectional representations of unlabeled text by co-conditioning both left and right context in all layers. A different contextual embeddings is produced by BERT according to the input sentence [12]. Nevertheless, BERT corrupts the input with masks, suffers a discrepancy between pre-training and fine-tuning, and ignores the interdependency between masked positions, thus leading to the loss of important information [13]- [15].…”
Section: Introductionmentioning
confidence: 99%