2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2019
DOI: 10.1109/bibm47256.2019.8983157
|View full text |Cite
|
Sign up to set email alerts
|

Active Learning with Deep Pre-trained Models for Sequence Tagging of Clinical and Biomedical Texts

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 23 publications
(16 citation statements)
references
References 14 publications
0
16
0
Order By: Relevance
“…Deep pre-trained models are evaluated in the AL setting for NER by Shelmanov et al (2019). However, they perform the evaluation only on the specific biomedical datasets and do not consider the Bayesian query strategies.…”
Section: Related Workmentioning
confidence: 99%
“…Deep pre-trained models are evaluated in the AL setting for NER by Shelmanov et al (2019). However, they perform the evaluation only on the specific biomedical datasets and do not consider the Bayesian query strategies.…”
Section: Related Workmentioning
confidence: 99%
“…However, this work focuses on a single task, and does not address the effect of small and imbalanced data. Additionally, Shelmanov et al (2019) and Liu et al (2020) focused on particular variants of BERT (BioBERT and BERT-CRF) and studied a single or two specific tasks, with a small collection of AL strategies. To the best of our knowledge, this work is the first to systematically explore advanced strategies like Core-Set (Sener and Savarese, 2017), Dropout (Gal and Ghahramani, 2016), Expected Gradient Length (Huang et al, 2016) and Discriminative Active Learning (Gissin and Shalev-Shwartz, 2019) for BERT, in various settings and a diversity of tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Finally, more “elaborate” methods have been used to extend the fine-tuning process and push benchmark performances further. Examples include the use of active learning on top of pre-trained BERT models [ 45 ], complementing the base model with a transfer learning framework [ 46 ], or a graph NN architecture [ 47 ].…”
Section: Resultsmentioning
confidence: 99%