Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1259
|View full text |Cite
|
Sign up to set email alerts
|

PubMedQA: A Dataset for Biomedical Research Question Answering

Abstract: We introduce PubMedQA, a novel biomedical question answering (QA) dataset collected from PubMed abstracts. The task of Pub-MedQA is to answer research questions with yes/no/maybe (e.g.: Do preoperative statins reduce atrial fibrillation after coronary artery bypass grafting?) using the corresponding abstracts. PubMedQA has 1k expert-annotated, 61.2k unlabeled and 211.3k artificially generated QA instances. Each PubMedQA instance is composed of (1) a question which is either an existing research article title o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
171
0
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 209 publications
(173 citation statements)
references
References 29 publications
1
171
0
1
Order By: Relevance
“…Other than for the BERT experiments, all experiments were conducted using the Flair framework [5] which goes on top of Theano providing convenient way to experiment with different combinations of word embeddings. It provides off-the-shelf neural-based system supporting the entity extraction.…”
Section: Entity Extractionmentioning
confidence: 99%
See 1 more Smart Citation
“…Other than for the BERT experiments, all experiments were conducted using the Flair framework [5] which goes on top of Theano providing convenient way to experiment with different combinations of word embeddings. It provides off-the-shelf neural-based system supporting the entity extraction.…”
Section: Entity Extractionmentioning
confidence: 99%
“…We train a Long Short Term Memory (LSTM) network with a hidden state of 256 dimensions, learning rate 0.1, mini-batch size of 8, and is optimized with Adam. We train for 150 epochs, and the model that performs best on the [4] SciELO.org [5] https://github.com/zalandoresearch/flair validation set provided by the organizers of the competition during training is used to prevent overfitting. We were unable to conveniently experiment with BERT embeddings using the Flair framework but preferred the Google Cloud TensorFlow TPU set up for both training contextualized word embeddings and the downstream task fine-tuning and predictions as it works much faster [6] .…”
Section: Entity Extractionmentioning
confidence: 99%
“…It consists of 6,686 questions and explores the i2b2 dataset to generate QA pairs. Recently, [15] released the PubMedQA dataset, where they derive questions based on article titles and answer them with their respective abstracts. Fig.…”
Section: Biomedical Qamentioning
confidence: 99%
“…Extraction of biomedical entities from these narratives is relevant to a number of NLP tasks such as adverse drug and drug-drug interaction extraction [2], [3], biomedical concept normalization, knowledge base population [4], and question answering [5].…”
Section: Introductionmentioning
confidence: 99%