2022
DOI: 10.1093/bib/bbac409
|View full text |Cite
|
Sign up to set email alerts
|

BioGPT: generative pre-trained transformer for biomedical text generation and mining

Abstract: Pre-trained language models have attracted increasing attention in the biomedical domain, inspired by their great success in the general natural language domain. Among the two main branches of pre-trained language models in the general language domain, i.e. BERT (and its variants) and GPT (and its variants), the first one has been extensively studied in the biomedical domain, such as BioBERT and PubMedBERT. While they have achieved great success on a variety of discriminative downstream biomedical tasks, the l… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
159
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 393 publications
(242 citation statements)
references
References 46 publications
1
159
0
1
Order By: Relevance
“… 7 BioGPT, a LLM developed by Microsoft and trained on biomedical data, has achieved human parity in certain tasks of biomedical text generation and mining. 8 Similar outcomes have been reported with other state-of-the-art LLMs for standardized tests in law and education. This capability, paired with the conversational ease of human-machine interaction and the sheer limitless scope of prompt complexity and variability which LLMs can handle is an impressive accomplishment.…”
Section: A Linguistic Stunt Takes Over the Worldsupporting
confidence: 73%
“… 7 BioGPT, a LLM developed by Microsoft and trained on biomedical data, has achieved human parity in certain tasks of biomedical text generation and mining. 8 Similar outcomes have been reported with other state-of-the-art LLMs for standardized tests in law and education. This capability, paired with the conversational ease of human-machine interaction and the sheer limitless scope of prompt complexity and variability which LLMs can handle is an impressive accomplishment.…”
Section: A Linguistic Stunt Takes Over the Worldsupporting
confidence: 73%
“…[34] Another example is BioGPT, a GPT-like model trained on PubMed articles. [35] In addition, UpToDate is an important source of CDS content for alert development and should also be considered for adding to the language model. Second, based on the Reinforcement Learning from Human Feedback (RLHF) framework, researchers could train a model specifically for this task on the improved language model.…”
Section: Discussionmentioning
confidence: 99%
“…The code is available, though the reader should know that topic modeling is not the most sophisticated language model available. Future iterations of language models about deep brain stimulation could use a more advanced approach, such as a generative pre-trained transformer model validated by experts in the field [ 50 ].…”
Section: Discussionmentioning
confidence: 99%