2023
DOI: 10.1186/s12859-023-05411-z
|View full text |Cite
|
Sign up to set email alerts
|

Leveraging pre-trained language models for mining microbiome-disease relationships

Nikitha Karkera,
Sathwik Acharya,
Sucheendra K. Palaniappan

Abstract: Background The growing recognition of the microbiome’s impact on human health and well-being has prompted extensive research into discovering the links between microbiome dysbiosis and disease (healthy) states. However, this valuable information is scattered in unstructured form within biomedical literature. The structured extraction and qualification of microbe-disease interactions are important. In parallel, recent advancements in deep-learning-based natural language processing algorithms hav… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(6 citation statements)
references
References 38 publications
0
6
0
Order By: Relevance
“…Beyond ChatGPT, other generative AI frameworks, such as DR.BENCH [62], were employed for clinical diagnostic reasoning tasks [62]. Moreover, various pre-trained LLMs can extract microbe-disease relationships from biomedical texts in zero-shot/few-shot contexts with high accuracy, with an average F1 score, precision, and recall greater than 80% [63]. In addition, ChatGPT was the best LLM when predicting high acuity cases than predicting low acuity cases according to emergency severity index (ESI), with a sensitivity of 76.2%, a specificity of 93.1%, compared to the overall sensitivity of 57.1%, specificity of 34.5% [64].…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Beyond ChatGPT, other generative AI frameworks, such as DR.BENCH [62], were employed for clinical diagnostic reasoning tasks [62]. Moreover, various pre-trained LLMs can extract microbe-disease relationships from biomedical texts in zero-shot/few-shot contexts with high accuracy, with an average F1 score, precision, and recall greater than 80% [63]. In addition, ChatGPT was the best LLM when predicting high acuity cases than predicting low acuity cases according to emergency severity index (ESI), with a sensitivity of 76.2%, a specificity of 93.1%, compared to the overall sensitivity of 57.1%, specificity of 34.5% [64].…”
Section: Resultsmentioning
confidence: 99%
“…Since recent LLMs are trained based on human-generated texts from the Internet, they also tend to provide biased answers [4]. Besides, algorithms may reinforce current health disparities and inequities [63]. Indeed, outputs from ChatGPT have been shown to be biased in terms of gender, race, and religion [4].…”
Section: Resultsmentioning
confidence: 99%
“…The data extraction ability of LLMs can also be enhanced through fine-tuning. This includes pre-trained LLMs in the generative and discriminative setting, i.e., they can generate responses to a question when prompted in a given context and classify input data into predefined labels [59]. Domain-specific LLMs, such as BioMedLM and BioGPT, are trained with data from the biomedical literature on PubMed and can be fine-tuned with gold standard oncology corpora [60,61].…”
Section: Oncology Researchmentioning
confidence: 99%
“…Validation processes have emphasized the importance of contextual relevance and applicability of the knowledge to specific tasks or domains [58,59,60,61]. Peer review mechanisms and consensus algorithms have been utilized to enhance the trustworthiness of collaboratively curated knowledge bases [62,61,63]. The application of cryptographic techniques for knowledge verification has provided a means to secure the integrity of information in distributed systems [64,65].…”
Section: Knowledge Verification and Validationmentioning
confidence: 99%