2017
DOI: 10.1007/978-3-319-59569-6_4
|View full text |Cite
|
Sign up to set email alerts
|

Automated Lexicon and Feature Construction Using Word Embedding and Clustering for Classification of ASD Diagnoses Using EHR

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
8
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3
2
2
1

Relationship

3
5

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 3 publications
0
8
0
Order By: Relevance
“…Text classification assigns a label or a class to a document. In the medical domain, this can be used to determine if a text refers to a positive instance of a particular medical condition [16,17]. Support vector machines (SVM) is a popular algorithm that generally performs well in a variety of tasks, including clinical applications [13,14,18].…”
Section: Nlp and Healthcare Applicationsmentioning
confidence: 99%
“…Text classification assigns a label or a class to a document. In the medical domain, this can be used to determine if a text refers to a positive instance of a particular medical condition [16,17]. Support vector machines (SVM) is a popular algorithm that generally performs well in a variety of tasks, including clinical applications [13,14,18].…”
Section: Nlp and Healthcare Applicationsmentioning
confidence: 99%
“…For example, Lee et al [ 16 ] proposed BioBERT by pretraining BERT on a large biomedical corpus of PubMed abstracts, and demonstrated that it outperforms BERT on three representative biomedical text mining tasks. Alsentzer et al [ 17 ] attempted to adapt pretrained models for clinical text by training BioBERT on clinical notes, resulting in the creation of BioClinical_BERT [ 18 ]. Gururangan et al [ 19 ] illustrated the usefulness of DAPT by continuing training of pretrained models on domain-specific data from four different domains (biomedical and computer science publications, news, and reviews).…”
Section: Introductionmentioning
confidence: 99%
“…A summary of the specific contributions of this paper are as follows: We compare the performances of six models pretrained with texts from different domains and sourcesBERT [ 14 ] and RoBERTa [ 15 ] (generic text), BERTweet [ 22 ], and Twitter BERT (social media text, specifically Twitter) [ 20 ], BioClinical_BERT [ 18 ] (clinical text), and BioBERT [ 16 ] (biomedical literature text)—on 22 social media-based health-related text classification tasks. We perform TSPT using the masked language model (MLM) [ 40 ], and assess its impact on classification performance compared to other pretraining strategies for three tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Alsentzer et al (2019) 17 attempted to adapt pretrained models for clinical text by training BioBERT on clinical notes, resulting in the creation of BioClinical_BERT. 18 Gururangan et al (2020) 19 illustrated the usefulness of DAPT by continuing training of pretrained models on domain-specific data from four different domains (biomedical and computer science publications, news, and reviews). However, some studies, including our own pilot, demonstrated that DAPT is not guaranteed to achieve SOTA results for health-related NLP tasks involving social media data.…”
Section: Introductionmentioning
confidence: 99%