Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2016
DOI: 10.18653/v1/p16-1096
|View full text |Cite
|
Sign up to set email alerts
|

Normalising Medical Concepts in Social Media Texts by Learning Semantic Representation

Abstract: Automatically recognising medical concepts mentioned in social media messages (e.g. tweets) enables several applications for enhancing health quality of people in a community, e.g. real-time monitoring of infectious diseases in population. However, the discrepancy between the type of language used in social media and medical ontologies poses a major challenge. Existing studies deal with this challenge by employing techniques, such as lexical term matching and statistical machine translation. In this work, we h… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
118
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 107 publications
(118 citation statements)
references
References 21 publications
0
118
0
Order By: Relevance
“…Huang and Lu (2015) survey the work done in the organization of biomedical NLP (BioNLP) challenge evaluations up to 2014. These tasks are devoted to the normalization of (1) genes from scientific articles (BioCreative I-III in 2005-2011; (2) (Limsopatham and Collier, 2016) 73.39 ----CNN (Limsopatham and Collier, 2016) 81.41 ----RNN (Limsopatham and Collier, 2016) 79.98 ----Attentional Char-CNN (Niu et al, 2018) 84.65 ----Hierarchical Char-CNN (Han et al, 2017) ----87.7 Ensemble (Sarker et al, 2018) - The 2017 SMM4H shared task (Sarker et al, 2018) was the first effort for the evaluation of NLP methods for the normalization of health-related text from social media on publicly released data. Recent advances in neural networks have been utilized for concept normalization: recent studies have employed convolutional neural networks (Limsopatham and Collier, 2016;Niu et al, 2018) and recurrent neural networks (Belousov et al, 2017;Han et al, 2017).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Huang and Lu (2015) survey the work done in the organization of biomedical NLP (BioNLP) challenge evaluations up to 2014. These tasks are devoted to the normalization of (1) genes from scientific articles (BioCreative I-III in 2005-2011; (2) (Limsopatham and Collier, 2016) 73.39 ----CNN (Limsopatham and Collier, 2016) 81.41 ----RNN (Limsopatham and Collier, 2016) 79.98 ----Attentional Char-CNN (Niu et al, 2018) 84.65 ----Hierarchical Char-CNN (Han et al, 2017) ----87.7 Ensemble (Sarker et al, 2018) - The 2017 SMM4H shared task (Sarker et al, 2018) was the first effort for the evaluation of NLP methods for the normalization of health-related text from social media on publicly released data. Recent advances in neural networks have been utilized for concept normalization: recent studies have employed convolutional neural networks (Limsopatham and Collier, 2016;Niu et al, 2018) and recurrent neural networks (Belousov et al, 2017;Han et al, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…Recent works go beyond string matching: these works have tried to view the problem of matching a one-or multi-word expression against a knowledge base as a supervised sequence labeling problem. Limsopatham and Collier (2016) utilized convolutional neural networks (CNNs) for phrase normalization in user reviews, while Tutubalina et al (2018), Han et al (2017), andBelousov et al (2017) applied recurrent neural networks (RNNs) to UGTs, achieving similar results. These works were among the first applications of deep learning techniques to medical concept normalization.…”
Section: Introductionmentioning
confidence: 99%
“…Without a semantic clustering scheme, where semantically similar sentences and questions are paired based on background knowledge, extracting an answer for a question from an entire Reddit post using mere lexical matches leads to inaccurate responses [18]. Our novel approach of dividing a post into sentences provides better results because 1) sentences in a post that don't convey relevant information are removed improving focus in response extraction and 2) better responses are generated for a question if answer is looked for in the most relevant part of a Reddit post.…”
Section: A Semantic Layermentioning
confidence: 99%
“…This task is commonly known as named entity normalization or entity linking and various approaches ranging from Levenshtein edit distances to recurrent neural networks have been suggested as the plausible solutions (Tiftikci et al, 2016;Limsopatham and Collier, 2016).…”
Section: Named Entity Categorizationmentioning
confidence: 99%